[PATCH] kvm: NULL noise removal
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
parent
5b71bddb78
commit
8b6d44c7bd
|
@ -62,7 +62,7 @@ static struct kvm_stats_debugfs_item {
|
||||||
{ "halt_exits", &kvm_stat.halt_exits },
|
{ "halt_exits", &kvm_stat.halt_exits },
|
||||||
{ "request_irq", &kvm_stat.request_irq_exits },
|
{ "request_irq", &kvm_stat.request_irq_exits },
|
||||||
{ "irq_exits", &kvm_stat.irq_exits },
|
{ "irq_exits", &kvm_stat.irq_exits },
|
||||||
{ 0, 0 }
|
{ NULL, NULL }
|
||||||
};
|
};
|
||||||
|
|
||||||
static struct dentry *debugfs_dir;
|
static struct dentry *debugfs_dir;
|
||||||
|
@ -205,7 +205,7 @@ static struct kvm_vcpu *vcpu_load(struct kvm *kvm, int vcpu_slot)
|
||||||
mutex_lock(&vcpu->mutex);
|
mutex_lock(&vcpu->mutex);
|
||||||
if (unlikely(!vcpu->vmcs)) {
|
if (unlikely(!vcpu->vmcs)) {
|
||||||
mutex_unlock(&vcpu->mutex);
|
mutex_unlock(&vcpu->mutex);
|
||||||
return 0;
|
return NULL;
|
||||||
}
|
}
|
||||||
return kvm_arch_ops->vcpu_load(vcpu);
|
return kvm_arch_ops->vcpu_load(vcpu);
|
||||||
}
|
}
|
||||||
|
@ -257,9 +257,9 @@ static void kvm_free_physmem_slot(struct kvm_memory_slot *free,
|
||||||
if (!dont || free->dirty_bitmap != dont->dirty_bitmap)
|
if (!dont || free->dirty_bitmap != dont->dirty_bitmap)
|
||||||
vfree(free->dirty_bitmap);
|
vfree(free->dirty_bitmap);
|
||||||
|
|
||||||
free->phys_mem = 0;
|
free->phys_mem = NULL;
|
||||||
free->npages = 0;
|
free->npages = 0;
|
||||||
free->dirty_bitmap = 0;
|
free->dirty_bitmap = NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void kvm_free_physmem(struct kvm *kvm)
|
static void kvm_free_physmem(struct kvm *kvm)
|
||||||
|
@ -267,7 +267,7 @@ static void kvm_free_physmem(struct kvm *kvm)
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
for (i = 0; i < kvm->nmemslots; ++i)
|
for (i = 0; i < kvm->nmemslots; ++i)
|
||||||
kvm_free_physmem_slot(&kvm->memslots[i], 0);
|
kvm_free_physmem_slot(&kvm->memslots[i], NULL);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void kvm_free_vcpu(struct kvm_vcpu *vcpu)
|
static void kvm_free_vcpu(struct kvm_vcpu *vcpu)
|
||||||
|
@ -640,11 +640,11 @@ raced:
|
||||||
|
|
||||||
/* Deallocate if slot is being removed */
|
/* Deallocate if slot is being removed */
|
||||||
if (!npages)
|
if (!npages)
|
||||||
new.phys_mem = 0;
|
new.phys_mem = NULL;
|
||||||
|
|
||||||
/* Free page dirty bitmap if unneeded */
|
/* Free page dirty bitmap if unneeded */
|
||||||
if (!(new.flags & KVM_MEM_LOG_DIRTY_PAGES))
|
if (!(new.flags & KVM_MEM_LOG_DIRTY_PAGES))
|
||||||
new.dirty_bitmap = 0;
|
new.dirty_bitmap = NULL;
|
||||||
|
|
||||||
r = -ENOMEM;
|
r = -ENOMEM;
|
||||||
|
|
||||||
|
@ -799,14 +799,14 @@ struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn)
|
||||||
&& gfn < memslot->base_gfn + memslot->npages)
|
&& gfn < memslot->base_gfn + memslot->npages)
|
||||||
return memslot;
|
return memslot;
|
||||||
}
|
}
|
||||||
return 0;
|
return NULL;
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(gfn_to_memslot);
|
EXPORT_SYMBOL_GPL(gfn_to_memslot);
|
||||||
|
|
||||||
void mark_page_dirty(struct kvm *kvm, gfn_t gfn)
|
void mark_page_dirty(struct kvm *kvm, gfn_t gfn)
|
||||||
{
|
{
|
||||||
int i;
|
int i;
|
||||||
struct kvm_memory_slot *memslot = 0;
|
struct kvm_memory_slot *memslot = NULL;
|
||||||
unsigned long rel_gfn;
|
unsigned long rel_gfn;
|
||||||
|
|
||||||
for (i = 0; i < kvm->nmemslots; ++i) {
|
for (i = 0; i < kvm->nmemslots; ++i) {
|
||||||
|
@ -2015,7 +2015,7 @@ static int kvm_reboot(struct notifier_block *notifier, unsigned long val,
|
||||||
* in vmx root mode.
|
* in vmx root mode.
|
||||||
*/
|
*/
|
||||||
printk(KERN_INFO "kvm: exiting hardware virtualization\n");
|
printk(KERN_INFO "kvm: exiting hardware virtualization\n");
|
||||||
on_each_cpu(kvm_arch_ops->hardware_disable, 0, 0, 1);
|
on_each_cpu(kvm_arch_ops->hardware_disable, NULL, 0, 1);
|
||||||
}
|
}
|
||||||
return NOTIFY_OK;
|
return NOTIFY_OK;
|
||||||
}
|
}
|
||||||
|
@ -2029,7 +2029,7 @@ static __init void kvm_init_debug(void)
|
||||||
{
|
{
|
||||||
struct kvm_stats_debugfs_item *p;
|
struct kvm_stats_debugfs_item *p;
|
||||||
|
|
||||||
debugfs_dir = debugfs_create_dir("kvm", 0);
|
debugfs_dir = debugfs_create_dir("kvm", NULL);
|
||||||
for (p = debugfs_entries; p->name; ++p)
|
for (p = debugfs_entries; p->name; ++p)
|
||||||
p->dentry = debugfs_create_u32(p->name, 0444, debugfs_dir,
|
p->dentry = debugfs_create_u32(p->name, 0444, debugfs_dir,
|
||||||
p->data);
|
p->data);
|
||||||
|
@ -2070,7 +2070,7 @@ int kvm_init_arch(struct kvm_arch_ops *ops, struct module *module)
|
||||||
if (r < 0)
|
if (r < 0)
|
||||||
return r;
|
return r;
|
||||||
|
|
||||||
on_each_cpu(kvm_arch_ops->hardware_enable, 0, 0, 1);
|
on_each_cpu(kvm_arch_ops->hardware_enable, NULL, 0, 1);
|
||||||
register_reboot_notifier(&kvm_reboot_notifier);
|
register_reboot_notifier(&kvm_reboot_notifier);
|
||||||
|
|
||||||
kvm_chardev_ops.owner = module;
|
kvm_chardev_ops.owner = module;
|
||||||
|
@ -2085,7 +2085,7 @@ int kvm_init_arch(struct kvm_arch_ops *ops, struct module *module)
|
||||||
|
|
||||||
out_free:
|
out_free:
|
||||||
unregister_reboot_notifier(&kvm_reboot_notifier);
|
unregister_reboot_notifier(&kvm_reboot_notifier);
|
||||||
on_each_cpu(kvm_arch_ops->hardware_disable, 0, 0, 1);
|
on_each_cpu(kvm_arch_ops->hardware_disable, NULL, 0, 1);
|
||||||
kvm_arch_ops->hardware_unsetup();
|
kvm_arch_ops->hardware_unsetup();
|
||||||
return r;
|
return r;
|
||||||
}
|
}
|
||||||
|
@ -2095,7 +2095,7 @@ void kvm_exit_arch(void)
|
||||||
misc_deregister(&kvm_dev);
|
misc_deregister(&kvm_dev);
|
||||||
|
|
||||||
unregister_reboot_notifier(&kvm_reboot_notifier);
|
unregister_reboot_notifier(&kvm_reboot_notifier);
|
||||||
on_each_cpu(kvm_arch_ops->hardware_disable, 0, 0, 1);
|
on_each_cpu(kvm_arch_ops->hardware_disable, NULL, 0, 1);
|
||||||
kvm_arch_ops->hardware_unsetup();
|
kvm_arch_ops->hardware_unsetup();
|
||||||
kvm_arch_ops = NULL;
|
kvm_arch_ops = NULL;
|
||||||
}
|
}
|
||||||
|
|
|
@ -274,7 +274,7 @@ static void svm_hardware_disable(void *garbage)
|
||||||
wrmsrl(MSR_VM_HSAVE_PA, 0);
|
wrmsrl(MSR_VM_HSAVE_PA, 0);
|
||||||
rdmsrl(MSR_EFER, efer);
|
rdmsrl(MSR_EFER, efer);
|
||||||
wrmsrl(MSR_EFER, efer & ~MSR_EFER_SVME_MASK);
|
wrmsrl(MSR_EFER, efer & ~MSR_EFER_SVME_MASK);
|
||||||
per_cpu(svm_data, raw_smp_processor_id()) = 0;
|
per_cpu(svm_data, raw_smp_processor_id()) = NULL;
|
||||||
__free_page(svm_data->save_area);
|
__free_page(svm_data->save_area);
|
||||||
kfree(svm_data);
|
kfree(svm_data);
|
||||||
}
|
}
|
||||||
|
@ -642,7 +642,7 @@ static struct vmcb_seg *svm_seg(struct kvm_vcpu *vcpu, int seg)
|
||||||
case VCPU_SREG_LDTR: return &save->ldtr;
|
case VCPU_SREG_LDTR: return &save->ldtr;
|
||||||
}
|
}
|
||||||
BUG();
|
BUG();
|
||||||
return 0;
|
return NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
static u64 svm_get_segment_base(struct kvm_vcpu *vcpu, int seg)
|
static u64 svm_get_segment_base(struct kvm_vcpu *vcpu, int seg)
|
||||||
|
@ -934,7 +934,7 @@ static int io_get_override(struct kvm_vcpu *vcpu,
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
*addr_override = 0;
|
*addr_override = 0;
|
||||||
*seg = 0;
|
*seg = NULL;
|
||||||
for (i = 0; i < ins_length; i++)
|
for (i = 0; i < ins_length; i++)
|
||||||
switch (inst[i]) {
|
switch (inst[i]) {
|
||||||
case 0xf0:
|
case 0xf0:
|
||||||
|
@ -1087,7 +1087,7 @@ static int cpuid_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
|
||||||
|
|
||||||
static int emulate_on_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
|
static int emulate_on_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
|
||||||
{
|
{
|
||||||
if (emulate_instruction(vcpu, 0, 0, 0) != EMULATE_DONE)
|
if (emulate_instruction(vcpu, NULL, 0, 0) != EMULATE_DONE)
|
||||||
printk(KERN_ERR "%s: failed\n", __FUNCTION__);
|
printk(KERN_ERR "%s: failed\n", __FUNCTION__);
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
|
@ -98,7 +98,7 @@ static struct vmx_msr_entry *find_msr_entry(struct kvm_vcpu *vcpu, u32 msr)
|
||||||
for (i = 0; i < vcpu->nmsrs; ++i)
|
for (i = 0; i < vcpu->nmsrs; ++i)
|
||||||
if (vcpu->guest_msrs[i].index == msr)
|
if (vcpu->guest_msrs[i].index == msr)
|
||||||
return &vcpu->guest_msrs[i];
|
return &vcpu->guest_msrs[i];
|
||||||
return 0;
|
return NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void vmcs_clear(struct vmcs *vmcs)
|
static void vmcs_clear(struct vmcs *vmcs)
|
||||||
|
|
Loading…
Reference in New Issue