KVM: VMX: conditionally disable 2M pages
authorMarcelo Tosatti <mtosatti@redhat.com>
Thu, 11 Jun 2009 15:07:44 +0000 (12:07 -0300)
committerAvi Kivity <avi@redhat.com>
Thu, 10 Sep 2009 05:32:56 +0000 (08:32 +0300)
Disable usage of 2M pages if VMX_EPT_2MB_PAGE_BIT (bit 16) is clear
in MSR_IA32_VMX_EPT_VPID_CAP and EPT is enabled.

[avi: s/largepages_disabled/largepages_enabled/ to avoid negative logic]

Signed-off-by: Marcelo Tosatti <mtosatti@redhat.com>
Signed-off-by: Avi Kivity <avi@redhat.com>
arch/x86/kvm/vmx.c
include/linux/kvm_host.h
virt/kvm/kvm_main.c

index 94c07ad..fc8d49c 100644 (file)
@@ -1381,6 +1381,9 @@ static __init int hardware_setup(void)
        if (!cpu_has_vmx_tpr_shadow())
                kvm_x86_ops->update_cr8_intercept = NULL;
 
+       if (enable_ept && !cpu_has_vmx_ept_2m_page())
+               kvm_disable_largepages();
+
        return alloc_kvm_area();
 }
 
index c6e4d02..6988858 100644 (file)
@@ -224,6 +224,7 @@ int kvm_arch_set_memory_region(struct kvm *kvm,
                                struct kvm_userspace_memory_region *mem,
                                struct kvm_memory_slot old,
                                int user_alloc);
+void kvm_disable_largepages(void);
 void kvm_arch_flush_shadow(struct kvm *kvm);
 gfn_t unalias_gfn(struct kvm *kvm, gfn_t gfn);
 struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn);
index 777fe53..48d5e69 100644 (file)
@@ -85,6 +85,8 @@ static long kvm_vcpu_ioctl(struct file *file, unsigned int ioctl,
 
 static bool kvm_rebooting;
 
+static bool largepages_enabled = true;
+
 #ifdef KVM_CAP_DEVICE_ASSIGNMENT
 static struct kvm_assigned_dev_kernel *kvm_find_assigned_dev(struct list_head *head,
                                                      int assigned_dev_id)
@@ -1174,9 +1176,11 @@ int __kvm_set_memory_region(struct kvm *kvm,
                ugfn = new.userspace_addr >> PAGE_SHIFT;
                /*
                 * If the gfn and userspace address are not aligned wrt each
-                * other, disable large page support for this slot
+                * other, or if explicitly asked to, disable large page
+                * support for this slot
                 */
-               if ((base_gfn ^ ugfn) & (KVM_PAGES_PER_HPAGE - 1))
+               if ((base_gfn ^ ugfn) & (KVM_PAGES_PER_HPAGE - 1) ||
+                   !largepages_enabled)
                        for (i = 0; i < largepages; ++i)
                                new.lpage_info[i].write_count = 1;
        }
@@ -1291,6 +1295,12 @@ out:
        return r;
 }
 
+void kvm_disable_largepages(void)
+{
+       largepages_enabled = false;
+}
+EXPORT_SYMBOL_GPL(kvm_disable_largepages);
+
 int is_error_page(struct page *page)
 {
        return page == bad_page;