summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorRobert Mustacchi <rm@joyent.com>2011-06-02 11:49:59 -0700
committerRobert Mustacchi <rm@joyent.com>2011-06-02 14:08:31 -0700
commit6db183dc245e3177d3c41d06ecf299b237c70734 (patch)
treedc2c54ffdb83ac5e56c1bb48a315b3480f73fcfd
parent5aebcacd8bec7e3f67f71732d24871513e789588 (diff)
downloadillumos-kvm-6db183dc245e3177d3c41d06ecf299b237c70734.tar.gz
HVM-264 vmx code should live in its own file
HVM-265 kvm_ioapic should be make clean
-rw-r--r--Makefile12
-rw-r--r--kvm.c3685
-rw-r--r--kvm.h109
-rw-r--r--kvm_ioapic.c2
-rw-r--r--kvm_subr.c16
-rw-r--r--kvm_vmx.c4741
-rw-r--r--kvm_x86.c842
-rw-r--r--kvm_x86host.h9
8 files changed, 4794 insertions, 4622 deletions
diff --git a/Makefile b/Makefile
index ebf838c..5f9187b 100644
--- a/Makefile
+++ b/Makefile
@@ -16,7 +16,7 @@ CSTYLE=$(KERNEL_SOURCE)/usr/src/tools/scripts/cstyle
all: kvm kvm.so
-kvm: kvm.c kvm_x86.c kvm_emulate.c kvm.h kvm_x86host.h msr.h bitops.h kvm_subr.c kvm_irq.c kvm_i8254.c kvm_lapic.c kvm_mmu.c kvm_iodev.c kvm_ioapic.c
+kvm: kvm.c kvm_x86.c kvm_emulate.c kvm.h kvm_x86host.h msr.h bitops.h kvm_subr.c kvm_irq.c kvm_i8254.c kvm_lapic.c kvm_mmu.c kvm_iodev.c kvm_ioapic.c kvm_vmx.c
$(CC) $(CFLAGS) $(INCLUDEDIR) kvm.c
$(CC) $(CFLAGS) $(INCLUDEDIR) kvm_x86.c
$(CC) $(CFLAGS) $(INCLUDEDIR) kvm_emulate.c
@@ -27,6 +27,7 @@ kvm: kvm.c kvm_x86.c kvm_emulate.c kvm.h kvm_x86host.h msr.h bitops.h kvm_subr.c
$(CC) $(CFLAGS) $(INCLUDEDIR) kvm_mmu.c
$(CC) $(CFLAGS) $(INCLUDEDIR) kvm_iodev.c
$(CC) $(CFLAGS) $(INCLUDEDIR) kvm_ioapic.c
+ $(CC) $(CFLAGS) $(INCLUDEDIR) kvm_vmx.c
$(CTFCONVERT) -i -L VERSION kvm.o
$(CTFCONVERT) -i -L VERSION kvm_x86.o
$(CTFCONVERT) -i -L VERSION kvm_emulate.o
@@ -37,8 +38,9 @@ kvm: kvm.c kvm_x86.c kvm_emulate.c kvm.h kvm_x86host.h msr.h bitops.h kvm_subr.c
$(CTFCONVERT) -i -L VERSION kvm_mmu.o
$(CTFCONVERT) -i -L VERSION kvm_iodev.o
$(CTFCONVERT) -i -L VERSION kvm_ioapic.o
- $(LD) -r -o kvm kvm.o kvm_x86.o kvm_emulate.o kvm_subr.o kvm_irq.o kvm_i8254.o kvm_lapic.o kvm_mmu.o kvm_iodev.o kvm_ioapic.o
- $(CTFMERGE) -L VERSION -o kvm kvm.o kvm_x86.o kvm_emulate.o kvm_subr.o kvm_irq.o kvm_i8254.o kvm_lapic.o kvm_mmu.o kvm_iodev.o kvm_ioapic.o
+ $(CTFCONVERT) -i -L VERSION kvm_vmx.o
+ $(LD) -r -o kvm kvm.o kvm_x86.o kvm_emulate.o kvm_subr.o kvm_irq.o kvm_i8254.o kvm_lapic.o kvm_mmu.o kvm_iodev.o kvm_ioapic.o kvm_vmx.o
+ $(CTFMERGE) -L VERSION -o kvm kvm.o kvm_x86.o kvm_emulate.o kvm_subr.o kvm_irq.o kvm_i8254.o kvm_lapic.o kvm_mmu.o kvm_iodev.o kvm_ioapic.o kvm_vmx.o
kvm.so: kvm_mdb.c
gcc -m64 -shared \
@@ -51,8 +53,8 @@ install: kvm
@pfexec cp kvm.conf /usr/kernel/drv
check:
- @$(CSTYLE) kvm.c kvm_mdb.c kvm_emulate.c kvm_x86.c kvm_irq.c kvm_lapic.c kvm_i8254.c kvm_mmu.c kvm_iodev.c kvm_subr.c
- @./tools/xxxcheck kvm_x86.c kvm.c kvm_irq.c kvm_lapic.c kvm_i8254.c kvm_mmu.c kvm_iodev.c
+ @$(CSTYLE) kvm.c kvm_mdb.c kvm_emulate.c kvm_x86.c kvm_irq.c kvm_lapic.c kvm_i8254.c kvm_mmu.c kvm_iodev.c kvm_subr.c kvm_ioapic.c kvm_vmx.c
+ @./tools/xxxcheck kvm_x86.c kvm.c kvm_irq.c kvm_lapic.c kvm_i8254.c kvm_mmu.c kvm_iodev.c kvm_ioapic.c kvm_vmx.c
load: install
@echo "==> Loading kvm module"
diff --git a/kvm.c b/kvm.c
index 498966e..d4afa7b 100644
--- a/kvm.c
+++ b/kvm.c
@@ -84,7 +84,6 @@ static int kvm_usage_count;
static list_t vm_list;
kmutex_t kvm_lock;
kmem_cache_t *kvm_cache;
-struct vmx_capability vmx_capability;
/*
* Driver forward declarations
@@ -164,9 +163,8 @@ void vmx_vcpu_load(struct kvm_vcpu *vcpu, int cpu);
void vmx_vcpu_put(struct kvm_vcpu *vcpu);
extern void vmx_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0);
extern void vmx_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4);
-static int vmx_set_tss_addr(struct kvm *kvmp, caddr_t addr);
+extern int vmx_set_tss_addr(struct kvm *kvmp, caddr_t addr);
static int vmx_hardware_setup(void);
-extern int vmx_hardware_enable(void *garbage);
extern unsigned long vmx_get_rflags(struct kvm_vcpu *vcpu);
void vmcs_writel(unsigned long field, unsigned long value);
unsigned long vmcs_readl(unsigned long field);
@@ -229,18 +227,14 @@ static int kvm_avlmmucmp(const void *, const void *);
int get_ept_level(void);
static void vmx_cache_reg(struct kvm_vcpu *vcpu, enum kvm_reg reg);
-
-struct vcpu_vmx *
-to_vmx(struct kvm_vcpu *vcpu)
-{
-#ifdef XXX_KVM_DOESNTCOMPILE
- return (container_of(vcpu, struct vcpu_vmx, vcpu));
-#else
- /* assumes vcpu is first field in vcpu_vmx */
- /* because gcc with kernel flags complains about container_of */
- return ((struct vcpu_vmx *)vcpu);
-#endif
-}
+/*
+ * XXX
+ */
+extern int enable_vpid;
+extern struct kvm_x86_ops vmx_x86_ops;
+extern int vmx_init(void);
+extern uint32_t bit(int);
+extern struct kvm_shared_msrs **shared_msrs;
/*
* Find the first cleared bit in a memory region.
@@ -268,209 +262,12 @@ found:
return (result + ffz(tmp));
}
-static inline void
-__invvpid(int ext, uint16_t vpid, gva_t gva)
-{
- struct {
- uint64_t vpid:16;
- uint64_t rsvd:48;
- uint64_t gva;
- } operand = { vpid, 0, gva };
-
- /* BEGIN CSTYLED */
-#ifdef XXX_KVM_DOESNTCOMPILE
- __asm__ volatile (__ex(ASM_VMX_INVVPID)
-#else
- __asm__ volatile (ASM_VMX_INVVPID
-#endif /*XXX*/
- /* CF==1 or ZF==1 --> rc = -1 */
- "; ja 1f ; ud2 ; 1:"
- : : "a"(&operand), "c"(ext) : "cc", "memory");
- /* END CSTYLED */
-}
-
-inline void
-vpid_sync_vcpu_all(struct vcpu_vmx *vmx)
-{
- if (vmx->vpid == 0)
- return;
-
- __invvpid(VMX_VPID_EXTENT_SINGLE_CONTEXT, vmx->vpid, 0);
-}
-
-static inline void
-__invept(int ext, uint64_t eptp, gpa_t gpa)
-{
- struct {
- uint64_t eptp, gpa;
- } operand = {eptp, gpa};
-
- /* BEGIN CSTYLED */
- __asm__ volatile (ASM_VMX_INVEPT
- /* CF==1 or ZF==1 --> rc = -1 */
- "; ja 1f ; ud2 ; 1:\n"
- : : "a" (&operand), "c" (ext) : "cc", "memory");
- /* END CSTYLED */
-}
-
-static inline int
-cpu_has_vmx_invept_context(void)
-{
- return (!!(vmx_capability.ept & VMX_EPT_EXTENT_CONTEXT_BIT));
-}
-
-static inline int
-cpu_has_vmx_invept_global(void)
-{
- return (!!(vmx_capability.ept & VMX_EPT_EXTENT_GLOBAL_BIT));
-}
-
-inline void
-ept_sync_global(void)
-{
- if (cpu_has_vmx_invept_global())
- __invept(VMX_EPT_EXTENT_GLOBAL, 0, 0);
-}
-
-int enable_ept = 1; /* XXX */
-
-static inline void
-ept_sync_context(uint64_t eptp)
-{
- if (enable_ept) {
- if (cpu_has_vmx_invept_context())
- __invept(VMX_EPT_EXTENT_CONTEXT, eptp, 0);
- else
- ept_sync_global();
- }
-}
-
-static uint64_t
-construct_eptp(unsigned long root_hpa)
-{
- uint64_t eptp;
-
- /* TODO write the value reading from MSR */
- eptp = VMX_EPT_DEFAULT_MT |
- VMX_EPT_DEFAULT_GAW << VMX_EPT_GAW_EPTP_SHIFT;
- eptp |= (root_hpa & PAGEMASK);
-
- return (eptp);
-}
-
-
-static void vmx_flush_tlb(struct kvm_vcpu *vcpu)
-{
- vpid_sync_vcpu_all(to_vmx(vcpu));
- if (enable_ept)
- ept_sync_context(construct_eptp(vcpu->arch.mmu.root_hpa));
-}
-
-void
-vmcs_write64(unsigned long field, uint64_t value)
-{
- vmcs_writel(field, value);
-#ifndef CONFIG_X86_64
- /*CSTYLED*/
- __asm__ volatile ("");
- vmcs_writel(field + 1, value >> 32);
-#endif
-}
inline int is_pae(struct kvm_vcpu *vcpu);
extern int is_paging(struct kvm_vcpu *);
extern int is_long_mode(struct kvm_vcpu *);
-static void
-ept_load_pdptrs(struct kvm_vcpu *vcpu)
-{
- if (!test_bit(VCPU_EXREG_PDPTR,
- (unsigned long *)&vcpu->arch.regs_dirty))
- return;
-
- if (is_paging(vcpu) && is_pae(vcpu) && !is_long_mode(vcpu)) {
- vmcs_write64(GUEST_PDPTR0, vcpu->arch.pdptrs[0]);
- vmcs_write64(GUEST_PDPTR1, vcpu->arch.pdptrs[1]);
- vmcs_write64(GUEST_PDPTR2, vcpu->arch.pdptrs[2]);
- vmcs_write64(GUEST_PDPTR3, vcpu->arch.pdptrs[3]);
- }
-}
-
-static void
-vmx_set_cr3(struct kvm_vcpu *vcpu, unsigned long cr3)
-{
- unsigned long guest_cr3;
- uint64_t eptp;
-
- guest_cr3 = cr3;
-
- if (enable_ept) {
- eptp = construct_eptp(cr3);
- vmcs_write64(EPT_POINTER, eptp);
- guest_cr3 = is_paging(vcpu) ?
- vcpu->arch.cr3 : vcpu->kvm->arch.ept_identity_map_addr;
- ept_load_pdptrs(vcpu);
- }
-
- vmx_flush_tlb(vcpu);
- vmcs_writel(GUEST_CR3, guest_cr3);
-}
-
-#define _ER(x) { EXIT_REASON_##x, #x }
-
-struct trace_print_flags {
- unsigned long mask;
- const char *name;
-};
-
-static const struct trace_print_flags vmx_exit_reasons_str[] = {
- _ER(EXCEPTION_NMI),
- _ER(EXTERNAL_INTERRUPT),
- _ER(TRIPLE_FAULT),
- _ER(PENDING_INTERRUPT),
- _ER(NMI_WINDOW),
- _ER(TASK_SWITCH),
- _ER(CPUID),
- _ER(HLT),
- _ER(INVLPG),
- _ER(RDPMC),
- _ER(RDTSC),
- _ER(VMCALL),
- _ER(VMCLEAR),
- _ER(VMLAUNCH),
- _ER(VMPTRLD),
- _ER(VMPTRST),
- _ER(VMREAD),
- _ER(VMRESUME),
- _ER(VMWRITE),
- _ER(VMOFF),
- _ER(VMON),
- _ER(CR_ACCESS),
- _ER(DR_ACCESS),
- _ER(IO_INSTRUCTION),
- _ER(MSR_READ),
- _ER(MSR_WRITE),
- _ER(MWAIT_INSTRUCTION),
- _ER(MONITOR_INSTRUCTION),
- _ER(PAUSE_INSTRUCTION),
- _ER(MCE_DURING_VMENTRY),
- _ER(TPR_BELOW_THRESHOLD),
- _ER(APIC_ACCESS),
- _ER(EPT_VIOLATION),
- _ER(EPT_MISCONFIG),
- _ER(WBINVD),
- { -1, NULL }
-};
-
-#undef _ER
-
-static int flexpriority_enabled = 1;
-static inline int
-report_flexpriority(void)
-{
- return (flexpriority_enabled);
-}
/*
* The function is based on mtrr_type_lookup() in
@@ -586,194 +383,20 @@ kvm_get_guest_memory_type(struct kvm_vcpu *vcpu, gfn_t gfn)
return (mtrr);
}
-static uint64_t
-vmx_get_mt_mask(struct kvm_vcpu *vcpu, gfn_t gfn, int is_mmio)
-{
- /*
- * For VT-d and EPT combination
- * 1. MMIO: always map as UC
- * 2. EPT with VT-d:
- * a. VT-d without snooping control feature: can't guarantee the
- * result, try to trust guest.
- * b. VT-d with snooping control feature: snooping control feature of
- * VT-d engine can guarantee the cache correctness. Just set it
- * to WB to keep consistent with host. So the same as item 3.
- * 3. EPT without VT-d: always map as WB and set IPAT=1 to keep
- * consistent with host MTRR
- */
- if (is_mmio)
- return (MTRR_TYPE_UNCACHABLE << VMX_EPT_MT_EPTE_SHIFT);
-
- if (vcpu->kvm->arch.iommu_domain &&
- !(vcpu->kvm->arch.iommu_flags & KVM_IOMMU_CACHE_COHERENCY)) {
- return (kvm_get_guest_memory_type(vcpu, gfn) <<
- VMX_EPT_MT_EPTE_SHIFT);
- }
-
- return ((MTRR_TYPE_WRBACK << VMX_EPT_MT_EPTE_SHIFT) | VMX_EPT_IPAT_BIT);
-}
-
-static void
-vmx_patch_hypercall(struct kvm_vcpu *vcpu, unsigned char *hypercall)
-{
- /*
- * Patch in the VMCALL instruction:
- */
- hypercall[0] = 0x0f;
- hypercall[1] = 0x01;
- hypercall[2] = 0xc1;
-}
static void vmx_get_cs_db_l_bits(struct kvm_vcpu *vcpu, int *db, int *l);
extern void update_exception_bitmap(struct kvm_vcpu *vcpu);
-static void
-set_guest_debug(struct kvm_vcpu *vcpu, struct kvm_guest_debug *dbg)
-{
- if (vcpu->guest_debug & KVM_GUESTDBG_USE_HW_BP)
- vmcs_writel(GUEST_DR7, dbg->arch.debugreg[7]);
- else
- vmcs_writel(GUEST_DR7, vcpu->arch.dr7);
-
- update_exception_bitmap(vcpu);
-}
-
extern struct vmcs_config vmcs_config;
static int setup_vmcs_config(struct vmcs_config *vmcs_conf);
-static void vmx_check_processor_compat(void *rtn)
-{
- struct vmcs_config vmcs_conf;
-
- if (setup_vmcs_config(&vmcs_conf) < 0)
- *(int *)rtn |= EIO;
- if (memcmp(&vmcs_config, &vmcs_conf, sizeof (struct vmcs_config))
- != 0) {
- cmn_err(CE_WARN, "kvm: CPU %d feature inconsistency!\n",
- curthread->t_cpu->cpu_id);
- *(int *)rtn |= EIO;
- }
-}
-
-static struct kvm_x86_ops vmx_x86_ops = {
- .cpu_has_kvm_support = nulldev, /* XXX: cpu_has_kvm_support? */
- .disabled_by_bios = nulldev, /* XXX: vmx_disabled_by_bios? */
-
- .hardware_enable = vmx_hardware_enable,
- .hardware_disable = hardware_disable,
-
- .check_processor_compatibility = vmx_check_processor_compat,
-
- .hardware_setup = vmx_hardware_setup,
-
- .hardware_unsetup = (void(*)(void))nulldev, /* XXX: hardware_unsetup? */
-
- .cpu_has_accelerated_tpr = report_flexpriority,
- .vcpu_create = vmx_create_vcpu,
- .vcpu_free = vmx_destroy_vcpu, /* XXX */
- .vcpu_reset = vmx_vcpu_reset,
-
- .prepare_guest_switch = vmx_save_host_state,
- .vcpu_load = vmx_vcpu_load,
- .vcpu_put = vmx_vcpu_put,
-
- .set_guest_debug = set_guest_debug,
- .get_msr = vmx_get_msr,
- .set_msr = vmx_set_msr,
- .get_segment_base = vmx_get_segment_base,
- .get_segment = vmx_get_segment,
- .set_segment = vmx_set_segment,
- .get_cpl = vmx_get_cpl,
- .get_cs_db_l_bits = vmx_get_cs_db_l_bits,
- .decache_cr0_guest_bits = vmx_decache_cr0_guest_bits,
- .decache_cr4_guest_bits = vmx_decache_cr4_guest_bits,
- .set_cr0 = vmx_set_cr0,
- .set_cr3 = vmx_set_cr3,
- .set_cr4 = vmx_set_cr4,
- .set_efer = vmx_set_efer,
- .get_idt = vmx_get_idt,
- .set_idt = vmx_set_idt,
- .get_gdt = vmx_get_gdt,
- .set_gdt = vmx_set_gdt,
- .cache_reg = vmx_cache_reg,
- .get_rflags = vmx_get_rflags,
- .set_rflags = vmx_set_rflags,
- .fpu_activate = vmx_fpu_activate,
- .fpu_deactivate = vmx_fpu_deactivate,
-
- .tlb_flush = vmx_flush_tlb,
-
- .run = vmx_vcpu_run,
- .handle_exit = vmx_handle_exit,
- .skip_emulated_instruction = skip_emulated_instruction,
- .set_interrupt_shadow = vmx_set_interrupt_shadow,
- .get_interrupt_shadow = vmx_get_interrupt_shadow,
- .patch_hypercall = vmx_patch_hypercall,
- .set_irq = vmx_inject_irq,
- .set_nmi = vmx_inject_nmi,
- .queue_exception = vmx_queue_exception,
- .interrupt_allowed = vmx_interrupt_allowed,
- .nmi_allowed = vmx_nmi_allowed,
- .get_nmi_mask = vmx_get_nmi_mask,
- .set_nmi_mask = vmx_set_nmi_mask,
- .enable_nmi_window = enable_nmi_window,
- .enable_irq_window = enable_irq_window,
- .update_cr8_intercept = vmx_update_cr8_intercept,
-
- .set_tss_addr = vmx_set_tss_addr,
- .get_tdp_level = get_ept_level,
- .get_mt_mask = vmx_get_mt_mask,
-
- .exit_reasons_str = vmx_exit_reasons_str,
-
- .get_lpage_level = vmx_get_lpage_level,
-
- .cpuid_update = vmx_cpuid_update,
-
- .rdtscp_supported = vmx_rdtscp_supported
-};
struct kvm_x86_ops *kvm_x86_ops;
-uint32_t
-vmcs_read32(unsigned long field)
-{
- return (vmcs_readl(field));
-}
-
-static void
-vmx_get_cs_db_l_bits(struct kvm_vcpu *vcpu, int *db, int *l)
-{
- uint32_t ar = vmcs_read32(GUEST_CS_AR_BYTES);
-
- *db = (ar >> 14) & 1;
- *l = (ar >> 13) & 1;
-}
-
-void
-vmcs_write32(unsigned long field, uint32_t value)
-{
- vmcs_writel(field, value);
-}
-
-static void
-vmx_decache_cr0_guest_bits(struct kvm_vcpu *vcpu)
-{
- ulong cr0_guest_owned_bits = vcpu->arch.cr0_guest_owned_bits;
- vcpu->arch.cr0 &= ~cr0_guest_owned_bits;
- vcpu->arch.cr0 |= vmcs_readl(GUEST_CR0) & cr0_guest_owned_bits;
-}
-static void
-vmx_decache_cr4_guest_bits(struct kvm_vcpu *vcpu)
-{
- ulong cr4_guest_owned_bits = vcpu->arch.cr4_guest_owned_bits;
- vcpu->arch.cr4 &= ~cr4_guest_owned_bits;
- vcpu->arch.cr4 |= vmcs_readl(GUEST_CR4) & cr4_guest_owned_bits;
-}
inline ulong
kvm_read_cr0_bits(struct kvm_vcpu *vcpu, ulong mask)
@@ -786,154 +409,7 @@ kvm_read_cr0_bits(struct kvm_vcpu *vcpu, ulong mask)
return (vcpu->arch.cr0 & mask);
}
-static void
-vmcs_clear_bits(unsigned long field, uint32_t mask)
-{
- vmcs_writel(field, vmcs_readl(field) & ~mask);
-}
-
-static void
-vmcs_set_bits(unsigned long field, uint32_t mask)
-{
- vmcs_writel(field, vmcs_readl(field) | mask);
-}
-
-void
-vmx_fpu_activate(struct kvm_vcpu *vcpu)
-{
- ulong cr0;
-
- if (vcpu->fpu_active)
- return;
-
- vcpu->fpu_active = 1;
- cr0 = vmcs_readl(GUEST_CR0);
- cr0 &= ~(X86_CR0_TS | X86_CR0_MP);
- cr0 |= kvm_read_cr0_bits(vcpu, X86_CR0_TS | X86_CR0_MP);
- vmcs_writel(GUEST_CR0, cr0);
- update_exception_bitmap(vcpu);
- vcpu->arch.cr0_guest_owned_bits = X86_CR0_TS;
- vmcs_writel(CR0_GUEST_HOST_MASK, ~vcpu->arch.cr0_guest_owned_bits);
-}
-
-static void
-vmx_fpu_deactivate(struct kvm_vcpu *vcpu)
-{
- vmx_decache_cr0_guest_bits(vcpu);
- vmcs_set_bits(GUEST_CR0, X86_CR0_TS | X86_CR0_MP);
- update_exception_bitmap(vcpu);
- vcpu->arch.cr0_guest_owned_bits = 0;
- vmcs_writel(CR0_GUEST_HOST_MASK, ~vcpu->arch.cr0_guest_owned_bits);
- vmcs_writel(CR0_READ_SHADOW, vcpu->arch.cr0);
-}
-
-static inline uint32_t
-bit(int bitno)
-{
- return (1 << (bitno & 31));
-}
-
-static void
-vmx_cpuid_update(struct kvm_vcpu *vcpu)
-{
- struct kvm_cpuid_entry2 *best;
- struct vcpu_vmx *vmx = to_vmx(vcpu);
- uint32_t exec_control;
-
- vmx->rdtscp_enabled = 0;
-
- if (vmx_rdtscp_supported()) {
- exec_control = vmcs_read32(SECONDARY_VM_EXEC_CONTROL);
- if (exec_control & SECONDARY_EXEC_RDTSCP) {
- best = kvm_find_cpuid_entry(vcpu, 0x80000001, 0);
- if (best && (best->edx & bit(X86_FEATURE_RDTSCP)))
- vmx->rdtscp_enabled = 1;
- else {
- exec_control &= ~SECONDARY_EXEC_RDTSCP;
- vmcs_write32(SECONDARY_VM_EXEC_CONTROL,
- exec_control);
- }
- }
- }
-}
-
-static void
-enable_irq_window(struct kvm_vcpu *vcpu)
-{
- uint32_t cpu_based_vm_exec_control;
-
- cpu_based_vm_exec_control = vmcs_read32(CPU_BASED_VM_EXEC_CONTROL);
- cpu_based_vm_exec_control |= CPU_BASED_VIRTUAL_INTR_PENDING;
- vmcs_write32(CPU_BASED_VM_EXEC_CONTROL, cpu_based_vm_exec_control);
-}
-
-extern struct vmcs_config vmcs_config;
-
-static inline int
-cpu_has_virtual_nmis(void)
-{
- return (vmcs_config.pin_based_exec_ctrl & PIN_BASED_VIRTUAL_NMIS);
-}
-
-static void
-enable_nmi_window(struct kvm_vcpu *vcpu)
-{
- uint32_t cpu_based_vm_exec_control;
-
- if (!cpu_has_virtual_nmis()) {
- enable_irq_window(vcpu);
- return;
- }
-
- cpu_based_vm_exec_control = vmcs_read32(CPU_BASED_VM_EXEC_CONTROL);
- cpu_based_vm_exec_control |= CPU_BASED_VIRTUAL_NMI_PENDING;
- vmcs_write32(CPU_BASED_VM_EXEC_CONTROL, cpu_based_vm_exec_control);
-}
-
-static void
-vmx_set_nmi_mask(struct kvm_vcpu *vcpu, int masked)
-{
- struct vcpu_vmx *vmx = to_vmx(vcpu);
-
- if (!cpu_has_virtual_nmis()) {
- if (vmx->soft_vnmi_blocked != masked) {
- vmx->soft_vnmi_blocked = masked;
- vmx->vnmi_blocked_time = 0;
- }
-
- return;
- } else {
- if (masked) {
- vmcs_set_bits(GUEST_INTERRUPTIBILITY_INFO,
- GUEST_INTR_STATE_NMI);
- } else {
- vmcs_clear_bits(GUEST_INTERRUPTIBILITY_INFO,
- GUEST_INTR_STATE_NMI);
- }
- }
-}
-
-static int
-vmx_get_nmi_mask(struct kvm_vcpu *vcpu)
-{
- if (!cpu_has_virtual_nmis())
- return (to_vmx(vcpu)->soft_vnmi_blocked);
- else
- return (!!(vmcs_read32(GUEST_INTERRUPTIBILITY_INFO) &
- GUEST_INTR_STATE_NMI));
-}
-
-static int
-vmx_nmi_allowed(struct kvm_vcpu *vcpu)
-{
- if (!cpu_has_virtual_nmis() && to_vmx(vcpu)->soft_vnmi_blocked)
- return (0);
-
- return (!(vmcs_read32(GUEST_INTERRUPTIBILITY_INFO) &
- (GUEST_INTR_STATE_MOV_SS | GUEST_INTR_STATE_NMI)));
-}
-
-static inline unsigned long
+unsigned long
kvm_register_read(struct kvm_vcpu *vcpu, enum kvm_reg reg)
{
if (!test_bit(reg, (unsigned long *)&vcpu->arch.regs_avail))
@@ -962,400 +438,12 @@ kvm_rip_write(struct kvm_vcpu *vcpu, unsigned long val)
kvm_register_write(vcpu, VCPU_REGS_RIP, val);
}
-static inline int
+inline int
kvm_exception_is_soft(unsigned int nr)
{
return (nr == BP_VECTOR) || (nr == OF_VECTOR);
}
-static void
-vmx_queue_exception(struct kvm_vcpu *vcpu, unsigned nr,
- int has_error_code, uint32_t error_code)
-{
- struct vcpu_vmx *vmx = to_vmx(vcpu);
- uint32_t intr_info = nr | INTR_INFO_VALID_MASK;
-
- if (has_error_code) {
- vmcs_write32(VM_ENTRY_EXCEPTION_ERROR_CODE, error_code);
- intr_info |= INTR_INFO_DELIVER_CODE_MASK;
- }
-
- if (vmx->rmode.vm86_active) {
- vmx->rmode.irq.pending = 1;
- vmx->rmode.irq.vector = nr;
- vmx->rmode.irq.rip = kvm_rip_read(vcpu);
- if (kvm_exception_is_soft(nr))
- vmx->rmode.irq.rip +=
- vmx->vcpu.arch.event_exit_inst_len;
- intr_info |= INTR_TYPE_SOFT_INTR;
- vmcs_write32(VM_ENTRY_INTR_INFO_FIELD, intr_info);
- vmcs_write32(VM_ENTRY_INSTRUCTION_LEN, 1);
- kvm_rip_write(vcpu, vmx->rmode.irq.rip - 1);
- return;
- }
-
- if (kvm_exception_is_soft(nr)) {
- vmcs_write32(VM_ENTRY_INSTRUCTION_LEN,
- vmx->vcpu.arch.event_exit_inst_len);
- intr_info |= INTR_TYPE_SOFT_EXCEPTION;
- } else
- intr_info |= INTR_TYPE_HARD_EXCEPTION;
-
- vmcs_write32(VM_ENTRY_INTR_INFO_FIELD, intr_info);
-}
-
-static void
-vmx_inject_nmi(struct kvm_vcpu *vcpu)
-{
- struct vcpu_vmx *vmx = to_vmx(vcpu);
-
- if (!cpu_has_virtual_nmis()) {
- /*
- * Tracking the NMI-blocked state in software is built upon
- * finding the next open IRQ window. This, in turn, depends on
- * well-behaving guests: They have to keep IRQs disabled at
- * least as long as the NMI handler runs. Otherwise we may
- * cause NMI nesting, maybe breaking the guest. But as this is
- * highly unlikely, we can live with the residual risk.
- */
- vmx->soft_vnmi_blocked = 1;
- vmx->vnmi_blocked_time = 0;
- }
-
- KVM_VCPU_KSTAT_INC(vcpu, kvmvs_nmi_injections);
-
- if (vmx->rmode.vm86_active) {
- vmx->rmode.irq.pending = 1;
- vmx->rmode.irq.vector = NMI_VECTOR;
- vmx->rmode.irq.rip = kvm_rip_read(vcpu);
- vmcs_write32(VM_ENTRY_INTR_INFO_FIELD,
- NMI_VECTOR | INTR_TYPE_SOFT_INTR |
- INTR_INFO_VALID_MASK);
- vmcs_write32(VM_ENTRY_INSTRUCTION_LEN, 1);
- kvm_rip_write(vcpu, vmx->rmode.irq.rip - 1);
- return;
- }
- vmcs_write32(VM_ENTRY_INTR_INFO_FIELD,
- INTR_TYPE_NMI_INTR | INTR_INFO_VALID_MASK | NMI_VECTOR);
-}
-
-static void
-vmx_inject_irq(struct kvm_vcpu *vcpu)
-{
- struct vcpu_vmx *vmx = to_vmx(vcpu);
- uint32_t intr;
- int irq = vcpu->arch.interrupt.nr;
-
- KVM_TRACE1(inj__virq, int, irq);
- KVM_VCPU_KSTAT_INC(vcpu, kvmvs_irq_injections);
-
- if (vmx->rmode.vm86_active) {
- vmx->rmode.irq.pending = 1;
- vmx->rmode.irq.vector = irq;
- vmx->rmode.irq.rip = kvm_rip_read(vcpu);
- if (vcpu->arch.interrupt.soft)
- vmx->rmode.irq.rip +=
- vmx->vcpu.arch.event_exit_inst_len;
- vmcs_write32(VM_ENTRY_INTR_INFO_FIELD,
- irq | INTR_TYPE_SOFT_INTR | INTR_INFO_VALID_MASK);
- vmcs_write32(VM_ENTRY_INSTRUCTION_LEN, 1);
- kvm_rip_write(vcpu, vmx->rmode.irq.rip - 1);
- return;
- }
- intr = irq | INTR_INFO_VALID_MASK;
- if (vcpu->arch.interrupt.soft) {
- intr |= INTR_TYPE_SOFT_INTR;
- vmcs_write32(VM_ENTRY_INSTRUCTION_LEN,
- vmx->vcpu.arch.event_exit_inst_len);
- } else
- intr |= INTR_TYPE_EXT_INTR;
-
- vmcs_write32(VM_ENTRY_INTR_INFO_FIELD, intr);
-}
-
-
-static void
-vmx_get_idt(struct kvm_vcpu *vcpu, struct descriptor_table *dt)
-{
- dt->limit = vmcs_read32(GUEST_IDTR_LIMIT);
- dt->base = vmcs_readl(GUEST_IDTR_BASE);
-}
-
-static void
-vmx_set_idt(struct kvm_vcpu *vcpu, struct descriptor_table *dt)
-{
- vmcs_write32(GUEST_IDTR_LIMIT, dt->limit);
- vmcs_writel(GUEST_IDTR_BASE, dt->base);
-}
-
-static void
-vmx_get_gdt(struct kvm_vcpu *vcpu, struct descriptor_table *dt)
-{
- dt->limit = vmcs_read32(GUEST_GDTR_LIMIT);
- dt->base = vmcs_readl(GUEST_GDTR_BASE);
-}
-
-static void
-vmx_set_gdt(struct kvm_vcpu *vcpu, struct descriptor_table *dt)
-{
- vmcs_write32(GUEST_GDTR_LIMIT, dt->limit);
- vmcs_writel(GUEST_GDTR_BASE, dt->base);
-}
-
-static uint32_t
-vmx_get_interrupt_shadow(struct kvm_vcpu *vcpu, int mask)
-{
- uint32_t interruptibility = vmcs_read32(GUEST_INTERRUPTIBILITY_INFO);
- int ret = 0;
-
- if (interruptibility & GUEST_INTR_STATE_STI)
- ret |= X86_SHADOW_INT_STI;
- if (interruptibility & GUEST_INTR_STATE_MOV_SS)
- ret |= X86_SHADOW_INT_MOV_SS;
-
- return (ret & mask);
-}
-
-static void
-vmx_set_interrupt_shadow(struct kvm_vcpu *vcpu, int mask)
-{
- uint32_t old = vmcs_read32(GUEST_INTERRUPTIBILITY_INFO);
- uint32_t interruptibility = old;
-
- interruptibility &= ~(GUEST_INTR_STATE_STI | GUEST_INTR_STATE_MOV_SS);
-
- if (mask & X86_SHADOW_INT_MOV_SS)
- interruptibility |= GUEST_INTR_STATE_MOV_SS;
- if (mask & X86_SHADOW_INT_STI)
- interruptibility |= GUEST_INTR_STATE_STI;
-
- if ((interruptibility != old))
- vmcs_write32(GUEST_INTERRUPTIBILITY_INFO, interruptibility);
-}
-
-static void
-skip_emulated_instruction(struct kvm_vcpu *vcpu)
-{
- unsigned long rip;
-
- rip = kvm_rip_read(vcpu);
- rip += vmcs_read32(VM_EXIT_INSTRUCTION_LEN);
- kvm_rip_write(vcpu, rip);
-
- /* skipping an emulated instruction also counts */
- vmx_set_interrupt_shadow(vcpu, 0);
-}
-
-/*
- * In linux, there is a separate vmx kernel module from the kvm driver.
- * That may be a good idea, but we're going to do everything in
- * the kvm driver, for now.
- * The call to vmx_init() in _init() is done when the vmx module
- * is loaded on linux.
- */
-
-struct vmcs **vmxarea; /* 1 per cpu */
-struct vmcs **current_vmcs;
-static struct kvm_shared_msrs **shared_msrs;
-list_t **vcpus_on_cpu;
-
-uint64_t *vmxarea_pa; /* physical address of each vmxarea */
-
-static int
-alloc_kvm_area(void)
-{
- int i, j;
- pfn_t pfn;
-
- /*
- * linux seems to do the allocations in a numa-aware
- * fashion. We'll just allocate...
- */
- vmxarea = kmem_alloc(ncpus * sizeof (struct vmcs *), KM_SLEEP);
- vmxarea_pa = kmem_alloc(ncpus * sizeof (uint64_t *), KM_SLEEP);
- current_vmcs = kmem_alloc(ncpus * sizeof (struct vmcs *), KM_SLEEP);
- shared_msrs = kmem_alloc(ncpus * sizeof (struct kvm_shared_msrs *),
- KM_SLEEP);
- vcpus_on_cpu = kmem_alloc(ncpus * sizeof (list_t *), KM_SLEEP);
-
- for (i = 0; i < ncpus; i++) {
- struct vmcs *vmcs;
-
- /* XXX the following assumes PAGESIZE allocations */
- /* are PAGESIZE aligned. We could enforce this */
- /* via kmem_cache_create, but I'm lazy */
- vmcs = kmem_zalloc(PAGESIZE, KM_SLEEP);
- vmxarea[i] = vmcs;
- current_vmcs[i] = vmcs;
- pfn = hat_getpfnum(kas.a_hat, (caddr_t)vmcs);
- vmxarea_pa[i] = ((uint64_t)pfn << PAGESHIFT) |
- ((uint64_t)vmxarea[i] & PAGEOFFSET);
- shared_msrs[i] = kmem_zalloc(sizeof (struct kvm_shared_msrs),
- KM_SLEEP);
- vcpus_on_cpu[i] = kmem_alloc(sizeof (list_t), KM_SLEEP);
- list_create(vcpus_on_cpu[i], sizeof (struct vcpu_vmx),
- offsetof(struct vcpu_vmx, local_vcpus_link));
- }
-
- return (0);
-}
-
-static int
-adjust_vmx_controls(uint32_t ctl_min, uint32_t ctl_opt,
- uint32_t msr, uint32_t *result)
-{
- uint32_t vmx_msr_low, vmx_msr_high;
- uint32_t ctl = ctl_min | ctl_opt;
-
- rdmsr(msr, vmx_msr_low, vmx_msr_high);
-
- ctl &= vmx_msr_high; /* bit == 0 in high word ==> must be zero */
- ctl |= vmx_msr_low; /* bit == 1 in low word ==> must be one */
-
- /* Ensure minimum (required) set of control bits are supported. */
- if (ctl_min & ~ctl)
- return (EIO);
-
- *result = ctl;
- return (DDI_SUCCESS);
-}
-
-/* Pure 2^n version of get_order */
-static inline int
-get_order(unsigned long size)
-{
- int order;
-
- size = (size - 1) >> (PAGESHIFT - 1);
- order = -1;
- do {
- size >>= 1;
- order++;
- } while (size);
-
- return (order);
-}
-
-static int
-setup_vmcs_config(struct vmcs_config *vmcs_conf)
-{
- uint32_t vmx_msr_low, vmx_msr_high;
- uint32_t min, opt, min2, opt2;
- uint32_t _pin_based_exec_control = 0;
- uint32_t _cpu_based_exec_control = 0;
- uint32_t _cpu_based_2nd_exec_control = 0;
- uint32_t _vmexit_control = 0;
- uint32_t _vmentry_control = 0;
- uint32_t ept, vpid;
-
- min = PIN_BASED_EXT_INTR_MASK | PIN_BASED_NMI_EXITING;
- opt = PIN_BASED_VIRTUAL_NMIS;
- if (adjust_vmx_controls(min, opt, MSR_IA32_VMX_PINBASED_CTLS,
- &_pin_based_exec_control) != DDI_SUCCESS)
- return (EIO);
-
- min = CPU_BASED_HLT_EXITING |
-#ifdef CONFIG_X86_64
- CPU_BASED_CR8_LOAD_EXITING |
- CPU_BASED_CR8_STORE_EXITING |
-#endif
- CPU_BASED_CR3_LOAD_EXITING |
- CPU_BASED_CR3_STORE_EXITING |
- CPU_BASED_USE_IO_BITMAPS |
- CPU_BASED_MOV_DR_EXITING |
- CPU_BASED_USE_TSC_OFFSETING |
- CPU_BASED_MWAIT_EXITING |
- CPU_BASED_MONITOR_EXITING |
- CPU_BASED_INVLPG_EXITING;
-
- opt = CPU_BASED_TPR_SHADOW |
- CPU_BASED_USE_MSR_BITMAPS |
- CPU_BASED_ACTIVATE_SECONDARY_CONTROLS;
-
- if (adjust_vmx_controls(min, opt, MSR_IA32_VMX_PROCBASED_CTLS,
- &_cpu_based_exec_control) != DDI_SUCCESS)
- return (EIO);
-
-#ifdef CONFIG_X86_64
- if ((_cpu_based_exec_control & CPU_BASED_TPR_SHADOW))
- _cpu_based_exec_control &= ~CPU_BASED_CR8_LOAD_EXITING &
- ~CPU_BASED_CR8_STORE_EXITING;
-#endif
- if (_cpu_based_exec_control & CPU_BASED_ACTIVATE_SECONDARY_CONTROLS) {
- min2 = 0;
- opt2 = SECONDARY_EXEC_VIRTUALIZE_APIC_ACCESSES |
- SECONDARY_EXEC_WBINVD_EXITING |
- SECONDARY_EXEC_ENABLE_VPID |
- SECONDARY_EXEC_ENABLE_EPT |
- SECONDARY_EXEC_UNRESTRICTED_GUEST |
- SECONDARY_EXEC_PAUSE_LOOP_EXITING |
- SECONDARY_EXEC_RDTSCP;
-
- if (adjust_vmx_controls(min2, opt2,
- MSR_IA32_VMX_PROCBASED_CTLS2,
- &_cpu_based_2nd_exec_control) != DDI_SUCCESS)
- return (EIO);
- }
-#ifndef CONFIG_X86_64
- if (!(_cpu_based_2nd_exec_control &
- SECONDARY_EXEC_VIRTUALIZE_APIC_ACCESSES))
- _cpu_based_exec_control &= ~CPU_BASED_TPR_SHADOW;
-#endif
- if (_cpu_based_2nd_exec_control & SECONDARY_EXEC_ENABLE_EPT) {
- /*
- * CR3 accesses and invlpg don't need to cause VM Exits when EPT
- * enabled
- */
- _cpu_based_exec_control &= ~(CPU_BASED_CR3_LOAD_EXITING |
- CPU_BASED_CR3_STORE_EXITING | CPU_BASED_INVLPG_EXITING);
- rdmsr(MSR_IA32_VMX_EPT_VPID_CAP, vmx_capability.ept,
- vmx_capability.vpid);
- }
-
- min = 0;
-#ifdef CONFIG_X86_64
- min |= VM_EXIT_HOST_ADDR_SPACE_SIZE;
-#endif
- opt = VM_EXIT_SAVE_IA32_PAT | VM_EXIT_LOAD_IA32_PAT;
- if (adjust_vmx_controls(min, opt, MSR_IA32_VMX_EXIT_CTLS,
- &_vmexit_control) != DDI_SUCCESS)
- return (EIO);
-
- min = 0;
- opt = VM_ENTRY_LOAD_IA32_PAT;
- if (adjust_vmx_controls(min, opt, MSR_IA32_VMX_ENTRY_CTLS,
- &_vmentry_control) != DDI_SUCCESS)
- return (EIO);
-
- rdmsr(MSR_IA32_VMX_BASIC, vmx_msr_low, vmx_msr_high);
-
- /* IA-32 SDM Vol 3B: VMCS size is never greater than 4kB. */
- if ((vmx_msr_high & 0x1fff) > PAGESIZE)
- return (EIO);
-
-#ifdef CONFIG_X86_64
- /* IA-32 SDM Vol 3B: 64-bit CPUs always have VMX_BASIC_MSR[48]==0. */
- if (vmx_msr_high & (1u<<16))
- return (EIO);
-#endif
-
- /* Require Write-Back (WB) memory type for VMCS accesses. */
- if (((vmx_msr_high >> 18) & 15) != 6)
- return (EIO);
-
- vmcs_conf->size = vmx_msr_high & 0x1fff;
- vmcs_conf->order = get_order(vmcs_config.size);
- vmcs_conf->revision_id = vmx_msr_low;
-
- vmcs_conf->pin_based_exec_ctrl = _pin_based_exec_control;
- vmcs_conf->cpu_based_exec_ctrl = _cpu_based_exec_control;
- vmcs_conf->cpu_based_2nd_exec_ctrl = _cpu_based_2nd_exec_control;
- vmcs_conf->vmexit_ctrl = _vmexit_control;
- vmcs_conf->vmentry_ctrl = _vmentry_control;
-
- return (0);
-}
-
/*
* EFER defaults:
* - enable syscall per default because its emulated by KVM
@@ -1367,131 +455,18 @@ static uint64_t efer_reserved_bits = 0xfffffffffffffafeULL;
static uint64_t efer_reserved_bits = 0xfffffffffffffffeULL;
#endif
-static int bypass_guest_pf = 1;
-int enable_vpid = 1;
-int enable_unrestricted_guest = 1;
-int emulate_invalid_guest_state = 0;
-
void
kvm_enable_efer_bits(uint64_t mask)
{
efer_reserved_bits &= ~mask;
}
-static inline int
-cpu_has_vmx_vpid(void)
-{
- return (vmcs_config.cpu_based_2nd_exec_ctrl &
- SECONDARY_EXEC_ENABLE_VPID);
-}
-
-static inline int
-cpu_has_vmx_ept(void)
-{
- return (vmcs_config.cpu_based_2nd_exec_ctrl &
- SECONDARY_EXEC_ENABLE_EPT);
-}
-
-static inline int
-cpu_has_vmx_unrestricted_guest(void)
-{
- return (vmcs_config.cpu_based_2nd_exec_ctrl &
- SECONDARY_EXEC_UNRESTRICTED_GUEST);
-}
-
-inline int
-cpu_has_vmx_tpr_shadow(void)
-{
- return (vmcs_config.cpu_based_exec_ctrl & CPU_BASED_TPR_SHADOW);
-}
-
-static inline int
-cpu_has_vmx_virtualize_apic_accesses(void)
-{
- return (vmcs_config.cpu_based_2nd_exec_ctrl &
- SECONDARY_EXEC_VIRTUALIZE_APIC_ACCESSES);
-}
-
-static inline int
-cpu_has_vmx_flexpriority(void)
-{
- return (cpu_has_vmx_tpr_shadow() &&
- cpu_has_vmx_virtualize_apic_accesses());
-}
-
-static inline int
-cpu_has_vmx_ept_2m_page(void)
-{
- return (!!(vmx_capability.ept & VMX_EPT_2MB_PAGE_BIT));
-}
-
void
kvm_disable_largepages(void)
{
largepages_enabled = 0;
}
-static inline int
-cpu_has_vmx_ple(void)
-{
- return (vmcs_config.cpu_based_2nd_exec_ctrl &
- SECONDARY_EXEC_PAUSE_LOOP_EXITING);
-}
-
-/*
- * These 2 parameters are used to config the controls for Pause-Loop Exiting:
- * ple_gap: upper bound on the amount of time between two successive
- * executions of PAUSE in a loop. Also indicate if ple enabled.
- * According to test, this time is usually small than 41 cycles.
- * ple_window: upper bound on the amount of time a guest is allowed to execute
- * in a PAUSE loop. Tests indicate that most spinlocks are held for
- * less than 2^12 cycles
- * Time is measured based on a counter that runs at the same rate as the TSC,
- * refer SDM volume 3b section 21.6.13 & 22.1.3.
- */
-#define KVM_VMX_EFAULT_PLE_GAP 41
-#define KVM_VMX_DEFAULT_PLE_WINDOW 4096
-static int ple_gap = KVM_VMX_DEFAULT_PLE_GAP;
-static int ple_window = KVM_VMX_DEFAULT_PLE_WINDOW;
-
-static int
-vmx_hardware_setup(void)
-{
- if (setup_vmcs_config(&vmcs_config) != DDI_SUCCESS)
- return (EIO);
-#ifdef XXX
- if (boot_cpu_has(X86_FEATURE_NX))
-#else
- XXX_KVM_PROBE;
-#endif
- kvm_enable_efer_bits(EFER_NX);
-
- if (!cpu_has_vmx_vpid())
- enable_vpid = 0;
-
- if (!cpu_has_vmx_ept()) {
- enable_ept = 0;
- enable_unrestricted_guest = 0;
- }
-
- if (!cpu_has_vmx_unrestricted_guest())
- enable_unrestricted_guest = 0;
- if (!cpu_has_vmx_flexpriority())
- flexpriority_enabled = 0;
-
- if (!cpu_has_vmx_tpr_shadow())
- kvm_x86_ops->update_cr8_intercept = NULL;
-
- if (enable_ept && !cpu_has_vmx_ept_2m_page())
- kvm_disable_largepages();
-
- if (!cpu_has_vmx_ple())
- ple_gap = 0;
-
-
- return (alloc_kvm_area());
-}
-
int
kvm_arch_hardware_setup(void)
{
@@ -2955,47 +1930,7 @@ out_fail:
return (r);
}
-extern unsigned long vmx_io_bitmap_a[];
-extern unsigned long vmx_io_bitmap_b[];
-extern unsigned long vmx_msr_bitmap_legacy[];
-extern unsigned long vmx_msr_bitmap_longmode[];
-
-static inline int
-cpu_has_vmx_msr_bitmap(void)
-{
- return (vmcs_config.cpu_based_exec_ctrl & CPU_BASED_USE_MSR_BITMAPS);
-}
-
-static void
-__vmx_disable_intercept_for_msr(unsigned long *msr_bitmap, uint32_t msr)
-{
- int f = sizeof (unsigned long);
-
- if (!cpu_has_vmx_msr_bitmap())
- return;
-
- /*
- * See Intel PRM Vol. 3, 20.6.9 (MSR-Bitmap Address). Early manuals
- * have the write-low and read-high bitmap offsets the wrong way round.
- * We can control MSRs 0x00000000-0x00001fff and 0xc0000000-0xc0001fff.
- */
- if (msr <= 0x1fff) {
- __clear_bit(msr, msr_bitmap + 0x000 / f); /* read-low */
- __clear_bit(msr, msr_bitmap + 0x800 / f); /* write-low */
- } else if ((msr >= 0xc0000000) && (msr <= 0xc0001fff)) {
- msr &= 0x1fff;
- __clear_bit(msr, msr_bitmap + 0x400 / f); /* read-high */
- __clear_bit(msr, msr_bitmap + 0xc00 / f); /* write-high */
- }
-}
-static void
-vmx_disable_intercept_for_msr(uint32_t msr, int longmode_only)
-{
- if (!longmode_only)
- __vmx_disable_intercept_for_msr(vmx_msr_bitmap_legacy, msr);
- __vmx_disable_intercept_for_msr(vmx_msr_bitmap_longmode, msr);
-}
static struct kvm_shared_msrs_global shared_msrs_global;
@@ -3013,19 +1948,6 @@ kvm_define_shared_msr(unsigned slot, uint32_t msr)
#endif
}
-static uint64_t host_efer;
-
-/*
- * Keep MSR_K6_STAR at the end, as setup_msrs() will try to optimize it
- * away by decrementing the array size.
- */
-static const uint32_t vmx_msr_index[] = {
-#ifdef CONFIG_X86_64
- MSR_SYSCALL_MASK, MSR_LSTAR, MSR_CSTAR,
-#endif
- MSR_EFER, MSR_TSC_AUX, MSR_K6_STAR,
-};
-#define NR_VMX_MSR ARRAY_SIZE(vmx_msr_index)
#define VMX_NR_VPIDS (1 << 16)
ulong_t *vmx_vpid_bitmap;
size_t vpid_bitmap_words;
@@ -3043,78 +1965,6 @@ kvm_enable_tdp(void)
tdp_enabled = 1;
}
-static int
-vmx_init(void)
-{
- int r, i;
-
- rdmsrl_safe(MSR_EFER, (unsigned long long *)&host_efer);
-
- for (i = 0; i < NR_VMX_MSR; ++i)
- kvm_define_shared_msr(i, vmx_msr_index[i]);
-
-#ifdef XXX
- vmx_io_bitmap_a = kmem_zalloc(PAGESIZE, KM_SLEEP);
- vmx_io_bitmap_b = kmem_zalloc(PAGESIZE, KM_SLEEP);
- vmx_msr_bitmap_legacy = kmem_zalloc(PAGESIZE, KM_SLEEP);
- vmx_msr_bitmap_longmode = kmem_zalloc(PAGESIZE, KM_SLEEP);
-#else
- XXX_KVM_PROBE;
-#endif
-
- /*
- * Allow direct access to the PC debug port (it is often used for I/O
- * delays, but the vmexits simply slow things down).
- */
- memset(vmx_io_bitmap_a, 0xff, PAGESIZE);
- clear_bit(0x80, vmx_io_bitmap_a);
-
- memset(vmx_io_bitmap_b, 0xff, PAGESIZE);
-
- memset(vmx_msr_bitmap_legacy, 0xff, PAGESIZE);
- memset(vmx_msr_bitmap_longmode, 0xff, PAGESIZE);
-
- set_bit(0, vmx_vpid_bitmap); /* 0 is reserved for host */
-
- r = kvm_init(&vmx_x86_ops, sizeof (struct vcpu_vmx));
-
- if (r)
- goto out3;
-
- vmx_disable_intercept_for_msr(MSR_FS_BASE, 0);
- vmx_disable_intercept_for_msr(MSR_GS_BASE, 0);
- vmx_disable_intercept_for_msr(MSR_KERNEL_GS_BASE, 1);
- vmx_disable_intercept_for_msr(MSR_IA32_SYSENTER_CS, 0);
- vmx_disable_intercept_for_msr(MSR_IA32_SYSENTER_ESP, 0);
- vmx_disable_intercept_for_msr(MSR_IA32_SYSENTER_EIP, 0);
-
- if (enable_ept) {
- bypass_guest_pf = 0;
- kvm_mmu_set_base_ptes(VMX_EPT_READABLE_MASK |
- VMX_EPT_WRITABLE_MASK);
- kvm_mmu_set_mask_ptes(0ull, 0ull, 0ull, 0ull,
- VMX_EPT_EXECUTABLE_MASK);
- kvm_enable_tdp();
- } else
- kvm_disable_tdp();
-
- if (bypass_guest_pf)
- kvm_mmu_set_nonpresent_ptes(~0xffeull, 0ull);
-
- return (0);
-
-out3:
- kmem_free(vmx_msr_bitmap_longmode, PAGESIZE);
-out2:
- kmem_free(vmx_msr_bitmap_legacy, PAGESIZE);
-out1:
- kmem_free(vmx_io_bitmap_b, PAGESIZE);
-out:
- kmem_free(vmx_io_bitmap_a, PAGESIZE);
-
- return (r);
-}
-
int
_init(void)
{
@@ -3338,8 +2188,8 @@ kvm_close(dev_t dev, int flag, int otyp, cred_t *cred)
return (0);
}
-
-static void hardware_enable(void *junk)
+static void
+hardware_enable(void *junk)
{
int cpu;
int r;
@@ -3361,7 +2211,8 @@ static void hardware_enable(void *junk)
}
}
-static void hardware_disable(void *junk)
+static void
+hardware_disable(void *junk)
{
int cpu = curthread->t_cpu->cpu_id;
@@ -4217,27 +3068,6 @@ kvm_set_memory_region(kvm_t *kvm,
return (r);
}
-static int
-vmx_set_tss_addr(struct kvm *kvmp, caddr_t addr)
-{
- int ret;
-
- struct kvm_userspace_memory_region tss_mem = {
- .slot = TSS_PRIVATE_MEMSLOT,
- .guest_phys_addr = (uint64_t)addr,
- .memory_size = PAGESIZE * 3,
- .flags = 0,
- };
-
- ret = kvm_set_memory_region(kvmp, &tss_mem, 0);
-
- if (ret)
- return (ret);
-
- kvmp->arch.tss_addr = (uint64_t)addr;
-
- return (DDI_SUCCESS);
-}
static int
kvm_vm_ioctl_set_tss_addr(struct kvm *kvmp, caddr_t addr)
@@ -4246,7 +3076,7 @@ kvm_vm_ioctl_set_tss_addr(struct kvm *kvmp, caddr_t addr)
* XXX later, if adding other arch beside x86, need to do something
* else here
*/
- return (vmx_set_tss_addr(kvmp, addr));
+ return (kvm_x86_ops->set_tss_addr(kvmp, addr));
}
extern int kvm_vm_ioctl_create_vcpu(struct kvm *kvm, uint32_t id, int *rv);
@@ -4295,33 +3125,6 @@ is_efer_nx(void)
return (efer & EFER_NX);
}
-static inline int
-cpu_has_vmx_ept_1g_page(void)
-{
- return (!!(vmx_capability.ept & VMX_EPT_1GB_PAGE_BIT));
-}
-
-static int
-vmx_get_lpage_level(void)
-{
- if (enable_ept && !cpu_has_vmx_ept_1g_page())
- return (PT_DIRECTORY_LEVEL);
- else
- /* For shadow and EPT supported 1GB page */
- return (PT_PDPE_LEVEL);
-}
-
-static inline int
-cpu_has_vmx_rdtscp(void)
-{
- return (vmcs_config.cpu_based_2nd_exec_ctrl & SECONDARY_EXEC_RDTSCP);
-}
-
-static int
-vmx_rdtscp_supported(void)
-{
- return (cpu_has_vmx_rdtscp());
-}
#define F(x) bit(X86_FEATURE_##x)
@@ -4514,72 +3317,6 @@ out:
return (r);
}
-#define __ex(x) __kvm_handle_fault_on_reboot(x)
-
-void
-vmcs_clear(uint64_t vmcs_pa)
-{
- unsigned char error;
-
- /*CSTYLED*/
- __asm__ volatile (__ex(ASM_VMX_VMCLEAR_RAX) "\n\tsetna %0\n"
- : "=g"(error) : "a"(&vmcs_pa), "m"(vmcs_pa)
- : "cc", "memory");
-
- if (error)
- cmn_err(CE_PANIC, "kvm: vmclear fail: %lx\n",
- vmcs_pa);
-}
-
-void
-__vcpu_clear(void *arg)
-{
- struct vcpu_vmx *vmx = arg;
- int cpu = CPU->cpu_id;
-
- vmx->vmcs->revision_id = vmcs_config.revision_id;
-
- if (vmx->vcpu.cpu == cpu)
- vmcs_clear(vmx->vmcs_pa);
-
- if (current_vmcs[cpu] == vmx->vmcs)
- current_vmcs[cpu] = NULL;
- rdtscll(vmx->vcpu.arch.host_tsc);
-
- list_remove(vcpus_on_cpu[cpu], vmx);
-
- vmx->vcpu.cpu = -1;
- vmx->launched = 0;
-}
-
-static void
-vcpu_clear(struct vcpu_vmx *vmx)
-{
- if (vmx->vcpu.cpu == -1)
- return;
-
- /*
- * XXX: commented out below?
- *
- * smp_call_function_single(vmx->vcpu.cpu, __vcpu_clear, vmx, 1);
- */
- kvm_xcall(vmx->vcpu.cpu, __vcpu_clear, vmx);
-}
-
-
-uint16_t
-vmcs_read16(unsigned long field)
-{
- return (vmcs_readl(field));
-}
-
-static void
-vmwrite_error(unsigned long field, unsigned long value)
-{
- cmn_err(CE_WARN, "vmwrite error: reg %lx value %lx (err %x)\n",
- field, value, vmcs_read32(VM_INSTRUCTION_ERROR));
-}
-
static inline void
__vmwrite(unsigned long field, unsigned long value)
{
@@ -4604,80 +3341,6 @@ __vmwrite(unsigned long field, unsigned long value)
}
}
-void
-vmcs_writel(unsigned long field, unsigned long value)
-{
- unsigned char error = 0;
-#ifndef XXX
- /*CSTYLED*/
- __asm__ volatile (ASM_VMX_VMWRITE_RAX_RDX "\n\tsetna %0"
- : "=q"(error) : "a"(value), "d"(field) : "cc");
-
- if ((error))
- vmwrite_error(field, value);
-#else
- XXX_KVM_PROBE;
- __vmwrite(field, value);
-#endif
-}
-
-unsigned long
-vmcs_readl(unsigned long field)
-{
- unsigned long value;
-
- /*CSTYLED*/
- __asm__ volatile (ASM_VMX_VMREAD_RDX_RAX
- : "=a"(value) : "d"(field) : "cc");
-
- return (value);
-}
-
-uint64_t
-vmcs_read64(unsigned long field)
-{
-#ifdef CONFIG_X86_64
- return (vmcs_readl(field));
-#else
- return (vmcs_readl(field) | ((uint64_t)vmcs_readl(field + 1) << 32));
-#endif
-}
-
-void
-vmcs_write16(unsigned long field, uint16_t value)
-{
- vmcs_writel(field, value);
-}
-
-/*
- * writes 'guest_tsc' into guest's timestamp counter "register"
- * guest_tsc = host_tsc + tsc_offset ==> tsc_offset = guest_tsc - host_tsc
- */
-static void
-guest_write_tsc(uint64_t guest_tsc, uint64_t host_tsc)
-{
- vmcs_write64(TSC_OFFSET, guest_tsc - host_tsc);
-}
-
-static inline int
-cpu_has_secondary_exec_ctrls(void)
-{
- return (vmcs_config.cpu_based_exec_ctrl &
- CPU_BASED_ACTIVATE_SECONDARY_CONTROLS);
-}
-
-int
-vm_need_virtualize_apic_accesses(struct kvm *kvm)
-{
- return (flexpriority_enabled && irqchip_in_kernel(kvm));
-}
-
-inline int
-vm_need_tpr_shadow(struct kvm *kvm)
-{
- return ((cpu_has_vmx_tpr_shadow()) && (irqchip_in_kernel(kvm)));
-}
-
/*
* Volatile isn't enough to prevent the compiler from reordering the
* read/write functions for the control registers and messing everything up.
@@ -4687,7 +3350,7 @@ vm_need_tpr_shadow(struct kvm *kvm)
*/
static unsigned long __force_order;
-static inline unsigned long
+unsigned long
native_read_cr0(void)
{
unsigned long val;
@@ -4697,7 +3360,7 @@ native_read_cr0(void)
#define read_cr0() (native_read_cr0())
-static inline unsigned long
+unsigned long
native_read_cr4(void)
{
unsigned long val;
@@ -4707,7 +3370,7 @@ native_read_cr4(void)
#define read_cr4() (native_read_cr4())
-static inline unsigned long
+unsigned long
native_read_cr3(void)
{
unsigned long val;
@@ -4719,240 +3382,13 @@ native_read_cr3(void)
inline ulong kvm_read_cr4(struct kvm_vcpu *vcpu);
-/*
- * Sets up the vmcs for emulated real mode.
- */
-int
-vmx_vcpu_setup(struct vcpu_vmx *vmx)
-{
- uint32_t host_sysenter_cs, msr_low, msr_high;
- uint32_t junk;
- uint64_t host_pat, tsc_this, tsc_base;
- volatile uint64_t a;
- struct descriptor_table dt;
- int i;
- unsigned long kvm_vmx_return;
- uint32_t exec_control;
-
- /* I/O */
- vmcs_write64(IO_BITMAP_A, kvm_va2pa((caddr_t)vmx_io_bitmap_a));
- vmcs_write64(IO_BITMAP_B, kvm_va2pa((caddr_t)vmx_io_bitmap_b));
-
- if (cpu_has_vmx_msr_bitmap()) {
- vmcs_write64(MSR_BITMAP,
- kvm_va2pa((caddr_t)vmx_msr_bitmap_legacy));
- }
-
- vmcs_write64(VMCS_LINK_POINTER, -1ull); /* 22.3.1.5 */
-
- /* Control */
- vmcs_write32(PIN_BASED_VM_EXEC_CONTROL,
- vmcs_config.pin_based_exec_ctrl);
-
- exec_control = vmcs_config.cpu_based_exec_ctrl;
- if (!vm_need_tpr_shadow(vmx->vcpu.kvm)) {
- exec_control &= ~CPU_BASED_TPR_SHADOW;
-#ifdef CONFIG_X86_64
- exec_control |= CPU_BASED_CR8_STORE_EXITING |
- CPU_BASED_CR8_LOAD_EXITING;
-#endif
- }
-
- if (!enable_ept)
- exec_control |= CPU_BASED_CR3_STORE_EXITING |
- CPU_BASED_CR3_LOAD_EXITING |
- CPU_BASED_INVLPG_EXITING;
- vmcs_write32(CPU_BASED_VM_EXEC_CONTROL, exec_control);
-
- if (cpu_has_secondary_exec_ctrls()) {
- exec_control = vmcs_config.cpu_based_2nd_exec_ctrl;
- if (!vm_need_virtualize_apic_accesses(vmx->vcpu.kvm))
- exec_control &=
- ~SECONDARY_EXEC_VIRTUALIZE_APIC_ACCESSES;
- if (vmx->vpid == 0)
- exec_control &= ~SECONDARY_EXEC_ENABLE_VPID;
- if (!enable_ept) {
- exec_control &= ~SECONDARY_EXEC_ENABLE_EPT;
- enable_unrestricted_guest = 0;
- }
- if (!enable_unrestricted_guest)
- exec_control &= ~SECONDARY_EXEC_UNRESTRICTED_GUEST;
-
- if (!ple_gap)
- exec_control &= ~SECONDARY_EXEC_PAUSE_LOOP_EXITING;
- vmcs_write32(SECONDARY_VM_EXEC_CONTROL, exec_control);
- }
-
- if (ple_gap) {
- vmcs_write32(PLE_GAP, ple_gap);
- vmcs_write32(PLE_WINDOW, ple_window);
- }
-
- vmcs_write32(PAGE_FAULT_ERROR_CODE_MASK, !!bypass_guest_pf);
- vmcs_write32(PAGE_FAULT_ERROR_CODE_MATCH, !!bypass_guest_pf);
- vmcs_write32(CR3_TARGET_COUNT, 0); /* 22.2.1 */
-
- vmcs_writel(HOST_CR0, read_cr0()); /* 22.2.3 */
- vmcs_writel(HOST_CR4, read_cr4()); /* 22.2.3, 22.2.5 */
- vmcs_writel(HOST_CR3, read_cr3()); /* 22.2.3 FIXME: shadow tables */
-
- vmcs_write16(HOST_CS_SELECTOR, KCS_SEL); /* 22.2.4 */
-#ifndef XXX
- vmcs_write16(HOST_DS_SELECTOR, KDS_SEL); /* 22.2.4 */
- vmcs_write16(HOST_ES_SELECTOR, KDS_SEL); /* 22.2.4 */
- vmcs_write16(HOST_FS_SELECTOR, kvm_read_fs()); /* 22.2.4 */
- vmcs_write16(HOST_GS_SELECTOR, kvm_read_gs()); /* 22.2.4 */
-
-#else
- XXX_KVM_PROBE;
- vmcs_write16(HOST_DS_SELECTOR, 0x4b); /* 22.2.4 */
- vmcs_write16(HOST_ES_SELECTOR, 0x4b); /* 22.2.4 */
- vmcs_write16(HOST_FS_SELECTOR, 0); /* 22.2.4 */
- vmcs_write16(HOST_GS_SELECTOR, 0); /* 22.2.4 */
-#endif
- vmcs_write16(HOST_SS_SELECTOR, KDS_SEL); /* 22.2.4 */
-#ifdef CONFIG_X86_64
- rdmsrl(MSR_FS_BASE, a);
- vmcs_writel(HOST_FS_BASE, a); /* 22.2.4 */
- rdmsrl(MSR_GS_BASE, a);
- vmcs_writel(HOST_GS_BASE, a); /* 22.2.4 */
-#else
- vmcs_writel(HOST_FS_BASE, 0); /* 22.2.4 */
- vmcs_writel(HOST_GS_BASE, 0); /* 22.2.4 */
-#endif
-
- vmcs_write16(HOST_TR_SELECTOR, KTSS_SEL); /* 22.2.4 */
-
- kvm_get_idt(&dt);
- vmcs_writel(HOST_IDTR_BASE, dt.base); /* 22.2.4 */
-
- __asm__("mov $.Lkvm_vmx_return, %0" : "=r"(kvm_vmx_return));
- vmcs_writel(HOST_RIP, kvm_vmx_return); /* 22.2.5 */
- vmcs_write32(VM_EXIT_MSR_STORE_COUNT, 0);
- vmcs_write32(VM_EXIT_MSR_LOAD_COUNT, 0);
- vmcs_write32(VM_ENTRY_MSR_LOAD_COUNT, 0);
-
- rdmsr(MSR_IA32_SYSENTER_CS, host_sysenter_cs, junk);
- vmcs_write32(HOST_IA32_SYSENTER_CS, host_sysenter_cs);
- rdmsrl(MSR_IA32_SYSENTER_ESP, a);
- vmcs_writel(HOST_IA32_SYSENTER_ESP, a); /* 22.2.3 */
- rdmsrl(MSR_IA32_SYSENTER_EIP, a);
- vmcs_writel(HOST_IA32_SYSENTER_EIP, a); /* 22.2.3 */
-
- if (vmcs_config.vmexit_ctrl & VM_EXIT_LOAD_IA32_PAT) {
- rdmsr(MSR_IA32_CR_PAT, msr_low, msr_high);
- host_pat = msr_low | ((uint64_t) msr_high << 32);
- vmcs_write64(HOST_IA32_PAT, host_pat);
- }
- if (vmcs_config.vmentry_ctrl & VM_ENTRY_LOAD_IA32_PAT) {
- rdmsr(MSR_IA32_CR_PAT, msr_low, msr_high);
- host_pat = msr_low | ((uint64_t) msr_high << 32);
- /* Write the default value follow host pat */
- vmcs_write64(GUEST_IA32_PAT, host_pat);
- /* Keep arch.pat sync with GUEST_IA32_PAT */
- vmx->vcpu.arch.pat = host_pat;
- }
-
- for (i = 0; i < NR_VMX_MSR; ++i) {
- uint32_t index = vmx_msr_index[i];
- uint32_t data_low, data_high;
- int j = vmx->nmsrs;
-
- if (rdmsr_safe(index, &data_low, &data_high) < 0)
- continue;
- if (wrmsr_safe(index, data_low, data_high) < 0)
- continue;
- vmx->guest_msrs[j].index = i;
- vmx->guest_msrs[j].data = 0;
- vmx->guest_msrs[j].mask = -1ull;
- ++vmx->nmsrs;
- }
-
- vmcs_write32(VM_EXIT_CONTROLS, vmcs_config.vmexit_ctrl);
-
- /* 22.2.1, 20.8.1 */
- vmcs_write32(VM_ENTRY_CONTROLS, vmcs_config.vmentry_ctrl);
-
- vmcs_writel(CR0_GUEST_HOST_MASK, ~0UL);
- vmx->vcpu.arch.cr4_guest_owned_bits = KVM_CR4_GUEST_OWNED_BITS;
- if (enable_ept)
- vmx->vcpu.arch.cr4_guest_owned_bits |= X86_CR4_PGE;
- vmcs_writel(CR4_GUEST_HOST_MASK, ~vmx->vcpu.arch.cr4_guest_owned_bits);
-
- tsc_base = vmx->vcpu.kvm->arch.vm_init_tsc;
- rdtscll(tsc_this);
- if (tsc_this < vmx->vcpu.kvm->arch.vm_init_tsc)
- tsc_base = tsc_this;
-
- guest_write_tsc(0, tsc_base);
-
- return (0);
-}
-
-static void kvm_migrate_timers(struct kvm_vcpu *vcpu)
+void
+kvm_migrate_timers(struct kvm_vcpu *vcpu)
{
set_bit(KVM_REQ_MIGRATE_TIMER, &vcpu->requests);
}
-/*
- * Switches to specified vcpu, until a matching vcpu_put(), but assumes
- * vcpu mutex is already taken.
- */
-void
-vmx_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
-{
- struct vcpu_vmx *vmx = to_vmx(vcpu);
- uint64_t phys_addr = vmx->vmcs_pa;
- uint64_t tsc_this, delta, new_offset;
-
- if (vcpu->cpu != cpu) {
- vcpu_clear(vmx);
- kvm_migrate_timers(vcpu);
- set_bit(KVM_REQ_TLB_FLUSH, &vcpu->requests);
- kpreempt_disable();
- list_insert_head(vcpus_on_cpu[cpu], vmx);
- kpreempt_enable();
- }
-
- if (current_vmcs[cpu] != vmx->vmcs) {
- uint8_t error;
-
- current_vmcs[cpu] = vmx->vmcs;
-
- /*CSTYLED*/
- __asm__ volatile (ASM_VMX_VMPTRLD_RAX "; setna %0"
- : "=g"(error) : "a"(&phys_addr), "m"(phys_addr)
- : "cc");
- }
-
- if (vcpu->cpu != cpu) {
- struct descriptor_table dt;
- unsigned long sysenter_esp;
-
- vcpu->cpu = cpu;
-
- /*
- * Linux uses per-cpu TSS and GDT, so set these when switching
- * processors.
- */
- vmcs_writel(HOST_TR_BASE, kvm_read_tr_base()); /* 22.2.4 */
- kvm_get_gdt(&dt);
- vmcs_writel(HOST_GDTR_BASE, dt.base); /* 22.2.4 */
-
- rdmsrl(MSR_IA32_SYSENTER_ESP, sysenter_esp);
- vmcs_writel(HOST_IA32_SYSENTER_ESP, sysenter_esp); /* 22.2.3 */
- /*
- * Make sure the time stamp counter is monotonous.
- */
- rdtscll(tsc_this);
- if (tsc_this < vcpu->arch.host_tsc) {
- delta = vcpu->arch.host_tsc - tsc_this;
- new_offset = vmcs_read64(TSC_OFFSET) + delta;
- vmcs_write64(TSC_OFFSET, new_offset);
- }
- }
-}
static int
kvm_request_guest_time_update(struct kvm_vcpu *v)
@@ -5006,31 +3442,6 @@ ldt_load(void)
wr_ldtr(ULDT_SEL);
}
-
-static void reload_tss(void)
-{
- /*
- * VT restores TR but not its size. Useless.
- */
- struct descriptor_table gdt;
- struct desc_struct *descs;
-
- kvm_get_gdt(&gdt);
- descs = (void *)gdt.base;
- descs[GDT_KTSS].c.b.type = 9; /* available TSS */
- load_TR_desc();
-}
-
-int
-is_long_mode(struct kvm_vcpu *vcpu)
-{
-#ifdef CONFIG_X86_64
- return (vcpu->arch.efer & EFER_LMA);
-#else
- return (0);
-#endif
-}
-
inline ulong
kvm_read_cr4_bits(struct kvm_vcpu *vcpu, ulong mask)
{
@@ -5048,55 +3459,7 @@ is_pae(struct kvm_vcpu *vcpu)
return (kvm_read_cr4_bits(vcpu, X86_CR4_PAE));
}
-static void
-__vmx_load_host_state(struct vcpu_vmx *vmx)
-{
- unsigned long flags;
-
- if (!vmx->host_state.loaded)
- return;
-
- KVM_VCPU_KSTAT_INC(&vmx->vcpu, kvmvs_host_state_reload);
-
- vmx->host_state.loaded = 0;
- if (vmx->host_state.fs_reload_needed)
- kvm_load_fs(vmx->host_state.fs_sel);
- if (vmx->host_state.gs_ldt_reload_needed) {
- kvm_load_ldt(vmx->host_state.ldt_sel);
- /*
- * If we have to reload gs, we must take care to
- * preserve our gs base.
- */
- cli();
- kvm_load_gs(vmx->host_state.gs_sel);
-#ifdef CONFIG_X86_64
- wrmsrl(MSR_GS_BASE, vmcs_readl(HOST_GS_BASE));
-#endif
- sti();
- }
- reload_tss();
-#ifdef CONFIG_X86_64
- if (is_long_mode(&vmx->vcpu)) {
- rdmsrl(MSR_KERNEL_GS_BASE, vmx->msr_guest_kernel_gs_base);
- wrmsrl(MSR_KERNEL_GS_BASE, vmx->msr_host_kernel_gs_base);
- }
-#endif
-}
-
-static void
-vmx_load_host_state(struct vcpu_vmx *vmx)
-{
- kpreempt_disable();
- __vmx_load_host_state(vmx);
- kpreempt_enable();
-}
-
-void
-vmx_vcpu_put(struct kvm_vcpu *vcpu)
-{
- __vmx_load_host_state(to_vmx(vcpu));
-}
void
kvm_arch_vcpu_put(struct kvm_vcpu *vcpu)
@@ -5288,40 +3651,7 @@ kvm_vcpu_ioctl_get_cpuid2(struct kvm_vcpu *vcpu, struct kvm_cpuid2 *cpuid)
return (0);
}
-static void
-ept_save_pdptrs(struct kvm_vcpu *vcpu)
-{
- if (is_paging(vcpu) && is_pae(vcpu) && !is_long_mode(vcpu)) {
- vcpu->arch.pdptrs[0] = vmcs_read64(GUEST_PDPTR0);
- vcpu->arch.pdptrs[1] = vmcs_read64(GUEST_PDPTR1);
- vcpu->arch.pdptrs[2] = vmcs_read64(GUEST_PDPTR2);
- vcpu->arch.pdptrs[3] = vmcs_read64(GUEST_PDPTR3);
- }
-
- __set_bit(VCPU_EXREG_PDPTR, (unsigned long *)&vcpu->arch.regs_avail);
- __set_bit(VCPU_EXREG_PDPTR, (unsigned long *)&vcpu->arch.regs_dirty);
-}
-
-static void
-vmx_cache_reg(struct kvm_vcpu *vcpu, enum kvm_reg reg)
-{
- __set_bit(reg, (unsigned long *)&vcpu->arch.regs_avail);
- switch (reg) {
- case VCPU_REGS_RSP:
- vcpu->arch.regs[VCPU_REGS_RSP] = vmcs_readl(GUEST_RSP);
- break;
- case VCPU_REGS_RIP:
- vcpu->arch.regs[VCPU_REGS_RIP] = vmcs_readl(GUEST_RIP);
- break;
- case VCPU_EXREG_PDPTR:
- if (enable_ept)
- ept_save_pdptrs(vcpu);
- break;
- default:
- break;
- }
-}
unsigned long
kvm_get_rflags(struct kvm_vcpu *vcpu)
@@ -5368,116 +3698,10 @@ kvm_arch_vcpu_ioctl_get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
return (0);
}
-#define VMX_SEGMENT_FIELD(seg) \
- [VCPU_SREG_##seg] = { \
- .selector = GUEST_##seg##_SELECTOR, \
- .base = GUEST_##seg##_BASE, \
- .limit = GUEST_##seg##_LIMIT, \
- .ar_bytes = GUEST_##seg##_AR_BYTES, \
- }
-
-struct kvm_vmx_segment_field kvm_vmx_segment_fields[] = {
- VMX_SEGMENT_FIELD(CS),
- VMX_SEGMENT_FIELD(DS),
- VMX_SEGMENT_FIELD(ES),
- VMX_SEGMENT_FIELD(FS),
- VMX_SEGMENT_FIELD(GS),
- VMX_SEGMENT_FIELD(SS),
- VMX_SEGMENT_FIELD(TR),
- VMX_SEGMENT_FIELD(LDTR),
-};
-
-void vmx_get_segment(struct kvm_vcpu *vcpu,
- struct kvm_segment *var, int seg)
-{
- struct kvm_vmx_segment_field *sf = &kvm_vmx_segment_fields[seg];
- uint32_t ar;
-
- var->base = vmcs_readl(sf->base);
- var->limit = vmcs_read32(sf->limit);
- var->selector = vmcs_read16(sf->selector);
- ar = vmcs_read32(sf->ar_bytes);
-
- if ((ar & AR_UNUSABLE_MASK) && !emulate_invalid_guest_state)
- ar = 0;
- var->type = ar & 15;
- var->s = (ar >> 4) & 1;
- var->dpl = (ar >> 5) & 3;
- var->present = (ar >> 7) & 1;
- var->avl = (ar >> 12) & 1;
- var->l = (ar >> 13) & 1;
- var->db = (ar >> 14) & 1;
- var->g = (ar >> 15) & 1;
- var->unusable = (ar >> 16) & 1;
-}
-
-static uint32_t vmx_segment_access_rights(struct kvm_segment *var)
-{
- uint32_t ar;
-
- if (var->unusable)
- ar = 1 << 16;
- else {
- ar = var->type & 15;
- ar |= (var->s & 1) << 4;
- ar |= (var->dpl & 3) << 5;
- ar |= (var->present & 1) << 7;
- ar |= (var->avl & 1) << 12;
- ar |= (var->l & 1) << 13;
- ar |= (var->db & 1) << 14;
- ar |= (var->g & 1) << 15;
- }
- if (ar == 0) /* a 0 value means unusable */
- ar = AR_UNUSABLE_MASK;
-
- return (ar);
-}
-static void
-vmx_set_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg)
-{
- struct vcpu_vmx *vmx = to_vmx(vcpu);
- struct kvm_vmx_segment_field *sf = &kvm_vmx_segment_fields[seg];
- uint32_t ar;
- if (vmx->rmode.vm86_active && seg == VCPU_SREG_TR) {
- vmx->rmode.tr.selector = var->selector;
- vmx->rmode.tr.base = var->base;
- vmx->rmode.tr.limit = var->limit;
- vmx->rmode.tr.ar = vmx_segment_access_rights(var);
- return;
- }
- vmcs_writel(sf->base, var->base);
- vmcs_write32(sf->limit, var->limit);
- vmcs_write16(sf->selector, var->selector);
- if (vmx->rmode.vm86_active && var->s) {
- /*
- * Hack real-mode segments into vm86 compatibility.
- */
- if (var->base == 0xffff0000 && var->selector == 0xf000)
- vmcs_writel(sf->base, 0xf0000);
- ar = 0xf3;
- } else
- ar = vmx_segment_access_rights(var);
- /*
- * Fix the "Accessed" bit in AR field of segment registers for older
- * qemu binaries.
- * IA32 arch specifies that at the time of processor reset the
- * "Accessed" bit in the AR field of segment registers is 1. And qemu
- * is setting it to 0 in the usedland code. This causes invalid guest
- * state vmexit when "unrestricted guest" mode is turned on.
- * Fix for this setup issue in cpu_reset is being pushed in the qemu
- * tree. Newer qemu binaries with that qemu fix would not need this
- * kvm hack.
- */
- if (enable_unrestricted_guest && (seg != VCPU_SREG_LDTR))
- ar |= 0x1; /* Accessed */
-
- vmcs_write32(sf->ar_bytes, ar);
-}
-
void
kvm_get_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg)
{
@@ -5673,7 +3897,7 @@ kvm_mmu_reset_context(struct kvm_vcpu *vcpu)
return (init_kvm_mmu(vcpu));
}
-static inline void
+inline void
kvm_queue_interrupt(struct kvm_vcpu *vcpu, uint8_t vector, int soft)
{
vcpu->arch.interrupt.pending = 1;
@@ -5770,17 +3994,6 @@ out:
}
static void
-vmx_update_cr8_intercept(struct kvm_vcpu *vcpu, int tpr, int irr)
-{
- if (irr == -1 || tpr < irr) {
- vmcs_write32(TPR_THRESHOLD, 0);
- return;
- }
-
- vmcs_write32(TPR_THRESHOLD, irr);
-}
-
-static void
update_cr8_intercept(struct kvm_vcpu *vcpu)
{
int max_irr, tpr;
@@ -5802,153 +4015,10 @@ update_cr8_intercept(struct kvm_vcpu *vcpu)
kvm_x86_ops->update_cr8_intercept(vcpu, tpr, max_irr);
}
-static int
-__find_msr_index(struct vcpu_vmx *vmx, uint32_t msr)
-{
- int i;
-
- for (i = 0; i < vmx->nmsrs; i++) {
- if (vmx_msr_index[vmx->guest_msrs[i].index] == msr)
- return (i);
- }
-
- return (-1);
-}
-static struct shared_msr_entry *
-find_msr_entry(struct vcpu_vmx *vmx, uint32_t msr)
-{
- int i;
- i = __find_msr_index(vmx, msr);
- if (i >= 0)
- return (&vmx->guest_msrs[i]);
- return (NULL);
-}
-
-/*
- * Swap MSR entry in host/guest MSR entry array.
- */
-static void
-move_msr_up(struct vcpu_vmx *vmx, int from, int to)
-{
- struct shared_msr_entry tmp;
-
- tmp = vmx->guest_msrs[to];
- vmx->guest_msrs[to] = vmx->guest_msrs[from];
- vmx->guest_msrs[from] = tmp;
-}
-
-static int
-update_transition_efer(struct vcpu_vmx *vmx, int efer_offset)
-{
- uint64_t guest_efer;
- uint64_t ignore_bits;
-
- guest_efer = vmx->vcpu.arch.efer;
-
- /*
- * NX is emulated; LMA and LME handled by hardware; SCE meaninless
- * outside long mode
- */
- ignore_bits = EFER_NX | EFER_SCE;
-#ifdef CONFIG_X86_64
- ignore_bits |= EFER_LMA | EFER_LME;
- /* SCE is meaningful only in long mode on Intel */
- if (guest_efer & EFER_LMA)
- ignore_bits &= ~(uint64_t)EFER_SCE;
-#endif
- guest_efer &= ~ignore_bits;
- guest_efer |= host_efer & ignore_bits;
- vmx->guest_msrs[efer_offset].data = guest_efer;
- vmx->guest_msrs[efer_offset].mask = ~ignore_bits;
-
- return (1);
-}
-
-/*
- * Set up the vmcs to automatically save and restore system
- * msrs. Don't touch the 64-bit msrs if the guest is in legacy
- * mode, as fiddling with msrs is very expensive.
- */
-void
-setup_msrs(struct vcpu_vmx *vmx)
-{
- int save_nmsrs, index;
- unsigned long *msr_bitmap;
-
- vmx_load_host_state(vmx);
- save_nmsrs = 0;
-#ifdef CONFIG_X86_64
- if (is_long_mode(&vmx->vcpu)) {
- index = __find_msr_index(vmx, MSR_SYSCALL_MASK);
- if (index >= 0)
- move_msr_up(vmx, index, save_nmsrs++);
- index = __find_msr_index(vmx, MSR_LSTAR);
- if (index >= 0)
- move_msr_up(vmx, index, save_nmsrs++);
- index = __find_msr_index(vmx, MSR_CSTAR);
- if (index >= 0)
- move_msr_up(vmx, index, save_nmsrs++);
- index = __find_msr_index(vmx, MSR_TSC_AUX);
- if (index >= 0 && vmx->rdtscp_enabled)
- move_msr_up(vmx, index, save_nmsrs++);
- /*
- * MSR_K6_STAR is only needed on long mode guests, and only
- * if efer.sce is enabled.
- */
- index = __find_msr_index(vmx, MSR_K6_STAR);
- if ((index >= 0) && (vmx->vcpu.arch.efer & EFER_SCE))
- move_msr_up(vmx, index, save_nmsrs++);
- }
-#endif
- index = __find_msr_index(vmx, MSR_EFER);
- if (index >= 0 && update_transition_efer(vmx, index))
- move_msr_up(vmx, index, save_nmsrs++);
-
- vmx->save_nmsrs = save_nmsrs;
-
- if (cpu_has_vmx_msr_bitmap()) {
- if (is_long_mode(&vmx->vcpu))
- msr_bitmap = vmx_msr_bitmap_longmode;
- else
- msr_bitmap = vmx_msr_bitmap_legacy;
-
- vmcs_write64(MSR_BITMAP, kvm_va2pa((caddr_t)msr_bitmap));
- }
-}
-
-void
-vmx_set_efer(struct kvm_vcpu *vcpu, uint64_t efer)
-{
- struct vcpu_vmx *vmx = to_vmx(vcpu);
- struct shared_msr_entry *msr = find_msr_entry(vmx, MSR_EFER);
-
- if (!msr)
- return;
-
- /*
- * Force kernel_gs_base reloading before EFER changes, as control
- * of this msr depends on is_long_mode().
- */
- vmx_load_host_state(to_vmx(vcpu));
- vcpu->arch.efer = efer;
- if (efer & EFER_LMA) {
- vmcs_write32(VM_ENTRY_CONTROLS,
- vmcs_read32(VM_ENTRY_CONTROLS) | VM_ENTRY_IA32E_MODE);
- msr->data = efer;
- } else {
- vmcs_write32(VM_ENTRY_CONTROLS,
- vmcs_read32(VM_ENTRY_CONTROLS) & ~VM_ENTRY_IA32E_MODE);
-
- msr->data = efer & ~EFER_LME;
- }
-
- setup_msrs(vmx);
-}
-
-static inline int
+inline int
is_protmode(struct kvm_vcpu *vcpu)
{
return (kvm_read_cr0_bits(vcpu, X86_CR0_PE));
@@ -7002,82 +5072,6 @@ static int __msr_io(struct kvm_vcpu *vcpu, struct kvm_msrs *msrs,
return (i);
}
-/*
- * reads and returns guest's timestamp counter "register"
- * guest_tsc = host_tsc + tsc_offset -- 21.3
- */
-static uint64_t
-guest_read_tsc(void)
-{
- uint64_t host_tsc, tsc_offset;
-
- rdtscll(host_tsc);
- tsc_offset = vmcs_read64(TSC_OFFSET);
- return (host_tsc + tsc_offset);
-}
-
-/*
- * Reads an msr value (of 'msr_index') into 'pdata'.
- * Returns 0 on success, non-0 otherwise.
- * Assumes vcpu_load() was already called.
- */
-static int
-vmx_get_msr(struct kvm_vcpu *vcpu, uint32_t msr_index, uint64_t *pdata)
-{
- uint64_t data;
- struct shared_msr_entry *msr;
-
- if (!pdata) {
- cmn_err(CE_WARN, "BUG: get_msr called with NULL pdata\n");
- return (EINVAL);
- }
-
- switch (msr_index) {
-#ifdef CONFIG_X86_64
- case MSR_FS_BASE:
- data = vmcs_readl(GUEST_FS_BASE);
- break;
- case MSR_GS_BASE:
- data = vmcs_readl(GUEST_GS_BASE);
- break;
- case MSR_KERNEL_GS_BASE:
- vmx_load_host_state(to_vmx(vcpu));
- data = to_vmx(vcpu)->msr_guest_kernel_gs_base;
- break;
-#endif
- case MSR_EFER:
- return (kvm_get_msr_common(vcpu, msr_index, pdata));
- case MSR_IA32_TSC:
- data = guest_read_tsc();
- break;
- case MSR_IA32_SYSENTER_CS:
- data = vmcs_read32(GUEST_SYSENTER_CS);
- break;
- case MSR_IA32_SYSENTER_EIP:
- data = vmcs_readl(GUEST_SYSENTER_EIP);
- break;
- case MSR_IA32_SYSENTER_ESP:
- data = vmcs_readl(GUEST_SYSENTER_ESP);
- break;
- case MSR_TSC_AUX:
- if (!to_vmx(vcpu)->rdtscp_enabled)
- return (1);
- /* Otherwise falls through */
- default:
- vmx_load_host_state(to_vmx(vcpu));
- msr = find_msr_entry(to_vmx(vcpu), msr_index);
- if (msr) {
- vmx_load_host_state(to_vmx(vcpu));
- data = msr->data;
- break;
- }
- return (kvm_get_msr_common(vcpu, msr_index, pdata));
- }
-
- *pdata = data;
-
- return (0);
-}
/*
* Reads an msr value (of 'msr_index') into 'pdata'.
@@ -7091,76 +5085,6 @@ kvm_get_msr(struct kvm_vcpu *vcpu, uint32_t msr_index, uint64_t *pdata)
}
-/*
- * Writes msr value into into the appropriate "register".
- * Returns 0 on success, non-0 otherwise.
- * Assumes vcpu_load() was already called.
- */
-static int
-vmx_set_msr(struct kvm_vcpu *vcpu, uint32_t msr_index, uint64_t data)
-{
- struct vcpu_vmx *vmx = to_vmx(vcpu);
- struct shared_msr_entry *msr;
- uint64_t host_tsc;
- int ret = 0;
-
- switch (msr_index) {
- case MSR_EFER:
- vmx_load_host_state(vmx);
- ret = kvm_set_msr_common(vcpu, msr_index, data);
- break;
-#ifdef CONFIG_X86_64
- case MSR_FS_BASE:
- vmcs_writel(GUEST_FS_BASE, data);
- break;
- case MSR_GS_BASE:
- vmcs_writel(GUEST_GS_BASE, data);
- break;
- case MSR_KERNEL_GS_BASE:
- vmx_load_host_state(vmx);
- vmx->msr_guest_kernel_gs_base = data;
- break;
-#endif
- case MSR_IA32_SYSENTER_CS:
- vmcs_write32(GUEST_SYSENTER_CS, data);
- break;
- case MSR_IA32_SYSENTER_EIP:
- vmcs_writel(GUEST_SYSENTER_EIP, data);
- break;
- case MSR_IA32_SYSENTER_ESP:
- vmcs_writel(GUEST_SYSENTER_ESP, data);
- break;
- case MSR_IA32_TSC:
- rdtscll(host_tsc);
- guest_write_tsc(data, host_tsc);
- break;
- case MSR_IA32_CR_PAT:
- if (vmcs_config.vmentry_ctrl & VM_ENTRY_LOAD_IA32_PAT) {
- vmcs_write64(GUEST_IA32_PAT, data);
- vcpu->arch.pat = data;
- break;
- }
- ret = kvm_set_msr_common(vcpu, msr_index, data);
- break;
- case MSR_TSC_AUX:
- if (!vmx->rdtscp_enabled)
- return (1);
- /* Check reserved bit, higher 32 bits should be zero */
- if ((data >> 32) != 0)
- return (1);
- /* Otherwise falls through */
- default:
- msr = find_msr_entry(vmx, msr_index);
- if (msr) {
- vmx_load_host_state(vmx);
- msr->data = data;
- break;
- }
- ret = kvm_set_msr_common(vcpu, msr_index, data);
- }
-
- return (ret);
-}
/*
* Writes msr value into into the appropriate "register".
@@ -7182,32 +5106,6 @@ do_set_msr(struct kvm_vcpu *vcpu, unsigned index, uint64_t *data)
return (kvm_set_msr(vcpu, index, *data));
}
-static inline int
-is_machine_check(uint32_t intr_info)
-{
- return (intr_info & (INTR_INFO_INTR_TYPE_MASK | INTR_INFO_VECTOR_MASK |
- INTR_INFO_VALID_MASK)) == (INTR_TYPE_HARD_EXCEPTION |
- MC_VECTOR | INTR_INFO_VALID_MASK);
-}
-
-/*
- * Trigger machine check on the host. We assume all the MSRs are already set up
- * by the CPU and that we still run on the same CPU as the MCE occurred on.
- * We pass a fake environment to the machine check handler because we want
- * the guest to be always treated like user space, no matter what context
- * it used internally.
- */
-static void kvm_machine_check(void)
-{
-#if defined(CONFIG_X86_MCE) && defined(CONFIG_X86_64)
- struct pt_regs regs = {
- .cs = 3, /* Fake ring 3 no matter what the guest ran on */
- .flags = X86_EFLAGS_IF,
- };
-
- do_machine_check(&regs, 0);
-#endif
-}
#define EXCPT_BENIGN 0
#define EXCPT_CONTRIBUTORY 1
@@ -7286,316 +5184,18 @@ kvm_queue_exception_e(struct kvm_vcpu *vcpu, unsigned nr, uint32_t error_code)
kvm_multiple_exception(vcpu, nr, 1, error_code);
}
-static inline void
+inline void
kvm_clear_exception_queue(struct kvm_vcpu *vcpu)
{
vcpu->arch.exception.pending = 0;
}
-static inline void
+inline void
kvm_clear_interrupt_queue(struct kvm_vcpu *vcpu)
{
vcpu->arch.interrupt.pending = 0;
}
-static void
-vmx_complete_interrupts(struct vcpu_vmx *vmx)
-{
- uint32_t exit_intr_info;
- uint32_t idt_vectoring_info = vmx->idt_vectoring_info;
- int unblock_nmi;
- uint8_t vector;
- int type;
- int idtv_info_valid;
-
- exit_intr_info = vmcs_read32(VM_EXIT_INTR_INFO);
-
- vmx->exit_reason = vmcs_read32(VM_EXIT_REASON);
-
- /* Handle machine checks before interrupts are enabled */
- if ((vmx->exit_reason == EXIT_REASON_MCE_DURING_VMENTRY) ||
- (vmx->exit_reason == EXIT_REASON_EXCEPTION_NMI &&
- is_machine_check(exit_intr_info)))
- kvm_machine_check();
-
- /* We need to handle NMIs before interrupts are enabled */
- if ((exit_intr_info & INTR_INFO_INTR_TYPE_MASK) == INTR_TYPE_NMI_INTR &&
- (exit_intr_info & INTR_INFO_VALID_MASK))
- __asm__("int $2");
-
- idtv_info_valid = idt_vectoring_info & VECTORING_INFO_VALID_MASK;
-
- if (cpu_has_virtual_nmis()) {
- unblock_nmi = (exit_intr_info & INTR_INFO_UNBLOCK_NMI) != 0;
- vector = exit_intr_info & INTR_INFO_VECTOR_MASK;
- /*
- * SDM 3: 27.7.1.2 (September 2008)
- * Re-set bit "block by NMI" before VM entry if vmexit caused by
- * a guest IRET fault.
- * SDM 3: 23.2.2 (September 2008)
- * Bit 12 is undefined in any of the following cases:
- * If the VM exit sets the valid bit in the IDT-vectoring
- * information field.
- * If the VM exit is due to a double fault.
- */
- if ((exit_intr_info & INTR_INFO_VALID_MASK) && unblock_nmi &&
- vector != DF_VECTOR && !idtv_info_valid)
- vmcs_set_bits(GUEST_INTERRUPTIBILITY_INFO,
- GUEST_INTR_STATE_NMI);
- } else if (vmx->soft_vnmi_blocked) {
-#ifdef XXX
- vmx->vnmi_blocked_time +=
- ktime_to_ns(ktime_sub(ktime_get(), vmx->entry_time));
-#else
- vmx->vnmi_blocked_time +=
- gethrtime() - vmx->entry_time;
- XXX_KVM_PROBE;
-#endif
- }
-
- vmx->vcpu.arch.nmi_injected = 0;
- kvm_clear_exception_queue(&vmx->vcpu);
- kvm_clear_interrupt_queue(&vmx->vcpu);
-
- if (!idtv_info_valid)
- return;
-
- vector = idt_vectoring_info & VECTORING_INFO_VECTOR_MASK;
- type = idt_vectoring_info & VECTORING_INFO_TYPE_MASK;
-
- switch (type) {
- case INTR_TYPE_NMI_INTR:
- vmx->vcpu.arch.nmi_injected = 1;
- /*
- * SDM 3: 27.7.1.2 (September 2008)
- * Clear bit "block by NMI" before VM entry if a NMI
- * delivery faulted.
- */
- vmcs_clear_bits(GUEST_INTERRUPTIBILITY_INFO,
- GUEST_INTR_STATE_NMI);
- break;
- case INTR_TYPE_SOFT_EXCEPTION:
- vmx->vcpu.arch.event_exit_inst_len =
- vmcs_read32(VM_EXIT_INSTRUCTION_LEN);
- /* fall through */
- case INTR_TYPE_HARD_EXCEPTION:
- if (idt_vectoring_info & VECTORING_INFO_DELIVER_CODE_MASK) {
- uint32_t err = vmcs_read32(IDT_VECTORING_ERROR_CODE);
- kvm_queue_exception_e(&vmx->vcpu, vector, err);
- } else
- kvm_queue_exception(&vmx->vcpu, vector);
- break;
- case INTR_TYPE_SOFT_INTR:
- vmx->vcpu.arch.event_exit_inst_len =
- vmcs_read32(VM_EXIT_INSTRUCTION_LEN);
- /* fall through */
- case INTR_TYPE_EXT_INTR:
- kvm_queue_interrupt(&vmx->vcpu, vector,
- type == INTR_TYPE_SOFT_INTR);
- break;
- default:
- break;
- }
-}
-
-#ifdef CONFIG_X86_64
-#define R "r"
-#define Q "q"
-#else
-#define R "e"
-#define Q "l"
-#endif
-
-/*
- * Failure to inject an interrupt should give us the information
- * in IDT_VECTORING_INFO_FIELD. However, if the failure occurs
- * when fetching the interrupt redirection bitmap in the real-mode
- * tss, this doesn't happen. So we do it ourselves.
- */
-static void
-fixup_rmode_irq(struct vcpu_vmx *vmx)
-{
- vmx->rmode.irq.pending = 0;
- if (kvm_rip_read(&vmx->vcpu) + 1 != vmx->rmode.irq.rip)
- return;
-
- kvm_rip_write(&vmx->vcpu, vmx->rmode.irq.rip);
- if (vmx->idt_vectoring_info & VECTORING_INFO_VALID_MASK) {
- vmx->idt_vectoring_info &= ~VECTORING_INFO_TYPE_MASK;
- vmx->idt_vectoring_info |= INTR_TYPE_EXT_INTR;
- return;
- }
-
- vmx->idt_vectoring_info = VECTORING_INFO_VALID_MASK |
- INTR_TYPE_EXT_INTR | vmx->rmode.irq.vector;
-}
-
-static void
-vmx_vcpu_run(struct kvm_vcpu *vcpu)
-{
- struct vcpu_vmx *vmx = to_vmx(vcpu);
-
- /* Record the guest's net vcpu time for enforced NMI injections. */
- if (!cpu_has_virtual_nmis() && vmx->soft_vnmi_blocked) {
-#ifdef XXX
- vmx->entry_time = ktime_get();
-#else
- vmx->entry_time = gethrtime();
- XXX_KVM_PROBE;
-#endif
- }
-
- /*
- * Don't enter VMX if guest state is invalid, let the exit handler
- * start emulation until we arrive back to a valid state
- */
- if (vmx->emulation_required && emulate_invalid_guest_state)
- return;
-
- if (test_bit(VCPU_REGS_RSP, (unsigned long *)&vcpu->arch.regs_dirty))
- vmcs_writel(GUEST_RSP, vcpu->arch.regs[VCPU_REGS_RSP]);
- if (test_bit(VCPU_REGS_RIP, (unsigned long *)&vcpu->arch.regs_dirty))
- vmcs_writel(GUEST_RIP, vcpu->arch.regs[VCPU_REGS_RIP]);
-
- DTRACE_PROBE1(kvm__vrun, unsigned long, vcpu->arch.regs[VCPU_REGS_RIP]);
-
- /*
- * When single-stepping over STI and MOV SS, we must clear the
- * corresponding interruptibility bits in the guest state. Otherwise
- * vmentry fails as it then expects bit 14 (BS) in pending debug
- * exceptions being set, but that's not correct for the guest debugging
- * case.
- */
- if (vcpu->guest_debug & KVM_GUESTDBG_SINGLESTEP)
- vmx_set_interrupt_shadow(vcpu, 0);
-
- /*
- * Loading guest fpu may have cleared host cr0.ts
- */
- vmcs_writel(HOST_CR0, read_cr0());
-
- __asm__(
- /* Store host registers */
- "push %%"R"dx; push %%"R"bp;"
- "push %%"R"cx \n\t"
- "cmp %%"R"sp, %c[host_rsp](%0) \n\t"
- "je 1f \n\t"
- "mov %%"R"sp, %c[host_rsp](%0) \n\t"
- __ex(ASM_VMX_VMWRITE_RSP_RDX) "\n\t"
- "1: \n\t"
- /* Reload cr2 if changed */
- "mov %c[cr2](%0), %%"R"ax \n\t"
- "mov %%cr2, %%"R"dx \n\t"
- "cmp %%"R"ax, %%"R"dx \n\t"
- "je 2f \n\t"
- "mov %%"R"ax, %%cr2 \n\t"
- "2: \n\t"
- /* Check if vmlaunch of vmresume is needed */
- "cmpl $0, %c[launched](%0) \n\t"
- /* Load guest registers. Don't clobber flags. */
- "mov %c[rax](%0), %%"R"ax \n\t"
- "mov %c[rbx](%0), %%"R"bx \n\t"
- "mov %c[rdx](%0), %%"R"dx \n\t"
- "mov %c[rsi](%0), %%"R"si \n\t"
- "mov %c[rdi](%0), %%"R"di \n\t"
- "mov %c[rbp](%0), %%"R"bp \n\t"
-#ifdef CONFIG_X86_64
- "mov %c[r8](%0), %%r8 \n\t"
- "mov %c[r9](%0), %%r9 \n\t"
- "mov %c[r10](%0), %%r10 \n\t"
- "mov %c[r11](%0), %%r11 \n\t"
- "mov %c[r12](%0), %%r12 \n\t"
- "mov %c[r13](%0), %%r13 \n\t"
- "mov %c[r14](%0), %%r14 \n\t"
- "mov %c[r15](%0), %%r15 \n\t"
-#endif
- "mov %c[rcx](%0), %%"R"cx \n\t" /* kills %0 (ecx) */
-
- /* Enter guest mode */
- "jne .Llaunched \n\t"
- __ex(ASM_VMX_VMLAUNCH) "\n\t"
- "jmp .Lkvm_vmx_return \n\t"
- ".Llaunched: " __ex(ASM_VMX_VMRESUME) "\n\t"
- ".Lkvm_vmx_return: "
- /* Save guest registers, load host registers, keep flags */
- "xchg %0, (%%"R"sp) \n\t"
- "mov %%"R"ax, %c[rax](%0) \n\t"
- "mov %%"R"bx, %c[rbx](%0) \n\t"
- "push"Q" (%%"R"sp); pop"Q" %c[rcx](%0) \n\t"
- "mov %%"R"dx, %c[rdx](%0) \n\t"
- "mov %%"R"si, %c[rsi](%0) \n\t"
- "mov %%"R"di, %c[rdi](%0) \n\t"
- "mov %%"R"bp, %c[rbp](%0) \n\t"
-#ifdef CONFIG_X86_64
- "mov %%r8, %c[r8](%0) \n\t"
- "mov %%r9, %c[r9](%0) \n\t"
- "mov %%r10, %c[r10](%0) \n\t"
- "mov %%r11, %c[r11](%0) \n\t"
- "mov %%r12, %c[r12](%0) \n\t"
- "mov %%r13, %c[r13](%0) \n\t"
- "mov %%r14, %c[r14](%0) \n\t"
- "mov %%r15, %c[r15](%0) \n\t"
-#endif
- "mov %%cr2, %%"R"ax \n\t"
- "mov %%"R"ax, %c[cr2](%0) \n\t"
-
- "pop %%"R"bp; pop %%"R"bp; pop %%"R"dx \n\t"
- "setbe %c[fail](%0) \n\t"
- : : "c"(vmx), "d"((unsigned long)HOST_RSP),
- [launched]"i"(offsetof(struct vcpu_vmx, launched)),
- [fail]"i"(offsetof(struct vcpu_vmx, fail)),
- [host_rsp]"i"(offsetof(struct vcpu_vmx, host_rsp)),
- [rax]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_RAX])),
- [rbx]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_RBX])),
- [rcx]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_RCX])),
- [rdx]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_RDX])),
- [rsi]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_RSI])),
- [rdi]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_RDI])),
- [rbp]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_RBP])),
-#ifdef CONFIG_X86_64
- [r8]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R8])),
- [r9]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R9])),
- [r10]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R10])),
- [r11]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R11])),
- [r12]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R12])),
- [r13]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R13])),
- [r14]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R14])),
- [r15]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R15])),
-#endif
- [cr2]"i"(offsetof(struct vcpu_vmx, vcpu.arch.cr2))
- : "cc", "memory"
- /*CSTYLED*/
- , R"bx", R"di", R"si"
-#ifdef CONFIG_X86_64
- /*CSTYLED*/
- , "r8", "r9", "r10", "r11", "r12", "r13", "r14", "r15"
-#endif
- /*CSTYLED*/
- );
-
- vcpu->arch.regs_avail = ~((1 << VCPU_REGS_RIP) |
- (1 << VCPU_REGS_RSP) | (1 << VCPU_EXREG_PDPTR));
- vcpu->arch.regs_dirty = 0;
-
- vmx->idt_vectoring_info = vmcs_read32(IDT_VECTORING_INFO_FIELD);
-
- if (vmx->rmode.irq.pending)
- fixup_rmode_irq(vmx);
-
-#ifdef XXX
- __asm__("mov %0, %%ds; mov %0, %%es" :
- : "r"SEL_GDT(GDT_UDATA, SEL_UPL));
-#else
- XXX_KVM_PROBE;
- __asm__("mov %0, %%ds; mov %0, %%es" : : "r"KDS_SEL);
-#endif
- vmx->launched = 1;
-
- vmx_complete_interrupts(vmx);
-}
-
-#undef R
-#undef Q
static void kvm_on_user_return(struct kvm_vcpu *,
struct kvm_user_return_notifier *);
@@ -7649,65 +5249,6 @@ kvm_set_shared_msr(struct kvm_vcpu *vcpu, unsigned slot, uint64_t value,
}
}
-static void
-vmx_save_host_state(struct kvm_vcpu *vcpu)
-{
- struct vcpu_vmx *vmx = to_vmx(vcpu);
- int i;
-
- if (vmx->host_state.loaded)
- return;
-
- vmx->host_state.loaded = 1;
- /*
- * Set host fs and gs selectors. Unfortunately, 22.2.3 does not
- * allow segment selectors with cpl > 0 or ti == 1.
- */
- vmx->host_state.ldt_sel = kvm_read_ldt();
- vmx->host_state.gs_ldt_reload_needed = vmx->host_state.ldt_sel;
- vmx->host_state.fs_sel = kvm_read_fs();
- if (!(vmx->host_state.fs_sel & 7)) {
- vmcs_write16(HOST_FS_SELECTOR, vmx->host_state.fs_sel);
- vmx->host_state.fs_reload_needed = 0;
- } else {
- vmcs_write16(HOST_FS_SELECTOR, 0);
- vmx->host_state.fs_reload_needed = 1;
- }
- vmx->host_state.gs_sel = kvm_read_gs();
- if (!(vmx->host_state.gs_sel & 7))
- vmcs_write16(HOST_GS_SELECTOR, vmx->host_state.gs_sel);
- else {
- vmcs_write16(HOST_GS_SELECTOR, 0);
- vmx->host_state.gs_ldt_reload_needed = 1;
- }
-
-#ifdef CONFIG_X86_64
- vmcs_writel(HOST_FS_BASE, read_msr(MSR_FS_BASE));
- vmcs_writel(HOST_GS_BASE, read_msr(MSR_GS_BASE));
-#else
- vmcs_writel(HOST_FS_BASE, segment_base(vmx->host_state.fs_sel));
- vmcs_writel(HOST_GS_BASE, segment_base(vmx->host_state.gs_sel));
-#endif
-
-#ifdef CONFIG_X86_64
- if (is_long_mode(&vmx->vcpu)) {
- rdmsrl(MSR_KERNEL_GS_BASE, vmx->msr_host_kernel_gs_base);
- wrmsrl(MSR_KERNEL_GS_BASE, vmx->msr_guest_kernel_gs_base);
- }
-#endif
- for (i = 0; i < vmx->save_nmsrs; i++) {
- kvm_set_shared_msr(vcpu, vmx->guest_msrs[i].index,
- vmx->guest_msrs[i].data, vmx->guest_msrs[i].mask);
- }
-}
-
-int
-vmx_interrupt_allowed(struct kvm_vcpu *vcpu)
-{
- return ((vmcs_readl(GUEST_RFLAGS) & X86_EFLAGS_IF) &&
- !(vmcs_read32(GUEST_INTERRUPTIBILITY_INFO) &
- (GUEST_INTR_STATE_STI | GUEST_INTR_STATE_MOV_SS)));
-}
int
kvm_arch_interrupt_allowed(struct kvm_vcpu *vcpu)
@@ -7715,20 +5256,7 @@ kvm_arch_interrupt_allowed(struct kvm_vcpu *vcpu)
return (kvm_x86_ops->interrupt_allowed(vcpu));
}
-static int
-handle_machine_check(struct kvm_vcpu *vcpu)
-{
- /* already handled by vcpu_run */
- return (1);
-}
-static inline int
-is_page_fault(uint32_t intr_info)
-{
- return ((intr_info & (INTR_INFO_INTR_TYPE_MASK | INTR_INFO_VECTOR_MASK |
- INTR_INFO_VALID_MASK)) == (INTR_TYPE_HARD_EXCEPTION |
- PF_VECTOR | INTR_INFO_VALID_MASK));
-}
static int
kvm_read_guest_virt_helper(gva_t addr, void *val, unsigned int bytes,
@@ -7778,17 +5306,6 @@ kvm_read_guest_virt_system(gva_t addr, void *val, unsigned int bytes,
return (kvm_read_guest_virt_helper(addr, val, bytes, vcpu, 0, error));
}
-static int vmx_get_cpl(struct kvm_vcpu *vcpu)
-{
- if (!is_protmode(vcpu))
- return (0);
-
- if (vmx_get_rflags(vcpu) & X86_EFLAGS_VM) /* if virtual 8086 */
- return (3);
-
- return (vmcs_read16(GUEST_CS_SELECTOR) & 3);
-}
-
/* used for instruction fetching */
static int
kvm_fetch_guest_virt(gva_t addr, void *val, unsigned int bytes,
@@ -8556,91 +6073,14 @@ kvm_mmu_page_fault(struct kvm_vcpu *vcpu, gva_t cr2, uint32_t error_code)
return (0);
}
-static inline int
-is_no_device(uint32_t intr_info)
-{
- return ((intr_info & (INTR_INFO_INTR_TYPE_MASK | INTR_INFO_VECTOR_MASK |
- INTR_INFO_VALID_MASK)) == (INTR_TYPE_HARD_EXCEPTION | NM_VECTOR |
- INTR_INFO_VALID_MASK));
-}
-
-static inline int
-is_invalid_opcode(uint32_t intr_info)
-{
- return ((intr_info & (INTR_INFO_INTR_TYPE_MASK | INTR_INFO_VECTOR_MASK |
- INTR_INFO_VALID_MASK)) == (INTR_TYPE_HARD_EXCEPTION | UD_VECTOR |
- INTR_INFO_VALID_MASK));
-}
-
-static inline int
-is_external_interrupt(uint32_t intr_info)
-{
- return ((intr_info & (INTR_INFO_INTR_TYPE_MASK |
- INTR_INFO_VALID_MASK)) == (INTR_TYPE_EXT_INTR |
- INTR_INFO_VALID_MASK));
-}
-static inline int
+int
kvm_event_needs_reinjection(struct kvm_vcpu *vcpu)
{
return (vcpu->arch.exception.pending || vcpu->arch.interrupt.pending ||
vcpu->arch.nmi_injected);
}
-static int
-handle_rmode_exception(struct kvm_vcpu *vcpu, int vec, uint32_t err_code)
-{
- /*
- * Instruction with address size override prefix opcode 0x67
- * Cause the #SS fault with 0 error code in VM86 mode.
- */
- if (((vec == GP_VECTOR) || (vec == SS_VECTOR)) && err_code == 0) {
- if (emulate_instruction(vcpu, 0, 0, 0) == EMULATE_DONE)
- return (1);
- }
-
- /*
- * Forward all other exceptions that are valid in real mode.
- * FIXME: Breaks guest debugging in real mode, needs to be fixed with
- * the required debugging infrastructure rework.
- */
- switch (vec) {
- case DB_VECTOR:
- if (vcpu->guest_debug &
- (KVM_GUESTDBG_SINGLESTEP | KVM_GUESTDBG_USE_HW_BP)) {
- return (0);
- }
-
- kvm_queue_exception(vcpu, vec);
- return (1);
-
- case BP_VECTOR:
- /*
- * Update instruction length as we may reinject the exception
- * from user space while in guest debugging mode.
- */
- to_vmx(vcpu)->vcpu.arch.event_exit_inst_len =
- vmcs_read32(VM_EXIT_INSTRUCTION_LEN);
-
- if (vcpu->guest_debug & KVM_GUESTDBG_USE_SW_BP)
- return (0);
- /* fall through */
-
- case DE_VECTOR:
- case OF_VECTOR:
- case BR_VECTOR:
- case UD_VECTOR:
- case DF_VECTOR:
- case SS_VECTOR:
- case GP_VECTOR:
- case MF_VECTOR:
- kvm_queue_exception(vcpu, vec);
- return (1);
- }
-
- return (0);
-}
-
int
kvm_emulate_halt(struct kvm_vcpu *vcpu)
{
@@ -8655,128 +6095,8 @@ kvm_emulate_halt(struct kvm_vcpu *vcpu)
}
}
-/*
- * The guest has exited. See if we can fix it or if we need userspace
- * assistance.
- */
-static int
-handle_exception(struct kvm_vcpu *vcpu)
-{
- struct vcpu_vmx *vmx = to_vmx(vcpu);
- struct kvm_run *kvm_run = vcpu->run;
- uint32_t intr_info, ex_no, error_code;
- unsigned long cr2, rip, dr6;
- uint32_t vect_info;
- enum emulation_result er;
-
- vect_info = vmx->idt_vectoring_info;
- intr_info = vmcs_read32(VM_EXIT_INTR_INFO);
-
- if (is_machine_check(intr_info))
- return (handle_machine_check(vcpu));
-
- if ((vect_info & VECTORING_INFO_VALID_MASK) &&
- !is_page_fault(intr_info)) {
- vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
- vcpu->run->internal.suberror = KVM_INTERNAL_ERROR_SIMUL_EX;
- vcpu->run->internal.ndata = 2;
- vcpu->run->internal.data[0] = vect_info;
- vcpu->run->internal.data[1] = intr_info;
- return (0);
- }
-
- if ((intr_info & INTR_INFO_INTR_TYPE_MASK) == INTR_TYPE_NMI_INTR)
- return (1); /* already handled by vmx_vcpu_run() */
-
- if (is_no_device(intr_info)) {
- vmx_fpu_activate(vcpu);
- return (1);
- }
-
- if (is_invalid_opcode(intr_info)) {
- er = emulate_instruction(vcpu, 0, 0, EMULTYPE_TRAP_UD);
- if (er != EMULATE_DONE)
- kvm_queue_exception(vcpu, UD_VECTOR);
- return (1);
- }
-
- error_code = 0;
- rip = kvm_rip_read(vcpu);
-
- if (intr_info & INTR_INFO_DELIVER_CODE_MASK)
- error_code = vmcs_read32(VM_EXIT_INTR_ERROR_CODE);
-
- if (is_page_fault(intr_info)) {
- /* EPT won't cause page fault directly */
- if (enable_ept)
- cmn_err(CE_PANIC, "page fault with ept enabled\n");
- cr2 = vmcs_readl(EXIT_QUALIFICATION);
-
- KVM_TRACE2(page__fault, uintptr_t, cr2, uint32_t, error_code);
-
- if (kvm_event_needs_reinjection(vcpu))
- kvm_mmu_unprotect_page_virt(vcpu, cr2);
- return (kvm_mmu_page_fault(vcpu, cr2, error_code));
- }
-
- if (vmx->rmode.vm86_active && handle_rmode_exception(vcpu,
- intr_info & INTR_INFO_VECTOR_MASK, error_code)) {
- if (vcpu->arch.halt_request) {
- vcpu->arch.halt_request = 0;
- return (kvm_emulate_halt(vcpu));
- }
- return (1);
- }
- ex_no = intr_info & INTR_INFO_VECTOR_MASK;
- switch (ex_no) {
- case DB_VECTOR:
- dr6 = vmcs_readl(EXIT_QUALIFICATION);
- if (!(vcpu->guest_debug &
- (KVM_GUESTDBG_SINGLESTEP | KVM_GUESTDBG_USE_HW_BP))) {
- vcpu->arch.dr6 = dr6 | DR6_FIXED_1;
- kvm_queue_exception(vcpu, DB_VECTOR);
- return (1);
- }
-
- kvm_run->debug.arch.dr6 = dr6 | DR6_FIXED_1;
- kvm_run->debug.arch.dr7 = vmcs_readl(GUEST_DR7);
- /* fall through */
- case BP_VECTOR:
- /*
- * Update instruction length as we may reinject #BP from
- * user space while in guest debugging mode. Reading it for
- * #DB as well causes no harm, it is not used in that case.
- */
- vmx->vcpu.arch.event_exit_inst_len =
- vmcs_read32(VM_EXIT_INSTRUCTION_LEN);
- kvm_run->exit_reason = KVM_EXIT_DEBUG;
- kvm_run->debug.arch.pc = vmcs_readl(GUEST_CS_BASE) + rip;
- kvm_run->debug.arch.exception = ex_no;
- break;
- default:
- kvm_run->exit_reason = KVM_EXIT_EXCEPTION;
- kvm_run->ex.exception = ex_no;
- kvm_run->ex.error_code = error_code;
- break;
- }
-
- return (0);
-}
-static int
-handle_external_interrupt(struct kvm_vcpu *vcpu)
-{
- KVM_VCPU_KSTAT_INC(vcpu, kvmvs_irq_exits);
- return (1);
-}
-
-static int
-handle_triple_fault(struct kvm_vcpu *vcpu)
-{
- vcpu->run->exit_reason = KVM_EXIT_SHUTDOWN;
- return (0);
-}
static int
kvm_write_guest_virt(gva_t addr, void *val, unsigned int bytes,
@@ -8928,291 +6248,6 @@ kvm_emulate_pio(struct kvm_vcpu *vcpu, int in, int size, unsigned port)
return (0);
}
-static int
-handle_io(struct kvm_vcpu *vcpu)
-{
- unsigned long exit_qualification;
- int size, in, string;
- unsigned port;
-
- KVM_VCPU_KSTAT_INC(vcpu, kvmvs_io_exits);
-
- exit_qualification = vmcs_readl(EXIT_QUALIFICATION);
- string = (exit_qualification & 16) != 0;
-
- if (string) {
- if (emulate_instruction(vcpu, 0, 0, 0) == EMULATE_DO_MMIO)
- return (0);
- return (1);
- }
-
- size = (exit_qualification & 7) + 1;
- in = (exit_qualification & 8) != 0;
- port = exit_qualification >> 16;
- skip_emulated_instruction(vcpu);
-
- return (kvm_emulate_pio(vcpu, in, size, port));
-}
-
-static int
-handle_nmi_window(struct kvm_vcpu *vcpu)
-{
- uint32_t cpu_based_vm_exec_control;
-
- /* clear pending NMI */
- cpu_based_vm_exec_control = vmcs_read32(CPU_BASED_VM_EXEC_CONTROL);
- cpu_based_vm_exec_control &= ~CPU_BASED_VIRTUAL_NMI_PENDING;
- vmcs_write32(CPU_BASED_VM_EXEC_CONTROL, cpu_based_vm_exec_control);
-
- KVM_VCPU_KSTAT_INC(vcpu, kvmvs_nmi_window_exits);
-
- return (1);
-}
-
-static int
-code_segment_valid(struct kvm_vcpu *vcpu)
-{
- struct kvm_segment cs;
- unsigned int cs_rpl;
-
- vmx_get_segment(vcpu, &cs, VCPU_SREG_CS);
- cs_rpl = cs.selector & SELECTOR_RPL_MASK;
-
- if (cs.unusable)
- return (0);
- if (~cs.type & (AR_TYPE_CODE_MASK|AR_TYPE_ACCESSES_MASK))
- return (0);
- if (!cs.s)
- return (0);
-
- if (cs.type & AR_TYPE_WRITEABLE_MASK) {
- if (cs.dpl > cs_rpl)
- return (0);
- } else {
- if (cs.dpl != cs_rpl)
- return (0);
- }
-
- if (!cs.present)
- return (0);
-
- /*
- * TODO: Add Reserved field check, this'll require a new member in the
- * kvm_segment_field structure
- */
- return (1);
-}
-
-static int
-data_segment_valid(struct kvm_vcpu *vcpu, int seg)
-{
- struct kvm_segment var;
- unsigned int rpl;
-
- vmx_get_segment(vcpu, &var, seg);
- rpl = var.selector & SELECTOR_RPL_MASK;
-
- if (var.unusable)
- return (1);
-
- if (!var.s)
- return (0);
-
- if (!var.present)
- return (0);
-
- if (~var.type & (AR_TYPE_CODE_MASK|AR_TYPE_WRITEABLE_MASK)) {
- if (var.dpl < rpl) /* DPL < RPL */
- return (0);
- }
-
- /*
- * TODO: Add other members to kvm_segment_field to allow checking for
- * other access rights flags
- */
- return (1);
-}
-
-static int
-ldtr_valid(struct kvm_vcpu *vcpu)
-{
- struct kvm_segment ldtr;
-
- vmx_get_segment(vcpu, &ldtr, VCPU_SREG_LDTR);
-
- if (ldtr.unusable)
- return (1);
- if (ldtr.selector & SELECTOR_TI_MASK) /* TI = 1 */
- return (0);
- if (ldtr.type != 2)
- return (0);
- if (!ldtr.present)
- return (0);
-
- return (1);
-}
-
-static int
-tr_valid(struct kvm_vcpu *vcpu)
-{
- struct kvm_segment tr;
-
- vmx_get_segment(vcpu, &tr, VCPU_SREG_TR);
-
- if (tr.unusable)
- return (0);
- if (tr.selector & SELECTOR_TI_MASK) /* TI = 1 */
- return (0);
- if (tr.type != 3 && tr.type != 11)
- return (0); /* TODO: Check if guest is in IA32e mode */
- if (!tr.present)
- return (0);
-
- return (1);
-}
-
-static int
-cs_ss_rpl_check(struct kvm_vcpu *vcpu)
-{
- struct kvm_segment cs, ss;
-
- vmx_get_segment(vcpu, &cs, VCPU_SREG_CS);
- vmx_get_segment(vcpu, &ss, VCPU_SREG_SS);
-
- return ((cs.selector & SELECTOR_RPL_MASK) ==
- (ss.selector & SELECTOR_RPL_MASK));
-}
-
-static int
-rmode_segment_valid(struct kvm_vcpu *vcpu, int seg)
-{
- struct kvm_segment var;
- uint32_t ar;
-
- vmx_get_segment(vcpu, &var, seg);
- ar = vmx_segment_access_rights(&var);
-
- if (var.base != (var.selector << 4))
- return (0);
- if (var.limit != 0xffff)
- return (0);
- if (ar != 0xf3)
- return (0);
-
- return (1);
-}
-
-static int
-stack_segment_valid(struct kvm_vcpu *vcpu)
-{
- struct kvm_segment ss;
- unsigned int ss_rpl;
-
- vmx_get_segment(vcpu, &ss, VCPU_SREG_SS);
- ss_rpl = ss.selector & SELECTOR_RPL_MASK;
-
- if (ss.unusable)
- return (1);
- if (ss.type != 3 && ss.type != 7)
- return (0);
- if (!ss.s)
- return (0);
- if (ss.dpl != ss_rpl) /* DPL != RPL */
- return (0);
- if (!ss.present)
- return (0);
-
- return (1);
-}
-
-/*
- * Check if guest state is valid. Returns true if valid, false if
- * not.
- * We assume that registers are always usable
- */
-static int
-guest_state_valid(struct kvm_vcpu *vcpu)
-{
- if (!is_protmode(vcpu)) {
- /* real mode guest state checks */
- if (!rmode_segment_valid(vcpu, VCPU_SREG_CS))
- return (0);
- if (!rmode_segment_valid(vcpu, VCPU_SREG_SS))
- return (0);
- if (!rmode_segment_valid(vcpu, VCPU_SREG_DS))
- return (0);
- if (!rmode_segment_valid(vcpu, VCPU_SREG_ES))
- return (0);
- if (!rmode_segment_valid(vcpu, VCPU_SREG_FS))
- return (0);
- if (!rmode_segment_valid(vcpu, VCPU_SREG_GS))
- return (0);
- } else {
- /* protected mode guest state checks */
- if (!cs_ss_rpl_check(vcpu))
- return (0);
- if (!code_segment_valid(vcpu))
- return (0);
- if (!stack_segment_valid(vcpu))
- return (0);
- if (!data_segment_valid(vcpu, VCPU_SREG_DS))
- return (0);
- if (!data_segment_valid(vcpu, VCPU_SREG_ES))
- return (0);
- if (!data_segment_valid(vcpu, VCPU_SREG_FS))
- return (0);
- if (!data_segment_valid(vcpu, VCPU_SREG_GS))
- return (0);
- if (!tr_valid(vcpu))
- return (0);
- if (!ldtr_valid(vcpu))
- return (0);
- }
-
- /*
- * TODO:
- * - Add checks on RIP
- * - Add checks on RFLAGS
- */
-
- return (1);
-}
-
-static int handle_invalid_guest_state(struct kvm_vcpu *vcpu)
-{
- struct vcpu_vmx *vmx = to_vmx(vcpu);
- enum emulation_result err = EMULATE_DONE;
- int ret = 1;
-
- while (!guest_state_valid(vcpu)) {
- err = emulate_instruction(vcpu, 0, 0, 0);
-
- if (err == EMULATE_DO_MMIO) {
- ret = 0;
- goto out;
- }
-
- if (err != EMULATE_DONE) {
- vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
- vcpu->run->internal.suberror =
- KVM_INTERNAL_ERROR_EMULATION;
- vcpu->run->internal.ndata = 0;
- ret = 0;
- goto out;
- }
-
-#ifdef XXX
- if ((current))
- goto out;
-#else
- XXX_KVM_PROBE;
-#endif
- }
-
- vmx->emulation_required = 0;
-out:
- return (ret);
-}
void
kvm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0)
@@ -9376,107 +6411,7 @@ kvm_lmsw(struct kvm_vcpu *vcpu, unsigned long msw)
kvm_set_cr0(vcpu, kvm_read_cr0_bits(vcpu, ~0x0ful) | (msw & 0x0f));
}
-static int
-handle_cr(struct kvm_vcpu *vcpu)
-{
- unsigned long exit_qualification, val;
- int cr;
- int reg;
-
- exit_qualification = vmcs_readl(EXIT_QUALIFICATION);
- cr = exit_qualification & 15;
- reg = (exit_qualification >> 8) & 15;
- DTRACE_PROBE3(kvm__cr, int, cr, int, reg, int,
- (exit_qualification >> 4) & 3);
- switch ((exit_qualification >> 4) & 3) {
- case 0: /* mov to cr */
- val = kvm_register_read(vcpu, reg);
- KVM_TRACE2(cr__write, int, cr, unsigned long, val);
-
- switch (cr) {
- case 0:
- kvm_set_cr0(vcpu, val);
- skip_emulated_instruction(vcpu);
- return (1);
- case 3:
- kvm_set_cr3(vcpu, val);
- skip_emulated_instruction(vcpu);
- return (1);
- case 4:
- kvm_set_cr4(vcpu, val);
- skip_emulated_instruction(vcpu);
- return (1);
- case 8: {
- uint8_t cr8_prev = kvm_get_cr8(vcpu);
- uint8_t cr8 = kvm_register_read(vcpu, reg);
- kvm_set_cr8(vcpu, cr8);
- skip_emulated_instruction(vcpu);
-
- if (irqchip_in_kernel(vcpu->kvm))
- return (1);
-
- if (cr8_prev <= cr8)
- return (1);
-
- vcpu->run->exit_reason = KVM_EXIT_SET_TPR;
- return (0);
- }
- };
-
- break;
-
- case 2: /* clts */
- vmx_set_cr0(vcpu, kvm_read_cr0_bits(vcpu, ~X86_CR0_TS));
- KVM_TRACE2(cr__write, int, 0,
- unsigned long, kvm_read_cr0(vcpu));
-
- skip_emulated_instruction(vcpu);
- vmx_fpu_activate(vcpu);
- return (1);
- case 1: /* mov from cr */
- switch (cr) {
- case 3:
- kvm_register_write(vcpu, reg, vcpu->arch.cr3);
- KVM_TRACE2(cr__read, int, cr,
- unsigned long, vcpu->arch.cr3);
- skip_emulated_instruction(vcpu);
- return (1);
- case 8:
- val = kvm_get_cr8(vcpu);
- kvm_register_write(vcpu, reg, val);
- KVM_TRACE2(cr__read, int, cr, unsigned long, val);
- skip_emulated_instruction(vcpu);
- return (1);
- }
- break;
- case 3: /* lmsw */
- val = (exit_qualification >> LMSW_SOURCE_DATA_SHIFT) & 0x0f;
- KVM_TRACE2(cr__write, int, 0, unsigned long,
- (kvm_read_cr0(vcpu) & ~0xful) | val);
- kvm_lmsw(vcpu, val);
-
- skip_emulated_instruction(vcpu);
- return (1);
- default:
- break;
- }
- vcpu->run->exit_reason = 0;
- cmn_err(CE_WARN, "unhandled control register: op %d cr %d\n",
- (int)(exit_qualification >> 4) & 3, cr);
-
- return (0);
-}
-
-static int
-check_dr_alias(struct kvm_vcpu *vcpu)
-{
- if (kvm_read_cr4_bits(vcpu, X86_CR4_DE)) {
- kvm_queue_exception(vcpu, UD_VECTOR);
- return (-1);
- }
- return (0);
-}
/*
* Checks if cpl <= required_cpl; if true, return true. Otherwise queue
@@ -9491,108 +6426,6 @@ kvm_require_cpl(struct kvm_vcpu *vcpu, int required_cpl)
return (0);
}
-static int handle_dr(struct kvm_vcpu *vcpu)
-{
- unsigned long exit_qualification;
- unsigned long val;
- int dr, reg;
-
- /* Do not handle if the CPL > 0, will trigger GP on re-entry */
- if (!kvm_require_cpl(vcpu, 0))
- return (1);
-
- dr = vmcs_readl(GUEST_DR7);
-
- if (dr & DR7_GD) {
- /*
- * As the vm-exit takes precedence over the debug trap, we
- * need to emulate the latter, either for the host or the
- * guest debugging itself.
- */
- if (vcpu->guest_debug & KVM_GUESTDBG_USE_HW_BP) {
- vcpu->run->debug.arch.dr6 = vcpu->arch.dr6;
- vcpu->run->debug.arch.dr7 = dr;
- vcpu->run->debug.arch.pc =
- vmcs_readl(GUEST_CS_BASE) +
- vmcs_readl(GUEST_RIP);
- vcpu->run->debug.arch.exception = DB_VECTOR;
- vcpu->run->exit_reason = KVM_EXIT_DEBUG;
- return (0);
- } else {
- vcpu->arch.dr7 &= ~DR7_GD;
- vcpu->arch.dr6 |= DR6_BD;
- vmcs_writel(GUEST_DR7, vcpu->arch.dr7);
- kvm_queue_exception(vcpu, DB_VECTOR);
- return (1);
- }
- }
-
- exit_qualification = vmcs_readl(EXIT_QUALIFICATION);
- dr = exit_qualification & DEBUG_REG_ACCESS_NUM;
- reg = DEBUG_REG_ACCESS_REG(exit_qualification);
- if (exit_qualification & TYPE_MOV_FROM_DR) {
- switch (dr) {
- case 0 ... 3:
- val = vcpu->arch.db[dr];
- break;
- case 4:
- if (check_dr_alias(vcpu) < 0)
- return (1);
- /* fall through */
- case 6:
- val = vcpu->arch.dr6;
- break;
- case 5:
- if (check_dr_alias(vcpu) < 0)
- return (1);
- /* fall through */
- default: /* 7 */
- val = vcpu->arch.dr7;
- break;
- }
- kvm_register_write(vcpu, reg, val);
- } else {
- val = vcpu->arch.regs[reg];
- switch (dr) {
- case 0 ... 3:
- vcpu->arch.db[dr] = val;
- if (!(vcpu->guest_debug & KVM_GUESTDBG_USE_HW_BP))
- vcpu->arch.eff_db[dr] = val;
- break;
- case 4:
- if (check_dr_alias(vcpu) < 0)
- return (1);
- /* fall through */
- case 6:
- if (val & 0xffffffff00000000ULL) {
- kvm_inject_gp(vcpu, 0);
- return (1);
- }
- vcpu->arch.dr6 = (val & DR6_VOLATILE) | DR6_FIXED_1;
- break;
- case 5:
- if (check_dr_alias(vcpu) < 0)
- return (1);
- /* fall through */
- default: /* 7 */
- if (val & 0xffffffff00000000ULL) {
- kvm_inject_gp(vcpu, 0);
- return (1);
- }
- vcpu->arch.dr7 = (val & DR7_VOLATILE) | DR7_FIXED_1;
-
- if (!(vcpu->guest_debug & KVM_GUESTDBG_USE_HW_BP)) {
- vmcs_writel(GUEST_DR7, vcpu->arch.dr7);
- vcpu->arch.switch_db_regs =
- (val & DR7_BP_EN_MASK);
-
- }
- break;
- }
- }
- skip_emulated_instruction(vcpu);
- return (1);
-}
void
kvm_emulate_cpuid(struct kvm_vcpu *vcpu)
@@ -9623,58 +6456,6 @@ kvm_emulate_cpuid(struct kvm_vcpu *vcpu)
}
static int
-handle_cpuid(struct kvm_vcpu *vcpu)
-{
- kvm_emulate_cpuid(vcpu);
- return (1);
-}
-
-static int
-handle_rdmsr(struct kvm_vcpu *vcpu)
-{
- uint32_t ecx = vcpu->arch.regs[VCPU_REGS_RCX];
- uint64_t data;
-
- if (vmx_get_msr(vcpu, ecx, &data)) {
- KVM_TRACE1(msr__read__ex, uint32_t, ecx);
- kvm_inject_gp(vcpu, 0);
- return (1);
- }
-
- KVM_TRACE2(msr__read, uint32_t, ecx, uint64_t, data);
-
- /* FIXME: handling of bits 32:63 of rax, rdx */
- vcpu->arch.regs[VCPU_REGS_RAX] = data & -1u;
- vcpu->arch.regs[VCPU_REGS_RDX] = (data >> 32) & -1u;
- skip_emulated_instruction(vcpu);
- return (1);
-}
-
-static int
-handle_wrmsr(struct kvm_vcpu *vcpu)
-{
- uint32_t ecx = vcpu->arch.regs[VCPU_REGS_RCX];
- uint64_t data = (vcpu->arch.regs[VCPU_REGS_RAX] & -1u) |
- ((uint64_t)(vcpu->arch.regs[VCPU_REGS_RDX] & -1u) << 32);
-
- if (vmx_set_msr(vcpu, ecx, data) != 0) {
- KVM_TRACE2(msr__write__ex, uint32_t, ecx, uint64_t, data);
- kvm_inject_gp(vcpu, 0);
- return (1);
- }
-
- KVM_TRACE2(msr__write, uint32_t, ecx, uint64_t, data);
- skip_emulated_instruction(vcpu);
- return (1);
-}
-
-static int
-handle_tpr_below_threshold(struct kvm_vcpu *vcpu)
-{
- return (1);
-}
-
-static int
kvm_hv_hypercall_enabled(struct kvm *kvm)
{
return (kvm->arch.hv_hypercall & HV_X64_MSR_HYPERCALL_ENABLE);
@@ -9850,27 +6631,6 @@ out:
return (r);
}
-static int
-handle_halt(struct kvm_vcpu *vcpu)
-{
- skip_emulated_instruction(vcpu);
- return (kvm_emulate_halt(vcpu));
-}
-
-static int
-handle_vmcall(struct kvm_vcpu *vcpu)
-{
- skip_emulated_instruction(vcpu);
- kvm_emulate_hypercall(vcpu);
- return (1);
-}
-
-static int
-handle_vmx_insn(struct kvm_vcpu *vcpu)
-{
- kvm_queue_exception(vcpu, UD_VECTOR);
- return (1);
-}
void
kvm_mmu_invlpg(struct kvm_vcpu *vcpu, gva_t gva)
@@ -9880,43 +6640,7 @@ kvm_mmu_invlpg(struct kvm_vcpu *vcpu, gva_t gva)
KVM_VCPU_KSTAT_INC(vcpu, kvmvs_invlpg);
}
-static int
-handle_invlpg(struct kvm_vcpu *vcpu)
-{
- unsigned long exit_qualification = vmcs_readl(EXIT_QUALIFICATION);
-
- kvm_mmu_invlpg(vcpu, exit_qualification);
- skip_emulated_instruction(vcpu);
- return (1);
-}
-
-static int
-handle_wbinvd(struct kvm_vcpu *vcpu)
-{
- skip_emulated_instruction(vcpu);
- /* TODO: Add support for VT-d/pass-through device */
- return (1);
-}
-
-static int
-handle_apic_access(struct kvm_vcpu *vcpu)
-{
- unsigned long exit_qualification;
- enum emulation_result er;
- unsigned long offset;
-
- exit_qualification = vmcs_readl(EXIT_QUALIFICATION);
- offset = exit_qualification & 0xffful;
-
- er = emulate_instruction(vcpu, 0, 0, 0);
- if (er != EMULATE_DONE) {
- cmn_err(CE_PANIC, "Fail to handle apic access vmexit! "
- "Offset is 0x%lx\n", offset);
- }
-
- return (1);
-}
static int
is_vm86_segment(struct kvm_vcpu *vcpu, int seg)
@@ -10422,14 +7146,6 @@ out:
return (ret);
}
-static uint64_t
-vmx_get_segment_base(struct kvm_vcpu *vcpu, int seg)
-{
- struct kvm_vmx_segment_field *sf = &kvm_vmx_segment_fields[seg];
-
- return (vmcs_readl(sf->base));
-}
-
int
kvm_task_switch(struct kvm_vcpu *vcpu, uint16_t tss_selector, int reason)
{
@@ -10515,91 +7231,6 @@ out:
return (ret);
}
-static int
-handle_task_switch(struct kvm_vcpu *vcpu)
-{
- struct vcpu_vmx *vmx = to_vmx(vcpu);
- unsigned long exit_qualification;
- uint16_t tss_selector;
- int reason, type, idt_v;
-
- idt_v = (vmx->idt_vectoring_info & VECTORING_INFO_VALID_MASK);
- type = (vmx->idt_vectoring_info & VECTORING_INFO_TYPE_MASK);
- exit_qualification = vmcs_readl(EXIT_QUALIFICATION);
-
- reason = (uint32_t)exit_qualification >> 30;
- if (reason == TASK_SWITCH_GATE && idt_v) {
- switch (type) {
- case INTR_TYPE_NMI_INTR:
- vcpu->arch.nmi_injected = 0;
- if (cpu_has_virtual_nmis()) {
- vmcs_set_bits(GUEST_INTERRUPTIBILITY_INFO,
- GUEST_INTR_STATE_NMI);
- }
- break;
- case INTR_TYPE_EXT_INTR:
- case INTR_TYPE_SOFT_INTR:
- kvm_clear_interrupt_queue(vcpu);
- break;
- case INTR_TYPE_HARD_EXCEPTION:
- case INTR_TYPE_SOFT_EXCEPTION:
- kvm_clear_exception_queue(vcpu);
- break;
- default:
- break;
- }
- }
- tss_selector = exit_qualification;
-
- if (!idt_v || (type != INTR_TYPE_HARD_EXCEPTION &&
- type != INTR_TYPE_EXT_INTR && type != INTR_TYPE_NMI_INTR))
- skip_emulated_instruction(vcpu);
-
- if (!kvm_task_switch(vcpu, tss_selector, reason))
- return (0);
-
- /* clear all local breakpoint enable flags */
- vmcs_writel(GUEST_DR7, vmcs_readl(GUEST_DR7) & ~55);
-
- /*
- * TODO: What about debug traps on tss switch?
- * Are we supposed to inject them and update dr6?
- */
-
- return (1);
-}
-
-static int
-handle_ept_violation(struct kvm_vcpu *vcpu)
-{
- unsigned long exit_qualification;
- gpa_t gpa;
- int gla_validity;
-
- exit_qualification = vmcs_readl(EXIT_QUALIFICATION);
-
- if (exit_qualification & (1 << 6)) {
- cmn_err(CE_PANIC, "EPT: GPA exceeds GAW!\n");
- }
-
- gla_validity = (exit_qualification >> 7) & 0x3;
- if (gla_validity != 0x3 && gla_validity != 0x1 && gla_validity != 0) {
- cmn_err(CE_WARN, "EPT: Handling EPT violation failed!\n");
- cmn_err(CE_CONT, "EPT: GPA: 0x%lx, GVA: 0x%lx\n",
- (long unsigned int)vmcs_read64(GUEST_PHYSICAL_ADDRESS),
- vmcs_readl(GUEST_LINEAR_ADDRESS));
- cmn_err(CE_PANIC, "EPT: Exit qualification is 0x%lx\n",
- (long unsigned int)exit_qualification);
- vcpu->run->exit_reason = KVM_EXIT_UNKNOWN;
- vcpu->run->hw.hardware_exit_reason = EXIT_REASON_EPT_VIOLATION;
- return (0);
- }
-
- gpa = vmcs_read64(GUEST_PHYSICAL_ADDRESS);
- KVM_TRACE2(page__fault, gpa_t, gpa, unsigned long, exit_qualification);
-
- return (kvm_mmu_page_fault(vcpu, gpa & PAGEMASK, 0));
-}
int
kvm_mmu_get_spte_hierarchy(struct kvm_vcpu *vcpu,
@@ -10620,129 +7251,8 @@ kvm_mmu_get_spte_hierarchy(struct kvm_vcpu *vcpu,
return (nr_sptes);
}
-/* XXX - The following assumes we're running on the maximum sized box... */
-#define MAX_PHYSMEM_BITS 46
-static uint64_t ept_rsvd_mask(uint64_t spte, int level)
-{
- int i;
- uint64_t mask = 0;
-#ifdef XXX
- for (i = 51; i > boot_cpu_data.x86_phys_bits; i--)
- mask |= (1ULL << i);
-#else
- XXX_KVM_PROBE;
- for (i = 51; i > MAX_PHYSMEM_BITS; i--)
- mask |= (1ULL << i);
-#endif
-
- if (level > 2)
- /* bits 7:3 reserved */
- mask |= 0xf8;
- else if (level == 2) {
- if (spte & (1ULL << 7))
- /* 2MB ref, bits 20:12 reserved */
- mask |= 0x1ff000;
- else
- /* bits 6:3 reserved */
- mask |= 0x78;
- }
-
- return (mask);
-}
-
-static inline int
-cpu_has_vmx_ept_execute_only(void)
-{
- return (!!(vmx_capability.ept & VMX_EPT_EXECUTE_ONLY_BIT));
-}
-
-static void
-ept_misconfig_inspect_spte(struct kvm_vcpu *vcpu, uint64_t spte, int level)
-{
- cmn_err(CE_WARN, "%s: spte 0x%lx level %d\n", __func__, spte, level);
-
- /* 010b (write-only) */
- if ((spte & 0x7) == 0x2)
- cmn_err(CE_CONT, "%s: spte is write-only\n", __func__);
-
- /* 110b (write/execute) */
- if ((spte & 0x7) == 0x6)
- cmn_err(CE_CONT, "%s: spte is write-execute\n", __func__);
-
- /* 100b (execute-only) and value not supported by logical processor */
- if (!cpu_has_vmx_ept_execute_only()) {
- if ((spte & 0x7) == 0x4)
- cmn_err(CE_CONT,
- "%s: spte is execute-only\n", __func__);
- }
-
- /* not 000b */
- if ((spte & 0x7)) {
- uint64_t rsvd_bits = spte & ept_rsvd_mask(spte, level);
-
- if (rsvd_bits != 0) {
- cmn_err(CE_CONT, "%s: rsvd_bits = 0x%lx\n",
- __func__, rsvd_bits);
- }
-
- if (level == 1 || (level == 2 && (spte & (1ULL << 7)))) {
- uint64_t ept_mem_type = (spte & 0x38) >> 3;
-
- if (ept_mem_type == 2 || ept_mem_type == 3 ||
- ept_mem_type == 7) {
- cmn_err(CE_CONT, "%s: ept_mem_type=0x%lx\n",
- __func__, ept_mem_type);
- }
- }
- }
-}
-
-static int
-handle_ept_misconfig(struct kvm_vcpu *vcpu)
-{
- uint64_t sptes[4];
- int nr_sptes, i;
- gpa_t gpa;
-
- gpa = vmcs_read64(GUEST_PHYSICAL_ADDRESS);
-
- cmn_err(CE_WARN, "EPT: Misconfiguration.\n");
- cmn_err(CE_CONT, "EPT: GPA: 0x%lx\n", gpa);
- nr_sptes = kvm_mmu_get_spte_hierarchy(vcpu, gpa, sptes);
-
- for (i = PT64_ROOT_LEVEL; i > PT64_ROOT_LEVEL - nr_sptes; --i)
- ept_misconfig_inspect_spte(vcpu, sptes[i-1], i);
-
- vcpu->run->exit_reason = KVM_EXIT_UNKNOWN;
- vcpu->run->hw.hardware_exit_reason = EXIT_REASON_EPT_MISCONFIG;
-
- return (0);
-}
-
-/*
- * Indicate a busy-waiting vcpu in spinlock. We do not enable the PAUSE
- * exiting, so only get here on cpu with PAUSE-Loop-Exiting.
- */
-static int
-handle_pause(struct kvm_vcpu *vcpu)
-{
- skip_emulated_instruction(vcpu);
-#ifdef XXX
- kvm_vcpu_on_spin(vcpu);
-#else
- XXX_KVM_PROBE;
-#endif
- return (1);
-}
-
-static int
-handle_invalid_op(struct kvm_vcpu *vcpu)
-{
- kvm_queue_exception(vcpu, UD_VECTOR);
- return (1);
-}
/*
* check if there is pending interrupt without intack.
@@ -10765,150 +7275,7 @@ kvm_cpu_has_interrupt(struct kvm_vcpu *v)
return (1);
}
-static int
-handle_interrupt_window(struct kvm_vcpu *vcpu)
-{
- uint32_t cpu_based_vm_exec_control;
-
- /* clear pending irq */
- cpu_based_vm_exec_control = vmcs_read32(CPU_BASED_VM_EXEC_CONTROL);
- cpu_based_vm_exec_control &= ~CPU_BASED_VIRTUAL_INTR_PENDING;
- vmcs_write32(CPU_BASED_VM_EXEC_CONTROL, cpu_based_vm_exec_control);
-
- KVM_VCPU_KSTAT_INC(vcpu, kvmvs_irq_window_exits);
-
- /*
- * If the user space waits to inject interrupts, exit as soon as
- * possible
- */
- if (!irqchip_in_kernel(vcpu->kvm) &&
- vcpu->run->request_interrupt_window &&
- !kvm_cpu_has_interrupt(vcpu)) {
- vcpu->run->exit_reason = KVM_EXIT_IRQ_WINDOW_OPEN;
- return (0);
- }
- return (1);
-}
-
-/*
- * The exit handlers return 1 if the exit was handled fully and guest execution
- * may resume. Otherwise they set the kvm_run parameter to indicate what needs
- * to be done to userspace and return 0.
- */
-static int (*kvm_vmx_exit_handlers[])(struct kvm_vcpu *vcpu) = {
- [EXIT_REASON_EXCEPTION_NMI] = handle_exception,
- [EXIT_REASON_EXTERNAL_INTERRUPT] = handle_external_interrupt,
- [EXIT_REASON_TRIPLE_FAULT] = handle_triple_fault,
- [EXIT_REASON_NMI_WINDOW] = handle_nmi_window,
- [EXIT_REASON_IO_INSTRUCTION] = handle_io,
- [EXIT_REASON_CR_ACCESS] = handle_cr,
- [EXIT_REASON_DR_ACCESS] = handle_dr,
- [EXIT_REASON_CPUID] = handle_cpuid,
- [EXIT_REASON_MSR_READ] = handle_rdmsr,
- [EXIT_REASON_MSR_WRITE] = handle_wrmsr,
- [EXIT_REASON_PENDING_INTERRUPT] = handle_interrupt_window,
- [EXIT_REASON_HLT] = handle_halt,
- [EXIT_REASON_INVLPG] = handle_invlpg,
- [EXIT_REASON_VMCALL] = handle_vmcall,
- [EXIT_REASON_VMCLEAR] = handle_vmx_insn,
- [EXIT_REASON_VMLAUNCH] = handle_vmx_insn,
- [EXIT_REASON_VMPTRLD] = handle_vmx_insn,
- [EXIT_REASON_VMPTRST] = handle_vmx_insn,
- [EXIT_REASON_VMREAD] = handle_vmx_insn,
- [EXIT_REASON_VMRESUME] = handle_vmx_insn,
- [EXIT_REASON_VMWRITE] = handle_vmx_insn,
- [EXIT_REASON_VMOFF] = handle_vmx_insn,
- [EXIT_REASON_VMON] = handle_vmx_insn,
- [EXIT_REASON_TPR_BELOW_THRESHOLD] = handle_tpr_below_threshold,
- [EXIT_REASON_APIC_ACCESS] = handle_apic_access,
- [EXIT_REASON_WBINVD] = handle_wbinvd,
- [EXIT_REASON_TASK_SWITCH] = handle_task_switch,
- [EXIT_REASON_MCE_DURING_VMENTRY] = handle_machine_check,
- [EXIT_REASON_EPT_VIOLATION] = handle_ept_violation,
- [EXIT_REASON_EPT_MISCONFIG] = handle_ept_misconfig,
- [EXIT_REASON_PAUSE_INSTRUCTION] = handle_pause,
- [EXIT_REASON_MWAIT_INSTRUCTION] = handle_invalid_op,
- [EXIT_REASON_MONITOR_INSTRUCTION] = handle_invalid_op,
-};
-
-static const int kvm_vmx_max_exit_handlers =
- ARRAY_SIZE(kvm_vmx_exit_handlers);
-
-/*
- * The guest has exited. See if we can fix it or if we need userspace
- * assistance.
- */
-static int
-vmx_handle_exit(struct kvm_vcpu *vcpu)
-{
- struct vcpu_vmx *vmx = to_vmx(vcpu);
- uint32_t exit_reason = vmx->exit_reason;
- uint32_t vectoring_info = vmx->idt_vectoring_info;
- int rval;
- unsigned long rip;
-
- /* Always read the guest rip when exiting */
- rip = vmcs_readl(GUEST_RIP);
- DTRACE_PROBE2(kvm__vexit, unsigned long, rip, uint32_t, exit_reason);
-
- /* If guest state is invalid, start emulating */
- if (vmx->emulation_required && emulate_invalid_guest_state)
- return (handle_invalid_guest_state(vcpu));
-
- /*
- * Access CR3 don't cause VMExit in paging mode, so we need
- * to sync with guest real CR3.
- */
- if (enable_ept && is_paging(vcpu))
- vcpu->arch.cr3 = vmcs_readl(GUEST_CR3);
-
- if (vmx->fail) {
- vcpu->run->exit_reason = KVM_EXIT_FAIL_ENTRY;
- vcpu->run->fail_entry.hardware_entry_failure_reason
- = vmcs_read32(VM_INSTRUCTION_ERROR);
-
- return (0);
- }
-
- if ((vectoring_info & VECTORING_INFO_VALID_MASK) &&
- (exit_reason != EXIT_REASON_EXCEPTION_NMI &&
- exit_reason != EXIT_REASON_EPT_VIOLATION &&
- exit_reason != EXIT_REASON_TASK_SWITCH)) {
- cmn_err(CE_WARN, "%s: unexpected, valid vectoring info "
- "(0x%x) and exit reason is 0x%x\n",
- __func__, vectoring_info, exit_reason);
- }
-
- if (!cpu_has_virtual_nmis() && vmx->soft_vnmi_blocked) {
- if (vmx_interrupt_allowed(vcpu)) {
- vmx->soft_vnmi_blocked = 0;
- } else if (vmx->vnmi_blocked_time > 1000000000LL &&
- vcpu->arch.nmi_pending) {
- /*
- * This CPU don't support us in finding the end of an
- * NMI-blocked window if the guest runs with IRQs
- * disabled. So we pull the trigger after 1 s of
- * futile waiting, but inform the user about this.
- */
- cmn_err(CE_WARN, "%s: Breaking out of NMI-blocked "
- "state on VCPU %d after 1 s timeout\n",
- __func__, vcpu->vcpu_id);
- vmx->soft_vnmi_blocked = 0;
- }
- }
-
- if (exit_reason < kvm_vmx_max_exit_handlers &&
- kvm_vmx_exit_handlers[exit_reason]) {
- rval = kvm_vmx_exit_handlers[exit_reason](vcpu);
- return (rval);
- } else {
- vcpu->run->exit_reason = KVM_EXIT_UNKNOWN;
- vcpu->run->hw.hardware_exit_reason = exit_reason;
- }
-
- return (0);
-}
static inline void
kvm_guest_exit(void)
diff --git a/kvm.h b/kvm.h
index b47b4b6..600c7de 100644
--- a/kvm.h
+++ b/kvm.h
@@ -120,38 +120,6 @@ typedef void (*kvm_xcall_t)(void *);
#define KVM_MAX_MCE_BANKS 32
#define KVM_MCE_CAP_SUPPORTED MCG_CTL_P
-#define KVM_GUEST_CR0_MASK_UNRESTRICTED_GUEST \
- (X86_CR0_WP | X86_CR0_NE | X86_CR0_NW | X86_CR0_CD)
-#define KVM_GUEST_CR0_MASK \
- (KVM_GUEST_CR0_MASK_UNRESTRICTED_GUEST | X86_CR0_PG | X86_CR0_PE)
-#define KVM_VM_CR0_ALWAYS_ON_UNRESTRICTED_GUEST \
- (X86_CR0_WP | X86_CR0_NE)
-#define KVM_VM_CR0_ALWAYS_ON \
- (KVM_VM_CR0_ALWAYS_ON_UNRESTRICTED_GUEST | X86_CR0_PG | X86_CR0_PE)
-#define KVM_CR4_GUEST_OWNED_BITS \
- (X86_CR4_PVI | X86_CR4_DE | X86_CR4_PCE | X86_CR4_OSFXSR \
- | X86_CR4_OSXMMEXCPT)
-
-#define KVM_PMODE_VM_CR4_ALWAYS_ON (X86_CR4_PAE | X86_CR4_VMXE)
-#define KVM_RMODE_VM_CR4_ALWAYS_ON (X86_CR4_VME | X86_CR4_PAE | X86_CR4_VMXE)
-
-#define RMODE_GUEST_OWNED_EFLAGS_BITS (~(X86_EFLAGS_IOPL | X86_EFLAGS_VM))
-
-/*
- * These 2 parameters are used to config the controls for Pause-Loop Exiting:
- * ple_gap: upper bound on the amount of time between two successive
- * executions of PAUSE in a loop. Also indicate if ple enabled.
- * According to test, this time is usually small than 41 cycles.
- * ple_window: upper bound on the amount of time a guest is allowed to execute
- * in a PAUSE loop. Tests indicate that most spinlocks are held for
- * less than 2^12 cycles
- * Time is measured based on a counter that runs at the same rate as the TSC,
- * refer SDM volume 3b section 21.6.13 & 22.1.3.
- */
-#define KVM_VMX_DEFAULT_PLE_GAP 41
-#define KVM_VMX_DEFAULT_PLE_WINDOW 4096
-
-
#ifdef __ASSEMBLY__
# define __IA64_UL(x) (x)
# define __IA64_UL_CONST(x) x
@@ -1639,17 +1607,6 @@ typedef struct kvm_id_map_addr_ioc {
*/
#define KVM_CHECK_EXTENSION _IO(KVMIO, 0x03)
-typedef struct vmcs_config {
- int size;
- int order;
- uint32_t revision_id;
- uint32_t pin_based_exec_ctrl;
- uint32_t cpu_based_exec_ctrl;
- uint32_t cpu_based_2nd_exec_ctrl;
- uint32_t vmexit_ctrl;
- uint32_t vmentry_ctrl;
-} vmcs_config_t;
-
#define RMAP_EXT 4
typedef struct kvm_rmap_desc {
@@ -1658,17 +1615,6 @@ typedef struct kvm_rmap_desc {
} kvm_rmap_desc_t;
-typedef struct vmx_capability {
- uint32_t ept;
- uint32_t vpid;
-} vmx_capability_t;
-
-typedef struct vmcs {
- uint32_t revision_id;
- uint32_t abort;
- char data[1]; /* size is read from MSR */
-} vmcs_t;
-
/* for KVM_INTERRUPT */
typedef struct kvm_interrupt {
/* in */
@@ -1793,62 +1739,7 @@ struct ldttss_desc64 {
typedef struct ldttss_desc64 ldttss_desc64_t;
-typedef struct shared_msr_entry {
- unsigned index;
- uint64_t data;
- uint64_t mask;
-} shared_msr_entry_t;
-
#ifdef _KERNEL
-typedef struct vcpu_vmx {
- struct kvm_vcpu vcpu;
- list_t local_vcpus_link;
- unsigned long host_rsp;
- int launched;
- unsigned char fail;
- uint32_t idt_vectoring_info;
- struct shared_msr_entry *guest_msrs;
- int nmsrs;
- int save_nmsrs;
-#ifdef CONFIG_X86_64
- uint64_t msr_host_kernel_gs_base;
- uint64_t msr_guest_kernel_gs_base;
-#endif
- struct vmcs *vmcs;
- uint64_t vmcs_pa; /* physical address of vmcs for this vmx */
-
- struct {
- int loaded;
- unsigned short fs_sel, gs_sel, ldt_sel;
- int gs_ldt_reload_needed;
- int fs_reload_needed;
- } host_state;
- struct {
- int vm86_active;
- ulong save_rflags;
- struct kvm_save_segment {
- unsigned short selector;
- unsigned long base;
- uint32_t limit;
- uint32_t ar;
- } tr, es, ds, fs, gs;
- struct {
- char pending;
- unsigned char vector;
- unsigned rip;
- } irq;
- } rmode;
- int vpid;
- char emulation_required;
-
- /* Support for vnmi-less CPUs */
- int soft_vnmi_blocked;
- time_t entry_time;
- int64_t vnmi_blocked_time;
- uint32_t exit_reason;
-
- char rdtscp_enabled;
-} vcpu_vmx_t;
#define kvm_for_each_vcpu(idx, vcpup, kvm) \
for (idx = 0, vcpup = kvm_get_vcpu(kvm, idx); \
diff --git a/kvm_ioapic.c b/kvm_ioapic.c
index 0679da8..c8e6e3e 100644
--- a/kvm_ioapic.c
+++ b/kvm_ioapic.c
@@ -468,5 +468,5 @@ kvm_set_ioapic(struct kvm *kvm, struct kvm_ioapic_state *state)
struct kvm_ioapic *
ioapic_irqchip(struct kvm *kvm)
{
- return kvm->arch.vioapic;
+ return (kvm->arch.vioapic);
}
diff --git a/kvm_subr.c b/kvm_subr.c
index ac20a46..2fd2fca 100644
--- a/kvm_subr.c
+++ b/kvm_subr.c
@@ -206,3 +206,19 @@ kvm_xcall(processorid_t cpu, kvm_xcall_t func, void *arg)
(xc_func_t) kvm_xcall_func);
kpreempt_enable();
}
+
+uint32_t
+bit(int bitno)
+{
+ return (1 << (bitno & 31));
+}
+
+int
+is_long_mode(struct kvm_vcpu *vcpu)
+{
+#ifdef CONFIG_X86_64
+ return (vcpu->arch.efer & EFER_LMA);
+#else
+ return (0);
+#endif
+}
diff --git a/kvm_vmx.c b/kvm_vmx.c
new file mode 100644
index 0000000..0dcdb9d
--- /dev/null
+++ b/kvm_vmx.c
@@ -0,0 +1,4741 @@
+/*
+ * Kernel-based Virtual Machine driver for Linux
+ *
+ * This module enables machines with Intel VT-x extensions to run virtual
+ * machines without emulation or binary translation.
+ *
+ * Copyright (C) 2006 Qumranet, Inc.
+ *
+ * Authors:
+ * Avi Kivity <avi@qumranet.com>
+ * Yaniv Kamay <yaniv@qumranet.com>
+ *
+ * This work is licensed under the terms of the GNU GPL, version 2. See
+ * the COPYING file in the top-level directory.
+ *
+ * Copyright 2011 Joyent, Inc. All Rights Reserved.
+ *
+ */
+
+#include <sys/types.h>
+#include <sys/mach_mmu.h>
+
+/*
+ * XXX Need proper header files!
+ */
+#include "processor-flags.h"
+#include "msr.h"
+#include "irqflags.h"
+#include "kvm_host.h"
+#include "kvm_x86host.h"
+#include "iodev.h"
+#include "kvm.h"
+#include "apicdef.h"
+#include "kvm_ioapic.h"
+#include "irq.h"
+#include "irq.h"
+#include "vmx.h"
+
+/*
+ * XXX
+ * The fact that I'm externing these is a sign of failure
+ */
+extern void kvm_xcall(processorid_t, kvm_xcall_t, void *);
+extern int is_long_mode(struct kvm_vcpu *vcpu);
+extern void kvm_migrate_timers(struct kvm_vcpu *vcpu);
+extern ulong kvm_read_cr0_bits(struct kvm_vcpu *vcpu, ulong mask);
+extern void kvm_rip_write(struct kvm_vcpu *, unsigned long);
+extern int kvm_exception_is_soft(unsigned int);
+extern uint64_t kvm_va2pa(caddr_t va);
+extern int kvm_get_msr_common(struct kvm_vcpu *, uint32_t, uint64_t *);
+extern int kvm_set_msr_common(struct kvm_vcpu *, uint32_t, uint64_t);
+extern int getcr4(void);
+extern void setcr4(ulong_t val);
+extern void kvm_enable_efer_bits(uint64_t);
+extern int is_paging(struct kvm_vcpu *);
+extern int is_pae(struct kvm_vcpu *vcpu);
+extern ulong kvm_read_cr4(struct kvm_vcpu *);
+extern int is_protmode(struct kvm_vcpu *vcpu);
+extern kmutex_t vmx_vpid_lock;
+extern ulong_t *vmx_vpid_bitmap;
+extern size_t vpid_bitmap_words;
+extern long find_first_zero_bit(const unsigned long *, unsigned long);
+extern unsigned long native_read_cr0(void);
+#define read_cr0() (native_read_cr0())
+extern unsigned long native_read_cr4(void);
+#define read_cr4() (native_read_cr4())
+extern unsigned long native_read_cr3(void);
+#define read_cr3() (native_read_cr3())
+extern void kvm_set_cr8(struct kvm_vcpu *, unsigned long);
+extern void kvm_set_apic_base(struct kvm_vcpu *, uint64_t);
+extern void fx_init(struct kvm_vcpu *);
+extern void kvm_register_write(struct kvm_vcpu *vcpu,
+ enum kvm_reg reg, unsigned long val);
+extern ulong kvm_read_cr0(struct kvm_vcpu *vcpu);
+extern int emulate_instruction(struct kvm_vcpu *, unsigned long,
+ uint16_t, int);
+extern void kvm_queue_exception(struct kvm_vcpu *vcpu, unsigned nr);
+extern int kvm_event_needs_reinjection(struct kvm_vcpu *);
+extern int kvm_mmu_unprotect_page_virt(struct kvm_vcpu *, gva_t);
+extern int kvm_mmu_page_fault(struct kvm_vcpu *, gva_t, uint32_t);
+extern int kvm_emulate_halt(struct kvm_vcpu *);
+extern int kvm_emulate_pio(struct kvm_vcpu *, int, int, unsigned);
+extern unsigned long kvm_register_read(struct kvm_vcpu *, enum kvm_reg);
+extern void kvm_set_cr0(struct kvm_vcpu *, unsigned long);
+extern void kvm_set_cr3(struct kvm_vcpu *, unsigned long);
+extern void kvm_set_cr4(struct kvm_vcpu *, unsigned long);
+extern void kvm_set_cr8(struct kvm_vcpu *, unsigned long);
+extern unsigned long kvm_get_cr8(struct kvm_vcpu *);
+extern void kvm_lmsw(struct kvm_vcpu *, unsigned long);
+extern ulong kvm_read_cr4_bits(struct kvm_vcpu *, ulong);
+extern int kvm_require_cpl(struct kvm_vcpu *, int);
+extern void kvm_emulate_cpuid(struct kvm_vcpu *);
+extern int kvm_emulate_hypercall(struct kvm_vcpu *);
+extern void kvm_mmu_invlpg(struct kvm_vcpu *, gva_t);
+extern void kvm_clear_interrupt_queue(struct kvm_vcpu *);
+extern void kvm_clear_exception_queue(struct kvm_vcpu *);
+extern int kvm_task_switch(struct kvm_vcpu *, uint16_t, int);
+extern int kvm_mmu_get_spte_hierarchy(struct kvm_vcpu *,
+ uint64_t, uint64_t sptes[4]);
+extern void kvm_queue_interrupt(struct kvm_vcpu *, uint8_t, int);
+extern int kvm_vcpu_init(struct kvm_vcpu *, struct kvm *, unsigned);
+
+/* These are the region types */
+#define MTRR_TYPE_UNCACHABLE 0
+#define MTRR_TYPE_WRCOMB 1
+#define MTRR_TYPE_WRTHROUGH 4
+#define MTRR_TYPE_WRPROT 5
+#define MTRR_TYPE_WRBACK 6
+#define MTRR_NUM_TYPES 7
+
+extern uint8_t kvm_get_guest_memory_type(struct kvm_vcpu *, gfn_t);
+extern uint32_t bit(int);
+extern int kvm_init(void *, unsigned int);
+extern void kvm_enable_tdp(void);
+extern void kvm_disable_tdp(void);
+
+/*
+ * XXX These should be from <asm/cpu.h>
+ */
+extern void cli(void);
+extern void sti(void);
+
+static int bypass_guest_pf = 1;
+/* XXX This should be static */
+int enable_vpid = 1;
+static int flexpriority_enabled = 1;
+static int enable_ept = 1;
+static int enable_unrestricted_guest = 1;
+static int emulate_invalid_guest_state = 0;
+
+/*
+ * In linux, there is a separate vmx kernel module from the kvm driver.
+ * That may be a good idea, but we're going to do everything in
+ * the kvm driver, for now.
+ * The call to vmx_init() in _init() is done when the vmx module
+ * is loaded on linux.
+ */
+
+struct vmcs **vmxarea; /* 1 per cpu */
+struct vmcs **current_vmcs;
+struct kvm_shared_msrs **shared_msrs;
+list_t **vcpus_on_cpu;
+uint64_t *vmxarea_pa; /* physical address of each vmxarea */
+
+#define KVM_GUEST_CR0_MASK_UNRESTRICTED_GUEST \
+ (X86_CR0_WP | X86_CR0_NE | X86_CR0_NW | X86_CR0_CD)
+#define KVM_GUEST_CR0_MASK \
+ (KVM_GUEST_CR0_MASK_UNRESTRICTED_GUEST | X86_CR0_PG | X86_CR0_PE)
+#define KVM_VM_CR0_ALWAYS_ON_UNRESTRICTED_GUEST \
+ (X86_CR0_WP | X86_CR0_NE)
+#define KVM_VM_CR0_ALWAYS_ON \
+ (KVM_VM_CR0_ALWAYS_ON_UNRESTRICTED_GUEST | X86_CR0_PG | X86_CR0_PE)
+#define KVM_CR4_GUEST_OWNED_BITS \
+ (X86_CR4_PVI | X86_CR4_DE | X86_CR4_PCE | X86_CR4_OSFXSR \
+ | X86_CR4_OSXMMEXCPT)
+
+#define KVM_PMODE_VM_CR4_ALWAYS_ON (X86_CR4_PAE | X86_CR4_VMXE)
+#define KVM_RMODE_VM_CR4_ALWAYS_ON (X86_CR4_VME | X86_CR4_PAE | X86_CR4_VMXE)
+
+#define RMODE_GUEST_OWNED_EFLAGS_BITS (~(X86_EFLAGS_IOPL | X86_EFLAGS_VM))
+
+#define __kvm_handle_fault_on_reboot(insn) \
+ "666: " insn "\n\t" \
+ ".pushsection .fixup, \"ax\" \n" \
+ "667: \n\t" \
+ __ASM_SIZE(push) " $666b \n\t" \
+ ".popsection \n\t" \
+ ".pushsection __ex_table, \"a\" \n\t" \
+ _ASM_PTR " 666b, 667b \n\t" \
+ ".popsection \n\t"
+
+#define __ex(x) __kvm_handle_fault_on_reboot(x)
+
+#define page_to_phys(page) (page->p_pagenum << PAGESHIFT)
+
+/*
+ * These 2 parameters are used to config the controls for Pause-Loop Exiting:
+ * ple_gap: upper bound on the amount of time between two successive
+ * executions of PAUSE in a loop. Also indicate if ple enabled.
+ * According to test, this time is usually small than 41 cycles.
+ * ple_window: upper bound on the amount of time a guest is allowed to execute
+ * in a PAUSE loop. Tests indicate that most spinlocks are held for
+ * less than 2^12 cycles
+ * Time is measured based on a counter that runs at the same rate as the TSC,
+ * refer SDM volume 3b section 21.6.13 & 22.1.3.
+ */
+#define KVM_VMX_DEFAULT_PLE_GAP 41
+#define KVM_VMX_DEFAULT_PLE_WINDOW 4096
+
+static int ple_gap = KVM_VMX_DEFAULT_PLE_GAP;
+static int ple_window = KVM_VMX_DEFAULT_PLE_WINDOW;
+
+typedef struct vmcs {
+ uint32_t revision_id;
+ uint32_t abort;
+ char data[1]; /* size is read from MSR */
+} vmcs_t;
+
+typedef struct shared_msr_entry {
+ unsigned index;
+ uint64_t data;
+ uint64_t mask;
+} shared_msr_entry_t;
+
+typedef struct vcpu_vmx {
+ struct kvm_vcpu vcpu;
+ list_t local_vcpus_link;
+ unsigned long host_rsp;
+ int launched;
+ unsigned char fail;
+ uint32_t idt_vectoring_info;
+ struct shared_msr_entry *guest_msrs;
+ int nmsrs;
+ int save_nmsrs;
+#ifdef CONFIG_X86_64
+ uint64_t msr_host_kernel_gs_base;
+ uint64_t msr_guest_kernel_gs_base;
+#endif
+ struct vmcs *vmcs;
+ uint64_t vmcs_pa; /* physical address of vmx's vmcs */
+ struct {
+ int loaded;
+ unsigned short fs_sel, gs_sel, ldt_sel;
+ int gs_ldt_reload_needed;
+ int fs_reload_needed;
+ } host_state;
+ struct {
+ int vm86_active;
+ ulong save_rflags;
+ struct kvm_save_segment {
+ unsigned short selector;
+ unsigned long base;
+ uint32_t limit;
+ uint32_t ar;
+ } tr, es, ds, fs, gs;
+ struct {
+ char pending;
+ unsigned char vector;
+ unsigned rip;
+ } irq;
+ } rmode;
+ int vpid;
+ char emulation_required;
+
+ /* Support for vnmi-less CPUs */
+ int soft_vnmi_blocked;
+ time_t entry_time;
+ int64_t vnmi_blocked_time;
+ uint32_t exit_reason;
+
+ char rdtscp_enabled;
+} vcpu_vmx_t;
+
+static struct vcpu_vmx *
+to_vmx(struct kvm_vcpu *vcpu)
+{
+#ifdef XXX_KVM_DOESNTCOMPILE
+ return (container_of(vcpu, struct vcpu_vmx, vcpu));
+#else
+ /* assumes vcpu is first field in vcpu_vmx */
+ /* because gcc with kernel flags complains about container_of */
+ return ((struct vcpu_vmx *)vcpu);
+#endif
+}
+
+/* XXX Should be pointers, not arrays of unknown length! */
+static unsigned long vmx_io_bitmap_a[];
+static unsigned long vmx_io_bitmap_b[];
+static unsigned long vmx_msr_bitmap_legacy[];
+static unsigned long vmx_msr_bitmap_longmode[];
+
+typedef struct vmcs_config {
+ int size;
+ int order;
+ uint32_t revision_id;
+ uint32_t pin_based_exec_ctrl;
+ uint32_t cpu_based_exec_ctrl;
+ uint32_t cpu_based_2nd_exec_ctrl;
+ uint32_t vmexit_ctrl;
+ uint32_t vmentry_ctrl;
+} vmcs_config_t;
+
+typedef struct vmx_capability {
+ uint32_t ept;
+ uint32_t vpid;
+} vmx_capability_t;
+
+#define VMX_SEGMENT_FIELD(seg) \
+ [VCPU_SREG_##seg] = { \
+ .selector = GUEST_##seg##_SELECTOR, \
+ .base = GUEST_##seg##_BASE, \
+ .limit = GUEST_##seg##_LIMIT, \
+ .ar_bytes = GUEST_##seg##_AR_BYTES, \
+ }
+
+struct kvm_vmx_segment_field kvm_vmx_segment_fields[] = {
+ VMX_SEGMENT_FIELD(CS),
+ VMX_SEGMENT_FIELD(DS),
+ VMX_SEGMENT_FIELD(ES),
+ VMX_SEGMENT_FIELD(FS),
+ VMX_SEGMENT_FIELD(GS),
+ VMX_SEGMENT_FIELD(SS),
+ VMX_SEGMENT_FIELD(TR),
+ VMX_SEGMENT_FIELD(LDTR),
+};
+
+static vmcs_config_t vmcs_config;
+static vmx_capability_t vmx_capability;
+static uint64_t host_efer;
+
+static void ept_save_pdptrs(struct kvm_vcpu *);
+
+/*
+ * Keep MSR_K6_STAR at the end, as setup_msrs() will try to optimize it
+ * away by decrementing the array size.
+ */
+static const uint32_t vmx_msr_index[] = {
+#ifdef CONFIG_X86_64
+ MSR_SYSCALL_MASK, MSR_LSTAR, MSR_CSTAR,
+#endif
+ MSR_EFER, MSR_TSC_AUX, MSR_K6_STAR,
+};
+
+#define NR_VMX_MSR ARRAY_SIZE(vmx_msr_index)
+
+static int
+is_page_fault(uint32_t intr_info)
+{
+ return ((intr_info & (INTR_INFO_INTR_TYPE_MASK | INTR_INFO_VECTOR_MASK |
+ INTR_INFO_VALID_MASK)) == (INTR_TYPE_HARD_EXCEPTION |
+ PF_VECTOR | INTR_INFO_VALID_MASK));
+}
+
+static int
+is_no_device(uint32_t intr_info)
+{
+ return ((intr_info & (INTR_INFO_INTR_TYPE_MASK | INTR_INFO_VECTOR_MASK |
+ INTR_INFO_VALID_MASK)) == (INTR_TYPE_HARD_EXCEPTION | NM_VECTOR |
+ INTR_INFO_VALID_MASK));
+}
+
+static int
+is_invalid_opcode(uint32_t intr_info)
+{
+ return ((intr_info & (INTR_INFO_INTR_TYPE_MASK | INTR_INFO_VECTOR_MASK |
+ INTR_INFO_VALID_MASK)) == (INTR_TYPE_HARD_EXCEPTION | UD_VECTOR |
+ INTR_INFO_VALID_MASK));
+}
+
+static int
+is_external_interrupt(uint32_t intr_info)
+{
+ return ((intr_info & (INTR_INFO_INTR_TYPE_MASK |
+ INTR_INFO_VALID_MASK)) == (INTR_TYPE_EXT_INTR |
+ INTR_INFO_VALID_MASK));
+}
+
+static int
+is_machine_check(uint32_t intr_info)
+{
+ return (intr_info & (INTR_INFO_INTR_TYPE_MASK | INTR_INFO_VECTOR_MASK |
+ INTR_INFO_VALID_MASK)) == (INTR_TYPE_HARD_EXCEPTION |
+ MC_VECTOR | INTR_INFO_VALID_MASK);
+}
+
+static int
+cpu_has_vmx_msr_bitmap(void)
+{
+ return (vmcs_config.cpu_based_exec_ctrl & CPU_BASED_USE_MSR_BITMAPS);
+}
+
+static int
+cpu_has_vmx_tpr_shadow(void)
+{
+ return (vmcs_config.cpu_based_exec_ctrl & CPU_BASED_TPR_SHADOW);
+}
+
+static int
+vm_need_tpr_shadow(struct kvm *kvm)
+{
+ return ((cpu_has_vmx_tpr_shadow()) && (irqchip_in_kernel(kvm)));
+}
+
+static int
+cpu_has_secondary_exec_ctrls(void)
+{
+ return (vmcs_config.cpu_based_exec_ctrl &
+ CPU_BASED_ACTIVATE_SECONDARY_CONTROLS);
+}
+
+static int
+cpu_has_vmx_virtualize_apic_accesses(void)
+{
+ return (vmcs_config.cpu_based_2nd_exec_ctrl &
+ SECONDARY_EXEC_VIRTUALIZE_APIC_ACCESSES);
+}
+
+static int
+cpu_has_vmx_flexpriority(void)
+{
+ return (cpu_has_vmx_tpr_shadow() &&
+ cpu_has_vmx_virtualize_apic_accesses());
+}
+
+static int
+cpu_has_vmx_ept_execute_only(void)
+{
+ return (!!(vmx_capability.ept & VMX_EPT_EXECUTE_ONLY_BIT));
+}
+
+static int
+cpu_has_vmx_ept_2m_page(void)
+{
+ return (!!(vmx_capability.ept & VMX_EPT_2MB_PAGE_BIT));
+}
+
+static int
+cpu_has_vmx_ept_1g_page(void)
+{
+ return (!!(vmx_capability.ept & VMX_EPT_1GB_PAGE_BIT));
+}
+
+static int
+cpu_has_vmx_invept_context(void)
+{
+ return (!!(vmx_capability.ept & VMX_EPT_EXTENT_CONTEXT_BIT));
+}
+
+static int
+cpu_has_vmx_invept_global(void)
+{
+ return (!!(vmx_capability.ept & VMX_EPT_EXTENT_GLOBAL_BIT));
+}
+
+static int
+cpu_has_vmx_ept(void)
+{
+ return (vmcs_config.cpu_based_2nd_exec_ctrl &
+ SECONDARY_EXEC_ENABLE_EPT);
+}
+
+static int
+cpu_has_vmx_unrestricted_guest(void)
+{
+ return (vmcs_config.cpu_based_2nd_exec_ctrl &
+ SECONDARY_EXEC_UNRESTRICTED_GUEST);
+}
+
+static int
+cpu_has_vmx_ple(void)
+{
+ return (vmcs_config.cpu_based_2nd_exec_ctrl &
+ SECONDARY_EXEC_PAUSE_LOOP_EXITING);
+}
+
+static int
+vm_need_virtualize_apic_accesses(struct kvm *kvm)
+{
+ return (flexpriority_enabled && irqchip_in_kernel(kvm));
+}
+
+static inline int
+cpu_has_vmx_vpid(void)
+{
+ return (vmcs_config.cpu_based_2nd_exec_ctrl &
+ SECONDARY_EXEC_ENABLE_VPID);
+}
+
+static int
+cpu_has_vmx_rdtscp(void)
+{
+ return (vmcs_config.cpu_based_2nd_exec_ctrl & SECONDARY_EXEC_RDTSCP);
+}
+
+static int
+cpu_has_virtual_nmis(void)
+{
+ return (vmcs_config.pin_based_exec_ctrl & PIN_BASED_VIRTUAL_NMIS);
+}
+
+static int
+report_flexpriority(void)
+{
+ return (flexpriority_enabled);
+}
+
+static int
+__find_msr_index(struct vcpu_vmx *vmx, uint32_t msr)
+{
+ int i;
+
+ for (i = 0; i < vmx->nmsrs; i++) {
+ if (vmx_msr_index[vmx->guest_msrs[i].index] == msr)
+ return (i);
+ }
+
+ return (-1);
+}
+
+/* XXX These used to have an __ex around them, maybe add it back? */
+static inline void
+__invvpid(int ext, uint16_t vpid, gva_t gva)
+{
+ struct {
+ uint64_t vpid:16;
+ uint64_t rsvd:48;
+ uint64_t gva;
+ } operand = { vpid, 0, gva };
+
+ /* BEGIN CSTYLED */
+ __asm__ volatile (ASM_VMX_INVVPID
+ /* CF==1 or ZF==1 --> rc = -1 */
+ "; ja 1f ; ud2 ; 1:"
+ : : "a"(&operand), "c"(ext) : "cc", "memory");
+ /* END CSTYLED */
+}
+
+static inline void
+__invept(int ext, uint64_t eptp, gpa_t gpa)
+{
+ struct {
+ uint64_t eptp, gpa;
+ } operand = {eptp, gpa};
+
+ /* BEGIN CSTYLED */
+ __asm__ volatile (ASM_VMX_INVEPT
+ /* CF==1 or ZF==1 --> rc = -1 */
+ "; ja 1f ; ud2 ; 1:\n"
+ : : "a" (&operand), "c" (ext) : "cc", "memory");
+ /* END CSTYLED */
+}
+
+static struct shared_msr_entry *
+find_msr_entry(struct vcpu_vmx *vmx, uint32_t msr)
+{
+ int i;
+
+ i = __find_msr_index(vmx, msr);
+ if (i >= 0)
+ return (&vmx->guest_msrs[i]);
+
+ return (NULL);
+}
+
+static void
+vmcs_clear(uint64_t vmcs_pa)
+{
+ unsigned char error;
+
+ /*CSTYLED*/
+ __asm__ volatile (__ex(ASM_VMX_VMCLEAR_RAX) "\n\tsetna %0\n"
+ : "=g"(error) : "a"(&vmcs_pa), "m"(vmcs_pa)
+ : "cc", "memory");
+
+ if (error)
+ cmn_err(CE_PANIC, "kvm: vmclear fail: %lx\n",
+ vmcs_pa);
+}
+
+static void
+__vcpu_clear(void *arg)
+{
+ struct vcpu_vmx *vmx = arg;
+ int cpu = CPU->cpu_id;
+
+ vmx->vmcs->revision_id = vmcs_config.revision_id;
+
+ if (vmx->vcpu.cpu == cpu)
+ vmcs_clear(vmx->vmcs_pa);
+
+ if (current_vmcs[cpu] == vmx->vmcs)
+ current_vmcs[cpu] = NULL;
+ rdtscll(vmx->vcpu.arch.host_tsc);
+
+ list_remove(vcpus_on_cpu[cpu], vmx);
+
+ vmx->vcpu.cpu = -1;
+ vmx->launched = 0;
+}
+
+static void
+vcpu_clear(struct vcpu_vmx *vmx)
+{
+ if (vmx->vcpu.cpu == -1)
+ return;
+
+ /*
+ * XXX: commented out below?
+ *
+ * smp_call_function_single(vmx->vcpu.cpu, __vcpu_clear, vmx, 1);
+ */
+ kvm_xcall(vmx->vcpu.cpu, __vcpu_clear, vmx);
+}
+
+static void
+vpid_sync_vcpu_all(struct vcpu_vmx *vmx)
+{
+ if (vmx->vpid == 0)
+ return;
+
+ __invvpid(VMX_VPID_EXTENT_SINGLE_CONTEXT, vmx->vpid, 0);
+}
+
+static void
+ept_sync_global(void)
+{
+ if (cpu_has_vmx_invept_global())
+ __invept(VMX_EPT_EXTENT_GLOBAL, 0, 0);
+}
+
+static void
+ept_sync_context(uint64_t eptp)
+{
+ if (enable_ept) {
+ if (cpu_has_vmx_invept_context())
+ __invept(VMX_EPT_EXTENT_CONTEXT, eptp, 0);
+ else
+ ept_sync_global();
+ }
+}
+
+static unsigned long
+vmcs_readl(unsigned long field)
+{
+ unsigned long value;
+
+ /*CSTYLED*/
+ __asm__ volatile (ASM_VMX_VMREAD_RDX_RAX
+ : "=a"(value) : "d"(field) : "cc");
+
+ return (value);
+}
+
+static uint16_t
+vmcs_read16(unsigned long field)
+{
+ return (vmcs_readl(field));
+}
+
+static uint32_t
+vmcs_read32(unsigned long field)
+{
+ return (vmcs_readl(field));
+}
+
+static uint64_t
+vmcs_read64(unsigned long field)
+{
+#ifdef CONFIG_X86_64
+ return (vmcs_readl(field));
+#else
+ return (vmcs_readl(field) | ((uint64_t)vmcs_readl(field + 1) << 32));
+#endif
+}
+
+static void
+vmwrite_error(unsigned long field, unsigned long value)
+{
+ cmn_err(CE_WARN, "vmwrite error: reg %lx value %lx (err %x)\n",
+ field, value, vmcs_read32(VM_INSTRUCTION_ERROR));
+}
+
+/* XXX Should be static! */
+void
+vmcs_writel(unsigned long field, unsigned long value)
+{
+ unsigned char error = 0;
+#ifndef XXX
+ /*CSTYLED*/
+ __asm__ volatile (ASM_VMX_VMWRITE_RAX_RDX "\n\tsetna %0"
+ : "=q"(error) : "a"(value), "d"(field) : "cc");
+
+ if ((error))
+ vmwrite_error(field, value);
+#else
+ XXX_KVM_PROBE;
+ __vmwrite(field, value);
+#endif
+}
+
+static void
+vmcs_write16(unsigned long field, uint16_t value)
+{
+ vmcs_writel(field, value);
+}
+
+static void
+vmcs_write32(unsigned long field, uint32_t value)
+{
+ vmcs_writel(field, value);
+}
+
+static void
+vmcs_write64(unsigned long field, uint64_t value)
+{
+ vmcs_writel(field, value);
+#ifndef CONFIG_X86_64
+ /*CSTYLED*/
+ __asm__ volatile ("");
+ vmcs_writel(field + 1, value >> 32);
+#endif
+}
+
+static void
+vmcs_clear_bits(unsigned long field, uint32_t mask)
+{
+ vmcs_writel(field, vmcs_readl(field) & ~mask);
+}
+
+static void
+vmcs_set_bits(unsigned long field, uint32_t mask)
+{
+ vmcs_writel(field, vmcs_readl(field) | mask);
+}
+
+static void
+update_exception_bitmap(struct kvm_vcpu *vcpu)
+{
+ uint32_t eb;
+
+ eb = (1u << PF_VECTOR) | (1u << UD_VECTOR) | (1u << MC_VECTOR) |
+ (1u << NM_VECTOR) | (1u << DB_VECTOR);
+
+#ifndef XXX
+ if ((vcpu->guest_debug &
+ (KVM_GUESTDBG_ENABLE | KVM_GUESTDBG_USE_SW_BP)) ==
+ (KVM_GUESTDBG_ENABLE | KVM_GUESTDBG_USE_SW_BP))
+ eb |= 1u << BP_VECTOR;
+#endif
+ if (to_vmx(vcpu)->rmode.vm86_active)
+ eb = ~0;
+ if (enable_ept)
+ eb &= ~(1u << PF_VECTOR); /* bypass_guest_pf = 0 */
+ if (vcpu->fpu_active)
+ eb &= ~(1u << NM_VECTOR);
+ vmcs_write32(EXCEPTION_BITMAP, eb);
+}
+
+
+static void
+reload_tss(void)
+{
+ /*
+ * VT restores TR but not its size. Useless.
+ */
+ struct descriptor_table gdt;
+ struct desc_struct *descs;
+
+ kvm_get_gdt(&gdt);
+ descs = (void *)gdt.base;
+ descs[GDT_KTSS].c.b.type = 9; /* available TSS */
+ load_TR_desc();
+}
+
+static int
+update_transition_efer(struct vcpu_vmx *vmx, int efer_offset)
+{
+ uint64_t guest_efer;
+ uint64_t ignore_bits;
+
+ guest_efer = vmx->vcpu.arch.efer;
+
+ /*
+ * NX is emulated; LMA and LME handled by hardware; SCE meaninless
+ * outside long mode
+ */
+ ignore_bits = EFER_NX | EFER_SCE;
+#ifdef CONFIG_X86_64
+ ignore_bits |= EFER_LMA | EFER_LME;
+ /* SCE is meaningful only in long mode on Intel */
+ if (guest_efer & EFER_LMA)
+ ignore_bits &= ~(uint64_t)EFER_SCE;
+#endif
+ guest_efer &= ~ignore_bits;
+ guest_efer |= host_efer & ignore_bits;
+ vmx->guest_msrs[efer_offset].data = guest_efer;
+ vmx->guest_msrs[efer_offset].mask = ~ignore_bits;
+
+ return (1);
+}
+
+static void
+vmx_save_host_state(struct kvm_vcpu *vcpu)
+{
+ struct vcpu_vmx *vmx = to_vmx(vcpu);
+ int i;
+
+ if (vmx->host_state.loaded)
+ return;
+
+ vmx->host_state.loaded = 1;
+ /*
+ * Set host fs and gs selectors. Unfortunately, 22.2.3 does not
+ * allow segment selectors with cpl > 0 or ti == 1.
+ */
+ vmx->host_state.ldt_sel = kvm_read_ldt();
+ vmx->host_state.gs_ldt_reload_needed = vmx->host_state.ldt_sel;
+ vmx->host_state.fs_sel = kvm_read_fs();
+ if (!(vmx->host_state.fs_sel & 7)) {
+ vmcs_write16(HOST_FS_SELECTOR, vmx->host_state.fs_sel);
+ vmx->host_state.fs_reload_needed = 0;
+ } else {
+ vmcs_write16(HOST_FS_SELECTOR, 0);
+ vmx->host_state.fs_reload_needed = 1;
+ }
+ vmx->host_state.gs_sel = kvm_read_gs();
+ if (!(vmx->host_state.gs_sel & 7))
+ vmcs_write16(HOST_GS_SELECTOR, vmx->host_state.gs_sel);
+ else {
+ vmcs_write16(HOST_GS_SELECTOR, 0);
+ vmx->host_state.gs_ldt_reload_needed = 1;
+ }
+
+#ifdef CONFIG_X86_64
+ vmcs_writel(HOST_FS_BASE, read_msr(MSR_FS_BASE));
+ vmcs_writel(HOST_GS_BASE, read_msr(MSR_GS_BASE));
+#else
+ vmcs_writel(HOST_FS_BASE, segment_base(vmx->host_state.fs_sel));
+ vmcs_writel(HOST_GS_BASE, segment_base(vmx->host_state.gs_sel));
+#endif
+
+#ifdef CONFIG_X86_64
+ if (is_long_mode(&vmx->vcpu)) {
+ rdmsrl(MSR_KERNEL_GS_BASE, vmx->msr_host_kernel_gs_base);
+ wrmsrl(MSR_KERNEL_GS_BASE, vmx->msr_guest_kernel_gs_base);
+ }
+#endif
+ for (i = 0; i < vmx->save_nmsrs; i++) {
+ kvm_set_shared_msr(vcpu, vmx->guest_msrs[i].index,
+ vmx->guest_msrs[i].data, vmx->guest_msrs[i].mask);
+ }
+}
+
+static void
+__vmx_load_host_state(struct vcpu_vmx *vmx)
+{
+ unsigned long flags;
+
+ if (!vmx->host_state.loaded)
+ return;
+
+ KVM_VCPU_KSTAT_INC(&vmx->vcpu, kvmvs_host_state_reload);
+
+ vmx->host_state.loaded = 0;
+ if (vmx->host_state.fs_reload_needed)
+ kvm_load_fs(vmx->host_state.fs_sel);
+ if (vmx->host_state.gs_ldt_reload_needed) {
+ kvm_load_ldt(vmx->host_state.ldt_sel);
+ /*
+ * If we have to reload gs, we must take care to
+ * preserve our gs base.
+ */
+ cli();
+ kvm_load_gs(vmx->host_state.gs_sel);
+#ifdef CONFIG_X86_64
+ wrmsrl(MSR_GS_BASE, vmcs_readl(HOST_GS_BASE));
+#endif
+ sti();
+ }
+ reload_tss();
+
+#ifdef CONFIG_X86_64
+ if (is_long_mode(&vmx->vcpu)) {
+ rdmsrl(MSR_KERNEL_GS_BASE, vmx->msr_guest_kernel_gs_base);
+ wrmsrl(MSR_KERNEL_GS_BASE, vmx->msr_host_kernel_gs_base);
+ }
+#endif
+}
+
+static void
+vmx_load_host_state(struct vcpu_vmx *vmx)
+{
+ kpreempt_disable();
+ __vmx_load_host_state(vmx);
+ kpreempt_enable();
+}
+
+
+/*
+ * Switches to specified vcpu, until a matching vcpu_put(), but assumes
+ * vcpu mutex is already taken.
+ */
+static void
+vmx_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
+{
+ struct vcpu_vmx *vmx = to_vmx(vcpu);
+ uint64_t phys_addr = vmx->vmcs_pa;
+ uint64_t tsc_this, delta, new_offset;
+
+ if (vcpu->cpu != cpu) {
+ vcpu_clear(vmx);
+ kvm_migrate_timers(vcpu);
+ set_bit(KVM_REQ_TLB_FLUSH, &vcpu->requests);
+ kpreempt_disable();
+ list_insert_head(vcpus_on_cpu[cpu], vmx);
+ kpreempt_enable();
+ }
+
+ if (current_vmcs[cpu] != vmx->vmcs) {
+ uint8_t error;
+
+ current_vmcs[cpu] = vmx->vmcs;
+
+ /*CSTYLED*/
+ __asm__ volatile (ASM_VMX_VMPTRLD_RAX "; setna %0"
+ : "=g"(error) : "a"(&phys_addr), "m"(phys_addr)
+ : "cc");
+ }
+
+ if (vcpu->cpu != cpu) {
+ struct descriptor_table dt;
+ unsigned long sysenter_esp;
+
+ vcpu->cpu = cpu;
+
+ /*
+ * Linux uses per-cpu TSS and GDT, so set these when switching
+ * processors.
+ */
+ vmcs_writel(HOST_TR_BASE, kvm_read_tr_base()); /* 22.2.4 */
+ kvm_get_gdt(&dt);
+ vmcs_writel(HOST_GDTR_BASE, dt.base); /* 22.2.4 */
+
+ rdmsrl(MSR_IA32_SYSENTER_ESP, sysenter_esp);
+ vmcs_writel(HOST_IA32_SYSENTER_ESP, sysenter_esp); /* 22.2.3 */
+
+ /*
+ * Make sure the time stamp counter is monotonous.
+ */
+ rdtscll(tsc_this);
+ if (tsc_this < vcpu->arch.host_tsc) {
+ delta = vcpu->arch.host_tsc - tsc_this;
+ new_offset = vmcs_read64(TSC_OFFSET) + delta;
+ vmcs_write64(TSC_OFFSET, new_offset);
+ }
+ }
+}
+
+
+static void
+vmx_vcpu_put(struct kvm_vcpu *vcpu)
+{
+ __vmx_load_host_state(to_vmx(vcpu));
+}
+
+static void
+vmx_fpu_activate(struct kvm_vcpu *vcpu)
+{
+ ulong cr0;
+
+ if (vcpu->fpu_active)
+ return;
+
+ vcpu->fpu_active = 1;
+ cr0 = vmcs_readl(GUEST_CR0);
+ cr0 &= ~(X86_CR0_TS | X86_CR0_MP);
+ cr0 |= kvm_read_cr0_bits(vcpu, X86_CR0_TS | X86_CR0_MP);
+ vmcs_writel(GUEST_CR0, cr0);
+ update_exception_bitmap(vcpu);
+ vcpu->arch.cr0_guest_owned_bits = X86_CR0_TS;
+ vmcs_writel(CR0_GUEST_HOST_MASK, ~vcpu->arch.cr0_guest_owned_bits);
+}
+
+static void vmx_decache_cr0_guest_bits(struct kvm_vcpu *);
+
+static void
+vmx_fpu_deactivate(struct kvm_vcpu *vcpu)
+{
+ vmx_decache_cr0_guest_bits(vcpu);
+ vmcs_set_bits(GUEST_CR0, X86_CR0_TS | X86_CR0_MP);
+ update_exception_bitmap(vcpu);
+ vcpu->arch.cr0_guest_owned_bits = 0;
+ vmcs_writel(CR0_GUEST_HOST_MASK, ~vcpu->arch.cr0_guest_owned_bits);
+ vmcs_writel(CR0_READ_SHADOW, vcpu->arch.cr0);
+}
+
+static unsigned long
+vmx_get_rflags(struct kvm_vcpu *vcpu)
+{
+ unsigned long rflags, save_rflags;
+
+ rflags = vmcs_readl(GUEST_RFLAGS);
+ if (to_vmx(vcpu)->rmode.vm86_active) {
+ rflags &= RMODE_GUEST_OWNED_EFLAGS_BITS;
+ save_rflags = to_vmx(vcpu)->rmode.save_rflags;
+ rflags |= save_rflags & ~RMODE_GUEST_OWNED_EFLAGS_BITS;
+ }
+
+ return (rflags);
+}
+
+static void
+vmx_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags)
+{
+ if (to_vmx(vcpu)->rmode.vm86_active) {
+ to_vmx(vcpu)->rmode.save_rflags = rflags;
+ rflags |= X86_EFLAGS_IOPL | X86_EFLAGS_VM;
+ }
+
+ vmcs_writel(GUEST_RFLAGS, rflags);
+}
+
+static uint32_t
+vmx_get_interrupt_shadow(struct kvm_vcpu *vcpu, int mask)
+{
+ uint32_t interruptibility = vmcs_read32(GUEST_INTERRUPTIBILITY_INFO);
+ int ret = 0;
+
+ if (interruptibility & GUEST_INTR_STATE_STI)
+ ret |= X86_SHADOW_INT_STI;
+ if (interruptibility & GUEST_INTR_STATE_MOV_SS)
+ ret |= X86_SHADOW_INT_MOV_SS;
+
+ return (ret & mask);
+}
+
+static void
+vmx_set_interrupt_shadow(struct kvm_vcpu *vcpu, int mask)
+{
+ uint32_t old = vmcs_read32(GUEST_INTERRUPTIBILITY_INFO);
+ uint32_t interruptibility = old;
+
+ interruptibility &= ~(GUEST_INTR_STATE_STI | GUEST_INTR_STATE_MOV_SS);
+
+ if (mask & X86_SHADOW_INT_MOV_SS)
+ interruptibility |= GUEST_INTR_STATE_MOV_SS;
+ if (mask & X86_SHADOW_INT_STI)
+ interruptibility |= GUEST_INTR_STATE_STI;
+
+ if ((interruptibility != old))
+ vmcs_write32(GUEST_INTERRUPTIBILITY_INFO, interruptibility);
+}
+
+static void
+skip_emulated_instruction(struct kvm_vcpu *vcpu)
+{
+ unsigned long rip;
+
+ rip = kvm_rip_read(vcpu);
+ rip += vmcs_read32(VM_EXIT_INSTRUCTION_LEN);
+ kvm_rip_write(vcpu, rip);
+
+ /* skipping an emulated instruction also counts */
+ vmx_set_interrupt_shadow(vcpu, 0);
+}
+
+static void
+vmx_queue_exception(struct kvm_vcpu *vcpu, unsigned nr,
+ int has_error_code, uint32_t error_code)
+{
+ struct vcpu_vmx *vmx = to_vmx(vcpu);
+ uint32_t intr_info = nr | INTR_INFO_VALID_MASK;
+
+ if (has_error_code) {
+ vmcs_write32(VM_ENTRY_EXCEPTION_ERROR_CODE, error_code);
+ intr_info |= INTR_INFO_DELIVER_CODE_MASK;
+ }
+
+ if (vmx->rmode.vm86_active) {
+ vmx->rmode.irq.pending = 1;
+ vmx->rmode.irq.vector = nr;
+ vmx->rmode.irq.rip = kvm_rip_read(vcpu);
+ if (kvm_exception_is_soft(nr))
+ vmx->rmode.irq.rip +=
+ vmx->vcpu.arch.event_exit_inst_len;
+ intr_info |= INTR_TYPE_SOFT_INTR;
+ vmcs_write32(VM_ENTRY_INTR_INFO_FIELD, intr_info);
+ vmcs_write32(VM_ENTRY_INSTRUCTION_LEN, 1);
+ kvm_rip_write(vcpu, vmx->rmode.irq.rip - 1);
+ return;
+ }
+
+ if (kvm_exception_is_soft(nr)) {
+ vmcs_write32(VM_ENTRY_INSTRUCTION_LEN,
+ vmx->vcpu.arch.event_exit_inst_len);
+ intr_info |= INTR_TYPE_SOFT_EXCEPTION;
+ } else
+ intr_info |= INTR_TYPE_HARD_EXCEPTION;
+
+ vmcs_write32(VM_ENTRY_INTR_INFO_FIELD, intr_info);
+}
+
+static int
+vmx_rdtscp_supported(void)
+{
+ return (cpu_has_vmx_rdtscp());
+}
+
+/*
+ * Swap MSR entry in host/guest MSR entry array.
+ */
+static void
+move_msr_up(struct vcpu_vmx *vmx, int from, int to)
+{
+ struct shared_msr_entry tmp;
+
+ tmp = vmx->guest_msrs[to];
+ vmx->guest_msrs[to] = vmx->guest_msrs[from];
+ vmx->guest_msrs[from] = tmp;
+}
+
+
+/*
+ * Set up the vmcs to automatically save and restore system
+ * msrs. Don't touch the 64-bit msrs if the guest is in legacy
+ * mode, as fiddling with msrs is very expensive.
+ */
+void
+setup_msrs(struct vcpu_vmx *vmx)
+{
+ int save_nmsrs, index;
+ unsigned long *msr_bitmap;
+
+ vmx_load_host_state(vmx);
+ save_nmsrs = 0;
+#ifdef CONFIG_X86_64
+ if (is_long_mode(&vmx->vcpu)) {
+ index = __find_msr_index(vmx, MSR_SYSCALL_MASK);
+ if (index >= 0)
+ move_msr_up(vmx, index, save_nmsrs++);
+ index = __find_msr_index(vmx, MSR_LSTAR);
+ if (index >= 0)
+ move_msr_up(vmx, index, save_nmsrs++);
+ index = __find_msr_index(vmx, MSR_CSTAR);
+ if (index >= 0)
+ move_msr_up(vmx, index, save_nmsrs++);
+ index = __find_msr_index(vmx, MSR_TSC_AUX);
+ if (index >= 0 && vmx->rdtscp_enabled)
+ move_msr_up(vmx, index, save_nmsrs++);
+ /*
+ * MSR_K6_STAR is only needed on long mode guests, and only
+ * if efer.sce is enabled.
+ */
+ index = __find_msr_index(vmx, MSR_K6_STAR);
+ if ((index >= 0) && (vmx->vcpu.arch.efer & EFER_SCE))
+ move_msr_up(vmx, index, save_nmsrs++);
+ }
+#endif
+ index = __find_msr_index(vmx, MSR_EFER);
+ if (index >= 0 && update_transition_efer(vmx, index))
+ move_msr_up(vmx, index, save_nmsrs++);
+
+ vmx->save_nmsrs = save_nmsrs;
+
+ if (cpu_has_vmx_msr_bitmap()) {
+ if (is_long_mode(&vmx->vcpu))
+ msr_bitmap = vmx_msr_bitmap_longmode;
+ else
+ msr_bitmap = vmx_msr_bitmap_legacy;
+
+ vmcs_write64(MSR_BITMAP, kvm_va2pa((caddr_t)msr_bitmap));
+ }
+}
+
+/*
+ * reads and returns guest's timestamp counter "register"
+ * guest_tsc = host_tsc + tsc_offset -- 21.3
+ */
+static uint64_t
+guest_read_tsc(void)
+{
+ uint64_t host_tsc, tsc_offset;
+
+ rdtscll(host_tsc);
+ tsc_offset = vmcs_read64(TSC_OFFSET);
+ return (host_tsc + tsc_offset);
+}
+
+/*
+ * writes 'guest_tsc' into guest's timestamp counter "register"
+ * guest_tsc = host_tsc + tsc_offset ==> tsc_offset = guest_tsc - host_tsc
+ */
+static void
+guest_write_tsc(uint64_t guest_tsc, uint64_t host_tsc)
+{
+ vmcs_write64(TSC_OFFSET, guest_tsc - host_tsc);
+}
+
+/*
+ * Reads an msr value (of 'msr_index') into 'pdata'.
+ * Returns 0 on success, non-0 otherwise.
+ * Assumes vcpu_load() was already called.
+ */
+static int
+vmx_get_msr(struct kvm_vcpu *vcpu, uint32_t msr_index, uint64_t *pdata)
+{
+ uint64_t data;
+ struct shared_msr_entry *msr;
+
+ if (!pdata) {
+ cmn_err(CE_WARN, "BUG: get_msr called with NULL pdata\n");
+ return (EINVAL);
+ }
+
+ switch (msr_index) {
+#ifdef CONFIG_X86_64
+ case MSR_FS_BASE:
+ data = vmcs_readl(GUEST_FS_BASE);
+ break;
+ case MSR_GS_BASE:
+ data = vmcs_readl(GUEST_GS_BASE);
+ break;
+ case MSR_KERNEL_GS_BASE:
+ vmx_load_host_state(to_vmx(vcpu));
+ data = to_vmx(vcpu)->msr_guest_kernel_gs_base;
+ break;
+#endif
+ case MSR_EFER:
+ return (kvm_get_msr_common(vcpu, msr_index, pdata));
+ case MSR_IA32_TSC:
+ data = guest_read_tsc();
+ break;
+ case MSR_IA32_SYSENTER_CS:
+ data = vmcs_read32(GUEST_SYSENTER_CS);
+ break;
+ case MSR_IA32_SYSENTER_EIP:
+ data = vmcs_readl(GUEST_SYSENTER_EIP);
+ break;
+ case MSR_IA32_SYSENTER_ESP:
+ data = vmcs_readl(GUEST_SYSENTER_ESP);
+ break;
+ case MSR_TSC_AUX:
+ if (!to_vmx(vcpu)->rdtscp_enabled)
+ return (1);
+ /* Otherwise falls through */
+ default:
+ vmx_load_host_state(to_vmx(vcpu));
+ msr = find_msr_entry(to_vmx(vcpu), msr_index);
+ if (msr) {
+ vmx_load_host_state(to_vmx(vcpu));
+ data = msr->data;
+ break;
+ }
+ return (kvm_get_msr_common(vcpu, msr_index, pdata));
+ }
+
+ *pdata = data;
+
+ return (0);
+}
+
+/*
+ * Writes msr value into into the appropriate "register".
+ * Returns 0 on success, non-0 otherwise.
+ * Assumes vcpu_load() was already called.
+ */
+static int
+vmx_set_msr(struct kvm_vcpu *vcpu, uint32_t msr_index, uint64_t data)
+{
+ struct vcpu_vmx *vmx = to_vmx(vcpu);
+ struct shared_msr_entry *msr;
+ uint64_t host_tsc;
+ int ret = 0;
+
+ switch (msr_index) {
+ case MSR_EFER:
+ vmx_load_host_state(vmx);
+ ret = kvm_set_msr_common(vcpu, msr_index, data);
+ break;
+#ifdef CONFIG_X86_64
+ case MSR_FS_BASE:
+ vmcs_writel(GUEST_FS_BASE, data);
+ break;
+ case MSR_GS_BASE:
+ vmcs_writel(GUEST_GS_BASE, data);
+ break;
+ case MSR_KERNEL_GS_BASE:
+ vmx_load_host_state(vmx);
+ vmx->msr_guest_kernel_gs_base = data;
+ break;
+#endif
+ case MSR_IA32_SYSENTER_CS:
+ vmcs_write32(GUEST_SYSENTER_CS, data);
+ break;
+ case MSR_IA32_SYSENTER_EIP:
+ vmcs_writel(GUEST_SYSENTER_EIP, data);
+ break;
+ case MSR_IA32_SYSENTER_ESP:
+ vmcs_writel(GUEST_SYSENTER_ESP, data);
+ break;
+ case MSR_IA32_TSC:
+ rdtscll(host_tsc);
+ guest_write_tsc(data, host_tsc);
+ break;
+ case MSR_IA32_CR_PAT:
+ if (vmcs_config.vmentry_ctrl & VM_ENTRY_LOAD_IA32_PAT) {
+ vmcs_write64(GUEST_IA32_PAT, data);
+ vcpu->arch.pat = data;
+ break;
+ }
+ ret = kvm_set_msr_common(vcpu, msr_index, data);
+ break;
+ case MSR_TSC_AUX:
+ if (!vmx->rdtscp_enabled)
+ return (1);
+ /* Check reserved bit, higher 32 bits should be zero */
+ if ((data >> 32) != 0)
+ return (1);
+ /* Otherwise falls through */
+ default:
+ msr = find_msr_entry(vmx, msr_index);
+ if (msr) {
+ vmx_load_host_state(vmx);
+ msr->data = data;
+ break;
+ }
+ ret = kvm_set_msr_common(vcpu, msr_index, data);
+ }
+
+ return (ret);
+}
+
+static void
+vmx_cache_reg(struct kvm_vcpu *vcpu, enum kvm_reg reg)
+{
+ __set_bit(reg, (unsigned long *)&vcpu->arch.regs_avail);
+
+ switch (reg) {
+ case VCPU_REGS_RSP:
+ vcpu->arch.regs[VCPU_REGS_RSP] = vmcs_readl(GUEST_RSP);
+ break;
+ case VCPU_REGS_RIP:
+ vcpu->arch.regs[VCPU_REGS_RIP] = vmcs_readl(GUEST_RIP);
+ break;
+ case VCPU_EXREG_PDPTR:
+ if (enable_ept)
+ ept_save_pdptrs(vcpu);
+ break;
+ default:
+ break;
+ }
+}
+
+static void
+set_guest_debug(struct kvm_vcpu *vcpu, struct kvm_guest_debug *dbg)
+{
+ if (vcpu->guest_debug & KVM_GUESTDBG_USE_HW_BP)
+ vmcs_writel(GUEST_DR7, dbg->arch.debugreg[7]);
+ else
+ vmcs_writel(GUEST_DR7, vcpu->arch.dr7);
+
+ update_exception_bitmap(vcpu);
+}
+
+static int
+vmx_hardware_enable(void *garbage)
+{
+ int cpu = curthread->t_cpu->cpu_seqid;
+ pfn_t pfn;
+ uint64_t old;
+#ifdef XXX
+ uint64_t phys_addr = kvtop(per_cpu(vmxarea, cpu));
+#else
+ uint64_t phys_addr;
+ XXX_KVM_PROBE;
+ phys_addr = vmxarea_pa[cpu];
+
+#endif
+
+ ((struct vmcs *)(vmxarea[cpu]))->revision_id = vmcs_config.revision_id;
+
+ if (getcr4() & X86_CR4_VMXE)
+ return (DDI_FAILURE);
+
+#ifdef XXX
+ INIT_LIST_HEAD(&per_cpu(vcpus_on_cpu, cpu));
+#else
+ XXX_KVM_PROBE;
+#endif
+ rdmsrl(MSR_IA32_FEATURE_CONTROL, old);
+ if ((old & (FEATURE_CONTROL_LOCKED | FEATURE_CONTROL_VMXON_ENABLED)) !=
+ (FEATURE_CONTROL_LOCKED | FEATURE_CONTROL_VMXON_ENABLED)) {
+ /* enable and lock */
+ wrmsrl(MSR_IA32_FEATURE_CONTROL, old | FEATURE_CONTROL_LOCKED |
+ FEATURE_CONTROL_VMXON_ENABLED);
+ }
+
+ setcr4(getcr4() | X86_CR4_VMXE); /* FIXME: not cpu hotplug safe */
+ /* BEGIN CSTYLED */
+ __asm__ volatile (ASM_VMX_VMXON_RAX
+ : : "a"(&phys_addr), "m"(phys_addr)
+ : "memory", "cc");
+ /* END CSTYLED */
+
+ ept_sync_global();
+
+ return (0);
+}
+
+static void
+vmclear_local_vcpus(void)
+{
+ int cpu = CPU->cpu_id;
+ struct vcpu_vmx *vmx, *n;
+
+ /*
+ * list_for_each_entry_safe(vmx, n, &per_cpu(vcpus_on_cpu, cpu),
+ * local_vcpus_link)
+ * __vcpu_clear(vmx);
+ */
+ for (vmx = list_head(vcpus_on_cpu[cpu]); vmx;
+ vmx = list_next(vcpus_on_cpu[cpu], vmx))
+ __vcpu_clear(vmx);
+}
+
+/*
+ * Just like cpu_vmxoff(), but with the __kvm_handle_fault_on_reboot()
+ * tricks.
+ */
+static void
+kvm_cpu_vmxoff(void)
+{
+ /* BEGIN CSTYLED */
+ __asm__ volatile ((ASM_VMX_VMXOFF) : : : "cc");
+ /* END CSTYLED */
+ setcr4(getcr4() & ~X86_CR4_VMXE);
+}
+
+static void vmx_hardware_disable(void *garbage)
+{
+ vmclear_local_vcpus();
+ kvm_cpu_vmxoff();
+}
+
+static int
+adjust_vmx_controls(uint32_t ctl_min, uint32_t ctl_opt,
+ uint32_t msr, uint32_t *result)
+{
+ uint32_t vmx_msr_low, vmx_msr_high;
+ uint32_t ctl = ctl_min | ctl_opt;
+
+ rdmsr(msr, vmx_msr_low, vmx_msr_high);
+
+ ctl &= vmx_msr_high; /* bit == 0 in high word ==> must be zero */
+ ctl |= vmx_msr_low; /* bit == 1 in low word ==> must be one */
+
+ /* Ensure minimum (required) set of control bits are supported. */
+ if (ctl_min & ~ctl)
+ return (EIO);
+
+ *result = ctl;
+ return (DDI_SUCCESS);
+}
+
+/* Pure 2^n version of get_order */
+static inline int
+get_order(unsigned long size)
+{
+ int order;
+
+ size = (size - 1) >> (PAGESHIFT - 1);
+ order = -1;
+ do {
+ size >>= 1;
+ order++;
+ } while (size);
+
+ return (order);
+}
+
+static int
+setup_vmcs_config(struct vmcs_config *vmcs_conf)
+{
+ uint32_t vmx_msr_low, vmx_msr_high;
+ uint32_t min, opt, min2, opt2;
+ uint32_t _pin_based_exec_control = 0;
+ uint32_t _cpu_based_exec_control = 0;
+ uint32_t _cpu_based_2nd_exec_control = 0;
+ uint32_t _vmexit_control = 0;
+ uint32_t _vmentry_control = 0;
+ uint32_t ept, vpid;
+
+ min = PIN_BASED_EXT_INTR_MASK | PIN_BASED_NMI_EXITING;
+ opt = PIN_BASED_VIRTUAL_NMIS;
+ if (adjust_vmx_controls(min, opt, MSR_IA32_VMX_PINBASED_CTLS,
+ &_pin_based_exec_control) != DDI_SUCCESS)
+ return (EIO);
+
+ min = CPU_BASED_HLT_EXITING |
+#ifdef CONFIG_X86_64
+ CPU_BASED_CR8_LOAD_EXITING |
+ CPU_BASED_CR8_STORE_EXITING |
+#endif
+ CPU_BASED_CR3_LOAD_EXITING |
+ CPU_BASED_CR3_STORE_EXITING |
+ CPU_BASED_USE_IO_BITMAPS |
+ CPU_BASED_MOV_DR_EXITING |
+ CPU_BASED_USE_TSC_OFFSETING |
+ CPU_BASED_MWAIT_EXITING |
+ CPU_BASED_MONITOR_EXITING |
+ CPU_BASED_INVLPG_EXITING;
+
+ opt = CPU_BASED_TPR_SHADOW |
+ CPU_BASED_USE_MSR_BITMAPS |
+ CPU_BASED_ACTIVATE_SECONDARY_CONTROLS;
+
+ if (adjust_vmx_controls(min, opt, MSR_IA32_VMX_PROCBASED_CTLS,
+ &_cpu_based_exec_control) != DDI_SUCCESS)
+ return (EIO);
+
+#ifdef CONFIG_X86_64
+ if ((_cpu_based_exec_control & CPU_BASED_TPR_SHADOW))
+ _cpu_based_exec_control &= ~CPU_BASED_CR8_LOAD_EXITING &
+ ~CPU_BASED_CR8_STORE_EXITING;
+#endif
+ if (_cpu_based_exec_control & CPU_BASED_ACTIVATE_SECONDARY_CONTROLS) {
+ min2 = 0;
+ opt2 = SECONDARY_EXEC_VIRTUALIZE_APIC_ACCESSES |
+ SECONDARY_EXEC_WBINVD_EXITING |
+ SECONDARY_EXEC_ENABLE_VPID |
+ SECONDARY_EXEC_ENABLE_EPT |
+ SECONDARY_EXEC_UNRESTRICTED_GUEST |
+ SECONDARY_EXEC_PAUSE_LOOP_EXITING |
+ SECONDARY_EXEC_RDTSCP;
+
+ if (adjust_vmx_controls(min2, opt2,
+ MSR_IA32_VMX_PROCBASED_CTLS2,
+ &_cpu_based_2nd_exec_control) != DDI_SUCCESS)
+ return (EIO);
+ }
+#ifndef CONFIG_X86_64
+ if (!(_cpu_based_2nd_exec_control &
+ SECONDARY_EXEC_VIRTUALIZE_APIC_ACCESSES))
+ _cpu_based_exec_control &= ~CPU_BASED_TPR_SHADOW;
+#endif
+ if (_cpu_based_2nd_exec_control & SECONDARY_EXEC_ENABLE_EPT) {
+ /*
+ * CR3 accesses and invlpg don't need to cause VM Exits when EPT
+ * enabled
+ */
+ _cpu_based_exec_control &= ~(CPU_BASED_CR3_LOAD_EXITING |
+ CPU_BASED_CR3_STORE_EXITING | CPU_BASED_INVLPG_EXITING);
+ rdmsr(MSR_IA32_VMX_EPT_VPID_CAP, vmx_capability.ept,
+ vmx_capability.vpid);
+ }
+
+ min = 0;
+#ifdef CONFIG_X86_64
+ min |= VM_EXIT_HOST_ADDR_SPACE_SIZE;
+#endif
+ opt = VM_EXIT_SAVE_IA32_PAT | VM_EXIT_LOAD_IA32_PAT;
+ if (adjust_vmx_controls(min, opt, MSR_IA32_VMX_EXIT_CTLS,
+ &_vmexit_control) != DDI_SUCCESS)
+ return (EIO);
+
+ min = 0;
+ opt = VM_ENTRY_LOAD_IA32_PAT;
+ if (adjust_vmx_controls(min, opt, MSR_IA32_VMX_ENTRY_CTLS,
+ &_vmentry_control) != DDI_SUCCESS)
+ return (EIO);
+
+ rdmsr(MSR_IA32_VMX_BASIC, vmx_msr_low, vmx_msr_high);
+
+ /* IA-32 SDM Vol 3B: VMCS size is never greater than 4kB. */
+ if ((vmx_msr_high & 0x1fff) > PAGESIZE)
+ return (EIO);
+
+#ifdef CONFIG_X86_64
+ /* IA-32 SDM Vol 3B: 64-bit CPUs always have VMX_BASIC_MSR[48]==0. */
+ if (vmx_msr_high & (1u<<16))
+ return (EIO);
+#endif
+
+ /* Require Write-Back (WB) memory type for VMCS accesses. */
+ if (((vmx_msr_high >> 18) & 15) != 6)
+ return (EIO);
+
+ vmcs_conf->size = vmx_msr_high & 0x1fff;
+ vmcs_conf->order = get_order(vmcs_config.size);
+ vmcs_conf->revision_id = vmx_msr_low;
+
+ vmcs_conf->pin_based_exec_ctrl = _pin_based_exec_control;
+ vmcs_conf->cpu_based_exec_ctrl = _cpu_based_exec_control;
+ vmcs_conf->cpu_based_2nd_exec_ctrl = _cpu_based_2nd_exec_control;
+ vmcs_conf->vmexit_ctrl = _vmexit_control;
+ vmcs_conf->vmentry_ctrl = _vmentry_control;
+
+ return (0);
+}
+
+static int
+alloc_kvm_area(void)
+{
+ int i, j;
+ pfn_t pfn;
+
+ /*
+ * linux seems to do the allocations in a numa-aware
+ * fashion. We'll just allocate...
+ */
+ vmxarea = kmem_alloc(ncpus * sizeof (struct vmcs *), KM_SLEEP);
+ vmxarea_pa = kmem_alloc(ncpus * sizeof (uint64_t *), KM_SLEEP);
+ current_vmcs = kmem_alloc(ncpus * sizeof (struct vmcs *), KM_SLEEP);
+ shared_msrs = kmem_alloc(ncpus * sizeof (struct kvm_shared_msrs *),
+ KM_SLEEP);
+ vcpus_on_cpu = kmem_alloc(ncpus * sizeof (list_t *), KM_SLEEP);
+
+ for (i = 0; i < ncpus; i++) {
+ struct vmcs *vmcs;
+
+ /* XXX the following assumes PAGESIZE allocations */
+ /* are PAGESIZE aligned. We could enforce this */
+ /* via kmem_cache_create, but I'm lazy */
+ vmcs = kmem_zalloc(PAGESIZE, KM_SLEEP);
+ vmxarea[i] = vmcs;
+ current_vmcs[i] = vmcs;
+ pfn = hat_getpfnum(kas.a_hat, (caddr_t)vmcs);
+ vmxarea_pa[i] = ((uint64_t)pfn << PAGESHIFT) |
+ ((uint64_t)vmxarea[i] & PAGEOFFSET);
+ shared_msrs[i] = kmem_zalloc(sizeof (struct kvm_shared_msrs),
+ KM_SLEEP);
+ vcpus_on_cpu[i] = kmem_alloc(sizeof (list_t), KM_SLEEP);
+ list_create(vcpus_on_cpu[i], sizeof (struct vcpu_vmx),
+ offsetof(struct vcpu_vmx, local_vcpus_link));
+ }
+
+ return (0);
+}
+
+
+static int
+vmx_hardware_setup(void)
+{
+ if (setup_vmcs_config(&vmcs_config) != DDI_SUCCESS)
+ return (EIO);
+#ifdef XXX
+ if (boot_cpu_has(X86_FEATURE_NX))
+#else
+ XXX_KVM_PROBE;
+#endif
+ kvm_enable_efer_bits(EFER_NX);
+
+ if (!cpu_has_vmx_vpid())
+ enable_vpid = 0;
+
+ if (!cpu_has_vmx_ept()) {
+ enable_ept = 0;
+ enable_unrestricted_guest = 0;
+ }
+
+ if (!cpu_has_vmx_unrestricted_guest())
+ enable_unrestricted_guest = 0;
+ if (!cpu_has_vmx_flexpriority())
+ flexpriority_enabled = 0;
+
+ if (!cpu_has_vmx_tpr_shadow())
+ kvm_x86_ops->update_cr8_intercept = NULL;
+
+ if (enable_ept && !cpu_has_vmx_ept_2m_page())
+ kvm_disable_largepages();
+
+ if (!cpu_has_vmx_ple())
+ ple_gap = 0;
+
+
+ return (alloc_kvm_area());
+}
+
+static void
+fix_pmode_dataseg(int seg, struct kvm_save_segment *save)
+{
+ struct kvm_vmx_segment_field *sf = &kvm_vmx_segment_fields[seg];
+
+ if (vmcs_readl(sf->base) == save->base && (save->base & AR_S_MASK)) {
+ vmcs_write16(sf->selector, save->selector);
+ vmcs_writel(sf->base, save->base);
+ vmcs_write32(sf->limit, save->limit);
+ vmcs_write32(sf->ar_bytes, save->ar);
+ } else {
+ uint32_t dpl = (vmcs_read16(sf->selector) & SELECTOR_RPL_MASK)
+ << AR_DPL_SHIFT;
+ vmcs_write32(sf->ar_bytes, 0x93 | dpl);
+ }
+}
+
+static void enter_pmode(struct kvm_vcpu *vcpu)
+{
+ unsigned long flags;
+ struct vcpu_vmx *vmx = to_vmx(vcpu);
+
+ vmx->emulation_required = 1;
+ vmx->rmode.vm86_active = 0;
+
+ vmcs_writel(GUEST_TR_BASE, vmx->rmode.tr.base);
+ vmcs_write32(GUEST_TR_LIMIT, vmx->rmode.tr.limit);
+ vmcs_write32(GUEST_TR_AR_BYTES, vmx->rmode.tr.ar);
+
+ flags = vmcs_readl(GUEST_RFLAGS);
+ flags &= RMODE_GUEST_OWNED_EFLAGS_BITS;
+ flags |= vmx->rmode.save_rflags & ~RMODE_GUEST_OWNED_EFLAGS_BITS;
+ vmcs_writel(GUEST_RFLAGS, flags);
+
+ vmcs_writel(GUEST_CR4, (vmcs_readl(GUEST_CR4) & ~X86_CR4_VME) |
+ (vmcs_readl(CR4_READ_SHADOW) & X86_CR4_VME));
+
+ update_exception_bitmap(vcpu);
+
+ if (emulate_invalid_guest_state)
+ return;
+
+ fix_pmode_dataseg(VCPU_SREG_ES, &vmx->rmode.es);
+ fix_pmode_dataseg(VCPU_SREG_DS, &vmx->rmode.ds);
+ fix_pmode_dataseg(VCPU_SREG_GS, &vmx->rmode.gs);
+ fix_pmode_dataseg(VCPU_SREG_FS, &vmx->rmode.fs);
+
+ vmcs_write16(GUEST_SS_SELECTOR, 0);
+ vmcs_write32(GUEST_SS_AR_BYTES, 0x93);
+
+ vmcs_write16(GUEST_CS_SELECTOR,
+ vmcs_read16(GUEST_CS_SELECTOR) & ~SELECTOR_RPL_MASK);
+ vmcs_write32(GUEST_CS_AR_BYTES, 0x9b);
+}
+
+static gva_t
+rmode_tss_base(struct kvm *kvm)
+{
+ if (!kvm->arch.tss_addr) {
+ struct kvm_memslots *slots;
+ gfn_t base_gfn;
+
+#ifdef XXX
+ slots = rcu_dereference(kvm->memslots);
+#else
+ XXX_KVM_PROBE;
+ slots = kvm->memslots;
+#endif
+ base_gfn = kvm->memslots->memslots[0].base_gfn +
+ kvm->memslots->memslots[0].npages - 3;
+ return (base_gfn << PAGESHIFT);
+ }
+
+ return (kvm->arch.tss_addr);
+}
+
+static void
+fix_rmode_seg(int seg, struct kvm_save_segment *save)
+{
+ struct kvm_vmx_segment_field *sf = &kvm_vmx_segment_fields[seg];
+
+ save->selector = vmcs_read16(sf->selector);
+ save->base = vmcs_readl(sf->base);
+ save->limit = vmcs_read32(sf->limit);
+ save->ar = vmcs_read32(sf->ar_bytes);
+ vmcs_write16(sf->selector, save->base >> 4);
+ vmcs_write32(sf->base, save->base & 0xfffff);
+ vmcs_write32(sf->limit, 0xffff);
+ vmcs_write32(sf->ar_bytes, 0xf3);
+}
+
+static int init_rmode(struct kvm *);
+
+static void
+enter_rmode(struct kvm_vcpu *vcpu)
+{
+ unsigned long flags;
+ struct vcpu_vmx *vmx = to_vmx(vcpu);
+
+ if (enable_unrestricted_guest)
+ return;
+
+ vmx->emulation_required = 1;
+ vmx->rmode.vm86_active = 1;
+
+ vmx->rmode.tr.base = vmcs_readl(GUEST_TR_BASE);
+ vmcs_writel(GUEST_TR_BASE, rmode_tss_base(vcpu->kvm));
+
+ vmx->rmode.tr.limit = vmcs_read32(GUEST_TR_LIMIT);
+ vmcs_write32(GUEST_TR_LIMIT, RMODE_TSS_SIZE - 1);
+
+ vmx->rmode.tr.ar = vmcs_read32(GUEST_TR_AR_BYTES);
+ vmcs_write32(GUEST_TR_AR_BYTES, 0x008b);
+
+ flags = vmcs_readl(GUEST_RFLAGS);
+ vmx->rmode.save_rflags = flags;
+
+ flags |= X86_EFLAGS_IOPL | X86_EFLAGS_VM;
+
+ vmcs_writel(GUEST_RFLAGS, flags);
+ vmcs_writel(GUEST_CR4, vmcs_readl(GUEST_CR4) | X86_CR4_VME);
+ update_exception_bitmap(vcpu);
+
+ if (emulate_invalid_guest_state)
+ goto continue_rmode;
+
+ vmcs_write16(GUEST_SS_SELECTOR, vmcs_readl(GUEST_SS_BASE) >> 4);
+ vmcs_write32(GUEST_SS_LIMIT, 0xffff);
+ vmcs_write32(GUEST_SS_AR_BYTES, 0xf3);
+
+ vmcs_write32(GUEST_CS_AR_BYTES, 0xf3);
+ vmcs_write32(GUEST_CS_LIMIT, 0xffff);
+ if (vmcs_readl(GUEST_CS_BASE) == 0xffff0000)
+ vmcs_writel(GUEST_CS_BASE, 0xf0000);
+ vmcs_write16(GUEST_CS_SELECTOR, vmcs_readl(GUEST_CS_BASE) >> 4);
+
+ fix_rmode_seg(VCPU_SREG_ES, &vmx->rmode.es);
+ fix_rmode_seg(VCPU_SREG_DS, &vmx->rmode.ds);
+ fix_rmode_seg(VCPU_SREG_GS, &vmx->rmode.gs);
+ fix_rmode_seg(VCPU_SREG_FS, &vmx->rmode.fs);
+
+continue_rmode:
+ kvm_mmu_reset_context(vcpu);
+ init_rmode(vcpu->kvm);
+}
+
+static void
+vmx_set_efer(struct kvm_vcpu *vcpu, uint64_t efer)
+{
+ struct vcpu_vmx *vmx = to_vmx(vcpu);
+ struct shared_msr_entry *msr = find_msr_entry(vmx, MSR_EFER);
+
+ if (!msr)
+ return;
+
+ /*
+ * Force kernel_gs_base reloading before EFER changes, as control
+ * of this msr depends on is_long_mode().
+ */
+ vmx_load_host_state(to_vmx(vcpu));
+ vcpu->arch.efer = efer;
+ if (efer & EFER_LMA) {
+ vmcs_write32(VM_ENTRY_CONTROLS,
+ vmcs_read32(VM_ENTRY_CONTROLS) | VM_ENTRY_IA32E_MODE);
+ msr->data = efer;
+ } else {
+ vmcs_write32(VM_ENTRY_CONTROLS,
+ vmcs_read32(VM_ENTRY_CONTROLS) & ~VM_ENTRY_IA32E_MODE);
+
+ msr->data = efer & ~EFER_LME;
+ }
+
+ setup_msrs(vmx);
+}
+
+#ifdef CONFIG_X86_64
+
+static void
+enter_lmode(struct kvm_vcpu *vcpu)
+{
+ uint32_t guest_tr_ar;
+
+ guest_tr_ar = vmcs_read32(GUEST_TR_AR_BYTES);
+ if ((guest_tr_ar & AR_TYPE_MASK) != AR_TYPE_BUSY_64_TSS) {
+ cmn_err(CE_NOTE, "%s: tss fixup for long mode. \n",
+ __func__);
+ vmcs_write32(GUEST_TR_AR_BYTES,
+ (guest_tr_ar & ~AR_TYPE_MASK) | AR_TYPE_BUSY_64_TSS);
+ }
+ vcpu->arch.efer |= EFER_LMA;
+ vmx_set_efer(vcpu, vcpu->arch.efer);
+}
+
+static void
+exit_lmode(struct kvm_vcpu *vcpu)
+{
+ vcpu->arch.efer &= ~EFER_LMA;
+
+ vmcs_write32(VM_ENTRY_CONTROLS,
+ vmcs_read32(VM_ENTRY_CONTROLS) & ~VM_ENTRY_IA32E_MODE);
+}
+
+#endif
+
+static uint64_t construct_eptp(unsigned long);
+
+static void
+vmx_flush_tlb(struct kvm_vcpu *vcpu)
+{
+ vpid_sync_vcpu_all(to_vmx(vcpu));
+ if (enable_ept)
+ ept_sync_context(construct_eptp(vcpu->arch.mmu.root_hpa));
+}
+
+static void
+vmx_decache_cr0_guest_bits(struct kvm_vcpu *vcpu)
+{
+ ulong cr0_guest_owned_bits = vcpu->arch.cr0_guest_owned_bits;
+
+ vcpu->arch.cr0 &= ~cr0_guest_owned_bits;
+ vcpu->arch.cr0 |= vmcs_readl(GUEST_CR0) & cr0_guest_owned_bits;
+}
+
+static void
+vmx_decache_cr4_guest_bits(struct kvm_vcpu *vcpu)
+{
+ ulong cr4_guest_owned_bits = vcpu->arch.cr4_guest_owned_bits;
+
+ vcpu->arch.cr4 &= ~cr4_guest_owned_bits;
+ vcpu->arch.cr4 |= vmcs_readl(GUEST_CR4) & cr4_guest_owned_bits;
+}
+
+static void
+ept_load_pdptrs(struct kvm_vcpu *vcpu)
+{
+ if (!test_bit(VCPU_EXREG_PDPTR,
+ (unsigned long *)&vcpu->arch.regs_dirty))
+ return;
+
+ if (is_paging(vcpu) && is_pae(vcpu) && !is_long_mode(vcpu)) {
+ vmcs_write64(GUEST_PDPTR0, vcpu->arch.pdptrs[0]);
+ vmcs_write64(GUEST_PDPTR1, vcpu->arch.pdptrs[1]);
+ vmcs_write64(GUEST_PDPTR2, vcpu->arch.pdptrs[2]);
+ vmcs_write64(GUEST_PDPTR3, vcpu->arch.pdptrs[3]);
+ }
+}
+
+static void
+ept_save_pdptrs(struct kvm_vcpu *vcpu)
+{
+ if (is_paging(vcpu) && is_pae(vcpu) && !is_long_mode(vcpu)) {
+ vcpu->arch.pdptrs[0] = vmcs_read64(GUEST_PDPTR0);
+ vcpu->arch.pdptrs[1] = vmcs_read64(GUEST_PDPTR1);
+ vcpu->arch.pdptrs[2] = vmcs_read64(GUEST_PDPTR2);
+ vcpu->arch.pdptrs[3] = vmcs_read64(GUEST_PDPTR3);
+ }
+
+ __set_bit(VCPU_EXREG_PDPTR, (unsigned long *)&vcpu->arch.regs_avail);
+ __set_bit(VCPU_EXREG_PDPTR, (unsigned long *)&vcpu->arch.regs_dirty);
+}
+
+static void vmx_set_cr4(struct kvm_vcpu *, unsigned long);
+
+static void
+ept_update_paging_mode_cr0(unsigned long *hw_cr0,
+ unsigned long cr0, struct kvm_vcpu *vcpu)
+{
+ if (!(cr0 & X86_CR0_PG)) {
+ /* From paging/starting to nonpaging */
+ vmcs_write32(CPU_BASED_VM_EXEC_CONTROL,
+ vmcs_read32(CPU_BASED_VM_EXEC_CONTROL) |
+ (CPU_BASED_CR3_LOAD_EXITING |
+ CPU_BASED_CR3_STORE_EXITING));
+ vcpu->arch.cr0 = cr0;
+ vmx_set_cr4(vcpu, kvm_read_cr4(vcpu));
+ } else if (!is_paging(vcpu)) {
+ /* From nonpaging to paging */
+ vmcs_write32(CPU_BASED_VM_EXEC_CONTROL,
+ vmcs_read32(CPU_BASED_VM_EXEC_CONTROL) &
+ ~(CPU_BASED_CR3_LOAD_EXITING |
+ CPU_BASED_CR3_STORE_EXITING));
+ vcpu->arch.cr0 = cr0;
+ vmx_set_cr4(vcpu, kvm_read_cr4(vcpu));
+ }
+
+ if (!(cr0 & X86_CR0_WP))
+ *hw_cr0 &= ~X86_CR0_WP;
+}
+
+static void
+vmx_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0)
+{
+ struct vcpu_vmx *vmx = to_vmx(vcpu);
+ unsigned long hw_cr0;
+
+ if (enable_unrestricted_guest) {
+ hw_cr0 = (cr0 & ~KVM_GUEST_CR0_MASK_UNRESTRICTED_GUEST) |
+ KVM_VM_CR0_ALWAYS_ON_UNRESTRICTED_GUEST;
+ } else {
+ hw_cr0 = (cr0 & ~KVM_GUEST_CR0_MASK) | KVM_VM_CR0_ALWAYS_ON;
+ }
+
+ if (vmx->rmode.vm86_active && (cr0 & X86_CR0_PE))
+ enter_pmode(vcpu);
+
+ if (!vmx->rmode.vm86_active && !(cr0 & X86_CR0_PE))
+ enter_rmode(vcpu);
+
+#ifdef CONFIG_X86_64
+ if (vcpu->arch.efer & EFER_LME) {
+ if (!is_paging(vcpu) && (cr0 & X86_CR0_PG))
+ enter_lmode(vcpu);
+ if (is_paging(vcpu) && !(cr0 & X86_CR0_PG))
+ exit_lmode(vcpu);
+ }
+#endif
+
+ if (enable_ept)
+ ept_update_paging_mode_cr0(&hw_cr0, cr0, vcpu);
+
+ if (!vcpu->fpu_active)
+ hw_cr0 |= X86_CR0_TS | X86_CR0_MP;
+
+ vmcs_writel(CR0_READ_SHADOW, cr0);
+ vmcs_writel(GUEST_CR0, hw_cr0);
+ vcpu->arch.cr0 = cr0;
+}
+
+static uint64_t
+construct_eptp(unsigned long root_hpa)
+{
+ uint64_t eptp;
+
+ /* TODO write the value reading from MSR */
+ eptp = VMX_EPT_DEFAULT_MT |
+ VMX_EPT_DEFAULT_GAW << VMX_EPT_GAW_EPTP_SHIFT;
+ eptp |= (root_hpa & PAGEMASK);
+
+ return (eptp);
+}
+
+static void
+vmx_set_cr3(struct kvm_vcpu *vcpu, unsigned long cr3)
+{
+ unsigned long guest_cr3;
+ uint64_t eptp;
+
+ guest_cr3 = cr3;
+
+ if (enable_ept) {
+ eptp = construct_eptp(cr3);
+ vmcs_write64(EPT_POINTER, eptp);
+ guest_cr3 = is_paging(vcpu) ?
+ vcpu->arch.cr3 : vcpu->kvm->arch.ept_identity_map_addr;
+ ept_load_pdptrs(vcpu);
+ }
+
+ vmx_flush_tlb(vcpu);
+ vmcs_writel(GUEST_CR3, guest_cr3);
+}
+
+static void
+vmx_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
+{
+ unsigned long hw_cr4 = cr4 | (to_vmx(vcpu)->rmode.vm86_active ?
+ KVM_RMODE_VM_CR4_ALWAYS_ON : KVM_PMODE_VM_CR4_ALWAYS_ON);
+
+ vcpu->arch.cr4 = cr4;
+ if (enable_ept) {
+ if (!is_paging(vcpu)) {
+ hw_cr4 &= ~X86_CR4_PAE;
+ hw_cr4 |= X86_CR4_PSE;
+ } else if (!(cr4 & X86_CR4_PAE)) {
+ hw_cr4 &= ~X86_CR4_PAE;
+ }
+ }
+
+ vmcs_writel(CR4_READ_SHADOW, cr4);
+ vmcs_writel(GUEST_CR4, hw_cr4);
+}
+
+static uint64_t
+vmx_get_segment_base(struct kvm_vcpu *vcpu, int seg)
+{
+ struct kvm_vmx_segment_field *sf = &kvm_vmx_segment_fields[seg];
+
+ return (vmcs_readl(sf->base));
+}
+
+static void
+vmx_get_segment(struct kvm_vcpu *vcpu,
+ struct kvm_segment *var, int seg)
+{
+ struct kvm_vmx_segment_field *sf = &kvm_vmx_segment_fields[seg];
+ uint32_t ar;
+
+ var->base = vmcs_readl(sf->base);
+ var->limit = vmcs_read32(sf->limit);
+ var->selector = vmcs_read16(sf->selector);
+ ar = vmcs_read32(sf->ar_bytes);
+
+ if ((ar & AR_UNUSABLE_MASK) && !emulate_invalid_guest_state)
+ ar = 0;
+ var->type = ar & 15;
+ var->s = (ar >> 4) & 1;
+ var->dpl = (ar >> 5) & 3;
+ var->present = (ar >> 7) & 1;
+ var->avl = (ar >> 12) & 1;
+ var->l = (ar >> 13) & 1;
+ var->db = (ar >> 14) & 1;
+ var->g = (ar >> 15) & 1;
+ var->unusable = (ar >> 16) & 1;
+}
+
+
+static int
+vmx_get_cpl(struct kvm_vcpu *vcpu)
+{
+ if (!is_protmode(vcpu))
+ return (0);
+
+ if (vmx_get_rflags(vcpu) & X86_EFLAGS_VM) /* if virtual 8086 */
+ return (3);
+
+ return (vmcs_read16(GUEST_CS_SELECTOR) & 3);
+}
+
+static uint32_t
+vmx_segment_access_rights(struct kvm_segment *var)
+{
+ uint32_t ar;
+
+ if (var->unusable)
+ ar = 1 << 16;
+ else {
+ ar = var->type & 15;
+ ar |= (var->s & 1) << 4;
+ ar |= (var->dpl & 3) << 5;
+ ar |= (var->present & 1) << 7;
+ ar |= (var->avl & 1) << 12;
+ ar |= (var->l & 1) << 13;
+ ar |= (var->db & 1) << 14;
+ ar |= (var->g & 1) << 15;
+ }
+ if (ar == 0) /* a 0 value means unusable */
+ ar = AR_UNUSABLE_MASK;
+
+ return (ar);
+}
+
+static void
+vmx_set_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg)
+{
+ struct vcpu_vmx *vmx = to_vmx(vcpu);
+ struct kvm_vmx_segment_field *sf = &kvm_vmx_segment_fields[seg];
+ uint32_t ar;
+
+ if (vmx->rmode.vm86_active && seg == VCPU_SREG_TR) {
+ vmx->rmode.tr.selector = var->selector;
+ vmx->rmode.tr.base = var->base;
+ vmx->rmode.tr.limit = var->limit;
+ vmx->rmode.tr.ar = vmx_segment_access_rights(var);
+ return;
+ }
+ vmcs_writel(sf->base, var->base);
+ vmcs_write32(sf->limit, var->limit);
+ vmcs_write16(sf->selector, var->selector);
+ if (vmx->rmode.vm86_active && var->s) {
+ /*
+ * Hack real-mode segments into vm86 compatibility.
+ */
+ if (var->base == 0xffff0000 && var->selector == 0xf000)
+ vmcs_writel(sf->base, 0xf0000);
+ ar = 0xf3;
+ } else
+ ar = vmx_segment_access_rights(var);
+
+
+ /*
+ * Fix the "Accessed" bit in AR field of segment registers for older
+ * qemu binaries.
+ * IA32 arch specifies that at the time of processor reset the
+ * "Accessed" bit in the AR field of segment registers is 1. And qemu
+ * is setting it to 0 in the usedland code. This causes invalid guest
+ * state vmexit when "unrestricted guest" mode is turned on.
+ * Fix for this setup issue in cpu_reset is being pushed in the qemu
+ * tree. Newer qemu binaries with that qemu fix would not need this
+ * kvm hack.
+ */
+ if (enable_unrestricted_guest && (seg != VCPU_SREG_LDTR))
+ ar |= 0x1; /* Accessed */
+
+ vmcs_write32(sf->ar_bytes, ar);
+}
+
+static void
+vmx_get_cs_db_l_bits(struct kvm_vcpu *vcpu, int *db, int *l)
+{
+ uint32_t ar = vmcs_read32(GUEST_CS_AR_BYTES);
+
+ *db = (ar >> 14) & 1;
+ *l = (ar >> 13) & 1;
+}
+
+static void
+vmx_get_idt(struct kvm_vcpu *vcpu, struct descriptor_table *dt)
+{
+ dt->limit = vmcs_read32(GUEST_IDTR_LIMIT);
+ dt->base = vmcs_readl(GUEST_IDTR_BASE);
+}
+
+static void
+vmx_set_idt(struct kvm_vcpu *vcpu, struct descriptor_table *dt)
+{
+ vmcs_write32(GUEST_IDTR_LIMIT, dt->limit);
+ vmcs_writel(GUEST_IDTR_BASE, dt->base);
+}
+
+static void
+vmx_get_gdt(struct kvm_vcpu *vcpu, struct descriptor_table *dt)
+{
+ dt->limit = vmcs_read32(GUEST_GDTR_LIMIT);
+ dt->base = vmcs_readl(GUEST_GDTR_BASE);
+}
+
+static void
+vmx_set_gdt(struct kvm_vcpu *vcpu, struct descriptor_table *dt)
+{
+ vmcs_write32(GUEST_GDTR_LIMIT, dt->limit);
+ vmcs_writel(GUEST_GDTR_BASE, dt->base);
+}
+
+static int
+rmode_segment_valid(struct kvm_vcpu *vcpu, int seg)
+{
+ struct kvm_segment var;
+ uint32_t ar;
+
+ vmx_get_segment(vcpu, &var, seg);
+ ar = vmx_segment_access_rights(&var);
+
+ if (var.base != (var.selector << 4))
+ return (0);
+ if (var.limit != 0xffff)
+ return (0);
+ if (ar != 0xf3)
+ return (0);
+
+ return (1);
+}
+
+static int
+code_segment_valid(struct kvm_vcpu *vcpu)
+{
+ struct kvm_segment cs;
+ unsigned int cs_rpl;
+
+ vmx_get_segment(vcpu, &cs, VCPU_SREG_CS);
+ cs_rpl = cs.selector & SELECTOR_RPL_MASK;
+
+ if (cs.unusable)
+ return (0);
+ if (~cs.type & (AR_TYPE_CODE_MASK|AR_TYPE_ACCESSES_MASK))
+ return (0);
+ if (!cs.s)
+ return (0);
+
+ if (cs.type & AR_TYPE_WRITEABLE_MASK) {
+ if (cs.dpl > cs_rpl)
+ return (0);
+ } else {
+ if (cs.dpl != cs_rpl)
+ return (0);
+ }
+
+ if (!cs.present)
+ return (0);
+
+ /*
+ * TODO: Add Reserved field check, this'll require a new member in the
+ * kvm_segment_field structure
+ */
+ return (1);
+}
+
+static int
+stack_segment_valid(struct kvm_vcpu *vcpu)
+{
+ struct kvm_segment ss;
+ unsigned int ss_rpl;
+
+ vmx_get_segment(vcpu, &ss, VCPU_SREG_SS);
+ ss_rpl = ss.selector & SELECTOR_RPL_MASK;
+
+ if (ss.unusable)
+ return (1);
+ if (ss.type != 3 && ss.type != 7)
+ return (0);
+ if (!ss.s)
+ return (0);
+ if (ss.dpl != ss_rpl) /* DPL != RPL */
+ return (0);
+ if (!ss.present)
+ return (0);
+
+ return (1);
+}
+
+static int
+data_segment_valid(struct kvm_vcpu *vcpu, int seg)
+{
+ struct kvm_segment var;
+ unsigned int rpl;
+
+ vmx_get_segment(vcpu, &var, seg);
+ rpl = var.selector & SELECTOR_RPL_MASK;
+
+ if (var.unusable)
+ return (1);
+
+ if (!var.s)
+ return (0);
+
+ if (!var.present)
+ return (0);
+
+ if (~var.type & (AR_TYPE_CODE_MASK|AR_TYPE_WRITEABLE_MASK)) {
+ if (var.dpl < rpl) /* DPL < RPL */
+ return (0);
+ }
+
+ /*
+ * TODO: Add other members to kvm_segment_field to allow checking for
+ * other access rights flags
+ */
+ return (1);
+}
+
+static int
+tr_valid(struct kvm_vcpu *vcpu)
+{
+ struct kvm_segment tr;
+
+ vmx_get_segment(vcpu, &tr, VCPU_SREG_TR);
+
+ if (tr.unusable)
+ return (0);
+ if (tr.selector & SELECTOR_TI_MASK) /* TI = 1 */
+ return (0);
+ if (tr.type != 3 && tr.type != 11)
+ return (0); /* TODO: Check if guest is in IA32e mode */
+ if (!tr.present)
+ return (0);
+
+ return (1);
+}
+
+static int
+ldtr_valid(struct kvm_vcpu *vcpu)
+{
+ struct kvm_segment ldtr;
+
+ vmx_get_segment(vcpu, &ldtr, VCPU_SREG_LDTR);
+
+ if (ldtr.unusable)
+ return (1);
+ if (ldtr.selector & SELECTOR_TI_MASK) /* TI = 1 */
+ return (0);
+ if (ldtr.type != 2)
+ return (0);
+ if (!ldtr.present)
+ return (0);
+
+ return (1);
+}
+
+static int
+cs_ss_rpl_check(struct kvm_vcpu *vcpu)
+{
+ struct kvm_segment cs, ss;
+
+ vmx_get_segment(vcpu, &cs, VCPU_SREG_CS);
+ vmx_get_segment(vcpu, &ss, VCPU_SREG_SS);
+
+ return ((cs.selector & SELECTOR_RPL_MASK) ==
+ (ss.selector & SELECTOR_RPL_MASK));
+}
+
+/*
+ * Check if guest state is valid. Returns true if valid, false if
+ * not.
+ * We assume that registers are always usable
+ */
+static int
+guest_state_valid(struct kvm_vcpu *vcpu)
+{
+ if (!is_protmode(vcpu)) {
+ /* real mode guest state checks */
+ if (!rmode_segment_valid(vcpu, VCPU_SREG_CS))
+ return (0);
+ if (!rmode_segment_valid(vcpu, VCPU_SREG_SS))
+ return (0);
+ if (!rmode_segment_valid(vcpu, VCPU_SREG_DS))
+ return (0);
+ if (!rmode_segment_valid(vcpu, VCPU_SREG_ES))
+ return (0);
+ if (!rmode_segment_valid(vcpu, VCPU_SREG_FS))
+ return (0);
+ if (!rmode_segment_valid(vcpu, VCPU_SREG_GS))
+ return (0);
+ } else {
+ /* protected mode guest state checks */
+ if (!cs_ss_rpl_check(vcpu))
+ return (0);
+ if (!code_segment_valid(vcpu))
+ return (0);
+ if (!stack_segment_valid(vcpu))
+ return (0);
+ if (!data_segment_valid(vcpu, VCPU_SREG_DS))
+ return (0);
+ if (!data_segment_valid(vcpu, VCPU_SREG_ES))
+ return (0);
+ if (!data_segment_valid(vcpu, VCPU_SREG_FS))
+ return (0);
+ if (!data_segment_valid(vcpu, VCPU_SREG_GS))
+ return (0);
+ if (!tr_valid(vcpu))
+ return (0);
+ if (!ldtr_valid(vcpu))
+ return (0);
+ }
+
+ /*
+ * TODO:
+ * - Add checks on RIP
+ * - Add checks on RFLAGS
+ */
+
+ return (1);
+}
+
+static int
+init_rmode_tss(struct kvm *kvm)
+{
+ gfn_t fn = rmode_tss_base(kvm) >> PAGESHIFT;
+ uint16_t data = 0;
+ int ret = 0;
+ int r;
+
+ r = kvm_clear_guest_page(kvm, fn, 0, PAGESIZE);
+ if (r < 0)
+ goto out;
+ data = TSS_BASE_SIZE + TSS_REDIRECTION_SIZE;
+ r = kvm_write_guest_page(kvm, fn++, &data,
+ TSS_IOPB_BASE_OFFSET, sizeof (uint16_t));
+
+ if (r < 0)
+ goto out;
+ r = kvm_clear_guest_page(kvm, fn++, 0, PAGESIZE);
+ if (r < 0)
+ goto out;
+ r = kvm_clear_guest_page(kvm, fn, 0, PAGESIZE);
+ if (r < 0)
+ goto out;
+ data = ~0;
+ r = kvm_write_guest_page(kvm, fn, &data,
+ RMODE_TSS_SIZE - 2 * PAGESIZE - 1, sizeof (uint8_t));
+
+ if (r < 0)
+ goto out;
+
+ ret = 1;
+out:
+ return (ret);
+}
+
+
+static int
+init_rmode_identity_map(struct kvm *kvm)
+{
+ int i, r, ret;
+ pfn_t identity_map_pfn;
+ uint32_t tmp;
+
+ if (!enable_ept)
+ return (1);
+ if ((!kvm->arch.ept_identity_pagetable)) {
+ cmn_err(CE_WARN, "EPT: identity-mapping pagetable "
+ "haven't been allocated!\n");
+ return (0);
+ }
+ if ((kvm->arch.ept_identity_pagetable_done))
+ return (1);
+
+ ret = 0;
+ identity_map_pfn = kvm->arch.ept_identity_map_addr >> PAGESHIFT;
+ r = kvm_clear_guest_page(kvm, identity_map_pfn, 0, PAGESIZE);
+ if (r < 0)
+ goto out;
+
+ /* Set up identity-mapping pagetable for EPT in real mode */
+ for (i = 0; i < PT32_ENT_PER_PAGE; i++) {
+ tmp = (i << 22) + (PT_VALID | PT_WRITABLE | PT_USER |
+ PT_REF | PT_MOD | PT_PAGESIZE);
+
+ r = kvm_write_guest_page(kvm, identity_map_pfn,
+ &tmp, i * sizeof (tmp), sizeof (tmp));
+
+ if (r < 0)
+ goto out;
+ }
+ kvm->arch.ept_identity_pagetable_done = 1;
+ ret = 1;
+out:
+ return (ret);
+}
+
+static void
+seg_setup(int seg)
+{
+ struct kvm_vmx_segment_field *sf = &kvm_vmx_segment_fields[seg];
+ unsigned int ar;
+
+ vmcs_write16(sf->selector, 0);
+ vmcs_writel(sf->base, 0);
+ vmcs_write32(sf->limit, 0xffff);
+
+ if (enable_unrestricted_guest) {
+ ar = 0x93;
+ if (seg == VCPU_SREG_CS)
+ ar |= 0x08; /* code segment */
+ } else
+ ar = 0xf3;
+
+ vmcs_write32(sf->ar_bytes, ar);
+}
+
+static int
+alloc_apic_access_page(struct kvm *kvm)
+{
+ struct kvm_userspace_memory_region kvm_userspace_mem;
+ int r = 0;
+
+ memset(&kvm_userspace_mem, 0,
+ sizeof (struct kvm_userspace_memory_region));
+
+ mutex_enter(&kvm->slots_lock);
+ if (kvm->arch.apic_access_page)
+ goto out;
+ kvm_userspace_mem.slot = APIC_ACCESS_PAGE_PRIVATE_MEMSLOT;
+ kvm_userspace_mem.flags = 0;
+ kvm_userspace_mem.guest_phys_addr = 0xfee00000ULL;
+ kvm_userspace_mem.memory_size = PAGESIZE;
+ r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, 0);
+ if (r)
+ goto out;
+
+ kvm->arch.apic_access_page = gfn_to_page(kvm, 0xfee00);
+out:
+ mutex_exit(&kvm->slots_lock);
+ return (r);
+}
+
+static int
+alloc_identity_pagetable(struct kvm *kvm)
+{
+ struct kvm_userspace_memory_region kvm_userspace_mem;
+ int r = 0;
+
+ mutex_enter(&kvm->slots_lock);
+ if (kvm->arch.ept_identity_pagetable)
+ goto out;
+ kvm_userspace_mem.slot = IDENTITY_PAGETABLE_PRIVATE_MEMSLOT;
+ kvm_userspace_mem.flags = 0;
+ kvm_userspace_mem.guest_phys_addr =
+ kvm->arch.ept_identity_map_addr;
+ kvm_userspace_mem.memory_size = PAGESIZE;
+
+ kvm_userspace_mem.userspace_addr = 0;
+ r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, 0);
+ if (r)
+ goto out;
+
+ kvm->arch.ept_identity_pagetable = gfn_to_page(kvm,
+ kvm->arch.ept_identity_map_addr >> PAGESHIFT);
+out:
+ mutex_exit(&kvm->slots_lock);
+ return (r);
+}
+
+static void
+allocate_vpid(struct vcpu_vmx *vmx)
+{
+ int vpid;
+
+ vmx->vpid = 0;
+ if (!enable_vpid)
+ return;
+ mutex_enter(&vmx_vpid_lock);
+ vpid = find_first_zero_bit(vmx_vpid_bitmap, VMX_NR_VPIDS);
+ if (vpid < VMX_NR_VPIDS) {
+ vmx->vpid = vpid;
+ __set_bit(vpid, vmx_vpid_bitmap);
+ }
+ mutex_exit(&vmx_vpid_lock);
+}
+
+static void
+__vmx_disable_intercept_for_msr(unsigned long *msr_bitmap, uint32_t msr)
+{
+ int f = sizeof (unsigned long);
+
+ if (!cpu_has_vmx_msr_bitmap())
+ return;
+
+ /*
+ * See Intel PRM Vol. 3, 20.6.9 (MSR-Bitmap Address). Early manuals
+ * have the write-low and read-high bitmap offsets the wrong way round.
+ * We can control MSRs 0x00000000-0x00001fff and 0xc0000000-0xc0001fff.
+ */
+ if (msr <= 0x1fff) {
+ __clear_bit(msr, msr_bitmap + 0x000 / f); /* read-low */
+ __clear_bit(msr, msr_bitmap + 0x800 / f); /* write-low */
+ } else if ((msr >= 0xc0000000) && (msr <= 0xc0001fff)) {
+ msr &= 0x1fff;
+ __clear_bit(msr, msr_bitmap + 0x400 / f); /* read-high */
+ __clear_bit(msr, msr_bitmap + 0xc00 / f); /* write-high */
+ }
+}
+
+static void
+vmx_disable_intercept_for_msr(uint32_t msr, int longmode_only)
+{
+ if (!longmode_only)
+ __vmx_disable_intercept_for_msr(vmx_msr_bitmap_legacy, msr);
+ __vmx_disable_intercept_for_msr(vmx_msr_bitmap_longmode, msr);
+}
+
+/*
+ * Sets up the vmcs for emulated real mode.
+ */
+static int
+vmx_vcpu_setup(struct vcpu_vmx *vmx)
+{
+ uint32_t host_sysenter_cs, msr_low, msr_high;
+ uint32_t junk;
+ uint64_t host_pat, tsc_this, tsc_base;
+ volatile uint64_t a;
+ struct descriptor_table dt;
+ int i;
+ unsigned long kvm_vmx_return;
+ uint32_t exec_control;
+
+ /* I/O */
+ vmcs_write64(IO_BITMAP_A, kvm_va2pa((caddr_t)vmx_io_bitmap_a));
+ vmcs_write64(IO_BITMAP_B, kvm_va2pa((caddr_t)vmx_io_bitmap_b));
+
+ if (cpu_has_vmx_msr_bitmap()) {
+ vmcs_write64(MSR_BITMAP,
+ kvm_va2pa((caddr_t)vmx_msr_bitmap_legacy));
+ }
+
+ vmcs_write64(VMCS_LINK_POINTER, -1ull); /* 22.3.1.5 */
+
+ /* Control */
+ vmcs_write32(PIN_BASED_VM_EXEC_CONTROL,
+ vmcs_config.pin_based_exec_ctrl);
+
+ exec_control = vmcs_config.cpu_based_exec_ctrl;
+ if (!vm_need_tpr_shadow(vmx->vcpu.kvm)) {
+ exec_control &= ~CPU_BASED_TPR_SHADOW;
+#ifdef CONFIG_X86_64
+ exec_control |= CPU_BASED_CR8_STORE_EXITING |
+ CPU_BASED_CR8_LOAD_EXITING;
+#endif
+ }
+
+ if (!enable_ept)
+ exec_control |= CPU_BASED_CR3_STORE_EXITING |
+ CPU_BASED_CR3_LOAD_EXITING |
+ CPU_BASED_INVLPG_EXITING;
+ vmcs_write32(CPU_BASED_VM_EXEC_CONTROL, exec_control);
+
+ if (cpu_has_secondary_exec_ctrls()) {
+ exec_control = vmcs_config.cpu_based_2nd_exec_ctrl;
+ if (!vm_need_virtualize_apic_accesses(vmx->vcpu.kvm))
+ exec_control &=
+ ~SECONDARY_EXEC_VIRTUALIZE_APIC_ACCESSES;
+ if (vmx->vpid == 0)
+ exec_control &= ~SECONDARY_EXEC_ENABLE_VPID;
+ if (!enable_ept) {
+ exec_control &= ~SECONDARY_EXEC_ENABLE_EPT;
+ enable_unrestricted_guest = 0;
+ }
+ if (!enable_unrestricted_guest)
+ exec_control &= ~SECONDARY_EXEC_UNRESTRICTED_GUEST;
+
+ if (!ple_gap)
+ exec_control &= ~SECONDARY_EXEC_PAUSE_LOOP_EXITING;
+ vmcs_write32(SECONDARY_VM_EXEC_CONTROL, exec_control);
+ }
+
+ if (ple_gap) {
+ vmcs_write32(PLE_GAP, ple_gap);
+ vmcs_write32(PLE_WINDOW, ple_window);
+ }
+
+ vmcs_write32(PAGE_FAULT_ERROR_CODE_MASK, !!bypass_guest_pf);
+ vmcs_write32(PAGE_FAULT_ERROR_CODE_MATCH, !!bypass_guest_pf);
+ vmcs_write32(CR3_TARGET_COUNT, 0); /* 22.2.1 */
+
+ vmcs_writel(HOST_CR0, read_cr0()); /* 22.2.3 */
+ vmcs_writel(HOST_CR4, read_cr4()); /* 22.2.3, 22.2.5 */
+ vmcs_writel(HOST_CR3, read_cr3()); /* 22.2.3 FIXME: shadow tables */
+
+ vmcs_write16(HOST_CS_SELECTOR, KCS_SEL); /* 22.2.4 */
+#ifndef XXX
+ vmcs_write16(HOST_DS_SELECTOR, KDS_SEL); /* 22.2.4 */
+ vmcs_write16(HOST_ES_SELECTOR, KDS_SEL); /* 22.2.4 */
+ vmcs_write16(HOST_FS_SELECTOR, kvm_read_fs()); /* 22.2.4 */
+ vmcs_write16(HOST_GS_SELECTOR, kvm_read_gs()); /* 22.2.4 */
+
+#else
+ XXX_KVM_PROBE;
+ vmcs_write16(HOST_DS_SELECTOR, 0x4b); /* 22.2.4 */
+ vmcs_write16(HOST_ES_SELECTOR, 0x4b); /* 22.2.4 */
+ vmcs_write16(HOST_FS_SELECTOR, 0); /* 22.2.4 */
+ vmcs_write16(HOST_GS_SELECTOR, 0); /* 22.2.4 */
+#endif
+ vmcs_write16(HOST_SS_SELECTOR, KDS_SEL); /* 22.2.4 */
+#ifdef CONFIG_X86_64
+ rdmsrl(MSR_FS_BASE, a);
+ vmcs_writel(HOST_FS_BASE, a); /* 22.2.4 */
+ rdmsrl(MSR_GS_BASE, a);
+ vmcs_writel(HOST_GS_BASE, a); /* 22.2.4 */
+#else
+ vmcs_writel(HOST_FS_BASE, 0); /* 22.2.4 */
+ vmcs_writel(HOST_GS_BASE, 0); /* 22.2.4 */
+#endif
+
+ vmcs_write16(HOST_TR_SELECTOR, KTSS_SEL); /* 22.2.4 */
+
+ kvm_get_idt(&dt);
+ vmcs_writel(HOST_IDTR_BASE, dt.base); /* 22.2.4 */
+
+ __asm__("mov $.Lkvm_vmx_return, %0" : "=r"(kvm_vmx_return));
+ vmcs_writel(HOST_RIP, kvm_vmx_return); /* 22.2.5 */
+ vmcs_write32(VM_EXIT_MSR_STORE_COUNT, 0);
+ vmcs_write32(VM_EXIT_MSR_LOAD_COUNT, 0);
+ vmcs_write32(VM_ENTRY_MSR_LOAD_COUNT, 0);
+
+ rdmsr(MSR_IA32_SYSENTER_CS, host_sysenter_cs, junk);
+ vmcs_write32(HOST_IA32_SYSENTER_CS, host_sysenter_cs);
+ rdmsrl(MSR_IA32_SYSENTER_ESP, a);
+ vmcs_writel(HOST_IA32_SYSENTER_ESP, a); /* 22.2.3 */
+ rdmsrl(MSR_IA32_SYSENTER_EIP, a);
+ vmcs_writel(HOST_IA32_SYSENTER_EIP, a); /* 22.2.3 */
+
+ if (vmcs_config.vmexit_ctrl & VM_EXIT_LOAD_IA32_PAT) {
+ rdmsr(MSR_IA32_CR_PAT, msr_low, msr_high);
+ host_pat = msr_low | ((uint64_t) msr_high << 32);
+ vmcs_write64(HOST_IA32_PAT, host_pat);
+ }
+ if (vmcs_config.vmentry_ctrl & VM_ENTRY_LOAD_IA32_PAT) {
+ rdmsr(MSR_IA32_CR_PAT, msr_low, msr_high);
+ host_pat = msr_low | ((uint64_t) msr_high << 32);
+ /* Write the default value follow host pat */
+ vmcs_write64(GUEST_IA32_PAT, host_pat);
+ /* Keep arch.pat sync with GUEST_IA32_PAT */
+ vmx->vcpu.arch.pat = host_pat;
+ }
+
+ for (i = 0; i < NR_VMX_MSR; ++i) {
+ uint32_t index = vmx_msr_index[i];
+ uint32_t data_low, data_high;
+ int j = vmx->nmsrs;
+
+ if (rdmsr_safe(index, &data_low, &data_high) < 0)
+ continue;
+ if (wrmsr_safe(index, data_low, data_high) < 0)
+ continue;
+ vmx->guest_msrs[j].index = i;
+ vmx->guest_msrs[j].data = 0;
+ vmx->guest_msrs[j].mask = -1ull;
+ ++vmx->nmsrs;
+ }
+
+ vmcs_write32(VM_EXIT_CONTROLS, vmcs_config.vmexit_ctrl);
+
+ /* 22.2.1, 20.8.1 */
+ vmcs_write32(VM_ENTRY_CONTROLS, vmcs_config.vmentry_ctrl);
+
+ vmcs_writel(CR0_GUEST_HOST_MASK, ~0UL);
+ vmx->vcpu.arch.cr4_guest_owned_bits = KVM_CR4_GUEST_OWNED_BITS;
+ if (enable_ept)
+ vmx->vcpu.arch.cr4_guest_owned_bits |= X86_CR4_PGE;
+ vmcs_writel(CR4_GUEST_HOST_MASK, ~vmx->vcpu.arch.cr4_guest_owned_bits);
+
+ tsc_base = vmx->vcpu.kvm->arch.vm_init_tsc;
+ rdtscll(tsc_this);
+ if (tsc_this < vmx->vcpu.kvm->arch.vm_init_tsc)
+ tsc_base = tsc_this;
+
+ guest_write_tsc(0, tsc_base);
+
+ return (0);
+}
+
+static int
+init_rmode(struct kvm *kvm)
+{
+ if (!init_rmode_tss(kvm))
+ return (0);
+
+ if (!init_rmode_identity_map(kvm))
+ return (0);
+
+ return (1);
+}
+
+static int
+vmx_vcpu_reset(struct kvm_vcpu *vcpu)
+{
+ struct vcpu_vmx *vmx = to_vmx(vcpu);
+ uint64_t msr;
+ int ret, idx;
+ page_t *ptp;
+
+ vcpu->arch.regs_avail = ~((1 << VCPU_REGS_RIP) | (1 << VCPU_REGS_RSP));
+#ifdef XXX
+ idx = srcu_read_lock(&vcpu->kvm->srcu);
+#else
+ XXX_KVM_SYNC_PROBE;
+#endif
+
+ if (!init_rmode(vmx->vcpu.kvm)) {
+ ret = -ENOMEM;
+ goto out;
+ }
+
+ vmx->rmode.vm86_active = 0;
+ vmx->soft_vnmi_blocked = 0;
+
+ vmx->vcpu.arch.regs[VCPU_REGS_RDX] = get_rdx_init_val();
+ kvm_set_cr8(&vmx->vcpu, 0);
+ msr = 0xfee00000 | MSR_IA32_APICBASE_ENABLE;
+
+ if (kvm_vcpu_is_bsp(&vmx->vcpu))
+ msr |= MSR_IA32_APICBASE_BSP;
+
+ kvm_set_apic_base(&vmx->vcpu, msr);
+
+ fx_init(&vmx->vcpu);
+
+ seg_setup(VCPU_SREG_CS);
+ /*
+ * GUEST_CS_BASE should really be 0xffff0000, but VT vm86 mode
+ * insists on having GUEST_CS_BASE == GUEST_CS_SELECTOR << 4. Sigh.
+ */
+ if (kvm_vcpu_is_bsp(&vmx->vcpu)) {
+ vmcs_write16(GUEST_CS_SELECTOR, 0xf000);
+#ifndef XXX
+ vmcs_writel(GUEST_CS_BASE, 0x000f0000);
+#else
+ vmcs_writel(GUEST_CS_BASE, 0xffff0000);
+#endif
+ } else {
+ vmcs_write16(GUEST_CS_SELECTOR,
+ vmx->vcpu.arch.sipi_vector << 8);
+ vmcs_writel(GUEST_CS_BASE, vmx->vcpu.arch.sipi_vector << 12);
+ }
+
+ seg_setup(VCPU_SREG_DS);
+ seg_setup(VCPU_SREG_ES);
+ seg_setup(VCPU_SREG_FS);
+ seg_setup(VCPU_SREG_GS);
+ seg_setup(VCPU_SREG_SS);
+
+ vmcs_write16(GUEST_TR_SELECTOR, 0);
+ vmcs_writel(GUEST_TR_BASE, 0);
+ vmcs_write32(GUEST_TR_LIMIT, 0xffff);
+ vmcs_write32(GUEST_TR_AR_BYTES, 0x008b);
+
+ vmcs_write16(GUEST_LDTR_SELECTOR, 0);
+ vmcs_writel(GUEST_LDTR_BASE, 0);
+ vmcs_write32(GUEST_LDTR_LIMIT, 0xffff);
+ vmcs_write32(GUEST_LDTR_AR_BYTES, 0x00082);
+
+ vmcs_write32(GUEST_SYSENTER_CS, 0);
+ vmcs_writel(GUEST_SYSENTER_ESP, 0);
+ vmcs_writel(GUEST_SYSENTER_EIP, 0);
+
+ vmcs_writel(GUEST_RFLAGS, 0x02);
+
+ if (kvm_vcpu_is_bsp(&vmx->vcpu))
+ kvm_rip_write(vcpu, 0xfff0);
+ else
+ kvm_rip_write(vcpu, 0);
+
+ kvm_register_write(vcpu, VCPU_REGS_RSP, 0);
+
+ vmcs_writel(GUEST_DR7, 0x400);
+
+ vmcs_writel(GUEST_GDTR_BASE, 0);
+ vmcs_write32(GUEST_GDTR_LIMIT, 0xffff);
+
+ vmcs_writel(GUEST_IDTR_BASE, 0);
+ vmcs_write32(GUEST_IDTR_LIMIT, 0xffff);
+
+ vmcs_write32(GUEST_ACTIVITY_STATE, 0);
+ vmcs_write32(GUEST_INTERRUPTIBILITY_INFO, 0);
+ vmcs_write32(GUEST_PENDING_DBG_EXCEPTIONS, 0);
+
+ /* Special registers */
+ vmcs_write64(GUEST_IA32_DEBUGCTL, 0);
+
+ setup_msrs(vmx);
+
+ vmcs_write32(VM_ENTRY_INTR_INFO_FIELD, 0); /* 22.2.1 */
+
+ if (cpu_has_vmx_tpr_shadow()) {
+ vmcs_write64(VIRTUAL_APIC_PAGE_ADDR, 0);
+ if (vm_need_tpr_shadow(vmx->vcpu.kvm)) {
+ ptp = page_numtopp_nolock(hat_getpfnum(kas.a_hat,
+ vmx->vcpu.arch.apic->regs));
+ vmcs_write64(VIRTUAL_APIC_PAGE_ADDR, page_to_phys(ptp));
+ }
+
+ vmcs_write32(TPR_THRESHOLD, 0);
+ }
+
+
+ if (vm_need_virtualize_apic_accesses(vmx->vcpu.kvm)) {
+ vmcs_write64(APIC_ACCESS_ADDR,
+ page_to_phys(vmx->vcpu.kvm->arch.apic_access_page));
+ }
+
+ if (vmx->vpid != 0)
+ vmcs_write16(VIRTUAL_PROCESSOR_ID, vmx->vpid);
+
+ vmx->vcpu.arch.cr0 = X86_CR0_NW | X86_CR0_CD | X86_CR0_ET;
+ vmx_set_cr0(&vmx->vcpu, kvm_read_cr0(vcpu)); /* enter rmode */
+ vmx_set_cr4(&vmx->vcpu, 0);
+ vmx_set_efer(&vmx->vcpu, 0);
+
+ vmx_fpu_activate(&vmx->vcpu);
+ update_exception_bitmap(&vmx->vcpu);
+ vpid_sync_vcpu_all(vmx);
+
+ ret = 0;
+
+ /* HACK: Don't enable emulation on guest boot/reset */
+ vmx->emulation_required = 0;
+
+out:
+#ifdef XXX
+ srcu_read_unlock(&vcpu->kvm->srcu, idx);
+#else
+ XXX_KVM_SYNC_PROBE;
+#endif
+ return (ret);
+}
+
+static void
+enable_irq_window(struct kvm_vcpu *vcpu)
+{
+ uint32_t cpu_based_vm_exec_control;
+
+ cpu_based_vm_exec_control = vmcs_read32(CPU_BASED_VM_EXEC_CONTROL);
+ cpu_based_vm_exec_control |= CPU_BASED_VIRTUAL_INTR_PENDING;
+ vmcs_write32(CPU_BASED_VM_EXEC_CONTROL, cpu_based_vm_exec_control);
+}
+
+static void
+enable_nmi_window(struct kvm_vcpu *vcpu)
+{
+ uint32_t cpu_based_vm_exec_control;
+
+ if (!cpu_has_virtual_nmis()) {
+ enable_irq_window(vcpu);
+ return;
+ }
+
+ cpu_based_vm_exec_control = vmcs_read32(CPU_BASED_VM_EXEC_CONTROL);
+ cpu_based_vm_exec_control |= CPU_BASED_VIRTUAL_NMI_PENDING;
+ vmcs_write32(CPU_BASED_VM_EXEC_CONTROL, cpu_based_vm_exec_control);
+}
+
+static void
+vmx_inject_irq(struct kvm_vcpu *vcpu)
+{
+ struct vcpu_vmx *vmx = to_vmx(vcpu);
+ uint32_t intr;
+ int irq = vcpu->arch.interrupt.nr;
+
+ KVM_TRACE1(inj__virq, int, irq);
+ KVM_VCPU_KSTAT_INC(vcpu, kvmvs_irq_injections);
+
+ if (vmx->rmode.vm86_active) {
+ vmx->rmode.irq.pending = 1;
+ vmx->rmode.irq.vector = irq;
+ vmx->rmode.irq.rip = kvm_rip_read(vcpu);
+ if (vcpu->arch.interrupt.soft)
+ vmx->rmode.irq.rip +=
+ vmx->vcpu.arch.event_exit_inst_len;
+ vmcs_write32(VM_ENTRY_INTR_INFO_FIELD,
+ irq | INTR_TYPE_SOFT_INTR | INTR_INFO_VALID_MASK);
+ vmcs_write32(VM_ENTRY_INSTRUCTION_LEN, 1);
+ kvm_rip_write(vcpu, vmx->rmode.irq.rip - 1);
+ return;
+ }
+ intr = irq | INTR_INFO_VALID_MASK;
+ if (vcpu->arch.interrupt.soft) {
+ intr |= INTR_TYPE_SOFT_INTR;
+ vmcs_write32(VM_ENTRY_INSTRUCTION_LEN,
+ vmx->vcpu.arch.event_exit_inst_len);
+ } else
+ intr |= INTR_TYPE_EXT_INTR;
+
+ vmcs_write32(VM_ENTRY_INTR_INFO_FIELD, intr);
+}
+
+static void
+vmx_inject_nmi(struct kvm_vcpu *vcpu)
+{
+ struct vcpu_vmx *vmx = to_vmx(vcpu);
+
+ if (!cpu_has_virtual_nmis()) {
+ /*
+ * Tracking the NMI-blocked state in software is built upon
+ * finding the next open IRQ window. This, in turn, depends on
+ * well-behaving guests: They have to keep IRQs disabled at
+ * least as long as the NMI handler runs. Otherwise we may
+ * cause NMI nesting, maybe breaking the guest. But as this is
+ * highly unlikely, we can live with the residual risk.
+ */
+ vmx->soft_vnmi_blocked = 1;
+ vmx->vnmi_blocked_time = 0;
+ }
+
+ KVM_VCPU_KSTAT_INC(vcpu, kvmvs_nmi_injections);
+
+ if (vmx->rmode.vm86_active) {
+ vmx->rmode.irq.pending = 1;
+ vmx->rmode.irq.vector = NMI_VECTOR;
+ vmx->rmode.irq.rip = kvm_rip_read(vcpu);
+ vmcs_write32(VM_ENTRY_INTR_INFO_FIELD,
+ NMI_VECTOR | INTR_TYPE_SOFT_INTR |
+ INTR_INFO_VALID_MASK);
+ vmcs_write32(VM_ENTRY_INSTRUCTION_LEN, 1);
+ kvm_rip_write(vcpu, vmx->rmode.irq.rip - 1);
+ return;
+ }
+ vmcs_write32(VM_ENTRY_INTR_INFO_FIELD,
+ INTR_TYPE_NMI_INTR | INTR_INFO_VALID_MASK | NMI_VECTOR);
+}
+
+static int
+vmx_nmi_allowed(struct kvm_vcpu *vcpu)
+{
+ if (!cpu_has_virtual_nmis() && to_vmx(vcpu)->soft_vnmi_blocked)
+ return (0);
+
+ return (!(vmcs_read32(GUEST_INTERRUPTIBILITY_INFO) &
+ (GUEST_INTR_STATE_MOV_SS | GUEST_INTR_STATE_NMI)));
+}
+
+static int
+vmx_get_nmi_mask(struct kvm_vcpu *vcpu)
+{
+ if (!cpu_has_virtual_nmis())
+ return (to_vmx(vcpu)->soft_vnmi_blocked);
+ else
+ return (!!(vmcs_read32(GUEST_INTERRUPTIBILITY_INFO) &
+ GUEST_INTR_STATE_NMI));
+}
+
+static void
+vmx_set_nmi_mask(struct kvm_vcpu *vcpu, int masked)
+{
+ struct vcpu_vmx *vmx = to_vmx(vcpu);
+
+ if (!cpu_has_virtual_nmis()) {
+ if (vmx->soft_vnmi_blocked != masked) {
+ vmx->soft_vnmi_blocked = masked;
+ vmx->vnmi_blocked_time = 0;
+ }
+
+ return;
+ } else {
+ if (masked) {
+ vmcs_set_bits(GUEST_INTERRUPTIBILITY_INFO,
+ GUEST_INTR_STATE_NMI);
+ } else {
+ vmcs_clear_bits(GUEST_INTERRUPTIBILITY_INFO,
+ GUEST_INTR_STATE_NMI);
+ }
+ }
+}
+
+static int
+vmx_interrupt_allowed(struct kvm_vcpu *vcpu)
+{
+ return ((vmcs_readl(GUEST_RFLAGS) & X86_EFLAGS_IF) &&
+ !(vmcs_read32(GUEST_INTERRUPTIBILITY_INFO) &
+ (GUEST_INTR_STATE_STI | GUEST_INTR_STATE_MOV_SS)));
+}
+
+/* XXX Should be static! */
+int
+vmx_set_tss_addr(struct kvm *kvmp, caddr_t addr)
+{
+ int ret;
+
+ struct kvm_userspace_memory_region tss_mem = {
+ .slot = TSS_PRIVATE_MEMSLOT,
+ .guest_phys_addr = (uint64_t)addr,
+ .memory_size = PAGESIZE * 3,
+ .flags = 0,
+ };
+
+ ret = kvm_set_memory_region(kvmp, &tss_mem, 0);
+
+ if (ret)
+ return (ret);
+
+ kvmp->arch.tss_addr = (uint64_t)addr;
+
+ return (DDI_SUCCESS);
+}
+
+static int
+handle_rmode_exception(struct kvm_vcpu *vcpu, int vec, uint32_t err_code)
+{
+ /*
+ * Instruction with address size override prefix opcode 0x67
+ * Cause the #SS fault with 0 error code in VM86 mode.
+ */
+ if (((vec == GP_VECTOR) || (vec == SS_VECTOR)) && err_code == 0) {
+ if (emulate_instruction(vcpu, 0, 0, 0) == EMULATE_DONE)
+ return (1);
+ }
+
+ /*
+ * Forward all other exceptions that are valid in real mode.
+ * FIXME: Breaks guest debugging in real mode, needs to be fixed with
+ * the required debugging infrastructure rework.
+ */
+ switch (vec) {
+ case DB_VECTOR:
+ if (vcpu->guest_debug &
+ (KVM_GUESTDBG_SINGLESTEP | KVM_GUESTDBG_USE_HW_BP)) {
+ return (0);
+ }
+
+ kvm_queue_exception(vcpu, vec);
+ return (1);
+
+ case BP_VECTOR:
+ /*
+ * Update instruction length as we may reinject the exception
+ * from user space while in guest debugging mode.
+ */
+ to_vmx(vcpu)->vcpu.arch.event_exit_inst_len =
+ vmcs_read32(VM_EXIT_INSTRUCTION_LEN);
+
+ if (vcpu->guest_debug & KVM_GUESTDBG_USE_SW_BP)
+ return (0);
+ /* fall through */
+
+ case DE_VECTOR:
+ case OF_VECTOR:
+ case BR_VECTOR:
+ case UD_VECTOR:
+ case DF_VECTOR:
+ case SS_VECTOR:
+ case GP_VECTOR:
+ case MF_VECTOR:
+ kvm_queue_exception(vcpu, vec);
+ return (1);
+ }
+
+ return (0);
+}
+
+/*
+ * Trigger machine check on the host. We assume all the MSRs are already set up
+ * by the CPU and that we still run on the same CPU as the MCE occurred on.
+ * We pass a fake environment to the machine check handler because we want
+ * the guest to be always treated like user space, no matter what context
+ * it used internally.
+ */
+static void kvm_machine_check(void)
+{
+#if defined(CONFIG_X86_MCE) && defined(CONFIG_X86_64)
+ struct pt_regs regs = {
+ .cs = 3, /* Fake ring 3 no matter what the guest ran on */
+ .flags = X86_EFLAGS_IF,
+ };
+
+ do_machine_check(&regs, 0);
+#endif
+}
+
+static int
+handle_machine_check(struct kvm_vcpu *vcpu)
+{
+ /* already handled by vcpu_run */
+ return (1);
+}
+
+/*
+ * The guest has exited. See if we can fix it or if we need userspace
+ * assistance.
+ */
+static int
+handle_exception(struct kvm_vcpu *vcpu)
+{
+ struct vcpu_vmx *vmx = to_vmx(vcpu);
+ struct kvm_run *kvm_run = vcpu->run;
+ uint32_t intr_info, ex_no, error_code;
+ unsigned long cr2, rip, dr6;
+ uint32_t vect_info;
+ enum emulation_result er;
+
+ vect_info = vmx->idt_vectoring_info;
+ intr_info = vmcs_read32(VM_EXIT_INTR_INFO);
+
+ if (is_machine_check(intr_info))
+ return (handle_machine_check(vcpu));
+
+ if ((vect_info & VECTORING_INFO_VALID_MASK) &&
+ !is_page_fault(intr_info)) {
+ vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
+ vcpu->run->internal.suberror = KVM_INTERNAL_ERROR_SIMUL_EX;
+ vcpu->run->internal.ndata = 2;
+ vcpu->run->internal.data[0] = vect_info;
+ vcpu->run->internal.data[1] = intr_info;
+ return (0);
+ }
+
+ if ((intr_info & INTR_INFO_INTR_TYPE_MASK) == INTR_TYPE_NMI_INTR)
+ return (1); /* already handled by vmx_vcpu_run() */
+
+ if (is_no_device(intr_info)) {
+ vmx_fpu_activate(vcpu);
+ return (1);
+ }
+
+ if (is_invalid_opcode(intr_info)) {
+ er = emulate_instruction(vcpu, 0, 0, EMULTYPE_TRAP_UD);
+ if (er != EMULATE_DONE)
+ kvm_queue_exception(vcpu, UD_VECTOR);
+ return (1);
+ }
+
+ error_code = 0;
+ rip = kvm_rip_read(vcpu);
+
+ if (intr_info & INTR_INFO_DELIVER_CODE_MASK)
+ error_code = vmcs_read32(VM_EXIT_INTR_ERROR_CODE);
+
+ if (is_page_fault(intr_info)) {
+ /* EPT won't cause page fault directly */
+ if (enable_ept)
+ cmn_err(CE_PANIC, "page fault with ept enabled\n");
+ cr2 = vmcs_readl(EXIT_QUALIFICATION);
+
+ KVM_TRACE2(page__fault, uintptr_t, cr2, uint32_t, error_code);
+
+ if (kvm_event_needs_reinjection(vcpu))
+ kvm_mmu_unprotect_page_virt(vcpu, cr2);
+ return (kvm_mmu_page_fault(vcpu, cr2, error_code));
+ }
+
+ if (vmx->rmode.vm86_active && handle_rmode_exception(vcpu,
+ intr_info & INTR_INFO_VECTOR_MASK, error_code)) {
+ if (vcpu->arch.halt_request) {
+ vcpu->arch.halt_request = 0;
+ return (kvm_emulate_halt(vcpu));
+ }
+ return (1);
+ }
+
+ ex_no = intr_info & INTR_INFO_VECTOR_MASK;
+ switch (ex_no) {
+ case DB_VECTOR:
+ dr6 = vmcs_readl(EXIT_QUALIFICATION);
+ if (!(vcpu->guest_debug &
+ (KVM_GUESTDBG_SINGLESTEP | KVM_GUESTDBG_USE_HW_BP))) {
+ vcpu->arch.dr6 = dr6 | DR6_FIXED_1;
+ kvm_queue_exception(vcpu, DB_VECTOR);
+ return (1);
+ }
+
+ kvm_run->debug.arch.dr6 = dr6 | DR6_FIXED_1;
+ kvm_run->debug.arch.dr7 = vmcs_readl(GUEST_DR7);
+ /* fall through */
+ case BP_VECTOR:
+ /*
+ * Update instruction length as we may reinject #BP from
+ * user space while in guest debugging mode. Reading it for
+ * #DB as well causes no harm, it is not used in that case.
+ */
+ vmx->vcpu.arch.event_exit_inst_len =
+ vmcs_read32(VM_EXIT_INSTRUCTION_LEN);
+ kvm_run->exit_reason = KVM_EXIT_DEBUG;
+ kvm_run->debug.arch.pc = vmcs_readl(GUEST_CS_BASE) + rip;
+ kvm_run->debug.arch.exception = ex_no;
+ break;
+ default:
+ kvm_run->exit_reason = KVM_EXIT_EXCEPTION;
+ kvm_run->ex.exception = ex_no;
+ kvm_run->ex.error_code = error_code;
+ break;
+ }
+
+ return (0);
+}
+
+static int
+handle_external_interrupt(struct kvm_vcpu *vcpu)
+{
+ KVM_VCPU_KSTAT_INC(vcpu, kvmvs_irq_exits);
+ return (1);
+}
+
+static int
+handle_triple_fault(struct kvm_vcpu *vcpu)
+{
+ vcpu->run->exit_reason = KVM_EXIT_SHUTDOWN;
+ return (0);
+}
+
+static int
+handle_io(struct kvm_vcpu *vcpu)
+{
+ unsigned long exit_qualification;
+ int size, in, string;
+ unsigned port;
+
+ KVM_VCPU_KSTAT_INC(vcpu, kvmvs_io_exits);
+
+ exit_qualification = vmcs_readl(EXIT_QUALIFICATION);
+ string = (exit_qualification & 16) != 0;
+
+ if (string) {
+ if (emulate_instruction(vcpu, 0, 0, 0) == EMULATE_DO_MMIO)
+ return (0);
+ return (1);
+ }
+
+ size = (exit_qualification & 7) + 1;
+ in = (exit_qualification & 8) != 0;
+ port = exit_qualification >> 16;
+ skip_emulated_instruction(vcpu);
+
+ return (kvm_emulate_pio(vcpu, in, size, port));
+}
+
+static void
+vmx_patch_hypercall(struct kvm_vcpu *vcpu, unsigned char *hypercall)
+{
+ /*
+ * Patch in the VMCALL instruction:
+ */
+ hypercall[0] = 0x0f;
+ hypercall[1] = 0x01;
+ hypercall[2] = 0xc1;
+}
+
+static int
+handle_cr(struct kvm_vcpu *vcpu)
+{
+ unsigned long exit_qualification, val;
+ int cr;
+ int reg;
+
+ exit_qualification = vmcs_readl(EXIT_QUALIFICATION);
+ cr = exit_qualification & 15;
+ reg = (exit_qualification >> 8) & 15;
+ DTRACE_PROBE3(kvm__cr, int, cr, int, reg, int,
+ (exit_qualification >> 4) & 3);
+ switch ((exit_qualification >> 4) & 3) {
+ case 0: /* mov to cr */
+ val = kvm_register_read(vcpu, reg);
+ KVM_TRACE2(cr__write, int, cr, unsigned long, val);
+
+ switch (cr) {
+ case 0:
+ kvm_set_cr0(vcpu, val);
+ skip_emulated_instruction(vcpu);
+ return (1);
+ case 3:
+ kvm_set_cr3(vcpu, val);
+ skip_emulated_instruction(vcpu);
+ return (1);
+ case 4:
+ kvm_set_cr4(vcpu, val);
+ skip_emulated_instruction(vcpu);
+ return (1);
+ case 8: {
+ uint8_t cr8_prev = kvm_get_cr8(vcpu);
+ uint8_t cr8 = kvm_register_read(vcpu, reg);
+ kvm_set_cr8(vcpu, cr8);
+ skip_emulated_instruction(vcpu);
+
+ if (irqchip_in_kernel(vcpu->kvm))
+ return (1);
+
+ if (cr8_prev <= cr8)
+ return (1);
+
+ vcpu->run->exit_reason = KVM_EXIT_SET_TPR;
+ return (0);
+ }
+ };
+
+ break;
+
+ case 2: /* clts */
+ vmx_set_cr0(vcpu, kvm_read_cr0_bits(vcpu, ~X86_CR0_TS));
+ KVM_TRACE2(cr__write, int, 0,
+ unsigned long, kvm_read_cr0(vcpu));
+
+ skip_emulated_instruction(vcpu);
+ vmx_fpu_activate(vcpu);
+ return (1);
+ case 1: /* mov from cr */
+ switch (cr) {
+ case 3:
+ kvm_register_write(vcpu, reg, vcpu->arch.cr3);
+ KVM_TRACE2(cr__read, int, cr,
+ unsigned long, vcpu->arch.cr3);
+ skip_emulated_instruction(vcpu);
+ return (1);
+ case 8:
+ val = kvm_get_cr8(vcpu);
+ kvm_register_write(vcpu, reg, val);
+ KVM_TRACE2(cr__read, int, cr, unsigned long, val);
+ skip_emulated_instruction(vcpu);
+ return (1);
+ }
+ break;
+ case 3: /* lmsw */
+ val = (exit_qualification >> LMSW_SOURCE_DATA_SHIFT) & 0x0f;
+ KVM_TRACE2(cr__write, int, 0, unsigned long,
+ (kvm_read_cr0(vcpu) & ~0xful) | val);
+ kvm_lmsw(vcpu, val);
+
+ skip_emulated_instruction(vcpu);
+ return (1);
+ default:
+ break;
+ }
+ vcpu->run->exit_reason = 0;
+ cmn_err(CE_WARN, "unhandled control register: op %d cr %d\n",
+ (int)(exit_qualification >> 4) & 3, cr);
+
+ return (0);
+}
+
+static int
+check_dr_alias(struct kvm_vcpu *vcpu)
+{
+ if (kvm_read_cr4_bits(vcpu, X86_CR4_DE)) {
+ kvm_queue_exception(vcpu, UD_VECTOR);
+ return (-1);
+ }
+
+ return (0);
+}
+
+static int
+handle_dr(struct kvm_vcpu *vcpu)
+{
+ unsigned long exit_qualification;
+ unsigned long val;
+ int dr, reg;
+
+ /* Do not handle if the CPL > 0, will trigger GP on re-entry */
+ if (!kvm_require_cpl(vcpu, 0))
+ return (1);
+
+ dr = vmcs_readl(GUEST_DR7);
+
+ if (dr & DR7_GD) {
+ /*
+ * As the vm-exit takes precedence over the debug trap, we
+ * need to emulate the latter, either for the host or the
+ * guest debugging itself.
+ */
+ if (vcpu->guest_debug & KVM_GUESTDBG_USE_HW_BP) {
+ vcpu->run->debug.arch.dr6 = vcpu->arch.dr6;
+ vcpu->run->debug.arch.dr7 = dr;
+ vcpu->run->debug.arch.pc =
+ vmcs_readl(GUEST_CS_BASE) +
+ vmcs_readl(GUEST_RIP);
+ vcpu->run->debug.arch.exception = DB_VECTOR;
+ vcpu->run->exit_reason = KVM_EXIT_DEBUG;
+ return (0);
+ } else {
+ vcpu->arch.dr7 &= ~DR7_GD;
+ vcpu->arch.dr6 |= DR6_BD;
+ vmcs_writel(GUEST_DR7, vcpu->arch.dr7);
+ kvm_queue_exception(vcpu, DB_VECTOR);
+ return (1);
+ }
+ }
+
+ exit_qualification = vmcs_readl(EXIT_QUALIFICATION);
+ dr = exit_qualification & DEBUG_REG_ACCESS_NUM;
+ reg = DEBUG_REG_ACCESS_REG(exit_qualification);
+ if (exit_qualification & TYPE_MOV_FROM_DR) {
+ switch (dr) {
+ case 0 ... 3:
+ val = vcpu->arch.db[dr];
+ break;
+ case 4:
+ if (check_dr_alias(vcpu) < 0)
+ return (1);
+ /* fall through */
+ case 6:
+ val = vcpu->arch.dr6;
+ break;
+ case 5:
+ if (check_dr_alias(vcpu) < 0)
+ return (1);
+ /* fall through */
+ default: /* 7 */
+ val = vcpu->arch.dr7;
+ break;
+ }
+ kvm_register_write(vcpu, reg, val);
+ } else {
+ val = vcpu->arch.regs[reg];
+ switch (dr) {
+ case 0 ... 3:
+ vcpu->arch.db[dr] = val;
+ if (!(vcpu->guest_debug & KVM_GUESTDBG_USE_HW_BP))
+ vcpu->arch.eff_db[dr] = val;
+ break;
+ case 4:
+ if (check_dr_alias(vcpu) < 0)
+ return (1);
+ /* fall through */
+ case 6:
+ if (val & 0xffffffff00000000ULL) {
+ kvm_inject_gp(vcpu, 0);
+ return (1);
+ }
+ vcpu->arch.dr6 = (val & DR6_VOLATILE) | DR6_FIXED_1;
+ break;
+ case 5:
+ if (check_dr_alias(vcpu) < 0)
+ return (1);
+ /* fall through */
+ default: /* 7 */
+ if (val & 0xffffffff00000000ULL) {
+ kvm_inject_gp(vcpu, 0);
+ return (1);
+ }
+ vcpu->arch.dr7 = (val & DR7_VOLATILE) | DR7_FIXED_1;
+
+ if (!(vcpu->guest_debug & KVM_GUESTDBG_USE_HW_BP)) {
+ vmcs_writel(GUEST_DR7, vcpu->arch.dr7);
+ vcpu->arch.switch_db_regs =
+ (val & DR7_BP_EN_MASK);
+
+ }
+ break;
+ }
+ }
+ skip_emulated_instruction(vcpu);
+ return (1);
+}
+
+static int
+handle_cpuid(struct kvm_vcpu *vcpu)
+{
+ kvm_emulate_cpuid(vcpu);
+ return (1);
+}
+
+static int
+handle_rdmsr(struct kvm_vcpu *vcpu)
+{
+ uint32_t ecx = vcpu->arch.regs[VCPU_REGS_RCX];
+ uint64_t data;
+
+ if (vmx_get_msr(vcpu, ecx, &data)) {
+ KVM_TRACE1(msr__read__ex, uint32_t, ecx);
+ kvm_inject_gp(vcpu, 0);
+ return (1);
+ }
+
+ KVM_TRACE2(msr__read, uint32_t, ecx, uint64_t, data);
+
+ /* FIXME: handling of bits 32:63 of rax, rdx */
+ vcpu->arch.regs[VCPU_REGS_RAX] = data & -1u;
+ vcpu->arch.regs[VCPU_REGS_RDX] = (data >> 32) & -1u;
+ skip_emulated_instruction(vcpu);
+ return (1);
+}
+
+static int
+handle_wrmsr(struct kvm_vcpu *vcpu)
+{
+ uint32_t ecx = vcpu->arch.regs[VCPU_REGS_RCX];
+ uint64_t data = (vcpu->arch.regs[VCPU_REGS_RAX] & -1u) |
+ ((uint64_t)(vcpu->arch.regs[VCPU_REGS_RDX] & -1u) << 32);
+
+ if (vmx_set_msr(vcpu, ecx, data) != 0) {
+ KVM_TRACE2(msr__write__ex, uint32_t, ecx, uint64_t, data);
+ kvm_inject_gp(vcpu, 0);
+ return (1);
+ }
+
+ KVM_TRACE2(msr__write, uint32_t, ecx, uint64_t, data);
+ skip_emulated_instruction(vcpu);
+ return (1);
+}
+
+static int
+handle_tpr_below_threshold(struct kvm_vcpu *vcpu)
+{
+ return (1);
+}
+
+static int
+handle_interrupt_window(struct kvm_vcpu *vcpu)
+{
+ uint32_t cpu_based_vm_exec_control;
+
+ /* clear pending irq */
+ cpu_based_vm_exec_control = vmcs_read32(CPU_BASED_VM_EXEC_CONTROL);
+ cpu_based_vm_exec_control &= ~CPU_BASED_VIRTUAL_INTR_PENDING;
+ vmcs_write32(CPU_BASED_VM_EXEC_CONTROL, cpu_based_vm_exec_control);
+
+ KVM_VCPU_KSTAT_INC(vcpu, kvmvs_irq_window_exits);
+
+ /*
+ * If the user space waits to inject interrupts, exit as soon as
+ * possible
+ */
+ if (!irqchip_in_kernel(vcpu->kvm) &&
+ vcpu->run->request_interrupt_window &&
+ !kvm_cpu_has_interrupt(vcpu)) {
+ vcpu->run->exit_reason = KVM_EXIT_IRQ_WINDOW_OPEN;
+ return (0);
+ }
+ return (1);
+}
+
+static int
+handle_halt(struct kvm_vcpu *vcpu)
+{
+ skip_emulated_instruction(vcpu);
+ return (kvm_emulate_halt(vcpu));
+}
+
+static int
+handle_vmcall(struct kvm_vcpu *vcpu)
+{
+ skip_emulated_instruction(vcpu);
+ kvm_emulate_hypercall(vcpu);
+ return (1);
+}
+
+static int
+handle_vmx_insn(struct kvm_vcpu *vcpu)
+{
+ kvm_queue_exception(vcpu, UD_VECTOR);
+ return (1);
+}
+
+static int
+handle_invlpg(struct kvm_vcpu *vcpu)
+{
+ unsigned long exit_qualification = vmcs_readl(EXIT_QUALIFICATION);
+
+ kvm_mmu_invlpg(vcpu, exit_qualification);
+ skip_emulated_instruction(vcpu);
+ return (1);
+}
+
+static int
+handle_wbinvd(struct kvm_vcpu *vcpu)
+{
+ skip_emulated_instruction(vcpu);
+ /* TODO: Add support for VT-d/pass-through device */
+ return (1);
+}
+
+static int
+handle_apic_access(struct kvm_vcpu *vcpu)
+{
+ unsigned long exit_qualification;
+ enum emulation_result er;
+ unsigned long offset;
+
+ exit_qualification = vmcs_readl(EXIT_QUALIFICATION);
+ offset = exit_qualification & 0xffful;
+
+ er = emulate_instruction(vcpu, 0, 0, 0);
+
+ if (er != EMULATE_DONE) {
+ cmn_err(CE_PANIC, "Fail to handle apic access vmexit! "
+ "Offset is 0x%lx\n", offset);
+ }
+
+ return (1);
+}
+
+static int
+handle_task_switch(struct kvm_vcpu *vcpu)
+{
+ struct vcpu_vmx *vmx = to_vmx(vcpu);
+ unsigned long exit_qualification;
+ uint16_t tss_selector;
+ int reason, type, idt_v;
+
+ idt_v = (vmx->idt_vectoring_info & VECTORING_INFO_VALID_MASK);
+ type = (vmx->idt_vectoring_info & VECTORING_INFO_TYPE_MASK);
+ exit_qualification = vmcs_readl(EXIT_QUALIFICATION);
+
+ reason = (uint32_t)exit_qualification >> 30;
+ if (reason == TASK_SWITCH_GATE && idt_v) {
+ switch (type) {
+ case INTR_TYPE_NMI_INTR:
+ vcpu->arch.nmi_injected = 0;
+ if (cpu_has_virtual_nmis()) {
+ vmcs_set_bits(GUEST_INTERRUPTIBILITY_INFO,
+ GUEST_INTR_STATE_NMI);
+ }
+ break;
+ case INTR_TYPE_EXT_INTR:
+ case INTR_TYPE_SOFT_INTR:
+ kvm_clear_interrupt_queue(vcpu);
+ break;
+ case INTR_TYPE_HARD_EXCEPTION:
+ case INTR_TYPE_SOFT_EXCEPTION:
+ kvm_clear_exception_queue(vcpu);
+ break;
+ default:
+ break;
+ }
+ }
+ tss_selector = exit_qualification;
+
+ if (!idt_v || (type != INTR_TYPE_HARD_EXCEPTION &&
+ type != INTR_TYPE_EXT_INTR && type != INTR_TYPE_NMI_INTR))
+ skip_emulated_instruction(vcpu);
+
+ if (!kvm_task_switch(vcpu, tss_selector, reason))
+ return (0);
+
+ /* clear all local breakpoint enable flags */
+ vmcs_writel(GUEST_DR7, vmcs_readl(GUEST_DR7) & ~55);
+
+ /*
+ * TODO: What about debug traps on tss switch?
+ * Are we supposed to inject them and update dr6?
+ */
+
+ return (1);
+}
+
+static int
+handle_ept_violation(struct kvm_vcpu *vcpu)
+{
+ unsigned long exit_qualification;
+ gpa_t gpa;
+ int gla_validity;
+
+ exit_qualification = vmcs_readl(EXIT_QUALIFICATION);
+
+ if (exit_qualification & (1 << 6)) {
+ cmn_err(CE_PANIC, "EPT: GPA exceeds GAW!\n");
+ }
+
+ gla_validity = (exit_qualification >> 7) & 0x3;
+ if (gla_validity != 0x3 && gla_validity != 0x1 && gla_validity != 0) {
+ cmn_err(CE_WARN, "EPT: Handling EPT violation failed!\n");
+ cmn_err(CE_CONT, "EPT: GPA: 0x%lx, GVA: 0x%lx\n",
+ (long unsigned int)vmcs_read64(GUEST_PHYSICAL_ADDRESS),
+ vmcs_readl(GUEST_LINEAR_ADDRESS));
+ cmn_err(CE_PANIC, "EPT: Exit qualification is 0x%lx\n",
+ (long unsigned int)exit_qualification);
+ vcpu->run->exit_reason = KVM_EXIT_UNKNOWN;
+ vcpu->run->hw.hardware_exit_reason = EXIT_REASON_EPT_VIOLATION;
+ return (0);
+ }
+
+ gpa = vmcs_read64(GUEST_PHYSICAL_ADDRESS);
+ KVM_TRACE2(page__fault, gpa_t, gpa, unsigned long, exit_qualification);
+
+ return (kvm_mmu_page_fault(vcpu, gpa & PAGEMASK, 0));
+}
+
+/* XXX - The following assumes we're running on the maximum sized box... */
+#define MAX_PHYSMEM_BITS 46
+static uint64_t
+ept_rsvd_mask(uint64_t spte, int level)
+{
+ int i;
+ uint64_t mask = 0;
+
+#ifdef XXX
+ for (i = 51; i > boot_cpu_data.x86_phys_bits; i--)
+ mask |= (1ULL << i);
+#else
+ XXX_KVM_PROBE;
+ for (i = 51; i > MAX_PHYSMEM_BITS; i--)
+ mask |= (1ULL << i);
+#endif
+
+ if (level > 2)
+ /* bits 7:3 reserved */
+ mask |= 0xf8;
+ else if (level == 2) {
+ if (spte & (1ULL << 7))
+ /* 2MB ref, bits 20:12 reserved */
+ mask |= 0x1ff000;
+ else
+ /* bits 6:3 reserved */
+ mask |= 0x78;
+ }
+
+ return (mask);
+}
+
+
+static void
+ept_misconfig_inspect_spte(struct kvm_vcpu *vcpu, uint64_t spte, int level)
+{
+ cmn_err(CE_WARN, "%s: spte 0x%lx level %d\n", __func__, spte, level);
+
+ /* 010b (write-only) */
+ if ((spte & 0x7) == 0x2)
+ cmn_err(CE_CONT, "%s: spte is write-only\n", __func__);
+
+ /* 110b (write/execute) */
+ if ((spte & 0x7) == 0x6)
+ cmn_err(CE_CONT, "%s: spte is write-execute\n", __func__);
+
+ /* 100b (execute-only) and value not supported by logical processor */
+ if (!cpu_has_vmx_ept_execute_only()) {
+ if ((spte & 0x7) == 0x4)
+ cmn_err(CE_CONT,
+ "%s: spte is execute-only\n", __func__);
+ }
+
+ /* not 000b */
+ if ((spte & 0x7)) {
+ uint64_t rsvd_bits = spte & ept_rsvd_mask(spte, level);
+
+ if (rsvd_bits != 0) {
+ cmn_err(CE_CONT, "%s: rsvd_bits = 0x%lx\n",
+ __func__, rsvd_bits);
+ }
+
+ if (level == 1 || (level == 2 && (spte & (1ULL << 7)))) {
+ uint64_t ept_mem_type = (spte & 0x38) >> 3;
+
+ if (ept_mem_type == 2 || ept_mem_type == 3 ||
+ ept_mem_type == 7) {
+ cmn_err(CE_CONT, "%s: ept_mem_type=0x%lx\n",
+ __func__, ept_mem_type);
+ }
+ }
+ }
+}
+
+static int
+handle_ept_misconfig(struct kvm_vcpu *vcpu)
+{
+ uint64_t sptes[4];
+ int nr_sptes, i;
+ gpa_t gpa;
+
+ gpa = vmcs_read64(GUEST_PHYSICAL_ADDRESS);
+
+ cmn_err(CE_WARN, "EPT: Misconfiguration.\n");
+ cmn_err(CE_CONT, "EPT: GPA: 0x%lx\n", gpa);
+ nr_sptes = kvm_mmu_get_spte_hierarchy(vcpu, gpa, sptes);
+
+ for (i = PT64_ROOT_LEVEL; i > PT64_ROOT_LEVEL - nr_sptes; --i)
+ ept_misconfig_inspect_spte(vcpu, sptes[i-1], i);
+
+ vcpu->run->exit_reason = KVM_EXIT_UNKNOWN;
+ vcpu->run->hw.hardware_exit_reason = EXIT_REASON_EPT_MISCONFIG;
+
+ return (0);
+}
+
+static int
+handle_nmi_window(struct kvm_vcpu *vcpu)
+{
+ uint32_t cpu_based_vm_exec_control;
+
+ /* clear pending NMI */
+ cpu_based_vm_exec_control = vmcs_read32(CPU_BASED_VM_EXEC_CONTROL);
+ cpu_based_vm_exec_control &= ~CPU_BASED_VIRTUAL_NMI_PENDING;
+ vmcs_write32(CPU_BASED_VM_EXEC_CONTROL, cpu_based_vm_exec_control);
+
+ KVM_VCPU_KSTAT_INC(vcpu, kvmvs_nmi_window_exits);
+
+ return (1);
+}
+
+
+static int handle_invalid_guest_state(struct kvm_vcpu *vcpu)
+{
+ struct vcpu_vmx *vmx = to_vmx(vcpu);
+ enum emulation_result err = EMULATE_DONE;
+ int ret = 1;
+
+ while (!guest_state_valid(vcpu)) {
+ err = emulate_instruction(vcpu, 0, 0, 0);
+
+ if (err == EMULATE_DO_MMIO) {
+ ret = 0;
+ goto out;
+ }
+
+ if (err != EMULATE_DONE) {
+ vcpu->run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
+ vcpu->run->internal.suberror =
+ KVM_INTERNAL_ERROR_EMULATION;
+ vcpu->run->internal.ndata = 0;
+ ret = 0;
+ goto out;
+ }
+
+#ifdef XXX
+ if ((current))
+ goto out;
+#else
+ XXX_KVM_PROBE;
+#endif
+ }
+
+ vmx->emulation_required = 0;
+out:
+ return (ret);
+}
+
+/*
+ * Indicate a busy-waiting vcpu in spinlock. We do not enable the PAUSE
+ * exiting, so only get here on cpu with PAUSE-Loop-Exiting.
+ */
+static int
+handle_pause(struct kvm_vcpu *vcpu)
+{
+ skip_emulated_instruction(vcpu);
+#ifdef XXX
+ kvm_vcpu_on_spin(vcpu);
+#else
+ XXX_KVM_PROBE;
+#endif
+ return (1);
+}
+
+static int
+handle_invalid_op(struct kvm_vcpu *vcpu)
+{
+ kvm_queue_exception(vcpu, UD_VECTOR);
+ return (1);
+}
+
+/*
+ * The exit handlers return 1 if the exit was handled fully and guest execution
+ * may resume. Otherwise they set the kvm_run parameter to indicate what needs
+ * to be done to userspace and return 0.
+ */
+static int (*kvm_vmx_exit_handlers[])(struct kvm_vcpu *vcpu) = {
+ [EXIT_REASON_EXCEPTION_NMI] = handle_exception,
+ [EXIT_REASON_EXTERNAL_INTERRUPT] = handle_external_interrupt,
+ [EXIT_REASON_TRIPLE_FAULT] = handle_triple_fault,
+ [EXIT_REASON_NMI_WINDOW] = handle_nmi_window,
+ [EXIT_REASON_IO_INSTRUCTION] = handle_io,
+ [EXIT_REASON_CR_ACCESS] = handle_cr,
+ [EXIT_REASON_DR_ACCESS] = handle_dr,
+ [EXIT_REASON_CPUID] = handle_cpuid,
+ [EXIT_REASON_MSR_READ] = handle_rdmsr,
+ [EXIT_REASON_MSR_WRITE] = handle_wrmsr,
+ [EXIT_REASON_PENDING_INTERRUPT] = handle_interrupt_window,
+ [EXIT_REASON_HLT] = handle_halt,
+ [EXIT_REASON_INVLPG] = handle_invlpg,
+ [EXIT_REASON_VMCALL] = handle_vmcall,
+ [EXIT_REASON_VMCLEAR] = handle_vmx_insn,
+ [EXIT_REASON_VMLAUNCH] = handle_vmx_insn,
+ [EXIT_REASON_VMPTRLD] = handle_vmx_insn,
+ [EXIT_REASON_VMPTRST] = handle_vmx_insn,
+ [EXIT_REASON_VMREAD] = handle_vmx_insn,
+ [EXIT_REASON_VMRESUME] = handle_vmx_insn,
+ [EXIT_REASON_VMWRITE] = handle_vmx_insn,
+ [EXIT_REASON_VMOFF] = handle_vmx_insn,
+ [EXIT_REASON_VMON] = handle_vmx_insn,
+ [EXIT_REASON_TPR_BELOW_THRESHOLD] = handle_tpr_below_threshold,
+ [EXIT_REASON_APIC_ACCESS] = handle_apic_access,
+ [EXIT_REASON_WBINVD] = handle_wbinvd,
+ [EXIT_REASON_TASK_SWITCH] = handle_task_switch,
+ [EXIT_REASON_MCE_DURING_VMENTRY] = handle_machine_check,
+ [EXIT_REASON_EPT_VIOLATION] = handle_ept_violation,
+ [EXIT_REASON_EPT_MISCONFIG] = handle_ept_misconfig,
+ [EXIT_REASON_PAUSE_INSTRUCTION] = handle_pause,
+ [EXIT_REASON_MWAIT_INSTRUCTION] = handle_invalid_op,
+ [EXIT_REASON_MONITOR_INSTRUCTION] = handle_invalid_op,
+};
+
+static const int kvm_vmx_max_exit_handlers =
+ ARRAY_SIZE(kvm_vmx_exit_handlers);
+
+/*
+ * The guest has exited. See if we can fix it or if we need userspace
+ * assistance.
+ */
+static int
+vmx_handle_exit(struct kvm_vcpu *vcpu)
+{
+ struct vcpu_vmx *vmx = to_vmx(vcpu);
+ uint32_t exit_reason = vmx->exit_reason;
+ uint32_t vectoring_info = vmx->idt_vectoring_info;
+ int rval;
+ unsigned long rip;
+
+ /* Always read the guest rip when exiting */
+ rip = vmcs_readl(GUEST_RIP);
+ DTRACE_PROBE2(kvm__vexit, unsigned long, rip, uint32_t, exit_reason);
+
+ /* If guest state is invalid, start emulating */
+ if (vmx->emulation_required && emulate_invalid_guest_state)
+ return (handle_invalid_guest_state(vcpu));
+
+ /*
+ * Access CR3 don't cause VMExit in paging mode, so we need
+ * to sync with guest real CR3.
+ */
+ if (enable_ept && is_paging(vcpu))
+ vcpu->arch.cr3 = vmcs_readl(GUEST_CR3);
+
+ if (vmx->fail) {
+ vcpu->run->exit_reason = KVM_EXIT_FAIL_ENTRY;
+ vcpu->run->fail_entry.hardware_entry_failure_reason
+ = vmcs_read32(VM_INSTRUCTION_ERROR);
+
+ return (0);
+ }
+
+ if ((vectoring_info & VECTORING_INFO_VALID_MASK) &&
+ (exit_reason != EXIT_REASON_EXCEPTION_NMI &&
+ exit_reason != EXIT_REASON_EPT_VIOLATION &&
+ exit_reason != EXIT_REASON_TASK_SWITCH)) {
+ cmn_err(CE_WARN, "%s: unexpected, valid vectoring info "
+ "(0x%x) and exit reason is 0x%x\n",
+ __func__, vectoring_info, exit_reason);
+ }
+
+ if (!cpu_has_virtual_nmis() && vmx->soft_vnmi_blocked) {
+ if (vmx_interrupt_allowed(vcpu)) {
+ vmx->soft_vnmi_blocked = 0;
+ } else if (vmx->vnmi_blocked_time > 1000000000LL &&
+ vcpu->arch.nmi_pending) {
+ /*
+ * This CPU don't support us in finding the end of an
+ * NMI-blocked window if the guest runs with IRQs
+ * disabled. So we pull the trigger after 1 s of
+ * futile waiting, but inform the user about this.
+ */
+ cmn_err(CE_WARN, "%s: Breaking out of NMI-blocked "
+ "state on VCPU %d after 1 s timeout\n",
+ __func__, vcpu->vcpu_id);
+ vmx->soft_vnmi_blocked = 0;
+ }
+ }
+
+ if (exit_reason < kvm_vmx_max_exit_handlers &&
+ kvm_vmx_exit_handlers[exit_reason]) {
+ rval = kvm_vmx_exit_handlers[exit_reason](vcpu);
+ return (rval);
+ } else {
+ vcpu->run->exit_reason = KVM_EXIT_UNKNOWN;
+ vcpu->run->hw.hardware_exit_reason = exit_reason;
+ }
+
+ return (0);
+}
+
+static void
+vmx_update_cr8_intercept(struct kvm_vcpu *vcpu, int tpr, int irr)
+{
+ if (irr == -1 || tpr < irr) {
+ vmcs_write32(TPR_THRESHOLD, 0);
+ return;
+ }
+
+ vmcs_write32(TPR_THRESHOLD, irr);
+}
+
+static void
+vmx_complete_interrupts(struct vcpu_vmx *vmx)
+{
+ uint32_t exit_intr_info;
+ uint32_t idt_vectoring_info = vmx->idt_vectoring_info;
+ int unblock_nmi;
+ uint8_t vector;
+ int type;
+ int idtv_info_valid;
+
+ exit_intr_info = vmcs_read32(VM_EXIT_INTR_INFO);
+
+ vmx->exit_reason = vmcs_read32(VM_EXIT_REASON);
+
+ /* Handle machine checks before interrupts are enabled */
+ if ((vmx->exit_reason == EXIT_REASON_MCE_DURING_VMENTRY) ||
+ (vmx->exit_reason == EXIT_REASON_EXCEPTION_NMI &&
+ is_machine_check(exit_intr_info)))
+ kvm_machine_check();
+
+ /* We need to handle NMIs before interrupts are enabled */
+ if ((exit_intr_info & INTR_INFO_INTR_TYPE_MASK) == INTR_TYPE_NMI_INTR &&
+ (exit_intr_info & INTR_INFO_VALID_MASK))
+ __asm__("int $2");
+
+ idtv_info_valid = idt_vectoring_info & VECTORING_INFO_VALID_MASK;
+
+ if (cpu_has_virtual_nmis()) {
+ unblock_nmi = (exit_intr_info & INTR_INFO_UNBLOCK_NMI) != 0;
+ vector = exit_intr_info & INTR_INFO_VECTOR_MASK;
+ /*
+ * SDM 3: 27.7.1.2 (September 2008)
+ * Re-set bit "block by NMI" before VM entry if vmexit caused by
+ * a guest IRET fault.
+ * SDM 3: 23.2.2 (September 2008)
+ * Bit 12 is undefined in any of the following cases:
+ * If the VM exit sets the valid bit in the IDT-vectoring
+ * information field.
+ * If the VM exit is due to a double fault.
+ */
+ if ((exit_intr_info & INTR_INFO_VALID_MASK) && unblock_nmi &&
+ vector != DF_VECTOR && !idtv_info_valid)
+ vmcs_set_bits(GUEST_INTERRUPTIBILITY_INFO,
+ GUEST_INTR_STATE_NMI);
+ } else if (vmx->soft_vnmi_blocked) {
+#ifdef XXX
+ vmx->vnmi_blocked_time +=
+ ktime_to_ns(ktime_sub(ktime_get(), vmx->entry_time));
+#else
+ vmx->vnmi_blocked_time +=
+ gethrtime() - vmx->entry_time;
+ XXX_KVM_PROBE;
+#endif
+ }
+
+ vmx->vcpu.arch.nmi_injected = 0;
+ kvm_clear_exception_queue(&vmx->vcpu);
+ kvm_clear_interrupt_queue(&vmx->vcpu);
+
+ if (!idtv_info_valid)
+ return;
+
+ vector = idt_vectoring_info & VECTORING_INFO_VECTOR_MASK;
+ type = idt_vectoring_info & VECTORING_INFO_TYPE_MASK;
+
+ switch (type) {
+ case INTR_TYPE_NMI_INTR:
+ vmx->vcpu.arch.nmi_injected = 1;
+ /*
+ * SDM 3: 27.7.1.2 (September 2008)
+ * Clear bit "block by NMI" before VM entry if a NMI
+ * delivery faulted.
+ */
+ vmcs_clear_bits(GUEST_INTERRUPTIBILITY_INFO,
+ GUEST_INTR_STATE_NMI);
+ break;
+ case INTR_TYPE_SOFT_EXCEPTION:
+ vmx->vcpu.arch.event_exit_inst_len =
+ vmcs_read32(VM_EXIT_INSTRUCTION_LEN);
+ /* fall through */
+ case INTR_TYPE_HARD_EXCEPTION:
+ if (idt_vectoring_info & VECTORING_INFO_DELIVER_CODE_MASK) {
+ uint32_t err = vmcs_read32(IDT_VECTORING_ERROR_CODE);
+ kvm_queue_exception_e(&vmx->vcpu, vector, err);
+ } else
+ kvm_queue_exception(&vmx->vcpu, vector);
+ break;
+ case INTR_TYPE_SOFT_INTR:
+ vmx->vcpu.arch.event_exit_inst_len =
+ vmcs_read32(VM_EXIT_INSTRUCTION_LEN);
+ /* fall through */
+ case INTR_TYPE_EXT_INTR:
+ kvm_queue_interrupt(&vmx->vcpu, vector,
+ type == INTR_TYPE_SOFT_INTR);
+ break;
+ default:
+ break;
+ }
+}
+
+/*
+ * Failure to inject an interrupt should give us the information
+ * in IDT_VECTORING_INFO_FIELD. However, if the failure occurs
+ * when fetching the interrupt redirection bitmap in the real-mode
+ * tss, this doesn't happen. So we do it ourselves.
+ */
+static void
+fixup_rmode_irq(struct vcpu_vmx *vmx)
+{
+ vmx->rmode.irq.pending = 0;
+ if (kvm_rip_read(&vmx->vcpu) + 1 != vmx->rmode.irq.rip)
+ return;
+
+ kvm_rip_write(&vmx->vcpu, vmx->rmode.irq.rip);
+ if (vmx->idt_vectoring_info & VECTORING_INFO_VALID_MASK) {
+ vmx->idt_vectoring_info &= ~VECTORING_INFO_TYPE_MASK;
+ vmx->idt_vectoring_info |= INTR_TYPE_EXT_INTR;
+ return;
+ }
+
+ vmx->idt_vectoring_info = VECTORING_INFO_VALID_MASK |
+ INTR_TYPE_EXT_INTR | vmx->rmode.irq.vector;
+}
+
+#ifdef CONFIG_X86_64
+#define R "r"
+#define Q "q"
+#else
+#define R "e"
+#define Q "l"
+#endif
+
+static void
+vmx_vcpu_run(struct kvm_vcpu *vcpu)
+{
+ struct vcpu_vmx *vmx = to_vmx(vcpu);
+
+ /* Record the guest's net vcpu time for enforced NMI injections. */
+ if (!cpu_has_virtual_nmis() && vmx->soft_vnmi_blocked) {
+#ifdef XXX
+ vmx->entry_time = ktime_get();
+#else
+ vmx->entry_time = gethrtime();
+ XXX_KVM_PROBE;
+#endif
+ }
+
+ /*
+ * Don't enter VMX if guest state is invalid, let the exit handler
+ * start emulation until we arrive back to a valid state
+ */
+ if (vmx->emulation_required && emulate_invalid_guest_state)
+ return;
+
+ if (test_bit(VCPU_REGS_RSP, (unsigned long *)&vcpu->arch.regs_dirty))
+ vmcs_writel(GUEST_RSP, vcpu->arch.regs[VCPU_REGS_RSP]);
+ if (test_bit(VCPU_REGS_RIP, (unsigned long *)&vcpu->arch.regs_dirty))
+ vmcs_writel(GUEST_RIP, vcpu->arch.regs[VCPU_REGS_RIP]);
+
+ DTRACE_PROBE1(kvm__vrun, unsigned long, vcpu->arch.regs[VCPU_REGS_RIP]);
+
+ /*
+ * When single-stepping over STI and MOV SS, we must clear the
+ * corresponding interruptibility bits in the guest state. Otherwise
+ * vmentry fails as it then expects bit 14 (BS) in pending debug
+ * exceptions being set, but that's not correct for the guest debugging
+ * case.
+ */
+ if (vcpu->guest_debug & KVM_GUESTDBG_SINGLESTEP)
+ vmx_set_interrupt_shadow(vcpu, 0);
+
+ /*
+ * Loading guest fpu may have cleared host cr0.ts
+ */
+ vmcs_writel(HOST_CR0, read_cr0());
+
+ __asm__(
+ /* Store host registers */
+ "push %%"R"dx; push %%"R"bp;"
+ "push %%"R"cx \n\t"
+ "cmp %%"R"sp, %c[host_rsp](%0) \n\t"
+ "je 1f \n\t"
+ "mov %%"R"sp, %c[host_rsp](%0) \n\t"
+ __ex(ASM_VMX_VMWRITE_RSP_RDX) "\n\t"
+ "1: \n\t"
+ /* Reload cr2 if changed */
+ "mov %c[cr2](%0), %%"R"ax \n\t"
+ "mov %%cr2, %%"R"dx \n\t"
+ "cmp %%"R"ax, %%"R"dx \n\t"
+ "je 2f \n\t"
+ "mov %%"R"ax, %%cr2 \n\t"
+ "2: \n\t"
+ /* Check if vmlaunch of vmresume is needed */
+ "cmpl $0, %c[launched](%0) \n\t"
+ /* Load guest registers. Don't clobber flags. */
+ "mov %c[rax](%0), %%"R"ax \n\t"
+ "mov %c[rbx](%0), %%"R"bx \n\t"
+ "mov %c[rdx](%0), %%"R"dx \n\t"
+ "mov %c[rsi](%0), %%"R"si \n\t"
+ "mov %c[rdi](%0), %%"R"di \n\t"
+ "mov %c[rbp](%0), %%"R"bp \n\t"
+#ifdef CONFIG_X86_64
+ "mov %c[r8](%0), %%r8 \n\t"
+ "mov %c[r9](%0), %%r9 \n\t"
+ "mov %c[r10](%0), %%r10 \n\t"
+ "mov %c[r11](%0), %%r11 \n\t"
+ "mov %c[r12](%0), %%r12 \n\t"
+ "mov %c[r13](%0), %%r13 \n\t"
+ "mov %c[r14](%0), %%r14 \n\t"
+ "mov %c[r15](%0), %%r15 \n\t"
+#endif
+ "mov %c[rcx](%0), %%"R"cx \n\t" /* kills %0 (ecx) */
+
+ /* Enter guest mode */
+ "jne .Llaunched \n\t"
+ __ex(ASM_VMX_VMLAUNCH) "\n\t"
+ "jmp .Lkvm_vmx_return \n\t"
+ ".Llaunched: " __ex(ASM_VMX_VMRESUME) "\n\t"
+ ".Lkvm_vmx_return: "
+ /* Save guest registers, load host registers, keep flags */
+ "xchg %0, (%%"R"sp) \n\t"
+ "mov %%"R"ax, %c[rax](%0) \n\t"
+ "mov %%"R"bx, %c[rbx](%0) \n\t"
+ "push"Q" (%%"R"sp); pop"Q" %c[rcx](%0) \n\t"
+ "mov %%"R"dx, %c[rdx](%0) \n\t"
+ "mov %%"R"si, %c[rsi](%0) \n\t"
+ "mov %%"R"di, %c[rdi](%0) \n\t"
+ "mov %%"R"bp, %c[rbp](%0) \n\t"
+#ifdef CONFIG_X86_64
+ "mov %%r8, %c[r8](%0) \n\t"
+ "mov %%r9, %c[r9](%0) \n\t"
+ "mov %%r10, %c[r10](%0) \n\t"
+ "mov %%r11, %c[r11](%0) \n\t"
+ "mov %%r12, %c[r12](%0) \n\t"
+ "mov %%r13, %c[r13](%0) \n\t"
+ "mov %%r14, %c[r14](%0) \n\t"
+ "mov %%r15, %c[r15](%0) \n\t"
+#endif
+ "mov %%cr2, %%"R"ax \n\t"
+ "mov %%"R"ax, %c[cr2](%0) \n\t"
+
+ "pop %%"R"bp; pop %%"R"bp; pop %%"R"dx \n\t"
+ "setbe %c[fail](%0) \n\t"
+ : : "c"(vmx), "d"((unsigned long)HOST_RSP),
+ [launched]"i"(offsetof(struct vcpu_vmx, launched)),
+ [fail]"i"(offsetof(struct vcpu_vmx, fail)),
+ [host_rsp]"i"(offsetof(struct vcpu_vmx, host_rsp)),
+ [rax]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_RAX])),
+ [rbx]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_RBX])),
+ [rcx]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_RCX])),
+ [rdx]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_RDX])),
+ [rsi]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_RSI])),
+ [rdi]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_RDI])),
+ [rbp]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_RBP])),
+#ifdef CONFIG_X86_64
+ [r8]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R8])),
+ [r9]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R9])),
+ [r10]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R10])),
+ [r11]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R11])),
+ [r12]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R12])),
+ [r13]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R13])),
+ [r14]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R14])),
+ [r15]"i"(offsetof(struct vcpu_vmx, vcpu.arch.regs[VCPU_REGS_R15])),
+#endif
+ [cr2]"i"(offsetof(struct vcpu_vmx, vcpu.arch.cr2))
+ : "cc", "memory"
+ /*CSTYLED*/
+ , R"bx", R"di", R"si"
+#ifdef CONFIG_X86_64
+ /*CSTYLED*/
+ , "r8", "r9", "r10", "r11", "r12", "r13", "r14", "r15"
+#endif
+ /*CSTYLED*/
+ );
+
+ vcpu->arch.regs_avail = ~((1 << VCPU_REGS_RIP) |
+ (1 << VCPU_REGS_RSP) | (1 << VCPU_EXREG_PDPTR));
+ vcpu->arch.regs_dirty = 0;
+
+ vmx->idt_vectoring_info = vmcs_read32(IDT_VECTORING_INFO_FIELD);
+
+ if (vmx->rmode.irq.pending)
+ fixup_rmode_irq(vmx);
+
+#ifdef XXX
+ __asm__("mov %0, %%ds; mov %0, %%es" :
+ : "r"SEL_GDT(GDT_UDATA, SEL_UPL));
+#else
+ XXX_KVM_PROBE;
+ __asm__("mov %0, %%ds; mov %0, %%es" : : "r"KDS_SEL);
+#endif
+ vmx->launched = 1;
+
+ vmx_complete_interrupts(vmx);
+}
+
+#undef R
+#undef Q
+
+static void
+vmx_destroy_vcpu(struct kvm_vcpu *vcpu)
+{
+ /* XXX don't assume it's the first element */
+ vcpu_vmx_t *vmx = (vcpu_vmx_t *)vcpu;
+
+ if (vmx->vmcs != NULL) {
+ /*
+ * XXX This should probably be just vcpu_clear. However, we need
+ * to get the per cpu lists working properly before we can do
+ * that.
+ */
+ __vcpu_clear(vmx);
+ kmem_free(vmx->vmcs, PAGESIZE);
+ vmx->vmcs = NULL;
+ }
+ if (vmx->guest_msrs != NULL)
+ kmem_free(vmx->guest_msrs, PAGESIZE);
+ kvm_vcpu_uninit(vcpu);
+ mutex_enter(&vmx_vpid_lock);
+ if (vmx->vpid != 0)
+ __clear_bit(vmx->vpid, vmx_vpid_bitmap);
+ mutex_exit(&vmx_vpid_lock);
+ kmem_cache_free(kvm_vcpu_cache, vmx);
+}
+
+struct kvm_vcpu *
+vmx_create_vcpu(struct kvm *kvm, unsigned int id)
+{
+ int err;
+ struct vcpu_vmx *vmx = kmem_cache_alloc(kvm_vcpu_cache, KM_SLEEP);
+ int cpu;
+
+ if (!vmx)
+ return (NULL);
+
+ allocate_vpid(vmx);
+ err = kvm_vcpu_init(&vmx->vcpu, kvm, id);
+ if (err) {
+ kmem_cache_free(kvm_vcpu_cache, vmx);
+ return (NULL);
+ }
+
+ vmx->guest_msrs = kmem_zalloc(PAGESIZE, KM_SLEEP);
+ if (!vmx->guest_msrs) {
+ return (NULL); /* XXX - need cleanup here */
+ }
+
+ vmx->vmcs = kmem_zalloc(PAGESIZE, KM_SLEEP);
+ if (!vmx->vmcs) {
+ kmem_cache_free(kvm_vcpu_cache, vmx);
+ vmx = NULL;
+ return (NULL);
+ }
+
+ vmx->vmcs_pa = (hat_getpfnum(kas.a_hat, (caddr_t)vmx->vmcs) <<
+ PAGESHIFT) | ((int64_t)(vmx->vmcs) & 0xfff);
+
+ kpreempt_disable();
+
+ cpu = curthread->t_cpu->cpu_seqid;
+
+ cmn_err(CE_NOTE, "vmcs revision_id = %x\n", vmcs_config.revision_id);
+ vmx->vmcs->revision_id = vmcs_config.revision_id;
+
+ vmcs_clear(vmx->vmcs_pa);
+
+ vmx_vcpu_load(&vmx->vcpu, cpu);
+ err = vmx_vcpu_setup(vmx);
+ vmx_vcpu_put(&vmx->vcpu);
+
+ kpreempt_enable();
+ if (err)
+ vmx->vmcs = NULL;
+ if (vm_need_virtualize_apic_accesses(kvm))
+ if (alloc_apic_access_page(kvm) != 0)
+ goto free_vmcs;
+
+ if (enable_ept) {
+ if (!kvm->arch.ept_identity_map_addr)
+ kvm->arch.ept_identity_map_addr =
+ VMX_EPT_IDENTITY_PAGETABLE_ADDR;
+ if (alloc_identity_pagetable(kvm) != 0)
+ goto free_vmcs;
+ }
+
+ return (&vmx->vcpu);
+
+free_vmcs:
+ kmem_free(vmx->vmcs, PAGESIZE);
+ vmx->vmcs = 0;
+#ifdef XXX
+free_msrs:
+ kfree(vmx->guest_msrs);
+uninit_vcpu:
+ kvm_vcpu_uninit(&vmx->vcpu);
+free_vcpu:
+ kmem_cache_free(kvm_vcpu_cache, vmx);
+#else
+ XXX_KVM_PROBE;
+#endif
+ return (NULL);
+}
+
+static void
+vmx_check_processor_compat(void *rtn)
+{
+ struct vmcs_config vmcs_conf;
+
+ if (setup_vmcs_config(&vmcs_conf) < 0)
+ *(int *)rtn |= EIO;
+ if (memcmp(&vmcs_config, &vmcs_conf, sizeof (struct vmcs_config))
+ != 0) {
+ cmn_err(CE_WARN, "kvm: CPU %d feature inconsistency!\n",
+ curthread->t_cpu->cpu_id);
+ *(int *)rtn |= EIO;
+ }
+}
+
+static int
+get_ept_level(void)
+{
+ return (VMX_EPT_DEFAULT_GAW + 1);
+}
+
+static uint64_t
+vmx_get_mt_mask(struct kvm_vcpu *vcpu, gfn_t gfn, int is_mmio)
+{
+ /*
+ * For VT-d and EPT combination
+ * 1. MMIO: always map as UC
+ * 2. EPT with VT-d:
+ * a. VT-d without snooping control feature: can't guarantee the
+ * result, try to trust guest.
+ * b. VT-d with snooping control feature: snooping control feature of
+ * VT-d engine can guarantee the cache correctness. Just set it
+ * to WB to keep consistent with host. So the same as item 3.
+ * 3. EPT without VT-d: always map as WB and set IPAT=1 to keep
+ * consistent with host MTRR
+ */
+ if (is_mmio)
+ return (MTRR_TYPE_UNCACHABLE << VMX_EPT_MT_EPTE_SHIFT);
+
+ if (vcpu->kvm->arch.iommu_domain &&
+ !(vcpu->kvm->arch.iommu_flags & KVM_IOMMU_CACHE_COHERENCY)) {
+ return (kvm_get_guest_memory_type(vcpu, gfn) <<
+ VMX_EPT_MT_EPTE_SHIFT);
+ }
+
+ return ((MTRR_TYPE_WRBACK << VMX_EPT_MT_EPTE_SHIFT) | VMX_EPT_IPAT_BIT);
+}
+
+#define _ER(x) { EXIT_REASON_##x, #x }
+
+struct trace_print_flags {
+ unsigned long mask;
+ const char *name;
+};
+
+static const struct trace_print_flags vmx_exit_reasons_str[] = {
+ _ER(EXCEPTION_NMI),
+ _ER(EXTERNAL_INTERRUPT),
+ _ER(TRIPLE_FAULT),
+ _ER(PENDING_INTERRUPT),
+ _ER(NMI_WINDOW),
+ _ER(TASK_SWITCH),
+ _ER(CPUID),
+ _ER(HLT),
+ _ER(INVLPG),
+ _ER(RDPMC),
+ _ER(RDTSC),
+ _ER(VMCALL),
+ _ER(VMCLEAR),
+ _ER(VMLAUNCH),
+ _ER(VMPTRLD),
+ _ER(VMPTRST),
+ _ER(VMREAD),
+ _ER(VMRESUME),
+ _ER(VMWRITE),
+ _ER(VMOFF),
+ _ER(VMON),
+ _ER(CR_ACCESS),
+ _ER(DR_ACCESS),
+ _ER(IO_INSTRUCTION),
+ _ER(MSR_READ),
+ _ER(MSR_WRITE),
+ _ER(MWAIT_INSTRUCTION),
+ _ER(MONITOR_INSTRUCTION),
+ _ER(PAUSE_INSTRUCTION),
+ _ER(MCE_DURING_VMENTRY),
+ _ER(TPR_BELOW_THRESHOLD),
+ _ER(APIC_ACCESS),
+ _ER(EPT_VIOLATION),
+ _ER(EPT_MISCONFIG),
+ _ER(WBINVD),
+ { -1, NULL }
+};
+
+#undef _ER
+
+static int
+vmx_get_lpage_level(void)
+{
+ if (enable_ept && !cpu_has_vmx_ept_1g_page())
+ return (PT_DIRECTORY_LEVEL);
+ else
+ /* For shadow and EPT supported 1GB page */
+ return (PT_PDPE_LEVEL);
+}
+
+static void
+vmx_cpuid_update(struct kvm_vcpu *vcpu)
+{
+ struct kvm_cpuid_entry2 *best;
+ struct vcpu_vmx *vmx = to_vmx(vcpu);
+ uint32_t exec_control;
+
+ vmx->rdtscp_enabled = 0;
+
+ if (vmx_rdtscp_supported()) {
+ exec_control = vmcs_read32(SECONDARY_VM_EXEC_CONTROL);
+ if (exec_control & SECONDARY_EXEC_RDTSCP) {
+ best = kvm_find_cpuid_entry(vcpu, 0x80000001, 0);
+ if (best && (best->edx & bit(X86_FEATURE_RDTSCP)))
+ vmx->rdtscp_enabled = 1;
+ else {
+ exec_control &= ~SECONDARY_EXEC_RDTSCP;
+ vmcs_write32(SECONDARY_VM_EXEC_CONTROL,
+ exec_control);
+ }
+ }
+ }
+}
+
+struct kvm_x86_ops vmx_x86_ops = {
+ .cpu_has_kvm_support = nulldev, /* XXX: cpu_has_kvm_support? */
+ .disabled_by_bios = nulldev, /* XXX: vmx_disabled_by_bios? */
+
+ .hardware_enable = vmx_hardware_enable,
+ .hardware_disable = vmx_hardware_disable,
+
+ .check_processor_compatibility = vmx_check_processor_compat,
+
+ .hardware_setup = vmx_hardware_setup,
+
+ .hardware_unsetup = (void(*)(void))nulldev, /* XXX: hardware_unsetup? */
+
+ .cpu_has_accelerated_tpr = report_flexpriority,
+ .vcpu_create = vmx_create_vcpu,
+ .vcpu_free = vmx_destroy_vcpu, /* XXX */
+ .vcpu_reset = vmx_vcpu_reset,
+
+ .prepare_guest_switch = vmx_save_host_state,
+ .vcpu_load = vmx_vcpu_load,
+ .vcpu_put = vmx_vcpu_put,
+
+ .set_guest_debug = set_guest_debug,
+ .get_msr = vmx_get_msr,
+ .set_msr = vmx_set_msr,
+ .get_segment_base = vmx_get_segment_base,
+ .get_segment = vmx_get_segment,
+ .set_segment = vmx_set_segment,
+ .get_cpl = vmx_get_cpl,
+ .get_cs_db_l_bits = vmx_get_cs_db_l_bits,
+ .decache_cr0_guest_bits = vmx_decache_cr0_guest_bits,
+ .decache_cr4_guest_bits = vmx_decache_cr4_guest_bits,
+ .set_cr0 = vmx_set_cr0,
+ .set_cr3 = vmx_set_cr3,
+ .set_cr4 = vmx_set_cr4,
+ .set_efer = vmx_set_efer,
+ .get_idt = vmx_get_idt,
+ .set_idt = vmx_set_idt,
+ .get_gdt = vmx_get_gdt,
+ .set_gdt = vmx_set_gdt,
+ .cache_reg = vmx_cache_reg,
+ .get_rflags = vmx_get_rflags,
+ .set_rflags = vmx_set_rflags,
+ .fpu_activate = vmx_fpu_activate,
+ .fpu_deactivate = vmx_fpu_deactivate,
+
+ .tlb_flush = vmx_flush_tlb,
+
+ .run = vmx_vcpu_run,
+ .handle_exit = vmx_handle_exit,
+ .skip_emulated_instruction = skip_emulated_instruction,
+ .set_interrupt_shadow = vmx_set_interrupt_shadow,
+ .get_interrupt_shadow = vmx_get_interrupt_shadow,
+ .patch_hypercall = vmx_patch_hypercall,
+ .set_irq = vmx_inject_irq,
+ .set_nmi = vmx_inject_nmi,
+ .queue_exception = vmx_queue_exception,
+ .interrupt_allowed = vmx_interrupt_allowed,
+ .nmi_allowed = vmx_nmi_allowed,
+ .get_nmi_mask = vmx_get_nmi_mask,
+ .set_nmi_mask = vmx_set_nmi_mask,
+ .enable_nmi_window = enable_nmi_window,
+ .enable_irq_window = enable_irq_window,
+ .update_cr8_intercept = vmx_update_cr8_intercept,
+
+ .set_tss_addr = vmx_set_tss_addr,
+ .get_tdp_level = get_ept_level,
+ .get_mt_mask = vmx_get_mt_mask,
+
+ .exit_reasons_str = vmx_exit_reasons_str,
+
+ .get_lpage_level = vmx_get_lpage_level,
+
+ .cpuid_update = vmx_cpuid_update,
+
+ .rdtscp_supported = vmx_rdtscp_supported
+};
+
+int
+vmx_init(void)
+{
+ int r, i;
+
+ rdmsrl_safe(MSR_EFER, (unsigned long long *)&host_efer);
+
+ for (i = 0; i < NR_VMX_MSR; ++i)
+ kvm_define_shared_msr(i, vmx_msr_index[i]);
+
+#ifdef XXX
+ vmx_io_bitmap_a = kmem_zalloc(PAGESIZE, KM_SLEEP);
+ vmx_io_bitmap_b = kmem_zalloc(PAGESIZE, KM_SLEEP);
+ vmx_msr_bitmap_legacy = kmem_zalloc(PAGESIZE, KM_SLEEP);
+ vmx_msr_bitmap_longmode = kmem_zalloc(PAGESIZE, KM_SLEEP);
+#else
+ XXX_KVM_PROBE;
+#endif
+
+ /*
+ * Allow direct access to the PC debug port (it is often used for I/O
+ * delays, but the vmexits simply slow things down).
+ */
+ memset(vmx_io_bitmap_a, 0xff, PAGESIZE);
+ clear_bit(0x80, vmx_io_bitmap_a);
+
+ memset(vmx_io_bitmap_b, 0xff, PAGESIZE);
+
+ memset(vmx_msr_bitmap_legacy, 0xff, PAGESIZE);
+ memset(vmx_msr_bitmap_longmode, 0xff, PAGESIZE);
+
+ set_bit(0, vmx_vpid_bitmap); /* 0 is reserved for host */
+
+ r = kvm_init(&vmx_x86_ops, sizeof (struct vcpu_vmx));
+
+ if (r)
+ goto out3;
+
+ vmx_disable_intercept_for_msr(MSR_FS_BASE, 0);
+ vmx_disable_intercept_for_msr(MSR_GS_BASE, 0);
+ vmx_disable_intercept_for_msr(MSR_KERNEL_GS_BASE, 1);
+ vmx_disable_intercept_for_msr(MSR_IA32_SYSENTER_CS, 0);
+ vmx_disable_intercept_for_msr(MSR_IA32_SYSENTER_ESP, 0);
+ vmx_disable_intercept_for_msr(MSR_IA32_SYSENTER_EIP, 0);
+
+ if (enable_ept) {
+ bypass_guest_pf = 0;
+ kvm_mmu_set_base_ptes(VMX_EPT_READABLE_MASK |
+ VMX_EPT_WRITABLE_MASK);
+ kvm_mmu_set_mask_ptes(0ull, 0ull, 0ull, 0ull,
+ VMX_EPT_EXECUTABLE_MASK);
+ kvm_enable_tdp();
+ } else
+ kvm_disable_tdp();
+
+ if (bypass_guest_pf)
+ kvm_mmu_set_nonpresent_ptes(~0xffeull, 0ull);
+
+ return (0);
+
+out3:
+ kmem_free(vmx_msr_bitmap_longmode, PAGESIZE);
+out2:
+ kmem_free(vmx_msr_bitmap_legacy, PAGESIZE);
+out1:
+ kmem_free(vmx_io_bitmap_b, PAGESIZE);
+out:
+ kmem_free(vmx_io_bitmap_a, PAGESIZE);
+
+ return (r);
+}
diff --git a/kvm_x86.c b/kvm_x86.c
index 7d51fae..9a24eee 100644
--- a/kvm_x86.c
+++ b/kvm_x86.c
@@ -63,7 +63,6 @@ extern int kvm_is_mmio_pfn(pfn_t pfn);
extern ulong kvm_read_cr4_bits(struct kvm_vcpu *vcpu, ulong mask);
extern int is_long_mode(struct kvm_vcpu *vcpu);
extern void kvm_mmu_unload(struct kvm_vcpu *);
-extern void __vcpu_clear(void *);
extern void kvm_free_physmem_slot(struct kvm_memory_slot *,
struct kvm_memory_slot *);
@@ -320,85 +319,15 @@ extern uint64_t shadow_accessed_mask;
extern uint64_t shadow_dirty_mask;
extern pfn_t hat_getpfnum(hat_t *hat, caddr_t addr);
-struct vmcs_config vmcs_config;
extern inline void ept_sync_global(void);
extern uint64_t *vmxarea_pa;
extern list_t **vcpus_on_cpu;
-int
-vmx_hardware_enable(void *garbage)
-{
- int cpu = curthread->t_cpu->cpu_seqid;
- pfn_t pfn;
- uint64_t old;
-#ifdef XXX
- uint64_t phys_addr = kvtop(per_cpu(vmxarea, cpu));
-#else
- uint64_t phys_addr;
- XXX_KVM_PROBE;
- phys_addr = vmxarea_pa[cpu];
-
-#endif
-
- ((struct vmcs *)(vmxarea[cpu]))->revision_id = vmcs_config.revision_id;
-
- if (getcr4() & X86_CR4_VMXE)
- return (DDI_FAILURE);
-
-#ifdef XXX
- INIT_LIST_HEAD(&per_cpu(vcpus_on_cpu, cpu));
-#else
- XXX_KVM_PROBE;
-#endif
- rdmsrl(MSR_IA32_FEATURE_CONTROL, old);
- if ((old & (FEATURE_CONTROL_LOCKED | FEATURE_CONTROL_VMXON_ENABLED)) !=
- (FEATURE_CONTROL_LOCKED | FEATURE_CONTROL_VMXON_ENABLED)) {
- /* enable and lock */
- wrmsrl(MSR_IA32_FEATURE_CONTROL, old | FEATURE_CONTROL_LOCKED |
- FEATURE_CONTROL_VMXON_ENABLED);
- }
-
- setcr4(getcr4() | X86_CR4_VMXE); /* FIXME: not cpu hotplug safe */
- /* BEGIN CSTYLED */
- __asm__ volatile (ASM_VMX_VMXON_RAX
- : : "a"(&phys_addr), "m"(phys_addr)
- : "memory", "cc");
- /* END CSTYLED */
-
- ept_sync_global();
-
- return (0);
-}
extern struct vcpu_vmx *to_vmx(struct kvm_vcpu *vcpu);
extern void vmcs_writel(unsigned long field, unsigned long value);
extern unsigned long vmcs_readl(unsigned long field);
-unsigned long
-vmx_get_rflags(struct kvm_vcpu *vcpu)
-{
- unsigned long rflags, save_rflags;
-
- rflags = vmcs_readl(GUEST_RFLAGS);
- if (to_vmx(vcpu)->rmode.vm86_active) {
- rflags &= RMODE_GUEST_OWNED_EFLAGS_BITS;
- save_rflags = to_vmx(vcpu)->rmode.save_rflags;
- rflags |= save_rflags & ~RMODE_GUEST_OWNED_EFLAGS_BITS;
- }
-
- return (rflags);
-}
-
-void
-vmx_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags)
-{
- if (to_vmx(vcpu)->rmode.vm86_active) {
- to_vmx(vcpu)->rmode.save_rflags = rflags;
- rflags |= X86_EFLAGS_IOPL | X86_EFLAGS_VM;
- }
-
- vmcs_writel(GUEST_RFLAGS, rflags);
-}
extern void kvm_shared_msr_cpu_online(void);
@@ -419,48 +348,13 @@ kvm_arch_hardware_enable(void *garbage)
#endif
kvm_shared_msr_cpu_online();
- return (vmx_hardware_enable(garbage));
-}
-
-static void vmclear_local_vcpus(void)
-{
- int cpu = CPU->cpu_id;
- struct vcpu_vmx *vmx, *n;
-
- /*
- * list_for_each_entry_safe(vmx, n, &per_cpu(vcpus_on_cpu, cpu),
- * local_vcpus_link)
- * __vcpu_clear(vmx);
- */
- for (vmx = list_head(vcpus_on_cpu[cpu]); vmx;
- vmx = list_next(vcpus_on_cpu[cpu], vmx))
- __vcpu_clear(vmx);
-}
-
-
-/*
- * Just like cpu_vmxoff(), but with the __kvm_handle_fault_on_reboot()
- * tricks.
- */
-static void kvm_cpu_vmxoff(void)
-{
- /* BEGIN CSTYLED */
- __asm__ volatile ((ASM_VMX_VMXOFF) : : : "cc");
- /* END CSTYLED */
- setcr4(getcr4() & ~X86_CR4_VMXE);
-}
-
-static void vmx_hardware_disable(void *garbage)
-{
- vmclear_local_vcpus();
- kvm_cpu_vmxoff();
+ return (kvm_x86_ops->hardware_enable(garbage));
}
-
void
kvm_arch_hardware_disable(void *garbage)
{
- vmx_hardware_disable(garbage);
+ kvm_x86_ops->hardware_disable(garbage);
#if defined(CONFIG_MMU_NOTIFIER) && defined(KVM_ARCH_WANT_MMU_NOTIFIER)
drop_user_return_notifiers(garbage);
#endif
@@ -938,192 +832,8 @@ extern int enable_vpid;
extern ulong_t *vmx_vpid_bitmap;
extern kmutex_t vmx_vpid_lock;
-static void
-allocate_vpid(struct vcpu_vmx *vmx)
-{
- int vpid;
-
- vmx->vpid = 0;
- if (!enable_vpid)
- return;
- mutex_enter(&vmx_vpid_lock);
- vpid = find_first_zero_bit(vmx_vpid_bitmap, VMX_NR_VPIDS);
- if (vpid < VMX_NR_VPIDS) {
- vmx->vpid = vpid;
- __set_bit(vpid, vmx_vpid_bitmap);
- }
- mutex_exit(&vmx_vpid_lock);
-}
-
extern page_t *gfn_to_page(struct kvm *kvm, gfn_t gfn);
-static int
-alloc_identity_pagetable(struct kvm *kvm)
-{
- struct kvm_userspace_memory_region kvm_userspace_mem;
- int r = 0;
-
- mutex_enter(&kvm->slots_lock);
- if (kvm->arch.ept_identity_pagetable)
- goto out;
- kvm_userspace_mem.slot = IDENTITY_PAGETABLE_PRIVATE_MEMSLOT;
- kvm_userspace_mem.flags = 0;
- kvm_userspace_mem.guest_phys_addr =
- kvm->arch.ept_identity_map_addr;
- kvm_userspace_mem.memory_size = PAGESIZE;
-
- kvm_userspace_mem.userspace_addr = 0;
- r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, 0);
- if (r)
- goto out;
-
- kvm->arch.ept_identity_pagetable = gfn_to_page(kvm,
- kvm->arch.ept_identity_map_addr >> PAGESHIFT);
-out:
- mutex_exit(&kvm->slots_lock);
- return (r);
-}
-
-static int
-alloc_apic_access_page(struct kvm *kvm)
-{
- struct kvm_userspace_memory_region kvm_userspace_mem;
- int r = 0;
-
- memset(&kvm_userspace_mem, 0,
- sizeof (struct kvm_userspace_memory_region));
-
- mutex_enter(&kvm->slots_lock);
- if (kvm->arch.apic_access_page)
- goto out;
- kvm_userspace_mem.slot = APIC_ACCESS_PAGE_PRIVATE_MEMSLOT;
- kvm_userspace_mem.flags = 0;
- kvm_userspace_mem.guest_phys_addr = 0xfee00000ULL;
- kvm_userspace_mem.memory_size = PAGESIZE;
- r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, 0);
- if (r)
- goto out;
-
- kvm->arch.apic_access_page = gfn_to_page(kvm, 0xfee00);
-out:
- mutex_exit(&kvm->slots_lock);
- return (r);
-}
-
-static void
-vcpu_clear(struct vcpu_vmx *vmx)
-{
- if (vmx->vcpu.cpu == -1)
- return;
-
- XXX_KVM_SYNC_PROBE;
- __vcpu_clear(vmx);
-}
-
-struct kvm_vcpu *
-vmx_create_vcpu(struct kvm *kvm, unsigned int id)
-{
- int err;
- struct vcpu_vmx *vmx = kmem_cache_alloc(kvm_vcpu_cache, KM_SLEEP);
- int cpu;
-
- if (!vmx)
- return (NULL);
-
- allocate_vpid(vmx);
- err = kvm_vcpu_init(&vmx->vcpu, kvm, id);
- if (err) {
- kmem_cache_free(kvm_vcpu_cache, vmx);
- return (NULL);
- }
-
- vmx->guest_msrs = kmem_zalloc(PAGESIZE, KM_SLEEP);
- if (!vmx->guest_msrs) {
- return (NULL); /* XXX - need cleanup here */
- }
-
- vmx->vmcs = kmem_zalloc(PAGESIZE, KM_SLEEP);
- if (!vmx->vmcs) {
- kmem_cache_free(kvm_vcpu_cache, vmx);
- vmx = NULL;
- return (NULL);
- }
-
- vmx->vmcs_pa = (hat_getpfnum(kas.a_hat, (caddr_t)vmx->vmcs) <<
- PAGESHIFT) | ((int64_t)(vmx->vmcs) & 0xfff);
-
- kpreempt_disable();
-
- cpu = curthread->t_cpu->cpu_seqid;
-
- cmn_err(CE_NOTE, "vmcs revision_id = %x\n", vmcs_config.revision_id);
- vmx->vmcs->revision_id = vmcs_config.revision_id;
-
- vmcs_clear(vmx->vmcs_pa);
-
- vmx_vcpu_load(&vmx->vcpu, cpu);
- err = vmx_vcpu_setup(vmx);
- vmx_vcpu_put(&vmx->vcpu);
-
- kpreempt_enable();
- if (err)
- vmx->vmcs = NULL;
- if (vm_need_virtualize_apic_accesses(kvm))
- if (alloc_apic_access_page(kvm) != 0)
- goto free_vmcs;
-
- if (enable_ept) {
- if (!kvm->arch.ept_identity_map_addr)
- kvm->arch.ept_identity_map_addr =
- VMX_EPT_IDENTITY_PAGETABLE_ADDR;
- if (alloc_identity_pagetable(kvm) != 0)
- goto free_vmcs;
- }
-
- return (&vmx->vcpu);
-
-free_vmcs:
- kmem_free(vmx->vmcs, PAGESIZE);
- vmx->vmcs = 0;
-#ifdef XXX
-free_msrs:
- kfree(vmx->guest_msrs);
-uninit_vcpu:
- kvm_vcpu_uninit(&vmx->vcpu);
-free_vcpu:
- kmem_cache_free(kvm_vcpu_cache, vmx);
-#else
- XXX_KVM_PROBE;
-#endif
- return (NULL);
-}
-
-void
-vmx_destroy_vcpu(struct kvm_vcpu *vcpu)
-{
- /* XXX don't assume it's the first element */
- vcpu_vmx_t *vmx = (vcpu_vmx_t *)vcpu;
-
- if (vmx->vmcs != NULL) {
- /*
- * XXX This should probably be just vcpu_clear. However, we need
- * to get the per cpu lists working properly before we can do
- * that.
- */
- __vcpu_clear(vmx);
- kmem_free(vmx->vmcs, PAGESIZE);
- vmx->vmcs = NULL;
- }
- if (vmx->guest_msrs != NULL)
- kmem_free(vmx->guest_msrs, PAGESIZE);
- kvm_vcpu_uninit(vcpu);
- mutex_enter(&vmx_vpid_lock);
- if (vmx->vpid != 0)
- __clear_bit(vmx->vpid, vmx_vpid_bitmap);
- mutex_exit(&vmx_vpid_lock);
- kmem_cache_free(kvm_vcpu_cache, vmx);
-}
-
struct kvm_vcpu *
kvm_arch_vcpu_create(struct kvm *kvm, unsigned int id)
{
@@ -1197,29 +907,6 @@ kvm_arch_vcpu_free(struct kvm_vcpu *vcpu)
kvm_x86_ops->vcpu_free(vcpu);
}
-void
-update_exception_bitmap(struct kvm_vcpu *vcpu)
-{
- uint32_t eb;
-
- eb = (1u << PF_VECTOR) | (1u << UD_VECTOR) | (1u << MC_VECTOR) |
- (1u << NM_VECTOR) | (1u << DB_VECTOR);
-
-#ifndef XXX
- if ((vcpu->guest_debug &
- (KVM_GUESTDBG_ENABLE | KVM_GUESTDBG_USE_SW_BP)) ==
- (KVM_GUESTDBG_ENABLE | KVM_GUESTDBG_USE_SW_BP))
- eb |= 1u << BP_VECTOR;
-#endif
- if (to_vmx(vcpu)->rmode.vm86_active)
- eb = ~0;
- if (enable_ept)
- eb &= ~(1u << PF_VECTOR); /* bypass_guest_pf = 0 */
- if (vcpu->fpu_active)
- eb &= ~(1u << NM_VECTOR);
- vmcs_write32(EXCEPTION_BITMAP, eb);
-}
-
uint64_t
kvm_get_apic_base(struct kvm_vcpu *vcpu)
@@ -1262,382 +949,9 @@ is_paging(struct kvm_vcpu *vcpu)
return (kvm_read_cr0_bits(vcpu, X86_CR0_PG));
}
-void
-vmx_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
-{
- unsigned long hw_cr4 = cr4 | (to_vmx(vcpu)->rmode.vm86_active ?
- KVM_RMODE_VM_CR4_ALWAYS_ON : KVM_PMODE_VM_CR4_ALWAYS_ON);
-
- vcpu->arch.cr4 = cr4;
- if (enable_ept) {
- if (!is_paging(vcpu)) {
- hw_cr4 &= ~X86_CR4_PAE;
- hw_cr4 |= X86_CR4_PSE;
- } else if (!(cr4 & X86_CR4_PAE)) {
- hw_cr4 &= ~X86_CR4_PAE;
- }
- }
-
- vmcs_writel(CR4_READ_SHADOW, cr4);
- vmcs_writel(GUEST_CR4, hw_cr4);
-}
-
-static void
-ept_update_paging_mode_cr0(unsigned long *hw_cr0,
- unsigned long cr0, struct kvm_vcpu *vcpu)
-{
- if (!(cr0 & X86_CR0_PG)) {
- /* From paging/starting to nonpaging */
- vmcs_write32(CPU_BASED_VM_EXEC_CONTROL,
- vmcs_read32(CPU_BASED_VM_EXEC_CONTROL) |
- (CPU_BASED_CR3_LOAD_EXITING |
- CPU_BASED_CR3_STORE_EXITING));
- vcpu->arch.cr0 = cr0;
- vmx_set_cr4(vcpu, kvm_read_cr4(vcpu));
- } else if (!is_paging(vcpu)) {
- /* From nonpaging to paging */
- vmcs_write32(CPU_BASED_VM_EXEC_CONTROL,
- vmcs_read32(CPU_BASED_VM_EXEC_CONTROL) &
- ~(CPU_BASED_CR3_LOAD_EXITING |
- CPU_BASED_CR3_STORE_EXITING));
- vcpu->arch.cr0 = cr0;
- vmx_set_cr4(vcpu, kvm_read_cr4(vcpu));
- }
-
- if (!(cr0 & X86_CR0_WP))
- *hw_cr0 &= ~X86_CR0_WP;
-}
-
-#define VMX_SEGMENT_FIELD(seg) \
- [VCPU_SREG_##seg] = { \
- .selector = GUEST_##seg##_SELECTOR, \
- .base = GUEST_##seg##_BASE, \
- .limit = GUEST_##seg##_LIMIT, \
- .ar_bytes = GUEST_##seg##_AR_BYTES, \
- }
-
-extern struct kvm_vmx_segment_field kvm_vmx_segment_fields[];
-
-static void fix_pmode_dataseg(int seg, struct kvm_save_segment *save)
-{
- struct kvm_vmx_segment_field *sf = &kvm_vmx_segment_fields[seg];
-
- if (vmcs_readl(sf->base) == save->base && (save->base & AR_S_MASK)) {
- vmcs_write16(sf->selector, save->selector);
- vmcs_writel(sf->base, save->base);
- vmcs_write32(sf->limit, save->limit);
- vmcs_write32(sf->ar_bytes, save->ar);
- } else {
- uint32_t dpl = (vmcs_read16(sf->selector) & SELECTOR_RPL_MASK)
- << AR_DPL_SHIFT;
- vmcs_write32(sf->ar_bytes, 0x93 | dpl);
- }
-}
-
-static void enter_pmode(struct kvm_vcpu *vcpu)
-{
- unsigned long flags;
- struct vcpu_vmx *vmx = to_vmx(vcpu);
-
- vmx->emulation_required = 1;
- vmx->rmode.vm86_active = 0;
-
- vmcs_writel(GUEST_TR_BASE, vmx->rmode.tr.base);
- vmcs_write32(GUEST_TR_LIMIT, vmx->rmode.tr.limit);
- vmcs_write32(GUEST_TR_AR_BYTES, vmx->rmode.tr.ar);
-
- flags = vmcs_readl(GUEST_RFLAGS);
- flags &= RMODE_GUEST_OWNED_EFLAGS_BITS;
- flags |= vmx->rmode.save_rflags & ~RMODE_GUEST_OWNED_EFLAGS_BITS;
- vmcs_writel(GUEST_RFLAGS, flags);
-
- vmcs_writel(GUEST_CR4, (vmcs_readl(GUEST_CR4) & ~X86_CR4_VME) |
- (vmcs_readl(CR4_READ_SHADOW) & X86_CR4_VME));
-
- update_exception_bitmap(vcpu);
-
- if (emulate_invalid_guest_state)
- return;
-
- fix_pmode_dataseg(VCPU_SREG_ES, &vmx->rmode.es);
- fix_pmode_dataseg(VCPU_SREG_DS, &vmx->rmode.ds);
- fix_pmode_dataseg(VCPU_SREG_GS, &vmx->rmode.gs);
- fix_pmode_dataseg(VCPU_SREG_FS, &vmx->rmode.fs);
-
- vmcs_write16(GUEST_SS_SELECTOR, 0);
- vmcs_write32(GUEST_SS_AR_BYTES, 0x93);
-
- vmcs_write16(GUEST_CS_SELECTOR,
- vmcs_read16(GUEST_CS_SELECTOR) & ~SELECTOR_RPL_MASK);
- vmcs_write32(GUEST_CS_AR_BYTES, 0x9b);
-}
-
-static gva_t
-rmode_tss_base(struct kvm *kvm)
-{
- if (!kvm->arch.tss_addr) {
- struct kvm_memslots *slots;
- gfn_t base_gfn;
-
-#ifdef XXX
- slots = rcu_dereference(kvm->memslots);
-#else
- XXX_KVM_PROBE;
- slots = kvm->memslots;
-#endif
- base_gfn = kvm->memslots->memslots[0].base_gfn +
- kvm->memslots->memslots[0].npages - 3;
- return (base_gfn << PAGESHIFT);
- }
-
- return (kvm->arch.tss_addr);
-}
-
-static void
-fix_rmode_seg(int seg, struct kvm_save_segment *save)
-{
- struct kvm_vmx_segment_field *sf = &kvm_vmx_segment_fields[seg];
-
- save->selector = vmcs_read16(sf->selector);
- save->base = vmcs_readl(sf->base);
- save->limit = vmcs_read32(sf->limit);
- save->ar = vmcs_read32(sf->ar_bytes);
- vmcs_write16(sf->selector, save->base >> 4);
- vmcs_write32(sf->base, save->base & 0xfffff);
- vmcs_write32(sf->limit, 0xffff);
- vmcs_write32(sf->ar_bytes, 0xf3);
-}
-
-static int
-init_rmode_tss(struct kvm *kvm)
-{
- gfn_t fn = rmode_tss_base(kvm) >> PAGESHIFT;
- uint16_t data = 0;
- int ret = 0;
- int r;
-
- r = kvm_clear_guest_page(kvm, fn, 0, PAGESIZE);
- if (r < 0)
- goto out;
- data = TSS_BASE_SIZE + TSS_REDIRECTION_SIZE;
- r = kvm_write_guest_page(kvm, fn++, &data,
- TSS_IOPB_BASE_OFFSET, sizeof (uint16_t));
-
- if (r < 0)
- goto out;
- r = kvm_clear_guest_page(kvm, fn++, 0, PAGESIZE);
- if (r < 0)
- goto out;
- r = kvm_clear_guest_page(kvm, fn, 0, PAGESIZE);
- if (r < 0)
- goto out;
- data = ~0;
- r = kvm_write_guest_page(kvm, fn, &data,
- RMODE_TSS_SIZE - 2 * PAGESIZE - 1, sizeof (uint8_t));
-
- if (r < 0)
- goto out;
-
- ret = 1;
-out:
- return (ret);
-}
-
-static int
-init_rmode_identity_map(struct kvm *kvm)
-{
- int i, r, ret;
- pfn_t identity_map_pfn;
- uint32_t tmp;
-
- if (!enable_ept)
- return (1);
- if ((!kvm->arch.ept_identity_pagetable)) {
- cmn_err(CE_WARN, "EPT: identity-mapping pagetable "
- "haven't been allocated!\n");
- return (0);
- }
- if ((kvm->arch.ept_identity_pagetable_done))
- return (1);
-
- ret = 0;
- identity_map_pfn = kvm->arch.ept_identity_map_addr >> PAGESHIFT;
- r = kvm_clear_guest_page(kvm, identity_map_pfn, 0, PAGESIZE);
- if (r < 0)
- goto out;
-
- /* Set up identity-mapping pagetable for EPT in real mode */
- for (i = 0; i < PT32_ENT_PER_PAGE; i++) {
- tmp = (i << 22) + (PT_VALID | PT_WRITABLE | PT_USER |
- PT_REF | PT_MOD | PT_PAGESIZE);
-
- r = kvm_write_guest_page(kvm, identity_map_pfn,
- &tmp, i * sizeof (tmp), sizeof (tmp));
-
- if (r < 0)
- goto out;
- }
- kvm->arch.ept_identity_pagetable_done = 1;
- ret = 1;
-out:
- return (ret);
-}
-
-static int
-init_rmode(struct kvm *kvm)
-{
- if (!init_rmode_tss(kvm))
- return (0);
-
- if (!init_rmode_identity_map(kvm))
- return (0);
-
- return (1);
-}
-
-
-static void enter_rmode(struct kvm_vcpu *vcpu)
-{
- unsigned long flags;
- struct vcpu_vmx *vmx = to_vmx(vcpu);
-
- if (enable_unrestricted_guest)
- return;
-
- vmx->emulation_required = 1;
- vmx->rmode.vm86_active = 1;
-
- vmx->rmode.tr.base = vmcs_readl(GUEST_TR_BASE);
- vmcs_writel(GUEST_TR_BASE, rmode_tss_base(vcpu->kvm));
-
- vmx->rmode.tr.limit = vmcs_read32(GUEST_TR_LIMIT);
- vmcs_write32(GUEST_TR_LIMIT, RMODE_TSS_SIZE - 1);
-
- vmx->rmode.tr.ar = vmcs_read32(GUEST_TR_AR_BYTES);
- vmcs_write32(GUEST_TR_AR_BYTES, 0x008b);
-
- flags = vmcs_readl(GUEST_RFLAGS);
- vmx->rmode.save_rflags = flags;
-
- flags |= X86_EFLAGS_IOPL | X86_EFLAGS_VM;
-
- vmcs_writel(GUEST_RFLAGS, flags);
- vmcs_writel(GUEST_CR4, vmcs_readl(GUEST_CR4) | X86_CR4_VME);
- update_exception_bitmap(vcpu);
-
- if (emulate_invalid_guest_state)
- goto continue_rmode;
-
- vmcs_write16(GUEST_SS_SELECTOR, vmcs_readl(GUEST_SS_BASE) >> 4);
- vmcs_write32(GUEST_SS_LIMIT, 0xffff);
- vmcs_write32(GUEST_SS_AR_BYTES, 0xf3);
-
- vmcs_write32(GUEST_CS_AR_BYTES, 0xf3);
- vmcs_write32(GUEST_CS_LIMIT, 0xffff);
- if (vmcs_readl(GUEST_CS_BASE) == 0xffff0000)
- vmcs_writel(GUEST_CS_BASE, 0xf0000);
- vmcs_write16(GUEST_CS_SELECTOR, vmcs_readl(GUEST_CS_BASE) >> 4);
-
- fix_rmode_seg(VCPU_SREG_ES, &vmx->rmode.es);
- fix_rmode_seg(VCPU_SREG_DS, &vmx->rmode.ds);
- fix_rmode_seg(VCPU_SREG_GS, &vmx->rmode.gs);
- fix_rmode_seg(VCPU_SREG_FS, &vmx->rmode.fs);
-
-continue_rmode:
- kvm_mmu_reset_context(vcpu);
- init_rmode(vcpu->kvm);
-}
extern void vmx_set_efer(struct kvm_vcpu *vcpu, uint64_t efer);
-#ifdef CONFIG_X86_64
-
-static void
-enter_lmode(struct kvm_vcpu *vcpu)
-{
- uint32_t guest_tr_ar;
-
- guest_tr_ar = vmcs_read32(GUEST_TR_AR_BYTES);
- if ((guest_tr_ar & AR_TYPE_MASK) != AR_TYPE_BUSY_64_TSS) {
- cmn_err(CE_NOTE, "%s: tss fixup for long mode. \n",
- __func__);
- vmcs_write32(GUEST_TR_AR_BYTES,
- (guest_tr_ar & ~AR_TYPE_MASK) | AR_TYPE_BUSY_64_TSS);
- }
- vcpu->arch.efer |= EFER_LMA;
- vmx_set_efer(vcpu, vcpu->arch.efer);
-}
-
-static void
-exit_lmode(struct kvm_vcpu *vcpu)
-{
- vcpu->arch.efer &= ~EFER_LMA;
-
- vmcs_write32(VM_ENTRY_CONTROLS,
- vmcs_read32(VM_ENTRY_CONTROLS) & ~VM_ENTRY_IA32E_MODE);
-}
-
-#endif
-
-void
-vmx_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0)
-{
- struct vcpu_vmx *vmx = to_vmx(vcpu);
- unsigned long hw_cr0;
-
- if (enable_unrestricted_guest) {
- hw_cr0 = (cr0 & ~KVM_GUEST_CR0_MASK_UNRESTRICTED_GUEST) |
- KVM_VM_CR0_ALWAYS_ON_UNRESTRICTED_GUEST;
- } else {
- hw_cr0 = (cr0 & ~KVM_GUEST_CR0_MASK) | KVM_VM_CR0_ALWAYS_ON;
- }
-
- if (vmx->rmode.vm86_active && (cr0 & X86_CR0_PE))
- enter_pmode(vcpu);
-
- if (!vmx->rmode.vm86_active && !(cr0 & X86_CR0_PE))
- enter_rmode(vcpu);
-
-#ifdef CONFIG_X86_64
- if (vcpu->arch.efer & EFER_LME) {
- if (!is_paging(vcpu) && (cr0 & X86_CR0_PG))
- enter_lmode(vcpu);
- if (is_paging(vcpu) && !(cr0 & X86_CR0_PG))
- exit_lmode(vcpu);
- }
-#endif
-
- if (enable_ept)
- ept_update_paging_mode_cr0(&hw_cr0, cr0, vcpu);
-
- if (!vcpu->fpu_active)
- hw_cr0 |= X86_CR0_TS | X86_CR0_MP;
-
- vmcs_writel(CR0_READ_SHADOW, cr0);
- vmcs_writel(GUEST_CR0, hw_cr0);
- vcpu->arch.cr0 = cr0;
-}
-
-static void
-seg_setup(int seg)
-{
- struct kvm_vmx_segment_field *sf = &kvm_vmx_segment_fields[seg];
- unsigned int ar;
-
- vmcs_write16(sf->selector, 0);
- vmcs_writel(sf->base, 0);
- vmcs_write32(sf->limit, 0xffff);
-
- if (enable_unrestricted_guest) {
- ar = 0x93;
- if (seg == VCPU_SREG_CS)
- ar |= 0x08; /* code segment */
- } else
- ar = 0xf3;
-
- vmcs_write32(sf->ar_bytes, ar);
-}
-
extern int kvm_write_guest_page(struct kvm *kvm,
gfn_t gfn, const void *data, int offset, int len);
@@ -1693,150 +1007,7 @@ extern void vmx_fpu_activate(struct kvm_vcpu *vcpu);
extern inline int vm_need_tpr_shadow(struct kvm *kvm);
extern inline int cpu_has_vmx_tpr_shadow(void);
-#define page_to_phys(page) (page->p_pagenum << PAGESHIFT)
-
-int
-vmx_vcpu_reset(struct kvm_vcpu *vcpu)
-{
- struct vcpu_vmx *vmx = to_vmx(vcpu);
- uint64_t msr;
- int ret, idx;
- page_t *ptp;
-
- vcpu->arch.regs_avail = ~((1 << VCPU_REGS_RIP) | (1 << VCPU_REGS_RSP));
-#ifdef XXX
- idx = srcu_read_lock(&vcpu->kvm->srcu);
-#else
- XXX_KVM_SYNC_PROBE;
-#endif
-
- if (!init_rmode(vmx->vcpu.kvm)) {
- ret = -ENOMEM;
- goto out;
- }
-
- vmx->rmode.vm86_active = 0;
- vmx->soft_vnmi_blocked = 0;
-
- vmx->vcpu.arch.regs[VCPU_REGS_RDX] = get_rdx_init_val();
- kvm_set_cr8(&vmx->vcpu, 0);
- msr = 0xfee00000 | MSR_IA32_APICBASE_ENABLE;
- if (kvm_vcpu_is_bsp(&vmx->vcpu))
- msr |= MSR_IA32_APICBASE_BSP;
-
- kvm_set_apic_base(&vmx->vcpu, msr);
-
- fx_init(&vmx->vcpu);
-
- seg_setup(VCPU_SREG_CS);
- /*
- * GUEST_CS_BASE should really be 0xffff0000, but VT vm86 mode
- * insists on having GUEST_CS_BASE == GUEST_CS_SELECTOR << 4. Sigh.
- */
- if (kvm_vcpu_is_bsp(&vmx->vcpu)) {
- vmcs_write16(GUEST_CS_SELECTOR, 0xf000);
-#ifndef XXX
- vmcs_writel(GUEST_CS_BASE, 0x000f0000);
-#else
- vmcs_writel(GUEST_CS_BASE, 0xffff0000);
-#endif
- } else {
- vmcs_write16(GUEST_CS_SELECTOR,
- vmx->vcpu.arch.sipi_vector << 8);
- vmcs_writel(GUEST_CS_BASE, vmx->vcpu.arch.sipi_vector << 12);
- }
-
- seg_setup(VCPU_SREG_DS);
- seg_setup(VCPU_SREG_ES);
- seg_setup(VCPU_SREG_FS);
- seg_setup(VCPU_SREG_GS);
- seg_setup(VCPU_SREG_SS);
-
- vmcs_write16(GUEST_TR_SELECTOR, 0);
- vmcs_writel(GUEST_TR_BASE, 0);
- vmcs_write32(GUEST_TR_LIMIT, 0xffff);
- vmcs_write32(GUEST_TR_AR_BYTES, 0x008b);
-
- vmcs_write16(GUEST_LDTR_SELECTOR, 0);
- vmcs_writel(GUEST_LDTR_BASE, 0);
- vmcs_write32(GUEST_LDTR_LIMIT, 0xffff);
- vmcs_write32(GUEST_LDTR_AR_BYTES, 0x00082);
-
- vmcs_write32(GUEST_SYSENTER_CS, 0);
- vmcs_writel(GUEST_SYSENTER_ESP, 0);
- vmcs_writel(GUEST_SYSENTER_EIP, 0);
-
- vmcs_writel(GUEST_RFLAGS, 0x02);
-
- if (kvm_vcpu_is_bsp(&vmx->vcpu))
- kvm_rip_write(vcpu, 0xfff0);
- else
- kvm_rip_write(vcpu, 0);
-
- kvm_register_write(vcpu, VCPU_REGS_RSP, 0);
-
- vmcs_writel(GUEST_DR7, 0x400);
-
- vmcs_writel(GUEST_GDTR_BASE, 0);
- vmcs_write32(GUEST_GDTR_LIMIT, 0xffff);
-
- vmcs_writel(GUEST_IDTR_BASE, 0);
- vmcs_write32(GUEST_IDTR_LIMIT, 0xffff);
-
- vmcs_write32(GUEST_ACTIVITY_STATE, 0);
- vmcs_write32(GUEST_INTERRUPTIBILITY_INFO, 0);
- vmcs_write32(GUEST_PENDING_DBG_EXCEPTIONS, 0);
-
- /* Special registers */
- vmcs_write64(GUEST_IA32_DEBUGCTL, 0);
-
- setup_msrs(vmx);
-
- vmcs_write32(VM_ENTRY_INTR_INFO_FIELD, 0); /* 22.2.1 */
-
- if (cpu_has_vmx_tpr_shadow()) {
- vmcs_write64(VIRTUAL_APIC_PAGE_ADDR, 0);
- if (vm_need_tpr_shadow(vmx->vcpu.kvm)) {
- ptp = page_numtopp_nolock(hat_getpfnum(kas.a_hat,
- vmx->vcpu.arch.apic->regs));
- vmcs_write64(VIRTUAL_APIC_PAGE_ADDR, page_to_phys(ptp));
- }
-
- vmcs_write32(TPR_THRESHOLD, 0);
- }
-
-
- if (vm_need_virtualize_apic_accesses(vmx->vcpu.kvm)) {
- vmcs_write64(APIC_ACCESS_ADDR,
- page_to_phys(vmx->vcpu.kvm->arch.apic_access_page));
- }
-
- if (vmx->vpid != 0)
- vmcs_write16(VIRTUAL_PROCESSOR_ID, vmx->vpid);
-
- vmx->vcpu.arch.cr0 = X86_CR0_NW | X86_CR0_CD | X86_CR0_ET;
- vmx_set_cr0(&vmx->vcpu, kvm_read_cr0(vcpu)); /* enter rmode */
- vmx_set_cr4(&vmx->vcpu, 0);
- vmx_set_efer(&vmx->vcpu, 0);
-
- vmx_fpu_activate(&vmx->vcpu);
- update_exception_bitmap(&vmx->vcpu);
- vpid_sync_vcpu_all(vmx);
-
- ret = 0;
-
- /* HACK: Don't enable emulation on guest boot/reset */
- vmx->emulation_required = 0;
-
-out:
-#ifdef XXX
- srcu_read_unlock(&vcpu->kvm->srcu, idx);
-#else
- XXX_KVM_SYNC_PROBE;
-#endif
- return (ret);
-}
int
kvm_arch_vcpu_reset(struct kvm_vcpu *vcpu)
@@ -1849,8 +1020,7 @@ kvm_arch_vcpu_reset(struct kvm_vcpu *vcpu)
vcpu->arch.dr6 = DR6_FIXED_1;
vcpu->arch.dr7 = DR7_FIXED_1;
- /* XXX: return kvm_x86_ops->vcpu_reset(vcpu); */
- return (vmx_vcpu_reset(vcpu));
+ return (kvm_x86_ops->vcpu_reset(vcpu));
}
extern void vcpu_load(struct kvm_vcpu *vcpu);
@@ -2883,12 +2053,6 @@ static void
nonpaging_invlpg(struct kvm_vcpu *vcpu, gva_t gva)
{}
-int
-get_ept_level(void)
-{
- return (VMX_EPT_DEFAULT_GAW + 1);
-}
-
static gpa_t
nonpaging_gva_to_gpa(struct kvm_vcpu *vcpu, gva_t vaddr,
uint32_t access, uint32_t *error)
diff --git a/kvm_x86host.h b/kvm_x86host.h
index 6640c30..ef0b3a3 100644
--- a/kvm_x86host.h
+++ b/kvm_x86host.h
@@ -989,15 +989,6 @@ enum {
/*asmlinkage*/ void kvm_handle_fault_on_reboot(void);
-#define __kvm_handle_fault_on_reboot(insn) \
- "666: " insn "\n\t" \
- ".pushsection .fixup, \"ax\" \n" \
- "667: \n\t" \
- __ASM_SIZE(push) " $666b \n\t" \
- ".popsection \n\t" \
- ".pushsection __ex_table, \"a\" \n\t" \
- _ASM_PTR " 666b, 667b \n\t" \
- ".popsection \n\t"
#define KVM_ARCH_WANT_MMU_NOTIFIER