mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-23 20:53:53 +08:00
68cda40d9f
Use an enum for passing around the failure code for a failed VM-Enter
that results in VM-Exit to provide a level of indirection from the final
resting place of the failure code, vmcs.EXIT_QUALIFICATION. The exit
qualification field is an unsigned long, e.g. passing around
'u32 exit_qual' throws up red flags as it suggests KVM may be dropping
bits when reporting errors to L1. This is a red herring because the
only defined failure codes are 0, 2, 3, and 4, i.e. don't come remotely
close to overflowing a u32.
Setting vmcs.EXIT_QUALIFICATION on entry failure is further complicated
by the MSR load list, which returns the (1-based) entry that failed, and
the number of MSRs to load is a 32-bit VMCS field. At first blush, it
would appear that overflowing a u32 is possible, but the number of MSRs
that can be loaded is hardcapped at 4096 (limited by MSR_IA32_VMX_MISC).
In other words, there are two completely disparate types of data that
eventually get stuffed into vmcs.EXIT_QUALIFICATION, neither of which is
an 'unsigned long' in nature. This was presumably the reasoning for
switching to 'u32' when the related code was refactored in commit
ca0bde28f2
("kvm: nVMX: Split VMCS checks from nested_vmx_run()").
Using an enum for the failure code addresses the technically-possible-
but-will-never-happen scenario where Intel defines a failure code that
doesn't fit in a 32-bit integer. The enum variables and values will
either be automatically sized (gcc 5.4 behavior) or be subjected to some
combination of truncation. The former case will simply work, while the
latter will trigger a compile-time warning unless the compiler is being
particularly unhelpful.
Separating the failure code from the failed MSR entry allows for
disassociating both from vmcs.EXIT_QUALIFICATION, which avoids the
conundrum where KVM has to choose between 'u32 exit_qual' and tracking
values as 'unsigned long' that have no business being tracked as such.
To cement the split, set vmcs12->exit_qualification directly from the
entry error code or failed MSR index instead of bouncing through a local
variable.
Opportunistically rename the variables in load_vmcs12_host_state() and
vmx_set_nested_state() to call out that they're ignored, set exit_reason
on demand on nested VM-Enter failure, and add a comment in
nested_vmx_load_msr() to call out that returning 'i + 1' can't wrap.
No functional change intended.
Reported-by: Vitaly Kuznetsov <vkuznets@redhat.com>
Cc: Jim Mattson <jmattson@google.com>
Signed-off-by: Sean Christopherson <sean.j.christopherson@intel.com>
Message-Id: <20200511220529.11402-1-sean.j.christopherson@intel.com>
Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
612 lines
25 KiB
C
612 lines
25 KiB
C
/* SPDX-License-Identifier: GPL-2.0-only */
|
|
/*
|
|
* vmx.h: VMX Architecture related definitions
|
|
* Copyright (c) 2004, Intel Corporation.
|
|
*
|
|
* A few random additions are:
|
|
* Copyright (C) 2006 Qumranet
|
|
* Avi Kivity <avi@qumranet.com>
|
|
* Yaniv Kamay <yaniv@qumranet.com>
|
|
*/
|
|
#ifndef VMX_H
|
|
#define VMX_H
|
|
|
|
|
|
#include <linux/bitops.h>
|
|
#include <linux/types.h>
|
|
#include <uapi/asm/vmx.h>
|
|
#include <asm/vmxfeatures.h>
|
|
|
|
#define VMCS_CONTROL_BIT(x) BIT(VMX_FEATURE_##x & 0x1f)
|
|
|
|
/*
|
|
* Definitions of Primary Processor-Based VM-Execution Controls.
|
|
*/
|
|
#define CPU_BASED_INTR_WINDOW_EXITING VMCS_CONTROL_BIT(INTR_WINDOW_EXITING)
|
|
#define CPU_BASED_USE_TSC_OFFSETTING VMCS_CONTROL_BIT(USE_TSC_OFFSETTING)
|
|
#define CPU_BASED_HLT_EXITING VMCS_CONTROL_BIT(HLT_EXITING)
|
|
#define CPU_BASED_INVLPG_EXITING VMCS_CONTROL_BIT(INVLPG_EXITING)
|
|
#define CPU_BASED_MWAIT_EXITING VMCS_CONTROL_BIT(MWAIT_EXITING)
|
|
#define CPU_BASED_RDPMC_EXITING VMCS_CONTROL_BIT(RDPMC_EXITING)
|
|
#define CPU_BASED_RDTSC_EXITING VMCS_CONTROL_BIT(RDTSC_EXITING)
|
|
#define CPU_BASED_CR3_LOAD_EXITING VMCS_CONTROL_BIT(CR3_LOAD_EXITING)
|
|
#define CPU_BASED_CR3_STORE_EXITING VMCS_CONTROL_BIT(CR3_STORE_EXITING)
|
|
#define CPU_BASED_CR8_LOAD_EXITING VMCS_CONTROL_BIT(CR8_LOAD_EXITING)
|
|
#define CPU_BASED_CR8_STORE_EXITING VMCS_CONTROL_BIT(CR8_STORE_EXITING)
|
|
#define CPU_BASED_TPR_SHADOW VMCS_CONTROL_BIT(VIRTUAL_TPR)
|
|
#define CPU_BASED_NMI_WINDOW_EXITING VMCS_CONTROL_BIT(NMI_WINDOW_EXITING)
|
|
#define CPU_BASED_MOV_DR_EXITING VMCS_CONTROL_BIT(MOV_DR_EXITING)
|
|
#define CPU_BASED_UNCOND_IO_EXITING VMCS_CONTROL_BIT(UNCOND_IO_EXITING)
|
|
#define CPU_BASED_USE_IO_BITMAPS VMCS_CONTROL_BIT(USE_IO_BITMAPS)
|
|
#define CPU_BASED_MONITOR_TRAP_FLAG VMCS_CONTROL_BIT(MONITOR_TRAP_FLAG)
|
|
#define CPU_BASED_USE_MSR_BITMAPS VMCS_CONTROL_BIT(USE_MSR_BITMAPS)
|
|
#define CPU_BASED_MONITOR_EXITING VMCS_CONTROL_BIT(MONITOR_EXITING)
|
|
#define CPU_BASED_PAUSE_EXITING VMCS_CONTROL_BIT(PAUSE_EXITING)
|
|
#define CPU_BASED_ACTIVATE_SECONDARY_CONTROLS VMCS_CONTROL_BIT(SEC_CONTROLS)
|
|
|
|
#define CPU_BASED_ALWAYSON_WITHOUT_TRUE_MSR 0x0401e172
|
|
|
|
/*
|
|
* Definitions of Secondary Processor-Based VM-Execution Controls.
|
|
*/
|
|
#define SECONDARY_EXEC_VIRTUALIZE_APIC_ACCESSES VMCS_CONTROL_BIT(VIRT_APIC_ACCESSES)
|
|
#define SECONDARY_EXEC_ENABLE_EPT VMCS_CONTROL_BIT(EPT)
|
|
#define SECONDARY_EXEC_DESC VMCS_CONTROL_BIT(DESC_EXITING)
|
|
#define SECONDARY_EXEC_RDTSCP VMCS_CONTROL_BIT(RDTSCP)
|
|
#define SECONDARY_EXEC_VIRTUALIZE_X2APIC_MODE VMCS_CONTROL_BIT(VIRTUAL_X2APIC)
|
|
#define SECONDARY_EXEC_ENABLE_VPID VMCS_CONTROL_BIT(VPID)
|
|
#define SECONDARY_EXEC_WBINVD_EXITING VMCS_CONTROL_BIT(WBINVD_EXITING)
|
|
#define SECONDARY_EXEC_UNRESTRICTED_GUEST VMCS_CONTROL_BIT(UNRESTRICTED_GUEST)
|
|
#define SECONDARY_EXEC_APIC_REGISTER_VIRT VMCS_CONTROL_BIT(APIC_REGISTER_VIRT)
|
|
#define SECONDARY_EXEC_VIRTUAL_INTR_DELIVERY VMCS_CONTROL_BIT(VIRT_INTR_DELIVERY)
|
|
#define SECONDARY_EXEC_PAUSE_LOOP_EXITING VMCS_CONTROL_BIT(PAUSE_LOOP_EXITING)
|
|
#define SECONDARY_EXEC_RDRAND_EXITING VMCS_CONTROL_BIT(RDRAND_EXITING)
|
|
#define SECONDARY_EXEC_ENABLE_INVPCID VMCS_CONTROL_BIT(INVPCID)
|
|
#define SECONDARY_EXEC_ENABLE_VMFUNC VMCS_CONTROL_BIT(VMFUNC)
|
|
#define SECONDARY_EXEC_SHADOW_VMCS VMCS_CONTROL_BIT(SHADOW_VMCS)
|
|
#define SECONDARY_EXEC_ENCLS_EXITING VMCS_CONTROL_BIT(ENCLS_EXITING)
|
|
#define SECONDARY_EXEC_RDSEED_EXITING VMCS_CONTROL_BIT(RDSEED_EXITING)
|
|
#define SECONDARY_EXEC_ENABLE_PML VMCS_CONTROL_BIT(PAGE_MOD_LOGGING)
|
|
#define SECONDARY_EXEC_PT_CONCEAL_VMX VMCS_CONTROL_BIT(PT_CONCEAL_VMX)
|
|
#define SECONDARY_EXEC_XSAVES VMCS_CONTROL_BIT(XSAVES)
|
|
#define SECONDARY_EXEC_MODE_BASED_EPT_EXEC VMCS_CONTROL_BIT(MODE_BASED_EPT_EXEC)
|
|
#define SECONDARY_EXEC_PT_USE_GPA VMCS_CONTROL_BIT(PT_USE_GPA)
|
|
#define SECONDARY_EXEC_TSC_SCALING VMCS_CONTROL_BIT(TSC_SCALING)
|
|
#define SECONDARY_EXEC_ENABLE_USR_WAIT_PAUSE VMCS_CONTROL_BIT(USR_WAIT_PAUSE)
|
|
|
|
#define PIN_BASED_EXT_INTR_MASK VMCS_CONTROL_BIT(INTR_EXITING)
|
|
#define PIN_BASED_NMI_EXITING VMCS_CONTROL_BIT(NMI_EXITING)
|
|
#define PIN_BASED_VIRTUAL_NMIS VMCS_CONTROL_BIT(VIRTUAL_NMIS)
|
|
#define PIN_BASED_VMX_PREEMPTION_TIMER VMCS_CONTROL_BIT(PREEMPTION_TIMER)
|
|
#define PIN_BASED_POSTED_INTR VMCS_CONTROL_BIT(POSTED_INTR)
|
|
|
|
#define PIN_BASED_ALWAYSON_WITHOUT_TRUE_MSR 0x00000016
|
|
|
|
#define VM_EXIT_SAVE_DEBUG_CONTROLS 0x00000004
|
|
#define VM_EXIT_HOST_ADDR_SPACE_SIZE 0x00000200
|
|
#define VM_EXIT_LOAD_IA32_PERF_GLOBAL_CTRL 0x00001000
|
|
#define VM_EXIT_ACK_INTR_ON_EXIT 0x00008000
|
|
#define VM_EXIT_SAVE_IA32_PAT 0x00040000
|
|
#define VM_EXIT_LOAD_IA32_PAT 0x00080000
|
|
#define VM_EXIT_SAVE_IA32_EFER 0x00100000
|
|
#define VM_EXIT_LOAD_IA32_EFER 0x00200000
|
|
#define VM_EXIT_SAVE_VMX_PREEMPTION_TIMER 0x00400000
|
|
#define VM_EXIT_CLEAR_BNDCFGS 0x00800000
|
|
#define VM_EXIT_PT_CONCEAL_PIP 0x01000000
|
|
#define VM_EXIT_CLEAR_IA32_RTIT_CTL 0x02000000
|
|
|
|
#define VM_EXIT_ALWAYSON_WITHOUT_TRUE_MSR 0x00036dff
|
|
|
|
#define VM_ENTRY_LOAD_DEBUG_CONTROLS 0x00000004
|
|
#define VM_ENTRY_IA32E_MODE 0x00000200
|
|
#define VM_ENTRY_SMM 0x00000400
|
|
#define VM_ENTRY_DEACT_DUAL_MONITOR 0x00000800
|
|
#define VM_ENTRY_LOAD_IA32_PERF_GLOBAL_CTRL 0x00002000
|
|
#define VM_ENTRY_LOAD_IA32_PAT 0x00004000
|
|
#define VM_ENTRY_LOAD_IA32_EFER 0x00008000
|
|
#define VM_ENTRY_LOAD_BNDCFGS 0x00010000
|
|
#define VM_ENTRY_PT_CONCEAL_PIP 0x00020000
|
|
#define VM_ENTRY_LOAD_IA32_RTIT_CTL 0x00040000
|
|
|
|
#define VM_ENTRY_ALWAYSON_WITHOUT_TRUE_MSR 0x000011ff
|
|
|
|
#define VMX_MISC_PREEMPTION_TIMER_RATE_MASK 0x0000001f
|
|
#define VMX_MISC_SAVE_EFER_LMA 0x00000020
|
|
#define VMX_MISC_ACTIVITY_HLT 0x00000040
|
|
#define VMX_MISC_ZERO_LEN_INS 0x40000000
|
|
#define VMX_MISC_MSR_LIST_MULTIPLIER 512
|
|
|
|
/* VMFUNC functions */
|
|
#define VMFUNC_CONTROL_BIT(x) BIT((VMX_FEATURE_##x & 0x1f) - 28)
|
|
|
|
#define VMX_VMFUNC_EPTP_SWITCHING VMFUNC_CONTROL_BIT(EPTP_SWITCHING)
|
|
#define VMFUNC_EPTP_ENTRIES 512
|
|
|
|
static inline u32 vmx_basic_vmcs_revision_id(u64 vmx_basic)
|
|
{
|
|
return vmx_basic & GENMASK_ULL(30, 0);
|
|
}
|
|
|
|
static inline u32 vmx_basic_vmcs_size(u64 vmx_basic)
|
|
{
|
|
return (vmx_basic & GENMASK_ULL(44, 32)) >> 32;
|
|
}
|
|
|
|
static inline int vmx_misc_preemption_timer_rate(u64 vmx_misc)
|
|
{
|
|
return vmx_misc & VMX_MISC_PREEMPTION_TIMER_RATE_MASK;
|
|
}
|
|
|
|
static inline int vmx_misc_cr3_count(u64 vmx_misc)
|
|
{
|
|
return (vmx_misc & GENMASK_ULL(24, 16)) >> 16;
|
|
}
|
|
|
|
static inline int vmx_misc_max_msr(u64 vmx_misc)
|
|
{
|
|
return (vmx_misc & GENMASK_ULL(27, 25)) >> 25;
|
|
}
|
|
|
|
static inline int vmx_misc_mseg_revid(u64 vmx_misc)
|
|
{
|
|
return (vmx_misc & GENMASK_ULL(63, 32)) >> 32;
|
|
}
|
|
|
|
/* VMCS Encodings */
|
|
enum vmcs_field {
|
|
VIRTUAL_PROCESSOR_ID = 0x00000000,
|
|
POSTED_INTR_NV = 0x00000002,
|
|
GUEST_ES_SELECTOR = 0x00000800,
|
|
GUEST_CS_SELECTOR = 0x00000802,
|
|
GUEST_SS_SELECTOR = 0x00000804,
|
|
GUEST_DS_SELECTOR = 0x00000806,
|
|
GUEST_FS_SELECTOR = 0x00000808,
|
|
GUEST_GS_SELECTOR = 0x0000080a,
|
|
GUEST_LDTR_SELECTOR = 0x0000080c,
|
|
GUEST_TR_SELECTOR = 0x0000080e,
|
|
GUEST_INTR_STATUS = 0x00000810,
|
|
GUEST_PML_INDEX = 0x00000812,
|
|
HOST_ES_SELECTOR = 0x00000c00,
|
|
HOST_CS_SELECTOR = 0x00000c02,
|
|
HOST_SS_SELECTOR = 0x00000c04,
|
|
HOST_DS_SELECTOR = 0x00000c06,
|
|
HOST_FS_SELECTOR = 0x00000c08,
|
|
HOST_GS_SELECTOR = 0x00000c0a,
|
|
HOST_TR_SELECTOR = 0x00000c0c,
|
|
IO_BITMAP_A = 0x00002000,
|
|
IO_BITMAP_A_HIGH = 0x00002001,
|
|
IO_BITMAP_B = 0x00002002,
|
|
IO_BITMAP_B_HIGH = 0x00002003,
|
|
MSR_BITMAP = 0x00002004,
|
|
MSR_BITMAP_HIGH = 0x00002005,
|
|
VM_EXIT_MSR_STORE_ADDR = 0x00002006,
|
|
VM_EXIT_MSR_STORE_ADDR_HIGH = 0x00002007,
|
|
VM_EXIT_MSR_LOAD_ADDR = 0x00002008,
|
|
VM_EXIT_MSR_LOAD_ADDR_HIGH = 0x00002009,
|
|
VM_ENTRY_MSR_LOAD_ADDR = 0x0000200a,
|
|
VM_ENTRY_MSR_LOAD_ADDR_HIGH = 0x0000200b,
|
|
PML_ADDRESS = 0x0000200e,
|
|
PML_ADDRESS_HIGH = 0x0000200f,
|
|
TSC_OFFSET = 0x00002010,
|
|
TSC_OFFSET_HIGH = 0x00002011,
|
|
VIRTUAL_APIC_PAGE_ADDR = 0x00002012,
|
|
VIRTUAL_APIC_PAGE_ADDR_HIGH = 0x00002013,
|
|
APIC_ACCESS_ADDR = 0x00002014,
|
|
APIC_ACCESS_ADDR_HIGH = 0x00002015,
|
|
POSTED_INTR_DESC_ADDR = 0x00002016,
|
|
POSTED_INTR_DESC_ADDR_HIGH = 0x00002017,
|
|
VM_FUNCTION_CONTROL = 0x00002018,
|
|
VM_FUNCTION_CONTROL_HIGH = 0x00002019,
|
|
EPT_POINTER = 0x0000201a,
|
|
EPT_POINTER_HIGH = 0x0000201b,
|
|
EOI_EXIT_BITMAP0 = 0x0000201c,
|
|
EOI_EXIT_BITMAP0_HIGH = 0x0000201d,
|
|
EOI_EXIT_BITMAP1 = 0x0000201e,
|
|
EOI_EXIT_BITMAP1_HIGH = 0x0000201f,
|
|
EOI_EXIT_BITMAP2 = 0x00002020,
|
|
EOI_EXIT_BITMAP2_HIGH = 0x00002021,
|
|
EOI_EXIT_BITMAP3 = 0x00002022,
|
|
EOI_EXIT_BITMAP3_HIGH = 0x00002023,
|
|
EPTP_LIST_ADDRESS = 0x00002024,
|
|
EPTP_LIST_ADDRESS_HIGH = 0x00002025,
|
|
VMREAD_BITMAP = 0x00002026,
|
|
VMREAD_BITMAP_HIGH = 0x00002027,
|
|
VMWRITE_BITMAP = 0x00002028,
|
|
VMWRITE_BITMAP_HIGH = 0x00002029,
|
|
XSS_EXIT_BITMAP = 0x0000202C,
|
|
XSS_EXIT_BITMAP_HIGH = 0x0000202D,
|
|
ENCLS_EXITING_BITMAP = 0x0000202E,
|
|
ENCLS_EXITING_BITMAP_HIGH = 0x0000202F,
|
|
TSC_MULTIPLIER = 0x00002032,
|
|
TSC_MULTIPLIER_HIGH = 0x00002033,
|
|
GUEST_PHYSICAL_ADDRESS = 0x00002400,
|
|
GUEST_PHYSICAL_ADDRESS_HIGH = 0x00002401,
|
|
VMCS_LINK_POINTER = 0x00002800,
|
|
VMCS_LINK_POINTER_HIGH = 0x00002801,
|
|
GUEST_IA32_DEBUGCTL = 0x00002802,
|
|
GUEST_IA32_DEBUGCTL_HIGH = 0x00002803,
|
|
GUEST_IA32_PAT = 0x00002804,
|
|
GUEST_IA32_PAT_HIGH = 0x00002805,
|
|
GUEST_IA32_EFER = 0x00002806,
|
|
GUEST_IA32_EFER_HIGH = 0x00002807,
|
|
GUEST_IA32_PERF_GLOBAL_CTRL = 0x00002808,
|
|
GUEST_IA32_PERF_GLOBAL_CTRL_HIGH= 0x00002809,
|
|
GUEST_PDPTR0 = 0x0000280a,
|
|
GUEST_PDPTR0_HIGH = 0x0000280b,
|
|
GUEST_PDPTR1 = 0x0000280c,
|
|
GUEST_PDPTR1_HIGH = 0x0000280d,
|
|
GUEST_PDPTR2 = 0x0000280e,
|
|
GUEST_PDPTR2_HIGH = 0x0000280f,
|
|
GUEST_PDPTR3 = 0x00002810,
|
|
GUEST_PDPTR3_HIGH = 0x00002811,
|
|
GUEST_BNDCFGS = 0x00002812,
|
|
GUEST_BNDCFGS_HIGH = 0x00002813,
|
|
GUEST_IA32_RTIT_CTL = 0x00002814,
|
|
GUEST_IA32_RTIT_CTL_HIGH = 0x00002815,
|
|
HOST_IA32_PAT = 0x00002c00,
|
|
HOST_IA32_PAT_HIGH = 0x00002c01,
|
|
HOST_IA32_EFER = 0x00002c02,
|
|
HOST_IA32_EFER_HIGH = 0x00002c03,
|
|
HOST_IA32_PERF_GLOBAL_CTRL = 0x00002c04,
|
|
HOST_IA32_PERF_GLOBAL_CTRL_HIGH = 0x00002c05,
|
|
PIN_BASED_VM_EXEC_CONTROL = 0x00004000,
|
|
CPU_BASED_VM_EXEC_CONTROL = 0x00004002,
|
|
EXCEPTION_BITMAP = 0x00004004,
|
|
PAGE_FAULT_ERROR_CODE_MASK = 0x00004006,
|
|
PAGE_FAULT_ERROR_CODE_MATCH = 0x00004008,
|
|
CR3_TARGET_COUNT = 0x0000400a,
|
|
VM_EXIT_CONTROLS = 0x0000400c,
|
|
VM_EXIT_MSR_STORE_COUNT = 0x0000400e,
|
|
VM_EXIT_MSR_LOAD_COUNT = 0x00004010,
|
|
VM_ENTRY_CONTROLS = 0x00004012,
|
|
VM_ENTRY_MSR_LOAD_COUNT = 0x00004014,
|
|
VM_ENTRY_INTR_INFO_FIELD = 0x00004016,
|
|
VM_ENTRY_EXCEPTION_ERROR_CODE = 0x00004018,
|
|
VM_ENTRY_INSTRUCTION_LEN = 0x0000401a,
|
|
TPR_THRESHOLD = 0x0000401c,
|
|
SECONDARY_VM_EXEC_CONTROL = 0x0000401e,
|
|
PLE_GAP = 0x00004020,
|
|
PLE_WINDOW = 0x00004022,
|
|
VM_INSTRUCTION_ERROR = 0x00004400,
|
|
VM_EXIT_REASON = 0x00004402,
|
|
VM_EXIT_INTR_INFO = 0x00004404,
|
|
VM_EXIT_INTR_ERROR_CODE = 0x00004406,
|
|
IDT_VECTORING_INFO_FIELD = 0x00004408,
|
|
IDT_VECTORING_ERROR_CODE = 0x0000440a,
|
|
VM_EXIT_INSTRUCTION_LEN = 0x0000440c,
|
|
VMX_INSTRUCTION_INFO = 0x0000440e,
|
|
GUEST_ES_LIMIT = 0x00004800,
|
|
GUEST_CS_LIMIT = 0x00004802,
|
|
GUEST_SS_LIMIT = 0x00004804,
|
|
GUEST_DS_LIMIT = 0x00004806,
|
|
GUEST_FS_LIMIT = 0x00004808,
|
|
GUEST_GS_LIMIT = 0x0000480a,
|
|
GUEST_LDTR_LIMIT = 0x0000480c,
|
|
GUEST_TR_LIMIT = 0x0000480e,
|
|
GUEST_GDTR_LIMIT = 0x00004810,
|
|
GUEST_IDTR_LIMIT = 0x00004812,
|
|
GUEST_ES_AR_BYTES = 0x00004814,
|
|
GUEST_CS_AR_BYTES = 0x00004816,
|
|
GUEST_SS_AR_BYTES = 0x00004818,
|
|
GUEST_DS_AR_BYTES = 0x0000481a,
|
|
GUEST_FS_AR_BYTES = 0x0000481c,
|
|
GUEST_GS_AR_BYTES = 0x0000481e,
|
|
GUEST_LDTR_AR_BYTES = 0x00004820,
|
|
GUEST_TR_AR_BYTES = 0x00004822,
|
|
GUEST_INTERRUPTIBILITY_INFO = 0x00004824,
|
|
GUEST_ACTIVITY_STATE = 0X00004826,
|
|
GUEST_SYSENTER_CS = 0x0000482A,
|
|
VMX_PREEMPTION_TIMER_VALUE = 0x0000482E,
|
|
HOST_IA32_SYSENTER_CS = 0x00004c00,
|
|
CR0_GUEST_HOST_MASK = 0x00006000,
|
|
CR4_GUEST_HOST_MASK = 0x00006002,
|
|
CR0_READ_SHADOW = 0x00006004,
|
|
CR4_READ_SHADOW = 0x00006006,
|
|
CR3_TARGET_VALUE0 = 0x00006008,
|
|
CR3_TARGET_VALUE1 = 0x0000600a,
|
|
CR3_TARGET_VALUE2 = 0x0000600c,
|
|
CR3_TARGET_VALUE3 = 0x0000600e,
|
|
EXIT_QUALIFICATION = 0x00006400,
|
|
GUEST_LINEAR_ADDRESS = 0x0000640a,
|
|
GUEST_CR0 = 0x00006800,
|
|
GUEST_CR3 = 0x00006802,
|
|
GUEST_CR4 = 0x00006804,
|
|
GUEST_ES_BASE = 0x00006806,
|
|
GUEST_CS_BASE = 0x00006808,
|
|
GUEST_SS_BASE = 0x0000680a,
|
|
GUEST_DS_BASE = 0x0000680c,
|
|
GUEST_FS_BASE = 0x0000680e,
|
|
GUEST_GS_BASE = 0x00006810,
|
|
GUEST_LDTR_BASE = 0x00006812,
|
|
GUEST_TR_BASE = 0x00006814,
|
|
GUEST_GDTR_BASE = 0x00006816,
|
|
GUEST_IDTR_BASE = 0x00006818,
|
|
GUEST_DR7 = 0x0000681a,
|
|
GUEST_RSP = 0x0000681c,
|
|
GUEST_RIP = 0x0000681e,
|
|
GUEST_RFLAGS = 0x00006820,
|
|
GUEST_PENDING_DBG_EXCEPTIONS = 0x00006822,
|
|
GUEST_SYSENTER_ESP = 0x00006824,
|
|
GUEST_SYSENTER_EIP = 0x00006826,
|
|
HOST_CR0 = 0x00006c00,
|
|
HOST_CR3 = 0x00006c02,
|
|
HOST_CR4 = 0x00006c04,
|
|
HOST_FS_BASE = 0x00006c06,
|
|
HOST_GS_BASE = 0x00006c08,
|
|
HOST_TR_BASE = 0x00006c0a,
|
|
HOST_GDTR_BASE = 0x00006c0c,
|
|
HOST_IDTR_BASE = 0x00006c0e,
|
|
HOST_IA32_SYSENTER_ESP = 0x00006c10,
|
|
HOST_IA32_SYSENTER_EIP = 0x00006c12,
|
|
HOST_RSP = 0x00006c14,
|
|
HOST_RIP = 0x00006c16,
|
|
};
|
|
|
|
/*
|
|
* Interruption-information format
|
|
*/
|
|
#define INTR_INFO_VECTOR_MASK 0xff /* 7:0 */
|
|
#define INTR_INFO_INTR_TYPE_MASK 0x700 /* 10:8 */
|
|
#define INTR_INFO_DELIVER_CODE_MASK 0x800 /* 11 */
|
|
#define INTR_INFO_UNBLOCK_NMI 0x1000 /* 12 */
|
|
#define INTR_INFO_VALID_MASK 0x80000000 /* 31 */
|
|
#define INTR_INFO_RESVD_BITS_MASK 0x7ffff000
|
|
|
|
#define VECTORING_INFO_VECTOR_MASK INTR_INFO_VECTOR_MASK
|
|
#define VECTORING_INFO_TYPE_MASK INTR_INFO_INTR_TYPE_MASK
|
|
#define VECTORING_INFO_DELIVER_CODE_MASK INTR_INFO_DELIVER_CODE_MASK
|
|
#define VECTORING_INFO_VALID_MASK INTR_INFO_VALID_MASK
|
|
|
|
#define INTR_TYPE_EXT_INTR (0 << 8) /* external interrupt */
|
|
#define INTR_TYPE_RESERVED (1 << 8) /* reserved */
|
|
#define INTR_TYPE_NMI_INTR (2 << 8) /* NMI */
|
|
#define INTR_TYPE_HARD_EXCEPTION (3 << 8) /* processor exception */
|
|
#define INTR_TYPE_SOFT_INTR (4 << 8) /* software interrupt */
|
|
#define INTR_TYPE_PRIV_SW_EXCEPTION (5 << 8) /* ICE breakpoint - undocumented */
|
|
#define INTR_TYPE_SOFT_EXCEPTION (6 << 8) /* software exception */
|
|
#define INTR_TYPE_OTHER_EVENT (7 << 8) /* other event */
|
|
|
|
/* GUEST_INTERRUPTIBILITY_INFO flags. */
|
|
#define GUEST_INTR_STATE_STI 0x00000001
|
|
#define GUEST_INTR_STATE_MOV_SS 0x00000002
|
|
#define GUEST_INTR_STATE_SMI 0x00000004
|
|
#define GUEST_INTR_STATE_NMI 0x00000008
|
|
|
|
/* GUEST_ACTIVITY_STATE flags */
|
|
#define GUEST_ACTIVITY_ACTIVE 0
|
|
#define GUEST_ACTIVITY_HLT 1
|
|
#define GUEST_ACTIVITY_SHUTDOWN 2
|
|
#define GUEST_ACTIVITY_WAIT_SIPI 3
|
|
|
|
/*
|
|
* Exit Qualifications for MOV for Control Register Access
|
|
*/
|
|
#define CONTROL_REG_ACCESS_NUM 0x7 /* 2:0, number of control reg.*/
|
|
#define CONTROL_REG_ACCESS_TYPE 0x30 /* 5:4, access type */
|
|
#define CONTROL_REG_ACCESS_REG 0xf00 /* 10:8, general purpose reg. */
|
|
#define LMSW_SOURCE_DATA_SHIFT 16
|
|
#define LMSW_SOURCE_DATA (0xFFFF << LMSW_SOURCE_DATA_SHIFT) /* 16:31 lmsw source */
|
|
#define REG_EAX (0 << 8)
|
|
#define REG_ECX (1 << 8)
|
|
#define REG_EDX (2 << 8)
|
|
#define REG_EBX (3 << 8)
|
|
#define REG_ESP (4 << 8)
|
|
#define REG_EBP (5 << 8)
|
|
#define REG_ESI (6 << 8)
|
|
#define REG_EDI (7 << 8)
|
|
#define REG_R8 (8 << 8)
|
|
#define REG_R9 (9 << 8)
|
|
#define REG_R10 (10 << 8)
|
|
#define REG_R11 (11 << 8)
|
|
#define REG_R12 (12 << 8)
|
|
#define REG_R13 (13 << 8)
|
|
#define REG_R14 (14 << 8)
|
|
#define REG_R15 (15 << 8)
|
|
|
|
/*
|
|
* Exit Qualifications for MOV for Debug Register Access
|
|
*/
|
|
#define DEBUG_REG_ACCESS_NUM 0x7 /* 2:0, number of debug reg. */
|
|
#define DEBUG_REG_ACCESS_TYPE 0x10 /* 4, direction of access */
|
|
#define TYPE_MOV_TO_DR (0 << 4)
|
|
#define TYPE_MOV_FROM_DR (1 << 4)
|
|
#define DEBUG_REG_ACCESS_REG(eq) (((eq) >> 8) & 0xf) /* 11:8, general purpose reg. */
|
|
|
|
|
|
/*
|
|
* Exit Qualifications for APIC-Access
|
|
*/
|
|
#define APIC_ACCESS_OFFSET 0xfff /* 11:0, offset within the APIC page */
|
|
#define APIC_ACCESS_TYPE 0xf000 /* 15:12, access type */
|
|
#define TYPE_LINEAR_APIC_INST_READ (0 << 12)
|
|
#define TYPE_LINEAR_APIC_INST_WRITE (1 << 12)
|
|
#define TYPE_LINEAR_APIC_INST_FETCH (2 << 12)
|
|
#define TYPE_LINEAR_APIC_EVENT (3 << 12)
|
|
#define TYPE_PHYSICAL_APIC_EVENT (10 << 12)
|
|
#define TYPE_PHYSICAL_APIC_INST (15 << 12)
|
|
|
|
/* segment AR in VMCS -- these are different from what LAR reports */
|
|
#define VMX_SEGMENT_AR_L_MASK (1 << 13)
|
|
|
|
#define VMX_AR_TYPE_ACCESSES_MASK 1
|
|
#define VMX_AR_TYPE_READABLE_MASK (1 << 1)
|
|
#define VMX_AR_TYPE_WRITEABLE_MASK (1 << 2)
|
|
#define VMX_AR_TYPE_CODE_MASK (1 << 3)
|
|
#define VMX_AR_TYPE_MASK 0x0f
|
|
#define VMX_AR_TYPE_BUSY_64_TSS 11
|
|
#define VMX_AR_TYPE_BUSY_32_TSS 11
|
|
#define VMX_AR_TYPE_BUSY_16_TSS 3
|
|
#define VMX_AR_TYPE_LDT 2
|
|
|
|
#define VMX_AR_UNUSABLE_MASK (1 << 16)
|
|
#define VMX_AR_S_MASK (1 << 4)
|
|
#define VMX_AR_P_MASK (1 << 7)
|
|
#define VMX_AR_L_MASK (1 << 13)
|
|
#define VMX_AR_DB_MASK (1 << 14)
|
|
#define VMX_AR_G_MASK (1 << 15)
|
|
#define VMX_AR_DPL_SHIFT 5
|
|
#define VMX_AR_DPL(ar) (((ar) >> VMX_AR_DPL_SHIFT) & 3)
|
|
|
|
#define VMX_AR_RESERVD_MASK 0xfffe0f00
|
|
|
|
#define TSS_PRIVATE_MEMSLOT (KVM_USER_MEM_SLOTS + 0)
|
|
#define APIC_ACCESS_PAGE_PRIVATE_MEMSLOT (KVM_USER_MEM_SLOTS + 1)
|
|
#define IDENTITY_PAGETABLE_PRIVATE_MEMSLOT (KVM_USER_MEM_SLOTS + 2)
|
|
|
|
#define VMX_NR_VPIDS (1 << 16)
|
|
#define VMX_VPID_EXTENT_INDIVIDUAL_ADDR 0
|
|
#define VMX_VPID_EXTENT_SINGLE_CONTEXT 1
|
|
#define VMX_VPID_EXTENT_ALL_CONTEXT 2
|
|
#define VMX_VPID_EXTENT_SINGLE_NON_GLOBAL 3
|
|
|
|
#define VMX_EPT_EXTENT_CONTEXT 1
|
|
#define VMX_EPT_EXTENT_GLOBAL 2
|
|
#define VMX_EPT_EXTENT_SHIFT 24
|
|
|
|
#define VMX_EPT_EXECUTE_ONLY_BIT (1ull)
|
|
#define VMX_EPT_PAGE_WALK_4_BIT (1ull << 6)
|
|
#define VMX_EPT_PAGE_WALK_5_BIT (1ull << 7)
|
|
#define VMX_EPTP_UC_BIT (1ull << 8)
|
|
#define VMX_EPTP_WB_BIT (1ull << 14)
|
|
#define VMX_EPT_2MB_PAGE_BIT (1ull << 16)
|
|
#define VMX_EPT_1GB_PAGE_BIT (1ull << 17)
|
|
#define VMX_EPT_INVEPT_BIT (1ull << 20)
|
|
#define VMX_EPT_AD_BIT (1ull << 21)
|
|
#define VMX_EPT_EXTENT_CONTEXT_BIT (1ull << 25)
|
|
#define VMX_EPT_EXTENT_GLOBAL_BIT (1ull << 26)
|
|
|
|
#define VMX_VPID_INVVPID_BIT (1ull << 0) /* (32 - 32) */
|
|
#define VMX_VPID_EXTENT_INDIVIDUAL_ADDR_BIT (1ull << 8) /* (40 - 32) */
|
|
#define VMX_VPID_EXTENT_SINGLE_CONTEXT_BIT (1ull << 9) /* (41 - 32) */
|
|
#define VMX_VPID_EXTENT_GLOBAL_CONTEXT_BIT (1ull << 10) /* (42 - 32) */
|
|
#define VMX_VPID_EXTENT_SINGLE_NON_GLOBAL_BIT (1ull << 11) /* (43 - 32) */
|
|
|
|
#define VMX_EPT_MT_EPTE_SHIFT 3
|
|
#define VMX_EPTP_PWL_MASK 0x38ull
|
|
#define VMX_EPTP_PWL_4 0x18ull
|
|
#define VMX_EPTP_PWL_5 0x20ull
|
|
#define VMX_EPTP_AD_ENABLE_BIT (1ull << 6)
|
|
#define VMX_EPTP_MT_MASK 0x7ull
|
|
#define VMX_EPTP_MT_WB 0x6ull
|
|
#define VMX_EPTP_MT_UC 0x0ull
|
|
#define VMX_EPT_READABLE_MASK 0x1ull
|
|
#define VMX_EPT_WRITABLE_MASK 0x2ull
|
|
#define VMX_EPT_EXECUTABLE_MASK 0x4ull
|
|
#define VMX_EPT_IPAT_BIT (1ull << 6)
|
|
#define VMX_EPT_ACCESS_BIT (1ull << 8)
|
|
#define VMX_EPT_DIRTY_BIT (1ull << 9)
|
|
#define VMX_EPT_RWX_MASK (VMX_EPT_READABLE_MASK | \
|
|
VMX_EPT_WRITABLE_MASK | \
|
|
VMX_EPT_EXECUTABLE_MASK)
|
|
#define VMX_EPT_MT_MASK (7ull << VMX_EPT_MT_EPTE_SHIFT)
|
|
|
|
static inline u8 vmx_eptp_page_walk_level(u64 eptp)
|
|
{
|
|
u64 encoded_level = eptp & VMX_EPTP_PWL_MASK;
|
|
|
|
if (encoded_level == VMX_EPTP_PWL_5)
|
|
return 5;
|
|
|
|
/* @eptp must be pre-validated by the caller. */
|
|
WARN_ON_ONCE(encoded_level != VMX_EPTP_PWL_4);
|
|
return 4;
|
|
}
|
|
|
|
/* The mask to use to trigger an EPT Misconfiguration in order to track MMIO */
|
|
#define VMX_EPT_MISCONFIG_WX_VALUE (VMX_EPT_WRITABLE_MASK | \
|
|
VMX_EPT_EXECUTABLE_MASK)
|
|
|
|
#define VMX_EPT_IDENTITY_PAGETABLE_ADDR 0xfffbc000ul
|
|
|
|
struct vmx_msr_entry {
|
|
u32 index;
|
|
u32 reserved;
|
|
u64 value;
|
|
} __aligned(16);
|
|
|
|
/*
|
|
* Exit Qualifications for entry failure during or after loading guest state
|
|
*/
|
|
enum vm_entry_failure_code {
|
|
ENTRY_FAIL_DEFAULT = 0,
|
|
ENTRY_FAIL_PDPTE = 2,
|
|
ENTRY_FAIL_NMI = 3,
|
|
ENTRY_FAIL_VMCS_LINK_PTR = 4,
|
|
};
|
|
|
|
/*
|
|
* Exit Qualifications for EPT Violations
|
|
*/
|
|
#define EPT_VIOLATION_ACC_READ_BIT 0
|
|
#define EPT_VIOLATION_ACC_WRITE_BIT 1
|
|
#define EPT_VIOLATION_ACC_INSTR_BIT 2
|
|
#define EPT_VIOLATION_READABLE_BIT 3
|
|
#define EPT_VIOLATION_WRITABLE_BIT 4
|
|
#define EPT_VIOLATION_EXECUTABLE_BIT 5
|
|
#define EPT_VIOLATION_GVA_TRANSLATED_BIT 8
|
|
#define EPT_VIOLATION_ACC_READ (1 << EPT_VIOLATION_ACC_READ_BIT)
|
|
#define EPT_VIOLATION_ACC_WRITE (1 << EPT_VIOLATION_ACC_WRITE_BIT)
|
|
#define EPT_VIOLATION_ACC_INSTR (1 << EPT_VIOLATION_ACC_INSTR_BIT)
|
|
#define EPT_VIOLATION_READABLE (1 << EPT_VIOLATION_READABLE_BIT)
|
|
#define EPT_VIOLATION_WRITABLE (1 << EPT_VIOLATION_WRITABLE_BIT)
|
|
#define EPT_VIOLATION_EXECUTABLE (1 << EPT_VIOLATION_EXECUTABLE_BIT)
|
|
#define EPT_VIOLATION_GVA_TRANSLATED (1 << EPT_VIOLATION_GVA_TRANSLATED_BIT)
|
|
|
|
/*
|
|
* VM-instruction error numbers
|
|
*/
|
|
enum vm_instruction_error_number {
|
|
VMXERR_VMCALL_IN_VMX_ROOT_OPERATION = 1,
|
|
VMXERR_VMCLEAR_INVALID_ADDRESS = 2,
|
|
VMXERR_VMCLEAR_VMXON_POINTER = 3,
|
|
VMXERR_VMLAUNCH_NONCLEAR_VMCS = 4,
|
|
VMXERR_VMRESUME_NONLAUNCHED_VMCS = 5,
|
|
VMXERR_VMRESUME_AFTER_VMXOFF = 6,
|
|
VMXERR_ENTRY_INVALID_CONTROL_FIELD = 7,
|
|
VMXERR_ENTRY_INVALID_HOST_STATE_FIELD = 8,
|
|
VMXERR_VMPTRLD_INVALID_ADDRESS = 9,
|
|
VMXERR_VMPTRLD_VMXON_POINTER = 10,
|
|
VMXERR_VMPTRLD_INCORRECT_VMCS_REVISION_ID = 11,
|
|
VMXERR_UNSUPPORTED_VMCS_COMPONENT = 12,
|
|
VMXERR_VMWRITE_READ_ONLY_VMCS_COMPONENT = 13,
|
|
VMXERR_VMXON_IN_VMX_ROOT_OPERATION = 15,
|
|
VMXERR_ENTRY_INVALID_EXECUTIVE_VMCS_POINTER = 16,
|
|
VMXERR_ENTRY_NONLAUNCHED_EXECUTIVE_VMCS = 17,
|
|
VMXERR_ENTRY_EXECUTIVE_VMCS_POINTER_NOT_VMXON_POINTER = 18,
|
|
VMXERR_VMCALL_NONCLEAR_VMCS = 19,
|
|
VMXERR_VMCALL_INVALID_VM_EXIT_CONTROL_FIELDS = 20,
|
|
VMXERR_VMCALL_INCORRECT_MSEG_REVISION_ID = 22,
|
|
VMXERR_VMXOFF_UNDER_DUAL_MONITOR_TREATMENT_OF_SMIS_AND_SMM = 23,
|
|
VMXERR_VMCALL_INVALID_SMM_MONITOR_FEATURES = 24,
|
|
VMXERR_ENTRY_INVALID_VM_EXECUTION_CONTROL_FIELDS_IN_EXECUTIVE_VMCS = 25,
|
|
VMXERR_ENTRY_EVENTS_BLOCKED_BY_MOV_SS = 26,
|
|
VMXERR_INVALID_OPERAND_TO_INVEPT_INVVPID = 28,
|
|
};
|
|
|
|
/*
|
|
* VM-instruction errors that can be encountered on VM-Enter, used to trace
|
|
* nested VM-Enter failures reported by hardware. Errors unique to VM-Enter
|
|
* from a SMI Transfer Monitor are not included as things have gone seriously
|
|
* sideways if we get one of those...
|
|
*/
|
|
#define VMX_VMENTER_INSTRUCTION_ERRORS \
|
|
{ VMXERR_VMLAUNCH_NONCLEAR_VMCS, "VMLAUNCH_NONCLEAR_VMCS" }, \
|
|
{ VMXERR_VMRESUME_NONLAUNCHED_VMCS, "VMRESUME_NONLAUNCHED_VMCS" }, \
|
|
{ VMXERR_VMRESUME_AFTER_VMXOFF, "VMRESUME_AFTER_VMXOFF" }, \
|
|
{ VMXERR_ENTRY_INVALID_CONTROL_FIELD, "VMENTRY_INVALID_CONTROL_FIELD" }, \
|
|
{ VMXERR_ENTRY_INVALID_HOST_STATE_FIELD, "VMENTRY_INVALID_HOST_STATE_FIELD" }, \
|
|
{ VMXERR_ENTRY_EVENTS_BLOCKED_BY_MOV_SS, "VMENTRY_EVENTS_BLOCKED_BY_MOV_SS" }
|
|
|
|
enum vmx_l1d_flush_state {
|
|
VMENTER_L1D_FLUSH_AUTO,
|
|
VMENTER_L1D_FLUSH_NEVER,
|
|
VMENTER_L1D_FLUSH_COND,
|
|
VMENTER_L1D_FLUSH_ALWAYS,
|
|
VMENTER_L1D_FLUSH_EPT_DISABLED,
|
|
VMENTER_L1D_FLUSH_NOT_REQUIRED,
|
|
};
|
|
|
|
extern enum vmx_l1d_flush_state l1tf_vmx_mitigation;
|
|
|
|
#endif
|