mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-18 04:06:54 +07:00
a45ff5994c
- Add support for chained PMU counters in guests - Improve SError handling - Handle Neoverse N1 erratum #1349291 - Allow side-channel mitigation status to be migrated - Standardise most AArch64 system register accesses to msr_s/mrs_s - Fix host MPIDR corruption on 32bit -----BEGIN PGP SIGNATURE----- iQJJBAABCgAzFiEEn9UcU+C1Yxj9lZw9I9DQutE9ekMFAl0kge4VHG1hcmMuenlu Z2llckBhcm0uY29tAAoJECPQ0LrRPXpDYyQP/3XY5tFcLKkp/h9rnGaCXwAxhNzn TyF/IZEFBKFTSoDMXKLLc8KllvoPQ7aUl03heYbuayYpyKR1+LCx7lDwu1MYyEf+ aSSuOKlbG//tLUEGp09pTRCgjs2mhhZYqOj5GF2mZ7xpovFVSNOPzTazbXDNQ7tw zUAs43YNg+bUMwj+SLWpBlizjrLr7T34utIr6daKJE/GSfmIrcYXhGbZqUh0zbO0 z5LNasebws8/pHyeGI7+/yoMIKaQ8foMgywTpsRpBsx6YI+AbOLjEmCk2IBOPcEK pm9KkSIBZEO2CSxZKl3NQiEow/Qd/lnz2xLMCSfh4XrYoI2Th4gNcsbJpiBDWP5a 0eZ5jSiexxKngIbM+to7jR3m0yc9RgcuzceJg3Uly7Ya0vb5RqKwOX4Ge4XP4VDT DzIVFdQjxDKdVIf3EvGp1cj4P7dRUU3xbZcbzyuRPEmT3vgjEnbxawmPLs3QMAl1 31Wd2wIsPB86kSxzSMel27Vs5VgMhgyHE26zN91R745CvhDXaDKydIWjGjdVMHsB GuX/h2kL+ohx+N/OpZPgwsVUAGLSOQFP3pE/EcGtqc2kkfqa+bx12DKcZ3zdmJvy +cu5ixU8q5thPH/pZob/C3hKUY/eLy02emS34RK0Jh2sZHbQgAOtMsiqUxNHEjUm 6TkpdWa5SRd7CtGV =yfCs -----END PGP SIGNATURE----- Merge tag 'kvm-arm-for-5.3' of git://git.kernel.org/pub/scm/linux/kernel/git/kvmarm/kvmarm into HEAD KVM/arm updates for 5.3 - Add support for chained PMU counters in guests - Improve SError handling - Handle Neoverse N1 erratum #1349291 - Allow side-channel mitigation status to be migrated - Standardise most AArch64 system register accesses to msr_s/mrs_s - Fix host MPIDR corruption on 32bit
443 lines
14 KiB
C
443 lines
14 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
#ifndef __KVM_X86_VMX_VMCS12_H
|
|
#define __KVM_X86_VMX_VMCS12_H
|
|
|
|
#include <linux/build_bug.h>
|
|
|
|
#include "vmcs.h"
|
|
|
|
/*
|
|
* struct vmcs12 describes the state that our guest hypervisor (L1) keeps for a
|
|
* single nested guest (L2), hence the name vmcs12. Any VMX implementation has
|
|
* a VMCS structure, and vmcs12 is our emulated VMX's VMCS. This structure is
|
|
* stored in guest memory specified by VMPTRLD, but is opaque to the guest,
|
|
* which must access it using VMREAD/VMWRITE/VMCLEAR instructions.
|
|
* More than one of these structures may exist, if L1 runs multiple L2 guests.
|
|
* nested_vmx_run() will use the data here to build the vmcs02: a VMCS for the
|
|
* underlying hardware which will be used to run L2.
|
|
* This structure is packed to ensure that its layout is identical across
|
|
* machines (necessary for live migration).
|
|
*
|
|
* IMPORTANT: Changing the layout of existing fields in this structure
|
|
* will break save/restore compatibility with older kvm releases. When
|
|
* adding new fields, either use space in the reserved padding* arrays
|
|
* or add the new fields to the end of the structure.
|
|
*/
|
|
typedef u64 natural_width;
|
|
struct __packed vmcs12 {
|
|
/* According to the Intel spec, a VMCS region must start with the
|
|
* following two fields. Then follow implementation-specific data.
|
|
*/
|
|
struct vmcs_hdr hdr;
|
|
u32 abort;
|
|
|
|
u32 launch_state; /* set to 0 by VMCLEAR, to 1 by VMLAUNCH */
|
|
u32 padding[7]; /* room for future expansion */
|
|
|
|
u64 io_bitmap_a;
|
|
u64 io_bitmap_b;
|
|
u64 msr_bitmap;
|
|
u64 vm_exit_msr_store_addr;
|
|
u64 vm_exit_msr_load_addr;
|
|
u64 vm_entry_msr_load_addr;
|
|
u64 tsc_offset;
|
|
u64 virtual_apic_page_addr;
|
|
u64 apic_access_addr;
|
|
u64 posted_intr_desc_addr;
|
|
u64 ept_pointer;
|
|
u64 eoi_exit_bitmap0;
|
|
u64 eoi_exit_bitmap1;
|
|
u64 eoi_exit_bitmap2;
|
|
u64 eoi_exit_bitmap3;
|
|
u64 xss_exit_bitmap;
|
|
u64 guest_physical_address;
|
|
u64 vmcs_link_pointer;
|
|
u64 guest_ia32_debugctl;
|
|
u64 guest_ia32_pat;
|
|
u64 guest_ia32_efer;
|
|
u64 guest_ia32_perf_global_ctrl;
|
|
u64 guest_pdptr0;
|
|
u64 guest_pdptr1;
|
|
u64 guest_pdptr2;
|
|
u64 guest_pdptr3;
|
|
u64 guest_bndcfgs;
|
|
u64 host_ia32_pat;
|
|
u64 host_ia32_efer;
|
|
u64 host_ia32_perf_global_ctrl;
|
|
u64 vmread_bitmap;
|
|
u64 vmwrite_bitmap;
|
|
u64 vm_function_control;
|
|
u64 eptp_list_address;
|
|
u64 pml_address;
|
|
u64 padding64[3]; /* room for future expansion */
|
|
/*
|
|
* To allow migration of L1 (complete with its L2 guests) between
|
|
* machines of different natural widths (32 or 64 bit), we cannot have
|
|
* unsigned long fields with no explicit size. We use u64 (aliased
|
|
* natural_width) instead. Luckily, x86 is little-endian.
|
|
*/
|
|
natural_width cr0_guest_host_mask;
|
|
natural_width cr4_guest_host_mask;
|
|
natural_width cr0_read_shadow;
|
|
natural_width cr4_read_shadow;
|
|
natural_width cr3_target_value0;
|
|
natural_width cr3_target_value1;
|
|
natural_width cr3_target_value2;
|
|
natural_width cr3_target_value3;
|
|
natural_width exit_qualification;
|
|
natural_width guest_linear_address;
|
|
natural_width guest_cr0;
|
|
natural_width guest_cr3;
|
|
natural_width guest_cr4;
|
|
natural_width guest_es_base;
|
|
natural_width guest_cs_base;
|
|
natural_width guest_ss_base;
|
|
natural_width guest_ds_base;
|
|
natural_width guest_fs_base;
|
|
natural_width guest_gs_base;
|
|
natural_width guest_ldtr_base;
|
|
natural_width guest_tr_base;
|
|
natural_width guest_gdtr_base;
|
|
natural_width guest_idtr_base;
|
|
natural_width guest_dr7;
|
|
natural_width guest_rsp;
|
|
natural_width guest_rip;
|
|
natural_width guest_rflags;
|
|
natural_width guest_pending_dbg_exceptions;
|
|
natural_width guest_sysenter_esp;
|
|
natural_width guest_sysenter_eip;
|
|
natural_width host_cr0;
|
|
natural_width host_cr3;
|
|
natural_width host_cr4;
|
|
natural_width host_fs_base;
|
|
natural_width host_gs_base;
|
|
natural_width host_tr_base;
|
|
natural_width host_gdtr_base;
|
|
natural_width host_idtr_base;
|
|
natural_width host_ia32_sysenter_esp;
|
|
natural_width host_ia32_sysenter_eip;
|
|
natural_width host_rsp;
|
|
natural_width host_rip;
|
|
natural_width paddingl[8]; /* room for future expansion */
|
|
u32 pin_based_vm_exec_control;
|
|
u32 cpu_based_vm_exec_control;
|
|
u32 exception_bitmap;
|
|
u32 page_fault_error_code_mask;
|
|
u32 page_fault_error_code_match;
|
|
u32 cr3_target_count;
|
|
u32 vm_exit_controls;
|
|
u32 vm_exit_msr_store_count;
|
|
u32 vm_exit_msr_load_count;
|
|
u32 vm_entry_controls;
|
|
u32 vm_entry_msr_load_count;
|
|
u32 vm_entry_intr_info_field;
|
|
u32 vm_entry_exception_error_code;
|
|
u32 vm_entry_instruction_len;
|
|
u32 tpr_threshold;
|
|
u32 secondary_vm_exec_control;
|
|
u32 vm_instruction_error;
|
|
u32 vm_exit_reason;
|
|
u32 vm_exit_intr_info;
|
|
u32 vm_exit_intr_error_code;
|
|
u32 idt_vectoring_info_field;
|
|
u32 idt_vectoring_error_code;
|
|
u32 vm_exit_instruction_len;
|
|
u32 vmx_instruction_info;
|
|
u32 guest_es_limit;
|
|
u32 guest_cs_limit;
|
|
u32 guest_ss_limit;
|
|
u32 guest_ds_limit;
|
|
u32 guest_fs_limit;
|
|
u32 guest_gs_limit;
|
|
u32 guest_ldtr_limit;
|
|
u32 guest_tr_limit;
|
|
u32 guest_gdtr_limit;
|
|
u32 guest_idtr_limit;
|
|
u32 guest_es_ar_bytes;
|
|
u32 guest_cs_ar_bytes;
|
|
u32 guest_ss_ar_bytes;
|
|
u32 guest_ds_ar_bytes;
|
|
u32 guest_fs_ar_bytes;
|
|
u32 guest_gs_ar_bytes;
|
|
u32 guest_ldtr_ar_bytes;
|
|
u32 guest_tr_ar_bytes;
|
|
u32 guest_interruptibility_info;
|
|
u32 guest_activity_state;
|
|
u32 guest_sysenter_cs;
|
|
u32 host_ia32_sysenter_cs;
|
|
u32 vmx_preemption_timer_value;
|
|
u32 padding32[7]; /* room for future expansion */
|
|
u16 virtual_processor_id;
|
|
u16 posted_intr_nv;
|
|
u16 guest_es_selector;
|
|
u16 guest_cs_selector;
|
|
u16 guest_ss_selector;
|
|
u16 guest_ds_selector;
|
|
u16 guest_fs_selector;
|
|
u16 guest_gs_selector;
|
|
u16 guest_ldtr_selector;
|
|
u16 guest_tr_selector;
|
|
u16 guest_intr_status;
|
|
u16 host_es_selector;
|
|
u16 host_cs_selector;
|
|
u16 host_ss_selector;
|
|
u16 host_ds_selector;
|
|
u16 host_fs_selector;
|
|
u16 host_gs_selector;
|
|
u16 host_tr_selector;
|
|
u16 guest_pml_index;
|
|
};
|
|
|
|
/*
|
|
* VMCS12_REVISION is an arbitrary id that should be changed if the content or
|
|
* layout of struct vmcs12 is changed. MSR_IA32_VMX_BASIC returns this id, and
|
|
* VMPTRLD verifies that the VMCS region that L1 is loading contains this id.
|
|
*
|
|
* IMPORTANT: Changing this value will break save/restore compatibility with
|
|
* older kvm releases.
|
|
*/
|
|
#define VMCS12_REVISION 0x11e57ed0
|
|
|
|
/*
|
|
* VMCS12_SIZE is the number of bytes L1 should allocate for the VMXON region
|
|
* and any VMCS region. Although only sizeof(struct vmcs12) are used by the
|
|
* current implementation, 4K are reserved to avoid future complications and
|
|
* to preserve userspace ABI.
|
|
*/
|
|
#define VMCS12_SIZE KVM_STATE_NESTED_VMX_VMCS_SIZE
|
|
|
|
/*
|
|
* VMCS12_MAX_FIELD_INDEX is the highest index value used in any
|
|
* supported VMCS12 field encoding.
|
|
*/
|
|
#define VMCS12_MAX_FIELD_INDEX 0x17
|
|
|
|
/*
|
|
* For save/restore compatibility, the vmcs12 field offsets must not change.
|
|
*/
|
|
#define CHECK_OFFSET(field, loc) \
|
|
BUILD_BUG_ON_MSG(offsetof(struct vmcs12, field) != (loc), \
|
|
"Offset of " #field " in struct vmcs12 has changed.")
|
|
|
|
static inline void vmx_check_vmcs12_offsets(void)
|
|
{
|
|
CHECK_OFFSET(hdr, 0);
|
|
CHECK_OFFSET(abort, 4);
|
|
CHECK_OFFSET(launch_state, 8);
|
|
CHECK_OFFSET(io_bitmap_a, 40);
|
|
CHECK_OFFSET(io_bitmap_b, 48);
|
|
CHECK_OFFSET(msr_bitmap, 56);
|
|
CHECK_OFFSET(vm_exit_msr_store_addr, 64);
|
|
CHECK_OFFSET(vm_exit_msr_load_addr, 72);
|
|
CHECK_OFFSET(vm_entry_msr_load_addr, 80);
|
|
CHECK_OFFSET(tsc_offset, 88);
|
|
CHECK_OFFSET(virtual_apic_page_addr, 96);
|
|
CHECK_OFFSET(apic_access_addr, 104);
|
|
CHECK_OFFSET(posted_intr_desc_addr, 112);
|
|
CHECK_OFFSET(ept_pointer, 120);
|
|
CHECK_OFFSET(eoi_exit_bitmap0, 128);
|
|
CHECK_OFFSET(eoi_exit_bitmap1, 136);
|
|
CHECK_OFFSET(eoi_exit_bitmap2, 144);
|
|
CHECK_OFFSET(eoi_exit_bitmap3, 152);
|
|
CHECK_OFFSET(xss_exit_bitmap, 160);
|
|
CHECK_OFFSET(guest_physical_address, 168);
|
|
CHECK_OFFSET(vmcs_link_pointer, 176);
|
|
CHECK_OFFSET(guest_ia32_debugctl, 184);
|
|
CHECK_OFFSET(guest_ia32_pat, 192);
|
|
CHECK_OFFSET(guest_ia32_efer, 200);
|
|
CHECK_OFFSET(guest_ia32_perf_global_ctrl, 208);
|
|
CHECK_OFFSET(guest_pdptr0, 216);
|
|
CHECK_OFFSET(guest_pdptr1, 224);
|
|
CHECK_OFFSET(guest_pdptr2, 232);
|
|
CHECK_OFFSET(guest_pdptr3, 240);
|
|
CHECK_OFFSET(guest_bndcfgs, 248);
|
|
CHECK_OFFSET(host_ia32_pat, 256);
|
|
CHECK_OFFSET(host_ia32_efer, 264);
|
|
CHECK_OFFSET(host_ia32_perf_global_ctrl, 272);
|
|
CHECK_OFFSET(vmread_bitmap, 280);
|
|
CHECK_OFFSET(vmwrite_bitmap, 288);
|
|
CHECK_OFFSET(vm_function_control, 296);
|
|
CHECK_OFFSET(eptp_list_address, 304);
|
|
CHECK_OFFSET(pml_address, 312);
|
|
CHECK_OFFSET(cr0_guest_host_mask, 344);
|
|
CHECK_OFFSET(cr4_guest_host_mask, 352);
|
|
CHECK_OFFSET(cr0_read_shadow, 360);
|
|
CHECK_OFFSET(cr4_read_shadow, 368);
|
|
CHECK_OFFSET(cr3_target_value0, 376);
|
|
CHECK_OFFSET(cr3_target_value1, 384);
|
|
CHECK_OFFSET(cr3_target_value2, 392);
|
|
CHECK_OFFSET(cr3_target_value3, 400);
|
|
CHECK_OFFSET(exit_qualification, 408);
|
|
CHECK_OFFSET(guest_linear_address, 416);
|
|
CHECK_OFFSET(guest_cr0, 424);
|
|
CHECK_OFFSET(guest_cr3, 432);
|
|
CHECK_OFFSET(guest_cr4, 440);
|
|
CHECK_OFFSET(guest_es_base, 448);
|
|
CHECK_OFFSET(guest_cs_base, 456);
|
|
CHECK_OFFSET(guest_ss_base, 464);
|
|
CHECK_OFFSET(guest_ds_base, 472);
|
|
CHECK_OFFSET(guest_fs_base, 480);
|
|
CHECK_OFFSET(guest_gs_base, 488);
|
|
CHECK_OFFSET(guest_ldtr_base, 496);
|
|
CHECK_OFFSET(guest_tr_base, 504);
|
|
CHECK_OFFSET(guest_gdtr_base, 512);
|
|
CHECK_OFFSET(guest_idtr_base, 520);
|
|
CHECK_OFFSET(guest_dr7, 528);
|
|
CHECK_OFFSET(guest_rsp, 536);
|
|
CHECK_OFFSET(guest_rip, 544);
|
|
CHECK_OFFSET(guest_rflags, 552);
|
|
CHECK_OFFSET(guest_pending_dbg_exceptions, 560);
|
|
CHECK_OFFSET(guest_sysenter_esp, 568);
|
|
CHECK_OFFSET(guest_sysenter_eip, 576);
|
|
CHECK_OFFSET(host_cr0, 584);
|
|
CHECK_OFFSET(host_cr3, 592);
|
|
CHECK_OFFSET(host_cr4, 600);
|
|
CHECK_OFFSET(host_fs_base, 608);
|
|
CHECK_OFFSET(host_gs_base, 616);
|
|
CHECK_OFFSET(host_tr_base, 624);
|
|
CHECK_OFFSET(host_gdtr_base, 632);
|
|
CHECK_OFFSET(host_idtr_base, 640);
|
|
CHECK_OFFSET(host_ia32_sysenter_esp, 648);
|
|
CHECK_OFFSET(host_ia32_sysenter_eip, 656);
|
|
CHECK_OFFSET(host_rsp, 664);
|
|
CHECK_OFFSET(host_rip, 672);
|
|
CHECK_OFFSET(pin_based_vm_exec_control, 744);
|
|
CHECK_OFFSET(cpu_based_vm_exec_control, 748);
|
|
CHECK_OFFSET(exception_bitmap, 752);
|
|
CHECK_OFFSET(page_fault_error_code_mask, 756);
|
|
CHECK_OFFSET(page_fault_error_code_match, 760);
|
|
CHECK_OFFSET(cr3_target_count, 764);
|
|
CHECK_OFFSET(vm_exit_controls, 768);
|
|
CHECK_OFFSET(vm_exit_msr_store_count, 772);
|
|
CHECK_OFFSET(vm_exit_msr_load_count, 776);
|
|
CHECK_OFFSET(vm_entry_controls, 780);
|
|
CHECK_OFFSET(vm_entry_msr_load_count, 784);
|
|
CHECK_OFFSET(vm_entry_intr_info_field, 788);
|
|
CHECK_OFFSET(vm_entry_exception_error_code, 792);
|
|
CHECK_OFFSET(vm_entry_instruction_len, 796);
|
|
CHECK_OFFSET(tpr_threshold, 800);
|
|
CHECK_OFFSET(secondary_vm_exec_control, 804);
|
|
CHECK_OFFSET(vm_instruction_error, 808);
|
|
CHECK_OFFSET(vm_exit_reason, 812);
|
|
CHECK_OFFSET(vm_exit_intr_info, 816);
|
|
CHECK_OFFSET(vm_exit_intr_error_code, 820);
|
|
CHECK_OFFSET(idt_vectoring_info_field, 824);
|
|
CHECK_OFFSET(idt_vectoring_error_code, 828);
|
|
CHECK_OFFSET(vm_exit_instruction_len, 832);
|
|
CHECK_OFFSET(vmx_instruction_info, 836);
|
|
CHECK_OFFSET(guest_es_limit, 840);
|
|
CHECK_OFFSET(guest_cs_limit, 844);
|
|
CHECK_OFFSET(guest_ss_limit, 848);
|
|
CHECK_OFFSET(guest_ds_limit, 852);
|
|
CHECK_OFFSET(guest_fs_limit, 856);
|
|
CHECK_OFFSET(guest_gs_limit, 860);
|
|
CHECK_OFFSET(guest_ldtr_limit, 864);
|
|
CHECK_OFFSET(guest_tr_limit, 868);
|
|
CHECK_OFFSET(guest_gdtr_limit, 872);
|
|
CHECK_OFFSET(guest_idtr_limit, 876);
|
|
CHECK_OFFSET(guest_es_ar_bytes, 880);
|
|
CHECK_OFFSET(guest_cs_ar_bytes, 884);
|
|
CHECK_OFFSET(guest_ss_ar_bytes, 888);
|
|
CHECK_OFFSET(guest_ds_ar_bytes, 892);
|
|
CHECK_OFFSET(guest_fs_ar_bytes, 896);
|
|
CHECK_OFFSET(guest_gs_ar_bytes, 900);
|
|
CHECK_OFFSET(guest_ldtr_ar_bytes, 904);
|
|
CHECK_OFFSET(guest_tr_ar_bytes, 908);
|
|
CHECK_OFFSET(guest_interruptibility_info, 912);
|
|
CHECK_OFFSET(guest_activity_state, 916);
|
|
CHECK_OFFSET(guest_sysenter_cs, 920);
|
|
CHECK_OFFSET(host_ia32_sysenter_cs, 924);
|
|
CHECK_OFFSET(vmx_preemption_timer_value, 928);
|
|
CHECK_OFFSET(virtual_processor_id, 960);
|
|
CHECK_OFFSET(posted_intr_nv, 962);
|
|
CHECK_OFFSET(guest_es_selector, 964);
|
|
CHECK_OFFSET(guest_cs_selector, 966);
|
|
CHECK_OFFSET(guest_ss_selector, 968);
|
|
CHECK_OFFSET(guest_ds_selector, 970);
|
|
CHECK_OFFSET(guest_fs_selector, 972);
|
|
CHECK_OFFSET(guest_gs_selector, 974);
|
|
CHECK_OFFSET(guest_ldtr_selector, 976);
|
|
CHECK_OFFSET(guest_tr_selector, 978);
|
|
CHECK_OFFSET(guest_intr_status, 980);
|
|
CHECK_OFFSET(host_es_selector, 982);
|
|
CHECK_OFFSET(host_cs_selector, 984);
|
|
CHECK_OFFSET(host_ss_selector, 986);
|
|
CHECK_OFFSET(host_ds_selector, 988);
|
|
CHECK_OFFSET(host_fs_selector, 990);
|
|
CHECK_OFFSET(host_gs_selector, 992);
|
|
CHECK_OFFSET(host_tr_selector, 994);
|
|
CHECK_OFFSET(guest_pml_index, 996);
|
|
}
|
|
|
|
extern const unsigned short vmcs_field_to_offset_table[];
|
|
extern const unsigned int nr_vmcs12_fields;
|
|
|
|
#define ROL16(val, n) ((u16)(((u16)(val) << (n)) | ((u16)(val) >> (16 - (n)))))
|
|
|
|
static inline short vmcs_field_to_offset(unsigned long field)
|
|
{
|
|
unsigned short offset;
|
|
unsigned int index;
|
|
|
|
if (field >> 15)
|
|
return -ENOENT;
|
|
|
|
index = ROL16(field, 6);
|
|
if (index >= nr_vmcs12_fields)
|
|
return -ENOENT;
|
|
|
|
index = array_index_nospec(index, nr_vmcs12_fields);
|
|
offset = vmcs_field_to_offset_table[index];
|
|
if (offset == 0)
|
|
return -ENOENT;
|
|
return offset;
|
|
}
|
|
|
|
#undef ROL16
|
|
|
|
static inline u64 vmcs12_read_any(struct vmcs12 *vmcs12, unsigned long field,
|
|
u16 offset)
|
|
{
|
|
char *p = (char *)vmcs12 + offset;
|
|
|
|
switch (vmcs_field_width(field)) {
|
|
case VMCS_FIELD_WIDTH_NATURAL_WIDTH:
|
|
return *((natural_width *)p);
|
|
case VMCS_FIELD_WIDTH_U16:
|
|
return *((u16 *)p);
|
|
case VMCS_FIELD_WIDTH_U32:
|
|
return *((u32 *)p);
|
|
case VMCS_FIELD_WIDTH_U64:
|
|
return *((u64 *)p);
|
|
default:
|
|
WARN_ON_ONCE(1);
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
static inline void vmcs12_write_any(struct vmcs12 *vmcs12, unsigned long field,
|
|
u16 offset, u64 field_value)
|
|
{
|
|
char *p = (char *)vmcs12 + offset;
|
|
|
|
switch (vmcs_field_width(field)) {
|
|
case VMCS_FIELD_WIDTH_U16:
|
|
*(u16 *)p = field_value;
|
|
break;
|
|
case VMCS_FIELD_WIDTH_U32:
|
|
*(u32 *)p = field_value;
|
|
break;
|
|
case VMCS_FIELD_WIDTH_U64:
|
|
*(u64 *)p = field_value;
|
|
break;
|
|
case VMCS_FIELD_WIDTH_NATURAL_WIDTH:
|
|
*(natural_width *)p = field_value;
|
|
break;
|
|
default:
|
|
WARN_ON_ONCE(1);
|
|
break;
|
|
}
|
|
}
|
|
|
|
#endif /* __KVM_X86_VMX_VMCS12_H */
|