mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-25 00:06:28 +07:00
6f6a657c99
Hyper-V provides direct tlb flush function which helps L1 Hypervisor to handle Hyper-V tlb flush request from L2 guest. Add the function support for VMX. Signed-off-by: Vitaly Kuznetsov <vkuznets@redhat.com> Signed-off-by: Tianyu Lan <Tianyu.Lan@microsoft.com> Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
206 lines
5.7 KiB
C
206 lines
5.7 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
#ifndef __KVM_X86_VMX_EVMCS_H
|
|
#define __KVM_X86_VMX_EVMCS_H
|
|
|
|
#include <linux/jump_label.h>
|
|
|
|
#include <asm/hyperv-tlfs.h>
|
|
#include <asm/mshyperv.h>
|
|
#include <asm/vmx.h>
|
|
|
|
#include "capabilities.h"
|
|
#include "vmcs.h"
|
|
|
|
struct vmcs_config;
|
|
|
|
DECLARE_STATIC_KEY_FALSE(enable_evmcs);
|
|
|
|
#define current_evmcs ((struct hv_enlightened_vmcs *)this_cpu_read(current_vmcs))
|
|
|
|
#define KVM_EVMCS_VERSION 1
|
|
|
|
/*
|
|
* Enlightened VMCSv1 doesn't support these:
|
|
*
|
|
* POSTED_INTR_NV = 0x00000002,
|
|
* GUEST_INTR_STATUS = 0x00000810,
|
|
* APIC_ACCESS_ADDR = 0x00002014,
|
|
* POSTED_INTR_DESC_ADDR = 0x00002016,
|
|
* EOI_EXIT_BITMAP0 = 0x0000201c,
|
|
* EOI_EXIT_BITMAP1 = 0x0000201e,
|
|
* EOI_EXIT_BITMAP2 = 0x00002020,
|
|
* EOI_EXIT_BITMAP3 = 0x00002022,
|
|
* GUEST_PML_INDEX = 0x00000812,
|
|
* PML_ADDRESS = 0x0000200e,
|
|
* VM_FUNCTION_CONTROL = 0x00002018,
|
|
* EPTP_LIST_ADDRESS = 0x00002024,
|
|
* VMREAD_BITMAP = 0x00002026,
|
|
* VMWRITE_BITMAP = 0x00002028,
|
|
*
|
|
* TSC_MULTIPLIER = 0x00002032,
|
|
* PLE_GAP = 0x00004020,
|
|
* PLE_WINDOW = 0x00004022,
|
|
* VMX_PREEMPTION_TIMER_VALUE = 0x0000482E,
|
|
* GUEST_IA32_PERF_GLOBAL_CTRL = 0x00002808,
|
|
* HOST_IA32_PERF_GLOBAL_CTRL = 0x00002c04,
|
|
*
|
|
* Currently unsupported in KVM:
|
|
* GUEST_IA32_RTIT_CTL = 0x00002814,
|
|
*/
|
|
#define EVMCS1_UNSUPPORTED_PINCTRL (PIN_BASED_POSTED_INTR | \
|
|
PIN_BASED_VMX_PREEMPTION_TIMER)
|
|
#define EVMCS1_UNSUPPORTED_2NDEXEC \
|
|
(SECONDARY_EXEC_VIRTUAL_INTR_DELIVERY | \
|
|
SECONDARY_EXEC_VIRTUALIZE_APIC_ACCESSES | \
|
|
SECONDARY_EXEC_APIC_REGISTER_VIRT | \
|
|
SECONDARY_EXEC_ENABLE_PML | \
|
|
SECONDARY_EXEC_ENABLE_VMFUNC | \
|
|
SECONDARY_EXEC_SHADOW_VMCS | \
|
|
SECONDARY_EXEC_TSC_SCALING | \
|
|
SECONDARY_EXEC_PAUSE_LOOP_EXITING)
|
|
#define EVMCS1_UNSUPPORTED_VMEXIT_CTRL (VM_EXIT_LOAD_IA32_PERF_GLOBAL_CTRL)
|
|
#define EVMCS1_UNSUPPORTED_VMENTRY_CTRL (VM_ENTRY_LOAD_IA32_PERF_GLOBAL_CTRL)
|
|
#define EVMCS1_UNSUPPORTED_VMFUNC (VMX_VMFUNC_EPTP_SWITCHING)
|
|
|
|
#if IS_ENABLED(CONFIG_HYPERV)
|
|
|
|
struct evmcs_field {
|
|
u16 offset;
|
|
u16 clean_field;
|
|
};
|
|
|
|
extern const struct evmcs_field vmcs_field_to_evmcs_1[];
|
|
extern const unsigned int nr_evmcs_1_fields;
|
|
|
|
#define ROL16(val, n) ((u16)(((u16)(val) << (n)) | ((u16)(val) >> (16 - (n)))))
|
|
|
|
static __always_inline int get_evmcs_offset(unsigned long field,
|
|
u16 *clean_field)
|
|
{
|
|
unsigned int index = ROL16(field, 6);
|
|
const struct evmcs_field *evmcs_field;
|
|
|
|
if (unlikely(index >= nr_evmcs_1_fields)) {
|
|
WARN_ONCE(1, "KVM: accessing unsupported EVMCS field %lx\n",
|
|
field);
|
|
return -ENOENT;
|
|
}
|
|
|
|
evmcs_field = &vmcs_field_to_evmcs_1[index];
|
|
|
|
if (clean_field)
|
|
*clean_field = evmcs_field->clean_field;
|
|
|
|
return evmcs_field->offset;
|
|
}
|
|
|
|
#undef ROL16
|
|
|
|
static inline void evmcs_write64(unsigned long field, u64 value)
|
|
{
|
|
u16 clean_field;
|
|
int offset = get_evmcs_offset(field, &clean_field);
|
|
|
|
if (offset < 0)
|
|
return;
|
|
|
|
*(u64 *)((char *)current_evmcs + offset) = value;
|
|
|
|
current_evmcs->hv_clean_fields &= ~clean_field;
|
|
}
|
|
|
|
static inline void evmcs_write32(unsigned long field, u32 value)
|
|
{
|
|
u16 clean_field;
|
|
int offset = get_evmcs_offset(field, &clean_field);
|
|
|
|
if (offset < 0)
|
|
return;
|
|
|
|
*(u32 *)((char *)current_evmcs + offset) = value;
|
|
current_evmcs->hv_clean_fields &= ~clean_field;
|
|
}
|
|
|
|
static inline void evmcs_write16(unsigned long field, u16 value)
|
|
{
|
|
u16 clean_field;
|
|
int offset = get_evmcs_offset(field, &clean_field);
|
|
|
|
if (offset < 0)
|
|
return;
|
|
|
|
*(u16 *)((char *)current_evmcs + offset) = value;
|
|
current_evmcs->hv_clean_fields &= ~clean_field;
|
|
}
|
|
|
|
static inline u64 evmcs_read64(unsigned long field)
|
|
{
|
|
int offset = get_evmcs_offset(field, NULL);
|
|
|
|
if (offset < 0)
|
|
return 0;
|
|
|
|
return *(u64 *)((char *)current_evmcs + offset);
|
|
}
|
|
|
|
static inline u32 evmcs_read32(unsigned long field)
|
|
{
|
|
int offset = get_evmcs_offset(field, NULL);
|
|
|
|
if (offset < 0)
|
|
return 0;
|
|
|
|
return *(u32 *)((char *)current_evmcs + offset);
|
|
}
|
|
|
|
static inline u16 evmcs_read16(unsigned long field)
|
|
{
|
|
int offset = get_evmcs_offset(field, NULL);
|
|
|
|
if (offset < 0)
|
|
return 0;
|
|
|
|
return *(u16 *)((char *)current_evmcs + offset);
|
|
}
|
|
|
|
static inline void evmcs_touch_msr_bitmap(void)
|
|
{
|
|
if (unlikely(!current_evmcs))
|
|
return;
|
|
|
|
if (current_evmcs->hv_enlightenments_control.msr_bitmap)
|
|
current_evmcs->hv_clean_fields &=
|
|
~HV_VMX_ENLIGHTENED_CLEAN_FIELD_MSR_BITMAP;
|
|
}
|
|
|
|
static inline void evmcs_load(u64 phys_addr)
|
|
{
|
|
struct hv_vp_assist_page *vp_ap =
|
|
hv_get_vp_assist_page(smp_processor_id());
|
|
|
|
if (current_evmcs->hv_enlightenments_control.nested_flush_hypercall)
|
|
vp_ap->nested_control.features.directhypercall = 1;
|
|
vp_ap->current_nested_vmcs = phys_addr;
|
|
vp_ap->enlighten_vmentry = 1;
|
|
}
|
|
|
|
void evmcs_sanitize_exec_ctrls(struct vmcs_config *vmcs_conf);
|
|
#else /* !IS_ENABLED(CONFIG_HYPERV) */
|
|
static inline void evmcs_write64(unsigned long field, u64 value) {}
|
|
static inline void evmcs_write32(unsigned long field, u32 value) {}
|
|
static inline void evmcs_write16(unsigned long field, u16 value) {}
|
|
static inline u64 evmcs_read64(unsigned long field) { return 0; }
|
|
static inline u32 evmcs_read32(unsigned long field) { return 0; }
|
|
static inline u16 evmcs_read16(unsigned long field) { return 0; }
|
|
static inline void evmcs_load(u64 phys_addr) {}
|
|
static inline void evmcs_sanitize_exec_ctrls(struct vmcs_config *vmcs_conf) {}
|
|
static inline void evmcs_touch_msr_bitmap(void) {}
|
|
#endif /* IS_ENABLED(CONFIG_HYPERV) */
|
|
|
|
bool nested_enlightened_vmentry(struct kvm_vcpu *vcpu, u64 *evmcs_gpa);
|
|
uint16_t nested_get_evmcs_version(struct kvm_vcpu *vcpu);
|
|
int nested_enable_evmcs(struct kvm_vcpu *vcpu,
|
|
uint16_t *vmcs_version);
|
|
|
|
#endif /* __KVM_X86_VMX_EVMCS_H */
|