mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-28 11:18:45 +07:00
99dae3bad2
Now that we have the vcpu floating-point and vector state stored in the same type of struct as the main kernel uses, we can load that state directly from the vcpu struct instead of having extra copies to/from the thread_struct. Similarly, when the guest state needs to be saved, we can have it saved it directly to the vcpu struct by setting the current->thread.fp_save_area and current->thread.vr_save_area pointers. That also means that we don't need to back up and restore userspace's FP/vector state. This all makes the code simpler and faster. Note that it's not necessary to save or modify current->thread.fpexc_mode, since nothing in KVM uses or is affected by its value. Nor is it necessary to touch used_vr or used_vsr. Signed-off-by: Paul Mackerras <paulus@samba.org> Signed-off-by: Alexander Graf <agraf@suse.de>
165 lines
5.7 KiB
C
165 lines
5.7 KiB
C
/*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License, version 2, as
|
|
* published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; if not, write to the Free Software
|
|
* Foundation, 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
|
|
*
|
|
* Copyright IBM Corp. 2008
|
|
*
|
|
* Authors: Hollis Blanchard <hollisb@us.ibm.com>
|
|
*/
|
|
|
|
#ifndef __KVM_BOOKE_H__
|
|
#define __KVM_BOOKE_H__
|
|
|
|
#include <linux/types.h>
|
|
#include <linux/kvm_host.h>
|
|
#include <asm/kvm_ppc.h>
|
|
#include <asm/switch_to.h>
|
|
#include "timing.h"
|
|
|
|
/* interrupt priortity ordering */
|
|
#define BOOKE_IRQPRIO_DATA_STORAGE 0
|
|
#define BOOKE_IRQPRIO_INST_STORAGE 1
|
|
#define BOOKE_IRQPRIO_ALIGNMENT 2
|
|
#define BOOKE_IRQPRIO_PROGRAM 3
|
|
#define BOOKE_IRQPRIO_FP_UNAVAIL 4
|
|
#define BOOKE_IRQPRIO_SPE_UNAVAIL 5
|
|
#define BOOKE_IRQPRIO_SPE_FP_DATA 6
|
|
#define BOOKE_IRQPRIO_SPE_FP_ROUND 7
|
|
#define BOOKE_IRQPRIO_SYSCALL 8
|
|
#define BOOKE_IRQPRIO_AP_UNAVAIL 9
|
|
#define BOOKE_IRQPRIO_DTLB_MISS 10
|
|
#define BOOKE_IRQPRIO_ITLB_MISS 11
|
|
#define BOOKE_IRQPRIO_MACHINE_CHECK 12
|
|
#define BOOKE_IRQPRIO_DEBUG 13
|
|
#define BOOKE_IRQPRIO_CRITICAL 14
|
|
#define BOOKE_IRQPRIO_WATCHDOG 15
|
|
#define BOOKE_IRQPRIO_EXTERNAL 16
|
|
#define BOOKE_IRQPRIO_FIT 17
|
|
#define BOOKE_IRQPRIO_DECREMENTER 18
|
|
#define BOOKE_IRQPRIO_PERFORMANCE_MONITOR 19
|
|
/* Internal pseudo-irqprio for level triggered externals */
|
|
#define BOOKE_IRQPRIO_EXTERNAL_LEVEL 20
|
|
#define BOOKE_IRQPRIO_DBELL 21
|
|
#define BOOKE_IRQPRIO_DBELL_CRIT 22
|
|
#define BOOKE_IRQPRIO_MAX 23
|
|
|
|
#define BOOKE_IRQMASK_EE ((1 << BOOKE_IRQPRIO_EXTERNAL_LEVEL) | \
|
|
(1 << BOOKE_IRQPRIO_PERFORMANCE_MONITOR) | \
|
|
(1 << BOOKE_IRQPRIO_DBELL) | \
|
|
(1 << BOOKE_IRQPRIO_DECREMENTER) | \
|
|
(1 << BOOKE_IRQPRIO_FIT) | \
|
|
(1 << BOOKE_IRQPRIO_EXTERNAL))
|
|
|
|
#define BOOKE_IRQMASK_CE ((1 << BOOKE_IRQPRIO_DBELL_CRIT) | \
|
|
(1 << BOOKE_IRQPRIO_WATCHDOG) | \
|
|
(1 << BOOKE_IRQPRIO_CRITICAL))
|
|
|
|
extern unsigned long kvmppc_booke_handlers;
|
|
extern unsigned long kvmppc_booke_handler_addr[];
|
|
|
|
void kvmppc_set_msr(struct kvm_vcpu *vcpu, u32 new_msr);
|
|
void kvmppc_mmu_msr_notify(struct kvm_vcpu *vcpu, u32 old_msr);
|
|
|
|
void kvmppc_set_epcr(struct kvm_vcpu *vcpu, u32 new_epcr);
|
|
void kvmppc_set_tcr(struct kvm_vcpu *vcpu, u32 new_tcr);
|
|
void kvmppc_set_tsr_bits(struct kvm_vcpu *vcpu, u32 tsr_bits);
|
|
void kvmppc_clr_tsr_bits(struct kvm_vcpu *vcpu, u32 tsr_bits);
|
|
|
|
int kvmppc_booke_emulate_op(struct kvm_run *run, struct kvm_vcpu *vcpu,
|
|
unsigned int inst, int *advance);
|
|
int kvmppc_booke_emulate_mfspr(struct kvm_vcpu *vcpu, int sprn, ulong *spr_val);
|
|
int kvmppc_booke_emulate_mtspr(struct kvm_vcpu *vcpu, int sprn, ulong spr_val);
|
|
|
|
/* low-level asm code to transfer guest state */
|
|
void kvmppc_load_guest_spe(struct kvm_vcpu *vcpu);
|
|
void kvmppc_save_guest_spe(struct kvm_vcpu *vcpu);
|
|
|
|
/* high-level function, manages flags, host state */
|
|
void kvmppc_vcpu_disable_spe(struct kvm_vcpu *vcpu);
|
|
|
|
void kvmppc_booke_vcpu_load(struct kvm_vcpu *vcpu, int cpu);
|
|
void kvmppc_booke_vcpu_put(struct kvm_vcpu *vcpu);
|
|
|
|
enum int_class {
|
|
INT_CLASS_NONCRIT,
|
|
INT_CLASS_CRIT,
|
|
INT_CLASS_MC,
|
|
INT_CLASS_DBG,
|
|
};
|
|
|
|
void kvmppc_set_pending_interrupt(struct kvm_vcpu *vcpu, enum int_class type);
|
|
|
|
extern void kvmppc_mmu_destroy_44x(struct kvm_vcpu *vcpu);
|
|
extern int kvmppc_core_emulate_op_44x(struct kvm_run *run, struct kvm_vcpu *vcpu,
|
|
unsigned int inst, int *advance);
|
|
extern int kvmppc_core_emulate_mtspr_44x(struct kvm_vcpu *vcpu, int sprn,
|
|
ulong spr_val);
|
|
extern int kvmppc_core_emulate_mfspr_44x(struct kvm_vcpu *vcpu, int sprn,
|
|
ulong *spr_val);
|
|
extern void kvmppc_mmu_destroy_e500(struct kvm_vcpu *vcpu);
|
|
extern int kvmppc_core_emulate_op_e500(struct kvm_run *run,
|
|
struct kvm_vcpu *vcpu,
|
|
unsigned int inst, int *advance);
|
|
extern int kvmppc_core_emulate_mtspr_e500(struct kvm_vcpu *vcpu, int sprn,
|
|
ulong spr_val);
|
|
extern int kvmppc_core_emulate_mfspr_e500(struct kvm_vcpu *vcpu, int sprn,
|
|
ulong *spr_val);
|
|
extern void kvmppc_mmu_destroy_e500(struct kvm_vcpu *vcpu);
|
|
extern int kvmppc_core_emulate_op_e500(struct kvm_run *run,
|
|
struct kvm_vcpu *vcpu,
|
|
unsigned int inst, int *advance);
|
|
extern int kvmppc_core_emulate_mtspr_e500(struct kvm_vcpu *vcpu, int sprn,
|
|
ulong spr_val);
|
|
extern int kvmppc_core_emulate_mfspr_e500(struct kvm_vcpu *vcpu, int sprn,
|
|
ulong *spr_val);
|
|
|
|
/*
|
|
* Load up guest vcpu FP state if it's needed.
|
|
* It also set the MSR_FP in thread so that host know
|
|
* we're holding FPU, and then host can help to save
|
|
* guest vcpu FP state if other threads require to use FPU.
|
|
* This simulates an FP unavailable fault.
|
|
*
|
|
* It requires to be called with preemption disabled.
|
|
*/
|
|
static inline void kvmppc_load_guest_fp(struct kvm_vcpu *vcpu)
|
|
{
|
|
#ifdef CONFIG_PPC_FPU
|
|
if (vcpu->fpu_active && !(current->thread.regs->msr & MSR_FP)) {
|
|
enable_kernel_fp();
|
|
load_fp_state(&vcpu->arch.fp);
|
|
current->thread.fp_save_area = &vcpu->arch.fp;
|
|
current->thread.regs->msr |= MSR_FP;
|
|
}
|
|
#endif
|
|
}
|
|
|
|
/*
|
|
* Save guest vcpu FP state into thread.
|
|
* It requires to be called with preemption disabled.
|
|
*/
|
|
static inline void kvmppc_save_guest_fp(struct kvm_vcpu *vcpu)
|
|
{
|
|
#ifdef CONFIG_PPC_FPU
|
|
if (vcpu->fpu_active && (current->thread.regs->msr & MSR_FP))
|
|
giveup_fpu(current);
|
|
current->thread.fp_save_area = NULL;
|
|
#endif
|
|
}
|
|
|
|
static inline void kvmppc_clear_dbsr(void)
|
|
{
|
|
mtspr(SPRN_DBSR, mfspr(SPRN_DBSR));
|
|
}
|
|
#endif /* __KVM_BOOKE_H__ */
|