mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-24 02:35:16 +07:00
6f515d842e
This patch adds the ability to be able to save the VEC registers to the thread struct without giving up (disabling the facility) next time the process returns to userspace. This patch builds on a previous optimisation for the FPU registers in the thread copy path to avoid a possibly pointless reload of VEC state. Signed-off-by: Cyril Bur <cyrilbur@gmail.com> Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
99 lines
2.7 KiB
C
99 lines
2.7 KiB
C
/*
|
|
* Copyright (C) 1999 Cort Dougan <cort@cs.nmt.edu>
|
|
*/
|
|
#ifndef _ASM_POWERPC_SWITCH_TO_H
|
|
#define _ASM_POWERPC_SWITCH_TO_H
|
|
|
|
#include <asm/reg.h>
|
|
|
|
struct thread_struct;
|
|
struct task_struct;
|
|
struct pt_regs;
|
|
|
|
extern struct task_struct *__switch_to(struct task_struct *,
|
|
struct task_struct *);
|
|
#define switch_to(prev, next, last) ((last) = __switch_to((prev), (next)))
|
|
|
|
extern struct task_struct *_switch(struct thread_struct *prev,
|
|
struct thread_struct *next);
|
|
|
|
extern void switch_booke_debug_regs(struct debug_reg *new_debug);
|
|
|
|
extern int emulate_altivec(struct pt_regs *);
|
|
|
|
extern void flush_all_to_thread(struct task_struct *);
|
|
extern void giveup_all(struct task_struct *);
|
|
|
|
#ifdef CONFIG_PPC_FPU
|
|
extern void enable_kernel_fp(void);
|
|
extern void flush_fp_to_thread(struct task_struct *);
|
|
extern void giveup_fpu(struct task_struct *);
|
|
extern void save_fpu(struct task_struct *);
|
|
static inline void disable_kernel_fp(void)
|
|
{
|
|
msr_check_and_clear(MSR_FP);
|
|
}
|
|
#else
|
|
static inline void __giveup_fpu(struct task_struct *t) { }
|
|
static inline void save_fpu(struct task_struct *t) { }
|
|
static inline void flush_fp_to_thread(struct task_struct *t) { }
|
|
#endif
|
|
|
|
#ifdef CONFIG_ALTIVEC
|
|
extern void enable_kernel_altivec(void);
|
|
extern void flush_altivec_to_thread(struct task_struct *);
|
|
extern void giveup_altivec(struct task_struct *);
|
|
extern void save_altivec(struct task_struct *);
|
|
static inline void disable_kernel_altivec(void)
|
|
{
|
|
msr_check_and_clear(MSR_VEC);
|
|
}
|
|
#else
|
|
static inline void save_altivec(struct task_struct *t) { }
|
|
static inline void __giveup_altivec(struct task_struct *t) { }
|
|
#endif
|
|
|
|
#ifdef CONFIG_VSX
|
|
extern void enable_kernel_vsx(void);
|
|
extern void flush_vsx_to_thread(struct task_struct *);
|
|
extern void giveup_vsx(struct task_struct *);
|
|
extern void __giveup_vsx(struct task_struct *);
|
|
static inline void disable_kernel_vsx(void)
|
|
{
|
|
msr_check_and_clear(MSR_FP|MSR_VEC|MSR_VSX);
|
|
}
|
|
#else
|
|
static inline void __giveup_vsx(struct task_struct *t) { }
|
|
#endif
|
|
|
|
#ifdef CONFIG_SPE
|
|
extern void enable_kernel_spe(void);
|
|
extern void flush_spe_to_thread(struct task_struct *);
|
|
extern void giveup_spe(struct task_struct *);
|
|
extern void __giveup_spe(struct task_struct *);
|
|
static inline void disable_kernel_spe(void)
|
|
{
|
|
msr_check_and_clear(MSR_SPE);
|
|
}
|
|
#else
|
|
static inline void __giveup_spe(struct task_struct *t) { }
|
|
#endif
|
|
|
|
static inline void clear_task_ebb(struct task_struct *t)
|
|
{
|
|
#ifdef CONFIG_PPC_BOOK3S_64
|
|
/* EBB perf events are not inherited, so clear all EBB state. */
|
|
t->thread.ebbrr = 0;
|
|
t->thread.ebbhr = 0;
|
|
t->thread.bescr = 0;
|
|
t->thread.mmcr2 = 0;
|
|
t->thread.mmcr0 = 0;
|
|
t->thread.siar = 0;
|
|
t->thread.sdar = 0;
|
|
t->thread.sier = 0;
|
|
t->thread.used_ebb = 0;
|
|
#endif
|
|
}
|
|
|
|
#endif /* _ASM_POWERPC_SWITCH_TO_H */
|