2008-07-30 00:29:19 +07:00
|
|
|
#ifndef __ASM_X86_XSAVE_H
|
|
|
|
#define __ASM_X86_XSAVE_H
|
|
|
|
|
2008-07-30 07:23:16 +07:00
|
|
|
#include <linux/types.h>
|
2008-07-30 00:29:19 +07:00
|
|
|
#include <asm/processor.h>
|
|
|
|
|
2010-07-22 00:03:54 +07:00
|
|
|
#define XSTATE_CPUID 0x0000000d
|
2008-07-30 00:29:19 +07:00
|
|
|
|
2014-02-21 04:24:51 +07:00
|
|
|
#define XSTATE_FP 0x1
|
|
|
|
#define XSTATE_SSE 0x2
|
|
|
|
#define XSTATE_YMM 0x4
|
|
|
|
#define XSTATE_BNDREGS 0x8
|
|
|
|
#define XSTATE_BNDCSR 0x10
|
|
|
|
#define XSTATE_OPMASK 0x20
|
|
|
|
#define XSTATE_ZMM_Hi256 0x40
|
|
|
|
#define XSTATE_Hi16_ZMM 0x80
|
2008-07-30 00:29:19 +07:00
|
|
|
|
|
|
|
#define XSTATE_FPSSE (XSTATE_FP | XSTATE_SSE)
|
2014-10-22 16:35:24 +07:00
|
|
|
#define XSTATE_AVX512 (XSTATE_OPMASK | XSTATE_ZMM_Hi256 | XSTATE_Hi16_ZMM)
|
2014-02-22 00:39:02 +07:00
|
|
|
/* Bit 63 of XCR0 is reserved for future expansion */
|
|
|
|
#define XSTATE_EXTEND_MASK (~(XSTATE_FPSSE | (1ULL << 63)))
|
2008-07-30 00:29:19 +07:00
|
|
|
|
|
|
|
#define FXSAVE_SIZE 512
|
|
|
|
|
2010-06-13 16:29:39 +07:00
|
|
|
#define XSAVE_HDR_SIZE 64
|
|
|
|
#define XSAVE_HDR_OFFSET FXSAVE_SIZE
|
|
|
|
|
|
|
|
#define XSAVE_YMM_SIZE 256
|
|
|
|
#define XSAVE_YMM_OFFSET (XSAVE_HDR_SIZE + XSAVE_HDR_OFFSET)
|
2010-05-17 16:22:23 +07:00
|
|
|
|
2013-12-05 16:15:34 +07:00
|
|
|
/* Supported features which support lazy state saving */
|
2014-02-21 04:24:51 +07:00
|
|
|
#define XSTATE_LAZY (XSTATE_FP | XSTATE_SSE | XSTATE_YMM \
|
|
|
|
| XSTATE_OPMASK | XSTATE_ZMM_Hi256 | XSTATE_Hi16_ZMM)
|
2013-12-05 16:15:34 +07:00
|
|
|
|
|
|
|
/* Supported features which require eager state saving */
|
|
|
|
#define XSTATE_EAGER (XSTATE_BNDREGS | XSTATE_BNDCSR)
|
|
|
|
|
|
|
|
/* All currently supported features */
|
|
|
|
#define XCNTXT_MASK (XSTATE_LAZY | XSTATE_EAGER)
|
2008-07-30 00:29:19 +07:00
|
|
|
|
2008-07-30 00:29:20 +07:00
|
|
|
#ifdef CONFIG_X86_64
|
|
|
|
#define REX_PREFIX "0x48, "
|
|
|
|
#else
|
|
|
|
#define REX_PREFIX
|
|
|
|
#endif
|
|
|
|
|
2008-07-30 07:23:16 +07:00
|
|
|
extern unsigned int xstate_size;
|
|
|
|
extern u64 pcntxt_mask;
|
2010-02-12 02:50:59 +07:00
|
|
|
extern u64 xstate_fx_sw_bytes[USER_XSTATE_FX_SW_WORDS];
|
2012-08-25 04:13:02 +07:00
|
|
|
extern struct xsave_struct *init_xstate_buf;
|
2008-07-30 00:29:19 +07:00
|
|
|
|
|
|
|
extern void xsave_init(void);
|
2010-02-12 02:50:59 +07:00
|
|
|
extern void update_regset_xstate_info(unsigned int size, u64 xstate_mask);
|
2008-07-30 00:29:20 +07:00
|
|
|
extern int init_fpu(struct task_struct *child);
|
|
|
|
|
2014-05-30 22:19:21 +07:00
|
|
|
/* These macros all use (%edi)/(%rdi) as the single memory argument. */
|
2014-05-30 01:12:34 +07:00
|
|
|
#define XSAVE ".byte " REX_PREFIX "0x0f,0xae,0x27"
|
|
|
|
#define XSAVEOPT ".byte " REX_PREFIX "0x0f,0xae,0x37"
|
|
|
|
#define XSAVES ".byte " REX_PREFIX "0x0f,0xc7,0x2f"
|
|
|
|
#define XRSTOR ".byte " REX_PREFIX "0x0f,0xae,0x2f"
|
|
|
|
#define XRSTORS ".byte " REX_PREFIX "0x0f,0xc7,0x1f"
|
|
|
|
|
2014-05-30 01:12:35 +07:00
|
|
|
#define xstate_fault ".section .fixup,\"ax\"\n" \
|
|
|
|
"3: movl $-1,%[err]\n" \
|
|
|
|
" jmp 2b\n" \
|
|
|
|
".previous\n" \
|
|
|
|
_ASM_EXTABLE(1b, 3b) \
|
|
|
|
: [err] "=r" (err)
|
|
|
|
|
2014-05-30 01:12:40 +07:00
|
|
|
/*
|
|
|
|
* This function is called only during boot time when x86 caps are not set
|
|
|
|
* up and alternative can not be used yet.
|
|
|
|
*/
|
2014-06-17 17:21:08 +07:00
|
|
|
static inline int xsave_state_booting(struct xsave_struct *fx, u64 mask)
|
2014-05-30 01:12:40 +07:00
|
|
|
{
|
|
|
|
u32 lmask = mask;
|
|
|
|
u32 hmask = mask >> 32;
|
|
|
|
int err = 0;
|
|
|
|
|
|
|
|
WARN_ON(system_state != SYSTEM_BOOTING);
|
|
|
|
|
|
|
|
if (boot_cpu_has(X86_FEATURE_XSAVES))
|
|
|
|
asm volatile("1:"XSAVES"\n\t"
|
|
|
|
"2:\n\t"
|
|
|
|
: : "D" (fx), "m" (*fx), "a" (lmask), "d" (hmask)
|
|
|
|
: "memory");
|
|
|
|
else
|
|
|
|
asm volatile("1:"XSAVE"\n\t"
|
|
|
|
"2:\n\t"
|
|
|
|
: : "D" (fx), "m" (*fx), "a" (lmask), "d" (hmask)
|
|
|
|
: "memory");
|
|
|
|
|
|
|
|
asm volatile(xstate_fault
|
|
|
|
: "0" (0)
|
|
|
|
: "memory");
|
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* This function is called only during boot time when x86 caps are not set
|
|
|
|
* up and alternative can not be used yet.
|
|
|
|
*/
|
|
|
|
static inline int xrstor_state_booting(struct xsave_struct *fx, u64 mask)
|
|
|
|
{
|
|
|
|
u32 lmask = mask;
|
|
|
|
u32 hmask = mask >> 32;
|
|
|
|
int err = 0;
|
|
|
|
|
|
|
|
WARN_ON(system_state != SYSTEM_BOOTING);
|
|
|
|
|
|
|
|
if (boot_cpu_has(X86_FEATURE_XSAVES))
|
|
|
|
asm volatile("1:"XRSTORS"\n\t"
|
|
|
|
"2:\n\t"
|
|
|
|
: : "D" (fx), "m" (*fx), "a" (lmask), "d" (hmask)
|
|
|
|
: "memory");
|
|
|
|
else
|
|
|
|
asm volatile("1:"XRSTOR"\n\t"
|
|
|
|
"2:\n\t"
|
|
|
|
: : "D" (fx), "m" (*fx), "a" (lmask), "d" (hmask)
|
|
|
|
: "memory");
|
|
|
|
|
|
|
|
asm volatile(xstate_fault
|
|
|
|
: "0" (0)
|
|
|
|
: "memory");
|
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2014-05-30 01:12:36 +07:00
|
|
|
/*
|
|
|
|
* Save processor xstate to xsave area.
|
|
|
|
*/
|
|
|
|
static inline int xsave_state(struct xsave_struct *fx, u64 mask)
|
|
|
|
{
|
|
|
|
u32 lmask = mask;
|
|
|
|
u32 hmask = mask >> 32;
|
|
|
|
int err = 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If xsaves is enabled, xsaves replaces xsaveopt because
|
|
|
|
* it supports compact format and supervisor states in addition to
|
|
|
|
* modified optimization in xsaveopt.
|
|
|
|
*
|
|
|
|
* Otherwise, if xsaveopt is enabled, xsaveopt replaces xsave
|
|
|
|
* because xsaveopt supports modified optimization which is not
|
|
|
|
* supported by xsave.
|
|
|
|
*
|
|
|
|
* If none of xsaves and xsaveopt is enabled, use xsave.
|
|
|
|
*/
|
|
|
|
alternative_input_2(
|
|
|
|
"1:"XSAVE,
|
|
|
|
"1:"XSAVEOPT,
|
|
|
|
X86_FEATURE_XSAVEOPT,
|
|
|
|
"1:"XSAVES,
|
|
|
|
X86_FEATURE_XSAVES,
|
|
|
|
[fx] "D" (fx), "a" (lmask), "d" (hmask) :
|
|
|
|
"memory");
|
|
|
|
asm volatile("2:\n\t"
|
|
|
|
xstate_fault
|
|
|
|
: "0" (0)
|
|
|
|
: "memory");
|
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Restore processor xstate from xsave area.
|
|
|
|
*/
|
|
|
|
static inline int xrstor_state(struct xsave_struct *fx, u64 mask)
|
|
|
|
{
|
|
|
|
int err = 0;
|
|
|
|
u32 lmask = mask;
|
|
|
|
u32 hmask = mask >> 32;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Use xrstors to restore context if it is enabled. xrstors supports
|
|
|
|
* compacted format of xsave area which is not supported by xrstor.
|
|
|
|
*/
|
|
|
|
alternative_input(
|
|
|
|
"1: " XRSTOR,
|
|
|
|
"1: " XRSTORS,
|
|
|
|
X86_FEATURE_XSAVES,
|
|
|
|
"D" (fx), "m" (*fx), "a" (lmask), "d" (hmask)
|
|
|
|
: "memory");
|
|
|
|
|
|
|
|
asm volatile("2:\n"
|
|
|
|
xstate_fault
|
|
|
|
: "0" (0)
|
|
|
|
: "memory");
|
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2014-05-30 01:12:37 +07:00
|
|
|
/*
|
|
|
|
* Save xstate context for old process during context switch.
|
|
|
|
*/
|
|
|
|
static inline void fpu_xsave(struct fpu *fpu)
|
2008-07-30 00:29:20 +07:00
|
|
|
{
|
2014-05-30 01:12:37 +07:00
|
|
|
xsave_state(&fpu->state->xsave, -1);
|
|
|
|
}
|
2008-07-30 00:29:20 +07:00
|
|
|
|
2014-05-30 01:12:37 +07:00
|
|
|
/*
|
|
|
|
* Restore xstate context for new process during context switch.
|
|
|
|
*/
|
|
|
|
static inline int fpu_xrstor_checking(struct xsave_struct *fx)
|
|
|
|
{
|
|
|
|
return xrstor_state(fx, -1);
|
2008-07-30 00:29:20 +07:00
|
|
|
}
|
|
|
|
|
2014-05-30 01:12:38 +07:00
|
|
|
/*
|
|
|
|
* Save xstate to user space xsave area.
|
|
|
|
*
|
|
|
|
* We don't use modified optimization because xrstor/xrstors might track
|
|
|
|
* a different application.
|
|
|
|
*
|
|
|
|
* We don't use compacted format xsave area for
|
|
|
|
* backward compatibility for old applications which don't understand
|
|
|
|
* compacted format of xsave area.
|
|
|
|
*/
|
2008-07-30 00:29:25 +07:00
|
|
|
static inline int xsave_user(struct xsave_struct __user *buf)
|
2008-07-30 00:29:23 +07:00
|
|
|
{
|
|
|
|
int err;
|
2010-06-23 06:23:37 +07:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Clear the xsave header first, so that reserved fields are
|
|
|
|
* initialized to zero.
|
|
|
|
*/
|
x86, fpu: Unify signal handling code paths for x86 and x86_64 kernels
Currently for x86 and x86_32 binaries, fpstate in the user sigframe is copied
to/from the fpstate in the task struct.
And in the case of signal delivery for x86_64 binaries, if the fpstate is live
in the CPU registers, then the live state is copied directly to the user
sigframe. Otherwise fpstate in the task struct is copied to the user sigframe.
During restore, fpstate in the user sigframe is restored directly to the live
CPU registers.
Historically, different code paths led to different bugs. For example,
x86_64 code path was not preemption safe till recently. Also there is lot
of code duplication for support of new features like xsave etc.
Unify signal handling code paths for x86 and x86_64 kernels.
New strategy is as follows:
Signal delivery: Both for 32/64-bit frames, align the core math frame area to
64bytes as needed by xsave (this where the main fpu/extended state gets copied
to and excludes the legacy compatibility fsave header for the 32-bit [f]xsave
frames). If the state is live, copy the register state directly to the user
frame. If not live, copy the state in the thread struct to the user frame. And
for 32-bit [f]xsave frames, construct the fsave header separately before
the actual [f]xsave area.
Signal return: As the 32-bit frames with [f]xstate has an additional
'fsave' header, copy everything back from the user sigframe to the
fpstate in the task structure and reconstruct the fxstate from the 'fsave'
header (Also user passed pointers may not be correctly aligned for
any attempt to directly restore any partial state). At the next fpstate usage,
everything will be restored to the live CPU registers.
For all the 64-bit frames and the 32-bit fsave frame, restore the state from
the user sigframe directly to the live CPU registers. 64-bit signals always
restored the math frame directly, so we can expect the math frame pointer
to be correctly aligned. For 32-bit fsave frames, there are no alignment
requirements, so we can restore the state directly.
"lat_sig catch" microbenchmark numbers (for x86, x86_64, x86_32 binaries) are
with in the noise range with this change.
Signed-off-by: Suresh Siddha <suresh.b.siddha@intel.com>
Link: http://lkml.kernel.org/r/1343171129-2747-4-git-send-email-suresh.b.siddha@intel.com
[ Merged in compilation fix ]
Link: http://lkml.kernel.org/r/1344544736.8326.17.camel@sbsiddha-desk.sc.intel.com
Signed-off-by: H. Peter Anvin <hpa@linux.intel.com>
2012-07-25 06:05:29 +07:00
|
|
|
err = __clear_user(&buf->xsave_hdr, sizeof(buf->xsave_hdr));
|
2010-06-23 06:23:37 +07:00
|
|
|
if (unlikely(err))
|
|
|
|
return -EFAULT;
|
|
|
|
|
2012-09-22 02:43:12 +07:00
|
|
|
__asm__ __volatile__(ASM_STAC "\n"
|
2014-05-30 01:12:38 +07:00
|
|
|
"1:"XSAVE"\n"
|
2012-09-22 02:43:12 +07:00
|
|
|
"2: " ASM_CLAC "\n"
|
2014-05-30 01:12:38 +07:00
|
|
|
xstate_fault
|
2008-07-30 00:29:23 +07:00
|
|
|
: "D" (buf), "a" (-1), "d" (-1), "0" (0)
|
|
|
|
: "memory");
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2014-05-30 01:12:38 +07:00
|
|
|
/*
|
|
|
|
* Restore xstate from user space xsave area.
|
|
|
|
*/
|
2008-07-30 07:23:16 +07:00
|
|
|
static inline int xrestore_user(struct xsave_struct __user *buf, u64 mask)
|
2008-07-30 00:29:23 +07:00
|
|
|
{
|
2014-05-30 01:12:38 +07:00
|
|
|
int err = 0;
|
2008-07-30 00:29:23 +07:00
|
|
|
struct xsave_struct *xstate = ((__force struct xsave_struct *)buf);
|
2008-07-30 07:23:16 +07:00
|
|
|
u32 lmask = mask;
|
|
|
|
u32 hmask = mask >> 32;
|
2008-07-30 00:29:23 +07:00
|
|
|
|
2012-09-22 02:43:12 +07:00
|
|
|
__asm__ __volatile__(ASM_STAC "\n"
|
2014-05-30 01:12:38 +07:00
|
|
|
"1:"XRSTOR"\n"
|
2012-09-22 02:43:12 +07:00
|
|
|
"2: " ASM_CLAC "\n"
|
2014-05-30 01:12:38 +07:00
|
|
|
xstate_fault
|
2008-07-30 00:29:23 +07:00
|
|
|
: "D" (xstate), "a" (lmask), "d" (hmask), "0" (0)
|
|
|
|
: "memory"); /* memory required? */
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2014-05-30 01:12:44 +07:00
|
|
|
void *get_xsave_addr(struct xsave_struct *xsave, int xstate);
|
|
|
|
void setup_xstate_comp(void);
|
|
|
|
|
2008-07-30 00:29:19 +07:00
|
|
|
#endif
|