mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-19 06:08:27 +07:00
1fc5dce78a
Manipulating the SVE architectural state, including the vector and predicate registers, first-fault register and the vector length, requires the use of dedicated instructions added by SVE. This patch adds suitable assembly functions for saving and restoring the SVE registers and querying the vector length. Setting of the vector length is done as part of register restore. Since people building kernels may not all get an SVE-enabled toolchain for a while, this patch uses macros that generate explicit opcodes in place of assembler mnemonics. Signed-off-by: Dave Martin <Dave.Martin@arm.com> Reviewed-by: Alex Bennée <alex.bennee@linaro.org> Acked-by: Catalin Marinas <catalin.marinas@arm.com> Signed-off-by: Will Deacon <will.deacon@arm.com>
226 lines
5.6 KiB
C
226 lines
5.6 KiB
C
/*
|
|
* FP/SIMD state saving and restoring macros
|
|
*
|
|
* Copyright (C) 2012 ARM Ltd.
|
|
* Author: Catalin Marinas <catalin.marinas@arm.com>
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License version 2 as
|
|
* published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
.macro fpsimd_save state, tmpnr
|
|
stp q0, q1, [\state, #16 * 0]
|
|
stp q2, q3, [\state, #16 * 2]
|
|
stp q4, q5, [\state, #16 * 4]
|
|
stp q6, q7, [\state, #16 * 6]
|
|
stp q8, q9, [\state, #16 * 8]
|
|
stp q10, q11, [\state, #16 * 10]
|
|
stp q12, q13, [\state, #16 * 12]
|
|
stp q14, q15, [\state, #16 * 14]
|
|
stp q16, q17, [\state, #16 * 16]
|
|
stp q18, q19, [\state, #16 * 18]
|
|
stp q20, q21, [\state, #16 * 20]
|
|
stp q22, q23, [\state, #16 * 22]
|
|
stp q24, q25, [\state, #16 * 24]
|
|
stp q26, q27, [\state, #16 * 26]
|
|
stp q28, q29, [\state, #16 * 28]
|
|
stp q30, q31, [\state, #16 * 30]!
|
|
mrs x\tmpnr, fpsr
|
|
str w\tmpnr, [\state, #16 * 2]
|
|
mrs x\tmpnr, fpcr
|
|
str w\tmpnr, [\state, #16 * 2 + 4]
|
|
.endm
|
|
|
|
.macro fpsimd_restore_fpcr state, tmp
|
|
/*
|
|
* Writes to fpcr may be self-synchronising, so avoid restoring
|
|
* the register if it hasn't changed.
|
|
*/
|
|
mrs \tmp, fpcr
|
|
cmp \tmp, \state
|
|
b.eq 9999f
|
|
msr fpcr, \state
|
|
9999:
|
|
.endm
|
|
|
|
/* Clobbers \state */
|
|
.macro fpsimd_restore state, tmpnr
|
|
ldp q0, q1, [\state, #16 * 0]
|
|
ldp q2, q3, [\state, #16 * 2]
|
|
ldp q4, q5, [\state, #16 * 4]
|
|
ldp q6, q7, [\state, #16 * 6]
|
|
ldp q8, q9, [\state, #16 * 8]
|
|
ldp q10, q11, [\state, #16 * 10]
|
|
ldp q12, q13, [\state, #16 * 12]
|
|
ldp q14, q15, [\state, #16 * 14]
|
|
ldp q16, q17, [\state, #16 * 16]
|
|
ldp q18, q19, [\state, #16 * 18]
|
|
ldp q20, q21, [\state, #16 * 20]
|
|
ldp q22, q23, [\state, #16 * 22]
|
|
ldp q24, q25, [\state, #16 * 24]
|
|
ldp q26, q27, [\state, #16 * 26]
|
|
ldp q28, q29, [\state, #16 * 28]
|
|
ldp q30, q31, [\state, #16 * 30]!
|
|
ldr w\tmpnr, [\state, #16 * 2]
|
|
msr fpsr, x\tmpnr
|
|
ldr w\tmpnr, [\state, #16 * 2 + 4]
|
|
fpsimd_restore_fpcr x\tmpnr, \state
|
|
.endm
|
|
|
|
/* Sanity-check macros to help avoid encoding garbage instructions */
|
|
|
|
.macro _check_general_reg nr
|
|
.if (\nr) < 0 || (\nr) > 30
|
|
.error "Bad register number \nr."
|
|
.endif
|
|
.endm
|
|
|
|
.macro _sve_check_zreg znr
|
|
.if (\znr) < 0 || (\znr) > 31
|
|
.error "Bad Scalable Vector Extension vector register number \znr."
|
|
.endif
|
|
.endm
|
|
|
|
.macro _sve_check_preg pnr
|
|
.if (\pnr) < 0 || (\pnr) > 15
|
|
.error "Bad Scalable Vector Extension predicate register number \pnr."
|
|
.endif
|
|
.endm
|
|
|
|
.macro _check_num n, min, max
|
|
.if (\n) < (\min) || (\n) > (\max)
|
|
.error "Number \n out of range [\min,\max]"
|
|
.endif
|
|
.endm
|
|
|
|
/* SVE instruction encodings for non-SVE-capable assemblers */
|
|
|
|
/* STR (vector): STR Z\nz, [X\nxbase, #\offset, MUL VL] */
|
|
.macro _sve_str_v nz, nxbase, offset=0
|
|
_sve_check_zreg \nz
|
|
_check_general_reg \nxbase
|
|
_check_num (\offset), -0x100, 0xff
|
|
.inst 0xe5804000 \
|
|
| (\nz) \
|
|
| ((\nxbase) << 5) \
|
|
| (((\offset) & 7) << 10) \
|
|
| (((\offset) & 0x1f8) << 13)
|
|
.endm
|
|
|
|
/* LDR (vector): LDR Z\nz, [X\nxbase, #\offset, MUL VL] */
|
|
.macro _sve_ldr_v nz, nxbase, offset=0
|
|
_sve_check_zreg \nz
|
|
_check_general_reg \nxbase
|
|
_check_num (\offset), -0x100, 0xff
|
|
.inst 0x85804000 \
|
|
| (\nz) \
|
|
| ((\nxbase) << 5) \
|
|
| (((\offset) & 7) << 10) \
|
|
| (((\offset) & 0x1f8) << 13)
|
|
.endm
|
|
|
|
/* STR (predicate): STR P\np, [X\nxbase, #\offset, MUL VL] */
|
|
.macro _sve_str_p np, nxbase, offset=0
|
|
_sve_check_preg \np
|
|
_check_general_reg \nxbase
|
|
_check_num (\offset), -0x100, 0xff
|
|
.inst 0xe5800000 \
|
|
| (\np) \
|
|
| ((\nxbase) << 5) \
|
|
| (((\offset) & 7) << 10) \
|
|
| (((\offset) & 0x1f8) << 13)
|
|
.endm
|
|
|
|
/* LDR (predicate): LDR P\np, [X\nxbase, #\offset, MUL VL] */
|
|
.macro _sve_ldr_p np, nxbase, offset=0
|
|
_sve_check_preg \np
|
|
_check_general_reg \nxbase
|
|
_check_num (\offset), -0x100, 0xff
|
|
.inst 0x85800000 \
|
|
| (\np) \
|
|
| ((\nxbase) << 5) \
|
|
| (((\offset) & 7) << 10) \
|
|
| (((\offset) & 0x1f8) << 13)
|
|
.endm
|
|
|
|
/* RDVL X\nx, #\imm */
|
|
.macro _sve_rdvl nx, imm
|
|
_check_general_reg \nx
|
|
_check_num (\imm), -0x20, 0x1f
|
|
.inst 0x04bf5000 \
|
|
| (\nx) \
|
|
| (((\imm) & 0x3f) << 5)
|
|
.endm
|
|
|
|
/* RDFFR (unpredicated): RDFFR P\np.B */
|
|
.macro _sve_rdffr np
|
|
_sve_check_preg \np
|
|
.inst 0x2519f000 \
|
|
| (\np)
|
|
.endm
|
|
|
|
/* WRFFR P\np.B */
|
|
.macro _sve_wrffr np
|
|
_sve_check_preg \np
|
|
.inst 0x25289000 \
|
|
| ((\np) << 5)
|
|
.endm
|
|
|
|
.macro __for from:req, to:req
|
|
.if (\from) == (\to)
|
|
_for__body \from
|
|
.else
|
|
__for \from, (\from) + ((\to) - (\from)) / 2
|
|
__for (\from) + ((\to) - (\from)) / 2 + 1, \to
|
|
.endif
|
|
.endm
|
|
|
|
.macro _for var:req, from:req, to:req, insn:vararg
|
|
.macro _for__body \var:req
|
|
\insn
|
|
.endm
|
|
|
|
__for \from, \to
|
|
|
|
.purgem _for__body
|
|
.endm
|
|
|
|
.macro sve_save nxbase, xpfpsr, nxtmp
|
|
_for n, 0, 31, _sve_str_v \n, \nxbase, \n - 34
|
|
_for n, 0, 15, _sve_str_p \n, \nxbase, \n - 16
|
|
_sve_rdffr 0
|
|
_sve_str_p 0, \nxbase
|
|
_sve_ldr_p 0, \nxbase, -16
|
|
|
|
mrs x\nxtmp, fpsr
|
|
str w\nxtmp, [\xpfpsr]
|
|
mrs x\nxtmp, fpcr
|
|
str w\nxtmp, [\xpfpsr, #4]
|
|
.endm
|
|
|
|
.macro sve_load nxbase, xpfpsr, xvqminus1, nxtmp
|
|
mrs_s x\nxtmp, SYS_ZCR_EL1
|
|
bic x\nxtmp, x\nxtmp, ZCR_ELx_LEN_MASK
|
|
orr x\nxtmp, x\nxtmp, \xvqminus1
|
|
msr_s SYS_ZCR_EL1, x\nxtmp // self-synchronising
|
|
|
|
_for n, 0, 31, _sve_ldr_v \n, \nxbase, \n - 34
|
|
_sve_ldr_p 0, \nxbase
|
|
_sve_wrffr 0
|
|
_for n, 0, 15, _sve_ldr_p \n, \nxbase, \n - 16
|
|
|
|
ldr w\nxtmp, [\xpfpsr]
|
|
msr fpsr, x\nxtmp
|
|
ldr w\nxtmp, [\xpfpsr, #4]
|
|
msr fpcr, x\nxtmp
|
|
.endm
|