mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-27 02:55:42 +07:00
159fd7b8d3
Writes to ZCR_EL1 are self-synchronising, and so may be expensive in typical implementations. This patch adopts the approach used for costly system register writes elsewhere in the kernel: the system register write is suppressed if it would not change the stored value. Since the common case will be that of switching between tasks that use the same vector length as one another, prediction hit rates on the conditional branch should be reasonably good, with lower expected amortised cost than the unconditional execution of a heavyweight self-synchronising instruction. Signed-off-by: Dave Martin <Dave.Martin@arm.com> Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
228 lines
5.7 KiB
C
228 lines
5.7 KiB
C
/*
|
|
* FP/SIMD state saving and restoring macros
|
|
*
|
|
* Copyright (C) 2012 ARM Ltd.
|
|
* Author: Catalin Marinas <catalin.marinas@arm.com>
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License version 2 as
|
|
* published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
.macro fpsimd_save state, tmpnr
|
|
stp q0, q1, [\state, #16 * 0]
|
|
stp q2, q3, [\state, #16 * 2]
|
|
stp q4, q5, [\state, #16 * 4]
|
|
stp q6, q7, [\state, #16 * 6]
|
|
stp q8, q9, [\state, #16 * 8]
|
|
stp q10, q11, [\state, #16 * 10]
|
|
stp q12, q13, [\state, #16 * 12]
|
|
stp q14, q15, [\state, #16 * 14]
|
|
stp q16, q17, [\state, #16 * 16]
|
|
stp q18, q19, [\state, #16 * 18]
|
|
stp q20, q21, [\state, #16 * 20]
|
|
stp q22, q23, [\state, #16 * 22]
|
|
stp q24, q25, [\state, #16 * 24]
|
|
stp q26, q27, [\state, #16 * 26]
|
|
stp q28, q29, [\state, #16 * 28]
|
|
stp q30, q31, [\state, #16 * 30]!
|
|
mrs x\tmpnr, fpsr
|
|
str w\tmpnr, [\state, #16 * 2]
|
|
mrs x\tmpnr, fpcr
|
|
str w\tmpnr, [\state, #16 * 2 + 4]
|
|
.endm
|
|
|
|
.macro fpsimd_restore_fpcr state, tmp
|
|
/*
|
|
* Writes to fpcr may be self-synchronising, so avoid restoring
|
|
* the register if it hasn't changed.
|
|
*/
|
|
mrs \tmp, fpcr
|
|
cmp \tmp, \state
|
|
b.eq 9999f
|
|
msr fpcr, \state
|
|
9999:
|
|
.endm
|
|
|
|
/* Clobbers \state */
|
|
.macro fpsimd_restore state, tmpnr
|
|
ldp q0, q1, [\state, #16 * 0]
|
|
ldp q2, q3, [\state, #16 * 2]
|
|
ldp q4, q5, [\state, #16 * 4]
|
|
ldp q6, q7, [\state, #16 * 6]
|
|
ldp q8, q9, [\state, #16 * 8]
|
|
ldp q10, q11, [\state, #16 * 10]
|
|
ldp q12, q13, [\state, #16 * 12]
|
|
ldp q14, q15, [\state, #16 * 14]
|
|
ldp q16, q17, [\state, #16 * 16]
|
|
ldp q18, q19, [\state, #16 * 18]
|
|
ldp q20, q21, [\state, #16 * 20]
|
|
ldp q22, q23, [\state, #16 * 22]
|
|
ldp q24, q25, [\state, #16 * 24]
|
|
ldp q26, q27, [\state, #16 * 26]
|
|
ldp q28, q29, [\state, #16 * 28]
|
|
ldp q30, q31, [\state, #16 * 30]!
|
|
ldr w\tmpnr, [\state, #16 * 2]
|
|
msr fpsr, x\tmpnr
|
|
ldr w\tmpnr, [\state, #16 * 2 + 4]
|
|
fpsimd_restore_fpcr x\tmpnr, \state
|
|
.endm
|
|
|
|
/* Sanity-check macros to help avoid encoding garbage instructions */
|
|
|
|
.macro _check_general_reg nr
|
|
.if (\nr) < 0 || (\nr) > 30
|
|
.error "Bad register number \nr."
|
|
.endif
|
|
.endm
|
|
|
|
.macro _sve_check_zreg znr
|
|
.if (\znr) < 0 || (\znr) > 31
|
|
.error "Bad Scalable Vector Extension vector register number \znr."
|
|
.endif
|
|
.endm
|
|
|
|
.macro _sve_check_preg pnr
|
|
.if (\pnr) < 0 || (\pnr) > 15
|
|
.error "Bad Scalable Vector Extension predicate register number \pnr."
|
|
.endif
|
|
.endm
|
|
|
|
.macro _check_num n, min, max
|
|
.if (\n) < (\min) || (\n) > (\max)
|
|
.error "Number \n out of range [\min,\max]"
|
|
.endif
|
|
.endm
|
|
|
|
/* SVE instruction encodings for non-SVE-capable assemblers */
|
|
|
|
/* STR (vector): STR Z\nz, [X\nxbase, #\offset, MUL VL] */
|
|
.macro _sve_str_v nz, nxbase, offset=0
|
|
_sve_check_zreg \nz
|
|
_check_general_reg \nxbase
|
|
_check_num (\offset), -0x100, 0xff
|
|
.inst 0xe5804000 \
|
|
| (\nz) \
|
|
| ((\nxbase) << 5) \
|
|
| (((\offset) & 7) << 10) \
|
|
| (((\offset) & 0x1f8) << 13)
|
|
.endm
|
|
|
|
/* LDR (vector): LDR Z\nz, [X\nxbase, #\offset, MUL VL] */
|
|
.macro _sve_ldr_v nz, nxbase, offset=0
|
|
_sve_check_zreg \nz
|
|
_check_general_reg \nxbase
|
|
_check_num (\offset), -0x100, 0xff
|
|
.inst 0x85804000 \
|
|
| (\nz) \
|
|
| ((\nxbase) << 5) \
|
|
| (((\offset) & 7) << 10) \
|
|
| (((\offset) & 0x1f8) << 13)
|
|
.endm
|
|
|
|
/* STR (predicate): STR P\np, [X\nxbase, #\offset, MUL VL] */
|
|
.macro _sve_str_p np, nxbase, offset=0
|
|
_sve_check_preg \np
|
|
_check_general_reg \nxbase
|
|
_check_num (\offset), -0x100, 0xff
|
|
.inst 0xe5800000 \
|
|
| (\np) \
|
|
| ((\nxbase) << 5) \
|
|
| (((\offset) & 7) << 10) \
|
|
| (((\offset) & 0x1f8) << 13)
|
|
.endm
|
|
|
|
/* LDR (predicate): LDR P\np, [X\nxbase, #\offset, MUL VL] */
|
|
.macro _sve_ldr_p np, nxbase, offset=0
|
|
_sve_check_preg \np
|
|
_check_general_reg \nxbase
|
|
_check_num (\offset), -0x100, 0xff
|
|
.inst 0x85800000 \
|
|
| (\np) \
|
|
| ((\nxbase) << 5) \
|
|
| (((\offset) & 7) << 10) \
|
|
| (((\offset) & 0x1f8) << 13)
|
|
.endm
|
|
|
|
/* RDVL X\nx, #\imm */
|
|
.macro _sve_rdvl nx, imm
|
|
_check_general_reg \nx
|
|
_check_num (\imm), -0x20, 0x1f
|
|
.inst 0x04bf5000 \
|
|
| (\nx) \
|
|
| (((\imm) & 0x3f) << 5)
|
|
.endm
|
|
|
|
/* RDFFR (unpredicated): RDFFR P\np.B */
|
|
.macro _sve_rdffr np
|
|
_sve_check_preg \np
|
|
.inst 0x2519f000 \
|
|
| (\np)
|
|
.endm
|
|
|
|
/* WRFFR P\np.B */
|
|
.macro _sve_wrffr np
|
|
_sve_check_preg \np
|
|
.inst 0x25289000 \
|
|
| ((\np) << 5)
|
|
.endm
|
|
|
|
.macro __for from:req, to:req
|
|
.if (\from) == (\to)
|
|
_for__body \from
|
|
.else
|
|
__for \from, (\from) + ((\to) - (\from)) / 2
|
|
__for (\from) + ((\to) - (\from)) / 2 + 1, \to
|
|
.endif
|
|
.endm
|
|
|
|
.macro _for var:req, from:req, to:req, insn:vararg
|
|
.macro _for__body \var:req
|
|
\insn
|
|
.endm
|
|
|
|
__for \from, \to
|
|
|
|
.purgem _for__body
|
|
.endm
|
|
|
|
.macro sve_save nxbase, xpfpsr, nxtmp
|
|
_for n, 0, 31, _sve_str_v \n, \nxbase, \n - 34
|
|
_for n, 0, 15, _sve_str_p \n, \nxbase, \n - 16
|
|
_sve_rdffr 0
|
|
_sve_str_p 0, \nxbase
|
|
_sve_ldr_p 0, \nxbase, -16
|
|
|
|
mrs x\nxtmp, fpsr
|
|
str w\nxtmp, [\xpfpsr]
|
|
mrs x\nxtmp, fpcr
|
|
str w\nxtmp, [\xpfpsr, #4]
|
|
.endm
|
|
|
|
.macro sve_load nxbase, xpfpsr, xvqminus1, nxtmp, xtmp2
|
|
mrs_s x\nxtmp, SYS_ZCR_EL1
|
|
bic \xtmp2, x\nxtmp, ZCR_ELx_LEN_MASK
|
|
orr \xtmp2, \xtmp2, \xvqminus1
|
|
cmp \xtmp2, x\nxtmp
|
|
b.eq 921f
|
|
msr_s SYS_ZCR_EL1, \xtmp2 // self-synchronising
|
|
921:
|
|
_for n, 0, 31, _sve_ldr_v \n, \nxbase, \n - 34
|
|
_sve_ldr_p 0, \nxbase
|
|
_sve_wrffr 0
|
|
_for n, 0, 15, _sve_ldr_p \n, \nxbase, \n - 16
|
|
|
|
ldr w\nxtmp, [\xpfpsr]
|
|
msr fpsr, x\nxtmp
|
|
ldr w\nxtmp, [\xpfpsr, #4]
|
|
msr fpcr, x\nxtmp
|
|
.endm
|