mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-22 18:32:14 +07:00
2874c5fd28
Based on 1 normalized pattern(s): this program is free software you can redistribute it and or modify it under the terms of the gnu general public license as published by the free software foundation either version 2 of the license or at your option any later version extracted by the scancode license scanner the SPDX license identifier GPL-2.0-or-later has been chosen to replace the boilerplate/reference in 3029 file(s). Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Reviewed-by: Allison Randal <allison@lohutok.net> Cc: linux-spdx@vger.kernel.org Link: https://lkml.kernel.org/r/20190527070032.746973796@linutronix.de Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
176 lines
5.0 KiB
ArmAsm
176 lines
5.0 KiB
ArmAsm
/* SPDX-License-Identifier: GPL-2.0-or-later */
|
|
/*
|
|
* Shared glue code for 128bit block ciphers, AVX2 assembler macros
|
|
*
|
|
* Copyright © 2012-2013 Jussi Kivilinna <jussi.kivilinna@mbnet.fi>
|
|
*/
|
|
|
|
#define load_16way(src, x0, x1, x2, x3, x4, x5, x6, x7) \
|
|
vmovdqu (0*32)(src), x0; \
|
|
vmovdqu (1*32)(src), x1; \
|
|
vmovdqu (2*32)(src), x2; \
|
|
vmovdqu (3*32)(src), x3; \
|
|
vmovdqu (4*32)(src), x4; \
|
|
vmovdqu (5*32)(src), x5; \
|
|
vmovdqu (6*32)(src), x6; \
|
|
vmovdqu (7*32)(src), x7;
|
|
|
|
#define store_16way(dst, x0, x1, x2, x3, x4, x5, x6, x7) \
|
|
vmovdqu x0, (0*32)(dst); \
|
|
vmovdqu x1, (1*32)(dst); \
|
|
vmovdqu x2, (2*32)(dst); \
|
|
vmovdqu x3, (3*32)(dst); \
|
|
vmovdqu x4, (4*32)(dst); \
|
|
vmovdqu x5, (5*32)(dst); \
|
|
vmovdqu x6, (6*32)(dst); \
|
|
vmovdqu x7, (7*32)(dst);
|
|
|
|
#define store_cbc_16way(src, dst, x0, x1, x2, x3, x4, x5, x6, x7, t0) \
|
|
vpxor t0, t0, t0; \
|
|
vinserti128 $1, (src), t0, t0; \
|
|
vpxor t0, x0, x0; \
|
|
vpxor (0*32+16)(src), x1, x1; \
|
|
vpxor (1*32+16)(src), x2, x2; \
|
|
vpxor (2*32+16)(src), x3, x3; \
|
|
vpxor (3*32+16)(src), x4, x4; \
|
|
vpxor (4*32+16)(src), x5, x5; \
|
|
vpxor (5*32+16)(src), x6, x6; \
|
|
vpxor (6*32+16)(src), x7, x7; \
|
|
store_16way(dst, x0, x1, x2, x3, x4, x5, x6, x7);
|
|
|
|
#define inc_le128(x, minus_one, tmp) \
|
|
vpcmpeqq minus_one, x, tmp; \
|
|
vpsubq minus_one, x, x; \
|
|
vpslldq $8, tmp, tmp; \
|
|
vpsubq tmp, x, x;
|
|
|
|
#define add2_le128(x, minus_one, minus_two, tmp1, tmp2) \
|
|
vpcmpeqq minus_one, x, tmp1; \
|
|
vpcmpeqq minus_two, x, tmp2; \
|
|
vpsubq minus_two, x, x; \
|
|
vpor tmp2, tmp1, tmp1; \
|
|
vpslldq $8, tmp1, tmp1; \
|
|
vpsubq tmp1, x, x;
|
|
|
|
#define load_ctr_16way(iv, bswap, x0, x1, x2, x3, x4, x5, x6, x7, t0, t0x, t1, \
|
|
t1x, t2, t2x, t3, t3x, t4, t5) \
|
|
vpcmpeqd t0, t0, t0; \
|
|
vpsrldq $8, t0, t0; /* ab: -1:0 ; cd: -1:0 */ \
|
|
vpaddq t0, t0, t4; /* ab: -2:0 ; cd: -2:0 */\
|
|
\
|
|
/* load IV and byteswap */ \
|
|
vmovdqu (iv), t2x; \
|
|
vmovdqa t2x, t3x; \
|
|
inc_le128(t2x, t0x, t1x); \
|
|
vbroadcasti128 bswap, t1; \
|
|
vinserti128 $1, t2x, t3, t2; /* ab: le0 ; cd: le1 */ \
|
|
vpshufb t1, t2, x0; \
|
|
\
|
|
/* construct IVs */ \
|
|
add2_le128(t2, t0, t4, t3, t5); /* ab: le2 ; cd: le3 */ \
|
|
vpshufb t1, t2, x1; \
|
|
add2_le128(t2, t0, t4, t3, t5); \
|
|
vpshufb t1, t2, x2; \
|
|
add2_le128(t2, t0, t4, t3, t5); \
|
|
vpshufb t1, t2, x3; \
|
|
add2_le128(t2, t0, t4, t3, t5); \
|
|
vpshufb t1, t2, x4; \
|
|
add2_le128(t2, t0, t4, t3, t5); \
|
|
vpshufb t1, t2, x5; \
|
|
add2_le128(t2, t0, t4, t3, t5); \
|
|
vpshufb t1, t2, x6; \
|
|
add2_le128(t2, t0, t4, t3, t5); \
|
|
vpshufb t1, t2, x7; \
|
|
vextracti128 $1, t2, t2x; \
|
|
inc_le128(t2x, t0x, t3x); \
|
|
vmovdqu t2x, (iv);
|
|
|
|
#define store_ctr_16way(src, dst, x0, x1, x2, x3, x4, x5, x6, x7) \
|
|
vpxor (0*32)(src), x0, x0; \
|
|
vpxor (1*32)(src), x1, x1; \
|
|
vpxor (2*32)(src), x2, x2; \
|
|
vpxor (3*32)(src), x3, x3; \
|
|
vpxor (4*32)(src), x4, x4; \
|
|
vpxor (5*32)(src), x5, x5; \
|
|
vpxor (6*32)(src), x6, x6; \
|
|
vpxor (7*32)(src), x7, x7; \
|
|
store_16way(dst, x0, x1, x2, x3, x4, x5, x6, x7);
|
|
|
|
#define gf128mul_x_ble(iv, mask, tmp) \
|
|
vpsrad $31, iv, tmp; \
|
|
vpaddq iv, iv, iv; \
|
|
vpshufd $0x13, tmp, tmp; \
|
|
vpand mask, tmp, tmp; \
|
|
vpxor tmp, iv, iv;
|
|
|
|
#define gf128mul_x2_ble(iv, mask1, mask2, tmp0, tmp1) \
|
|
vpsrad $31, iv, tmp0; \
|
|
vpaddq iv, iv, tmp1; \
|
|
vpsllq $2, iv, iv; \
|
|
vpshufd $0x13, tmp0, tmp0; \
|
|
vpsrad $31, tmp1, tmp1; \
|
|
vpand mask2, tmp0, tmp0; \
|
|
vpshufd $0x13, tmp1, tmp1; \
|
|
vpxor tmp0, iv, iv; \
|
|
vpand mask1, tmp1, tmp1; \
|
|
vpxor tmp1, iv, iv;
|
|
|
|
#define load_xts_16way(iv, src, dst, x0, x1, x2, x3, x4, x5, x6, x7, tiv, \
|
|
tivx, t0, t0x, t1, t1x, t2, t2x, t3, \
|
|
xts_gf128mul_and_shl1_mask_0, \
|
|
xts_gf128mul_and_shl1_mask_1) \
|
|
vbroadcasti128 xts_gf128mul_and_shl1_mask_0, t1; \
|
|
\
|
|
/* load IV and construct second IV */ \
|
|
vmovdqu (iv), tivx; \
|
|
vmovdqa tivx, t0x; \
|
|
gf128mul_x_ble(tivx, t1x, t2x); \
|
|
vbroadcasti128 xts_gf128mul_and_shl1_mask_1, t2; \
|
|
vinserti128 $1, tivx, t0, tiv; \
|
|
vpxor (0*32)(src), tiv, x0; \
|
|
vmovdqu tiv, (0*32)(dst); \
|
|
\
|
|
/* construct and store IVs, also xor with source */ \
|
|
gf128mul_x2_ble(tiv, t1, t2, t0, t3); \
|
|
vpxor (1*32)(src), tiv, x1; \
|
|
vmovdqu tiv, (1*32)(dst); \
|
|
\
|
|
gf128mul_x2_ble(tiv, t1, t2, t0, t3); \
|
|
vpxor (2*32)(src), tiv, x2; \
|
|
vmovdqu tiv, (2*32)(dst); \
|
|
\
|
|
gf128mul_x2_ble(tiv, t1, t2, t0, t3); \
|
|
vpxor (3*32)(src), tiv, x3; \
|
|
vmovdqu tiv, (3*32)(dst); \
|
|
\
|
|
gf128mul_x2_ble(tiv, t1, t2, t0, t3); \
|
|
vpxor (4*32)(src), tiv, x4; \
|
|
vmovdqu tiv, (4*32)(dst); \
|
|
\
|
|
gf128mul_x2_ble(tiv, t1, t2, t0, t3); \
|
|
vpxor (5*32)(src), tiv, x5; \
|
|
vmovdqu tiv, (5*32)(dst); \
|
|
\
|
|
gf128mul_x2_ble(tiv, t1, t2, t0, t3); \
|
|
vpxor (6*32)(src), tiv, x6; \
|
|
vmovdqu tiv, (6*32)(dst); \
|
|
\
|
|
gf128mul_x2_ble(tiv, t1, t2, t0, t3); \
|
|
vpxor (7*32)(src), tiv, x7; \
|
|
vmovdqu tiv, (7*32)(dst); \
|
|
\
|
|
vextracti128 $1, tiv, tivx; \
|
|
gf128mul_x_ble(tivx, t1x, t2x); \
|
|
vmovdqu tivx, (iv);
|
|
|
|
#define store_xts_16way(dst, x0, x1, x2, x3, x4, x5, x6, x7) \
|
|
vpxor (0*32)(dst), x0, x0; \
|
|
vpxor (1*32)(dst), x1, x1; \
|
|
vpxor (2*32)(dst), x2, x2; \
|
|
vpxor (3*32)(dst), x3, x3; \
|
|
vpxor (4*32)(dst), x4, x4; \
|
|
vpxor (5*32)(dst), x5, x5; \
|
|
vpxor (6*32)(dst), x6, x6; \
|
|
vpxor (7*32)(dst), x7, x7; \
|
|
store_16way(dst, x0, x1, x2, x3, x4, x5, x6, x7);
|