mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-04 11:16:48 +07:00
cf1521a1a5
Patch adds AVX2/x86-64 implementation of Twofish cipher, requiring 16 parallel blocks for input (256 bytes). Table look-ups are performed using vpgatherdd instruction directly from vector registers and thus should be faster than earlier implementations. Implementation also uses 256-bit wide YMM registers, which should give additional speed up compared to the AVX implementation. Signed-off-by: Jussi Kivilinna <jussi.kivilinna@iki.fi> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
601 lines
13 KiB
ArmAsm
601 lines
13 KiB
ArmAsm
/*
|
||
* x86_64/AVX2 assembler optimized version of Twofish
|
||
*
|
||
* Copyright © 2012-2013 Jussi Kivilinna <jussi.kivilinna@iki.fi>
|
||
*
|
||
* This program is free software; you can redistribute it and/or modify
|
||
* it under the terms of the GNU General Public License as published by
|
||
* the Free Software Foundation; either version 2 of the License, or
|
||
* (at your option) any later version.
|
||
*
|
||
*/
|
||
|
||
#include <linux/linkage.h>
|
||
#include "glue_helper-asm-avx2.S"
|
||
|
||
.file "twofish-avx2-asm_64.S"
|
||
|
||
.data
|
||
.align 16
|
||
|
||
.Lvpshufb_mask0:
|
||
.long 0x80808000
|
||
.long 0x80808004
|
||
.long 0x80808008
|
||
.long 0x8080800c
|
||
|
||
.Lbswap128_mask:
|
||
.byte 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0
|
||
.Lxts_gf128mul_and_shl1_mask_0:
|
||
.byte 0x87, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0
|
||
.Lxts_gf128mul_and_shl1_mask_1:
|
||
.byte 0x0e, 1, 0, 0, 0, 0, 0, 0, 2, 0, 0, 0, 0, 0, 0, 0
|
||
|
||
.text
|
||
|
||
/* structure of crypto context */
|
||
#define s0 0
|
||
#define s1 1024
|
||
#define s2 2048
|
||
#define s3 3072
|
||
#define w 4096
|
||
#define k 4128
|
||
|
||
/* register macros */
|
||
#define CTX %rdi
|
||
|
||
#define RS0 CTX
|
||
#define RS1 %r8
|
||
#define RS2 %r9
|
||
#define RS3 %r10
|
||
#define RK %r11
|
||
#define RW %rax
|
||
#define RROUND %r12
|
||
#define RROUNDd %r12d
|
||
|
||
#define RA0 %ymm8
|
||
#define RB0 %ymm9
|
||
#define RC0 %ymm10
|
||
#define RD0 %ymm11
|
||
#define RA1 %ymm12
|
||
#define RB1 %ymm13
|
||
#define RC1 %ymm14
|
||
#define RD1 %ymm15
|
||
|
||
/* temp regs */
|
||
#define RX0 %ymm0
|
||
#define RY0 %ymm1
|
||
#define RX1 %ymm2
|
||
#define RY1 %ymm3
|
||
#define RT0 %ymm4
|
||
#define RIDX %ymm5
|
||
|
||
#define RX0x %xmm0
|
||
#define RY0x %xmm1
|
||
#define RX1x %xmm2
|
||
#define RY1x %xmm3
|
||
#define RT0x %xmm4
|
||
|
||
/* vpgatherdd mask and '-1' */
|
||
#define RNOT %ymm6
|
||
|
||
/* byte mask, (-1 >> 24) */
|
||
#define RBYTE %ymm7
|
||
|
||
/**********************************************************************
|
||
16-way AVX2 twofish
|
||
**********************************************************************/
|
||
#define init_round_constants() \
|
||
vpcmpeqd RNOT, RNOT, RNOT; \
|
||
vpsrld $24, RNOT, RBYTE; \
|
||
leaq k(CTX), RK; \
|
||
leaq w(CTX), RW; \
|
||
leaq s1(CTX), RS1; \
|
||
leaq s2(CTX), RS2; \
|
||
leaq s3(CTX), RS3; \
|
||
|
||
#define g16(ab, rs0, rs1, rs2, rs3, xy) \
|
||
vpand RBYTE, ab ## 0, RIDX; \
|
||
vpgatherdd RNOT, (rs0, RIDX, 4), xy ## 0; \
|
||
vpcmpeqd RNOT, RNOT, RNOT; \
|
||
\
|
||
vpand RBYTE, ab ## 1, RIDX; \
|
||
vpgatherdd RNOT, (rs0, RIDX, 4), xy ## 1; \
|
||
vpcmpeqd RNOT, RNOT, RNOT; \
|
||
\
|
||
vpsrld $8, ab ## 0, RIDX; \
|
||
vpand RBYTE, RIDX, RIDX; \
|
||
vpgatherdd RNOT, (rs1, RIDX, 4), RT0; \
|
||
vpcmpeqd RNOT, RNOT, RNOT; \
|
||
vpxor RT0, xy ## 0, xy ## 0; \
|
||
\
|
||
vpsrld $8, ab ## 1, RIDX; \
|
||
vpand RBYTE, RIDX, RIDX; \
|
||
vpgatherdd RNOT, (rs1, RIDX, 4), RT0; \
|
||
vpcmpeqd RNOT, RNOT, RNOT; \
|
||
vpxor RT0, xy ## 1, xy ## 1; \
|
||
\
|
||
vpsrld $16, ab ## 0, RIDX; \
|
||
vpand RBYTE, RIDX, RIDX; \
|
||
vpgatherdd RNOT, (rs2, RIDX, 4), RT0; \
|
||
vpcmpeqd RNOT, RNOT, RNOT; \
|
||
vpxor RT0, xy ## 0, xy ## 0; \
|
||
\
|
||
vpsrld $16, ab ## 1, RIDX; \
|
||
vpand RBYTE, RIDX, RIDX; \
|
||
vpgatherdd RNOT, (rs2, RIDX, 4), RT0; \
|
||
vpcmpeqd RNOT, RNOT, RNOT; \
|
||
vpxor RT0, xy ## 1, xy ## 1; \
|
||
\
|
||
vpsrld $24, ab ## 0, RIDX; \
|
||
vpgatherdd RNOT, (rs3, RIDX, 4), RT0; \
|
||
vpcmpeqd RNOT, RNOT, RNOT; \
|
||
vpxor RT0, xy ## 0, xy ## 0; \
|
||
\
|
||
vpsrld $24, ab ## 1, RIDX; \
|
||
vpgatherdd RNOT, (rs3, RIDX, 4), RT0; \
|
||
vpcmpeqd RNOT, RNOT, RNOT; \
|
||
vpxor RT0, xy ## 1, xy ## 1;
|
||
|
||
#define g1_16(a, x) \
|
||
g16(a, RS0, RS1, RS2, RS3, x);
|
||
|
||
#define g2_16(b, y) \
|
||
g16(b, RS1, RS2, RS3, RS0, y);
|
||
|
||
#define encrypt_round_end16(a, b, c, d, nk) \
|
||
vpaddd RY0, RX0, RX0; \
|
||
vpaddd RX0, RY0, RY0; \
|
||
vpbroadcastd nk(RK,RROUND,8), RT0; \
|
||
vpaddd RT0, RX0, RX0; \
|
||
vpbroadcastd 4+nk(RK,RROUND,8), RT0; \
|
||
vpaddd RT0, RY0, RY0; \
|
||
\
|
||
vpxor RY0, d ## 0, d ## 0; \
|
||
\
|
||
vpxor RX0, c ## 0, c ## 0; \
|
||
vpsrld $1, c ## 0, RT0; \
|
||
vpslld $31, c ## 0, c ## 0; \
|
||
vpor RT0, c ## 0, c ## 0; \
|
||
\
|
||
vpaddd RY1, RX1, RX1; \
|
||
vpaddd RX1, RY1, RY1; \
|
||
vpbroadcastd nk(RK,RROUND,8), RT0; \
|
||
vpaddd RT0, RX1, RX1; \
|
||
vpbroadcastd 4+nk(RK,RROUND,8), RT0; \
|
||
vpaddd RT0, RY1, RY1; \
|
||
\
|
||
vpxor RY1, d ## 1, d ## 1; \
|
||
\
|
||
vpxor RX1, c ## 1, c ## 1; \
|
||
vpsrld $1, c ## 1, RT0; \
|
||
vpslld $31, c ## 1, c ## 1; \
|
||
vpor RT0, c ## 1, c ## 1; \
|
||
|
||
#define encrypt_round16(a, b, c, d, nk) \
|
||
g2_16(b, RY); \
|
||
\
|
||
vpslld $1, b ## 0, RT0; \
|
||
vpsrld $31, b ## 0, b ## 0; \
|
||
vpor RT0, b ## 0, b ## 0; \
|
||
\
|
||
vpslld $1, b ## 1, RT0; \
|
||
vpsrld $31, b ## 1, b ## 1; \
|
||
vpor RT0, b ## 1, b ## 1; \
|
||
\
|
||
g1_16(a, RX); \
|
||
\
|
||
encrypt_round_end16(a, b, c, d, nk);
|
||
|
||
#define encrypt_round_first16(a, b, c, d, nk) \
|
||
vpslld $1, d ## 0, RT0; \
|
||
vpsrld $31, d ## 0, d ## 0; \
|
||
vpor RT0, d ## 0, d ## 0; \
|
||
\
|
||
vpslld $1, d ## 1, RT0; \
|
||
vpsrld $31, d ## 1, d ## 1; \
|
||
vpor RT0, d ## 1, d ## 1; \
|
||
\
|
||
encrypt_round16(a, b, c, d, nk);
|
||
|
||
#define encrypt_round_last16(a, b, c, d, nk) \
|
||
g2_16(b, RY); \
|
||
\
|
||
g1_16(a, RX); \
|
||
\
|
||
encrypt_round_end16(a, b, c, d, nk);
|
||
|
||
#define decrypt_round_end16(a, b, c, d, nk) \
|
||
vpaddd RY0, RX0, RX0; \
|
||
vpaddd RX0, RY0, RY0; \
|
||
vpbroadcastd nk(RK,RROUND,8), RT0; \
|
||
vpaddd RT0, RX0, RX0; \
|
||
vpbroadcastd 4+nk(RK,RROUND,8), RT0; \
|
||
vpaddd RT0, RY0, RY0; \
|
||
\
|
||
vpxor RX0, c ## 0, c ## 0; \
|
||
\
|
||
vpxor RY0, d ## 0, d ## 0; \
|
||
vpsrld $1, d ## 0, RT0; \
|
||
vpslld $31, d ## 0, d ## 0; \
|
||
vpor RT0, d ## 0, d ## 0; \
|
||
\
|
||
vpaddd RY1, RX1, RX1; \
|
||
vpaddd RX1, RY1, RY1; \
|
||
vpbroadcastd nk(RK,RROUND,8), RT0; \
|
||
vpaddd RT0, RX1, RX1; \
|
||
vpbroadcastd 4+nk(RK,RROUND,8), RT0; \
|
||
vpaddd RT0, RY1, RY1; \
|
||
\
|
||
vpxor RX1, c ## 1, c ## 1; \
|
||
\
|
||
vpxor RY1, d ## 1, d ## 1; \
|
||
vpsrld $1, d ## 1, RT0; \
|
||
vpslld $31, d ## 1, d ## 1; \
|
||
vpor RT0, d ## 1, d ## 1;
|
||
|
||
#define decrypt_round16(a, b, c, d, nk) \
|
||
g1_16(a, RX); \
|
||
\
|
||
vpslld $1, a ## 0, RT0; \
|
||
vpsrld $31, a ## 0, a ## 0; \
|
||
vpor RT0, a ## 0, a ## 0; \
|
||
\
|
||
vpslld $1, a ## 1, RT0; \
|
||
vpsrld $31, a ## 1, a ## 1; \
|
||
vpor RT0, a ## 1, a ## 1; \
|
||
\
|
||
g2_16(b, RY); \
|
||
\
|
||
decrypt_round_end16(a, b, c, d, nk);
|
||
|
||
#define decrypt_round_first16(a, b, c, d, nk) \
|
||
vpslld $1, c ## 0, RT0; \
|
||
vpsrld $31, c ## 0, c ## 0; \
|
||
vpor RT0, c ## 0, c ## 0; \
|
||
\
|
||
vpslld $1, c ## 1, RT0; \
|
||
vpsrld $31, c ## 1, c ## 1; \
|
||
vpor RT0, c ## 1, c ## 1; \
|
||
\
|
||
decrypt_round16(a, b, c, d, nk)
|
||
|
||
#define decrypt_round_last16(a, b, c, d, nk) \
|
||
g1_16(a, RX); \
|
||
\
|
||
g2_16(b, RY); \
|
||
\
|
||
decrypt_round_end16(a, b, c, d, nk);
|
||
|
||
#define encrypt_cycle16() \
|
||
encrypt_round16(RA, RB, RC, RD, 0); \
|
||
encrypt_round16(RC, RD, RA, RB, 8);
|
||
|
||
#define encrypt_cycle_first16() \
|
||
encrypt_round_first16(RA, RB, RC, RD, 0); \
|
||
encrypt_round16(RC, RD, RA, RB, 8);
|
||
|
||
#define encrypt_cycle_last16() \
|
||
encrypt_round16(RA, RB, RC, RD, 0); \
|
||
encrypt_round_last16(RC, RD, RA, RB, 8);
|
||
|
||
#define decrypt_cycle16(n) \
|
||
decrypt_round16(RC, RD, RA, RB, 8); \
|
||
decrypt_round16(RA, RB, RC, RD, 0);
|
||
|
||
#define decrypt_cycle_first16(n) \
|
||
decrypt_round_first16(RC, RD, RA, RB, 8); \
|
||
decrypt_round16(RA, RB, RC, RD, 0);
|
||
|
||
#define decrypt_cycle_last16(n) \
|
||
decrypt_round16(RC, RD, RA, RB, 8); \
|
||
decrypt_round_last16(RA, RB, RC, RD, 0);
|
||
|
||
#define transpose_4x4(x0,x1,x2,x3,t1,t2) \
|
||
vpunpckhdq x1, x0, t2; \
|
||
vpunpckldq x1, x0, x0; \
|
||
\
|
||
vpunpckldq x3, x2, t1; \
|
||
vpunpckhdq x3, x2, x2; \
|
||
\
|
||
vpunpckhqdq t1, x0, x1; \
|
||
vpunpcklqdq t1, x0, x0; \
|
||
\
|
||
vpunpckhqdq x2, t2, x3; \
|
||
vpunpcklqdq x2, t2, x2;
|
||
|
||
#define read_blocks8(offs,a,b,c,d) \
|
||
transpose_4x4(a, b, c, d, RX0, RY0);
|
||
|
||
#define write_blocks8(offs,a,b,c,d) \
|
||
transpose_4x4(a, b, c, d, RX0, RY0);
|
||
|
||
#define inpack_enc8(a,b,c,d) \
|
||
vpbroadcastd 4*0(RW), RT0; \
|
||
vpxor RT0, a, a; \
|
||
\
|
||
vpbroadcastd 4*1(RW), RT0; \
|
||
vpxor RT0, b, b; \
|
||
\
|
||
vpbroadcastd 4*2(RW), RT0; \
|
||
vpxor RT0, c, c; \
|
||
\
|
||
vpbroadcastd 4*3(RW), RT0; \
|
||
vpxor RT0, d, d;
|
||
|
||
#define outunpack_enc8(a,b,c,d) \
|
||
vpbroadcastd 4*4(RW), RX0; \
|
||
vpbroadcastd 4*5(RW), RY0; \
|
||
vpxor RX0, c, RX0; \
|
||
vpxor RY0, d, RY0; \
|
||
\
|
||
vpbroadcastd 4*6(RW), RT0; \
|
||
vpxor RT0, a, c; \
|
||
vpbroadcastd 4*7(RW), RT0; \
|
||
vpxor RT0, b, d; \
|
||
\
|
||
vmovdqa RX0, a; \
|
||
vmovdqa RY0, b;
|
||
|
||
#define inpack_dec8(a,b,c,d) \
|
||
vpbroadcastd 4*4(RW), RX0; \
|
||
vpbroadcastd 4*5(RW), RY0; \
|
||
vpxor RX0, a, RX0; \
|
||
vpxor RY0, b, RY0; \
|
||
\
|
||
vpbroadcastd 4*6(RW), RT0; \
|
||
vpxor RT0, c, a; \
|
||
vpbroadcastd 4*7(RW), RT0; \
|
||
vpxor RT0, d, b; \
|
||
\
|
||
vmovdqa RX0, c; \
|
||
vmovdqa RY0, d;
|
||
|
||
#define outunpack_dec8(a,b,c,d) \
|
||
vpbroadcastd 4*0(RW), RT0; \
|
||
vpxor RT0, a, a; \
|
||
\
|
||
vpbroadcastd 4*1(RW), RT0; \
|
||
vpxor RT0, b, b; \
|
||
\
|
||
vpbroadcastd 4*2(RW), RT0; \
|
||
vpxor RT0, c, c; \
|
||
\
|
||
vpbroadcastd 4*3(RW), RT0; \
|
||
vpxor RT0, d, d;
|
||
|
||
#define read_blocks16(a,b,c,d) \
|
||
read_blocks8(0, a ## 0, b ## 0, c ## 0, d ## 0); \
|
||
read_blocks8(8, a ## 1, b ## 1, c ## 1, d ## 1);
|
||
|
||
#define write_blocks16(a,b,c,d) \
|
||
write_blocks8(0, a ## 0, b ## 0, c ## 0, d ## 0); \
|
||
write_blocks8(8, a ## 1, b ## 1, c ## 1, d ## 1);
|
||
|
||
#define xor_blocks16(a,b,c,d) \
|
||
xor_blocks8(0, a ## 0, b ## 0, c ## 0, d ## 0); \
|
||
xor_blocks8(8, a ## 1, b ## 1, c ## 1, d ## 1);
|
||
|
||
#define inpack_enc16(a,b,c,d) \
|
||
inpack_enc8(a ## 0, b ## 0, c ## 0, d ## 0); \
|
||
inpack_enc8(a ## 1, b ## 1, c ## 1, d ## 1);
|
||
|
||
#define outunpack_enc16(a,b,c,d) \
|
||
outunpack_enc8(a ## 0, b ## 0, c ## 0, d ## 0); \
|
||
outunpack_enc8(a ## 1, b ## 1, c ## 1, d ## 1);
|
||
|
||
#define inpack_dec16(a,b,c,d) \
|
||
inpack_dec8(a ## 0, b ## 0, c ## 0, d ## 0); \
|
||
inpack_dec8(a ## 1, b ## 1, c ## 1, d ## 1);
|
||
|
||
#define outunpack_dec16(a,b,c,d) \
|
||
outunpack_dec8(a ## 0, b ## 0, c ## 0, d ## 0); \
|
||
outunpack_dec8(a ## 1, b ## 1, c ## 1, d ## 1);
|
||
|
||
.align 8
|
||
__twofish_enc_blk16:
|
||
/* input:
|
||
* %rdi: ctx, CTX
|
||
* RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1: plaintext
|
||
* output:
|
||
* RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1: ciphertext
|
||
*/
|
||
init_round_constants();
|
||
|
||
read_blocks16(RA, RB, RC, RD);
|
||
inpack_enc16(RA, RB, RC, RD);
|
||
|
||
xorl RROUNDd, RROUNDd;
|
||
encrypt_cycle_first16();
|
||
movl $2, RROUNDd;
|
||
|
||
.align 4
|
||
.L__enc_loop:
|
||
encrypt_cycle16();
|
||
|
||
addl $2, RROUNDd;
|
||
cmpl $14, RROUNDd;
|
||
jne .L__enc_loop;
|
||
|
||
encrypt_cycle_last16();
|
||
|
||
outunpack_enc16(RA, RB, RC, RD);
|
||
write_blocks16(RA, RB, RC, RD);
|
||
|
||
ret;
|
||
ENDPROC(__twofish_enc_blk16)
|
||
|
||
.align 8
|
||
__twofish_dec_blk16:
|
||
/* input:
|
||
* %rdi: ctx, CTX
|
||
* RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1: ciphertext
|
||
* output:
|
||
* RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1: plaintext
|
||
*/
|
||
init_round_constants();
|
||
|
||
read_blocks16(RA, RB, RC, RD);
|
||
inpack_dec16(RA, RB, RC, RD);
|
||
|
||
movl $14, RROUNDd;
|
||
decrypt_cycle_first16();
|
||
movl $12, RROUNDd;
|
||
|
||
.align 4
|
||
.L__dec_loop:
|
||
decrypt_cycle16();
|
||
|
||
addl $-2, RROUNDd;
|
||
jnz .L__dec_loop;
|
||
|
||
decrypt_cycle_last16();
|
||
|
||
outunpack_dec16(RA, RB, RC, RD);
|
||
write_blocks16(RA, RB, RC, RD);
|
||
|
||
ret;
|
||
ENDPROC(__twofish_dec_blk16)
|
||
|
||
ENTRY(twofish_ecb_enc_16way)
|
||
/* input:
|
||
* %rdi: ctx, CTX
|
||
* %rsi: dst
|
||
* %rdx: src
|
||
*/
|
||
|
||
vzeroupper;
|
||
pushq %r12;
|
||
|
||
load_16way(%rdx, RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1);
|
||
|
||
call __twofish_enc_blk16;
|
||
|
||
store_16way(%rsi, RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1);
|
||
|
||
popq %r12;
|
||
vzeroupper;
|
||
|
||
ret;
|
||
ENDPROC(twofish_ecb_enc_16way)
|
||
|
||
ENTRY(twofish_ecb_dec_16way)
|
||
/* input:
|
||
* %rdi: ctx, CTX
|
||
* %rsi: dst
|
||
* %rdx: src
|
||
*/
|
||
|
||
vzeroupper;
|
||
pushq %r12;
|
||
|
||
load_16way(%rdx, RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1);
|
||
|
||
call __twofish_dec_blk16;
|
||
|
||
store_16way(%rsi, RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1);
|
||
|
||
popq %r12;
|
||
vzeroupper;
|
||
|
||
ret;
|
||
ENDPROC(twofish_ecb_dec_16way)
|
||
|
||
ENTRY(twofish_cbc_dec_16way)
|
||
/* input:
|
||
* %rdi: ctx, CTX
|
||
* %rsi: dst
|
||
* %rdx: src
|
||
*/
|
||
|
||
vzeroupper;
|
||
pushq %r12;
|
||
|
||
load_16way(%rdx, RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1);
|
||
|
||
call __twofish_dec_blk16;
|
||
|
||
store_cbc_16way(%rdx, %rsi, RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1,
|
||
RX0);
|
||
|
||
popq %r12;
|
||
vzeroupper;
|
||
|
||
ret;
|
||
ENDPROC(twofish_cbc_dec_16way)
|
||
|
||
ENTRY(twofish_ctr_16way)
|
||
/* input:
|
||
* %rdi: ctx, CTX
|
||
* %rsi: dst (16 blocks)
|
||
* %rdx: src (16 blocks)
|
||
* %rcx: iv (little endian, 128bit)
|
||
*/
|
||
|
||
vzeroupper;
|
||
pushq %r12;
|
||
|
||
load_ctr_16way(%rcx, .Lbswap128_mask, RA0, RB0, RC0, RD0, RA1, RB1, RC1,
|
||
RD1, RX0, RX0x, RX1, RX1x, RY0, RY0x, RY1, RY1x, RNOT,
|
||
RBYTE);
|
||
|
||
call __twofish_enc_blk16;
|
||
|
||
store_ctr_16way(%rdx, %rsi, RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1);
|
||
|
||
popq %r12;
|
||
vzeroupper;
|
||
|
||
ret;
|
||
ENDPROC(twofish_ctr_16way)
|
||
|
||
.align 8
|
||
twofish_xts_crypt_16way:
|
||
/* input:
|
||
* %rdi: ctx, CTX
|
||
* %rsi: dst (16 blocks)
|
||
* %rdx: src (16 blocks)
|
||
* %rcx: iv (t ⊕ αⁿ ∈ GF(2¹²⁸))
|
||
* %r8: pointer to __twofish_enc_blk16 or __twofish_dec_blk16
|
||
*/
|
||
|
||
vzeroupper;
|
||
pushq %r12;
|
||
|
||
load_xts_16way(%rcx, %rdx, %rsi, RA0, RB0, RC0, RD0, RA1, RB1, RC1,
|
||
RD1, RX0, RX0x, RX1, RX1x, RY0, RY0x, RY1, RY1x, RNOT,
|
||
.Lxts_gf128mul_and_shl1_mask_0,
|
||
.Lxts_gf128mul_and_shl1_mask_1);
|
||
|
||
call *%r8;
|
||
|
||
store_xts_16way(%rsi, RA0, RB0, RC0, RD0, RA1, RB1, RC1, RD1);
|
||
|
||
popq %r12;
|
||
vzeroupper;
|
||
|
||
ret;
|
||
ENDPROC(twofish_xts_crypt_16way)
|
||
|
||
ENTRY(twofish_xts_enc_16way)
|
||
/* input:
|
||
* %rdi: ctx, CTX
|
||
* %rsi: dst (16 blocks)
|
||
* %rdx: src (16 blocks)
|
||
* %rcx: iv (t ⊕ αⁿ ∈ GF(2¹²⁸))
|
||
*/
|
||
leaq __twofish_enc_blk16, %r8;
|
||
jmp twofish_xts_crypt_16way;
|
||
ENDPROC(twofish_xts_enc_16way)
|
||
|
||
ENTRY(twofish_xts_dec_16way)
|
||
/* input:
|
||
* %rdi: ctx, CTX
|
||
* %rsi: dst (16 blocks)
|
||
* %rdx: src (16 blocks)
|
||
* %rcx: iv (t ⊕ αⁿ ∈ GF(2¹²⁸))
|
||
*/
|
||
leaq __twofish_dec_blk16, %r8;
|
||
jmp twofish_xts_crypt_16way;
|
||
ENDPROC(twofish_xts_dec_16way)
|