mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-11-25 07:30:53 +07:00
dd165a658d
Based on 1 normalized pattern(s): this program is free software you can redistribute it and or modify it under the terms of the gnu general public license as published by the free software foundation inc 53 temple place ste 330 boston ma 02111 1307 usa either version 2 of the license or at your option any later version incorporated herein by reference extracted by the scancode license scanner the SPDX license identifier GPL-2.0-or-later has been chosen to replace the boilerplate/reference in 13 file(s). Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Reviewed-by: Allison Randal <allison@lohutok.net> Reviewed-by: Kate Stewart <kstewart@linuxfoundation.org> Cc: linux-spdx@vger.kernel.org Link: https://lkml.kernel.org/r/20190520170858.645641371@linutronix.de Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
475 lines
17 KiB
C
475 lines
17 KiB
C
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
/* -*- linux-c -*- ------------------------------------------------------- *
|
|
*
|
|
* Copyright (C) 2012 Intel Corporation
|
|
* Author: Yuanhan Liu <yuanhan.liu@linux.intel.com>
|
|
*
|
|
* Based on sse2.c: Copyright 2002 H. Peter Anvin - All Rights Reserved
|
|
*
|
|
* ----------------------------------------------------------------------- */
|
|
|
|
/*
|
|
* AVX2 implementation of RAID-6 syndrome functions
|
|
*
|
|
*/
|
|
|
|
#ifdef CONFIG_AS_AVX2
|
|
|
|
#include <linux/raid/pq.h>
|
|
#include "x86.h"
|
|
|
|
static const struct raid6_avx2_constants {
|
|
u64 x1d[4];
|
|
} raid6_avx2_constants __aligned(32) = {
|
|
{ 0x1d1d1d1d1d1d1d1dULL, 0x1d1d1d1d1d1d1d1dULL,
|
|
0x1d1d1d1d1d1d1d1dULL, 0x1d1d1d1d1d1d1d1dULL,},
|
|
};
|
|
|
|
static int raid6_have_avx2(void)
|
|
{
|
|
return boot_cpu_has(X86_FEATURE_AVX2) && boot_cpu_has(X86_FEATURE_AVX);
|
|
}
|
|
|
|
/*
|
|
* Plain AVX2 implementation
|
|
*/
|
|
static void raid6_avx21_gen_syndrome(int disks, size_t bytes, void **ptrs)
|
|
{
|
|
u8 **dptr = (u8 **)ptrs;
|
|
u8 *p, *q;
|
|
int d, z, z0;
|
|
|
|
z0 = disks - 3; /* Highest data disk */
|
|
p = dptr[z0+1]; /* XOR parity */
|
|
q = dptr[z0+2]; /* RS syndrome */
|
|
|
|
kernel_fpu_begin();
|
|
|
|
asm volatile("vmovdqa %0,%%ymm0" : : "m" (raid6_avx2_constants.x1d[0]));
|
|
asm volatile("vpxor %ymm3,%ymm3,%ymm3"); /* Zero temp */
|
|
|
|
for (d = 0; d < bytes; d += 32) {
|
|
asm volatile("prefetchnta %0" : : "m" (dptr[z0][d]));
|
|
asm volatile("vmovdqa %0,%%ymm2" : : "m" (dptr[z0][d]));/* P[0] */
|
|
asm volatile("prefetchnta %0" : : "m" (dptr[z0-1][d]));
|
|
asm volatile("vmovdqa %ymm2,%ymm4");/* Q[0] */
|
|
asm volatile("vmovdqa %0,%%ymm6" : : "m" (dptr[z0-1][d]));
|
|
for (z = z0-2; z >= 0; z--) {
|
|
asm volatile("prefetchnta %0" : : "m" (dptr[z][d]));
|
|
asm volatile("vpcmpgtb %ymm4,%ymm3,%ymm5");
|
|
asm volatile("vpaddb %ymm4,%ymm4,%ymm4");
|
|
asm volatile("vpand %ymm0,%ymm5,%ymm5");
|
|
asm volatile("vpxor %ymm5,%ymm4,%ymm4");
|
|
asm volatile("vpxor %ymm6,%ymm2,%ymm2");
|
|
asm volatile("vpxor %ymm6,%ymm4,%ymm4");
|
|
asm volatile("vmovdqa %0,%%ymm6" : : "m" (dptr[z][d]));
|
|
}
|
|
asm volatile("vpcmpgtb %ymm4,%ymm3,%ymm5");
|
|
asm volatile("vpaddb %ymm4,%ymm4,%ymm4");
|
|
asm volatile("vpand %ymm0,%ymm5,%ymm5");
|
|
asm volatile("vpxor %ymm5,%ymm4,%ymm4");
|
|
asm volatile("vpxor %ymm6,%ymm2,%ymm2");
|
|
asm volatile("vpxor %ymm6,%ymm4,%ymm4");
|
|
|
|
asm volatile("vmovntdq %%ymm2,%0" : "=m" (p[d]));
|
|
asm volatile("vpxor %ymm2,%ymm2,%ymm2");
|
|
asm volatile("vmovntdq %%ymm4,%0" : "=m" (q[d]));
|
|
asm volatile("vpxor %ymm4,%ymm4,%ymm4");
|
|
}
|
|
|
|
asm volatile("sfence" : : : "memory");
|
|
kernel_fpu_end();
|
|
}
|
|
|
|
static void raid6_avx21_xor_syndrome(int disks, int start, int stop,
|
|
size_t bytes, void **ptrs)
|
|
{
|
|
u8 **dptr = (u8 **)ptrs;
|
|
u8 *p, *q;
|
|
int d, z, z0;
|
|
|
|
z0 = stop; /* P/Q right side optimization */
|
|
p = dptr[disks-2]; /* XOR parity */
|
|
q = dptr[disks-1]; /* RS syndrome */
|
|
|
|
kernel_fpu_begin();
|
|
|
|
asm volatile("vmovdqa %0,%%ymm0" : : "m" (raid6_avx2_constants.x1d[0]));
|
|
|
|
for (d = 0 ; d < bytes ; d += 32) {
|
|
asm volatile("vmovdqa %0,%%ymm4" :: "m" (dptr[z0][d]));
|
|
asm volatile("vmovdqa %0,%%ymm2" : : "m" (p[d]));
|
|
asm volatile("vpxor %ymm4,%ymm2,%ymm2");
|
|
/* P/Q data pages */
|
|
for (z = z0-1 ; z >= start ; z--) {
|
|
asm volatile("vpxor %ymm5,%ymm5,%ymm5");
|
|
asm volatile("vpcmpgtb %ymm4,%ymm5,%ymm5");
|
|
asm volatile("vpaddb %ymm4,%ymm4,%ymm4");
|
|
asm volatile("vpand %ymm0,%ymm5,%ymm5");
|
|
asm volatile("vpxor %ymm5,%ymm4,%ymm4");
|
|
asm volatile("vmovdqa %0,%%ymm5" :: "m" (dptr[z][d]));
|
|
asm volatile("vpxor %ymm5,%ymm2,%ymm2");
|
|
asm volatile("vpxor %ymm5,%ymm4,%ymm4");
|
|
}
|
|
/* P/Q left side optimization */
|
|
for (z = start-1 ; z >= 0 ; z--) {
|
|
asm volatile("vpxor %ymm5,%ymm5,%ymm5");
|
|
asm volatile("vpcmpgtb %ymm4,%ymm5,%ymm5");
|
|
asm volatile("vpaddb %ymm4,%ymm4,%ymm4");
|
|
asm volatile("vpand %ymm0,%ymm5,%ymm5");
|
|
asm volatile("vpxor %ymm5,%ymm4,%ymm4");
|
|
}
|
|
asm volatile("vpxor %0,%%ymm4,%%ymm4" : : "m" (q[d]));
|
|
/* Don't use movntdq for r/w memory area < cache line */
|
|
asm volatile("vmovdqa %%ymm4,%0" : "=m" (q[d]));
|
|
asm volatile("vmovdqa %%ymm2,%0" : "=m" (p[d]));
|
|
}
|
|
|
|
asm volatile("sfence" : : : "memory");
|
|
kernel_fpu_end();
|
|
}
|
|
|
|
const struct raid6_calls raid6_avx2x1 = {
|
|
raid6_avx21_gen_syndrome,
|
|
raid6_avx21_xor_syndrome,
|
|
raid6_have_avx2,
|
|
"avx2x1",
|
|
1 /* Has cache hints */
|
|
};
|
|
|
|
/*
|
|
* Unrolled-by-2 AVX2 implementation
|
|
*/
|
|
static void raid6_avx22_gen_syndrome(int disks, size_t bytes, void **ptrs)
|
|
{
|
|
u8 **dptr = (u8 **)ptrs;
|
|
u8 *p, *q;
|
|
int d, z, z0;
|
|
|
|
z0 = disks - 3; /* Highest data disk */
|
|
p = dptr[z0+1]; /* XOR parity */
|
|
q = dptr[z0+2]; /* RS syndrome */
|
|
|
|
kernel_fpu_begin();
|
|
|
|
asm volatile("vmovdqa %0,%%ymm0" : : "m" (raid6_avx2_constants.x1d[0]));
|
|
asm volatile("vpxor %ymm1,%ymm1,%ymm1"); /* Zero temp */
|
|
|
|
/* We uniformly assume a single prefetch covers at least 32 bytes */
|
|
for (d = 0; d < bytes; d += 64) {
|
|
asm volatile("prefetchnta %0" : : "m" (dptr[z0][d]));
|
|
asm volatile("prefetchnta %0" : : "m" (dptr[z0][d+32]));
|
|
asm volatile("vmovdqa %0,%%ymm2" : : "m" (dptr[z0][d]));/* P[0] */
|
|
asm volatile("vmovdqa %0,%%ymm3" : : "m" (dptr[z0][d+32]));/* P[1] */
|
|
asm volatile("vmovdqa %ymm2,%ymm4"); /* Q[0] */
|
|
asm volatile("vmovdqa %ymm3,%ymm6"); /* Q[1] */
|
|
for (z = z0-1; z >= 0; z--) {
|
|
asm volatile("prefetchnta %0" : : "m" (dptr[z][d]));
|
|
asm volatile("prefetchnta %0" : : "m" (dptr[z][d+32]));
|
|
asm volatile("vpcmpgtb %ymm4,%ymm1,%ymm5");
|
|
asm volatile("vpcmpgtb %ymm6,%ymm1,%ymm7");
|
|
asm volatile("vpaddb %ymm4,%ymm4,%ymm4");
|
|
asm volatile("vpaddb %ymm6,%ymm6,%ymm6");
|
|
asm volatile("vpand %ymm0,%ymm5,%ymm5");
|
|
asm volatile("vpand %ymm0,%ymm7,%ymm7");
|
|
asm volatile("vpxor %ymm5,%ymm4,%ymm4");
|
|
asm volatile("vpxor %ymm7,%ymm6,%ymm6");
|
|
asm volatile("vmovdqa %0,%%ymm5" : : "m" (dptr[z][d]));
|
|
asm volatile("vmovdqa %0,%%ymm7" : : "m" (dptr[z][d+32]));
|
|
asm volatile("vpxor %ymm5,%ymm2,%ymm2");
|
|
asm volatile("vpxor %ymm7,%ymm3,%ymm3");
|
|
asm volatile("vpxor %ymm5,%ymm4,%ymm4");
|
|
asm volatile("vpxor %ymm7,%ymm6,%ymm6");
|
|
}
|
|
asm volatile("vmovntdq %%ymm2,%0" : "=m" (p[d]));
|
|
asm volatile("vmovntdq %%ymm3,%0" : "=m" (p[d+32]));
|
|
asm volatile("vmovntdq %%ymm4,%0" : "=m" (q[d]));
|
|
asm volatile("vmovntdq %%ymm6,%0" : "=m" (q[d+32]));
|
|
}
|
|
|
|
asm volatile("sfence" : : : "memory");
|
|
kernel_fpu_end();
|
|
}
|
|
|
|
static void raid6_avx22_xor_syndrome(int disks, int start, int stop,
|
|
size_t bytes, void **ptrs)
|
|
{
|
|
u8 **dptr = (u8 **)ptrs;
|
|
u8 *p, *q;
|
|
int d, z, z0;
|
|
|
|
z0 = stop; /* P/Q right side optimization */
|
|
p = dptr[disks-2]; /* XOR parity */
|
|
q = dptr[disks-1]; /* RS syndrome */
|
|
|
|
kernel_fpu_begin();
|
|
|
|
asm volatile("vmovdqa %0,%%ymm0" : : "m" (raid6_avx2_constants.x1d[0]));
|
|
|
|
for (d = 0 ; d < bytes ; d += 64) {
|
|
asm volatile("vmovdqa %0,%%ymm4" :: "m" (dptr[z0][d]));
|
|
asm volatile("vmovdqa %0,%%ymm6" :: "m" (dptr[z0][d+32]));
|
|
asm volatile("vmovdqa %0,%%ymm2" : : "m" (p[d]));
|
|
asm volatile("vmovdqa %0,%%ymm3" : : "m" (p[d+32]));
|
|
asm volatile("vpxor %ymm4,%ymm2,%ymm2");
|
|
asm volatile("vpxor %ymm6,%ymm3,%ymm3");
|
|
/* P/Q data pages */
|
|
for (z = z0-1 ; z >= start ; z--) {
|
|
asm volatile("vpxor %ymm5,%ymm5,%ymm5");
|
|
asm volatile("vpxor %ymm7,%ymm7,%ymm7");
|
|
asm volatile("vpcmpgtb %ymm4,%ymm5,%ymm5");
|
|
asm volatile("vpcmpgtb %ymm6,%ymm7,%ymm7");
|
|
asm volatile("vpaddb %ymm4,%ymm4,%ymm4");
|
|
asm volatile("vpaddb %ymm6,%ymm6,%ymm6");
|
|
asm volatile("vpand %ymm0,%ymm5,%ymm5");
|
|
asm volatile("vpand %ymm0,%ymm7,%ymm7");
|
|
asm volatile("vpxor %ymm5,%ymm4,%ymm4");
|
|
asm volatile("vpxor %ymm7,%ymm6,%ymm6");
|
|
asm volatile("vmovdqa %0,%%ymm5" :: "m" (dptr[z][d]));
|
|
asm volatile("vmovdqa %0,%%ymm7"
|
|
:: "m" (dptr[z][d+32]));
|
|
asm volatile("vpxor %ymm5,%ymm2,%ymm2");
|
|
asm volatile("vpxor %ymm7,%ymm3,%ymm3");
|
|
asm volatile("vpxor %ymm5,%ymm4,%ymm4");
|
|
asm volatile("vpxor %ymm7,%ymm6,%ymm6");
|
|
}
|
|
/* P/Q left side optimization */
|
|
for (z = start-1 ; z >= 0 ; z--) {
|
|
asm volatile("vpxor %ymm5,%ymm5,%ymm5");
|
|
asm volatile("vpxor %ymm7,%ymm7,%ymm7");
|
|
asm volatile("vpcmpgtb %ymm4,%ymm5,%ymm5");
|
|
asm volatile("vpcmpgtb %ymm6,%ymm7,%ymm7");
|
|
asm volatile("vpaddb %ymm4,%ymm4,%ymm4");
|
|
asm volatile("vpaddb %ymm6,%ymm6,%ymm6");
|
|
asm volatile("vpand %ymm0,%ymm5,%ymm5");
|
|
asm volatile("vpand %ymm0,%ymm7,%ymm7");
|
|
asm volatile("vpxor %ymm5,%ymm4,%ymm4");
|
|
asm volatile("vpxor %ymm7,%ymm6,%ymm6");
|
|
}
|
|
asm volatile("vpxor %0,%%ymm4,%%ymm4" : : "m" (q[d]));
|
|
asm volatile("vpxor %0,%%ymm6,%%ymm6" : : "m" (q[d+32]));
|
|
/* Don't use movntdq for r/w memory area < cache line */
|
|
asm volatile("vmovdqa %%ymm4,%0" : "=m" (q[d]));
|
|
asm volatile("vmovdqa %%ymm6,%0" : "=m" (q[d+32]));
|
|
asm volatile("vmovdqa %%ymm2,%0" : "=m" (p[d]));
|
|
asm volatile("vmovdqa %%ymm3,%0" : "=m" (p[d+32]));
|
|
}
|
|
|
|
asm volatile("sfence" : : : "memory");
|
|
kernel_fpu_end();
|
|
}
|
|
|
|
const struct raid6_calls raid6_avx2x2 = {
|
|
raid6_avx22_gen_syndrome,
|
|
raid6_avx22_xor_syndrome,
|
|
raid6_have_avx2,
|
|
"avx2x2",
|
|
1 /* Has cache hints */
|
|
};
|
|
|
|
#ifdef CONFIG_X86_64
|
|
|
|
/*
|
|
* Unrolled-by-4 AVX2 implementation
|
|
*/
|
|
static void raid6_avx24_gen_syndrome(int disks, size_t bytes, void **ptrs)
|
|
{
|
|
u8 **dptr = (u8 **)ptrs;
|
|
u8 *p, *q;
|
|
int d, z, z0;
|
|
|
|
z0 = disks - 3; /* Highest data disk */
|
|
p = dptr[z0+1]; /* XOR parity */
|
|
q = dptr[z0+2]; /* RS syndrome */
|
|
|
|
kernel_fpu_begin();
|
|
|
|
asm volatile("vmovdqa %0,%%ymm0" : : "m" (raid6_avx2_constants.x1d[0]));
|
|
asm volatile("vpxor %ymm1,%ymm1,%ymm1"); /* Zero temp */
|
|
asm volatile("vpxor %ymm2,%ymm2,%ymm2"); /* P[0] */
|
|
asm volatile("vpxor %ymm3,%ymm3,%ymm3"); /* P[1] */
|
|
asm volatile("vpxor %ymm4,%ymm4,%ymm4"); /* Q[0] */
|
|
asm volatile("vpxor %ymm6,%ymm6,%ymm6"); /* Q[1] */
|
|
asm volatile("vpxor %ymm10,%ymm10,%ymm10"); /* P[2] */
|
|
asm volatile("vpxor %ymm11,%ymm11,%ymm11"); /* P[3] */
|
|
asm volatile("vpxor %ymm12,%ymm12,%ymm12"); /* Q[2] */
|
|
asm volatile("vpxor %ymm14,%ymm14,%ymm14"); /* Q[3] */
|
|
|
|
for (d = 0; d < bytes; d += 128) {
|
|
for (z = z0; z >= 0; z--) {
|
|
asm volatile("prefetchnta %0" : : "m" (dptr[z][d]));
|
|
asm volatile("prefetchnta %0" : : "m" (dptr[z][d+32]));
|
|
asm volatile("prefetchnta %0" : : "m" (dptr[z][d+64]));
|
|
asm volatile("prefetchnta %0" : : "m" (dptr[z][d+96]));
|
|
asm volatile("vpcmpgtb %ymm4,%ymm1,%ymm5");
|
|
asm volatile("vpcmpgtb %ymm6,%ymm1,%ymm7");
|
|
asm volatile("vpcmpgtb %ymm12,%ymm1,%ymm13");
|
|
asm volatile("vpcmpgtb %ymm14,%ymm1,%ymm15");
|
|
asm volatile("vpaddb %ymm4,%ymm4,%ymm4");
|
|
asm volatile("vpaddb %ymm6,%ymm6,%ymm6");
|
|
asm volatile("vpaddb %ymm12,%ymm12,%ymm12");
|
|
asm volatile("vpaddb %ymm14,%ymm14,%ymm14");
|
|
asm volatile("vpand %ymm0,%ymm5,%ymm5");
|
|
asm volatile("vpand %ymm0,%ymm7,%ymm7");
|
|
asm volatile("vpand %ymm0,%ymm13,%ymm13");
|
|
asm volatile("vpand %ymm0,%ymm15,%ymm15");
|
|
asm volatile("vpxor %ymm5,%ymm4,%ymm4");
|
|
asm volatile("vpxor %ymm7,%ymm6,%ymm6");
|
|
asm volatile("vpxor %ymm13,%ymm12,%ymm12");
|
|
asm volatile("vpxor %ymm15,%ymm14,%ymm14");
|
|
asm volatile("vmovdqa %0,%%ymm5" : : "m" (dptr[z][d]));
|
|
asm volatile("vmovdqa %0,%%ymm7" : : "m" (dptr[z][d+32]));
|
|
asm volatile("vmovdqa %0,%%ymm13" : : "m" (dptr[z][d+64]));
|
|
asm volatile("vmovdqa %0,%%ymm15" : : "m" (dptr[z][d+96]));
|
|
asm volatile("vpxor %ymm5,%ymm2,%ymm2");
|
|
asm volatile("vpxor %ymm7,%ymm3,%ymm3");
|
|
asm volatile("vpxor %ymm13,%ymm10,%ymm10");
|
|
asm volatile("vpxor %ymm15,%ymm11,%ymm11");
|
|
asm volatile("vpxor %ymm5,%ymm4,%ymm4");
|
|
asm volatile("vpxor %ymm7,%ymm6,%ymm6");
|
|
asm volatile("vpxor %ymm13,%ymm12,%ymm12");
|
|
asm volatile("vpxor %ymm15,%ymm14,%ymm14");
|
|
}
|
|
asm volatile("vmovntdq %%ymm2,%0" : "=m" (p[d]));
|
|
asm volatile("vpxor %ymm2,%ymm2,%ymm2");
|
|
asm volatile("vmovntdq %%ymm3,%0" : "=m" (p[d+32]));
|
|
asm volatile("vpxor %ymm3,%ymm3,%ymm3");
|
|
asm volatile("vmovntdq %%ymm10,%0" : "=m" (p[d+64]));
|
|
asm volatile("vpxor %ymm10,%ymm10,%ymm10");
|
|
asm volatile("vmovntdq %%ymm11,%0" : "=m" (p[d+96]));
|
|
asm volatile("vpxor %ymm11,%ymm11,%ymm11");
|
|
asm volatile("vmovntdq %%ymm4,%0" : "=m" (q[d]));
|
|
asm volatile("vpxor %ymm4,%ymm4,%ymm4");
|
|
asm volatile("vmovntdq %%ymm6,%0" : "=m" (q[d+32]));
|
|
asm volatile("vpxor %ymm6,%ymm6,%ymm6");
|
|
asm volatile("vmovntdq %%ymm12,%0" : "=m" (q[d+64]));
|
|
asm volatile("vpxor %ymm12,%ymm12,%ymm12");
|
|
asm volatile("vmovntdq %%ymm14,%0" : "=m" (q[d+96]));
|
|
asm volatile("vpxor %ymm14,%ymm14,%ymm14");
|
|
}
|
|
|
|
asm volatile("sfence" : : : "memory");
|
|
kernel_fpu_end();
|
|
}
|
|
|
|
static void raid6_avx24_xor_syndrome(int disks, int start, int stop,
|
|
size_t bytes, void **ptrs)
|
|
{
|
|
u8 **dptr = (u8 **)ptrs;
|
|
u8 *p, *q;
|
|
int d, z, z0;
|
|
|
|
z0 = stop; /* P/Q right side optimization */
|
|
p = dptr[disks-2]; /* XOR parity */
|
|
q = dptr[disks-1]; /* RS syndrome */
|
|
|
|
kernel_fpu_begin();
|
|
|
|
asm volatile("vmovdqa %0,%%ymm0" :: "m" (raid6_avx2_constants.x1d[0]));
|
|
|
|
for (d = 0 ; d < bytes ; d += 128) {
|
|
asm volatile("vmovdqa %0,%%ymm4" :: "m" (dptr[z0][d]));
|
|
asm volatile("vmovdqa %0,%%ymm6" :: "m" (dptr[z0][d+32]));
|
|
asm volatile("vmovdqa %0,%%ymm12" :: "m" (dptr[z0][d+64]));
|
|
asm volatile("vmovdqa %0,%%ymm14" :: "m" (dptr[z0][d+96]));
|
|
asm volatile("vmovdqa %0,%%ymm2" : : "m" (p[d]));
|
|
asm volatile("vmovdqa %0,%%ymm3" : : "m" (p[d+32]));
|
|
asm volatile("vmovdqa %0,%%ymm10" : : "m" (p[d+64]));
|
|
asm volatile("vmovdqa %0,%%ymm11" : : "m" (p[d+96]));
|
|
asm volatile("vpxor %ymm4,%ymm2,%ymm2");
|
|
asm volatile("vpxor %ymm6,%ymm3,%ymm3");
|
|
asm volatile("vpxor %ymm12,%ymm10,%ymm10");
|
|
asm volatile("vpxor %ymm14,%ymm11,%ymm11");
|
|
/* P/Q data pages */
|
|
for (z = z0-1 ; z >= start ; z--) {
|
|
asm volatile("prefetchnta %0" :: "m" (dptr[z][d]));
|
|
asm volatile("prefetchnta %0" :: "m" (dptr[z][d+64]));
|
|
asm volatile("vpxor %ymm5,%ymm5,%ymm5");
|
|
asm volatile("vpxor %ymm7,%ymm7,%ymm7");
|
|
asm volatile("vpxor %ymm13,%ymm13,%ymm13");
|
|
asm volatile("vpxor %ymm15,%ymm15,%ymm15");
|
|
asm volatile("vpcmpgtb %ymm4,%ymm5,%ymm5");
|
|
asm volatile("vpcmpgtb %ymm6,%ymm7,%ymm7");
|
|
asm volatile("vpcmpgtb %ymm12,%ymm13,%ymm13");
|
|
asm volatile("vpcmpgtb %ymm14,%ymm15,%ymm15");
|
|
asm volatile("vpaddb %ymm4,%ymm4,%ymm4");
|
|
asm volatile("vpaddb %ymm6,%ymm6,%ymm6");
|
|
asm volatile("vpaddb %ymm12,%ymm12,%ymm12");
|
|
asm volatile("vpaddb %ymm14,%ymm14,%ymm14");
|
|
asm volatile("vpand %ymm0,%ymm5,%ymm5");
|
|
asm volatile("vpand %ymm0,%ymm7,%ymm7");
|
|
asm volatile("vpand %ymm0,%ymm13,%ymm13");
|
|
asm volatile("vpand %ymm0,%ymm15,%ymm15");
|
|
asm volatile("vpxor %ymm5,%ymm4,%ymm4");
|
|
asm volatile("vpxor %ymm7,%ymm6,%ymm6");
|
|
asm volatile("vpxor %ymm13,%ymm12,%ymm12");
|
|
asm volatile("vpxor %ymm15,%ymm14,%ymm14");
|
|
asm volatile("vmovdqa %0,%%ymm5" :: "m" (dptr[z][d]));
|
|
asm volatile("vmovdqa %0,%%ymm7"
|
|
:: "m" (dptr[z][d+32]));
|
|
asm volatile("vmovdqa %0,%%ymm13"
|
|
:: "m" (dptr[z][d+64]));
|
|
asm volatile("vmovdqa %0,%%ymm15"
|
|
:: "m" (dptr[z][d+96]));
|
|
asm volatile("vpxor %ymm5,%ymm2,%ymm2");
|
|
asm volatile("vpxor %ymm7,%ymm3,%ymm3");
|
|
asm volatile("vpxor %ymm13,%ymm10,%ymm10");
|
|
asm volatile("vpxor %ymm15,%ymm11,%ymm11");
|
|
asm volatile("vpxor %ymm5,%ymm4,%ymm4");
|
|
asm volatile("vpxor %ymm7,%ymm6,%ymm6");
|
|
asm volatile("vpxor %ymm13,%ymm12,%ymm12");
|
|
asm volatile("vpxor %ymm15,%ymm14,%ymm14");
|
|
}
|
|
asm volatile("prefetchnta %0" :: "m" (q[d]));
|
|
asm volatile("prefetchnta %0" :: "m" (q[d+64]));
|
|
/* P/Q left side optimization */
|
|
for (z = start-1 ; z >= 0 ; z--) {
|
|
asm volatile("vpxor %ymm5,%ymm5,%ymm5");
|
|
asm volatile("vpxor %ymm7,%ymm7,%ymm7");
|
|
asm volatile("vpxor %ymm13,%ymm13,%ymm13");
|
|
asm volatile("vpxor %ymm15,%ymm15,%ymm15");
|
|
asm volatile("vpcmpgtb %ymm4,%ymm5,%ymm5");
|
|
asm volatile("vpcmpgtb %ymm6,%ymm7,%ymm7");
|
|
asm volatile("vpcmpgtb %ymm12,%ymm13,%ymm13");
|
|
asm volatile("vpcmpgtb %ymm14,%ymm15,%ymm15");
|
|
asm volatile("vpaddb %ymm4,%ymm4,%ymm4");
|
|
asm volatile("vpaddb %ymm6,%ymm6,%ymm6");
|
|
asm volatile("vpaddb %ymm12,%ymm12,%ymm12");
|
|
asm volatile("vpaddb %ymm14,%ymm14,%ymm14");
|
|
asm volatile("vpand %ymm0,%ymm5,%ymm5");
|
|
asm volatile("vpand %ymm0,%ymm7,%ymm7");
|
|
asm volatile("vpand %ymm0,%ymm13,%ymm13");
|
|
asm volatile("vpand %ymm0,%ymm15,%ymm15");
|
|
asm volatile("vpxor %ymm5,%ymm4,%ymm4");
|
|
asm volatile("vpxor %ymm7,%ymm6,%ymm6");
|
|
asm volatile("vpxor %ymm13,%ymm12,%ymm12");
|
|
asm volatile("vpxor %ymm15,%ymm14,%ymm14");
|
|
}
|
|
asm volatile("vmovntdq %%ymm2,%0" : "=m" (p[d]));
|
|
asm volatile("vmovntdq %%ymm3,%0" : "=m" (p[d+32]));
|
|
asm volatile("vmovntdq %%ymm10,%0" : "=m" (p[d+64]));
|
|
asm volatile("vmovntdq %%ymm11,%0" : "=m" (p[d+96]));
|
|
asm volatile("vpxor %0,%%ymm4,%%ymm4" : : "m" (q[d]));
|
|
asm volatile("vpxor %0,%%ymm6,%%ymm6" : : "m" (q[d+32]));
|
|
asm volatile("vpxor %0,%%ymm12,%%ymm12" : : "m" (q[d+64]));
|
|
asm volatile("vpxor %0,%%ymm14,%%ymm14" : : "m" (q[d+96]));
|
|
asm volatile("vmovntdq %%ymm4,%0" : "=m" (q[d]));
|
|
asm volatile("vmovntdq %%ymm6,%0" : "=m" (q[d+32]));
|
|
asm volatile("vmovntdq %%ymm12,%0" : "=m" (q[d+64]));
|
|
asm volatile("vmovntdq %%ymm14,%0" : "=m" (q[d+96]));
|
|
}
|
|
asm volatile("sfence" : : : "memory");
|
|
kernel_fpu_end();
|
|
}
|
|
|
|
const struct raid6_calls raid6_avx2x4 = {
|
|
raid6_avx24_gen_syndrome,
|
|
raid6_avx24_xor_syndrome,
|
|
raid6_have_avx2,
|
|
"avx2x4",
|
|
1 /* Has cache hints */
|
|
};
|
|
#endif
|
|
|
|
#endif /* CONFIG_AS_AVX2 */
|