mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-16 00:46:47 +07:00
674f368a95
The CRYPTO_TFM_RES_BAD_KEY_LEN flag was apparently meant as a way to make the ->setkey() functions provide more information about errors. However, no one actually checks for this flag, which makes it pointless. Also, many algorithms fail to set this flag when given a bad length key. Reviewing just the generic implementations, this is the case for aes-fixed-time, cbcmac, echainiv, nhpoly1305, pcrypt, rfc3686, rfc4309, rfc7539, rfc7539esp, salsa20, seqiv, and xcbc. But there are probably many more in arch/*/crypto/ and drivers/crypto/. Some algorithms can even set this flag when the key is the correct length. For example, authenc and authencesn set it when the key payload is malformed in any way (not just a bad length), the atmel-sha and ccree drivers can set it if a memory allocation fails, and the chelsio driver sets it for bad auth tag lengths, not just bad key lengths. So even if someone actually wanted to start checking this flag (which seems unlikely, since it's been unused for a long time), there would be a lot of work needed to get it working correctly. But it would probably be much better to go back to the drawing board and just define different return values, like -EINVAL if the key is invalid for the algorithm vs. -EKEYREJECTED if the key was rejected by a policy like "no weak keys". That would be much simpler, less error-prone, and easier to test. So just remove this flag. Signed-off-by: Eric Biggers <ebiggers@google.com> Reviewed-by: Horia Geantă <horia.geanta@nxp.com> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
311 lines
8.2 KiB
C
311 lines
8.2 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Crypto-API module for CRC-32 algorithms implemented with the
|
|
* z/Architecture Vector Extension Facility.
|
|
*
|
|
* Copyright IBM Corp. 2015
|
|
* Author(s): Hendrik Brueckner <brueckner@linux.vnet.ibm.com>
|
|
*/
|
|
#define KMSG_COMPONENT "crc32-vx"
|
|
#define pr_fmt(fmt) KMSG_COMPONENT ": " fmt
|
|
|
|
#include <linux/module.h>
|
|
#include <linux/cpufeature.h>
|
|
#include <linux/crc32.h>
|
|
#include <crypto/internal/hash.h>
|
|
#include <asm/fpu/api.h>
|
|
|
|
|
|
#define CRC32_BLOCK_SIZE 1
|
|
#define CRC32_DIGEST_SIZE 4
|
|
|
|
#define VX_MIN_LEN 64
|
|
#define VX_ALIGNMENT 16L
|
|
#define VX_ALIGN_MASK (VX_ALIGNMENT - 1)
|
|
|
|
struct crc_ctx {
|
|
u32 key;
|
|
};
|
|
|
|
struct crc_desc_ctx {
|
|
u32 crc;
|
|
};
|
|
|
|
/* Prototypes for functions in assembly files */
|
|
u32 crc32_le_vgfm_16(u32 crc, unsigned char const *buf, size_t size);
|
|
u32 crc32_be_vgfm_16(u32 crc, unsigned char const *buf, size_t size);
|
|
u32 crc32c_le_vgfm_16(u32 crc, unsigned char const *buf, size_t size);
|
|
|
|
/*
|
|
* DEFINE_CRC32_VX() - Define a CRC-32 function using the vector extension
|
|
*
|
|
* Creates a function to perform a particular CRC-32 computation. Depending
|
|
* on the message buffer, the hardware-accelerated or software implementation
|
|
* is used. Note that the message buffer is aligned to improve fetch
|
|
* operations of VECTOR LOAD MULTIPLE instructions.
|
|
*
|
|
*/
|
|
#define DEFINE_CRC32_VX(___fname, ___crc32_vx, ___crc32_sw) \
|
|
static u32 __pure ___fname(u32 crc, \
|
|
unsigned char const *data, size_t datalen) \
|
|
{ \
|
|
struct kernel_fpu vxstate; \
|
|
unsigned long prealign, aligned, remaining; \
|
|
\
|
|
if (datalen < VX_MIN_LEN + VX_ALIGN_MASK) \
|
|
return ___crc32_sw(crc, data, datalen); \
|
|
\
|
|
if ((unsigned long)data & VX_ALIGN_MASK) { \
|
|
prealign = VX_ALIGNMENT - \
|
|
((unsigned long)data & VX_ALIGN_MASK); \
|
|
datalen -= prealign; \
|
|
crc = ___crc32_sw(crc, data, prealign); \
|
|
data = (void *)((unsigned long)data + prealign); \
|
|
} \
|
|
\
|
|
aligned = datalen & ~VX_ALIGN_MASK; \
|
|
remaining = datalen & VX_ALIGN_MASK; \
|
|
\
|
|
kernel_fpu_begin(&vxstate, KERNEL_VXR_LOW); \
|
|
crc = ___crc32_vx(crc, data, aligned); \
|
|
kernel_fpu_end(&vxstate, KERNEL_VXR_LOW); \
|
|
\
|
|
if (remaining) \
|
|
crc = ___crc32_sw(crc, data + aligned, remaining); \
|
|
\
|
|
return crc; \
|
|
}
|
|
|
|
DEFINE_CRC32_VX(crc32_le_vx, crc32_le_vgfm_16, crc32_le)
|
|
DEFINE_CRC32_VX(crc32_be_vx, crc32_be_vgfm_16, crc32_be)
|
|
DEFINE_CRC32_VX(crc32c_le_vx, crc32c_le_vgfm_16, __crc32c_le)
|
|
|
|
|
|
static int crc32_vx_cra_init_zero(struct crypto_tfm *tfm)
|
|
{
|
|
struct crc_ctx *mctx = crypto_tfm_ctx(tfm);
|
|
|
|
mctx->key = 0;
|
|
return 0;
|
|
}
|
|
|
|
static int crc32_vx_cra_init_invert(struct crypto_tfm *tfm)
|
|
{
|
|
struct crc_ctx *mctx = crypto_tfm_ctx(tfm);
|
|
|
|
mctx->key = ~0;
|
|
return 0;
|
|
}
|
|
|
|
static int crc32_vx_init(struct shash_desc *desc)
|
|
{
|
|
struct crc_ctx *mctx = crypto_shash_ctx(desc->tfm);
|
|
struct crc_desc_ctx *ctx = shash_desc_ctx(desc);
|
|
|
|
ctx->crc = mctx->key;
|
|
return 0;
|
|
}
|
|
|
|
static int crc32_vx_setkey(struct crypto_shash *tfm, const u8 *newkey,
|
|
unsigned int newkeylen)
|
|
{
|
|
struct crc_ctx *mctx = crypto_shash_ctx(tfm);
|
|
|
|
if (newkeylen != sizeof(mctx->key))
|
|
return -EINVAL;
|
|
mctx->key = le32_to_cpu(*(__le32 *)newkey);
|
|
return 0;
|
|
}
|
|
|
|
static int crc32be_vx_setkey(struct crypto_shash *tfm, const u8 *newkey,
|
|
unsigned int newkeylen)
|
|
{
|
|
struct crc_ctx *mctx = crypto_shash_ctx(tfm);
|
|
|
|
if (newkeylen != sizeof(mctx->key))
|
|
return -EINVAL;
|
|
mctx->key = be32_to_cpu(*(__be32 *)newkey);
|
|
return 0;
|
|
}
|
|
|
|
static int crc32le_vx_final(struct shash_desc *desc, u8 *out)
|
|
{
|
|
struct crc_desc_ctx *ctx = shash_desc_ctx(desc);
|
|
|
|
*(__le32 *)out = cpu_to_le32p(&ctx->crc);
|
|
return 0;
|
|
}
|
|
|
|
static int crc32be_vx_final(struct shash_desc *desc, u8 *out)
|
|
{
|
|
struct crc_desc_ctx *ctx = shash_desc_ctx(desc);
|
|
|
|
*(__be32 *)out = cpu_to_be32p(&ctx->crc);
|
|
return 0;
|
|
}
|
|
|
|
static int crc32c_vx_final(struct shash_desc *desc, u8 *out)
|
|
{
|
|
struct crc_desc_ctx *ctx = shash_desc_ctx(desc);
|
|
|
|
/*
|
|
* Perform a final XOR with 0xFFFFFFFF to be in sync
|
|
* with the generic crc32c shash implementation.
|
|
*/
|
|
*(__le32 *)out = ~cpu_to_le32p(&ctx->crc);
|
|
return 0;
|
|
}
|
|
|
|
static int __crc32le_vx_finup(u32 *crc, const u8 *data, unsigned int len,
|
|
u8 *out)
|
|
{
|
|
*(__le32 *)out = cpu_to_le32(crc32_le_vx(*crc, data, len));
|
|
return 0;
|
|
}
|
|
|
|
static int __crc32be_vx_finup(u32 *crc, const u8 *data, unsigned int len,
|
|
u8 *out)
|
|
{
|
|
*(__be32 *)out = cpu_to_be32(crc32_be_vx(*crc, data, len));
|
|
return 0;
|
|
}
|
|
|
|
static int __crc32c_vx_finup(u32 *crc, const u8 *data, unsigned int len,
|
|
u8 *out)
|
|
{
|
|
/*
|
|
* Perform a final XOR with 0xFFFFFFFF to be in sync
|
|
* with the generic crc32c shash implementation.
|
|
*/
|
|
*(__le32 *)out = ~cpu_to_le32(crc32c_le_vx(*crc, data, len));
|
|
return 0;
|
|
}
|
|
|
|
|
|
#define CRC32_VX_FINUP(alg, func) \
|
|
static int alg ## _vx_finup(struct shash_desc *desc, const u8 *data, \
|
|
unsigned int datalen, u8 *out) \
|
|
{ \
|
|
return __ ## alg ## _vx_finup(shash_desc_ctx(desc), \
|
|
data, datalen, out); \
|
|
}
|
|
|
|
CRC32_VX_FINUP(crc32le, crc32_le_vx)
|
|
CRC32_VX_FINUP(crc32be, crc32_be_vx)
|
|
CRC32_VX_FINUP(crc32c, crc32c_le_vx)
|
|
|
|
#define CRC32_VX_DIGEST(alg, func) \
|
|
static int alg ## _vx_digest(struct shash_desc *desc, const u8 *data, \
|
|
unsigned int len, u8 *out) \
|
|
{ \
|
|
return __ ## alg ## _vx_finup(crypto_shash_ctx(desc->tfm), \
|
|
data, len, out); \
|
|
}
|
|
|
|
CRC32_VX_DIGEST(crc32le, crc32_le_vx)
|
|
CRC32_VX_DIGEST(crc32be, crc32_be_vx)
|
|
CRC32_VX_DIGEST(crc32c, crc32c_le_vx)
|
|
|
|
#define CRC32_VX_UPDATE(alg, func) \
|
|
static int alg ## _vx_update(struct shash_desc *desc, const u8 *data, \
|
|
unsigned int datalen) \
|
|
{ \
|
|
struct crc_desc_ctx *ctx = shash_desc_ctx(desc); \
|
|
ctx->crc = func(ctx->crc, data, datalen); \
|
|
return 0; \
|
|
}
|
|
|
|
CRC32_VX_UPDATE(crc32le, crc32_le_vx)
|
|
CRC32_VX_UPDATE(crc32be, crc32_be_vx)
|
|
CRC32_VX_UPDATE(crc32c, crc32c_le_vx)
|
|
|
|
|
|
static struct shash_alg crc32_vx_algs[] = {
|
|
/* CRC-32 LE */
|
|
{
|
|
.init = crc32_vx_init,
|
|
.setkey = crc32_vx_setkey,
|
|
.update = crc32le_vx_update,
|
|
.final = crc32le_vx_final,
|
|
.finup = crc32le_vx_finup,
|
|
.digest = crc32le_vx_digest,
|
|
.descsize = sizeof(struct crc_desc_ctx),
|
|
.digestsize = CRC32_DIGEST_SIZE,
|
|
.base = {
|
|
.cra_name = "crc32",
|
|
.cra_driver_name = "crc32-vx",
|
|
.cra_priority = 200,
|
|
.cra_flags = CRYPTO_ALG_OPTIONAL_KEY,
|
|
.cra_blocksize = CRC32_BLOCK_SIZE,
|
|
.cra_ctxsize = sizeof(struct crc_ctx),
|
|
.cra_module = THIS_MODULE,
|
|
.cra_init = crc32_vx_cra_init_zero,
|
|
},
|
|
},
|
|
/* CRC-32 BE */
|
|
{
|
|
.init = crc32_vx_init,
|
|
.setkey = crc32be_vx_setkey,
|
|
.update = crc32be_vx_update,
|
|
.final = crc32be_vx_final,
|
|
.finup = crc32be_vx_finup,
|
|
.digest = crc32be_vx_digest,
|
|
.descsize = sizeof(struct crc_desc_ctx),
|
|
.digestsize = CRC32_DIGEST_SIZE,
|
|
.base = {
|
|
.cra_name = "crc32be",
|
|
.cra_driver_name = "crc32be-vx",
|
|
.cra_priority = 200,
|
|
.cra_flags = CRYPTO_ALG_OPTIONAL_KEY,
|
|
.cra_blocksize = CRC32_BLOCK_SIZE,
|
|
.cra_ctxsize = sizeof(struct crc_ctx),
|
|
.cra_module = THIS_MODULE,
|
|
.cra_init = crc32_vx_cra_init_zero,
|
|
},
|
|
},
|
|
/* CRC-32C LE */
|
|
{
|
|
.init = crc32_vx_init,
|
|
.setkey = crc32_vx_setkey,
|
|
.update = crc32c_vx_update,
|
|
.final = crc32c_vx_final,
|
|
.finup = crc32c_vx_finup,
|
|
.digest = crc32c_vx_digest,
|
|
.descsize = sizeof(struct crc_desc_ctx),
|
|
.digestsize = CRC32_DIGEST_SIZE,
|
|
.base = {
|
|
.cra_name = "crc32c",
|
|
.cra_driver_name = "crc32c-vx",
|
|
.cra_priority = 200,
|
|
.cra_flags = CRYPTO_ALG_OPTIONAL_KEY,
|
|
.cra_blocksize = CRC32_BLOCK_SIZE,
|
|
.cra_ctxsize = sizeof(struct crc_ctx),
|
|
.cra_module = THIS_MODULE,
|
|
.cra_init = crc32_vx_cra_init_invert,
|
|
},
|
|
},
|
|
};
|
|
|
|
|
|
static int __init crc_vx_mod_init(void)
|
|
{
|
|
return crypto_register_shashes(crc32_vx_algs,
|
|
ARRAY_SIZE(crc32_vx_algs));
|
|
}
|
|
|
|
static void __exit crc_vx_mod_exit(void)
|
|
{
|
|
crypto_unregister_shashes(crc32_vx_algs, ARRAY_SIZE(crc32_vx_algs));
|
|
}
|
|
|
|
module_cpu_feature_match(VXRS, crc_vx_mod_init);
|
|
module_exit(crc_vx_mod_exit);
|
|
|
|
MODULE_AUTHOR("Hendrik Brueckner <brueckner@linux.vnet.ibm.com>");
|
|
MODULE_LICENSE("GPL");
|
|
|
|
MODULE_ALIAS_CRYPTO("crc32");
|
|
MODULE_ALIAS_CRYPTO("crc32-vx");
|
|
MODULE_ALIAS_CRYPTO("crc32c");
|
|
MODULE_ALIAS_CRYPTO("crc32c-vx");
|