mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-19 10:06:45 +07:00
a876ca4dde
Now that the SPDX tag is in all arch/s390/crypto/ files, that identifies the license in a specific and legally-defined manner. So the extra GPL text wording can be removed as it is no longer needed at all. This is done on a quest to remove the 700+ different ways that files in the kernel describe the GPL license text. And there's unneeded stuff like the address (sometimes incorrect) for the FSF which is never needed. No copyright headers or other non-license-description text was removed. Cc: Herbert Xu <herbert@gondor.apana.org.au> Cc: "David S. Miller" <davem@davemloft.net> Cc: Martin Schwidefsky <schwidefsky@de.ibm.com> Cc: Heiko Carstens <heiko.carstens@de.ibm.com> Cc: "Paul E. McKenney" <paulmck@linux.vnet.ibm.com> Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org> Signed-off-by: Martin Schwidefsky <schwidefsky@de.ibm.com>
619 lines
17 KiB
C
619 lines
17 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Cryptographic API.
|
|
*
|
|
* s390 implementation of the AES Cipher Algorithm with protected keys.
|
|
*
|
|
* s390 Version:
|
|
* Copyright IBM Corp. 2017
|
|
* Author(s): Martin Schwidefsky <schwidefsky@de.ibm.com>
|
|
* Harald Freudenberger <freude@de.ibm.com>
|
|
*/
|
|
|
|
#define KMSG_COMPONENT "paes_s390"
|
|
#define pr_fmt(fmt) KMSG_COMPONENT ": " fmt
|
|
|
|
#include <crypto/aes.h>
|
|
#include <crypto/algapi.h>
|
|
#include <linux/bug.h>
|
|
#include <linux/err.h>
|
|
#include <linux/module.h>
|
|
#include <linux/cpufeature.h>
|
|
#include <linux/init.h>
|
|
#include <linux/spinlock.h>
|
|
#include <crypto/xts.h>
|
|
#include <asm/cpacf.h>
|
|
#include <asm/pkey.h>
|
|
|
|
static u8 *ctrblk;
|
|
static DEFINE_SPINLOCK(ctrblk_lock);
|
|
|
|
static cpacf_mask_t km_functions, kmc_functions, kmctr_functions;
|
|
|
|
struct s390_paes_ctx {
|
|
struct pkey_seckey sk;
|
|
struct pkey_protkey pk;
|
|
unsigned long fc;
|
|
};
|
|
|
|
struct s390_pxts_ctx {
|
|
struct pkey_seckey sk[2];
|
|
struct pkey_protkey pk[2];
|
|
unsigned long fc;
|
|
};
|
|
|
|
static inline int __paes_convert_key(struct pkey_seckey *sk,
|
|
struct pkey_protkey *pk)
|
|
{
|
|
int i, ret;
|
|
|
|
/* try three times in case of failure */
|
|
for (i = 0; i < 3; i++) {
|
|
ret = pkey_skey2pkey(sk, pk);
|
|
if (ret == 0)
|
|
break;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int __paes_set_key(struct s390_paes_ctx *ctx)
|
|
{
|
|
unsigned long fc;
|
|
|
|
if (__paes_convert_key(&ctx->sk, &ctx->pk))
|
|
return -EINVAL;
|
|
|
|
/* Pick the correct function code based on the protected key type */
|
|
fc = (ctx->pk.type == PKEY_KEYTYPE_AES_128) ? CPACF_KM_PAES_128 :
|
|
(ctx->pk.type == PKEY_KEYTYPE_AES_192) ? CPACF_KM_PAES_192 :
|
|
(ctx->pk.type == PKEY_KEYTYPE_AES_256) ? CPACF_KM_PAES_256 : 0;
|
|
|
|
/* Check if the function code is available */
|
|
ctx->fc = (fc && cpacf_test_func(&km_functions, fc)) ? fc : 0;
|
|
|
|
return ctx->fc ? 0 : -EINVAL;
|
|
}
|
|
|
|
static int ecb_paes_set_key(struct crypto_tfm *tfm, const u8 *in_key,
|
|
unsigned int key_len)
|
|
{
|
|
struct s390_paes_ctx *ctx = crypto_tfm_ctx(tfm);
|
|
|
|
if (key_len != SECKEYBLOBSIZE)
|
|
return -EINVAL;
|
|
|
|
memcpy(ctx->sk.seckey, in_key, SECKEYBLOBSIZE);
|
|
if (__paes_set_key(ctx)) {
|
|
tfm->crt_flags |= CRYPTO_TFM_RES_BAD_KEY_LEN;
|
|
return -EINVAL;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int ecb_paes_crypt(struct blkcipher_desc *desc,
|
|
unsigned long modifier,
|
|
struct blkcipher_walk *walk)
|
|
{
|
|
struct s390_paes_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
|
|
unsigned int nbytes, n, k;
|
|
int ret;
|
|
|
|
ret = blkcipher_walk_virt(desc, walk);
|
|
while ((nbytes = walk->nbytes) >= AES_BLOCK_SIZE) {
|
|
/* only use complete blocks */
|
|
n = nbytes & ~(AES_BLOCK_SIZE - 1);
|
|
k = cpacf_km(ctx->fc | modifier, ctx->pk.protkey,
|
|
walk->dst.virt.addr, walk->src.virt.addr, n);
|
|
if (k)
|
|
ret = blkcipher_walk_done(desc, walk, nbytes - k);
|
|
if (k < n) {
|
|
if (__paes_set_key(ctx) != 0)
|
|
return blkcipher_walk_done(desc, walk, -EIO);
|
|
}
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static int ecb_paes_encrypt(struct blkcipher_desc *desc,
|
|
struct scatterlist *dst, struct scatterlist *src,
|
|
unsigned int nbytes)
|
|
{
|
|
struct blkcipher_walk walk;
|
|
|
|
blkcipher_walk_init(&walk, dst, src, nbytes);
|
|
return ecb_paes_crypt(desc, CPACF_ENCRYPT, &walk);
|
|
}
|
|
|
|
static int ecb_paes_decrypt(struct blkcipher_desc *desc,
|
|
struct scatterlist *dst, struct scatterlist *src,
|
|
unsigned int nbytes)
|
|
{
|
|
struct blkcipher_walk walk;
|
|
|
|
blkcipher_walk_init(&walk, dst, src, nbytes);
|
|
return ecb_paes_crypt(desc, CPACF_DECRYPT, &walk);
|
|
}
|
|
|
|
static struct crypto_alg ecb_paes_alg = {
|
|
.cra_name = "ecb(paes)",
|
|
.cra_driver_name = "ecb-paes-s390",
|
|
.cra_priority = 400, /* combo: aes + ecb */
|
|
.cra_flags = CRYPTO_ALG_TYPE_BLKCIPHER,
|
|
.cra_blocksize = AES_BLOCK_SIZE,
|
|
.cra_ctxsize = sizeof(struct s390_paes_ctx),
|
|
.cra_type = &crypto_blkcipher_type,
|
|
.cra_module = THIS_MODULE,
|
|
.cra_list = LIST_HEAD_INIT(ecb_paes_alg.cra_list),
|
|
.cra_u = {
|
|
.blkcipher = {
|
|
.min_keysize = SECKEYBLOBSIZE,
|
|
.max_keysize = SECKEYBLOBSIZE,
|
|
.setkey = ecb_paes_set_key,
|
|
.encrypt = ecb_paes_encrypt,
|
|
.decrypt = ecb_paes_decrypt,
|
|
}
|
|
}
|
|
};
|
|
|
|
static int __cbc_paes_set_key(struct s390_paes_ctx *ctx)
|
|
{
|
|
unsigned long fc;
|
|
|
|
if (__paes_convert_key(&ctx->sk, &ctx->pk))
|
|
return -EINVAL;
|
|
|
|
/* Pick the correct function code based on the protected key type */
|
|
fc = (ctx->pk.type == PKEY_KEYTYPE_AES_128) ? CPACF_KMC_PAES_128 :
|
|
(ctx->pk.type == PKEY_KEYTYPE_AES_192) ? CPACF_KMC_PAES_192 :
|
|
(ctx->pk.type == PKEY_KEYTYPE_AES_256) ? CPACF_KMC_PAES_256 : 0;
|
|
|
|
/* Check if the function code is available */
|
|
ctx->fc = (fc && cpacf_test_func(&kmc_functions, fc)) ? fc : 0;
|
|
|
|
return ctx->fc ? 0 : -EINVAL;
|
|
}
|
|
|
|
static int cbc_paes_set_key(struct crypto_tfm *tfm, const u8 *in_key,
|
|
unsigned int key_len)
|
|
{
|
|
struct s390_paes_ctx *ctx = crypto_tfm_ctx(tfm);
|
|
|
|
memcpy(ctx->sk.seckey, in_key, SECKEYBLOBSIZE);
|
|
if (__cbc_paes_set_key(ctx)) {
|
|
tfm->crt_flags |= CRYPTO_TFM_RES_BAD_KEY_LEN;
|
|
return -EINVAL;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int cbc_paes_crypt(struct blkcipher_desc *desc, unsigned long modifier,
|
|
struct blkcipher_walk *walk)
|
|
{
|
|
struct s390_paes_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
|
|
unsigned int nbytes, n, k;
|
|
int ret;
|
|
struct {
|
|
u8 iv[AES_BLOCK_SIZE];
|
|
u8 key[MAXPROTKEYSIZE];
|
|
} param;
|
|
|
|
ret = blkcipher_walk_virt(desc, walk);
|
|
memcpy(param.iv, walk->iv, AES_BLOCK_SIZE);
|
|
memcpy(param.key, ctx->pk.protkey, MAXPROTKEYSIZE);
|
|
while ((nbytes = walk->nbytes) >= AES_BLOCK_SIZE) {
|
|
/* only use complete blocks */
|
|
n = nbytes & ~(AES_BLOCK_SIZE - 1);
|
|
k = cpacf_kmc(ctx->fc | modifier, ¶m,
|
|
walk->dst.virt.addr, walk->src.virt.addr, n);
|
|
if (k)
|
|
ret = blkcipher_walk_done(desc, walk, nbytes - k);
|
|
if (n < k) {
|
|
if (__cbc_paes_set_key(ctx) != 0)
|
|
return blkcipher_walk_done(desc, walk, -EIO);
|
|
memcpy(param.key, ctx->pk.protkey, MAXPROTKEYSIZE);
|
|
}
|
|
}
|
|
memcpy(walk->iv, param.iv, AES_BLOCK_SIZE);
|
|
return ret;
|
|
}
|
|
|
|
static int cbc_paes_encrypt(struct blkcipher_desc *desc,
|
|
struct scatterlist *dst, struct scatterlist *src,
|
|
unsigned int nbytes)
|
|
{
|
|
struct blkcipher_walk walk;
|
|
|
|
blkcipher_walk_init(&walk, dst, src, nbytes);
|
|
return cbc_paes_crypt(desc, 0, &walk);
|
|
}
|
|
|
|
static int cbc_paes_decrypt(struct blkcipher_desc *desc,
|
|
struct scatterlist *dst, struct scatterlist *src,
|
|
unsigned int nbytes)
|
|
{
|
|
struct blkcipher_walk walk;
|
|
|
|
blkcipher_walk_init(&walk, dst, src, nbytes);
|
|
return cbc_paes_crypt(desc, CPACF_DECRYPT, &walk);
|
|
}
|
|
|
|
static struct crypto_alg cbc_paes_alg = {
|
|
.cra_name = "cbc(paes)",
|
|
.cra_driver_name = "cbc-paes-s390",
|
|
.cra_priority = 400, /* combo: aes + cbc */
|
|
.cra_flags = CRYPTO_ALG_TYPE_BLKCIPHER,
|
|
.cra_blocksize = AES_BLOCK_SIZE,
|
|
.cra_ctxsize = sizeof(struct s390_paes_ctx),
|
|
.cra_type = &crypto_blkcipher_type,
|
|
.cra_module = THIS_MODULE,
|
|
.cra_list = LIST_HEAD_INIT(cbc_paes_alg.cra_list),
|
|
.cra_u = {
|
|
.blkcipher = {
|
|
.min_keysize = SECKEYBLOBSIZE,
|
|
.max_keysize = SECKEYBLOBSIZE,
|
|
.ivsize = AES_BLOCK_SIZE,
|
|
.setkey = cbc_paes_set_key,
|
|
.encrypt = cbc_paes_encrypt,
|
|
.decrypt = cbc_paes_decrypt,
|
|
}
|
|
}
|
|
};
|
|
|
|
static int __xts_paes_set_key(struct s390_pxts_ctx *ctx)
|
|
{
|
|
unsigned long fc;
|
|
|
|
if (__paes_convert_key(&ctx->sk[0], &ctx->pk[0]) ||
|
|
__paes_convert_key(&ctx->sk[1], &ctx->pk[1]))
|
|
return -EINVAL;
|
|
|
|
if (ctx->pk[0].type != ctx->pk[1].type)
|
|
return -EINVAL;
|
|
|
|
/* Pick the correct function code based on the protected key type */
|
|
fc = (ctx->pk[0].type == PKEY_KEYTYPE_AES_128) ? CPACF_KM_PXTS_128 :
|
|
(ctx->pk[0].type == PKEY_KEYTYPE_AES_256) ?
|
|
CPACF_KM_PXTS_256 : 0;
|
|
|
|
/* Check if the function code is available */
|
|
ctx->fc = (fc && cpacf_test_func(&km_functions, fc)) ? fc : 0;
|
|
|
|
return ctx->fc ? 0 : -EINVAL;
|
|
}
|
|
|
|
static int xts_paes_set_key(struct crypto_tfm *tfm, const u8 *in_key,
|
|
unsigned int key_len)
|
|
{
|
|
struct s390_pxts_ctx *ctx = crypto_tfm_ctx(tfm);
|
|
u8 ckey[2 * AES_MAX_KEY_SIZE];
|
|
unsigned int ckey_len;
|
|
|
|
memcpy(ctx->sk[0].seckey, in_key, SECKEYBLOBSIZE);
|
|
memcpy(ctx->sk[1].seckey, in_key + SECKEYBLOBSIZE, SECKEYBLOBSIZE);
|
|
if (__xts_paes_set_key(ctx)) {
|
|
tfm->crt_flags |= CRYPTO_TFM_RES_BAD_KEY_LEN;
|
|
return -EINVAL;
|
|
}
|
|
|
|
/*
|
|
* xts_check_key verifies the key length is not odd and makes
|
|
* sure that the two keys are not the same. This can be done
|
|
* on the two protected keys as well
|
|
*/
|
|
ckey_len = (ctx->pk[0].type == PKEY_KEYTYPE_AES_128) ?
|
|
AES_KEYSIZE_128 : AES_KEYSIZE_256;
|
|
memcpy(ckey, ctx->pk[0].protkey, ckey_len);
|
|
memcpy(ckey + ckey_len, ctx->pk[1].protkey, ckey_len);
|
|
return xts_check_key(tfm, ckey, 2*ckey_len);
|
|
}
|
|
|
|
static int xts_paes_crypt(struct blkcipher_desc *desc, unsigned long modifier,
|
|
struct blkcipher_walk *walk)
|
|
{
|
|
struct s390_pxts_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
|
|
unsigned int keylen, offset, nbytes, n, k;
|
|
int ret;
|
|
struct {
|
|
u8 key[MAXPROTKEYSIZE]; /* key + verification pattern */
|
|
u8 tweak[16];
|
|
u8 block[16];
|
|
u8 bit[16];
|
|
u8 xts[16];
|
|
} pcc_param;
|
|
struct {
|
|
u8 key[MAXPROTKEYSIZE]; /* key + verification pattern */
|
|
u8 init[16];
|
|
} xts_param;
|
|
|
|
ret = blkcipher_walk_virt(desc, walk);
|
|
keylen = (ctx->pk[0].type == PKEY_KEYTYPE_AES_128) ? 48 : 64;
|
|
offset = (ctx->pk[0].type == PKEY_KEYTYPE_AES_128) ? 16 : 0;
|
|
retry:
|
|
memset(&pcc_param, 0, sizeof(pcc_param));
|
|
memcpy(pcc_param.tweak, walk->iv, sizeof(pcc_param.tweak));
|
|
memcpy(pcc_param.key + offset, ctx->pk[1].protkey, keylen);
|
|
cpacf_pcc(ctx->fc, pcc_param.key + offset);
|
|
|
|
memcpy(xts_param.key + offset, ctx->pk[0].protkey, keylen);
|
|
memcpy(xts_param.init, pcc_param.xts, 16);
|
|
|
|
while ((nbytes = walk->nbytes) >= AES_BLOCK_SIZE) {
|
|
/* only use complete blocks */
|
|
n = nbytes & ~(AES_BLOCK_SIZE - 1);
|
|
k = cpacf_km(ctx->fc | modifier, xts_param.key + offset,
|
|
walk->dst.virt.addr, walk->src.virt.addr, n);
|
|
if (k)
|
|
ret = blkcipher_walk_done(desc, walk, nbytes - k);
|
|
if (k < n) {
|
|
if (__xts_paes_set_key(ctx) != 0)
|
|
return blkcipher_walk_done(desc, walk, -EIO);
|
|
goto retry;
|
|
}
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static int xts_paes_encrypt(struct blkcipher_desc *desc,
|
|
struct scatterlist *dst, struct scatterlist *src,
|
|
unsigned int nbytes)
|
|
{
|
|
struct blkcipher_walk walk;
|
|
|
|
blkcipher_walk_init(&walk, dst, src, nbytes);
|
|
return xts_paes_crypt(desc, 0, &walk);
|
|
}
|
|
|
|
static int xts_paes_decrypt(struct blkcipher_desc *desc,
|
|
struct scatterlist *dst, struct scatterlist *src,
|
|
unsigned int nbytes)
|
|
{
|
|
struct blkcipher_walk walk;
|
|
|
|
blkcipher_walk_init(&walk, dst, src, nbytes);
|
|
return xts_paes_crypt(desc, CPACF_DECRYPT, &walk);
|
|
}
|
|
|
|
static struct crypto_alg xts_paes_alg = {
|
|
.cra_name = "xts(paes)",
|
|
.cra_driver_name = "xts-paes-s390",
|
|
.cra_priority = 400, /* combo: aes + xts */
|
|
.cra_flags = CRYPTO_ALG_TYPE_BLKCIPHER,
|
|
.cra_blocksize = AES_BLOCK_SIZE,
|
|
.cra_ctxsize = sizeof(struct s390_pxts_ctx),
|
|
.cra_type = &crypto_blkcipher_type,
|
|
.cra_module = THIS_MODULE,
|
|
.cra_list = LIST_HEAD_INIT(xts_paes_alg.cra_list),
|
|
.cra_u = {
|
|
.blkcipher = {
|
|
.min_keysize = 2 * SECKEYBLOBSIZE,
|
|
.max_keysize = 2 * SECKEYBLOBSIZE,
|
|
.ivsize = AES_BLOCK_SIZE,
|
|
.setkey = xts_paes_set_key,
|
|
.encrypt = xts_paes_encrypt,
|
|
.decrypt = xts_paes_decrypt,
|
|
}
|
|
}
|
|
};
|
|
|
|
static int __ctr_paes_set_key(struct s390_paes_ctx *ctx)
|
|
{
|
|
unsigned long fc;
|
|
|
|
if (__paes_convert_key(&ctx->sk, &ctx->pk))
|
|
return -EINVAL;
|
|
|
|
/* Pick the correct function code based on the protected key type */
|
|
fc = (ctx->pk.type == PKEY_KEYTYPE_AES_128) ? CPACF_KMCTR_PAES_128 :
|
|
(ctx->pk.type == PKEY_KEYTYPE_AES_192) ? CPACF_KMCTR_PAES_192 :
|
|
(ctx->pk.type == PKEY_KEYTYPE_AES_256) ?
|
|
CPACF_KMCTR_PAES_256 : 0;
|
|
|
|
/* Check if the function code is available */
|
|
ctx->fc = (fc && cpacf_test_func(&kmctr_functions, fc)) ? fc : 0;
|
|
|
|
return ctx->fc ? 0 : -EINVAL;
|
|
}
|
|
|
|
static int ctr_paes_set_key(struct crypto_tfm *tfm, const u8 *in_key,
|
|
unsigned int key_len)
|
|
{
|
|
struct s390_paes_ctx *ctx = crypto_tfm_ctx(tfm);
|
|
|
|
memcpy(ctx->sk.seckey, in_key, key_len);
|
|
if (__ctr_paes_set_key(ctx)) {
|
|
tfm->crt_flags |= CRYPTO_TFM_RES_BAD_KEY_LEN;
|
|
return -EINVAL;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static unsigned int __ctrblk_init(u8 *ctrptr, u8 *iv, unsigned int nbytes)
|
|
{
|
|
unsigned int i, n;
|
|
|
|
/* only use complete blocks, max. PAGE_SIZE */
|
|
memcpy(ctrptr, iv, AES_BLOCK_SIZE);
|
|
n = (nbytes > PAGE_SIZE) ? PAGE_SIZE : nbytes & ~(AES_BLOCK_SIZE - 1);
|
|
for (i = (n / AES_BLOCK_SIZE) - 1; i > 0; i--) {
|
|
memcpy(ctrptr + AES_BLOCK_SIZE, ctrptr, AES_BLOCK_SIZE);
|
|
crypto_inc(ctrptr + AES_BLOCK_SIZE, AES_BLOCK_SIZE);
|
|
ctrptr += AES_BLOCK_SIZE;
|
|
}
|
|
return n;
|
|
}
|
|
|
|
static int ctr_paes_crypt(struct blkcipher_desc *desc, unsigned long modifier,
|
|
struct blkcipher_walk *walk)
|
|
{
|
|
struct s390_paes_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
|
|
u8 buf[AES_BLOCK_SIZE], *ctrptr;
|
|
unsigned int nbytes, n, k;
|
|
int ret, locked;
|
|
|
|
locked = spin_trylock(&ctrblk_lock);
|
|
|
|
ret = blkcipher_walk_virt_block(desc, walk, AES_BLOCK_SIZE);
|
|
while ((nbytes = walk->nbytes) >= AES_BLOCK_SIZE) {
|
|
n = AES_BLOCK_SIZE;
|
|
if (nbytes >= 2*AES_BLOCK_SIZE && locked)
|
|
n = __ctrblk_init(ctrblk, walk->iv, nbytes);
|
|
ctrptr = (n > AES_BLOCK_SIZE) ? ctrblk : walk->iv;
|
|
k = cpacf_kmctr(ctx->fc | modifier, ctx->pk.protkey,
|
|
walk->dst.virt.addr, walk->src.virt.addr,
|
|
n, ctrptr);
|
|
if (k) {
|
|
if (ctrptr == ctrblk)
|
|
memcpy(walk->iv, ctrptr + k - AES_BLOCK_SIZE,
|
|
AES_BLOCK_SIZE);
|
|
crypto_inc(walk->iv, AES_BLOCK_SIZE);
|
|
ret = blkcipher_walk_done(desc, walk, nbytes - n);
|
|
}
|
|
if (k < n) {
|
|
if (__ctr_paes_set_key(ctx) != 0) {
|
|
if (locked)
|
|
spin_unlock(&ctrblk_lock);
|
|
return blkcipher_walk_done(desc, walk, -EIO);
|
|
}
|
|
}
|
|
}
|
|
if (locked)
|
|
spin_unlock(&ctrblk_lock);
|
|
/*
|
|
* final block may be < AES_BLOCK_SIZE, copy only nbytes
|
|
*/
|
|
if (nbytes) {
|
|
while (1) {
|
|
if (cpacf_kmctr(ctx->fc | modifier,
|
|
ctx->pk.protkey, buf,
|
|
walk->src.virt.addr, AES_BLOCK_SIZE,
|
|
walk->iv) == AES_BLOCK_SIZE)
|
|
break;
|
|
if (__ctr_paes_set_key(ctx) != 0)
|
|
return blkcipher_walk_done(desc, walk, -EIO);
|
|
}
|
|
memcpy(walk->dst.virt.addr, buf, nbytes);
|
|
crypto_inc(walk->iv, AES_BLOCK_SIZE);
|
|
ret = blkcipher_walk_done(desc, walk, 0);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ctr_paes_encrypt(struct blkcipher_desc *desc,
|
|
struct scatterlist *dst, struct scatterlist *src,
|
|
unsigned int nbytes)
|
|
{
|
|
struct blkcipher_walk walk;
|
|
|
|
blkcipher_walk_init(&walk, dst, src, nbytes);
|
|
return ctr_paes_crypt(desc, 0, &walk);
|
|
}
|
|
|
|
static int ctr_paes_decrypt(struct blkcipher_desc *desc,
|
|
struct scatterlist *dst, struct scatterlist *src,
|
|
unsigned int nbytes)
|
|
{
|
|
struct blkcipher_walk walk;
|
|
|
|
blkcipher_walk_init(&walk, dst, src, nbytes);
|
|
return ctr_paes_crypt(desc, CPACF_DECRYPT, &walk);
|
|
}
|
|
|
|
static struct crypto_alg ctr_paes_alg = {
|
|
.cra_name = "ctr(paes)",
|
|
.cra_driver_name = "ctr-paes-s390",
|
|
.cra_priority = 400, /* combo: aes + ctr */
|
|
.cra_flags = CRYPTO_ALG_TYPE_BLKCIPHER,
|
|
.cra_blocksize = 1,
|
|
.cra_ctxsize = sizeof(struct s390_paes_ctx),
|
|
.cra_type = &crypto_blkcipher_type,
|
|
.cra_module = THIS_MODULE,
|
|
.cra_list = LIST_HEAD_INIT(ctr_paes_alg.cra_list),
|
|
.cra_u = {
|
|
.blkcipher = {
|
|
.min_keysize = SECKEYBLOBSIZE,
|
|
.max_keysize = SECKEYBLOBSIZE,
|
|
.ivsize = AES_BLOCK_SIZE,
|
|
.setkey = ctr_paes_set_key,
|
|
.encrypt = ctr_paes_encrypt,
|
|
.decrypt = ctr_paes_decrypt,
|
|
}
|
|
}
|
|
};
|
|
|
|
static inline void __crypto_unregister_alg(struct crypto_alg *alg)
|
|
{
|
|
if (!list_empty(&alg->cra_list))
|
|
crypto_unregister_alg(alg);
|
|
}
|
|
|
|
static void paes_s390_fini(void)
|
|
{
|
|
if (ctrblk)
|
|
free_page((unsigned long) ctrblk);
|
|
__crypto_unregister_alg(&ctr_paes_alg);
|
|
__crypto_unregister_alg(&xts_paes_alg);
|
|
__crypto_unregister_alg(&cbc_paes_alg);
|
|
__crypto_unregister_alg(&ecb_paes_alg);
|
|
}
|
|
|
|
static int __init paes_s390_init(void)
|
|
{
|
|
int ret;
|
|
|
|
/* Query available functions for KM, KMC and KMCTR */
|
|
cpacf_query(CPACF_KM, &km_functions);
|
|
cpacf_query(CPACF_KMC, &kmc_functions);
|
|
cpacf_query(CPACF_KMCTR, &kmctr_functions);
|
|
|
|
if (cpacf_test_func(&km_functions, CPACF_KM_PAES_128) ||
|
|
cpacf_test_func(&km_functions, CPACF_KM_PAES_192) ||
|
|
cpacf_test_func(&km_functions, CPACF_KM_PAES_256)) {
|
|
ret = crypto_register_alg(&ecb_paes_alg);
|
|
if (ret)
|
|
goto out_err;
|
|
}
|
|
|
|
if (cpacf_test_func(&kmc_functions, CPACF_KMC_PAES_128) ||
|
|
cpacf_test_func(&kmc_functions, CPACF_KMC_PAES_192) ||
|
|
cpacf_test_func(&kmc_functions, CPACF_KMC_PAES_256)) {
|
|
ret = crypto_register_alg(&cbc_paes_alg);
|
|
if (ret)
|
|
goto out_err;
|
|
}
|
|
|
|
if (cpacf_test_func(&km_functions, CPACF_KM_PXTS_128) ||
|
|
cpacf_test_func(&km_functions, CPACF_KM_PXTS_256)) {
|
|
ret = crypto_register_alg(&xts_paes_alg);
|
|
if (ret)
|
|
goto out_err;
|
|
}
|
|
|
|
if (cpacf_test_func(&kmctr_functions, CPACF_KMCTR_PAES_128) ||
|
|
cpacf_test_func(&kmctr_functions, CPACF_KMCTR_PAES_192) ||
|
|
cpacf_test_func(&kmctr_functions, CPACF_KMCTR_PAES_256)) {
|
|
ret = crypto_register_alg(&ctr_paes_alg);
|
|
if (ret)
|
|
goto out_err;
|
|
ctrblk = (u8 *) __get_free_page(GFP_KERNEL);
|
|
if (!ctrblk) {
|
|
ret = -ENOMEM;
|
|
goto out_err;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
out_err:
|
|
paes_s390_fini();
|
|
return ret;
|
|
}
|
|
|
|
module_init(paes_s390_init);
|
|
module_exit(paes_s390_fini);
|
|
|
|
MODULE_ALIAS_CRYPTO("paes");
|
|
|
|
MODULE_DESCRIPTION("Rijndael (AES) Cipher Algorithm with protected keys");
|
|
MODULE_LICENSE("GPL");
|