mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-11-24 16:50:54 +07:00
crypto: skcipher - make chunksize and walksize accessors internal
The 'chunksize' and 'walksize' properties of skcipher algorithms are implementation details that users of the skcipher API should not be looking at. So move their accessor functions from <crypto/skcipher.h> to <crypto/internal/skcipher.h>. Signed-off-by: Eric Biggers <ebiggers@google.com> Acked-by: Ard Biesheuvel <ard.biesheuvel@linaro.org> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
This commit is contained in:
parent
81bcbb1ee7
commit
314d0f0ea6
@ -205,6 +205,66 @@ static inline unsigned int crypto_skcipher_alg_max_keysize(
|
||||
return alg->max_keysize;
|
||||
}
|
||||
|
||||
static inline unsigned int crypto_skcipher_alg_chunksize(
|
||||
struct skcipher_alg *alg)
|
||||
{
|
||||
if ((alg->base.cra_flags & CRYPTO_ALG_TYPE_MASK) ==
|
||||
CRYPTO_ALG_TYPE_BLKCIPHER)
|
||||
return alg->base.cra_blocksize;
|
||||
|
||||
if (alg->base.cra_ablkcipher.encrypt)
|
||||
return alg->base.cra_blocksize;
|
||||
|
||||
return alg->chunksize;
|
||||
}
|
||||
|
||||
static inline unsigned int crypto_skcipher_alg_walksize(
|
||||
struct skcipher_alg *alg)
|
||||
{
|
||||
if ((alg->base.cra_flags & CRYPTO_ALG_TYPE_MASK) ==
|
||||
CRYPTO_ALG_TYPE_BLKCIPHER)
|
||||
return alg->base.cra_blocksize;
|
||||
|
||||
if (alg->base.cra_ablkcipher.encrypt)
|
||||
return alg->base.cra_blocksize;
|
||||
|
||||
return alg->walksize;
|
||||
}
|
||||
|
||||
/**
|
||||
* crypto_skcipher_chunksize() - obtain chunk size
|
||||
* @tfm: cipher handle
|
||||
*
|
||||
* The block size is set to one for ciphers such as CTR. However,
|
||||
* you still need to provide incremental updates in multiples of
|
||||
* the underlying block size as the IV does not have sub-block
|
||||
* granularity. This is known in this API as the chunk size.
|
||||
*
|
||||
* Return: chunk size in bytes
|
||||
*/
|
||||
static inline unsigned int crypto_skcipher_chunksize(
|
||||
struct crypto_skcipher *tfm)
|
||||
{
|
||||
return crypto_skcipher_alg_chunksize(crypto_skcipher_alg(tfm));
|
||||
}
|
||||
|
||||
/**
|
||||
* crypto_skcipher_walksize() - obtain walk size
|
||||
* @tfm: cipher handle
|
||||
*
|
||||
* In some cases, algorithms can only perform optimally when operating on
|
||||
* multiple blocks in parallel. This is reflected by the walksize, which
|
||||
* must be a multiple of the chunksize (or equal if the concern does not
|
||||
* apply)
|
||||
*
|
||||
* Return: walk size in bytes
|
||||
*/
|
||||
static inline unsigned int crypto_skcipher_walksize(
|
||||
struct crypto_skcipher *tfm)
|
||||
{
|
||||
return crypto_skcipher_alg_walksize(crypto_skcipher_alg(tfm));
|
||||
}
|
||||
|
||||
/* Helpers for simple block cipher modes of operation */
|
||||
struct skcipher_ctx_simple {
|
||||
struct crypto_cipher *cipher; /* underlying block cipher */
|
||||
|
@ -293,66 +293,6 @@ static inline unsigned int crypto_sync_skcipher_ivsize(
|
||||
return crypto_skcipher_ivsize(&tfm->base);
|
||||
}
|
||||
|
||||
static inline unsigned int crypto_skcipher_alg_chunksize(
|
||||
struct skcipher_alg *alg)
|
||||
{
|
||||
if ((alg->base.cra_flags & CRYPTO_ALG_TYPE_MASK) ==
|
||||
CRYPTO_ALG_TYPE_BLKCIPHER)
|
||||
return alg->base.cra_blocksize;
|
||||
|
||||
if (alg->base.cra_ablkcipher.encrypt)
|
||||
return alg->base.cra_blocksize;
|
||||
|
||||
return alg->chunksize;
|
||||
}
|
||||
|
||||
static inline unsigned int crypto_skcipher_alg_walksize(
|
||||
struct skcipher_alg *alg)
|
||||
{
|
||||
if ((alg->base.cra_flags & CRYPTO_ALG_TYPE_MASK) ==
|
||||
CRYPTO_ALG_TYPE_BLKCIPHER)
|
||||
return alg->base.cra_blocksize;
|
||||
|
||||
if (alg->base.cra_ablkcipher.encrypt)
|
||||
return alg->base.cra_blocksize;
|
||||
|
||||
return alg->walksize;
|
||||
}
|
||||
|
||||
/**
|
||||
* crypto_skcipher_chunksize() - obtain chunk size
|
||||
* @tfm: cipher handle
|
||||
*
|
||||
* The block size is set to one for ciphers such as CTR. However,
|
||||
* you still need to provide incremental updates in multiples of
|
||||
* the underlying block size as the IV does not have sub-block
|
||||
* granularity. This is known in this API as the chunk size.
|
||||
*
|
||||
* Return: chunk size in bytes
|
||||
*/
|
||||
static inline unsigned int crypto_skcipher_chunksize(
|
||||
struct crypto_skcipher *tfm)
|
||||
{
|
||||
return crypto_skcipher_alg_chunksize(crypto_skcipher_alg(tfm));
|
||||
}
|
||||
|
||||
/**
|
||||
* crypto_skcipher_walksize() - obtain walk size
|
||||
* @tfm: cipher handle
|
||||
*
|
||||
* In some cases, algorithms can only perform optimally when operating on
|
||||
* multiple blocks in parallel. This is reflected by the walksize, which
|
||||
* must be a multiple of the chunksize (or equal if the concern does not
|
||||
* apply)
|
||||
*
|
||||
* Return: walk size in bytes
|
||||
*/
|
||||
static inline unsigned int crypto_skcipher_walksize(
|
||||
struct crypto_skcipher *tfm)
|
||||
{
|
||||
return crypto_skcipher_alg_walksize(crypto_skcipher_alg(tfm));
|
||||
}
|
||||
|
||||
/**
|
||||
* crypto_skcipher_blocksize() - obtain block size of cipher
|
||||
* @tfm: cipher handle
|
||||
|
Loading…
Reference in New Issue
Block a user