crypto: nx - convert AES-CTR to skcipher API
Convert the PowerPC Nest (NX) implementation of AES-CTR from the deprecated "blkcipher" API to the "skcipher" API. This is needed in order for the blkcipher API to be removed. Signed-off-by: Eric Biggers <ebiggers@google.com> Reviewed-by: Ard Biesheuvel <ard.biesheuvel@linaro.org> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
This commit is contained in:
parent
c1e9c386c9
commit
4ba802980c
|
@ -19,11 +19,11 @@
|
||||||
#include "nx.h"
|
#include "nx.h"
|
||||||
|
|
||||||
|
|
||||||
static int ctr_aes_nx_set_key(struct crypto_tfm *tfm,
|
static int ctr_aes_nx_set_key(struct crypto_skcipher *tfm,
|
||||||
const u8 *in_key,
|
const u8 *in_key,
|
||||||
unsigned int key_len)
|
unsigned int key_len)
|
||||||
{
|
{
|
||||||
struct nx_crypto_ctx *nx_ctx = crypto_tfm_ctx(tfm);
|
struct nx_crypto_ctx *nx_ctx = crypto_skcipher_ctx(tfm);
|
||||||
struct nx_csbcpb *csbcpb = nx_ctx->csbcpb;
|
struct nx_csbcpb *csbcpb = nx_ctx->csbcpb;
|
||||||
|
|
||||||
nx_ctx_init(nx_ctx, HCOP_FC_AES);
|
nx_ctx_init(nx_ctx, HCOP_FC_AES);
|
||||||
|
@ -51,11 +51,11 @@ static int ctr_aes_nx_set_key(struct crypto_tfm *tfm,
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int ctr3686_aes_nx_set_key(struct crypto_tfm *tfm,
|
static int ctr3686_aes_nx_set_key(struct crypto_skcipher *tfm,
|
||||||
const u8 *in_key,
|
const u8 *in_key,
|
||||||
unsigned int key_len)
|
unsigned int key_len)
|
||||||
{
|
{
|
||||||
struct nx_crypto_ctx *nx_ctx = crypto_tfm_ctx(tfm);
|
struct nx_crypto_ctx *nx_ctx = crypto_skcipher_ctx(tfm);
|
||||||
|
|
||||||
if (key_len < CTR_RFC3686_NONCE_SIZE)
|
if (key_len < CTR_RFC3686_NONCE_SIZE)
|
||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
|
@ -69,12 +69,10 @@ static int ctr3686_aes_nx_set_key(struct crypto_tfm *tfm,
|
||||||
return ctr_aes_nx_set_key(tfm, in_key, key_len);
|
return ctr_aes_nx_set_key(tfm, in_key, key_len);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int ctr_aes_nx_crypt(struct blkcipher_desc *desc,
|
static int ctr_aes_nx_crypt(struct skcipher_request *req, u8 *iv)
|
||||||
struct scatterlist *dst,
|
|
||||||
struct scatterlist *src,
|
|
||||||
unsigned int nbytes)
|
|
||||||
{
|
{
|
||||||
struct nx_crypto_ctx *nx_ctx = crypto_blkcipher_ctx(desc->tfm);
|
struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
|
||||||
|
struct nx_crypto_ctx *nx_ctx = crypto_skcipher_ctx(tfm);
|
||||||
struct nx_csbcpb *csbcpb = nx_ctx->csbcpb;
|
struct nx_csbcpb *csbcpb = nx_ctx->csbcpb;
|
||||||
unsigned long irq_flags;
|
unsigned long irq_flags;
|
||||||
unsigned int processed = 0, to_process;
|
unsigned int processed = 0, to_process;
|
||||||
|
@ -83,9 +81,9 @@ static int ctr_aes_nx_crypt(struct blkcipher_desc *desc,
|
||||||
spin_lock_irqsave(&nx_ctx->lock, irq_flags);
|
spin_lock_irqsave(&nx_ctx->lock, irq_flags);
|
||||||
|
|
||||||
do {
|
do {
|
||||||
to_process = nbytes - processed;
|
to_process = req->cryptlen - processed;
|
||||||
|
|
||||||
rc = nx_build_sg_lists(nx_ctx, desc->info, dst, src,
|
rc = nx_build_sg_lists(nx_ctx, iv, req->dst, req->src,
|
||||||
&to_process, processed,
|
&to_process, processed,
|
||||||
csbcpb->cpb.aes_ctr.iv);
|
csbcpb->cpb.aes_ctr.iv);
|
||||||
if (rc)
|
if (rc)
|
||||||
|
@ -97,59 +95,51 @@ static int ctr_aes_nx_crypt(struct blkcipher_desc *desc,
|
||||||
}
|
}
|
||||||
|
|
||||||
rc = nx_hcall_sync(nx_ctx, &nx_ctx->op,
|
rc = nx_hcall_sync(nx_ctx, &nx_ctx->op,
|
||||||
desc->flags & CRYPTO_TFM_REQ_MAY_SLEEP);
|
req->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP);
|
||||||
if (rc)
|
if (rc)
|
||||||
goto out;
|
goto out;
|
||||||
|
|
||||||
memcpy(desc->info, csbcpb->cpb.aes_cbc.cv, AES_BLOCK_SIZE);
|
memcpy(iv, csbcpb->cpb.aes_cbc.cv, AES_BLOCK_SIZE);
|
||||||
|
|
||||||
atomic_inc(&(nx_ctx->stats->aes_ops));
|
atomic_inc(&(nx_ctx->stats->aes_ops));
|
||||||
atomic64_add(csbcpb->csb.processed_byte_count,
|
atomic64_add(csbcpb->csb.processed_byte_count,
|
||||||
&(nx_ctx->stats->aes_bytes));
|
&(nx_ctx->stats->aes_bytes));
|
||||||
|
|
||||||
processed += to_process;
|
processed += to_process;
|
||||||
} while (processed < nbytes);
|
} while (processed < req->cryptlen);
|
||||||
out:
|
out:
|
||||||
spin_unlock_irqrestore(&nx_ctx->lock, irq_flags);
|
spin_unlock_irqrestore(&nx_ctx->lock, irq_flags);
|
||||||
return rc;
|
return rc;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int ctr3686_aes_nx_crypt(struct blkcipher_desc *desc,
|
static int ctr3686_aes_nx_crypt(struct skcipher_request *req)
|
||||||
struct scatterlist *dst,
|
|
||||||
struct scatterlist *src,
|
|
||||||
unsigned int nbytes)
|
|
||||||
{
|
{
|
||||||
struct nx_crypto_ctx *nx_ctx = crypto_blkcipher_ctx(desc->tfm);
|
struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
|
||||||
|
struct nx_crypto_ctx *nx_ctx = crypto_skcipher_ctx(tfm);
|
||||||
u8 iv[16];
|
u8 iv[16];
|
||||||
|
|
||||||
memcpy(iv, nx_ctx->priv.ctr.nonce, CTR_RFC3686_IV_SIZE);
|
memcpy(iv, nx_ctx->priv.ctr.nonce, CTR_RFC3686_IV_SIZE);
|
||||||
memcpy(iv + CTR_RFC3686_NONCE_SIZE,
|
memcpy(iv + CTR_RFC3686_NONCE_SIZE, req->iv, CTR_RFC3686_IV_SIZE);
|
||||||
desc->info, CTR_RFC3686_IV_SIZE);
|
|
||||||
iv[12] = iv[13] = iv[14] = 0;
|
iv[12] = iv[13] = iv[14] = 0;
|
||||||
iv[15] = 1;
|
iv[15] = 1;
|
||||||
|
|
||||||
desc->info = iv;
|
return ctr_aes_nx_crypt(req, iv);
|
||||||
|
|
||||||
return ctr_aes_nx_crypt(desc, dst, src, nbytes);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
struct crypto_alg nx_ctr3686_aes_alg = {
|
struct skcipher_alg nx_ctr3686_aes_alg = {
|
||||||
.cra_name = "rfc3686(ctr(aes))",
|
.base.cra_name = "rfc3686(ctr(aes))",
|
||||||
.cra_driver_name = "rfc3686-ctr-aes-nx",
|
.base.cra_driver_name = "rfc3686-ctr-aes-nx",
|
||||||
.cra_priority = 300,
|
.base.cra_priority = 300,
|
||||||
.cra_flags = CRYPTO_ALG_TYPE_BLKCIPHER,
|
.base.cra_blocksize = 1,
|
||||||
.cra_blocksize = 1,
|
.base.cra_ctxsize = sizeof(struct nx_crypto_ctx),
|
||||||
.cra_ctxsize = sizeof(struct nx_crypto_ctx),
|
.base.cra_module = THIS_MODULE,
|
||||||
.cra_type = &crypto_blkcipher_type,
|
.init = nx_crypto_ctx_aes_ctr_init,
|
||||||
.cra_module = THIS_MODULE,
|
.exit = nx_crypto_ctx_skcipher_exit,
|
||||||
.cra_init = nx_crypto_ctx_aes_ctr_init,
|
|
||||||
.cra_exit = nx_crypto_ctx_exit,
|
|
||||||
.cra_blkcipher = {
|
|
||||||
.min_keysize = AES_MIN_KEY_SIZE + CTR_RFC3686_NONCE_SIZE,
|
.min_keysize = AES_MIN_KEY_SIZE + CTR_RFC3686_NONCE_SIZE,
|
||||||
.max_keysize = AES_MAX_KEY_SIZE + CTR_RFC3686_NONCE_SIZE,
|
.max_keysize = AES_MAX_KEY_SIZE + CTR_RFC3686_NONCE_SIZE,
|
||||||
.ivsize = CTR_RFC3686_IV_SIZE,
|
.ivsize = CTR_RFC3686_IV_SIZE,
|
||||||
.setkey = ctr3686_aes_nx_set_key,
|
.setkey = ctr3686_aes_nx_set_key,
|
||||||
.encrypt = ctr3686_aes_nx_crypt,
|
.encrypt = ctr3686_aes_nx_crypt,
|
||||||
.decrypt = ctr3686_aes_nx_crypt,
|
.decrypt = ctr3686_aes_nx_crypt,
|
||||||
}
|
.chunksize = AES_BLOCK_SIZE,
|
||||||
};
|
};
|
||||||
|
|
|
@ -511,12 +511,6 @@ static bool nx_check_props(struct device *dev, u32 fc, u32 mode)
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int nx_register_alg(struct crypto_alg *alg, u32 fc, u32 mode)
|
|
||||||
{
|
|
||||||
return nx_check_props(&nx_driver.viodev->dev, fc, mode) ?
|
|
||||||
crypto_register_alg(alg) : 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
static int nx_register_skcipher(struct skcipher_alg *alg, u32 fc, u32 mode)
|
static int nx_register_skcipher(struct skcipher_alg *alg, u32 fc, u32 mode)
|
||||||
{
|
{
|
||||||
return nx_check_props(&nx_driver.viodev->dev, fc, mode) ?
|
return nx_check_props(&nx_driver.viodev->dev, fc, mode) ?
|
||||||
|
@ -537,12 +531,6 @@ static int nx_register_shash(struct shash_alg *alg, u32 fc, u32 mode, int slot)
|
||||||
crypto_register_shash(alg) : 0;
|
crypto_register_shash(alg) : 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void nx_unregister_alg(struct crypto_alg *alg, u32 fc, u32 mode)
|
|
||||||
{
|
|
||||||
if (nx_check_props(NULL, fc, mode))
|
|
||||||
crypto_unregister_alg(alg);
|
|
||||||
}
|
|
||||||
|
|
||||||
static void nx_unregister_skcipher(struct skcipher_alg *alg, u32 fc, u32 mode)
|
static void nx_unregister_skcipher(struct skcipher_alg *alg, u32 fc, u32 mode)
|
||||||
{
|
{
|
||||||
if (nx_check_props(NULL, fc, mode))
|
if (nx_check_props(NULL, fc, mode))
|
||||||
|
@ -593,7 +581,8 @@ static int nx_register_algs(void)
|
||||||
if (rc)
|
if (rc)
|
||||||
goto out_unreg_ecb;
|
goto out_unreg_ecb;
|
||||||
|
|
||||||
rc = nx_register_alg(&nx_ctr3686_aes_alg, NX_FC_AES, NX_MODE_AES_CTR);
|
rc = nx_register_skcipher(&nx_ctr3686_aes_alg, NX_FC_AES,
|
||||||
|
NX_MODE_AES_CTR);
|
||||||
if (rc)
|
if (rc)
|
||||||
goto out_unreg_cbc;
|
goto out_unreg_cbc;
|
||||||
|
|
||||||
|
@ -645,7 +634,7 @@ static int nx_register_algs(void)
|
||||||
out_unreg_gcm:
|
out_unreg_gcm:
|
||||||
nx_unregister_aead(&nx_gcm_aes_alg, NX_FC_AES, NX_MODE_AES_GCM);
|
nx_unregister_aead(&nx_gcm_aes_alg, NX_FC_AES, NX_MODE_AES_GCM);
|
||||||
out_unreg_ctr3686:
|
out_unreg_ctr3686:
|
||||||
nx_unregister_alg(&nx_ctr3686_aes_alg, NX_FC_AES, NX_MODE_AES_CTR);
|
nx_unregister_skcipher(&nx_ctr3686_aes_alg, NX_FC_AES, NX_MODE_AES_CTR);
|
||||||
out_unreg_cbc:
|
out_unreg_cbc:
|
||||||
nx_unregister_skcipher(&nx_cbc_aes_alg, NX_FC_AES, NX_MODE_AES_CBC);
|
nx_unregister_skcipher(&nx_cbc_aes_alg, NX_FC_AES, NX_MODE_AES_CBC);
|
||||||
out_unreg_ecb:
|
out_unreg_ecb:
|
||||||
|
@ -716,9 +705,9 @@ int nx_crypto_ctx_aes_gcm_init(struct crypto_aead *tfm)
|
||||||
NX_MODE_AES_GCM);
|
NX_MODE_AES_GCM);
|
||||||
}
|
}
|
||||||
|
|
||||||
int nx_crypto_ctx_aes_ctr_init(struct crypto_tfm *tfm)
|
int nx_crypto_ctx_aes_ctr_init(struct crypto_skcipher *tfm)
|
||||||
{
|
{
|
||||||
return nx_crypto_ctx_init(crypto_tfm_ctx(tfm), NX_FC_AES,
|
return nx_crypto_ctx_init(crypto_skcipher_ctx(tfm), NX_FC_AES,
|
||||||
NX_MODE_AES_CTR);
|
NX_MODE_AES_CTR);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -815,7 +804,7 @@ static int nx_remove(struct vio_dev *viodev)
|
||||||
NX_FC_AES, NX_MODE_AES_GCM);
|
NX_FC_AES, NX_MODE_AES_GCM);
|
||||||
nx_unregister_aead(&nx_gcm_aes_alg,
|
nx_unregister_aead(&nx_gcm_aes_alg,
|
||||||
NX_FC_AES, NX_MODE_AES_GCM);
|
NX_FC_AES, NX_MODE_AES_GCM);
|
||||||
nx_unregister_alg(&nx_ctr3686_aes_alg,
|
nx_unregister_skcipher(&nx_ctr3686_aes_alg,
|
||||||
NX_FC_AES, NX_MODE_AES_CTR);
|
NX_FC_AES, NX_MODE_AES_CTR);
|
||||||
nx_unregister_skcipher(&nx_cbc_aes_alg, NX_FC_AES,
|
nx_unregister_skcipher(&nx_cbc_aes_alg, NX_FC_AES,
|
||||||
NX_MODE_AES_CBC);
|
NX_MODE_AES_CBC);
|
||||||
|
|
|
@ -145,7 +145,7 @@ struct crypto_aead;
|
||||||
int nx_crypto_ctx_aes_ccm_init(struct crypto_aead *tfm);
|
int nx_crypto_ctx_aes_ccm_init(struct crypto_aead *tfm);
|
||||||
int nx_crypto_ctx_aes_gcm_init(struct crypto_aead *tfm);
|
int nx_crypto_ctx_aes_gcm_init(struct crypto_aead *tfm);
|
||||||
int nx_crypto_ctx_aes_xcbc_init(struct crypto_tfm *tfm);
|
int nx_crypto_ctx_aes_xcbc_init(struct crypto_tfm *tfm);
|
||||||
int nx_crypto_ctx_aes_ctr_init(struct crypto_tfm *tfm);
|
int nx_crypto_ctx_aes_ctr_init(struct crypto_skcipher *tfm);
|
||||||
int nx_crypto_ctx_aes_cbc_init(struct crypto_skcipher *tfm);
|
int nx_crypto_ctx_aes_cbc_init(struct crypto_skcipher *tfm);
|
||||||
int nx_crypto_ctx_aes_ecb_init(struct crypto_skcipher *tfm);
|
int nx_crypto_ctx_aes_ecb_init(struct crypto_skcipher *tfm);
|
||||||
int nx_crypto_ctx_sha_init(struct crypto_tfm *tfm);
|
int nx_crypto_ctx_sha_init(struct crypto_tfm *tfm);
|
||||||
|
@ -180,7 +180,7 @@ extern struct skcipher_alg nx_cbc_aes_alg;
|
||||||
extern struct skcipher_alg nx_ecb_aes_alg;
|
extern struct skcipher_alg nx_ecb_aes_alg;
|
||||||
extern struct aead_alg nx_gcm_aes_alg;
|
extern struct aead_alg nx_gcm_aes_alg;
|
||||||
extern struct aead_alg nx_gcm4106_aes_alg;
|
extern struct aead_alg nx_gcm4106_aes_alg;
|
||||||
extern struct crypto_alg nx_ctr3686_aes_alg;
|
extern struct skcipher_alg nx_ctr3686_aes_alg;
|
||||||
extern struct aead_alg nx_ccm_aes_alg;
|
extern struct aead_alg nx_ccm_aes_alg;
|
||||||
extern struct aead_alg nx_ccm4309_aes_alg;
|
extern struct aead_alg nx_ccm4309_aes_alg;
|
||||||
extern struct shash_alg nx_shash_aes_xcbc_alg;
|
extern struct shash_alg nx_shash_aes_xcbc_alg;
|
||||||
|
|
Loading…
Reference in New Issue