You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
373 lines
13 KiB
373 lines
13 KiB
From 9e0a0b3a192af20193f074ed2ad9dd85a2e48d00 Mon Sep 17 00:00:00 2001
|
|
From: Christian Lamparter <chunkeey@googlemail.com>
|
|
Date: Fri, 25 Aug 2017 15:47:25 +0200
|
|
Subject: [PATCH 12/25] crypto: crypto4xx - pointer arithmetic overhaul
|
|
|
|
This patch improves the readability of various functions,
|
|
by replacing various void* pointers declarations with
|
|
their respective structs *. This makes it possible to go
|
|
for the eye-friendly array-indexing methods.
|
|
|
|
Signed-off-by: Christian Lamparter <chunkeey@googlemail.com>
|
|
Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
|
|
---
|
|
drivers/crypto/amcc/crypto4xx_alg.c | 26 ++++++++--------
|
|
drivers/crypto/amcc/crypto4xx_core.c | 60 +++++++++++++++---------------------
|
|
drivers/crypto/amcc/crypto4xx_core.h | 41 +++++++++++++-----------
|
|
3 files changed, 59 insertions(+), 68 deletions(-)
|
|
|
|
--- a/drivers/crypto/amcc/crypto4xx_alg.c
|
|
+++ b/drivers/crypto/amcc/crypto4xx_alg.c
|
|
@@ -134,7 +134,7 @@ static int crypto4xx_setkey_aes(struct c
|
|
}
|
|
}
|
|
/* Setup SA */
|
|
- sa = (struct dynamic_sa_ctl *) ctx->sa_in;
|
|
+ sa = ctx->sa_in;
|
|
|
|
set_dynamic_sa_command_0(sa, SA_NOT_SAVE_HASH, SA_NOT_SAVE_IV,
|
|
SA_LOAD_HASH_FROM_SA, SA_LOAD_IV_FROM_STATE,
|
|
@@ -159,7 +159,7 @@ static int crypto4xx_setkey_aes(struct c
|
|
ctx->offset_to_sr_ptr = get_dynamic_sa_offset_state_ptr_field(sa);
|
|
|
|
memcpy(ctx->sa_out, ctx->sa_in, ctx->sa_len * 4);
|
|
- sa = (struct dynamic_sa_ctl *) ctx->sa_out;
|
|
+ sa = ctx->sa_out;
|
|
sa->sa_command_0.bf.dir = DIR_OUTBOUND;
|
|
|
|
return 0;
|
|
@@ -248,8 +248,7 @@ static int crypto4xx_hash_alg_init(struc
|
|
struct crypto_alg *alg = tfm->__crt_alg;
|
|
struct crypto4xx_alg *my_alg = crypto_alg_to_crypto4xx_alg(alg);
|
|
struct crypto4xx_ctx *ctx = crypto_tfm_ctx(tfm);
|
|
- struct dynamic_sa_ctl *sa;
|
|
- struct dynamic_sa_hash160 *sa_in;
|
|
+ struct dynamic_sa_hash160 *sa;
|
|
int rc;
|
|
|
|
ctx->dev = my_alg->dev;
|
|
@@ -273,25 +272,24 @@ static int crypto4xx_hash_alg_init(struc
|
|
|
|
crypto_ahash_set_reqsize(__crypto_ahash_cast(tfm),
|
|
sizeof(struct crypto4xx_ctx));
|
|
- sa = (struct dynamic_sa_ctl *) ctx->sa_in;
|
|
- set_dynamic_sa_command_0(sa, SA_SAVE_HASH, SA_NOT_SAVE_IV,
|
|
+ sa = (struct dynamic_sa_hash160 *)ctx->sa_in;
|
|
+ set_dynamic_sa_command_0(&sa->ctrl, SA_SAVE_HASH, SA_NOT_SAVE_IV,
|
|
SA_NOT_LOAD_HASH, SA_LOAD_IV_FROM_SA,
|
|
SA_NO_HEADER_PROC, ha, SA_CIPHER_ALG_NULL,
|
|
SA_PAD_TYPE_ZERO, SA_OP_GROUP_BASIC,
|
|
SA_OPCODE_HASH, DIR_INBOUND);
|
|
- set_dynamic_sa_command_1(sa, 0, SA_HASH_MODE_HASH,
|
|
+ set_dynamic_sa_command_1(&sa->ctrl, 0, SA_HASH_MODE_HASH,
|
|
CRYPTO_FEEDBACK_MODE_NO_FB, SA_EXTENDED_SN_OFF,
|
|
SA_SEQ_MASK_OFF, SA_MC_ENABLE,
|
|
SA_NOT_COPY_PAD, SA_NOT_COPY_PAYLOAD,
|
|
SA_NOT_COPY_HDR);
|
|
ctx->direction = DIR_INBOUND;
|
|
- sa->sa_contents.w = SA_HASH160_CONTENTS;
|
|
- sa_in = (struct dynamic_sa_hash160 *) ctx->sa_in;
|
|
/* Need to zero hash digest in SA */
|
|
- memset(sa_in->inner_digest, 0, sizeof(sa_in->inner_digest));
|
|
- memset(sa_in->outer_digest, 0, sizeof(sa_in->outer_digest));
|
|
- sa_in->state_ptr = ctx->state_record_dma_addr;
|
|
- ctx->offset_to_sr_ptr = get_dynamic_sa_offset_state_ptr_field(sa);
|
|
+ memset(sa->inner_digest, 0, sizeof(sa->inner_digest));
|
|
+ memset(sa->outer_digest, 0, sizeof(sa->outer_digest));
|
|
+ sa->state_ptr = ctx->state_record_dma_addr;
|
|
+ ctx->offset_to_sr_ptr =
|
|
+ get_dynamic_sa_offset_state_ptr_field(&sa->ctrl);
|
|
|
|
return 0;
|
|
}
|
|
@@ -302,7 +300,7 @@ int crypto4xx_hash_init(struct ahash_req
|
|
int ds;
|
|
struct dynamic_sa_ctl *sa;
|
|
|
|
- sa = (struct dynamic_sa_ctl *) ctx->sa_in;
|
|
+ sa = ctx->sa_in;
|
|
ds = crypto_ahash_digestsize(
|
|
__crypto_ahash_cast(req->base.tfm));
|
|
sa->sa_command_0.bf.digest_len = ds >> 2;
|
|
--- a/drivers/crypto/amcc/crypto4xx_core.c
|
|
+++ b/drivers/crypto/amcc/crypto4xx_core.c
|
|
@@ -211,7 +211,7 @@ static u32 crypto4xx_build_pdr(struct cr
|
|
}
|
|
memset(dev->pdr, 0, sizeof(struct ce_pd) * PPC4XX_NUM_PD);
|
|
dev->shadow_sa_pool = dma_alloc_coherent(dev->core_dev->device,
|
|
- 256 * PPC4XX_NUM_PD,
|
|
+ sizeof(union shadow_sa_buf) * PPC4XX_NUM_PD,
|
|
&dev->shadow_sa_pool_pa,
|
|
GFP_ATOMIC);
|
|
if (!dev->shadow_sa_pool)
|
|
@@ -223,16 +223,14 @@ static u32 crypto4xx_build_pdr(struct cr
|
|
if (!dev->shadow_sr_pool)
|
|
return -ENOMEM;
|
|
for (i = 0; i < PPC4XX_NUM_PD; i++) {
|
|
- pd_uinfo = (struct pd_uinfo *) (dev->pdr_uinfo +
|
|
- sizeof(struct pd_uinfo) * i);
|
|
+ pd_uinfo = &dev->pdr_uinfo[i];
|
|
|
|
/* alloc 256 bytes which is enough for any kind of dynamic sa */
|
|
- pd_uinfo->sa_va = dev->shadow_sa_pool + 256 * i;
|
|
+ pd_uinfo->sa_va = &dev->shadow_sa_pool[i].sa;
|
|
pd_uinfo->sa_pa = dev->shadow_sa_pool_pa + 256 * i;
|
|
|
|
/* alloc state record */
|
|
- pd_uinfo->sr_va = dev->shadow_sr_pool +
|
|
- sizeof(struct sa_state_record) * i;
|
|
+ pd_uinfo->sr_va = &dev->shadow_sr_pool[i];
|
|
pd_uinfo->sr_pa = dev->shadow_sr_pool_pa +
|
|
sizeof(struct sa_state_record) * i;
|
|
}
|
|
@@ -248,8 +246,9 @@ static void crypto4xx_destroy_pdr(struct
|
|
dev->pdr, dev->pdr_pa);
|
|
|
|
if (dev->shadow_sa_pool)
|
|
- dma_free_coherent(dev->core_dev->device, 256 * PPC4XX_NUM_PD,
|
|
- dev->shadow_sa_pool, dev->shadow_sa_pool_pa);
|
|
+ dma_free_coherent(dev->core_dev->device,
|
|
+ sizeof(union shadow_sa_buf) * PPC4XX_NUM_PD,
|
|
+ dev->shadow_sa_pool, dev->shadow_sa_pool_pa);
|
|
|
|
if (dev->shadow_sr_pool)
|
|
dma_free_coherent(dev->core_dev->device,
|
|
@@ -277,11 +276,9 @@ static u32 crypto4xx_get_pd_from_pdr_nol
|
|
|
|
static u32 crypto4xx_put_pd_to_pdr(struct crypto4xx_device *dev, u32 idx)
|
|
{
|
|
- struct pd_uinfo *pd_uinfo;
|
|
+ struct pd_uinfo *pd_uinfo = &dev->pdr_uinfo[idx];
|
|
unsigned long flags;
|
|
|
|
- pd_uinfo = (struct pd_uinfo *)(dev->pdr_uinfo +
|
|
- sizeof(struct pd_uinfo) * idx);
|
|
spin_lock_irqsave(&dev->core_dev->lock, flags);
|
|
if (dev->pdr_tail != PPC4XX_LAST_PD)
|
|
dev->pdr_tail++;
|
|
@@ -298,7 +295,7 @@ static struct ce_pd *crypto4xx_get_pdp(s
|
|
{
|
|
*pd_dma = dev->pdr_pa + sizeof(struct ce_pd) * idx;
|
|
|
|
- return dev->pdr + sizeof(struct ce_pd) * idx;
|
|
+ return &dev->pdr[idx];
|
|
}
|
|
|
|
/**
|
|
@@ -376,7 +373,7 @@ static inline struct ce_gd *crypto4xx_ge
|
|
{
|
|
*gd_dma = dev->gdr_pa + sizeof(struct ce_gd) * idx;
|
|
|
|
- return (struct ce_gd *) (dev->gdr + sizeof(struct ce_gd) * idx);
|
|
+ return &dev->gdr[idx];
|
|
}
|
|
|
|
/**
|
|
@@ -387,7 +384,6 @@ static inline struct ce_gd *crypto4xx_ge
|
|
static u32 crypto4xx_build_sdr(struct crypto4xx_device *dev)
|
|
{
|
|
int i;
|
|
- struct ce_sd *sd_array;
|
|
|
|
/* alloc memory for scatter descriptor ring */
|
|
dev->sdr = dma_alloc_coherent(dev->core_dev->device,
|
|
@@ -407,10 +403,8 @@ static u32 crypto4xx_build_sdr(struct cr
|
|
return -ENOMEM;
|
|
}
|
|
|
|
- sd_array = dev->sdr;
|
|
-
|
|
for (i = 0; i < PPC4XX_NUM_SD; i++) {
|
|
- sd_array[i].ptr = dev->scatter_buffer_pa +
|
|
+ dev->sdr[i].ptr = dev->scatter_buffer_pa +
|
|
PPC4XX_SD_BUFFER_SIZE * i;
|
|
}
|
|
|
|
@@ -480,7 +474,7 @@ static inline struct ce_sd *crypto4xx_ge
|
|
{
|
|
*sd_dma = dev->sdr_pa + sizeof(struct ce_sd) * idx;
|
|
|
|
- return (struct ce_sd *)(dev->sdr + sizeof(struct ce_sd) * idx);
|
|
+ return &dev->sdr[idx];
|
|
}
|
|
|
|
static void crypto4xx_copy_pkt_to_dst(struct crypto4xx_device *dev,
|
|
@@ -529,11 +523,10 @@ static u32 crypto4xx_copy_digest_to_dst(
|
|
struct crypto4xx_ctx *ctx)
|
|
{
|
|
struct dynamic_sa_ctl *sa = (struct dynamic_sa_ctl *) ctx->sa_in;
|
|
- struct sa_state_record *state_record =
|
|
- (struct sa_state_record *) pd_uinfo->sr_va;
|
|
|
|
if (sa->sa_command_0.bf.hash_alg == SA_HASH_ALG_SHA1) {
|
|
- memcpy((void *) pd_uinfo->dest_va, state_record->save_digest,
|
|
+ memcpy((void *) pd_uinfo->dest_va,
|
|
+ pd_uinfo->sr_va->save_digest,
|
|
SA_HASH_ALG_SHA1_DIGEST_SIZE);
|
|
}
|
|
|
|
@@ -607,11 +600,9 @@ static u32 crypto4xx_ahash_done(struct c
|
|
|
|
static u32 crypto4xx_pd_done(struct crypto4xx_device *dev, u32 idx)
|
|
{
|
|
- struct ce_pd *pd;
|
|
- struct pd_uinfo *pd_uinfo;
|
|
+ struct ce_pd *pd = &dev->pdr[idx];
|
|
+ struct pd_uinfo *pd_uinfo = &dev->pdr_uinfo[idx];
|
|
|
|
- pd = dev->pdr + sizeof(struct ce_pd)*idx;
|
|
- pd_uinfo = dev->pdr_uinfo + sizeof(struct pd_uinfo)*idx;
|
|
if (crypto_tfm_alg_type(pd_uinfo->async_req->tfm) ==
|
|
CRYPTO_ALG_TYPE_ABLKCIPHER)
|
|
return crypto4xx_ablkcipher_done(dev, pd_uinfo, pd);
|
|
@@ -712,7 +703,6 @@ u32 crypto4xx_build_pd(struct crypto_asy
|
|
unsigned long flags;
|
|
struct pd_uinfo *pd_uinfo = NULL;
|
|
unsigned int nbytes = datalen, idx;
|
|
- unsigned int ivlen = 0;
|
|
u32 gd_idx = 0;
|
|
|
|
/* figure how many gd is needed */
|
|
@@ -771,17 +761,15 @@ u32 crypto4xx_build_pd(struct crypto_asy
|
|
}
|
|
spin_unlock_irqrestore(&dev->core_dev->lock, flags);
|
|
|
|
- pd_uinfo = (struct pd_uinfo *)(dev->pdr_uinfo +
|
|
- sizeof(struct pd_uinfo) * pd_entry);
|
|
+ pd_uinfo = &dev->pdr_uinfo[pd_entry];
|
|
pd = crypto4xx_get_pdp(dev, &pd_dma, pd_entry);
|
|
pd_uinfo->async_req = req;
|
|
pd_uinfo->num_gd = num_gd;
|
|
pd_uinfo->num_sd = num_sd;
|
|
|
|
if (iv_len || ctx->is_hash) {
|
|
- ivlen = iv_len;
|
|
pd->sa = pd_uinfo->sa_pa;
|
|
- sa = (struct dynamic_sa_ctl *) pd_uinfo->sa_va;
|
|
+ sa = pd_uinfo->sa_va;
|
|
if (ctx->direction == DIR_INBOUND)
|
|
memcpy(sa, ctx->sa_in, ctx->sa_len * 4);
|
|
else
|
|
@@ -791,14 +779,15 @@ u32 crypto4xx_build_pd(struct crypto_asy
|
|
&pd_uinfo->sr_pa, 4);
|
|
|
|
if (iv_len)
|
|
- crypto4xx_memcpy_le(pd_uinfo->sr_va, iv, iv_len);
|
|
+ crypto4xx_memcpy_le(pd_uinfo->sr_va->save_iv,
|
|
+ iv, iv_len);
|
|
} else {
|
|
if (ctx->direction == DIR_INBOUND) {
|
|
pd->sa = ctx->sa_in_dma_addr;
|
|
- sa = (struct dynamic_sa_ctl *) ctx->sa_in;
|
|
+ sa = ctx->sa_in;
|
|
} else {
|
|
pd->sa = ctx->sa_out_dma_addr;
|
|
- sa = (struct dynamic_sa_ctl *) ctx->sa_out;
|
|
+ sa = ctx->sa_out;
|
|
}
|
|
}
|
|
pd->sa_len = ctx->sa_len;
|
|
@@ -1006,9 +995,8 @@ static void crypto4xx_bh_tasklet_cb(unsi
|
|
|
|
while (core_dev->dev->pdr_head != core_dev->dev->pdr_tail) {
|
|
tail = core_dev->dev->pdr_tail;
|
|
- pd_uinfo = core_dev->dev->pdr_uinfo +
|
|
- sizeof(struct pd_uinfo)*tail;
|
|
- pd = core_dev->dev->pdr + sizeof(struct ce_pd) * tail;
|
|
+ pd_uinfo = &core_dev->dev->pdr_uinfo[tail];
|
|
+ pd = &core_dev->dev->pdr[tail];
|
|
if ((pd_uinfo->state == PD_ENTRY_INUSE) &&
|
|
pd->pd_ctl.bf.pe_done &&
|
|
!pd->pd_ctl.bf.host_ready) {
|
|
--- a/drivers/crypto/amcc/crypto4xx_core.h
|
|
+++ b/drivers/crypto/amcc/crypto4xx_core.h
|
|
@@ -23,6 +23,8 @@
|
|
#define __CRYPTO4XX_CORE_H__
|
|
|
|
#include <crypto/internal/hash.h>
|
|
+#include "crypto4xx_reg_def.h"
|
|
+#include "crypto4xx_sa.h"
|
|
|
|
#define MODULE_NAME "crypto4xx"
|
|
|
|
@@ -48,6 +50,13 @@
|
|
|
|
struct crypto4xx_device;
|
|
|
|
+union shadow_sa_buf {
|
|
+ struct dynamic_sa_ctl sa;
|
|
+
|
|
+ /* alloc 256 bytes which is enough for any kind of dynamic sa */
|
|
+ u8 buf[256];
|
|
+} __packed;
|
|
+
|
|
struct pd_uinfo {
|
|
struct crypto4xx_device *dev;
|
|
u32 state;
|
|
@@ -60,9 +69,9 @@ struct pd_uinfo {
|
|
used by this packet */
|
|
u32 num_sd; /* number of scatter discriptors
|
|
used by this packet */
|
|
- void *sa_va; /* shadow sa, when using cp from ctx->sa */
|
|
+ struct dynamic_sa_ctl *sa_va; /* shadow sa */
|
|
u32 sa_pa;
|
|
- void *sr_va; /* state record for shadow sa */
|
|
+ struct sa_state_record *sr_va; /* state record for shadow sa */
|
|
u32 sr_pa;
|
|
struct scatterlist *dest_va;
|
|
struct crypto_async_request *async_req; /* base crypto request
|
|
@@ -75,22 +84,18 @@ struct crypto4xx_device {
|
|
void __iomem *ce_base;
|
|
void __iomem *trng_base;
|
|
|
|
- void *pdr; /* base address of packet
|
|
- descriptor ring */
|
|
- dma_addr_t pdr_pa; /* physical address used to
|
|
- program ce pdr_base_register */
|
|
- void *gdr; /* gather descriptor ring */
|
|
- dma_addr_t gdr_pa; /* physical address used to
|
|
- program ce gdr_base_register */
|
|
- void *sdr; /* scatter descriptor ring */
|
|
- dma_addr_t sdr_pa; /* physical address used to
|
|
- program ce sdr_base_register */
|
|
+ struct ce_pd *pdr; /* base address of packet descriptor ring */
|
|
+ dma_addr_t pdr_pa; /* physical address of pdr_base_register */
|
|
+ struct ce_gd *gdr; /* gather descriptor ring */
|
|
+ dma_addr_t gdr_pa; /* physical address of gdr_base_register */
|
|
+ struct ce_sd *sdr; /* scatter descriptor ring */
|
|
+ dma_addr_t sdr_pa; /* physical address of sdr_base_register */
|
|
void *scatter_buffer_va;
|
|
dma_addr_t scatter_buffer_pa;
|
|
|
|
- void *shadow_sa_pool; /* pool of memory for sa in pd_uinfo */
|
|
+ union shadow_sa_buf *shadow_sa_pool;
|
|
dma_addr_t shadow_sa_pool_pa;
|
|
- void *shadow_sr_pool; /* pool of memory for sr in pd_uinfo */
|
|
+ struct sa_state_record *shadow_sr_pool;
|
|
dma_addr_t shadow_sr_pool_pa;
|
|
u32 pdr_tail;
|
|
u32 pdr_head;
|
|
@@ -98,7 +103,7 @@ struct crypto4xx_device {
|
|
u32 gdr_head;
|
|
u32 sdr_tail;
|
|
u32 sdr_head;
|
|
- void *pdr_uinfo;
|
|
+ struct pd_uinfo *pdr_uinfo;
|
|
struct list_head alg_list; /* List of algorithm supported
|
|
by this device */
|
|
};
|
|
@@ -116,11 +121,11 @@ struct crypto4xx_core_device {
|
|
|
|
struct crypto4xx_ctx {
|
|
struct crypto4xx_device *dev;
|
|
- void *sa_in;
|
|
+ struct dynamic_sa_ctl *sa_in;
|
|
dma_addr_t sa_in_dma_addr;
|
|
- void *sa_out;
|
|
+ struct dynamic_sa_ctl *sa_out;
|
|
dma_addr_t sa_out_dma_addr;
|
|
- void *state_record;
|
|
+ struct sa_state_record *state_record;
|
|
dma_addr_t state_record_dma_addr;
|
|
u32 sa_len;
|
|
u32 offset_to_sr_ptr; /* offset to state ptr, in dynamic sa */
|
|
|