crypto: talitos - don't persistently map req_ctx->hw_context and req_ctx->buf
Commit49f9783b0c
("crypto: talitos - do hw_context DMA mapping outside the requests") introduced a persistent dma mapping of req_ctx->hw_context Commit37b5e8897e
("crypto: talitos - chain in buffered data for ahash on SEC1") introduced a persistent dma mapping of req_ctx->buf As there is no destructor for req_ctx (the request context), the associated dma handlers where set in ctx (the tfm context). This is wrong as several hash operations can run with the same ctx. This patch removes this persistent mapping. Reported-by: Horia Geanta <horia.geanta@nxp.com> Cc: <stable@vger.kernel.org> Fixes:49f9783b0c
("crypto: talitos - do hw_context DMA mapping outside the requests") Fixes:37b5e8897e
("crypto: talitos - chain in buffered data for ahash on SEC1") Signed-off-by: Christophe Leroy <christophe.leroy@c-s.fr> Tested-by: Horia Geantă <horia.geanta@nxp.com> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
This commit is contained in:
parent
cf03868711
commit
ad4cd51fb8
|
@ -832,8 +832,6 @@ struct talitos_ctx {
|
|||
unsigned int keylen;
|
||||
unsigned int enckeylen;
|
||||
unsigned int authkeylen;
|
||||
dma_addr_t dma_buf;
|
||||
dma_addr_t dma_hw_context;
|
||||
};
|
||||
|
||||
#define HASH_MAX_BLOCK_SIZE SHA512_BLOCK_SIZE
|
||||
|
@ -1690,9 +1688,30 @@ static void common_nonsnoop_hash_unmap(struct device *dev,
|
|||
struct ahash_request *areq)
|
||||
{
|
||||
struct talitos_ahash_req_ctx *req_ctx = ahash_request_ctx(areq);
|
||||
struct talitos_private *priv = dev_get_drvdata(dev);
|
||||
bool is_sec1 = has_ftr_sec1(priv);
|
||||
struct talitos_desc *desc = &edesc->desc;
|
||||
struct talitos_desc *desc2 = desc + 1;
|
||||
|
||||
unmap_single_talitos_ptr(dev, &edesc->desc.ptr[5], DMA_FROM_DEVICE);
|
||||
if (desc->next_desc &&
|
||||
desc->ptr[5].ptr != desc2->ptr[5].ptr)
|
||||
unmap_single_talitos_ptr(dev, &desc2->ptr[5], DMA_FROM_DEVICE);
|
||||
|
||||
talitos_sg_unmap(dev, edesc, req_ctx->psrc, NULL, 0, 0);
|
||||
|
||||
/* When using hashctx-in, must unmap it. */
|
||||
if (from_talitos_ptr_len(&edesc->desc.ptr[1], is_sec1))
|
||||
unmap_single_talitos_ptr(dev, &edesc->desc.ptr[1],
|
||||
DMA_TO_DEVICE);
|
||||
else if (desc->next_desc)
|
||||
unmap_single_talitos_ptr(dev, &desc2->ptr[1],
|
||||
DMA_TO_DEVICE);
|
||||
|
||||
if (is_sec1 && req_ctx->nbuf)
|
||||
unmap_single_talitos_ptr(dev, &desc->ptr[3],
|
||||
DMA_TO_DEVICE);
|
||||
|
||||
if (edesc->dma_len)
|
||||
dma_unmap_single(dev, edesc->dma_link_tbl, edesc->dma_len,
|
||||
DMA_BIDIRECTIONAL);
|
||||
|
@ -1766,8 +1785,10 @@ static int common_nonsnoop_hash(struct talitos_edesc *edesc,
|
|||
|
||||
/* hash context in */
|
||||
if (!req_ctx->first || req_ctx->swinit) {
|
||||
to_talitos_ptr(&desc->ptr[1], ctx->dma_hw_context,
|
||||
req_ctx->hw_context_size, is_sec1);
|
||||
map_single_talitos_ptr(dev, &desc->ptr[1],
|
||||
req_ctx->hw_context_size,
|
||||
(char *)req_ctx->hw_context,
|
||||
DMA_TO_DEVICE);
|
||||
req_ctx->swinit = 0;
|
||||
}
|
||||
/* Indicate next op is not the first. */
|
||||
|
@ -1793,10 +1814,9 @@ static int common_nonsnoop_hash(struct talitos_edesc *edesc,
|
|||
* data in
|
||||
*/
|
||||
if (is_sec1 && req_ctx->nbuf) {
|
||||
dma_addr_t dma_buf = ctx->dma_buf + req_ctx->buf_idx *
|
||||
HASH_MAX_BLOCK_SIZE;
|
||||
|
||||
to_talitos_ptr(&desc->ptr[3], dma_buf, req_ctx->nbuf, is_sec1);
|
||||
map_single_talitos_ptr(dev, &desc->ptr[3], req_ctx->nbuf,
|
||||
req_ctx->buf[req_ctx->buf_idx],
|
||||
DMA_TO_DEVICE);
|
||||
} else {
|
||||
sg_count = talitos_sg_map(dev, req_ctx->psrc, length, edesc,
|
||||
&desc->ptr[3], sg_count, offset, 0);
|
||||
|
@ -1812,8 +1832,9 @@ static int common_nonsnoop_hash(struct talitos_edesc *edesc,
|
|||
crypto_ahash_digestsize(tfm),
|
||||
areq->result, DMA_FROM_DEVICE);
|
||||
else
|
||||
to_talitos_ptr(&desc->ptr[5], ctx->dma_hw_context,
|
||||
req_ctx->hw_context_size, is_sec1);
|
||||
map_single_talitos_ptr(dev, &desc->ptr[5],
|
||||
req_ctx->hw_context_size,
|
||||
req_ctx->hw_context, DMA_FROM_DEVICE);
|
||||
|
||||
/* last DWORD empty */
|
||||
|
||||
|
@ -1832,9 +1853,14 @@ static int common_nonsnoop_hash(struct talitos_edesc *edesc,
|
|||
desc->hdr |= DESC_HDR_MODE0_MDEU_CONT;
|
||||
desc->hdr &= ~DESC_HDR_DONE_NOTIFY;
|
||||
|
||||
to_talitos_ptr(&desc2->ptr[1], ctx->dma_hw_context,
|
||||
req_ctx->hw_context_size, is_sec1);
|
||||
|
||||
if (desc->ptr[1].ptr)
|
||||
copy_talitos_ptr(&desc2->ptr[1], &desc->ptr[1],
|
||||
is_sec1);
|
||||
else
|
||||
map_single_talitos_ptr(dev, &desc2->ptr[1],
|
||||
req_ctx->hw_context_size,
|
||||
req_ctx->hw_context,
|
||||
DMA_TO_DEVICE);
|
||||
copy_talitos_ptr(&desc2->ptr[2], &desc->ptr[2], is_sec1);
|
||||
sg_count = talitos_sg_map(dev, req_ctx->psrc, length, edesc,
|
||||
&desc2->ptr[3], sg_count, offset, 0);
|
||||
|
@ -1842,8 +1868,10 @@ static int common_nonsnoop_hash(struct talitos_edesc *edesc,
|
|||
sync_needed = true;
|
||||
copy_talitos_ptr(&desc2->ptr[5], &desc->ptr[5], is_sec1);
|
||||
if (req_ctx->last)
|
||||
to_talitos_ptr(&desc->ptr[5], ctx->dma_hw_context,
|
||||
req_ctx->hw_context_size, is_sec1);
|
||||
map_single_talitos_ptr(dev, &desc->ptr[5],
|
||||
req_ctx->hw_context_size,
|
||||
req_ctx->hw_context,
|
||||
DMA_FROM_DEVICE);
|
||||
|
||||
next_desc = dma_map_single(dev, &desc2->hdr1, TALITOS_DESC_SIZE,
|
||||
DMA_BIDIRECTIONAL);
|
||||
|
@ -1881,12 +1909,8 @@ static struct talitos_edesc *ahash_edesc_alloc(struct ahash_request *areq,
|
|||
static int ahash_init(struct ahash_request *areq)
|
||||
{
|
||||
struct crypto_ahash *tfm = crypto_ahash_reqtfm(areq);
|
||||
struct talitos_ctx *ctx = crypto_ahash_ctx(tfm);
|
||||
struct device *dev = ctx->dev;
|
||||
struct talitos_ahash_req_ctx *req_ctx = ahash_request_ctx(areq);
|
||||
unsigned int size;
|
||||
struct talitos_private *priv = dev_get_drvdata(dev);
|
||||
bool is_sec1 = has_ftr_sec1(priv);
|
||||
|
||||
/* Initialize the context */
|
||||
req_ctx->buf_idx = 0;
|
||||
|
@ -1898,18 +1922,6 @@ static int ahash_init(struct ahash_request *areq)
|
|||
: TALITOS_MDEU_CONTEXT_SIZE_SHA384_SHA512;
|
||||
req_ctx->hw_context_size = size;
|
||||
|
||||
if (ctx->dma_hw_context)
|
||||
dma_unmap_single(dev, ctx->dma_hw_context, size,
|
||||
DMA_BIDIRECTIONAL);
|
||||
ctx->dma_hw_context = dma_map_single(dev, req_ctx->hw_context, size,
|
||||
DMA_BIDIRECTIONAL);
|
||||
if (ctx->dma_buf)
|
||||
dma_unmap_single(dev, ctx->dma_buf, sizeof(req_ctx->buf),
|
||||
DMA_TO_DEVICE);
|
||||
if (is_sec1)
|
||||
ctx->dma_buf = dma_map_single(dev, req_ctx->buf,
|
||||
sizeof(req_ctx->buf),
|
||||
DMA_TO_DEVICE);
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
@ -1920,9 +1932,6 @@ static int ahash_init(struct ahash_request *areq)
|
|||
static int ahash_init_sha224_swinit(struct ahash_request *areq)
|
||||
{
|
||||
struct talitos_ahash_req_ctx *req_ctx = ahash_request_ctx(areq);
|
||||
struct crypto_ahash *tfm = crypto_ahash_reqtfm(areq);
|
||||
struct talitos_ctx *ctx = crypto_ahash_ctx(tfm);
|
||||
struct device *dev = ctx->dev;
|
||||
|
||||
ahash_init(areq);
|
||||
req_ctx->swinit = 1;/* prevent h/w initting context with sha256 values*/
|
||||
|
@ -1940,9 +1949,6 @@ static int ahash_init_sha224_swinit(struct ahash_request *areq)
|
|||
req_ctx->hw_context[8] = 0;
|
||||
req_ctx->hw_context[9] = 0;
|
||||
|
||||
dma_sync_single_for_device(dev, ctx->dma_hw_context,
|
||||
req_ctx->hw_context_size, DMA_TO_DEVICE);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
@ -2046,13 +2052,6 @@ static int ahash_process_req(struct ahash_request *areq, unsigned int nbytes)
|
|||
/* request SEC to INIT hash. */
|
||||
if (req_ctx->first && !req_ctx->swinit)
|
||||
edesc->desc.hdr |= DESC_HDR_MODE0_MDEU_INIT;
|
||||
if (is_sec1) {
|
||||
dma_addr_t dma_buf = ctx->dma_buf + req_ctx->buf_idx *
|
||||
HASH_MAX_BLOCK_SIZE;
|
||||
|
||||
dma_sync_single_for_device(dev, dma_buf,
|
||||
req_ctx->nbuf, DMA_TO_DEVICE);
|
||||
}
|
||||
|
||||
/* When the tfm context has a keylen, it's an HMAC.
|
||||
* A first or last (ie. not middle) descriptor must request HMAC.
|
||||
|
@ -2106,12 +2105,7 @@ static int ahash_export(struct ahash_request *areq, void *out)
|
|||
{
|
||||
struct talitos_ahash_req_ctx *req_ctx = ahash_request_ctx(areq);
|
||||
struct talitos_export_state *export = out;
|
||||
struct crypto_ahash *ahash = crypto_ahash_reqtfm(areq);
|
||||
struct talitos_ctx *ctx = crypto_ahash_ctx(ahash);
|
||||
struct device *dev = ctx->dev;
|
||||
|
||||
dma_sync_single_for_cpu(dev, ctx->dma_hw_context,
|
||||
req_ctx->hw_context_size, DMA_FROM_DEVICE);
|
||||
memcpy(export->hw_context, req_ctx->hw_context,
|
||||
req_ctx->hw_context_size);
|
||||
memcpy(export->buf, req_ctx->buf[req_ctx->buf_idx], req_ctx->nbuf);
|
||||
|
@ -2130,31 +2124,14 @@ static int ahash_import(struct ahash_request *areq, const void *in)
|
|||
struct crypto_ahash *tfm = crypto_ahash_reqtfm(areq);
|
||||
const struct talitos_export_state *export = in;
|
||||
unsigned int size;
|
||||
struct talitos_ctx *ctx = crypto_ahash_ctx(tfm);
|
||||
struct device *dev = ctx->dev;
|
||||
struct talitos_private *priv = dev_get_drvdata(dev);
|
||||
bool is_sec1 = has_ftr_sec1(priv);
|
||||
|
||||
memset(req_ctx, 0, sizeof(*req_ctx));
|
||||
size = (crypto_ahash_digestsize(tfm) <= SHA256_DIGEST_SIZE)
|
||||
? TALITOS_MDEU_CONTEXT_SIZE_MD5_SHA1_SHA256
|
||||
: TALITOS_MDEU_CONTEXT_SIZE_SHA384_SHA512;
|
||||
req_ctx->hw_context_size = size;
|
||||
if (ctx->dma_hw_context)
|
||||
dma_unmap_single(dev, ctx->dma_hw_context, size,
|
||||
DMA_BIDIRECTIONAL);
|
||||
|
||||
memcpy(req_ctx->hw_context, export->hw_context, size);
|
||||
ctx->dma_hw_context = dma_map_single(dev, req_ctx->hw_context, size,
|
||||
DMA_BIDIRECTIONAL);
|
||||
if (ctx->dma_buf)
|
||||
dma_unmap_single(dev, ctx->dma_buf, sizeof(req_ctx->buf),
|
||||
DMA_TO_DEVICE);
|
||||
memcpy(req_ctx->buf[0], export->buf, export->nbuf);
|
||||
if (is_sec1)
|
||||
ctx->dma_buf = dma_map_single(dev, req_ctx->buf,
|
||||
sizeof(req_ctx->buf),
|
||||
DMA_TO_DEVICE);
|
||||
req_ctx->swinit = export->swinit;
|
||||
req_ctx->first = export->first;
|
||||
req_ctx->last = export->last;
|
||||
|
@ -3064,27 +3041,6 @@ static void talitos_cra_exit(struct crypto_tfm *tfm)
|
|||
dma_unmap_single(dev, ctx->dma_key, ctx->keylen, DMA_TO_DEVICE);
|
||||
}
|
||||
|
||||
static void talitos_cra_exit_ahash(struct crypto_tfm *tfm)
|
||||
{
|
||||
struct talitos_ctx *ctx = crypto_tfm_ctx(tfm);
|
||||
struct device *dev = ctx->dev;
|
||||
unsigned int size;
|
||||
|
||||
talitos_cra_exit(tfm);
|
||||
|
||||
size = (crypto_ahash_digestsize(__crypto_ahash_cast(tfm)) <=
|
||||
SHA256_DIGEST_SIZE)
|
||||
? TALITOS_MDEU_CONTEXT_SIZE_MD5_SHA1_SHA256
|
||||
: TALITOS_MDEU_CONTEXT_SIZE_SHA384_SHA512;
|
||||
|
||||
if (ctx->dma_hw_context)
|
||||
dma_unmap_single(dev, ctx->dma_hw_context, size,
|
||||
DMA_BIDIRECTIONAL);
|
||||
if (ctx->dma_buf)
|
||||
dma_unmap_single(dev, ctx->dma_buf, HASH_MAX_BLOCK_SIZE * 2,
|
||||
DMA_TO_DEVICE);
|
||||
}
|
||||
|
||||
/*
|
||||
* given the alg's descriptor header template, determine whether descriptor
|
||||
* type and primary/secondary execution units required match the hw
|
||||
|
@ -3183,7 +3139,7 @@ static struct talitos_crypto_alg *talitos_alg_alloc(struct device *dev,
|
|||
case CRYPTO_ALG_TYPE_AHASH:
|
||||
alg = &t_alg->algt.alg.hash.halg.base;
|
||||
alg->cra_init = talitos_cra_init_ahash;
|
||||
alg->cra_exit = talitos_cra_exit_ahash;
|
||||
alg->cra_exit = talitos_cra_exit;
|
||||
alg->cra_type = &crypto_ahash_type;
|
||||
t_alg->algt.alg.hash.init = ahash_init;
|
||||
t_alg->algt.alg.hash.update = ahash_update;
|
||||
|
|
Loading…
Reference in New Issue