crypto: qat - refactor submission logic
All the algorithms in qat_algs.c and qat_asym_algs.c use the same pattern to submit messages to the HW queues. Move the submission loop to a new function, qat_alg_send_message(), and share it between the symmetric and the asymmetric algorithms. As part of this rework, since the number of retries before returning an error is inconsistent between the symmetric and asymmetric implementations, set it to a value that works for both (i.e. 20, was 10 in qat_algs.c and 100 in qat_asym_algs.c) In addition fix the return code reported when the HW queues are full. In that case return -ENOSPC instead of -EBUSY. Including stable in CC since (1) the error code returned if the HW queues are full is incorrect and (2) to facilitate the backport of the next fix "crypto: qat - add backlog mechanism". Cc: stable@vger.kernel.org Signed-off-by: Giovanni Cabiddu <giovanni.cabiddu@intel.com> Reviewed-by: Marco Chiappero <marco.chiappero@intel.com> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
This commit is contained in:
parent
e0831e7af4
commit
af88d3c109
|
@ -16,6 +16,7 @@ intel_qat-objs := adf_cfg.o \
|
|||
qat_crypto.o \
|
||||
qat_algs.o \
|
||||
qat_asym_algs.o \
|
||||
qat_algs_send.o \
|
||||
qat_uclo.o \
|
||||
qat_hal.o
|
||||
|
||||
|
|
|
@ -17,7 +17,7 @@
|
|||
#include <crypto/xts.h>
|
||||
#include <linux/dma-mapping.h>
|
||||
#include "adf_accel_devices.h"
|
||||
#include "adf_transport.h"
|
||||
#include "qat_algs_send.h"
|
||||
#include "adf_common_drv.h"
|
||||
#include "qat_crypto.h"
|
||||
#include "icp_qat_hw.h"
|
||||
|
@ -939,6 +939,17 @@ void qat_alg_callback(void *resp)
|
|||
qat_req->cb(qat_resp, qat_req);
|
||||
}
|
||||
|
||||
static int qat_alg_send_sym_message(struct qat_crypto_request *qat_req,
|
||||
struct qat_crypto_instance *inst)
|
||||
{
|
||||
struct qat_alg_req req;
|
||||
|
||||
req.fw_req = (u32 *)&qat_req->req;
|
||||
req.tx_ring = inst->sym_tx;
|
||||
|
||||
return qat_alg_send_message(&req);
|
||||
}
|
||||
|
||||
static int qat_alg_aead_dec(struct aead_request *areq)
|
||||
{
|
||||
struct crypto_aead *aead_tfm = crypto_aead_reqtfm(areq);
|
||||
|
@ -949,7 +960,7 @@ static int qat_alg_aead_dec(struct aead_request *areq)
|
|||
struct icp_qat_fw_la_auth_req_params *auth_param;
|
||||
struct icp_qat_fw_la_bulk_req *msg;
|
||||
int digst_size = crypto_aead_authsize(aead_tfm);
|
||||
int ret, ctr = 0;
|
||||
int ret;
|
||||
u32 cipher_len;
|
||||
|
||||
cipher_len = areq->cryptlen - digst_size;
|
||||
|
@ -975,15 +986,12 @@ static int qat_alg_aead_dec(struct aead_request *areq)
|
|||
auth_param = (void *)((u8 *)cipher_param + sizeof(*cipher_param));
|
||||
auth_param->auth_off = 0;
|
||||
auth_param->auth_len = areq->assoclen + cipher_param->cipher_length;
|
||||
do {
|
||||
ret = adf_send_message(ctx->inst->sym_tx, (u32 *)msg);
|
||||
} while (ret == -EAGAIN && ctr++ < 10);
|
||||
|
||||
if (ret == -EAGAIN) {
|
||||
ret = qat_alg_send_sym_message(qat_req, ctx->inst);
|
||||
if (ret == -ENOSPC)
|
||||
qat_alg_free_bufl(ctx->inst, qat_req);
|
||||
return -EBUSY;
|
||||
}
|
||||
return -EINPROGRESS;
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int qat_alg_aead_enc(struct aead_request *areq)
|
||||
|
@ -996,7 +1004,7 @@ static int qat_alg_aead_enc(struct aead_request *areq)
|
|||
struct icp_qat_fw_la_auth_req_params *auth_param;
|
||||
struct icp_qat_fw_la_bulk_req *msg;
|
||||
u8 *iv = areq->iv;
|
||||
int ret, ctr = 0;
|
||||
int ret;
|
||||
|
||||
if (areq->cryptlen % AES_BLOCK_SIZE != 0)
|
||||
return -EINVAL;
|
||||
|
@ -1023,15 +1031,11 @@ static int qat_alg_aead_enc(struct aead_request *areq)
|
|||
auth_param->auth_off = 0;
|
||||
auth_param->auth_len = areq->assoclen + areq->cryptlen;
|
||||
|
||||
do {
|
||||
ret = adf_send_message(ctx->inst->sym_tx, (u32 *)msg);
|
||||
} while (ret == -EAGAIN && ctr++ < 10);
|
||||
|
||||
if (ret == -EAGAIN) {
|
||||
ret = qat_alg_send_sym_message(qat_req, ctx->inst);
|
||||
if (ret == -ENOSPC)
|
||||
qat_alg_free_bufl(ctx->inst, qat_req);
|
||||
return -EBUSY;
|
||||
}
|
||||
return -EINPROGRESS;
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int qat_alg_skcipher_rekey(struct qat_alg_skcipher_ctx *ctx,
|
||||
|
@ -1184,7 +1188,7 @@ static int qat_alg_skcipher_encrypt(struct skcipher_request *req)
|
|||
struct qat_crypto_request *qat_req = skcipher_request_ctx(req);
|
||||
struct icp_qat_fw_la_cipher_req_params *cipher_param;
|
||||
struct icp_qat_fw_la_bulk_req *msg;
|
||||
int ret, ctr = 0;
|
||||
int ret;
|
||||
|
||||
if (req->cryptlen == 0)
|
||||
return 0;
|
||||
|
@ -1208,15 +1212,11 @@ static int qat_alg_skcipher_encrypt(struct skcipher_request *req)
|
|||
|
||||
qat_alg_set_req_iv(qat_req);
|
||||
|
||||
do {
|
||||
ret = adf_send_message(ctx->inst->sym_tx, (u32 *)msg);
|
||||
} while (ret == -EAGAIN && ctr++ < 10);
|
||||
|
||||
if (ret == -EAGAIN) {
|
||||
ret = qat_alg_send_sym_message(qat_req, ctx->inst);
|
||||
if (ret == -ENOSPC)
|
||||
qat_alg_free_bufl(ctx->inst, qat_req);
|
||||
return -EBUSY;
|
||||
}
|
||||
return -EINPROGRESS;
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int qat_alg_skcipher_blk_encrypt(struct skcipher_request *req)
|
||||
|
@ -1253,7 +1253,7 @@ static int qat_alg_skcipher_decrypt(struct skcipher_request *req)
|
|||
struct qat_crypto_request *qat_req = skcipher_request_ctx(req);
|
||||
struct icp_qat_fw_la_cipher_req_params *cipher_param;
|
||||
struct icp_qat_fw_la_bulk_req *msg;
|
||||
int ret, ctr = 0;
|
||||
int ret;
|
||||
|
||||
if (req->cryptlen == 0)
|
||||
return 0;
|
||||
|
@ -1278,15 +1278,11 @@ static int qat_alg_skcipher_decrypt(struct skcipher_request *req)
|
|||
qat_alg_set_req_iv(qat_req);
|
||||
qat_alg_update_iv(qat_req);
|
||||
|
||||
do {
|
||||
ret = adf_send_message(ctx->inst->sym_tx, (u32 *)msg);
|
||||
} while (ret == -EAGAIN && ctr++ < 10);
|
||||
|
||||
if (ret == -EAGAIN) {
|
||||
ret = qat_alg_send_sym_message(qat_req, ctx->inst);
|
||||
if (ret == -ENOSPC)
|
||||
qat_alg_free_bufl(ctx->inst, qat_req);
|
||||
return -EBUSY;
|
||||
}
|
||||
return -EINPROGRESS;
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int qat_alg_skcipher_blk_decrypt(struct skcipher_request *req)
|
||||
|
|
|
@ -0,0 +1,21 @@
|
|||
// SPDX-License-Identifier: (BSD-3-Clause OR GPL-2.0-only)
|
||||
/* Copyright(c) 2022 Intel Corporation */
|
||||
#include "adf_transport.h"
|
||||
#include "qat_algs_send.h"
|
||||
#include "qat_crypto.h"
|
||||
|
||||
#define ADF_MAX_RETRIES 20
|
||||
|
||||
int qat_alg_send_message(struct qat_alg_req *req)
|
||||
{
|
||||
int ret = 0, ctr = 0;
|
||||
|
||||
do {
|
||||
ret = adf_send_message(req->tx_ring, req->fw_req);
|
||||
} while (ret == -EAGAIN && ctr++ < ADF_MAX_RETRIES);
|
||||
|
||||
if (ret == -EAGAIN)
|
||||
return -ENOSPC;
|
||||
|
||||
return -EINPROGRESS;
|
||||
}
|
|
@ -0,0 +1,10 @@
|
|||
/* SPDX-License-Identifier: (BSD-3-Clause OR GPL-2.0-only) */
|
||||
/* Copyright(c) 2022 Intel Corporation */
|
||||
#ifndef QAT_ALGS_SEND_H
|
||||
#define QAT_ALGS_SEND_H
|
||||
|
||||
#include "qat_crypto.h"
|
||||
|
||||
int qat_alg_send_message(struct qat_alg_req *req);
|
||||
|
||||
#endif
|
|
@ -12,6 +12,7 @@
|
|||
#include <crypto/scatterwalk.h>
|
||||
#include "icp_qat_fw_pke.h"
|
||||
#include "adf_accel_devices.h"
|
||||
#include "qat_algs_send.h"
|
||||
#include "adf_transport.h"
|
||||
#include "adf_common_drv.h"
|
||||
#include "qat_crypto.h"
|
||||
|
@ -137,6 +138,17 @@ struct qat_asym_request {
|
|||
void (*cb)(struct icp_qat_fw_pke_resp *resp);
|
||||
} __aligned(64);
|
||||
|
||||
static int qat_alg_send_asym_message(struct qat_asym_request *qat_req,
|
||||
struct qat_crypto_instance *inst)
|
||||
{
|
||||
struct qat_alg_req req;
|
||||
|
||||
req.fw_req = (u32 *)&qat_req->req;
|
||||
req.tx_ring = inst->pke_tx;
|
||||
|
||||
return qat_alg_send_message(&req);
|
||||
}
|
||||
|
||||
static void qat_dh_cb(struct icp_qat_fw_pke_resp *resp)
|
||||
{
|
||||
struct qat_asym_request *req = (void *)(__force long)resp->opaque;
|
||||
|
@ -213,7 +225,7 @@ static int qat_dh_compute_value(struct kpp_request *req)
|
|||
struct qat_asym_request *qat_req =
|
||||
PTR_ALIGN(kpp_request_ctx(req), 64);
|
||||
struct icp_qat_fw_pke_request *msg = &qat_req->req;
|
||||
int ret, ctr = 0;
|
||||
int ret;
|
||||
int n_input_params = 0;
|
||||
|
||||
if (unlikely(!ctx->xa))
|
||||
|
@ -338,13 +350,13 @@ static int qat_dh_compute_value(struct kpp_request *req)
|
|||
msg->input_param_count = n_input_params;
|
||||
msg->output_param_count = 1;
|
||||
|
||||
do {
|
||||
ret = adf_send_message(ctx->inst->pke_tx, (u32 *)msg);
|
||||
} while (ret == -EBUSY && ctr++ < 100);
|
||||
ret = qat_alg_send_asym_message(qat_req, ctx->inst);
|
||||
if (ret == -ENOSPC)
|
||||
goto unmap_all;
|
||||
|
||||
if (!ret)
|
||||
return -EINPROGRESS;
|
||||
return ret;
|
||||
|
||||
unmap_all:
|
||||
if (!dma_mapping_error(dev, qat_req->phy_out))
|
||||
dma_unmap_single(dev, qat_req->phy_out,
|
||||
sizeof(struct qat_dh_output_params),
|
||||
|
@ -645,7 +657,7 @@ static int qat_rsa_enc(struct akcipher_request *req)
|
|||
struct qat_asym_request *qat_req =
|
||||
PTR_ALIGN(akcipher_request_ctx(req), 64);
|
||||
struct icp_qat_fw_pke_request *msg = &qat_req->req;
|
||||
int ret, ctr = 0;
|
||||
int ret;
|
||||
|
||||
if (unlikely(!ctx->n || !ctx->e))
|
||||
return -EINVAL;
|
||||
|
@ -735,13 +747,14 @@ static int qat_rsa_enc(struct akcipher_request *req)
|
|||
msg->pke_mid.opaque = (u64)(__force long)qat_req;
|
||||
msg->input_param_count = 3;
|
||||
msg->output_param_count = 1;
|
||||
do {
|
||||
ret = adf_send_message(ctx->inst->pke_tx, (u32 *)msg);
|
||||
} while (ret == -EBUSY && ctr++ < 100);
|
||||
|
||||
if (!ret)
|
||||
return -EINPROGRESS;
|
||||
ret = qat_alg_send_asym_message(qat_req, ctx->inst);
|
||||
if (ret == -ENOSPC)
|
||||
goto unmap_all;
|
||||
|
||||
return ret;
|
||||
|
||||
unmap_all:
|
||||
if (!dma_mapping_error(dev, qat_req->phy_out))
|
||||
dma_unmap_single(dev, qat_req->phy_out,
|
||||
sizeof(struct qat_rsa_output_params),
|
||||
|
@ -779,7 +792,7 @@ static int qat_rsa_dec(struct akcipher_request *req)
|
|||
struct qat_asym_request *qat_req =
|
||||
PTR_ALIGN(akcipher_request_ctx(req), 64);
|
||||
struct icp_qat_fw_pke_request *msg = &qat_req->req;
|
||||
int ret, ctr = 0;
|
||||
int ret;
|
||||
|
||||
if (unlikely(!ctx->n || !ctx->d))
|
||||
return -EINVAL;
|
||||
|
@ -887,13 +900,14 @@ static int qat_rsa_dec(struct akcipher_request *req)
|
|||
msg->input_param_count = 3;
|
||||
|
||||
msg->output_param_count = 1;
|
||||
do {
|
||||
ret = adf_send_message(ctx->inst->pke_tx, (u32 *)msg);
|
||||
} while (ret == -EBUSY && ctr++ < 100);
|
||||
|
||||
if (!ret)
|
||||
return -EINPROGRESS;
|
||||
ret = qat_alg_send_asym_message(qat_req, ctx->inst);
|
||||
if (ret == -ENOSPC)
|
||||
goto unmap_all;
|
||||
|
||||
return ret;
|
||||
|
||||
unmap_all:
|
||||
if (!dma_mapping_error(dev, qat_req->phy_out))
|
||||
dma_unmap_single(dev, qat_req->phy_out,
|
||||
sizeof(struct qat_rsa_output_params),
|
||||
|
|
|
@ -9,6 +9,11 @@
|
|||
#include "adf_accel_devices.h"
|
||||
#include "icp_qat_fw_la.h"
|
||||
|
||||
struct qat_alg_req {
|
||||
u32 *fw_req;
|
||||
struct adf_etr_ring_data *tx_ring;
|
||||
};
|
||||
|
||||
struct qat_crypto_instance {
|
||||
struct adf_etr_ring_data *sym_tx;
|
||||
struct adf_etr_ring_data *sym_rx;
|
||||
|
|
Loading…
Reference in New Issue