crypto: engine - permit to enqueue ashash_request
The current crypto engine allow only ablkcipher_request to be enqueued. Thus denying any use of it for hardware that also handle hash algo. This patch modify the API for allowing to enqueue ciphers and hash. Since omap-aes/omap-des are the only users, this patch also convert them to the new cryptoengine API. Signed-off-by: Corentin Labbe <clabbe.montjoie@gmail.com> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
This commit is contained in:
parent
2589ad8404
commit
4cba7cf025
|
@ -15,13 +15,11 @@
|
|||
#include <linux/err.h>
|
||||
#include <linux/delay.h>
|
||||
#include <crypto/engine.h>
|
||||
#include <crypto/internal/hash.h>
|
||||
#include "internal.h"
|
||||
|
||||
#define CRYPTO_ENGINE_MAX_QLEN 10
|
||||
|
||||
void crypto_finalize_request(struct crypto_engine *engine,
|
||||
struct ablkcipher_request *req, int err);
|
||||
|
||||
/**
|
||||
* crypto_pump_requests - dequeue one request from engine queue to process
|
||||
* @engine: the hardware engine
|
||||
|
@ -35,10 +33,11 @@ static void crypto_pump_requests(struct crypto_engine *engine,
|
|||
bool in_kthread)
|
||||
{
|
||||
struct crypto_async_request *async_req, *backlog;
|
||||
struct ablkcipher_request *req;
|
||||
struct ahash_request *hreq;
|
||||
struct ablkcipher_request *breq;
|
||||
unsigned long flags;
|
||||
bool was_busy = false;
|
||||
int ret;
|
||||
int ret, rtype;
|
||||
|
||||
spin_lock_irqsave(&engine->queue_lock, flags);
|
||||
|
||||
|
@ -83,9 +82,7 @@ static void crypto_pump_requests(struct crypto_engine *engine,
|
|||
if (!async_req)
|
||||
goto out;
|
||||
|
||||
req = ablkcipher_request_cast(async_req);
|
||||
|
||||
engine->cur_req = req;
|
||||
engine->cur_req = async_req;
|
||||
if (backlog)
|
||||
backlog->complete(backlog, -EINPROGRESS);
|
||||
|
||||
|
@ -96,6 +93,7 @@ static void crypto_pump_requests(struct crypto_engine *engine,
|
|||
|
||||
spin_unlock_irqrestore(&engine->queue_lock, flags);
|
||||
|
||||
rtype = crypto_tfm_alg_type(engine->cur_req->tfm);
|
||||
/* Until here we get the request need to be encrypted successfully */
|
||||
if (!was_busy && engine->prepare_crypt_hardware) {
|
||||
ret = engine->prepare_crypt_hardware(engine);
|
||||
|
@ -105,24 +103,55 @@ static void crypto_pump_requests(struct crypto_engine *engine,
|
|||
}
|
||||
}
|
||||
|
||||
if (engine->prepare_request) {
|
||||
ret = engine->prepare_request(engine, engine->cur_req);
|
||||
switch (rtype) {
|
||||
case CRYPTO_ALG_TYPE_AHASH:
|
||||
hreq = ahash_request_cast(engine->cur_req);
|
||||
if (engine->prepare_hash_request) {
|
||||
ret = engine->prepare_hash_request(engine, hreq);
|
||||
if (ret) {
|
||||
pr_err("failed to prepare request: %d\n", ret);
|
||||
goto req_err;
|
||||
}
|
||||
engine->cur_req_prepared = true;
|
||||
}
|
||||
ret = engine->hash_one_request(engine, hreq);
|
||||
if (ret) {
|
||||
pr_err("failed to prepare request: %d\n", ret);
|
||||
pr_err("failed to hash one request from queue\n");
|
||||
goto req_err;
|
||||
}
|
||||
engine->cur_req_prepared = true;
|
||||
return;
|
||||
case CRYPTO_ALG_TYPE_ABLKCIPHER:
|
||||
breq = ablkcipher_request_cast(engine->cur_req);
|
||||
if (engine->prepare_cipher_request) {
|
||||
ret = engine->prepare_cipher_request(engine, breq);
|
||||
if (ret) {
|
||||
pr_err("failed to prepare request: %d\n", ret);
|
||||
goto req_err;
|
||||
}
|
||||
engine->cur_req_prepared = true;
|
||||
}
|
||||
ret = engine->cipher_one_request(engine, breq);
|
||||
if (ret) {
|
||||
pr_err("failed to cipher one request from queue\n");
|
||||
goto req_err;
|
||||
}
|
||||
return;
|
||||
default:
|
||||
pr_err("failed to prepare request of unknown type\n");
|
||||
return;
|
||||
}
|
||||
|
||||
ret = engine->crypt_one_request(engine, engine->cur_req);
|
||||
if (ret) {
|
||||
pr_err("failed to crypt one request from queue\n");
|
||||
goto req_err;
|
||||
}
|
||||
return;
|
||||
|
||||
req_err:
|
||||
crypto_finalize_request(engine, engine->cur_req, ret);
|
||||
switch (rtype) {
|
||||
case CRYPTO_ALG_TYPE_AHASH:
|
||||
hreq = ahash_request_cast(engine->cur_req);
|
||||
crypto_finalize_hash_request(engine, hreq, ret);
|
||||
break;
|
||||
case CRYPTO_ALG_TYPE_ABLKCIPHER:
|
||||
breq = ablkcipher_request_cast(engine->cur_req);
|
||||
crypto_finalize_cipher_request(engine, breq, ret);
|
||||
break;
|
||||
}
|
||||
return;
|
||||
|
||||
out:
|
||||
|
@ -138,12 +167,14 @@ static void crypto_pump_work(struct kthread_work *work)
|
|||
}
|
||||
|
||||
/**
|
||||
* crypto_transfer_request - transfer the new request into the engine queue
|
||||
* crypto_transfer_cipher_request - transfer the new request into the
|
||||
* enginequeue
|
||||
* @engine: the hardware engine
|
||||
* @req: the request need to be listed into the engine queue
|
||||
*/
|
||||
int crypto_transfer_request(struct crypto_engine *engine,
|
||||
struct ablkcipher_request *req, bool need_pump)
|
||||
int crypto_transfer_cipher_request(struct crypto_engine *engine,
|
||||
struct ablkcipher_request *req,
|
||||
bool need_pump)
|
||||
{
|
||||
unsigned long flags;
|
||||
int ret;
|
||||
|
@ -163,46 +194,88 @@ int crypto_transfer_request(struct crypto_engine *engine,
|
|||
spin_unlock_irqrestore(&engine->queue_lock, flags);
|
||||
return ret;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(crypto_transfer_request);
|
||||
EXPORT_SYMBOL_GPL(crypto_transfer_cipher_request);
|
||||
|
||||
/**
|
||||
* crypto_transfer_request_to_engine - transfer one request to list into the
|
||||
* engine queue
|
||||
* crypto_transfer_cipher_request_to_engine - transfer one request to list
|
||||
* into the engine queue
|
||||
* @engine: the hardware engine
|
||||
* @req: the request need to be listed into the engine queue
|
||||
*/
|
||||
int crypto_transfer_request_to_engine(struct crypto_engine *engine,
|
||||
struct ablkcipher_request *req)
|
||||
int crypto_transfer_cipher_request_to_engine(struct crypto_engine *engine,
|
||||
struct ablkcipher_request *req)
|
||||
{
|
||||
return crypto_transfer_request(engine, req, true);
|
||||
return crypto_transfer_cipher_request(engine, req, true);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(crypto_transfer_request_to_engine);
|
||||
EXPORT_SYMBOL_GPL(crypto_transfer_cipher_request_to_engine);
|
||||
|
||||
/**
|
||||
* crypto_finalize_request - finalize one request if the request is done
|
||||
* crypto_transfer_hash_request - transfer the new request into the
|
||||
* enginequeue
|
||||
* @engine: the hardware engine
|
||||
* @req: the request need to be listed into the engine queue
|
||||
*/
|
||||
int crypto_transfer_hash_request(struct crypto_engine *engine,
|
||||
struct ahash_request *req, bool need_pump)
|
||||
{
|
||||
unsigned long flags;
|
||||
int ret;
|
||||
|
||||
spin_lock_irqsave(&engine->queue_lock, flags);
|
||||
|
||||
if (!engine->running) {
|
||||
spin_unlock_irqrestore(&engine->queue_lock, flags);
|
||||
return -ESHUTDOWN;
|
||||
}
|
||||
|
||||
ret = ahash_enqueue_request(&engine->queue, req);
|
||||
|
||||
if (!engine->busy && need_pump)
|
||||
queue_kthread_work(&engine->kworker, &engine->pump_requests);
|
||||
|
||||
spin_unlock_irqrestore(&engine->queue_lock, flags);
|
||||
return ret;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(crypto_transfer_hash_request);
|
||||
|
||||
/**
|
||||
* crypto_transfer_hash_request_to_engine - transfer one request to list
|
||||
* into the engine queue
|
||||
* @engine: the hardware engine
|
||||
* @req: the request need to be listed into the engine queue
|
||||
*/
|
||||
int crypto_transfer_hash_request_to_engine(struct crypto_engine *engine,
|
||||
struct ahash_request *req)
|
||||
{
|
||||
return crypto_transfer_hash_request(engine, req, true);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(crypto_transfer_hash_request_to_engine);
|
||||
|
||||
/**
|
||||
* crypto_finalize_cipher_request - finalize one request if the request is done
|
||||
* @engine: the hardware engine
|
||||
* @req: the request need to be finalized
|
||||
* @err: error number
|
||||
*/
|
||||
void crypto_finalize_request(struct crypto_engine *engine,
|
||||
struct ablkcipher_request *req, int err)
|
||||
void crypto_finalize_cipher_request(struct crypto_engine *engine,
|
||||
struct ablkcipher_request *req, int err)
|
||||
{
|
||||
unsigned long flags;
|
||||
bool finalize_cur_req = false;
|
||||
int ret;
|
||||
|
||||
spin_lock_irqsave(&engine->queue_lock, flags);
|
||||
if (engine->cur_req == req)
|
||||
if (engine->cur_req == &req->base)
|
||||
finalize_cur_req = true;
|
||||
spin_unlock_irqrestore(&engine->queue_lock, flags);
|
||||
|
||||
if (finalize_cur_req) {
|
||||
if (engine->cur_req_prepared && engine->unprepare_request) {
|
||||
ret = engine->unprepare_request(engine, req);
|
||||
if (engine->cur_req_prepared &&
|
||||
engine->unprepare_cipher_request) {
|
||||
ret = engine->unprepare_cipher_request(engine, req);
|
||||
if (ret)
|
||||
pr_err("failed to unprepare request\n");
|
||||
}
|
||||
|
||||
spin_lock_irqsave(&engine->queue_lock, flags);
|
||||
engine->cur_req = NULL;
|
||||
engine->cur_req_prepared = false;
|
||||
|
@ -213,7 +286,44 @@ void crypto_finalize_request(struct crypto_engine *engine,
|
|||
|
||||
queue_kthread_work(&engine->kworker, &engine->pump_requests);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(crypto_finalize_request);
|
||||
EXPORT_SYMBOL_GPL(crypto_finalize_cipher_request);
|
||||
|
||||
/**
|
||||
* crypto_finalize_hash_request - finalize one request if the request is done
|
||||
* @engine: the hardware engine
|
||||
* @req: the request need to be finalized
|
||||
* @err: error number
|
||||
*/
|
||||
void crypto_finalize_hash_request(struct crypto_engine *engine,
|
||||
struct ahash_request *req, int err)
|
||||
{
|
||||
unsigned long flags;
|
||||
bool finalize_cur_req = false;
|
||||
int ret;
|
||||
|
||||
spin_lock_irqsave(&engine->queue_lock, flags);
|
||||
if (engine->cur_req == &req->base)
|
||||
finalize_cur_req = true;
|
||||
spin_unlock_irqrestore(&engine->queue_lock, flags);
|
||||
|
||||
if (finalize_cur_req) {
|
||||
if (engine->cur_req_prepared &&
|
||||
engine->unprepare_hash_request) {
|
||||
ret = engine->unprepare_hash_request(engine, req);
|
||||
if (ret)
|
||||
pr_err("failed to unprepare request\n");
|
||||
}
|
||||
spin_lock_irqsave(&engine->queue_lock, flags);
|
||||
engine->cur_req = NULL;
|
||||
engine->cur_req_prepared = false;
|
||||
spin_unlock_irqrestore(&engine->queue_lock, flags);
|
||||
}
|
||||
|
||||
req->base.complete(&req->base, err);
|
||||
|
||||
queue_kthread_work(&engine->kworker, &engine->pump_requests);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(crypto_finalize_hash_request);
|
||||
|
||||
/**
|
||||
* crypto_engine_start - start the hardware engine
|
||||
|
@ -250,7 +360,7 @@ EXPORT_SYMBOL_GPL(crypto_engine_start);
|
|||
int crypto_engine_stop(struct crypto_engine *engine)
|
||||
{
|
||||
unsigned long flags;
|
||||
unsigned limit = 500;
|
||||
unsigned int limit = 500;
|
||||
int ret = 0;
|
||||
|
||||
spin_lock_irqsave(&engine->queue_lock, flags);
|
||||
|
|
|
@ -520,7 +520,7 @@ static void omap_aes_finish_req(struct omap_aes_dev *dd, int err)
|
|||
|
||||
pr_debug("err: %d\n", err);
|
||||
|
||||
crypto_finalize_request(dd->engine, req, err);
|
||||
crypto_finalize_cipher_request(dd->engine, req, err);
|
||||
}
|
||||
|
||||
static int omap_aes_crypt_dma_stop(struct omap_aes_dev *dd)
|
||||
|
@ -593,7 +593,7 @@ static int omap_aes_handle_queue(struct omap_aes_dev *dd,
|
|||
struct ablkcipher_request *req)
|
||||
{
|
||||
if (req)
|
||||
return crypto_transfer_request_to_engine(dd->engine, req);
|
||||
return crypto_transfer_cipher_request_to_engine(dd->engine, req);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
@ -1209,8 +1209,8 @@ static int omap_aes_probe(struct platform_device *pdev)
|
|||
if (!dd->engine)
|
||||
goto err_algs;
|
||||
|
||||
dd->engine->prepare_request = omap_aes_prepare_req;
|
||||
dd->engine->crypt_one_request = omap_aes_crypt_req;
|
||||
dd->engine->prepare_cipher_request = omap_aes_prepare_req;
|
||||
dd->engine->cipher_one_request = omap_aes_crypt_req;
|
||||
err = crypto_engine_start(dd->engine);
|
||||
if (err)
|
||||
goto err_engine;
|
||||
|
|
|
@ -507,7 +507,7 @@ static void omap_des_finish_req(struct omap_des_dev *dd, int err)
|
|||
pr_debug("err: %d\n", err);
|
||||
|
||||
pm_runtime_put(dd->dev);
|
||||
crypto_finalize_request(dd->engine, req, err);
|
||||
crypto_finalize_cipher_request(dd->engine, req, err);
|
||||
}
|
||||
|
||||
static int omap_des_crypt_dma_stop(struct omap_des_dev *dd)
|
||||
|
@ -575,7 +575,7 @@ static int omap_des_handle_queue(struct omap_des_dev *dd,
|
|||
struct ablkcipher_request *req)
|
||||
{
|
||||
if (req)
|
||||
return crypto_transfer_request_to_engine(dd->engine, req);
|
||||
return crypto_transfer_cipher_request_to_engine(dd->engine, req);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
@ -1099,8 +1099,8 @@ static int omap_des_probe(struct platform_device *pdev)
|
|||
if (!dd->engine)
|
||||
goto err_algs;
|
||||
|
||||
dd->engine->prepare_request = omap_des_prepare_req;
|
||||
dd->engine->crypt_one_request = omap_des_crypt_req;
|
||||
dd->engine->prepare_cipher_request = omap_des_prepare_req;
|
||||
dd->engine->cipher_one_request = omap_des_crypt_req;
|
||||
err = crypto_engine_start(dd->engine);
|
||||
if (err)
|
||||
goto err_engine;
|
||||
|
|
|
@ -17,6 +17,7 @@
|
|||
#include <linux/kernel.h>
|
||||
#include <linux/kthread.h>
|
||||
#include <crypto/algapi.h>
|
||||
#include <crypto/hash.h>
|
||||
|
||||
#define ENGINE_NAME_LEN 30
|
||||
/*
|
||||
|
@ -36,9 +37,12 @@
|
|||
* @unprepare_crypt_hardware: there are currently no more requests on the
|
||||
* queue so the subsystem notifies the driver that it may relax the
|
||||
* hardware by issuing this call
|
||||
* @prepare_request: do some prepare if need before handle the current request
|
||||
* @unprepare_request: undo any work done by prepare_message()
|
||||
* @crypt_one_request: do encryption for current request
|
||||
* @prepare_cipher_request: do some prepare if need before handle the current request
|
||||
* @unprepare_cipher_request: undo any work done by prepare_cipher_request()
|
||||
* @cipher_one_request: do encryption for current request
|
||||
* @prepare_hash_request: do some prepare if need before handle the current request
|
||||
* @unprepare_hash_request: undo any work done by prepare_hash_request()
|
||||
* @hash_one_request: do hash for current request
|
||||
* @kworker: thread struct for request pump
|
||||
* @kworker_task: pointer to task for request pump kworker thread
|
||||
* @pump_requests: work struct for scheduling work to the request pump
|
||||
|
@ -61,27 +65,40 @@ struct crypto_engine {
|
|||
int (*prepare_crypt_hardware)(struct crypto_engine *engine);
|
||||
int (*unprepare_crypt_hardware)(struct crypto_engine *engine);
|
||||
|
||||
int (*prepare_request)(struct crypto_engine *engine,
|
||||
struct ablkcipher_request *req);
|
||||
int (*unprepare_request)(struct crypto_engine *engine,
|
||||
struct ablkcipher_request *req);
|
||||
int (*crypt_one_request)(struct crypto_engine *engine,
|
||||
struct ablkcipher_request *req);
|
||||
int (*prepare_cipher_request)(struct crypto_engine *engine,
|
||||
struct ablkcipher_request *req);
|
||||
int (*unprepare_cipher_request)(struct crypto_engine *engine,
|
||||
struct ablkcipher_request *req);
|
||||
int (*prepare_hash_request)(struct crypto_engine *engine,
|
||||
struct ahash_request *req);
|
||||
int (*unprepare_hash_request)(struct crypto_engine *engine,
|
||||
struct ahash_request *req);
|
||||
int (*cipher_one_request)(struct crypto_engine *engine,
|
||||
struct ablkcipher_request *req);
|
||||
int (*hash_one_request)(struct crypto_engine *engine,
|
||||
struct ahash_request *req);
|
||||
|
||||
struct kthread_worker kworker;
|
||||
struct task_struct *kworker_task;
|
||||
struct kthread_work pump_requests;
|
||||
|
||||
void *priv_data;
|
||||
struct ablkcipher_request *cur_req;
|
||||
struct crypto_async_request *cur_req;
|
||||
};
|
||||
|
||||
int crypto_transfer_request(struct crypto_engine *engine,
|
||||
struct ablkcipher_request *req, bool need_pump);
|
||||
int crypto_transfer_request_to_engine(struct crypto_engine *engine,
|
||||
struct ablkcipher_request *req);
|
||||
void crypto_finalize_request(struct crypto_engine *engine,
|
||||
struct ablkcipher_request *req, int err);
|
||||
int crypto_transfer_cipher_request(struct crypto_engine *engine,
|
||||
struct ablkcipher_request *req,
|
||||
bool need_pump);
|
||||
int crypto_transfer_cipher_request_to_engine(struct crypto_engine *engine,
|
||||
struct ablkcipher_request *req);
|
||||
int crypto_transfer_hash_request(struct crypto_engine *engine,
|
||||
struct ahash_request *req, bool need_pump);
|
||||
int crypto_transfer_hash_request_to_engine(struct crypto_engine *engine,
|
||||
struct ahash_request *req);
|
||||
void crypto_finalize_cipher_request(struct crypto_engine *engine,
|
||||
struct ablkcipher_request *req, int err);
|
||||
void crypto_finalize_hash_request(struct crypto_engine *engine,
|
||||
struct ahash_request *req, int err);
|
||||
int crypto_engine_start(struct crypto_engine *engine);
|
||||
int crypto_engine_stop(struct crypto_engine *engine);
|
||||
struct crypto_engine *crypto_engine_alloc_init(struct device *dev, bool rt);
|
||||
|
|
Loading…
Reference in New Issue