1
0
Fork 0

crypto: engine - Permit to enqueue all async requests

The crypto engine could actually only enqueue hash and ablkcipher request.
This patch permit it to enqueue any type of crypto_async_request.

Signed-off-by: Corentin Labbe <clabbe.montjoie@gmail.com>
Tested-by: Fabien Dessenne <fabien.dessenne@st.com>
Tested-by: Fabien Dessenne <fabien.dessenne@st.com>
Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
hifive-unleashed-5.1
Corentin LABBE 2018-01-26 20:15:30 +01:00 committed by Herbert Xu
parent ce09a6c042
commit 218d1cc186
2 changed files with 207 additions and 170 deletions

View File

@ -15,12 +15,49 @@
#include <linux/err.h> #include <linux/err.h>
#include <linux/delay.h> #include <linux/delay.h>
#include <crypto/engine.h> #include <crypto/engine.h>
#include <crypto/internal/hash.h>
#include <uapi/linux/sched/types.h> #include <uapi/linux/sched/types.h>
#include "internal.h" #include "internal.h"
#define CRYPTO_ENGINE_MAX_QLEN 10 #define CRYPTO_ENGINE_MAX_QLEN 10
/**
* crypto_finalize_request - finalize one request if the request is done
* @engine: the hardware engine
* @req: the request need to be finalized
* @err: error number
*/
static void crypto_finalize_request(struct crypto_engine *engine,
struct crypto_async_request *req, int err)
{
unsigned long flags;
bool finalize_cur_req = false;
int ret;
struct crypto_engine_ctx *enginectx;
spin_lock_irqsave(&engine->queue_lock, flags);
if (engine->cur_req == req)
finalize_cur_req = true;
spin_unlock_irqrestore(&engine->queue_lock, flags);
if (finalize_cur_req) {
enginectx = crypto_tfm_ctx(req->tfm);
if (engine->cur_req_prepared &&
enginectx->op.unprepare_request) {
ret = enginectx->op.unprepare_request(engine, req);
if (ret)
dev_err(engine->dev, "failed to unprepare request\n");
}
spin_lock_irqsave(&engine->queue_lock, flags);
engine->cur_req = NULL;
engine->cur_req_prepared = false;
spin_unlock_irqrestore(&engine->queue_lock, flags);
}
req->complete(req, err);
kthread_queue_work(engine->kworker, &engine->pump_requests);
}
/** /**
* crypto_pump_requests - dequeue one request from engine queue to process * crypto_pump_requests - dequeue one request from engine queue to process
* @engine: the hardware engine * @engine: the hardware engine
@ -34,11 +71,10 @@ static void crypto_pump_requests(struct crypto_engine *engine,
bool in_kthread) bool in_kthread)
{ {
struct crypto_async_request *async_req, *backlog; struct crypto_async_request *async_req, *backlog;
struct ahash_request *hreq;
struct ablkcipher_request *breq;
unsigned long flags; unsigned long flags;
bool was_busy = false; bool was_busy = false;
int ret, rtype; int ret;
struct crypto_engine_ctx *enginectx;
spin_lock_irqsave(&engine->queue_lock, flags); spin_lock_irqsave(&engine->queue_lock, flags);
@ -94,7 +130,6 @@ static void crypto_pump_requests(struct crypto_engine *engine,
spin_unlock_irqrestore(&engine->queue_lock, flags); spin_unlock_irqrestore(&engine->queue_lock, flags);
rtype = crypto_tfm_alg_type(engine->cur_req->tfm);
/* Until here we get the request need to be encrypted successfully */ /* Until here we get the request need to be encrypted successfully */
if (!was_busy && engine->prepare_crypt_hardware) { if (!was_busy && engine->prepare_crypt_hardware) {
ret = engine->prepare_crypt_hardware(engine); ret = engine->prepare_crypt_hardware(engine);
@ -104,57 +139,31 @@ static void crypto_pump_requests(struct crypto_engine *engine,
} }
} }
switch (rtype) { enginectx = crypto_tfm_ctx(async_req->tfm);
case CRYPTO_ALG_TYPE_AHASH:
hreq = ahash_request_cast(engine->cur_req); if (enginectx->op.prepare_request) {
if (engine->prepare_hash_request) { ret = enginectx->op.prepare_request(engine, async_req);
ret = engine->prepare_hash_request(engine, hreq);
if (ret) {
dev_err(engine->dev, "failed to prepare request: %d\n",
ret);
goto req_err;
}
engine->cur_req_prepared = true;
}
ret = engine->hash_one_request(engine, hreq);
if (ret) { if (ret) {
dev_err(engine->dev, "failed to hash one request from queue\n"); dev_err(engine->dev, "failed to prepare request: %d\n",
ret);
goto req_err; goto req_err;
} }
return; engine->cur_req_prepared = true;
case CRYPTO_ALG_TYPE_ABLKCIPHER:
breq = ablkcipher_request_cast(engine->cur_req);
if (engine->prepare_cipher_request) {
ret = engine->prepare_cipher_request(engine, breq);
if (ret) {
dev_err(engine->dev, "failed to prepare request: %d\n",
ret);
goto req_err;
}
engine->cur_req_prepared = true;
}
ret = engine->cipher_one_request(engine, breq);
if (ret) {
dev_err(engine->dev, "failed to cipher one request from queue\n");
goto req_err;
}
return;
default:
dev_err(engine->dev, "failed to prepare request of unknown type\n");
return;
} }
if (!enginectx->op.do_one_request) {
dev_err(engine->dev, "failed to do request\n");
ret = -EINVAL;
goto req_err;
}
ret = enginectx->op.do_one_request(engine, async_req);
if (ret) {
dev_err(engine->dev, "Failed to do one request from queue: %d\n", ret);
goto req_err;
}
return;
req_err: req_err:
switch (rtype) { crypto_finalize_request(engine, async_req, ret);
case CRYPTO_ALG_TYPE_AHASH:
hreq = ahash_request_cast(engine->cur_req);
crypto_finalize_hash_request(engine, hreq, ret);
break;
case CRYPTO_ALG_TYPE_ABLKCIPHER:
breq = ablkcipher_request_cast(engine->cur_req);
crypto_finalize_cipher_request(engine, breq, ret);
break;
}
return; return;
out: out:
@ -170,13 +179,12 @@ static void crypto_pump_work(struct kthread_work *work)
} }
/** /**
* crypto_transfer_cipher_request - transfer the new request into the * crypto_transfer_request - transfer the new request into the engine queue
* enginequeue
* @engine: the hardware engine * @engine: the hardware engine
* @req: the request need to be listed into the engine queue * @req: the request need to be listed into the engine queue
*/ */
int crypto_transfer_cipher_request(struct crypto_engine *engine, static int crypto_transfer_request(struct crypto_engine *engine,
struct ablkcipher_request *req, struct crypto_async_request *req,
bool need_pump) bool need_pump)
{ {
unsigned long flags; unsigned long flags;
@ -189,7 +197,7 @@ int crypto_transfer_cipher_request(struct crypto_engine *engine,
return -ESHUTDOWN; return -ESHUTDOWN;
} }
ret = ablkcipher_enqueue_request(&engine->queue, req); ret = crypto_enqueue_request(&engine->queue, req);
if (!engine->busy && need_pump) if (!engine->busy && need_pump)
kthread_queue_work(engine->kworker, &engine->pump_requests); kthread_queue_work(engine->kworker, &engine->pump_requests);
@ -197,102 +205,131 @@ int crypto_transfer_cipher_request(struct crypto_engine *engine,
spin_unlock_irqrestore(&engine->queue_lock, flags); spin_unlock_irqrestore(&engine->queue_lock, flags);
return ret; return ret;
} }
EXPORT_SYMBOL_GPL(crypto_transfer_cipher_request);
/** /**
* crypto_transfer_cipher_request_to_engine - transfer one request to list * crypto_transfer_request_to_engine - transfer one request to list
* into the engine queue * into the engine queue
* @engine: the hardware engine * @engine: the hardware engine
* @req: the request need to be listed into the engine queue * @req: the request need to be listed into the engine queue
*/ */
int crypto_transfer_cipher_request_to_engine(struct crypto_engine *engine, static int crypto_transfer_request_to_engine(struct crypto_engine *engine,
struct ablkcipher_request *req) struct crypto_async_request *req)
{ {
return crypto_transfer_cipher_request(engine, req, true); return crypto_transfer_request(engine, req, true);
} }
EXPORT_SYMBOL_GPL(crypto_transfer_cipher_request_to_engine);
/** /**
* crypto_transfer_hash_request - transfer the new request into the * crypto_transfer_ablkcipher_request_to_engine - transfer one ablkcipher_request
* enginequeue * to list into the engine queue
* @engine: the hardware engine
* @req: the request need to be listed into the engine queue
* TODO: Remove this function when skcipher conversion is finished
*/
int crypto_transfer_ablkcipher_request_to_engine(struct crypto_engine *engine,
struct ablkcipher_request *req)
{
return crypto_transfer_request_to_engine(engine, &req->base);
}
EXPORT_SYMBOL_GPL(crypto_transfer_ablkcipher_request_to_engine);
/**
* crypto_transfer_aead_request_to_engine - transfer one aead_request
* to list into the engine queue
* @engine: the hardware engine * @engine: the hardware engine
* @req: the request need to be listed into the engine queue * @req: the request need to be listed into the engine queue
*/ */
int crypto_transfer_hash_request(struct crypto_engine *engine, int crypto_transfer_aead_request_to_engine(struct crypto_engine *engine,
struct ahash_request *req, bool need_pump) struct aead_request *req)
{ {
unsigned long flags; return crypto_transfer_request_to_engine(engine, &req->base);
int ret;
spin_lock_irqsave(&engine->queue_lock, flags);
if (!engine->running) {
spin_unlock_irqrestore(&engine->queue_lock, flags);
return -ESHUTDOWN;
}
ret = ahash_enqueue_request(&engine->queue, req);
if (!engine->busy && need_pump)
kthread_queue_work(engine->kworker, &engine->pump_requests);
spin_unlock_irqrestore(&engine->queue_lock, flags);
return ret;
} }
EXPORT_SYMBOL_GPL(crypto_transfer_hash_request); EXPORT_SYMBOL_GPL(crypto_transfer_aead_request_to_engine);
/** /**
* crypto_transfer_hash_request_to_engine - transfer one request to list * crypto_transfer_akcipher_request_to_engine - transfer one akcipher_request
* into the engine queue * to list into the engine queue
* @engine: the hardware engine
* @req: the request need to be listed into the engine queue
*/
int crypto_transfer_akcipher_request_to_engine(struct crypto_engine *engine,
struct akcipher_request *req)
{
return crypto_transfer_request_to_engine(engine, &req->base);
}
EXPORT_SYMBOL_GPL(crypto_transfer_akcipher_request_to_engine);
/**
* crypto_transfer_hash_request_to_engine - transfer one ahash_request
* to list into the engine queue
* @engine: the hardware engine * @engine: the hardware engine
* @req: the request need to be listed into the engine queue * @req: the request need to be listed into the engine queue
*/ */
int crypto_transfer_hash_request_to_engine(struct crypto_engine *engine, int crypto_transfer_hash_request_to_engine(struct crypto_engine *engine,
struct ahash_request *req) struct ahash_request *req)
{ {
return crypto_transfer_hash_request(engine, req, true); return crypto_transfer_request_to_engine(engine, &req->base);
} }
EXPORT_SYMBOL_GPL(crypto_transfer_hash_request_to_engine); EXPORT_SYMBOL_GPL(crypto_transfer_hash_request_to_engine);
/** /**
* crypto_finalize_cipher_request - finalize one request if the request is done * crypto_transfer_skcipher_request_to_engine - transfer one skcipher_request
* to list into the engine queue
* @engine: the hardware engine
* @req: the request need to be listed into the engine queue
*/
int crypto_transfer_skcipher_request_to_engine(struct crypto_engine *engine,
struct skcipher_request *req)
{
return crypto_transfer_request_to_engine(engine, &req->base);
}
EXPORT_SYMBOL_GPL(crypto_transfer_skcipher_request_to_engine);
/**
* crypto_finalize_ablkcipher_request - finalize one ablkcipher_request if
* the request is done
* @engine: the hardware engine
* @req: the request need to be finalized
* @err: error number
* TODO: Remove this function when skcipher conversion is finished
*/
void crypto_finalize_ablkcipher_request(struct crypto_engine *engine,
struct ablkcipher_request *req, int err)
{
return crypto_finalize_request(engine, &req->base, err);
}
EXPORT_SYMBOL_GPL(crypto_finalize_ablkcipher_request);
/**
* crypto_finalize_aead_request - finalize one aead_request if
* the request is done
* @engine: the hardware engine * @engine: the hardware engine
* @req: the request need to be finalized * @req: the request need to be finalized
* @err: error number * @err: error number
*/ */
void crypto_finalize_cipher_request(struct crypto_engine *engine, void crypto_finalize_aead_request(struct crypto_engine *engine,
struct ablkcipher_request *req, int err) struct aead_request *req, int err)
{ {
unsigned long flags; return crypto_finalize_request(engine, &req->base, err);
bool finalize_cur_req = false;
int ret;
spin_lock_irqsave(&engine->queue_lock, flags);
if (engine->cur_req == &req->base)
finalize_cur_req = true;
spin_unlock_irqrestore(&engine->queue_lock, flags);
if (finalize_cur_req) {
if (engine->cur_req_prepared &&
engine->unprepare_cipher_request) {
ret = engine->unprepare_cipher_request(engine, req);
if (ret)
dev_err(engine->dev, "failed to unprepare request\n");
}
spin_lock_irqsave(&engine->queue_lock, flags);
engine->cur_req = NULL;
engine->cur_req_prepared = false;
spin_unlock_irqrestore(&engine->queue_lock, flags);
}
req->base.complete(&req->base, err);
kthread_queue_work(engine->kworker, &engine->pump_requests);
} }
EXPORT_SYMBOL_GPL(crypto_finalize_cipher_request); EXPORT_SYMBOL_GPL(crypto_finalize_aead_request);
/** /**
* crypto_finalize_hash_request - finalize one request if the request is done * crypto_finalize_akcipher_request - finalize one akcipher_request if
* the request is done
* @engine: the hardware engine
* @req: the request need to be finalized
* @err: error number
*/
void crypto_finalize_akcipher_request(struct crypto_engine *engine,
struct akcipher_request *req, int err)
{
return crypto_finalize_request(engine, &req->base, err);
}
EXPORT_SYMBOL_GPL(crypto_finalize_akcipher_request);
/**
* crypto_finalize_hash_request - finalize one ahash_request if
* the request is done
* @engine: the hardware engine * @engine: the hardware engine
* @req: the request need to be finalized * @req: the request need to be finalized
* @err: error number * @err: error number
@ -300,34 +337,24 @@ EXPORT_SYMBOL_GPL(crypto_finalize_cipher_request);
void crypto_finalize_hash_request(struct crypto_engine *engine, void crypto_finalize_hash_request(struct crypto_engine *engine,
struct ahash_request *req, int err) struct ahash_request *req, int err)
{ {
unsigned long flags; return crypto_finalize_request(engine, &req->base, err);
bool finalize_cur_req = false;
int ret;
spin_lock_irqsave(&engine->queue_lock, flags);
if (engine->cur_req == &req->base)
finalize_cur_req = true;
spin_unlock_irqrestore(&engine->queue_lock, flags);
if (finalize_cur_req) {
if (engine->cur_req_prepared &&
engine->unprepare_hash_request) {
ret = engine->unprepare_hash_request(engine, req);
if (ret)
dev_err(engine->dev, "failed to unprepare request\n");
}
spin_lock_irqsave(&engine->queue_lock, flags);
engine->cur_req = NULL;
engine->cur_req_prepared = false;
spin_unlock_irqrestore(&engine->queue_lock, flags);
}
req->base.complete(&req->base, err);
kthread_queue_work(engine->kworker, &engine->pump_requests);
} }
EXPORT_SYMBOL_GPL(crypto_finalize_hash_request); EXPORT_SYMBOL_GPL(crypto_finalize_hash_request);
/**
* crypto_finalize_skcipher_request - finalize one skcipher_request if
* the request is done
* @engine: the hardware engine
* @req: the request need to be finalized
* @err: error number
*/
void crypto_finalize_skcipher_request(struct crypto_engine *engine,
struct skcipher_request *req, int err)
{
return crypto_finalize_request(engine, &req->base, err);
}
EXPORT_SYMBOL_GPL(crypto_finalize_skcipher_request);
/** /**
* crypto_engine_start - start the hardware engine * crypto_engine_start - start the hardware engine
* @engine: the hardware engine need to be started * @engine: the hardware engine need to be started

View File

@ -17,7 +17,10 @@
#include <linux/kernel.h> #include <linux/kernel.h>
#include <linux/kthread.h> #include <linux/kthread.h>
#include <crypto/algapi.h> #include <crypto/algapi.h>
#include <crypto/aead.h>
#include <crypto/akcipher.h>
#include <crypto/hash.h> #include <crypto/hash.h>
#include <crypto/skcipher.h>
#define ENGINE_NAME_LEN 30 #define ENGINE_NAME_LEN 30
/* /*
@ -37,12 +40,6 @@
* @unprepare_crypt_hardware: there are currently no more requests on the * @unprepare_crypt_hardware: there are currently no more requests on the
* queue so the subsystem notifies the driver that it may relax the * queue so the subsystem notifies the driver that it may relax the
* hardware by issuing this call * hardware by issuing this call
* @prepare_cipher_request: do some prepare if need before handle the current request
* @unprepare_cipher_request: undo any work done by prepare_cipher_request()
* @cipher_one_request: do encryption for current request
* @prepare_hash_request: do some prepare if need before handle the current request
* @unprepare_hash_request: undo any work done by prepare_hash_request()
* @hash_one_request: do hash for current request
* @kworker: kthread worker struct for request pump * @kworker: kthread worker struct for request pump
* @pump_requests: work struct for scheduling work to the request pump * @pump_requests: work struct for scheduling work to the request pump
* @priv_data: the engine private data * @priv_data: the engine private data
@ -65,19 +62,6 @@ struct crypto_engine {
int (*prepare_crypt_hardware)(struct crypto_engine *engine); int (*prepare_crypt_hardware)(struct crypto_engine *engine);
int (*unprepare_crypt_hardware)(struct crypto_engine *engine); int (*unprepare_crypt_hardware)(struct crypto_engine *engine);
int (*prepare_cipher_request)(struct crypto_engine *engine,
struct ablkcipher_request *req);
int (*unprepare_cipher_request)(struct crypto_engine *engine,
struct ablkcipher_request *req);
int (*prepare_hash_request)(struct crypto_engine *engine,
struct ahash_request *req);
int (*unprepare_hash_request)(struct crypto_engine *engine,
struct ahash_request *req);
int (*cipher_one_request)(struct crypto_engine *engine,
struct ablkcipher_request *req);
int (*hash_one_request)(struct crypto_engine *engine,
struct ahash_request *req);
struct kthread_worker *kworker; struct kthread_worker *kworker;
struct kthread_work pump_requests; struct kthread_work pump_requests;
@ -85,19 +69,45 @@ struct crypto_engine {
struct crypto_async_request *cur_req; struct crypto_async_request *cur_req;
}; };
int crypto_transfer_cipher_request(struct crypto_engine *engine, /*
struct ablkcipher_request *req, * struct crypto_engine_op - crypto hardware engine operations
bool need_pump); * @prepare__request: do some prepare if need before handle the current request
int crypto_transfer_cipher_request_to_engine(struct crypto_engine *engine, * @unprepare_request: undo any work done by prepare_request()
struct ablkcipher_request *req); * @do_one_request: do encryption for current request
int crypto_transfer_hash_request(struct crypto_engine *engine, */
struct ahash_request *req, bool need_pump); struct crypto_engine_op {
int (*prepare_request)(struct crypto_engine *engine,
void *areq);
int (*unprepare_request)(struct crypto_engine *engine,
void *areq);
int (*do_one_request)(struct crypto_engine *engine,
void *areq);
};
struct crypto_engine_ctx {
struct crypto_engine_op op;
};
int crypto_transfer_ablkcipher_request_to_engine(struct crypto_engine *engine,
struct ablkcipher_request *req);
int crypto_transfer_aead_request_to_engine(struct crypto_engine *engine,
struct aead_request *req);
int crypto_transfer_akcipher_request_to_engine(struct crypto_engine *engine,
struct akcipher_request *req);
int crypto_transfer_hash_request_to_engine(struct crypto_engine *engine, int crypto_transfer_hash_request_to_engine(struct crypto_engine *engine,
struct ahash_request *req); struct ahash_request *req);
void crypto_finalize_cipher_request(struct crypto_engine *engine, int crypto_transfer_skcipher_request_to_engine(struct crypto_engine *engine,
struct ablkcipher_request *req, int err); struct skcipher_request *req);
void crypto_finalize_ablkcipher_request(struct crypto_engine *engine,
struct ablkcipher_request *req, int err);
void crypto_finalize_aead_request(struct crypto_engine *engine,
struct aead_request *req, int err);
void crypto_finalize_akcipher_request(struct crypto_engine *engine,
struct akcipher_request *req, int err);
void crypto_finalize_hash_request(struct crypto_engine *engine, void crypto_finalize_hash_request(struct crypto_engine *engine,
struct ahash_request *req, int err); struct ahash_request *req, int err);
void crypto_finalize_skcipher_request(struct crypto_engine *engine,
struct skcipher_request *req, int err);
int crypto_engine_start(struct crypto_engine *engine); int crypto_engine_start(struct crypto_engine *engine);
int crypto_engine_stop(struct crypto_engine *engine); int crypto_engine_stop(struct crypto_engine *engine);
struct crypto_engine *crypto_engine_alloc_init(struct device *dev, bool rt); struct crypto_engine *crypto_engine_alloc_init(struct device *dev, bool rt);