1 /* SPDX-License-Identifier: GPL-2.0-or-later */
5 * Copyright (c) 2016 Baolin Wang <baolin.wang@linaro.org>
7 #ifndef _CRYPTO_ENGINE_H
8 #define _CRYPTO_ENGINE_H
10 #include <linux/crypto.h>
11 #include <linux/list.h>
12 #include <linux/kernel.h>
13 #include <linux/kthread.h>
14 #include <crypto/algapi.h>
15 #include <crypto/aead.h>
16 #include <crypto/akcipher.h>
17 #include <crypto/hash.h>
18 #include <crypto/skcipher.h>
20 #define ENGINE_NAME_LEN 30
22 * struct crypto_engine - crypto hardware engine
23 * @name: the engine name
24 * @idling: the engine is entering idle state
25 * @busy: request pump is busy
26 * @running: the engine is on working
27 * @cur_req_prepared: current request is prepared
28 * @list: link with the global crypto engine list
29 * @queue_lock: spinlock to syncronise access to request queue
30 * @queue: the crypto queue of the engine
31 * @rt: whether this queue is set to run as a realtime task
32 * @prepare_crypt_hardware: a request will soon arrive from the queue
33 * so the subsystem requests the driver to prepare the hardware
34 * by issuing this call
35 * @unprepare_crypt_hardware: there are currently no more requests on the
36 * queue so the subsystem notifies the driver that it may relax the
37 * hardware by issuing this call
38 * @kworker: kthread worker struct for request pump
39 * @pump_requests: work struct for scheduling work to the request pump
40 * @priv_data: the engine private data
41 * @cur_req: the current request which is on processing
43 struct crypto_engine
{
44 char name
[ENGINE_NAME_LEN
];
48 bool cur_req_prepared
;
50 struct list_head list
;
51 spinlock_t queue_lock
;
52 struct crypto_queue queue
;
57 int (*prepare_crypt_hardware
)(struct crypto_engine
*engine
);
58 int (*unprepare_crypt_hardware
)(struct crypto_engine
*engine
);
60 struct kthread_worker
*kworker
;
61 struct kthread_work pump_requests
;
64 struct crypto_async_request
*cur_req
;
68 * struct crypto_engine_op - crypto hardware engine operations
69 * @prepare__request: do some prepare if need before handle the current request
70 * @unprepare_request: undo any work done by prepare_request()
71 * @do_one_request: do encryption for current request
73 struct crypto_engine_op
{
74 int (*prepare_request
)(struct crypto_engine
*engine
,
76 int (*unprepare_request
)(struct crypto_engine
*engine
,
78 int (*do_one_request
)(struct crypto_engine
*engine
,
82 struct crypto_engine_ctx
{
83 struct crypto_engine_op op
;
86 int crypto_transfer_aead_request_to_engine(struct crypto_engine
*engine
,
87 struct aead_request
*req
);
88 int crypto_transfer_akcipher_request_to_engine(struct crypto_engine
*engine
,
89 struct akcipher_request
*req
);
90 int crypto_transfer_hash_request_to_engine(struct crypto_engine
*engine
,
91 struct ahash_request
*req
);
92 int crypto_transfer_skcipher_request_to_engine(struct crypto_engine
*engine
,
93 struct skcipher_request
*req
);
94 void crypto_finalize_aead_request(struct crypto_engine
*engine
,
95 struct aead_request
*req
, int err
);
96 void crypto_finalize_akcipher_request(struct crypto_engine
*engine
,
97 struct akcipher_request
*req
, int err
);
98 void crypto_finalize_hash_request(struct crypto_engine
*engine
,
99 struct ahash_request
*req
, int err
);
100 void crypto_finalize_skcipher_request(struct crypto_engine
*engine
,
101 struct skcipher_request
*req
, int err
);
102 int crypto_engine_start(struct crypto_engine
*engine
);
103 int crypto_engine_stop(struct crypto_engine
*engine
);
104 struct crypto_engine
*crypto_engine_alloc_init(struct device
*dev
, bool rt
);
105 int crypto_engine_exit(struct crypto_engine
*engine
);
107 #endif /* _CRYPTO_ENGINE_H */