1 /* SPDX-License-Identifier: GPL-2.0-or-later */
5 * Copyright (c) 2016 Baolin Wang <baolin.wang@linaro.org>
7 #ifndef _CRYPTO_ENGINE_H
8 #define _CRYPTO_ENGINE_H
10 #include <linux/crypto.h>
11 #include <linux/list.h>
12 #include <linux/kernel.h>
13 #include <linux/kthread.h>
14 #include <crypto/algapi.h>
15 #include <crypto/aead.h>
16 #include <crypto/akcipher.h>
17 #include <crypto/hash.h>
18 #include <crypto/skcipher.h>
20 #define ENGINE_NAME_LEN 30
22 * struct crypto_engine - crypto hardware engine
23 * @name: the engine name
24 * @idling: the engine is entering idle state
25 * @busy: request pump is busy
26 * @running: the engine is on working
27 * @retry_support: indication that the hardware allows re-execution
28 * of a failed backlog request
29 * crypto-engine, in head position to keep order
30 * @list: link with the global crypto engine list
31 * @queue_lock: spinlock to syncronise access to request queue
32 * @queue: the crypto queue of the engine
33 * @rt: whether this queue is set to run as a realtime task
34 * @prepare_crypt_hardware: a request will soon arrive from the queue
35 * so the subsystem requests the driver to prepare the hardware
36 * by issuing this call
37 * @unprepare_crypt_hardware: there are currently no more requests on the
38 * queue so the subsystem notifies the driver that it may relax the
39 * hardware by issuing this call
40 * @do_batch_requests: execute a batch of requests. Depends on multiple
42 * @kworker: kthread worker struct for request pump
43 * @pump_requests: work struct for scheduling work to the request pump
44 * @priv_data: the engine private data
45 * @cur_req: the current request which is on processing
47 struct crypto_engine
{
48 char name
[ENGINE_NAME_LEN
];
55 struct list_head list
;
56 spinlock_t queue_lock
;
57 struct crypto_queue queue
;
62 int (*prepare_crypt_hardware
)(struct crypto_engine
*engine
);
63 int (*unprepare_crypt_hardware
)(struct crypto_engine
*engine
);
64 int (*do_batch_requests
)(struct crypto_engine
*engine
);
67 struct kthread_worker
*kworker
;
68 struct kthread_work pump_requests
;
71 struct crypto_async_request
*cur_req
;
75 * struct crypto_engine_op - crypto hardware engine operations
76 * @prepare__request: do some prepare if need before handle the current request
77 * @unprepare_request: undo any work done by prepare_request()
78 * @do_one_request: do encryption for current request
80 struct crypto_engine_op
{
81 int (*prepare_request
)(struct crypto_engine
*engine
,
83 int (*unprepare_request
)(struct crypto_engine
*engine
,
85 int (*do_one_request
)(struct crypto_engine
*engine
,
89 struct crypto_engine_ctx
{
90 struct crypto_engine_op op
;
93 int crypto_transfer_aead_request_to_engine(struct crypto_engine
*engine
,
94 struct aead_request
*req
);
95 int crypto_transfer_akcipher_request_to_engine(struct crypto_engine
*engine
,
96 struct akcipher_request
*req
);
97 int crypto_transfer_hash_request_to_engine(struct crypto_engine
*engine
,
98 struct ahash_request
*req
);
99 int crypto_transfer_skcipher_request_to_engine(struct crypto_engine
*engine
,
100 struct skcipher_request
*req
);
101 void crypto_finalize_aead_request(struct crypto_engine
*engine
,
102 struct aead_request
*req
, int err
);
103 void crypto_finalize_akcipher_request(struct crypto_engine
*engine
,
104 struct akcipher_request
*req
, int err
);
105 void crypto_finalize_hash_request(struct crypto_engine
*engine
,
106 struct ahash_request
*req
, int err
);
107 void crypto_finalize_skcipher_request(struct crypto_engine
*engine
,
108 struct skcipher_request
*req
, int err
);
109 int crypto_engine_start(struct crypto_engine
*engine
);
110 int crypto_engine_stop(struct crypto_engine
*engine
);
111 struct crypto_engine
*crypto_engine_alloc_init(struct device
*dev
, bool rt
);
112 struct crypto_engine
*crypto_engine_alloc_init_and_set(struct device
*dev
,
114 int (*cbk_do_batch
)(struct crypto_engine
*engine
),
116 int crypto_engine_exit(struct crypto_engine
*engine
);
118 #endif /* _CRYPTO_ENGINE_H */