WIP FPC-III support
[linux/fpc-iii.git] / drivers / crypto / amlogic / amlogic-gxl-core.c
blob5bbeff433c8c09d03486e2f8956fa11789545ad8
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * amlgoic-core.c - hardware cryptographic offloader for Amlogic GXL SoC
5 * Copyright (C) 2018-2019 Corentin Labbe <clabbe@baylibre.com>
7 * Core file which registers crypto algorithms supported by the hardware.
8 */
9 #include <linux/clk.h>
10 #include <linux/crypto.h>
11 #include <linux/io.h>
12 #include <linux/interrupt.h>
13 #include <linux/irq.h>
14 #include <linux/module.h>
15 #include <linux/of.h>
16 #include <linux/of_device.h>
17 #include <linux/platform_device.h>
18 #include <crypto/internal/skcipher.h>
19 #include <linux/dma-mapping.h>
21 #include "amlogic-gxl.h"
23 static irqreturn_t meson_irq_handler(int irq, void *data)
25 struct meson_dev *mc = (struct meson_dev *)data;
26 int flow;
27 u32 p;
29 for (flow = 0; flow < MAXFLOW; flow++) {
30 if (mc->irqs[flow] == irq) {
31 p = readl(mc->base + ((0x04 + flow) << 2));
32 if (p) {
33 writel_relaxed(0xF, mc->base + ((0x4 + flow) << 2));
34 mc->chanlist[flow].status = 1;
35 complete(&mc->chanlist[flow].complete);
36 return IRQ_HANDLED;
38 dev_err(mc->dev, "%s %d Got irq for flow %d but ctrl is empty\n", __func__, irq, flow);
42 dev_err(mc->dev, "%s %d from unknown irq\n", __func__, irq);
43 return IRQ_HANDLED;
46 static struct meson_alg_template mc_algs[] = {
48 .type = CRYPTO_ALG_TYPE_SKCIPHER,
49 .blockmode = MESON_OPMODE_CBC,
50 .alg.skcipher = {
51 .base = {
52 .cra_name = "cbc(aes)",
53 .cra_driver_name = "cbc-aes-gxl",
54 .cra_priority = 400,
55 .cra_blocksize = AES_BLOCK_SIZE,
56 .cra_flags = CRYPTO_ALG_TYPE_SKCIPHER |
57 CRYPTO_ALG_ASYNC | CRYPTO_ALG_ALLOCATES_MEMORY |
58 CRYPTO_ALG_NEED_FALLBACK,
59 .cra_ctxsize = sizeof(struct meson_cipher_tfm_ctx),
60 .cra_module = THIS_MODULE,
61 .cra_alignmask = 0xf,
62 .cra_init = meson_cipher_init,
63 .cra_exit = meson_cipher_exit,
65 .min_keysize = AES_MIN_KEY_SIZE,
66 .max_keysize = AES_MAX_KEY_SIZE,
67 .ivsize = AES_BLOCK_SIZE,
68 .setkey = meson_aes_setkey,
69 .encrypt = meson_skencrypt,
70 .decrypt = meson_skdecrypt,
74 .type = CRYPTO_ALG_TYPE_SKCIPHER,
75 .blockmode = MESON_OPMODE_ECB,
76 .alg.skcipher = {
77 .base = {
78 .cra_name = "ecb(aes)",
79 .cra_driver_name = "ecb-aes-gxl",
80 .cra_priority = 400,
81 .cra_blocksize = AES_BLOCK_SIZE,
82 .cra_flags = CRYPTO_ALG_TYPE_SKCIPHER |
83 CRYPTO_ALG_ASYNC | CRYPTO_ALG_ALLOCATES_MEMORY |
84 CRYPTO_ALG_NEED_FALLBACK,
85 .cra_ctxsize = sizeof(struct meson_cipher_tfm_ctx),
86 .cra_module = THIS_MODULE,
87 .cra_alignmask = 0xf,
88 .cra_init = meson_cipher_init,
89 .cra_exit = meson_cipher_exit,
91 .min_keysize = AES_MIN_KEY_SIZE,
92 .max_keysize = AES_MAX_KEY_SIZE,
93 .setkey = meson_aes_setkey,
94 .encrypt = meson_skencrypt,
95 .decrypt = meson_skdecrypt,
100 #ifdef CONFIG_CRYPTO_DEV_AMLOGIC_GXL_DEBUG
101 static int meson_debugfs_show(struct seq_file *seq, void *v)
103 struct meson_dev *mc = seq->private;
104 int i;
106 for (i = 0; i < MAXFLOW; i++)
107 seq_printf(seq, "Channel %d: nreq %lu\n", i, mc->chanlist[i].stat_req);
109 for (i = 0; i < ARRAY_SIZE(mc_algs); i++) {
110 switch (mc_algs[i].type) {
111 case CRYPTO_ALG_TYPE_SKCIPHER:
112 seq_printf(seq, "%s %s %lu %lu\n",
113 mc_algs[i].alg.skcipher.base.cra_driver_name,
114 mc_algs[i].alg.skcipher.base.cra_name,
115 mc_algs[i].stat_req, mc_algs[i].stat_fb);
116 break;
119 return 0;
121 DEFINE_SHOW_ATTRIBUTE(meson_debugfs);
122 #endif
124 static void meson_free_chanlist(struct meson_dev *mc, int i)
126 while (i >= 0) {
127 crypto_engine_exit(mc->chanlist[i].engine);
128 if (mc->chanlist[i].tl)
129 dma_free_coherent(mc->dev, sizeof(struct meson_desc) * MAXDESC,
130 mc->chanlist[i].tl,
131 mc->chanlist[i].t_phy);
132 i--;
137 * Allocate the channel list structure
139 static int meson_allocate_chanlist(struct meson_dev *mc)
141 int i, err;
143 mc->chanlist = devm_kcalloc(mc->dev, MAXFLOW,
144 sizeof(struct meson_flow), GFP_KERNEL);
145 if (!mc->chanlist)
146 return -ENOMEM;
148 for (i = 0; i < MAXFLOW; i++) {
149 init_completion(&mc->chanlist[i].complete);
151 mc->chanlist[i].engine = crypto_engine_alloc_init(mc->dev, true);
152 if (!mc->chanlist[i].engine) {
153 dev_err(mc->dev, "Cannot allocate engine\n");
154 i--;
155 err = -ENOMEM;
156 goto error_engine;
158 err = crypto_engine_start(mc->chanlist[i].engine);
159 if (err) {
160 dev_err(mc->dev, "Cannot start engine\n");
161 goto error_engine;
163 mc->chanlist[i].tl = dma_alloc_coherent(mc->dev,
164 sizeof(struct meson_desc) * MAXDESC,
165 &mc->chanlist[i].t_phy,
166 GFP_KERNEL);
167 if (!mc->chanlist[i].tl) {
168 err = -ENOMEM;
169 goto error_engine;
172 return 0;
173 error_engine:
174 meson_free_chanlist(mc, i);
175 return err;
178 static int meson_register_algs(struct meson_dev *mc)
180 int err, i;
182 for (i = 0; i < ARRAY_SIZE(mc_algs); i++) {
183 mc_algs[i].mc = mc;
184 switch (mc_algs[i].type) {
185 case CRYPTO_ALG_TYPE_SKCIPHER:
186 err = crypto_register_skcipher(&mc_algs[i].alg.skcipher);
187 if (err) {
188 dev_err(mc->dev, "Fail to register %s\n",
189 mc_algs[i].alg.skcipher.base.cra_name);
190 mc_algs[i].mc = NULL;
191 return err;
193 break;
197 return 0;
200 static void meson_unregister_algs(struct meson_dev *mc)
202 int i;
204 for (i = 0; i < ARRAY_SIZE(mc_algs); i++) {
205 if (!mc_algs[i].mc)
206 continue;
207 switch (mc_algs[i].type) {
208 case CRYPTO_ALG_TYPE_SKCIPHER:
209 crypto_unregister_skcipher(&mc_algs[i].alg.skcipher);
210 break;
215 static int meson_crypto_probe(struct platform_device *pdev)
217 struct meson_dev *mc;
218 int err, i;
220 if (!pdev->dev.of_node)
221 return -ENODEV;
223 mc = devm_kzalloc(&pdev->dev, sizeof(*mc), GFP_KERNEL);
224 if (!mc)
225 return -ENOMEM;
227 mc->dev = &pdev->dev;
228 platform_set_drvdata(pdev, mc);
230 mc->base = devm_platform_ioremap_resource(pdev, 0);
231 if (IS_ERR(mc->base)) {
232 err = PTR_ERR(mc->base);
233 dev_err(&pdev->dev, "Cannot request MMIO err=%d\n", err);
234 return err;
236 mc->busclk = devm_clk_get(&pdev->dev, "blkmv");
237 if (IS_ERR(mc->busclk)) {
238 err = PTR_ERR(mc->busclk);
239 dev_err(&pdev->dev, "Cannot get core clock err=%d\n", err);
240 return err;
243 mc->irqs = devm_kcalloc(mc->dev, MAXFLOW, sizeof(int), GFP_KERNEL);
244 for (i = 0; i < MAXFLOW; i++) {
245 mc->irqs[i] = platform_get_irq(pdev, i);
246 if (mc->irqs[i] < 0)
247 return mc->irqs[i];
249 err = devm_request_irq(&pdev->dev, mc->irqs[i], meson_irq_handler, 0,
250 "gxl-crypto", mc);
251 if (err < 0) {
252 dev_err(mc->dev, "Cannot request IRQ for flow %d\n", i);
253 return err;
257 err = clk_prepare_enable(mc->busclk);
258 if (err != 0) {
259 dev_err(&pdev->dev, "Cannot prepare_enable busclk\n");
260 return err;
263 err = meson_allocate_chanlist(mc);
264 if (err)
265 goto error_flow;
267 err = meson_register_algs(mc);
268 if (err)
269 goto error_alg;
271 #ifdef CONFIG_CRYPTO_DEV_AMLOGIC_GXL_DEBUG
272 mc->dbgfs_dir = debugfs_create_dir("gxl-crypto", NULL);
273 debugfs_create_file("stats", 0444, mc->dbgfs_dir, mc, &meson_debugfs_fops);
274 #endif
276 return 0;
277 error_alg:
278 meson_unregister_algs(mc);
279 error_flow:
280 meson_free_chanlist(mc, MAXFLOW - 1);
281 clk_disable_unprepare(mc->busclk);
282 return err;
285 static int meson_crypto_remove(struct platform_device *pdev)
287 struct meson_dev *mc = platform_get_drvdata(pdev);
289 #ifdef CONFIG_CRYPTO_DEV_AMLOGIC_GXL_DEBUG
290 debugfs_remove_recursive(mc->dbgfs_dir);
291 #endif
293 meson_unregister_algs(mc);
295 meson_free_chanlist(mc, MAXFLOW - 1);
297 clk_disable_unprepare(mc->busclk);
298 return 0;
301 static const struct of_device_id meson_crypto_of_match_table[] = {
302 { .compatible = "amlogic,gxl-crypto", },
305 MODULE_DEVICE_TABLE(of, meson_crypto_of_match_table);
307 static struct platform_driver meson_crypto_driver = {
308 .probe = meson_crypto_probe,
309 .remove = meson_crypto_remove,
310 .driver = {
311 .name = "gxl-crypto",
312 .of_match_table = meson_crypto_of_match_table,
316 module_platform_driver(meson_crypto_driver);
318 MODULE_DESCRIPTION("Amlogic GXL cryptographic offloader");
319 MODULE_LICENSE("GPL");
320 MODULE_AUTHOR("Corentin Labbe <clabbe@baylibre.com>");