1 // SPDX-License-Identifier: GPL-2.0
2 /* Copyright (C) 2012-2019 ARM Limited or its affiliates. */
4 #include <linux/kernel.h>
5 #include <linux/module.h>
7 #include <linux/crypto.h>
8 #include <linux/moduleparam.h>
9 #include <linux/types.h>
10 #include <linux/interrupt.h>
11 #include <linux/platform_device.h>
12 #include <linux/slab.h>
13 #include <linux/spinlock.h>
15 #include <linux/clk.h>
16 #include <linux/of_address.h>
18 #include "cc_driver.h"
19 #include "cc_request_mgr.h"
20 #include "cc_buffer_mgr.h"
21 #include "cc_debugfs.h"
22 #include "cc_cipher.h"
25 #include "cc_sram_mgr.h"
30 module_param_named(dump_desc
, cc_dump_desc
, bool, 0600);
31 MODULE_PARM_DESC(cc_dump_desc
, "Dump descriptors to kernel log as debugging aid");
33 module_param_named(dump_bytes
, cc_dump_bytes
, bool, 0600);
34 MODULE_PARM_DESC(cc_dump_bytes
, "Dump buffers to kernel log as debugging aid");
36 static bool cc_sec_disable
;
37 module_param_named(sec_disable
, cc_sec_disable
, bool, 0600);
38 MODULE_PARM_DESC(cc_sec_disable
, "Disable security functions");
50 #define CC_HW_RESET_LOOP_COUNT 10
52 /* Note: PIDR3 holds CMOD/Rev so ignored for HW identification purposes */
53 static const u32 pidr_0124_offsets
[CC_NUM_IDRS
] = {
54 CC_REG(PERIPHERAL_ID_0
), CC_REG(PERIPHERAL_ID_1
),
55 CC_REG(PERIPHERAL_ID_2
), CC_REG(PERIPHERAL_ID_4
)
58 static const u32 cidr_0123_offsets
[CC_NUM_IDRS
] = {
59 CC_REG(COMPONENT_ID_0
), CC_REG(COMPONENT_ID_1
),
60 CC_REG(COMPONENT_ID_2
), CC_REG(COMPONENT_ID_3
)
63 /* Hardware revisions defs. */
65 /* The 703 is a OSCCA only variant of the 713 */
66 static const struct cc_hw_data cc703_hw
= {
67 .name
= "703", .rev
= CC_HW_REV_713
, .cidr_0123
= 0xB105F00DU
,
68 .pidr_0124
= 0x040BB0D0U
, .std_bodies
= CC_STD_OSCCA
71 static const struct cc_hw_data cc713_hw
= {
72 .name
= "713", .rev
= CC_HW_REV_713
, .cidr_0123
= 0xB105F00DU
,
73 .pidr_0124
= 0x040BB0D0U
, .std_bodies
= CC_STD_ALL
76 static const struct cc_hw_data cc712_hw
= {
77 .name
= "712", .rev
= CC_HW_REV_712
, .sig
= 0xDCC71200U
,
78 .std_bodies
= CC_STD_ALL
81 static const struct cc_hw_data cc710_hw
= {
82 .name
= "710", .rev
= CC_HW_REV_710
, .sig
= 0xDCC63200U
,
83 .std_bodies
= CC_STD_ALL
86 static const struct cc_hw_data cc630p_hw
= {
87 .name
= "630P", .rev
= CC_HW_REV_630
, .sig
= 0xDCC63000U
,
88 .std_bodies
= CC_STD_ALL
91 static const struct of_device_id arm_ccree_dev_of_match
[] = {
92 { .compatible
= "arm,cryptocell-703-ree", .data
= &cc703_hw
},
93 { .compatible
= "arm,cryptocell-713-ree", .data
= &cc713_hw
},
94 { .compatible
= "arm,cryptocell-712-ree", .data
= &cc712_hw
},
95 { .compatible
= "arm,cryptocell-710-ree", .data
= &cc710_hw
},
96 { .compatible
= "arm,cryptocell-630p-ree", .data
= &cc630p_hw
},
99 MODULE_DEVICE_TABLE(of
, arm_ccree_dev_of_match
);
101 static u32
cc_read_idr(struct cc_drvdata
*drvdata
, const u32
*idr_offsets
)
105 u8 regs
[CC_NUM_IDRS
];
109 for (i
= 0; i
< CC_NUM_IDRS
; ++i
)
110 idr
.regs
[i
] = cc_ioread(drvdata
, idr_offsets
[i
]);
112 return le32_to_cpu(idr
.val
);
115 void __dump_byte_array(const char *name
, const u8
*buf
, size_t len
)
122 snprintf(prefix
, sizeof(prefix
), "%s[%zu]: ", name
, len
);
124 print_hex_dump(KERN_DEBUG
, prefix
, DUMP_PREFIX_ADDRESS
, 16, 1, buf
,
128 static irqreturn_t
cc_isr(int irq
, void *dev_id
)
130 struct cc_drvdata
*drvdata
= (struct cc_drvdata
*)dev_id
;
131 struct device
*dev
= drvdata_to_dev(drvdata
);
135 /* STAT_OP_TYPE_GENERIC STAT_PHASE_0: Interrupt */
136 /* if driver suspended return, probably shared interrupt */
137 if (cc_pm_is_dev_suspended(dev
))
140 /* read the interrupt status */
141 irr
= cc_ioread(drvdata
, CC_REG(HOST_IRR
));
142 dev_dbg(dev
, "Got IRR=0x%08X\n", irr
);
144 if (irr
== 0) /* Probably shared interrupt line */
147 imr
= cc_ioread(drvdata
, CC_REG(HOST_IMR
));
149 /* clear interrupt - must be before processing events */
150 cc_iowrite(drvdata
, CC_REG(HOST_ICR
), irr
);
153 /* Completion interrupt - most probable */
154 if (irr
& drvdata
->comp_mask
) {
155 /* Mask all completion interrupts - will be unmasked in
156 * deferred service handler
158 cc_iowrite(drvdata
, CC_REG(HOST_IMR
), imr
| drvdata
->comp_mask
);
159 irr
&= ~drvdata
->comp_mask
;
160 complete_request(drvdata
);
162 #ifdef CONFIG_CRYPTO_FIPS
163 /* TEE FIPS interrupt */
164 if (irr
& CC_GPR0_IRQ_MASK
) {
165 /* Mask interrupt - will be unmasked in Deferred service
168 cc_iowrite(drvdata
, CC_REG(HOST_IMR
), imr
| CC_GPR0_IRQ_MASK
);
169 irr
&= ~CC_GPR0_IRQ_MASK
;
170 fips_handler(drvdata
);
173 /* AXI error interrupt */
174 if (irr
& CC_AXI_ERR_IRQ_MASK
) {
177 /* Read the AXI error ID */
178 axi_err
= cc_ioread(drvdata
, CC_REG(AXIM_MON_ERR
));
179 dev_dbg(dev
, "AXI completion error: axim_mon_err=0x%08X\n",
182 irr
&= ~CC_AXI_ERR_IRQ_MASK
;
186 dev_dbg_ratelimited(dev
, "IRR includes unknown cause bits (0x%08X)\n",
194 bool cc_wait_for_reset_completion(struct cc_drvdata
*drvdata
)
199 /* 712/710/63 has no reset completion indication, always return true */
200 if (drvdata
->hw_rev
<= CC_HW_REV_712
)
203 for (i
= 0; i
< CC_HW_RESET_LOOP_COUNT
; i
++) {
204 /* in cc7x3 NVM_IS_IDLE indicates that CC reset is
205 * completed and device is fully functional
207 val
= cc_ioread(drvdata
, CC_REG(NVM_IS_IDLE
));
208 if (val
& CC_NVM_IS_IDLE_MASK
) {
209 /* hw indicate reset completed */
212 /* allow scheduling other process on the processor */
215 /* reset not completed */
219 int init_cc_regs(struct cc_drvdata
*drvdata
, bool is_probe
)
221 unsigned int val
, cache_params
;
222 struct device
*dev
= drvdata_to_dev(drvdata
);
224 /* Unmask all AXI interrupt sources AXI_CFG1 register */
225 /* AXI interrupt config are obsoleted startign at cc7x3 */
226 if (drvdata
->hw_rev
<= CC_HW_REV_712
) {
227 val
= cc_ioread(drvdata
, CC_REG(AXIM_CFG
));
228 cc_iowrite(drvdata
, CC_REG(AXIM_CFG
), val
& ~CC_AXI_IRQ_MASK
);
229 dev_dbg(dev
, "AXIM_CFG=0x%08X\n",
230 cc_ioread(drvdata
, CC_REG(AXIM_CFG
)));
233 /* Clear all pending interrupts */
234 val
= cc_ioread(drvdata
, CC_REG(HOST_IRR
));
235 dev_dbg(dev
, "IRR=0x%08X\n", val
);
236 cc_iowrite(drvdata
, CC_REG(HOST_ICR
), val
);
238 /* Unmask relevant interrupt cause */
239 val
= drvdata
->comp_mask
| CC_AXI_ERR_IRQ_MASK
;
241 if (drvdata
->hw_rev
>= CC_HW_REV_712
)
242 val
|= CC_GPR0_IRQ_MASK
;
244 cc_iowrite(drvdata
, CC_REG(HOST_IMR
), ~val
);
246 cache_params
= (drvdata
->coherent
? CC_COHERENT_CACHE_PARAMS
: 0x0);
248 val
= cc_ioread(drvdata
, CC_REG(AXIM_CACHE_PARAMS
));
251 dev_dbg(dev
, "Cache params previous: 0x%08X\n", val
);
253 cc_iowrite(drvdata
, CC_REG(AXIM_CACHE_PARAMS
), cache_params
);
254 val
= cc_ioread(drvdata
, CC_REG(AXIM_CACHE_PARAMS
));
257 dev_dbg(dev
, "Cache params current: 0x%08X (expect: 0x%08X)\n",
263 static int init_cc_resources(struct platform_device
*plat_dev
)
265 struct resource
*req_mem_cc_regs
= NULL
;
266 struct cc_drvdata
*new_drvdata
;
267 struct device
*dev
= &plat_dev
->dev
;
268 struct device_node
*np
= dev
->of_node
;
269 u32 val
, hw_rev_pidr
, sig_cidr
;
271 const struct cc_hw_data
*hw_rev
;
272 const struct of_device_id
*dev_id
;
277 new_drvdata
= devm_kzalloc(dev
, sizeof(*new_drvdata
), GFP_KERNEL
);
281 dev_id
= of_match_node(arm_ccree_dev_of_match
, np
);
285 hw_rev
= (struct cc_hw_data
*)dev_id
->data
;
286 new_drvdata
->hw_rev_name
= hw_rev
->name
;
287 new_drvdata
->hw_rev
= hw_rev
->rev
;
288 new_drvdata
->std_bodies
= hw_rev
->std_bodies
;
290 if (hw_rev
->rev
>= CC_HW_REV_712
) {
291 new_drvdata
->axim_mon_offset
= CC_REG(AXIM_MON_COMP
);
292 new_drvdata
->sig_offset
= CC_REG(HOST_SIGNATURE_712
);
293 new_drvdata
->ver_offset
= CC_REG(HOST_VERSION_712
);
295 new_drvdata
->axim_mon_offset
= CC_REG(AXIM_MON_COMP8
);
296 new_drvdata
->sig_offset
= CC_REG(HOST_SIGNATURE_630
);
297 new_drvdata
->ver_offset
= CC_REG(HOST_VERSION_630
);
300 new_drvdata
->comp_mask
= CC_COMP_IRQ_MASK
;
302 platform_set_drvdata(plat_dev
, new_drvdata
);
303 new_drvdata
->plat_dev
= plat_dev
;
305 clk
= devm_clk_get(dev
, NULL
);
307 switch (PTR_ERR(clk
)) {
308 /* Clock is optional so this might be fine */
312 /* Clock not available, let's try again soon */
314 return -EPROBE_DEFER
;
317 dev_err(dev
, "Error getting clock: %ld\n",
321 new_drvdata
->clk
= clk
;
323 new_drvdata
->coherent
= of_dma_is_coherent(np
);
325 /* Get device resources */
326 /* First CC registers space */
327 req_mem_cc_regs
= platform_get_resource(plat_dev
, IORESOURCE_MEM
, 0);
328 /* Map registers space */
329 new_drvdata
->cc_base
= devm_ioremap_resource(dev
, req_mem_cc_regs
);
330 if (IS_ERR(new_drvdata
->cc_base
)) {
331 dev_err(dev
, "Failed to ioremap registers");
332 return PTR_ERR(new_drvdata
->cc_base
);
335 dev_dbg(dev
, "Got MEM resource (%s): %pR\n", req_mem_cc_regs
->name
,
337 dev_dbg(dev
, "CC registers mapped from %pa to 0x%p\n",
338 &req_mem_cc_regs
->start
, new_drvdata
->cc_base
);
341 irq
= platform_get_irq(plat_dev
, 0);
345 init_completion(&new_drvdata
->hw_queue_avail
);
347 if (!plat_dev
->dev
.dma_mask
)
348 plat_dev
->dev
.dma_mask
= &plat_dev
->dev
.coherent_dma_mask
;
350 dma_mask
= DMA_BIT_MASK(DMA_BIT_MASK_LEN
);
351 while (dma_mask
> 0x7fffffffUL
) {
352 if (dma_supported(&plat_dev
->dev
, dma_mask
)) {
353 rc
= dma_set_coherent_mask(&plat_dev
->dev
, dma_mask
);
361 dev_err(dev
, "Failed in dma_set_mask, mask=%llx\n", dma_mask
);
365 rc
= cc_clk_on(new_drvdata
);
367 dev_err(dev
, "Failed to enable clock");
371 new_drvdata
->sec_disabled
= cc_sec_disable
;
373 /* wait for Crytpcell reset completion */
374 if (!cc_wait_for_reset_completion(new_drvdata
)) {
375 dev_err(dev
, "Cryptocell reset not completed");
378 if (hw_rev
->rev
<= CC_HW_REV_712
) {
379 /* Verify correct mapping */
380 val
= cc_ioread(new_drvdata
, new_drvdata
->sig_offset
);
381 if (val
!= hw_rev
->sig
) {
382 dev_err(dev
, "Invalid CC signature: SIGNATURE=0x%08X != expected=0x%08X\n",
388 hw_rev_pidr
= cc_ioread(new_drvdata
, new_drvdata
->ver_offset
);
390 /* Verify correct mapping */
391 val
= cc_read_idr(new_drvdata
, pidr_0124_offsets
);
392 if (val
!= hw_rev
->pidr_0124
) {
393 dev_err(dev
, "Invalid CC PIDR: PIDR0124=0x%08X != expected=0x%08X\n",
394 val
, hw_rev
->pidr_0124
);
400 val
= cc_read_idr(new_drvdata
, cidr_0123_offsets
);
401 if (val
!= hw_rev
->cidr_0123
) {
402 dev_err(dev
, "Invalid CC CIDR: CIDR0123=0x%08X != expected=0x%08X\n",
403 val
, hw_rev
->cidr_0123
);
409 /* Check HW engine configuration */
410 val
= cc_ioread(new_drvdata
, CC_REG(HOST_REMOVE_INPUT_PINS
));
416 if (new_drvdata
->std_bodies
& CC_STD_NIST
) {
417 dev_warn(dev
, "703 mode forced due to HW configuration.\n");
418 new_drvdata
->std_bodies
= CC_STD_OSCCA
;
422 dev_err(dev
, "Unsupported engines configuration.\n");
427 /* Check security disable state */
428 val
= cc_ioread(new_drvdata
, CC_REG(SECURITY_DISABLED
));
429 val
&= CC_SECURITY_DISABLED_MASK
;
430 new_drvdata
->sec_disabled
|= !!val
;
432 if (!new_drvdata
->sec_disabled
) {
433 new_drvdata
->comp_mask
|= CC_CPP_SM4_ABORT_MASK
;
434 if (new_drvdata
->std_bodies
& CC_STD_NIST
)
435 new_drvdata
->comp_mask
|= CC_CPP_AES_ABORT_MASK
;
439 if (new_drvdata
->sec_disabled
)
440 dev_info(dev
, "Security Disabled mode is in effect. Security functions disabled.\n");
442 /* Display HW versions */
443 dev_info(dev
, "ARM CryptoCell %s Driver: HW version 0x%08X/0x%8X, Driver version %s\n",
444 hw_rev
->name
, hw_rev_pidr
, sig_cidr
, DRV_MODULE_VERSION
);
445 /* register the driver isr function */
446 rc
= devm_request_irq(dev
, irq
, cc_isr
, IRQF_SHARED
, "ccree",
449 dev_err(dev
, "Could not register to interrupt %d\n", irq
);
452 dev_dbg(dev
, "Registered to IRQ: %d\n", irq
);
454 rc
= init_cc_regs(new_drvdata
, true);
456 dev_err(dev
, "init_cc_regs failed\n");
460 rc
= cc_debugfs_init(new_drvdata
);
462 dev_err(dev
, "Failed registering debugfs interface\n");
466 rc
= cc_fips_init(new_drvdata
);
468 dev_err(dev
, "cc_fips_init failed 0x%x\n", rc
);
469 goto post_debugfs_err
;
471 rc
= cc_sram_mgr_init(new_drvdata
);
473 dev_err(dev
, "cc_sram_mgr_init failed\n");
474 goto post_fips_init_err
;
477 new_drvdata
->mlli_sram_addr
=
478 cc_sram_alloc(new_drvdata
, MAX_MLLI_BUFF_SIZE
);
479 if (new_drvdata
->mlli_sram_addr
== NULL_SRAM_ADDR
) {
480 dev_err(dev
, "Failed to alloc MLLI Sram buffer\n");
482 goto post_sram_mgr_err
;
485 rc
= cc_req_mgr_init(new_drvdata
);
487 dev_err(dev
, "cc_req_mgr_init failed\n");
488 goto post_sram_mgr_err
;
491 rc
= cc_buffer_mgr_init(new_drvdata
);
493 dev_err(dev
, "cc_buffer_mgr_init failed\n");
494 goto post_req_mgr_err
;
497 rc
= cc_pm_init(new_drvdata
);
499 dev_err(dev
, "cc_pm_init failed\n");
500 goto post_buf_mgr_err
;
503 /* Allocate crypto algs */
504 rc
= cc_cipher_alloc(new_drvdata
);
506 dev_err(dev
, "cc_cipher_alloc failed\n");
507 goto post_buf_mgr_err
;
510 /* hash must be allocated before aead since hash exports APIs */
511 rc
= cc_hash_alloc(new_drvdata
);
513 dev_err(dev
, "cc_hash_alloc failed\n");
514 goto post_cipher_err
;
517 rc
= cc_aead_alloc(new_drvdata
);
519 dev_err(dev
, "cc_aead_alloc failed\n");
523 /* All set, we can allow autosuspend */
524 cc_pm_go(new_drvdata
);
526 /* If we got here and FIPS mode is enabled
527 * it means all FIPS test passed, so let TEE
530 cc_set_ree_fips_status(new_drvdata
, true);
535 cc_hash_free(new_drvdata
);
537 cc_cipher_free(new_drvdata
);
539 cc_buffer_mgr_fini(new_drvdata
);
541 cc_req_mgr_fini(new_drvdata
);
543 cc_sram_mgr_fini(new_drvdata
);
545 cc_fips_fini(new_drvdata
);
547 cc_debugfs_fini(new_drvdata
);
549 fini_cc_regs(new_drvdata
);
551 cc_clk_off(new_drvdata
);
555 void fini_cc_regs(struct cc_drvdata
*drvdata
)
557 /* Mask all interrupts */
558 cc_iowrite(drvdata
, CC_REG(HOST_IMR
), 0xFFFFFFFF);
561 static void cleanup_cc_resources(struct platform_device
*plat_dev
)
563 struct cc_drvdata
*drvdata
=
564 (struct cc_drvdata
*)platform_get_drvdata(plat_dev
);
566 cc_aead_free(drvdata
);
567 cc_hash_free(drvdata
);
568 cc_cipher_free(drvdata
);
570 cc_buffer_mgr_fini(drvdata
);
571 cc_req_mgr_fini(drvdata
);
572 cc_sram_mgr_fini(drvdata
);
573 cc_fips_fini(drvdata
);
574 cc_debugfs_fini(drvdata
);
575 fini_cc_regs(drvdata
);
579 int cc_clk_on(struct cc_drvdata
*drvdata
)
581 struct clk
*clk
= drvdata
->clk
;
585 /* Not all devices have a clock associated with CCREE */
588 rc
= clk_prepare_enable(clk
);
595 unsigned int cc_get_default_hash_len(struct cc_drvdata
*drvdata
)
597 if (drvdata
->hw_rev
>= CC_HW_REV_712
)
598 return HASH_LEN_SIZE_712
;
600 return HASH_LEN_SIZE_630
;
603 void cc_clk_off(struct cc_drvdata
*drvdata
)
605 struct clk
*clk
= drvdata
->clk
;
608 /* Not all devices have a clock associated with CCREE */
611 clk_disable_unprepare(clk
);
614 static int ccree_probe(struct platform_device
*plat_dev
)
617 struct device
*dev
= &plat_dev
->dev
;
619 /* Map registers space */
620 rc
= init_cc_resources(plat_dev
);
624 dev_info(dev
, "ARM ccree device initialized\n");
629 static int ccree_remove(struct platform_device
*plat_dev
)
631 struct device
*dev
= &plat_dev
->dev
;
633 dev_dbg(dev
, "Releasing ccree resources...\n");
635 cleanup_cc_resources(plat_dev
);
637 dev_info(dev
, "ARM ccree device terminated\n");
642 static struct platform_driver ccree_driver
= {
645 .of_match_table
= arm_ccree_dev_of_match
,
650 .probe
= ccree_probe
,
651 .remove
= ccree_remove
,
654 static int __init
ccree_init(void)
656 cc_hash_global_init();
657 cc_debugfs_global_init();
659 return platform_driver_register(&ccree_driver
);
661 module_init(ccree_init
);
663 static void __exit
ccree_exit(void)
665 platform_driver_unregister(&ccree_driver
);
666 cc_debugfs_global_fini();
668 module_exit(ccree_exit
);
670 /* Module description */
671 MODULE_DESCRIPTION("ARM TrustZone CryptoCell REE Driver");
672 MODULE_VERSION(DRV_MODULE_VERSION
);
673 MODULE_AUTHOR("ARM");
674 MODULE_LICENSE("GPL v2");