1 // SPDX-License-Identifier: GPL-2.0
2 /* Copyright (C) 2019-2020 ARM Limited or its affiliates. */
4 #include <linux/kernel.h>
5 #include <linux/module.h>
7 #include <linux/hw_random.h>
9 #include <linux/platform_device.h>
10 #include <linux/pm_runtime.h>
11 #include <linux/interrupt.h>
12 #include <linux/irqreturn.h>
13 #include <linux/workqueue.h>
14 #include <linux/circ_buf.h>
15 #include <linux/completion.h>
17 #include <linux/bitfield.h>
18 #include <linux/fips.h>
22 #define CC_REG_LOW(name) (name ## _BIT_SHIFT)
23 #define CC_REG_HIGH(name) (CC_REG_LOW(name) + name ## _BIT_SIZE - 1)
24 #define CC_GENMASK(name) GENMASK(CC_REG_HIGH(name), CC_REG_LOW(name))
26 #define CC_REG_FLD_GET(reg_name, fld_name, reg_val) \
27 (FIELD_GET(CC_GENMASK(CC_ ## reg_name ## _ ## fld_name), reg_val))
29 #define CC_HW_RESET_LOOP_COUNT 10
30 #define CC_TRNG_SUSPEND_TIMEOUT 3000
32 /* data circular buffer in words must be:
33 * - of a power-of-2 size (limitation of circ_buf.h macros)
34 * - at least 6, the size generated in the EHR according to HW implementation
36 #define CCTRNG_DATA_BUF_WORDS 32
38 /* The timeout for the TRNG operation should be calculated with the formula:
39 * Timeout = EHR_NUM * VN_COEFF * EHR_LENGTH * SAMPLE_CNT * SCALE_VALUE
41 * - SAMPLE_CNT is input value from the characterisation process
42 * - all the rest are constants
46 #define EHR_LENGTH CC_TRNG_EHR_IN_BITS
48 #define CCTRNG_TIMEOUT(smpl_cnt) \
49 (EHR_NUM * VN_COEFF * EHR_LENGTH * smpl_cnt * SCALE_VALUE)
51 struct cctrng_drvdata
{
52 struct platform_device
*pdev
;
53 void __iomem
*cc_base
;
57 /* Sampling interval for each ring oscillator:
58 * count of ring oscillator cycles between consecutive bits sampling.
59 * Value of 0 indicates non-valid rosc
61 u32 smpl_ratio
[CC_TRNG_NUM_OF_ROSCS
];
63 u32 data_buf
[CCTRNG_DATA_BUF_WORDS
];
65 struct work_struct compwork
;
66 struct work_struct startwork
;
68 /* pending_hw - 1 when HW is pending, 0 when it is idle */
71 /* protects against multiple concurrent consumers of data_buf */
76 /* functions for write/read CC registers */
77 static inline void cc_iowrite(struct cctrng_drvdata
*drvdata
, u32 reg
, u32 val
)
79 iowrite32(val
, (drvdata
->cc_base
+ reg
));
81 static inline u32
cc_ioread(struct cctrng_drvdata
*drvdata
, u32 reg
)
83 return ioread32(drvdata
->cc_base
+ reg
);
87 static int cc_trng_pm_get(struct device
*dev
)
91 rc
= pm_runtime_get_sync(dev
);
93 /* pm_runtime_get_sync() can return 1 as a valid return code */
94 return (rc
== 1 ? 0 : rc
);
97 static void cc_trng_pm_put_suspend(struct device
*dev
)
101 pm_runtime_mark_last_busy(dev
);
102 rc
= pm_runtime_put_autosuspend(dev
);
104 dev_err(dev
, "pm_runtime_put_autosuspend returned %x\n", rc
);
107 static int cc_trng_pm_init(struct cctrng_drvdata
*drvdata
)
109 struct device
*dev
= &(drvdata
->pdev
->dev
);
111 /* must be before the enabling to avoid redundant suspending */
112 pm_runtime_set_autosuspend_delay(dev
, CC_TRNG_SUSPEND_TIMEOUT
);
113 pm_runtime_use_autosuspend(dev
);
114 /* set us as active - note we won't do PM ops until cc_trng_pm_go()! */
115 return pm_runtime_set_active(dev
);
118 static void cc_trng_pm_go(struct cctrng_drvdata
*drvdata
)
120 struct device
*dev
= &(drvdata
->pdev
->dev
);
122 /* enable the PM module*/
123 pm_runtime_enable(dev
);
126 static void cc_trng_pm_fini(struct cctrng_drvdata
*drvdata
)
128 struct device
*dev
= &(drvdata
->pdev
->dev
);
130 pm_runtime_disable(dev
);
134 static inline int cc_trng_parse_sampling_ratio(struct cctrng_drvdata
*drvdata
)
136 struct device
*dev
= &(drvdata
->pdev
->dev
);
137 struct device_node
*np
= drvdata
->pdev
->dev
.of_node
;
140 /* ret will be set to 0 if at least one rosc has (sampling ratio > 0) */
143 rc
= of_property_read_u32_array(np
, "arm,rosc-ratio",
145 CC_TRNG_NUM_OF_ROSCS
);
147 /* arm,rosc-ratio was not found in device tree */
151 /* verify that at least one rosc has (sampling ratio > 0) */
152 for (i
= 0; i
< CC_TRNG_NUM_OF_ROSCS
; ++i
) {
153 dev_dbg(dev
, "rosc %d sampling ratio %u",
154 i
, drvdata
->smpl_ratio
[i
]);
156 if (drvdata
->smpl_ratio
[i
] > 0)
163 static int cc_trng_change_rosc(struct cctrng_drvdata
*drvdata
)
165 struct device
*dev
= &(drvdata
->pdev
->dev
);
167 dev_dbg(dev
, "cctrng change rosc (was %d)\n", drvdata
->active_rosc
);
168 drvdata
->active_rosc
+= 1;
170 while (drvdata
->active_rosc
< CC_TRNG_NUM_OF_ROSCS
) {
171 if (drvdata
->smpl_ratio
[drvdata
->active_rosc
] > 0)
174 drvdata
->active_rosc
+= 1;
180 static void cc_trng_enable_rnd_source(struct cctrng_drvdata
*drvdata
)
184 /* Set watchdog threshold to maximal allowed time (in CPU cycles) */
185 max_cycles
= CCTRNG_TIMEOUT(drvdata
->smpl_ratio
[drvdata
->active_rosc
]);
186 cc_iowrite(drvdata
, CC_RNG_WATCHDOG_VAL_REG_OFFSET
, max_cycles
);
188 /* enable the RND source */
189 cc_iowrite(drvdata
, CC_RND_SOURCE_ENABLE_REG_OFFSET
, 0x1);
191 /* unmask RNG interrupts */
192 cc_iowrite(drvdata
, CC_RNG_IMR_REG_OFFSET
, (u32
)~CC_RNG_INT_MASK
);
196 /* increase circular data buffer index (head/tail) */
197 static inline void circ_idx_inc(int *idx
, int bytes
)
199 *idx
+= (bytes
+ 3) >> 2;
200 *idx
&= (CCTRNG_DATA_BUF_WORDS
- 1);
203 static inline size_t circ_buf_space(struct cctrng_drvdata
*drvdata
)
205 return CIRC_SPACE(drvdata
->circ
.head
,
206 drvdata
->circ
.tail
, CCTRNG_DATA_BUF_WORDS
);
210 static int cctrng_read(struct hwrng
*rng
, void *data
, size_t max
, bool wait
)
212 /* current implementation ignores "wait" */
214 struct cctrng_drvdata
*drvdata
= (struct cctrng_drvdata
*)rng
->priv
;
215 struct device
*dev
= &(drvdata
->pdev
->dev
);
216 u32
*buf
= (u32
*)drvdata
->circ
.buf
;
222 if (!spin_trylock(&drvdata
->read_lock
)) {
223 /* concurrent consumers from data_buf cannot be served */
224 dev_dbg_ratelimited(dev
, "unable to hold lock\n");
228 /* copy till end of data buffer (without wrap back) */
229 cnt_w
= CIRC_CNT_TO_END(drvdata
->circ
.head
,
230 drvdata
->circ
.tail
, CCTRNG_DATA_BUF_WORDS
);
231 size
= min((cnt_w
<<2), max
);
232 memcpy(data
, &(buf
[drvdata
->circ
.tail
]), size
);
234 circ_idx_inc(&drvdata
->circ
.tail
, size
);
235 /* copy rest of data in data buffer */
238 cnt_w
= CIRC_CNT(drvdata
->circ
.head
,
239 drvdata
->circ
.tail
, CCTRNG_DATA_BUF_WORDS
);
240 size
= min((cnt_w
<<2), left
);
241 memcpy(data
, &(buf
[drvdata
->circ
.tail
]), size
);
243 circ_idx_inc(&drvdata
->circ
.tail
, size
);
246 spin_unlock(&drvdata
->read_lock
);
248 if (circ_buf_space(drvdata
) >= CC_TRNG_EHR_IN_WORDS
) {
249 if (atomic_cmpxchg(&drvdata
->pending_hw
, 0, 1) == 0) {
250 /* re-check space in buffer to avoid potential race */
251 if (circ_buf_space(drvdata
) >= CC_TRNG_EHR_IN_WORDS
) {
252 /* increment device's usage counter */
253 int rc
= cc_trng_pm_get(dev
);
257 "cc_trng_pm_get returned %x\n",
262 /* schedule execution of deferred work handler
263 * for filling of data buffer
265 schedule_work(&drvdata
->startwork
);
267 atomic_set(&drvdata
->pending_hw
, 0);
275 static void cc_trng_hw_trigger(struct cctrng_drvdata
*drvdata
)
277 u32 tmp_smpl_cnt
= 0;
278 struct device
*dev
= &(drvdata
->pdev
->dev
);
280 dev_dbg(dev
, "cctrng hw trigger.\n");
282 /* enable the HW RND clock */
283 cc_iowrite(drvdata
, CC_RNG_CLK_ENABLE_REG_OFFSET
, 0x1);
285 /* do software reset */
286 cc_iowrite(drvdata
, CC_RNG_SW_RESET_REG_OFFSET
, 0x1);
287 /* in order to verify that the reset has completed,
288 * the sample count need to be verified
291 /* enable the HW RND clock */
292 cc_iowrite(drvdata
, CC_RNG_CLK_ENABLE_REG_OFFSET
, 0x1);
294 /* set sampling ratio (rng_clocks) between consecutive bits */
295 cc_iowrite(drvdata
, CC_SAMPLE_CNT1_REG_OFFSET
,
296 drvdata
->smpl_ratio
[drvdata
->active_rosc
]);
298 /* read the sampling ratio */
299 tmp_smpl_cnt
= cc_ioread(drvdata
, CC_SAMPLE_CNT1_REG_OFFSET
);
301 } while (tmp_smpl_cnt
!= drvdata
->smpl_ratio
[drvdata
->active_rosc
]);
303 /* disable the RND source for setting new parameters in HW */
304 cc_iowrite(drvdata
, CC_RND_SOURCE_ENABLE_REG_OFFSET
, 0);
306 cc_iowrite(drvdata
, CC_RNG_ICR_REG_OFFSET
, 0xFFFFFFFF);
308 cc_iowrite(drvdata
, CC_TRNG_CONFIG_REG_OFFSET
, drvdata
->active_rosc
);
310 /* Debug Control register: set to 0 - no bypasses */
311 cc_iowrite(drvdata
, CC_TRNG_DEBUG_CONTROL_REG_OFFSET
, 0);
313 cc_trng_enable_rnd_source(drvdata
);
316 static void cc_trng_compwork_handler(struct work_struct
*w
)
320 struct cctrng_drvdata
*drvdata
=
321 container_of(w
, struct cctrng_drvdata
, compwork
);
322 struct device
*dev
= &(drvdata
->pdev
->dev
);
325 /* stop DMA and the RNG source */
326 cc_iowrite(drvdata
, CC_RNG_DMA_ENABLE_REG_OFFSET
, 0);
327 cc_iowrite(drvdata
, CC_RND_SOURCE_ENABLE_REG_OFFSET
, 0);
329 /* read RNG_ISR and check for errors */
330 isr
= cc_ioread(drvdata
, CC_RNG_ISR_REG_OFFSET
);
331 ehr_valid
= CC_REG_FLD_GET(RNG_ISR
, EHR_VALID
, isr
);
332 dev_dbg(dev
, "Got RNG_ISR=0x%08X (EHR_VALID=%u)\n", isr
, ehr_valid
);
334 if (fips_enabled
&& CC_REG_FLD_GET(RNG_ISR
, CRNGT_ERR
, isr
)) {
336 /* FIPS error is fatal */
337 panic("Got HW CRNGT error while fips is enabled!\n");
340 /* Clear all pending RNG interrupts */
341 cc_iowrite(drvdata
, CC_RNG_ICR_REG_OFFSET
, isr
);
345 /* in case of AUTOCORR/TIMEOUT error, try the next ROSC */
346 if (CC_REG_FLD_GET(RNG_ISR
, AUTOCORR_ERR
, isr
) ||
347 CC_REG_FLD_GET(RNG_ISR
, WATCHDOG
, isr
)) {
348 dev_dbg(dev
, "cctrng autocorr/timeout error.\n");
352 /* in case of VN error, ignore it */
355 /* read EHR data from registers */
356 for (i
= 0; i
< CC_TRNG_EHR_IN_WORDS
; i
++) {
357 /* calc word ptr in data_buf */
358 u32
*buf
= (u32
*)drvdata
->circ
.buf
;
360 buf
[drvdata
->circ
.head
] = cc_ioread(drvdata
,
361 CC_EHR_DATA_0_REG_OFFSET
+ (i
*sizeof(u32
)));
363 /* EHR_DATA registers are cleared on read. In case 0 value was
364 * returned, restart the entropy collection.
366 if (buf
[drvdata
->circ
.head
] == 0) {
367 dev_dbg(dev
, "Got 0 value in EHR. active_rosc %u\n",
368 drvdata
->active_rosc
);
372 circ_idx_inc(&drvdata
->circ
.head
, 1<<2);
375 atomic_set(&drvdata
->pending_hw
, 0);
377 /* continue to fill data buffer if needed */
378 if (circ_buf_space(drvdata
) >= CC_TRNG_EHR_IN_WORDS
) {
379 if (atomic_cmpxchg(&drvdata
->pending_hw
, 0, 1) == 0) {
380 /* Re-enable rnd source */
381 cc_trng_enable_rnd_source(drvdata
);
386 cc_trng_pm_put_suspend(dev
);
388 dev_dbg(dev
, "compwork handler done\n");
392 if ((circ_buf_space(drvdata
) >= CC_TRNG_EHR_IN_WORDS
) &&
393 (cc_trng_change_rosc(drvdata
) == 0)) {
394 /* trigger trng hw with next rosc */
395 cc_trng_hw_trigger(drvdata
);
397 atomic_set(&drvdata
->pending_hw
, 0);
398 cc_trng_pm_put_suspend(dev
);
402 static irqreturn_t
cc_isr(int irq
, void *dev_id
)
404 struct cctrng_drvdata
*drvdata
= (struct cctrng_drvdata
*)dev_id
;
405 struct device
*dev
= &(drvdata
->pdev
->dev
);
408 /* if driver suspended return, probably shared interrupt */
409 if (pm_runtime_suspended(dev
))
412 /* read the interrupt status */
413 irr
= cc_ioread(drvdata
, CC_HOST_RGF_IRR_REG_OFFSET
);
414 dev_dbg(dev
, "Got IRR=0x%08X\n", irr
);
416 if (irr
== 0) /* Probably shared interrupt line */
419 /* clear interrupt - must be before processing events */
420 cc_iowrite(drvdata
, CC_HOST_RGF_ICR_REG_OFFSET
, irr
);
422 /* RNG interrupt - most probable */
423 if (irr
& CC_HOST_RNG_IRQ_MASK
) {
424 /* Mask RNG interrupts - will be unmasked in deferred work */
425 cc_iowrite(drvdata
, CC_RNG_IMR_REG_OFFSET
, 0xFFFFFFFF);
427 /* We clear RNG interrupt here,
428 * to avoid it from firing as we'll unmask RNG interrupts.
430 cc_iowrite(drvdata
, CC_HOST_RGF_ICR_REG_OFFSET
,
431 CC_HOST_RNG_IRQ_MASK
);
433 irr
&= ~CC_HOST_RNG_IRQ_MASK
;
435 /* schedule execution of deferred work handler */
436 schedule_work(&drvdata
->compwork
);
440 dev_dbg_ratelimited(dev
,
441 "IRR includes unknown cause bits (0x%08X)\n",
449 static void cc_trng_startwork_handler(struct work_struct
*w
)
451 struct cctrng_drvdata
*drvdata
=
452 container_of(w
, struct cctrng_drvdata
, startwork
);
454 drvdata
->active_rosc
= 0;
455 cc_trng_hw_trigger(drvdata
);
458 static int cctrng_probe(struct platform_device
*pdev
)
460 struct cctrng_drvdata
*drvdata
;
461 struct device
*dev
= &pdev
->dev
;
466 /* Compile time assertion checks */
467 BUILD_BUG_ON(CCTRNG_DATA_BUF_WORDS
< 6);
468 BUILD_BUG_ON((CCTRNG_DATA_BUF_WORDS
& (CCTRNG_DATA_BUF_WORDS
-1)) != 0);
470 drvdata
= devm_kzalloc(dev
, sizeof(*drvdata
), GFP_KERNEL
);
474 drvdata
->rng
.name
= devm_kstrdup(dev
, dev_name(dev
), GFP_KERNEL
);
475 if (!drvdata
->rng
.name
)
478 drvdata
->rng
.read
= cctrng_read
;
479 drvdata
->rng
.priv
= (unsigned long)drvdata
;
480 drvdata
->rng
.quality
= CC_TRNG_QUALITY
;
482 platform_set_drvdata(pdev
, drvdata
);
483 drvdata
->pdev
= pdev
;
485 drvdata
->circ
.buf
= (char *)drvdata
->data_buf
;
487 drvdata
->cc_base
= devm_platform_ioremap_resource(pdev
, 0);
488 if (IS_ERR(drvdata
->cc_base
))
489 return dev_err_probe(dev
, PTR_ERR(drvdata
->cc_base
), "Failed to ioremap registers");
492 irq
= platform_get_irq(pdev
, 0);
496 /* parse sampling rate from device tree */
497 rc
= cc_trng_parse_sampling_ratio(drvdata
);
499 return dev_err_probe(dev
, rc
, "Failed to get legal sampling ratio for rosc\n");
501 drvdata
->clk
= devm_clk_get_optional_enabled(dev
, NULL
);
502 if (IS_ERR(drvdata
->clk
))
503 return dev_err_probe(dev
, PTR_ERR(drvdata
->clk
),
504 "Failed to get or enable the clock\n");
506 INIT_WORK(&drvdata
->compwork
, cc_trng_compwork_handler
);
507 INIT_WORK(&drvdata
->startwork
, cc_trng_startwork_handler
);
508 spin_lock_init(&drvdata
->read_lock
);
510 /* register the driver isr function */
511 rc
= devm_request_irq(dev
, irq
, cc_isr
, IRQF_SHARED
, "cctrng", drvdata
);
513 return dev_err_probe(dev
, rc
, "Could not register to interrupt %d\n", irq
);
514 dev_dbg(dev
, "Registered to IRQ: %d\n", irq
);
516 /* Clear all pending interrupts */
517 val
= cc_ioread(drvdata
, CC_HOST_RGF_IRR_REG_OFFSET
);
518 dev_dbg(dev
, "IRR=0x%08X\n", val
);
519 cc_iowrite(drvdata
, CC_HOST_RGF_ICR_REG_OFFSET
, val
);
521 /* unmask HOST RNG interrupt */
522 cc_iowrite(drvdata
, CC_HOST_RGF_IMR_REG_OFFSET
,
523 cc_ioread(drvdata
, CC_HOST_RGF_IMR_REG_OFFSET
) &
524 ~CC_HOST_RNG_IRQ_MASK
);
527 rc
= cc_trng_pm_init(drvdata
);
529 return dev_err_probe(dev
, rc
, "cc_trng_pm_init failed\n");
531 /* increment device's usage counter */
532 rc
= cc_trng_pm_get(dev
);
534 return dev_err_probe(dev
, rc
, "cc_trng_pm_get returned %x\n", rc
);
536 /* set pending_hw to verify that HW won't be triggered from read */
537 atomic_set(&drvdata
->pending_hw
, 1);
539 /* registration of the hwrng device */
540 rc
= devm_hwrng_register(dev
, &drvdata
->rng
);
542 dev_err(dev
, "Could not register hwrng device.\n");
546 /* trigger HW to start generate data */
547 drvdata
->active_rosc
= 0;
548 cc_trng_hw_trigger(drvdata
);
550 /* All set, we can allow auto-suspend */
551 cc_trng_pm_go(drvdata
);
553 dev_info(dev
, "ARM cctrng device initialized\n");
558 cc_trng_pm_fini(drvdata
);
563 static void cctrng_remove(struct platform_device
*pdev
)
565 struct cctrng_drvdata
*drvdata
= platform_get_drvdata(pdev
);
566 struct device
*dev
= &pdev
->dev
;
568 dev_dbg(dev
, "Releasing cctrng resources...\n");
570 cc_trng_pm_fini(drvdata
);
572 dev_info(dev
, "ARM cctrng device terminated\n");
575 static int __maybe_unused
cctrng_suspend(struct device
*dev
)
577 struct cctrng_drvdata
*drvdata
= dev_get_drvdata(dev
);
579 dev_dbg(dev
, "set HOST_POWER_DOWN_EN\n");
580 cc_iowrite(drvdata
, CC_HOST_POWER_DOWN_EN_REG_OFFSET
,
583 clk_disable_unprepare(drvdata
->clk
);
588 static bool cctrng_wait_for_reset_completion(struct cctrng_drvdata
*drvdata
)
593 for (i
= 0; i
< CC_HW_RESET_LOOP_COUNT
; i
++) {
594 /* in cc7x3 NVM_IS_IDLE indicates that CC reset is
595 * completed and device is fully functional
597 val
= cc_ioread(drvdata
, CC_NVM_IS_IDLE_REG_OFFSET
);
598 if (val
& BIT(CC_NVM_IS_IDLE_VALUE_BIT_SHIFT
)) {
599 /* hw indicate reset completed */
602 /* allow scheduling other process on the processor */
605 /* reset not completed */
609 static int __maybe_unused
cctrng_resume(struct device
*dev
)
611 struct cctrng_drvdata
*drvdata
= dev_get_drvdata(dev
);
614 dev_dbg(dev
, "unset HOST_POWER_DOWN_EN\n");
615 /* Enables the device source clk */
616 rc
= clk_prepare_enable(drvdata
->clk
);
618 dev_err(dev
, "failed getting clock back on. We're toast.\n");
622 /* wait for Cryptocell reset completion */
623 if (!cctrng_wait_for_reset_completion(drvdata
)) {
624 dev_err(dev
, "Cryptocell reset not completed");
625 clk_disable_unprepare(drvdata
->clk
);
629 /* unmask HOST RNG interrupt */
630 cc_iowrite(drvdata
, CC_HOST_RGF_IMR_REG_OFFSET
,
631 cc_ioread(drvdata
, CC_HOST_RGF_IMR_REG_OFFSET
) &
632 ~CC_HOST_RNG_IRQ_MASK
);
634 cc_iowrite(drvdata
, CC_HOST_POWER_DOWN_EN_REG_OFFSET
,
640 static UNIVERSAL_DEV_PM_OPS(cctrng_pm
, cctrng_suspend
, cctrng_resume
, NULL
);
642 static const struct of_device_id arm_cctrng_dt_match
[] = {
643 { .compatible
= "arm,cryptocell-713-trng", },
644 { .compatible
= "arm,cryptocell-703-trng", },
647 MODULE_DEVICE_TABLE(of
, arm_cctrng_dt_match
);
649 static struct platform_driver cctrng_driver
= {
652 .of_match_table
= arm_cctrng_dt_match
,
655 .probe
= cctrng_probe
,
656 .remove
= cctrng_remove
,
659 module_platform_driver(cctrng_driver
);
661 /* Module description */
662 MODULE_DESCRIPTION("ARM CryptoCell TRNG Driver");
663 MODULE_AUTHOR("ARM");
664 MODULE_LICENSE("GPL v2");