1 // SPDX-License-Identifier: GPL-2.0-only
3 * Driver for Broadcom BRCMSTB, NSP, NS2, Cygnus SPI Controllers
5 * Copyright 2016 Broadcom
9 #include <linux/delay.h>
10 #include <linux/device.h>
11 #include <linux/init.h>
12 #include <linux/interrupt.h>
14 #include <linux/ioport.h>
15 #include <linux/kernel.h>
16 #include <linux/module.h>
18 #include <linux/of_irq.h>
19 #include <linux/platform_device.h>
20 #include <linux/slab.h>
21 #include <linux/spi/spi.h>
22 #include <linux/spi/spi-mem.h>
23 #include <linux/sysfs.h>
24 #include <linux/types.h>
25 #include "spi-bcm-qspi.h"
27 #define DRIVER_NAME "bcm_qspi"
30 /* BSPI register offsets */
31 #define BSPI_REVISION_ID 0x000
32 #define BSPI_SCRATCH 0x004
33 #define BSPI_MAST_N_BOOT_CTRL 0x008
34 #define BSPI_BUSY_STATUS 0x00c
35 #define BSPI_INTR_STATUS 0x010
36 #define BSPI_B0_STATUS 0x014
37 #define BSPI_B0_CTRL 0x018
38 #define BSPI_B1_STATUS 0x01c
39 #define BSPI_B1_CTRL 0x020
40 #define BSPI_STRAP_OVERRIDE_CTRL 0x024
41 #define BSPI_FLEX_MODE_ENABLE 0x028
42 #define BSPI_BITS_PER_CYCLE 0x02c
43 #define BSPI_BITS_PER_PHASE 0x030
44 #define BSPI_CMD_AND_MODE_BYTE 0x034
45 #define BSPI_BSPI_FLASH_UPPER_ADDR_BYTE 0x038
46 #define BSPI_BSPI_XOR_VALUE 0x03c
47 #define BSPI_BSPI_XOR_ENABLE 0x040
48 #define BSPI_BSPI_PIO_MODE_ENABLE 0x044
49 #define BSPI_BSPI_PIO_IODIR 0x048
50 #define BSPI_BSPI_PIO_DATA 0x04c
52 /* RAF register offsets */
53 #define BSPI_RAF_START_ADDR 0x100
54 #define BSPI_RAF_NUM_WORDS 0x104
55 #define BSPI_RAF_CTRL 0x108
56 #define BSPI_RAF_FULLNESS 0x10c
57 #define BSPI_RAF_WATERMARK 0x110
58 #define BSPI_RAF_STATUS 0x114
59 #define BSPI_RAF_READ_DATA 0x118
60 #define BSPI_RAF_WORD_CNT 0x11c
61 #define BSPI_RAF_CURR_ADDR 0x120
63 /* Override mode masks */
64 #define BSPI_STRAP_OVERRIDE_CTRL_OVERRIDE BIT(0)
65 #define BSPI_STRAP_OVERRIDE_CTRL_DATA_DUAL BIT(1)
66 #define BSPI_STRAP_OVERRIDE_CTRL_ADDR_4BYTE BIT(2)
67 #define BSPI_STRAP_OVERRIDE_CTRL_DATA_QUAD BIT(3)
68 #define BSPI_STRAP_OVERRIDE_CTRL_ENDAIN_MODE BIT(4)
70 #define BSPI_ADDRLEN_3BYTES 3
71 #define BSPI_ADDRLEN_4BYTES 4
73 #define BSPI_RAF_STATUS_FIFO_EMPTY_MASK BIT(1)
75 #define BSPI_RAF_CTRL_START_MASK BIT(0)
76 #define BSPI_RAF_CTRL_CLEAR_MASK BIT(1)
78 #define BSPI_BPP_MODE_SELECT_MASK BIT(8)
79 #define BSPI_BPP_ADDR_SELECT_MASK BIT(16)
81 #define BSPI_READ_LENGTH 256
83 /* MSPI register offsets */
84 #define MSPI_SPCR0_LSB 0x000
85 #define MSPI_SPCR0_MSB 0x004
86 #define MSPI_SPCR1_LSB 0x008
87 #define MSPI_SPCR1_MSB 0x00c
88 #define MSPI_NEWQP 0x010
89 #define MSPI_ENDQP 0x014
90 #define MSPI_SPCR2 0x018
91 #define MSPI_MSPI_STATUS 0x020
92 #define MSPI_CPTQP 0x024
93 #define MSPI_SPCR3 0x028
94 #define MSPI_REV 0x02c
95 #define MSPI_TXRAM 0x040
96 #define MSPI_RXRAM 0x0c0
97 #define MSPI_CDRAM 0x140
98 #define MSPI_WRITE_LOCK 0x180
100 #define MSPI_MASTER_BIT BIT(7)
102 #define MSPI_NUM_CDRAM 16
103 #define MSPI_CDRAM_CONT_BIT BIT(7)
104 #define MSPI_CDRAM_BITSE_BIT BIT(6)
105 #define MSPI_CDRAM_PCS 0xf
107 #define MSPI_SPCR2_SPE BIT(6)
108 #define MSPI_SPCR2_CONT_AFTER_CMD BIT(7)
110 #define MSPI_SPCR3_FASTBR BIT(0)
111 #define MSPI_SPCR3_FASTDT BIT(1)
112 #define MSPI_SPCR3_SYSCLKSEL_MASK GENMASK(11, 10)
113 #define MSPI_SPCR3_SYSCLKSEL_27 (MSPI_SPCR3_SYSCLKSEL_MASK & \
114 ~(BIT(10) | BIT(11)))
115 #define MSPI_SPCR3_SYSCLKSEL_108 (MSPI_SPCR3_SYSCLKSEL_MASK & \
118 #define MSPI_MSPI_STATUS_SPIF BIT(0)
120 #define INTR_BASE_BIT_SHIFT 0x02
121 #define INTR_COUNT 0x07
123 #define NUM_CHIPSELECT 4
124 #define QSPI_SPBR_MAX 255U
125 #define MSPI_BASE_FREQ 27000000UL
127 #define OPCODE_DIOR 0xBB
128 #define OPCODE_QIOR 0xEB
129 #define OPCODE_DIOR_4B 0xBC
130 #define OPCODE_QIOR_4B 0xEC
132 #define MAX_CMD_SIZE 6
134 #define ADDR_4MB_MASK GENMASK(22, 0)
136 /* stop at end of transfer, no other reason */
137 #define TRANS_STATUS_BREAK_NONE 0
138 /* stop at end of spi_message */
139 #define TRANS_STATUS_BREAK_EOM 1
140 /* stop at end of spi_transfer if delay */
141 #define TRANS_STATUS_BREAK_DELAY 2
142 /* stop at end of spi_transfer if cs_change */
143 #define TRANS_STATUS_BREAK_CS_CHANGE 4
144 /* stop if we run out of bytes */
145 #define TRANS_STATUS_BREAK_NO_BYTES 8
147 /* events that make us stop filling TX slots */
148 #define TRANS_STATUS_BREAK_TX (TRANS_STATUS_BREAK_EOM | \
149 TRANS_STATUS_BREAK_DELAY | \
150 TRANS_STATUS_BREAK_CS_CHANGE)
152 /* events that make us deassert CS */
153 #define TRANS_STATUS_BREAK_DESELECT (TRANS_STATUS_BREAK_EOM | \
154 TRANS_STATUS_BREAK_CS_CHANGE)
156 struct bcm_qspi_parms
{
162 struct bcm_xfer_mode
{
165 unsigned int addrlen
;
181 struct bcm_qspi_irq
{
182 const char *irq_name
;
183 const irq_handler_t irq_handler
;
188 struct bcm_qspi_dev_id
{
189 const struct bcm_qspi_irq
*irqp
;
195 struct spi_transfer
*trans
;
197 bool mspi_last_trans
;
201 struct platform_device
*pdev
;
202 struct spi_master
*master
;
206 void __iomem
*base
[BASEMAX
];
208 /* Some SoCs provide custom interrupt status register(s) */
209 struct bcm_qspi_soc_intc
*soc_intc
;
211 struct bcm_qspi_parms last_parms
;
212 struct qspi_trans trans_pos
;
217 const struct spi_mem_op
*bspi_rf_op
;
220 u32 bspi_rf_op_status
;
221 struct bcm_xfer_mode xfer_mode
;
222 u32 s3_strap_override_ctrl
;
226 struct bcm_qspi_dev_id
*dev_ids
;
227 struct completion mspi_done
;
228 struct completion bspi_done
;
231 bool mspi_spcr3_sysclk
;
234 static inline bool has_bspi(struct bcm_qspi
*qspi
)
236 return qspi
->bspi_mode
;
239 /* hardware supports spcr3 and fast baud-rate */
240 static inline bool bcm_qspi_has_fastbr(struct bcm_qspi
*qspi
)
242 if (!has_bspi(qspi
) &&
243 ((qspi
->mspi_maj_rev
>= 1) &&
244 (qspi
->mspi_min_rev
>= 5)))
250 /* hardware supports sys clk 108Mhz */
251 static inline bool bcm_qspi_has_sysclk_108(struct bcm_qspi
*qspi
)
253 if (!has_bspi(qspi
) && (qspi
->mspi_spcr3_sysclk
||
254 ((qspi
->mspi_maj_rev
>= 1) &&
255 (qspi
->mspi_min_rev
>= 6))))
261 static inline int bcm_qspi_spbr_min(struct bcm_qspi
*qspi
)
263 if (bcm_qspi_has_fastbr(qspi
))
269 /* Read qspi controller register*/
270 static inline u32
bcm_qspi_read(struct bcm_qspi
*qspi
, enum base_type type
,
273 return bcm_qspi_readl(qspi
->big_endian
, qspi
->base
[type
] + offset
);
276 /* Write qspi controller register*/
277 static inline void bcm_qspi_write(struct bcm_qspi
*qspi
, enum base_type type
,
278 unsigned int offset
, unsigned int data
)
280 bcm_qspi_writel(qspi
->big_endian
, data
, qspi
->base
[type
] + offset
);
284 static int bcm_qspi_bspi_busy_poll(struct bcm_qspi
*qspi
)
288 /* this should normally finish within 10us */
289 for (i
= 0; i
< 1000; i
++) {
290 if (!(bcm_qspi_read(qspi
, BSPI
, BSPI_BUSY_STATUS
) & 1))
294 dev_warn(&qspi
->pdev
->dev
, "timeout waiting for !busy_status\n");
298 static inline bool bcm_qspi_bspi_ver_three(struct bcm_qspi
*qspi
)
300 if (qspi
->bspi_maj_rev
< 4)
305 static void bcm_qspi_bspi_flush_prefetch_buffers(struct bcm_qspi
*qspi
)
307 bcm_qspi_bspi_busy_poll(qspi
);
308 /* Force rising edge for the b0/b1 'flush' field */
309 bcm_qspi_write(qspi
, BSPI
, BSPI_B0_CTRL
, 1);
310 bcm_qspi_write(qspi
, BSPI
, BSPI_B1_CTRL
, 1);
311 bcm_qspi_write(qspi
, BSPI
, BSPI_B0_CTRL
, 0);
312 bcm_qspi_write(qspi
, BSPI
, BSPI_B1_CTRL
, 0);
315 static int bcm_qspi_bspi_lr_is_fifo_empty(struct bcm_qspi
*qspi
)
317 return (bcm_qspi_read(qspi
, BSPI
, BSPI_RAF_STATUS
) &
318 BSPI_RAF_STATUS_FIFO_EMPTY_MASK
);
321 static inline u32
bcm_qspi_bspi_lr_read_fifo(struct bcm_qspi
*qspi
)
323 u32 data
= bcm_qspi_read(qspi
, BSPI
, BSPI_RAF_READ_DATA
);
325 /* BSPI v3 LR is LE only, convert data to host endianness */
326 if (bcm_qspi_bspi_ver_three(qspi
))
327 data
= le32_to_cpu(data
);
332 static inline void bcm_qspi_bspi_lr_start(struct bcm_qspi
*qspi
)
334 bcm_qspi_bspi_busy_poll(qspi
);
335 bcm_qspi_write(qspi
, BSPI
, BSPI_RAF_CTRL
,
336 BSPI_RAF_CTRL_START_MASK
);
339 static inline void bcm_qspi_bspi_lr_clear(struct bcm_qspi
*qspi
)
341 bcm_qspi_write(qspi
, BSPI
, BSPI_RAF_CTRL
,
342 BSPI_RAF_CTRL_CLEAR_MASK
);
343 bcm_qspi_bspi_flush_prefetch_buffers(qspi
);
346 static void bcm_qspi_bspi_lr_data_read(struct bcm_qspi
*qspi
)
348 u32
*buf
= (u32
*)qspi
->bspi_rf_op
->data
.buf
.in
;
351 dev_dbg(&qspi
->pdev
->dev
, "xfer %p rx %p rxlen %d\n", qspi
->bspi_rf_op
,
352 qspi
->bspi_rf_op
->data
.buf
.in
, qspi
->bspi_rf_op_len
);
353 while (!bcm_qspi_bspi_lr_is_fifo_empty(qspi
)) {
354 data
= bcm_qspi_bspi_lr_read_fifo(qspi
);
355 if (likely(qspi
->bspi_rf_op_len
>= 4) &&
356 IS_ALIGNED((uintptr_t)buf
, 4)) {
357 buf
[qspi
->bspi_rf_op_idx
++] = data
;
358 qspi
->bspi_rf_op_len
-= 4;
360 /* Read out remaining bytes, make sure*/
361 u8
*cbuf
= (u8
*)&buf
[qspi
->bspi_rf_op_idx
];
363 data
= cpu_to_le32(data
);
364 while (qspi
->bspi_rf_op_len
) {
367 qspi
->bspi_rf_op_len
--;
373 static void bcm_qspi_bspi_set_xfer_params(struct bcm_qspi
*qspi
, u8 cmd_byte
,
374 int bpp
, int bpc
, int flex_mode
)
376 bcm_qspi_write(qspi
, BSPI
, BSPI_FLEX_MODE_ENABLE
, 0);
377 bcm_qspi_write(qspi
, BSPI
, BSPI_BITS_PER_CYCLE
, bpc
);
378 bcm_qspi_write(qspi
, BSPI
, BSPI_BITS_PER_PHASE
, bpp
);
379 bcm_qspi_write(qspi
, BSPI
, BSPI_CMD_AND_MODE_BYTE
, cmd_byte
);
380 bcm_qspi_write(qspi
, BSPI
, BSPI_FLEX_MODE_ENABLE
, flex_mode
);
383 static int bcm_qspi_bspi_set_flex_mode(struct bcm_qspi
*qspi
,
384 const struct spi_mem_op
*op
, int hp
)
386 int bpc
= 0, bpp
= 0;
387 u8 command
= op
->cmd
.opcode
;
388 int width
= op
->data
.buswidth
? op
->data
.buswidth
: SPI_NBITS_SINGLE
;
389 int addrlen
= op
->addr
.nbytes
;
392 dev_dbg(&qspi
->pdev
->dev
, "set flex mode w %x addrlen %x hp %d\n",
395 if (addrlen
== BSPI_ADDRLEN_4BYTES
)
396 bpp
= BSPI_BPP_ADDR_SELECT_MASK
;
398 bpp
|= (op
->dummy
.nbytes
* 8) / op
->dummy
.buswidth
;
401 case SPI_NBITS_SINGLE
:
402 if (addrlen
== BSPI_ADDRLEN_3BYTES
)
403 /* default mode, does not need flex_cmd */
409 bpc
|= 0x00010100; /* address and mode are 2-bit */
410 bpp
= BSPI_BPP_MODE_SELECT_MASK
;
416 bpc
|= 0x00020200; /* address and mode are 4-bit */
417 bpp
|= BSPI_BPP_MODE_SELECT_MASK
;
424 bcm_qspi_bspi_set_xfer_params(qspi
, command
, bpp
, bpc
, flex_mode
);
429 static int bcm_qspi_bspi_set_override(struct bcm_qspi
*qspi
,
430 const struct spi_mem_op
*op
, int hp
)
432 int width
= op
->data
.buswidth
? op
->data
.buswidth
: SPI_NBITS_SINGLE
;
433 int addrlen
= op
->addr
.nbytes
;
434 u32 data
= bcm_qspi_read(qspi
, BSPI
, BSPI_STRAP_OVERRIDE_CTRL
);
436 dev_dbg(&qspi
->pdev
->dev
, "set override mode w %x addrlen %x hp %d\n",
440 case SPI_NBITS_SINGLE
:
441 /* clear quad/dual mode */
442 data
&= ~(BSPI_STRAP_OVERRIDE_CTRL_DATA_QUAD
|
443 BSPI_STRAP_OVERRIDE_CTRL_DATA_DUAL
);
446 /* clear dual mode and set quad mode */
447 data
&= ~BSPI_STRAP_OVERRIDE_CTRL_DATA_DUAL
;
448 data
|= BSPI_STRAP_OVERRIDE_CTRL_DATA_QUAD
;
451 /* clear quad mode set dual mode */
452 data
&= ~BSPI_STRAP_OVERRIDE_CTRL_DATA_QUAD
;
453 data
|= BSPI_STRAP_OVERRIDE_CTRL_DATA_DUAL
;
459 if (addrlen
== BSPI_ADDRLEN_4BYTES
)
461 data
|= BSPI_STRAP_OVERRIDE_CTRL_ADDR_4BYTE
;
463 /* clear 4 byte mode */
464 data
&= ~BSPI_STRAP_OVERRIDE_CTRL_ADDR_4BYTE
;
466 /* set the override mode */
467 data
|= BSPI_STRAP_OVERRIDE_CTRL_OVERRIDE
;
468 bcm_qspi_write(qspi
, BSPI
, BSPI_STRAP_OVERRIDE_CTRL
, data
);
469 bcm_qspi_bspi_set_xfer_params(qspi
, op
->cmd
.opcode
, 0, 0, 0);
474 static int bcm_qspi_bspi_set_mode(struct bcm_qspi
*qspi
,
475 const struct spi_mem_op
*op
, int hp
)
478 int width
= op
->data
.buswidth
? op
->data
.buswidth
: SPI_NBITS_SINGLE
;
479 int addrlen
= op
->addr
.nbytes
;
482 qspi
->xfer_mode
.flex_mode
= true;
484 if (!bcm_qspi_bspi_ver_three(qspi
)) {
487 val
= bcm_qspi_read(qspi
, BSPI
, BSPI_STRAP_OVERRIDE_CTRL
);
488 mask
= BSPI_STRAP_OVERRIDE_CTRL_OVERRIDE
;
489 if (val
& mask
|| qspi
->s3_strap_override_ctrl
& mask
) {
490 qspi
->xfer_mode
.flex_mode
= false;
491 bcm_qspi_write(qspi
, BSPI
, BSPI_FLEX_MODE_ENABLE
, 0);
492 error
= bcm_qspi_bspi_set_override(qspi
, op
, hp
);
496 if (qspi
->xfer_mode
.flex_mode
)
497 error
= bcm_qspi_bspi_set_flex_mode(qspi
, op
, hp
);
500 dev_warn(&qspi
->pdev
->dev
,
501 "INVALID COMBINATION: width=%d addrlen=%d hp=%d\n",
503 } else if (qspi
->xfer_mode
.width
!= width
||
504 qspi
->xfer_mode
.addrlen
!= addrlen
||
505 qspi
->xfer_mode
.hp
!= hp
) {
506 qspi
->xfer_mode
.width
= width
;
507 qspi
->xfer_mode
.addrlen
= addrlen
;
508 qspi
->xfer_mode
.hp
= hp
;
509 dev_dbg(&qspi
->pdev
->dev
,
510 "cs:%d %d-lane output, %d-byte address%s\n",
512 qspi
->xfer_mode
.width
,
513 qspi
->xfer_mode
.addrlen
,
514 qspi
->xfer_mode
.hp
!= -1 ? ", hp mode" : "");
520 static void bcm_qspi_enable_bspi(struct bcm_qspi
*qspi
)
525 qspi
->bspi_enabled
= 1;
526 if ((bcm_qspi_read(qspi
, BSPI
, BSPI_MAST_N_BOOT_CTRL
) & 1) == 0)
529 bcm_qspi_bspi_flush_prefetch_buffers(qspi
);
531 bcm_qspi_write(qspi
, BSPI
, BSPI_MAST_N_BOOT_CTRL
, 0);
535 static void bcm_qspi_disable_bspi(struct bcm_qspi
*qspi
)
540 qspi
->bspi_enabled
= 0;
541 if ((bcm_qspi_read(qspi
, BSPI
, BSPI_MAST_N_BOOT_CTRL
) & 1))
544 bcm_qspi_bspi_busy_poll(qspi
);
545 bcm_qspi_write(qspi
, BSPI
, BSPI_MAST_N_BOOT_CTRL
, 1);
549 static void bcm_qspi_chip_select(struct bcm_qspi
*qspi
, int cs
)
554 if (qspi
->base
[CHIP_SELECT
]) {
555 rd
= bcm_qspi_read(qspi
, CHIP_SELECT
, 0);
556 wr
= (rd
& ~0xff) | (1 << cs
);
559 bcm_qspi_write(qspi
, CHIP_SELECT
, 0, wr
);
560 usleep_range(10, 20);
563 dev_dbg(&qspi
->pdev
->dev
, "using cs:%d\n", cs
);
568 static void bcm_qspi_hw_set_parms(struct bcm_qspi
*qspi
,
569 const struct bcm_qspi_parms
*xp
)
574 spbr
= qspi
->base_clk
/ (2 * xp
->speed_hz
);
576 spcr
= clamp_val(spbr
, bcm_qspi_spbr_min(qspi
), QSPI_SPBR_MAX
);
577 bcm_qspi_write(qspi
, MSPI
, MSPI_SPCR0_LSB
, spcr
);
579 if (!qspi
->mspi_maj_rev
)
580 /* legacy controller */
581 spcr
= MSPI_MASTER_BIT
;
585 /* for 16 bit the data should be zero */
586 if (xp
->bits_per_word
!= 16)
587 spcr
|= xp
->bits_per_word
<< 2;
588 spcr
|= xp
->mode
& 3;
590 bcm_qspi_write(qspi
, MSPI
, MSPI_SPCR0_MSB
, spcr
);
592 if (bcm_qspi_has_fastbr(qspi
)) {
596 spcr
|= MSPI_SPCR3_FASTBR
;
598 if (bcm_qspi_has_sysclk_108(qspi
)) {
600 spcr
|= MSPI_SPCR3_SYSCLKSEL_108
;
601 qspi
->base_clk
= MSPI_BASE_FREQ
* 4;
602 /* Change spbr as we changed sysclk */
603 bcm_qspi_write(qspi
, MSPI
, MSPI_SPCR0_LSB
, 4);
606 bcm_qspi_write(qspi
, MSPI
, MSPI_SPCR3
, spcr
);
609 qspi
->last_parms
= *xp
;
612 static void bcm_qspi_update_parms(struct bcm_qspi
*qspi
,
613 struct spi_device
*spi
,
614 struct spi_transfer
*trans
)
616 struct bcm_qspi_parms xp
;
618 xp
.speed_hz
= trans
->speed_hz
;
619 xp
.bits_per_word
= trans
->bits_per_word
;
622 bcm_qspi_hw_set_parms(qspi
, &xp
);
625 static int bcm_qspi_setup(struct spi_device
*spi
)
627 struct bcm_qspi_parms
*xp
;
629 if (spi
->bits_per_word
> 16)
632 xp
= spi_get_ctldata(spi
);
634 xp
= kzalloc(sizeof(*xp
), GFP_KERNEL
);
637 spi_set_ctldata(spi
, xp
);
639 xp
->speed_hz
= spi
->max_speed_hz
;
640 xp
->mode
= spi
->mode
;
642 if (spi
->bits_per_word
)
643 xp
->bits_per_word
= spi
->bits_per_word
;
645 xp
->bits_per_word
= 8;
650 static bool bcm_qspi_mspi_transfer_is_last(struct bcm_qspi
*qspi
,
651 struct qspi_trans
*qt
)
653 if (qt
->mspi_last_trans
&&
654 spi_transfer_is_last(qspi
->master
, qt
->trans
))
660 static int update_qspi_trans_byte_count(struct bcm_qspi
*qspi
,
661 struct qspi_trans
*qt
, int flags
)
663 int ret
= TRANS_STATUS_BREAK_NONE
;
665 /* count the last transferred bytes */
666 if (qt
->trans
->bits_per_word
<= 8)
671 if (qt
->byte
>= qt
->trans
->len
) {
672 /* we're at the end of the spi_transfer */
673 /* in TX mode, need to pause for a delay or CS change */
674 if (qt
->trans
->delay_usecs
&&
675 (flags
& TRANS_STATUS_BREAK_DELAY
))
676 ret
|= TRANS_STATUS_BREAK_DELAY
;
677 if (qt
->trans
->cs_change
&&
678 (flags
& TRANS_STATUS_BREAK_CS_CHANGE
))
679 ret
|= TRANS_STATUS_BREAK_CS_CHANGE
;
681 if (bcm_qspi_mspi_transfer_is_last(qspi
, qt
))
682 ret
|= TRANS_STATUS_BREAK_EOM
;
684 ret
|= TRANS_STATUS_BREAK_NO_BYTES
;
689 dev_dbg(&qspi
->pdev
->dev
, "trans %p len %d byte %d ret %x\n",
690 qt
->trans
, qt
->trans
? qt
->trans
->len
: 0, qt
->byte
, ret
);
694 static inline u8
read_rxram_slot_u8(struct bcm_qspi
*qspi
, int slot
)
696 u32 slot_offset
= MSPI_RXRAM
+ (slot
<< 3) + 0x4;
698 /* mask out reserved bits */
699 return bcm_qspi_read(qspi
, MSPI
, slot_offset
) & 0xff;
702 static inline u16
read_rxram_slot_u16(struct bcm_qspi
*qspi
, int slot
)
704 u32 reg_offset
= MSPI_RXRAM
;
705 u32 lsb_offset
= reg_offset
+ (slot
<< 3) + 0x4;
706 u32 msb_offset
= reg_offset
+ (slot
<< 3);
708 return (bcm_qspi_read(qspi
, MSPI
, lsb_offset
) & 0xff) |
709 ((bcm_qspi_read(qspi
, MSPI
, msb_offset
) & 0xff) << 8);
712 static void read_from_hw(struct bcm_qspi
*qspi
, int slots
)
714 struct qspi_trans tp
;
717 bcm_qspi_disable_bspi(qspi
);
719 if (slots
> MSPI_NUM_CDRAM
) {
720 /* should never happen */
721 dev_err(&qspi
->pdev
->dev
, "%s: too many slots!\n", __func__
);
725 tp
= qspi
->trans_pos
;
727 for (slot
= 0; slot
< slots
; slot
++) {
728 if (tp
.trans
->bits_per_word
<= 8) {
729 u8
*buf
= tp
.trans
->rx_buf
;
732 buf
[tp
.byte
] = read_rxram_slot_u8(qspi
, slot
);
733 dev_dbg(&qspi
->pdev
->dev
, "RD %02x\n",
734 buf
? buf
[tp
.byte
] : 0x0);
736 u16
*buf
= tp
.trans
->rx_buf
;
739 buf
[tp
.byte
/ 2] = read_rxram_slot_u16(qspi
,
741 dev_dbg(&qspi
->pdev
->dev
, "RD %04x\n",
742 buf
? buf
[tp
.byte
/ 2] : 0x0);
745 update_qspi_trans_byte_count(qspi
, &tp
,
746 TRANS_STATUS_BREAK_NONE
);
749 qspi
->trans_pos
= tp
;
752 static inline void write_txram_slot_u8(struct bcm_qspi
*qspi
, int slot
,
755 u32 reg_offset
= MSPI_TXRAM
+ (slot
<< 3);
757 /* mask out reserved bits */
758 bcm_qspi_write(qspi
, MSPI
, reg_offset
, val
);
761 static inline void write_txram_slot_u16(struct bcm_qspi
*qspi
, int slot
,
764 u32 reg_offset
= MSPI_TXRAM
;
765 u32 msb_offset
= reg_offset
+ (slot
<< 3);
766 u32 lsb_offset
= reg_offset
+ (slot
<< 3) + 0x4;
768 bcm_qspi_write(qspi
, MSPI
, msb_offset
, (val
>> 8));
769 bcm_qspi_write(qspi
, MSPI
, lsb_offset
, (val
& 0xff));
772 static inline u32
read_cdram_slot(struct bcm_qspi
*qspi
, int slot
)
774 return bcm_qspi_read(qspi
, MSPI
, MSPI_CDRAM
+ (slot
<< 2));
777 static inline void write_cdram_slot(struct bcm_qspi
*qspi
, int slot
, u32 val
)
779 bcm_qspi_write(qspi
, MSPI
, (MSPI_CDRAM
+ (slot
<< 2)), val
);
782 /* Return number of slots written */
783 static int write_to_hw(struct bcm_qspi
*qspi
, struct spi_device
*spi
)
785 struct qspi_trans tp
;
786 int slot
= 0, tstatus
= 0;
789 bcm_qspi_disable_bspi(qspi
);
790 tp
= qspi
->trans_pos
;
791 bcm_qspi_update_parms(qspi
, spi
, tp
.trans
);
793 /* Run until end of transfer or reached the max data */
794 while (!tstatus
&& slot
< MSPI_NUM_CDRAM
) {
795 if (tp
.trans
->bits_per_word
<= 8) {
796 const u8
*buf
= tp
.trans
->tx_buf
;
797 u8 val
= buf
? buf
[tp
.byte
] : 0x00;
799 write_txram_slot_u8(qspi
, slot
, val
);
800 dev_dbg(&qspi
->pdev
->dev
, "WR %02x\n", val
);
802 const u16
*buf
= tp
.trans
->tx_buf
;
803 u16 val
= buf
? buf
[tp
.byte
/ 2] : 0x0000;
805 write_txram_slot_u16(qspi
, slot
, val
);
806 dev_dbg(&qspi
->pdev
->dev
, "WR %04x\n", val
);
808 mspi_cdram
= MSPI_CDRAM_CONT_BIT
;
813 mspi_cdram
|= (~(1 << spi
->chip_select
) &
816 mspi_cdram
|= ((tp
.trans
->bits_per_word
<= 8) ? 0 :
817 MSPI_CDRAM_BITSE_BIT
);
819 write_cdram_slot(qspi
, slot
, mspi_cdram
);
821 tstatus
= update_qspi_trans_byte_count(qspi
, &tp
,
822 TRANS_STATUS_BREAK_TX
);
827 dev_err(&qspi
->pdev
->dev
, "%s: no data to send?", __func__
);
831 dev_dbg(&qspi
->pdev
->dev
, "submitting %d slots\n", slot
);
832 bcm_qspi_write(qspi
, MSPI
, MSPI_NEWQP
, 0);
833 bcm_qspi_write(qspi
, MSPI
, MSPI_ENDQP
, slot
- 1);
836 * case 1) EOM =1, cs_change =0: SSb inactive
837 * case 2) EOM =1, cs_change =1: SSb stay active
838 * case 3) EOM =0, cs_change =0: SSb stay active
839 * case 4) EOM =0, cs_change =1: SSb inactive
841 if (((tstatus
& TRANS_STATUS_BREAK_DESELECT
)
842 == TRANS_STATUS_BREAK_CS_CHANGE
) ||
843 ((tstatus
& TRANS_STATUS_BREAK_DESELECT
)
844 == TRANS_STATUS_BREAK_EOM
)) {
845 mspi_cdram
= read_cdram_slot(qspi
, slot
- 1) &
846 ~MSPI_CDRAM_CONT_BIT
;
847 write_cdram_slot(qspi
, slot
- 1, mspi_cdram
);
851 bcm_qspi_write(qspi
, MSPI
, MSPI_WRITE_LOCK
, 1);
853 /* Must flush previous writes before starting MSPI operation */
855 /* Set cont | spe | spifie */
856 bcm_qspi_write(qspi
, MSPI
, MSPI_SPCR2
, 0xe0);
862 static int bcm_qspi_bspi_exec_mem_op(struct spi_device
*spi
,
863 const struct spi_mem_op
*op
)
865 struct bcm_qspi
*qspi
= spi_master_get_devdata(spi
->master
);
866 u32 addr
= 0, len
, rdlen
, len_words
, from
= 0;
868 unsigned long timeo
= msecs_to_jiffies(100);
869 struct bcm_qspi_soc_intc
*soc_intc
= qspi
->soc_intc
;
871 if (bcm_qspi_bspi_ver_three(qspi
))
872 if (op
->addr
.nbytes
== BSPI_ADDRLEN_4BYTES
)
877 bcm_qspi_chip_select(qspi
, spi
->chip_select
);
878 bcm_qspi_write(qspi
, MSPI
, MSPI_WRITE_LOCK
, 0);
881 * when using flex mode we need to send
882 * the upper address byte to bspi
884 if (bcm_qspi_bspi_ver_three(qspi
) == false) {
885 addr
= from
& 0xff000000;
886 bcm_qspi_write(qspi
, BSPI
,
887 BSPI_BSPI_FLASH_UPPER_ADDR_BYTE
, addr
);
890 if (!qspi
->xfer_mode
.flex_mode
)
893 addr
= from
& 0x00ffffff;
895 if (bcm_qspi_bspi_ver_three(qspi
) == true)
896 addr
= (addr
+ 0xc00000) & 0xffffff;
899 * read into the entire buffer by breaking the reads
900 * into RAF buffer read lengths
902 len
= op
->data
.nbytes
;
903 qspi
->bspi_rf_op_idx
= 0;
906 if (len
> BSPI_READ_LENGTH
)
907 rdlen
= BSPI_READ_LENGTH
;
911 reinit_completion(&qspi
->bspi_done
);
912 bcm_qspi_enable_bspi(qspi
);
913 len_words
= (rdlen
+ 3) >> 2;
914 qspi
->bspi_rf_op
= op
;
915 qspi
->bspi_rf_op_status
= 0;
916 qspi
->bspi_rf_op_len
= rdlen
;
917 dev_dbg(&qspi
->pdev
->dev
,
918 "bspi xfr addr 0x%x len 0x%x", addr
, rdlen
);
919 bcm_qspi_write(qspi
, BSPI
, BSPI_RAF_START_ADDR
, addr
);
920 bcm_qspi_write(qspi
, BSPI
, BSPI_RAF_NUM_WORDS
, len_words
);
921 bcm_qspi_write(qspi
, BSPI
, BSPI_RAF_WATERMARK
, 0);
922 if (qspi
->soc_intc
) {
924 * clear soc MSPI and BSPI interrupts and enable
927 soc_intc
->bcm_qspi_int_ack(soc_intc
, MSPI_BSPI_DONE
);
928 soc_intc
->bcm_qspi_int_set(soc_intc
, BSPI_DONE
, true);
931 /* Must flush previous writes before starting BSPI operation */
933 bcm_qspi_bspi_lr_start(qspi
);
934 if (!wait_for_completion_timeout(&qspi
->bspi_done
, timeo
)) {
935 dev_err(&qspi
->pdev
->dev
, "timeout waiting for BSPI\n");
940 /* set msg return length */
948 static int bcm_qspi_transfer_one(struct spi_master
*master
,
949 struct spi_device
*spi
,
950 struct spi_transfer
*trans
)
952 struct bcm_qspi
*qspi
= spi_master_get_devdata(master
);
954 unsigned long timeo
= msecs_to_jiffies(100);
957 bcm_qspi_chip_select(qspi
, spi
->chip_select
);
958 qspi
->trans_pos
.trans
= trans
;
959 qspi
->trans_pos
.byte
= 0;
961 while (qspi
->trans_pos
.byte
< trans
->len
) {
962 reinit_completion(&qspi
->mspi_done
);
964 slots
= write_to_hw(qspi
, spi
);
965 if (!wait_for_completion_timeout(&qspi
->mspi_done
, timeo
)) {
966 dev_err(&qspi
->pdev
->dev
, "timeout waiting for MSPI\n");
970 read_from_hw(qspi
, slots
);
972 bcm_qspi_enable_bspi(qspi
);
977 static int bcm_qspi_mspi_exec_mem_op(struct spi_device
*spi
,
978 const struct spi_mem_op
*op
)
980 struct spi_master
*master
= spi
->master
;
981 struct bcm_qspi
*qspi
= spi_master_get_devdata(master
);
982 struct spi_transfer t
[2];
986 memset(cmd
, 0, sizeof(cmd
));
987 memset(t
, 0, sizeof(t
));
990 /* opcode is in cmd[0] */
991 cmd
[0] = op
->cmd
.opcode
;
992 for (i
= 0; i
< op
->addr
.nbytes
; i
++)
993 cmd
[1 + i
] = op
->addr
.val
>> (8 * (op
->addr
.nbytes
- i
- 1));
996 t
[0].len
= op
->addr
.nbytes
+ op
->dummy
.nbytes
+ 1;
997 t
[0].bits_per_word
= spi
->bits_per_word
;
998 t
[0].tx_nbits
= op
->cmd
.buswidth
;
999 /* lets mspi know that this is not last transfer */
1000 qspi
->trans_pos
.mspi_last_trans
= false;
1001 ret
= bcm_qspi_transfer_one(master
, spi
, &t
[0]);
1004 qspi
->trans_pos
.mspi_last_trans
= true;
1007 t
[1].rx_buf
= op
->data
.buf
.in
;
1008 t
[1].len
= op
->data
.nbytes
;
1009 t
[1].rx_nbits
= op
->data
.buswidth
;
1010 t
[1].bits_per_word
= spi
->bits_per_word
;
1011 ret
= bcm_qspi_transfer_one(master
, spi
, &t
[1]);
1017 static int bcm_qspi_exec_mem_op(struct spi_mem
*mem
,
1018 const struct spi_mem_op
*op
)
1020 struct spi_device
*spi
= mem
->spi
;
1021 struct bcm_qspi
*qspi
= spi_master_get_devdata(spi
->master
);
1023 bool mspi_read
= false;
1027 if (!op
->data
.nbytes
|| !op
->addr
.nbytes
|| op
->addr
.nbytes
> 4 ||
1028 op
->data
.dir
!= SPI_MEM_DATA_IN
)
1031 buf
= op
->data
.buf
.in
;
1032 addr
= op
->addr
.val
;
1033 len
= op
->data
.nbytes
;
1035 if (bcm_qspi_bspi_ver_three(qspi
) == true) {
1037 * The address coming into this function is a raw flash offset.
1038 * But for BSPI <= V3, we need to convert it to a remapped BSPI
1039 * address. If it crosses a 4MB boundary, just revert back to
1042 addr
= (addr
+ 0xc00000) & 0xffffff;
1044 if ((~ADDR_4MB_MASK
& addr
) ^
1045 (~ADDR_4MB_MASK
& (addr
+ len
- 1)))
1049 /* non-aligned and very short transfers are handled by MSPI */
1050 if (!IS_ALIGNED((uintptr_t)addr
, 4) || !IS_ALIGNED((uintptr_t)buf
, 4) ||
1055 return bcm_qspi_mspi_exec_mem_op(spi
, op
);
1057 ret
= bcm_qspi_bspi_set_mode(qspi
, op
, 0);
1060 ret
= bcm_qspi_bspi_exec_mem_op(spi
, op
);
1065 static void bcm_qspi_cleanup(struct spi_device
*spi
)
1067 struct bcm_qspi_parms
*xp
= spi_get_ctldata(spi
);
1072 static irqreturn_t
bcm_qspi_mspi_l2_isr(int irq
, void *dev_id
)
1074 struct bcm_qspi_dev_id
*qspi_dev_id
= dev_id
;
1075 struct bcm_qspi
*qspi
= qspi_dev_id
->dev
;
1076 u32 status
= bcm_qspi_read(qspi
, MSPI
, MSPI_MSPI_STATUS
);
1078 if (status
& MSPI_MSPI_STATUS_SPIF
) {
1079 struct bcm_qspi_soc_intc
*soc_intc
= qspi
->soc_intc
;
1080 /* clear interrupt */
1081 status
&= ~MSPI_MSPI_STATUS_SPIF
;
1082 bcm_qspi_write(qspi
, MSPI
, MSPI_MSPI_STATUS
, status
);
1084 soc_intc
->bcm_qspi_int_ack(soc_intc
, MSPI_DONE
);
1085 complete(&qspi
->mspi_done
);
1092 static irqreturn_t
bcm_qspi_bspi_lr_l2_isr(int irq
, void *dev_id
)
1094 struct bcm_qspi_dev_id
*qspi_dev_id
= dev_id
;
1095 struct bcm_qspi
*qspi
= qspi_dev_id
->dev
;
1096 struct bcm_qspi_soc_intc
*soc_intc
= qspi
->soc_intc
;
1097 u32 status
= qspi_dev_id
->irqp
->mask
;
1099 if (qspi
->bspi_enabled
&& qspi
->bspi_rf_op
) {
1100 bcm_qspi_bspi_lr_data_read(qspi
);
1101 if (qspi
->bspi_rf_op_len
== 0) {
1102 qspi
->bspi_rf_op
= NULL
;
1103 if (qspi
->soc_intc
) {
1104 /* disable soc BSPI interrupt */
1105 soc_intc
->bcm_qspi_int_set(soc_intc
, BSPI_DONE
,
1108 status
= INTR_BSPI_LR_SESSION_DONE_MASK
;
1111 if (qspi
->bspi_rf_op_status
)
1112 bcm_qspi_bspi_lr_clear(qspi
);
1114 bcm_qspi_bspi_flush_prefetch_buffers(qspi
);
1118 /* clear soc BSPI interrupt */
1119 soc_intc
->bcm_qspi_int_ack(soc_intc
, BSPI_DONE
);
1122 status
&= INTR_BSPI_LR_SESSION_DONE_MASK
;
1123 if (qspi
->bspi_enabled
&& status
&& qspi
->bspi_rf_op_len
== 0)
1124 complete(&qspi
->bspi_done
);
1129 static irqreturn_t
bcm_qspi_bspi_lr_err_l2_isr(int irq
, void *dev_id
)
1131 struct bcm_qspi_dev_id
*qspi_dev_id
= dev_id
;
1132 struct bcm_qspi
*qspi
= qspi_dev_id
->dev
;
1133 struct bcm_qspi_soc_intc
*soc_intc
= qspi
->soc_intc
;
1135 dev_err(&qspi
->pdev
->dev
, "BSPI INT error\n");
1136 qspi
->bspi_rf_op_status
= -EIO
;
1138 /* clear soc interrupt */
1139 soc_intc
->bcm_qspi_int_ack(soc_intc
, BSPI_ERR
);
1141 complete(&qspi
->bspi_done
);
1145 static irqreturn_t
bcm_qspi_l1_isr(int irq
, void *dev_id
)
1147 struct bcm_qspi_dev_id
*qspi_dev_id
= dev_id
;
1148 struct bcm_qspi
*qspi
= qspi_dev_id
->dev
;
1149 struct bcm_qspi_soc_intc
*soc_intc
= qspi
->soc_intc
;
1150 irqreturn_t ret
= IRQ_NONE
;
1153 u32 status
= soc_intc
->bcm_qspi_get_int_status(soc_intc
);
1155 if (status
& MSPI_DONE
)
1156 ret
= bcm_qspi_mspi_l2_isr(irq
, dev_id
);
1157 else if (status
& BSPI_DONE
)
1158 ret
= bcm_qspi_bspi_lr_l2_isr(irq
, dev_id
);
1159 else if (status
& BSPI_ERR
)
1160 ret
= bcm_qspi_bspi_lr_err_l2_isr(irq
, dev_id
);
1166 static const struct bcm_qspi_irq qspi_irq_tab
[] = {
1168 .irq_name
= "spi_lr_fullness_reached",
1169 .irq_handler
= bcm_qspi_bspi_lr_l2_isr
,
1170 .mask
= INTR_BSPI_LR_FULLNESS_REACHED_MASK
,
1173 .irq_name
= "spi_lr_session_aborted",
1174 .irq_handler
= bcm_qspi_bspi_lr_err_l2_isr
,
1175 .mask
= INTR_BSPI_LR_SESSION_ABORTED_MASK
,
1178 .irq_name
= "spi_lr_impatient",
1179 .irq_handler
= bcm_qspi_bspi_lr_err_l2_isr
,
1180 .mask
= INTR_BSPI_LR_IMPATIENT_MASK
,
1183 .irq_name
= "spi_lr_session_done",
1184 .irq_handler
= bcm_qspi_bspi_lr_l2_isr
,
1185 .mask
= INTR_BSPI_LR_SESSION_DONE_MASK
,
1187 #ifdef QSPI_INT_DEBUG
1188 /* this interrupt is for debug purposes only, dont request irq */
1190 .irq_name
= "spi_lr_overread",
1191 .irq_handler
= bcm_qspi_bspi_lr_err_l2_isr
,
1192 .mask
= INTR_BSPI_LR_OVERREAD_MASK
,
1196 .irq_name
= "mspi_done",
1197 .irq_handler
= bcm_qspi_mspi_l2_isr
,
1198 .mask
= INTR_MSPI_DONE_MASK
,
1201 .irq_name
= "mspi_halted",
1202 .irq_handler
= bcm_qspi_mspi_l2_isr
,
1203 .mask
= INTR_MSPI_HALTED_MASK
,
1206 /* single muxed L1 interrupt source */
1207 .irq_name
= "spi_l1_intr",
1208 .irq_handler
= bcm_qspi_l1_isr
,
1209 .irq_source
= MUXED_L1
,
1210 .mask
= QSPI_INTERRUPTS_ALL
,
1214 static void bcm_qspi_bspi_init(struct bcm_qspi
*qspi
)
1218 val
= bcm_qspi_read(qspi
, BSPI
, BSPI_REVISION_ID
);
1219 qspi
->bspi_maj_rev
= (val
>> 8) & 0xff;
1220 qspi
->bspi_min_rev
= val
& 0xff;
1221 if (!(bcm_qspi_bspi_ver_three(qspi
))) {
1222 /* Force mapping of BSPI address -> flash offset */
1223 bcm_qspi_write(qspi
, BSPI
, BSPI_BSPI_XOR_VALUE
, 0);
1224 bcm_qspi_write(qspi
, BSPI
, BSPI_BSPI_XOR_ENABLE
, 1);
1226 qspi
->bspi_enabled
= 1;
1227 bcm_qspi_disable_bspi(qspi
);
1228 bcm_qspi_write(qspi
, BSPI
, BSPI_B0_CTRL
, 0);
1229 bcm_qspi_write(qspi
, BSPI
, BSPI_B1_CTRL
, 0);
1232 static void bcm_qspi_hw_init(struct bcm_qspi
*qspi
)
1234 struct bcm_qspi_parms parms
;
1236 bcm_qspi_write(qspi
, MSPI
, MSPI_SPCR1_LSB
, 0);
1237 bcm_qspi_write(qspi
, MSPI
, MSPI_SPCR1_MSB
, 0);
1238 bcm_qspi_write(qspi
, MSPI
, MSPI_NEWQP
, 0);
1239 bcm_qspi_write(qspi
, MSPI
, MSPI_ENDQP
, 0);
1240 bcm_qspi_write(qspi
, MSPI
, MSPI_SPCR2
, 0x20);
1242 parms
.mode
= SPI_MODE_3
;
1243 parms
.bits_per_word
= 8;
1244 parms
.speed_hz
= qspi
->max_speed_hz
;
1245 bcm_qspi_hw_set_parms(qspi
, &parms
);
1248 bcm_qspi_bspi_init(qspi
);
1251 static void bcm_qspi_hw_uninit(struct bcm_qspi
*qspi
)
1253 bcm_qspi_write(qspi
, MSPI
, MSPI_SPCR2
, 0);
1255 bcm_qspi_write(qspi
, MSPI
, MSPI_WRITE_LOCK
, 0);
1259 static const struct spi_controller_mem_ops bcm_qspi_mem_ops
= {
1260 .exec_op
= bcm_qspi_exec_mem_op
,
1263 struct bcm_qspi_data
{
1265 bool has_spcr3_sysclk
;
1268 static const struct bcm_qspi_data bcm_qspi_no_rev_data
= {
1269 .has_mspi_rev
= false,
1270 .has_spcr3_sysclk
= false,
1273 static const struct bcm_qspi_data bcm_qspi_rev_data
= {
1274 .has_mspi_rev
= true,
1275 .has_spcr3_sysclk
= false,
1278 static const struct bcm_qspi_data bcm_qspi_spcr3_data
= {
1279 .has_mspi_rev
= true,
1280 .has_spcr3_sysclk
= true,
1283 static const struct of_device_id bcm_qspi_of_match
[] = {
1285 .compatible
= "brcm,spi-bcm7445-qspi",
1286 .data
= &bcm_qspi_rev_data
,
1290 .compatible
= "brcm,spi-bcm-qspi",
1291 .data
= &bcm_qspi_no_rev_data
,
1294 .compatible
= "brcm,spi-bcm7216-qspi",
1295 .data
= &bcm_qspi_spcr3_data
,
1298 .compatible
= "brcm,spi-bcm7278-qspi",
1299 .data
= &bcm_qspi_spcr3_data
,
1303 MODULE_DEVICE_TABLE(of
, bcm_qspi_of_match
);
1305 int bcm_qspi_probe(struct platform_device
*pdev
,
1306 struct bcm_qspi_soc_intc
*soc_intc
)
1308 const struct of_device_id
*of_id
= NULL
;
1309 const struct bcm_qspi_data
*data
;
1310 struct device
*dev
= &pdev
->dev
;
1311 struct bcm_qspi
*qspi
;
1312 struct spi_master
*master
;
1313 struct resource
*res
;
1314 int irq
, ret
= 0, num_ints
= 0;
1317 const char *name
= NULL
;
1318 int num_irqs
= ARRAY_SIZE(qspi_irq_tab
);
1320 /* We only support device-tree instantiation */
1324 of_id
= of_match_node(bcm_qspi_of_match
, dev
->of_node
);
1330 master
= devm_spi_alloc_master(dev
, sizeof(struct bcm_qspi
));
1332 dev_err(dev
, "error allocating spi_master\n");
1336 qspi
= spi_master_get_devdata(master
);
1338 qspi
->clk
= devm_clk_get_optional(&pdev
->dev
, NULL
);
1339 if (IS_ERR(qspi
->clk
))
1340 return PTR_ERR(qspi
->clk
);
1343 qspi
->trans_pos
.trans
= NULL
;
1344 qspi
->trans_pos
.byte
= 0;
1345 qspi
->trans_pos
.mspi_last_trans
= true;
1346 qspi
->master
= master
;
1348 master
->bus_num
= -1;
1349 master
->mode_bits
= SPI_CPHA
| SPI_CPOL
| SPI_RX_DUAL
| SPI_RX_QUAD
;
1350 master
->setup
= bcm_qspi_setup
;
1351 master
->transfer_one
= bcm_qspi_transfer_one
;
1352 master
->mem_ops
= &bcm_qspi_mem_ops
;
1353 master
->cleanup
= bcm_qspi_cleanup
;
1354 master
->dev
.of_node
= dev
->of_node
;
1355 master
->num_chipselect
= NUM_CHIPSELECT
;
1356 master
->use_gpio_descriptors
= true;
1358 qspi
->big_endian
= of_device_is_big_endian(dev
->of_node
);
1360 if (!of_property_read_u32(dev
->of_node
, "num-cs", &val
))
1361 master
->num_chipselect
= val
;
1363 res
= platform_get_resource_byname(pdev
, IORESOURCE_MEM
, "hif_mspi");
1365 res
= platform_get_resource_byname(pdev
, IORESOURCE_MEM
,
1369 qspi
->base
[MSPI
] = devm_ioremap_resource(dev
, res
);
1370 if (IS_ERR(qspi
->base
[MSPI
]))
1371 return PTR_ERR(qspi
->base
[MSPI
]);
1376 res
= platform_get_resource_byname(pdev
, IORESOURCE_MEM
, "bspi");
1378 qspi
->base
[BSPI
] = devm_ioremap_resource(dev
, res
);
1379 if (IS_ERR(qspi
->base
[BSPI
]))
1380 return PTR_ERR(qspi
->base
[BSPI
]);
1381 qspi
->bspi_mode
= true;
1383 qspi
->bspi_mode
= false;
1386 dev_info(dev
, "using %smspi mode\n", qspi
->bspi_mode
? "bspi-" : "");
1388 res
= platform_get_resource_byname(pdev
, IORESOURCE_MEM
, "cs_reg");
1390 qspi
->base
[CHIP_SELECT
] = devm_ioremap_resource(dev
, res
);
1391 if (IS_ERR(qspi
->base
[CHIP_SELECT
]))
1392 return PTR_ERR(qspi
->base
[CHIP_SELECT
]);
1395 qspi
->dev_ids
= kcalloc(num_irqs
, sizeof(struct bcm_qspi_dev_id
),
1400 for (val
= 0; val
< num_irqs
; val
++) {
1402 name
= qspi_irq_tab
[val
].irq_name
;
1403 if (qspi_irq_tab
[val
].irq_source
== SINGLE_L2
) {
1404 /* get the l2 interrupts */
1405 irq
= platform_get_irq_byname_optional(pdev
, name
);
1406 } else if (!num_ints
&& soc_intc
) {
1407 /* all mspi, bspi intrs muxed to one L1 intr */
1408 irq
= platform_get_irq(pdev
, 0);
1412 ret
= devm_request_irq(&pdev
->dev
, irq
,
1413 qspi_irq_tab
[val
].irq_handler
, 0,
1415 &qspi
->dev_ids
[val
]);
1417 dev_err(&pdev
->dev
, "IRQ %s not found\n", name
);
1418 goto qspi_probe_err
;
1421 qspi
->dev_ids
[val
].dev
= qspi
;
1422 qspi
->dev_ids
[val
].irqp
= &qspi_irq_tab
[val
];
1424 dev_dbg(&pdev
->dev
, "registered IRQ %s %d\n",
1425 qspi_irq_tab
[val
].irq_name
,
1431 dev_err(&pdev
->dev
, "no IRQs registered, cannot init driver\n");
1433 goto qspi_probe_err
;
1437 * Some SoCs integrate spi controller (e.g., its interrupt bits)
1441 qspi
->soc_intc
= soc_intc
;
1442 soc_intc
->bcm_qspi_int_set(soc_intc
, MSPI_DONE
, true);
1444 qspi
->soc_intc
= NULL
;
1447 ret
= clk_prepare_enable(qspi
->clk
);
1449 dev_err(dev
, "failed to prepare clock\n");
1450 goto qspi_probe_err
;
1453 qspi
->base_clk
= clk_get_rate(qspi
->clk
);
1455 if (data
->has_mspi_rev
) {
1456 rev
= bcm_qspi_read(qspi
, MSPI
, MSPI_REV
);
1457 /* some older revs do not have a MSPI_REV register */
1458 if ((rev
& 0xff) == 0xff)
1462 qspi
->mspi_maj_rev
= (rev
>> 4) & 0xf;
1463 qspi
->mspi_min_rev
= rev
& 0xf;
1464 qspi
->mspi_spcr3_sysclk
= data
->has_spcr3_sysclk
;
1466 qspi
->max_speed_hz
= qspi
->base_clk
/ (bcm_qspi_spbr_min(qspi
) * 2);
1468 bcm_qspi_hw_init(qspi
);
1469 init_completion(&qspi
->mspi_done
);
1470 init_completion(&qspi
->bspi_done
);
1473 platform_set_drvdata(pdev
, qspi
);
1475 qspi
->xfer_mode
.width
= -1;
1476 qspi
->xfer_mode
.addrlen
= -1;
1477 qspi
->xfer_mode
.hp
= -1;
1479 ret
= spi_register_master(master
);
1481 dev_err(dev
, "can't register master\n");
1488 bcm_qspi_hw_uninit(qspi
);
1489 clk_disable_unprepare(qspi
->clk
);
1491 kfree(qspi
->dev_ids
);
1494 /* probe function to be called by SoC specific platform driver probe */
1495 EXPORT_SYMBOL_GPL(bcm_qspi_probe
);
1497 int bcm_qspi_remove(struct platform_device
*pdev
)
1499 struct bcm_qspi
*qspi
= platform_get_drvdata(pdev
);
1501 spi_unregister_master(qspi
->master
);
1502 bcm_qspi_hw_uninit(qspi
);
1503 clk_disable_unprepare(qspi
->clk
);
1504 kfree(qspi
->dev_ids
);
1508 /* function to be called by SoC specific platform driver remove() */
1509 EXPORT_SYMBOL_GPL(bcm_qspi_remove
);
1511 static int __maybe_unused
bcm_qspi_suspend(struct device
*dev
)
1513 struct bcm_qspi
*qspi
= dev_get_drvdata(dev
);
1515 /* store the override strap value */
1516 if (!bcm_qspi_bspi_ver_three(qspi
))
1517 qspi
->s3_strap_override_ctrl
=
1518 bcm_qspi_read(qspi
, BSPI
, BSPI_STRAP_OVERRIDE_CTRL
);
1520 spi_master_suspend(qspi
->master
);
1521 clk_disable_unprepare(qspi
->clk
);
1522 bcm_qspi_hw_uninit(qspi
);
1527 static int __maybe_unused
bcm_qspi_resume(struct device
*dev
)
1529 struct bcm_qspi
*qspi
= dev_get_drvdata(dev
);
1532 bcm_qspi_hw_init(qspi
);
1533 bcm_qspi_chip_select(qspi
, qspi
->curr_cs
);
1535 /* enable MSPI interrupt */
1536 qspi
->soc_intc
->bcm_qspi_int_set(qspi
->soc_intc
, MSPI_DONE
,
1539 ret
= clk_prepare_enable(qspi
->clk
);
1541 spi_master_resume(qspi
->master
);
1546 SIMPLE_DEV_PM_OPS(bcm_qspi_pm_ops
, bcm_qspi_suspend
, bcm_qspi_resume
);
1548 /* pm_ops to be called by SoC specific platform driver */
1549 EXPORT_SYMBOL_GPL(bcm_qspi_pm_ops
);
1551 MODULE_AUTHOR("Kamal Dasu");
1552 MODULE_DESCRIPTION("Broadcom QSPI driver");
1553 MODULE_LICENSE("GPL v2");
1554 MODULE_ALIAS("platform:" DRIVER_NAME
);