2 * Microchip PIC32 SPI controller driver.
4 * Purna Chandra Mandal <purna.mandal@microchip.com>
5 * Copyright (c) 2016, Microchip Technology Inc.
7 * This program is free software; you can distribute it and/or modify it
8 * under the terms of the GNU General Public License (Version 2) as
9 * published by the Free Software Foundation.
11 * This program is distributed in the hope it will be useful, but WITHOUT
12 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
13 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
17 #include <linux/clk.h>
18 #include <linux/clkdev.h>
19 #include <linux/delay.h>
20 #include <linux/dmaengine.h>
21 #include <linux/dma-mapping.h>
22 #include <linux/highmem.h>
23 #include <linux/module.h>
25 #include <linux/interrupt.h>
27 #include <linux/of_irq.h>
28 #include <linux/of_gpio.h>
29 #include <linux/of_address.h>
30 #include <linux/platform_device.h>
31 #include <linux/spi/spi.h>
33 /* SPI controller registers */
34 struct pic32_spi_regs
{
53 /* Bit fields of SPI Control Register */
54 #define CTRL_RX_INT_SHIFT 0 /* Rx interrupt generation */
55 #define RX_FIFO_EMPTY 0
56 #define RX_FIFO_NOT_EMPTY 1 /* not empty */
57 #define RX_FIFO_HALF_FULL 2 /* full by half or more */
58 #define RX_FIFO_FULL 3 /* completely full */
60 #define CTRL_TX_INT_SHIFT 2 /* TX interrupt generation */
61 #define TX_FIFO_ALL_EMPTY 0 /* completely empty */
62 #define TX_FIFO_EMPTY 1 /* empty */
63 #define TX_FIFO_HALF_EMPTY 2 /* empty by half or more */
64 #define TX_FIFO_NOT_FULL 3 /* atleast one empty */
66 #define CTRL_MSTEN BIT(5) /* enable master mode */
67 #define CTRL_CKP BIT(6) /* active low */
68 #define CTRL_CKE BIT(8) /* Tx on falling edge */
69 #define CTRL_SMP BIT(9) /* Rx at middle or end of tx */
70 #define CTRL_BPW_MASK 0x03 /* bits per word/sample */
71 #define CTRL_BPW_SHIFT 10
73 #define PIC32_BPW_16 1
74 #define PIC32_BPW_32 2
75 #define CTRL_SIDL BIT(13) /* sleep when idle */
76 #define CTRL_ON BIT(15) /* enable macro */
77 #define CTRL_ENHBUF BIT(16) /* enable enhanced buffering */
78 #define CTRL_MCLKSEL BIT(23) /* select clock source */
79 #define CTRL_MSSEN BIT(28) /* macro driven /SS */
80 #define CTRL_FRMEN BIT(31) /* enable framing mode */
82 /* Bit fields of SPI Status Register */
83 #define STAT_RF_EMPTY BIT(5) /* RX Fifo empty */
84 #define STAT_RX_OV BIT(6) /* err, s/w needs to clear */
85 #define STAT_TX_UR BIT(8) /* UR in Framed SPI modes */
86 #define STAT_FRM_ERR BIT(12) /* Multiple Frame Sync pulse */
87 #define STAT_TF_LVL_MASK 0x1F
88 #define STAT_TF_LVL_SHIFT 16
89 #define STAT_RF_LVL_MASK 0x1F
90 #define STAT_RF_LVL_SHIFT 24
92 /* Bit fields of SPI Baud Register */
93 #define BAUD_MASK 0x1ff
95 /* Bit fields of SPI Control2 Register */
96 #define CTRL2_TX_UR_EN BIT(10) /* Enable int on Tx under-run */
97 #define CTRL2_RX_OV_EN BIT(11) /* Enable int on Rx over-run */
98 #define CTRL2_FRM_ERR_EN BIT(12) /* Enable frame err int */
100 /* Minimum DMA transfer size */
101 #define PIC32_DMA_LEN_MIN 64
105 struct pic32_spi_regs __iomem
*regs
;
109 u32 fifo_n_byte
; /* FIFO depth in bytes */
111 struct spi_master
*master
;
112 /* Current controller setting */
113 u32 speed_hz
; /* spi-clk rate */
116 u32 fifo_n_elm
; /* FIFO depth in words */
117 #define PIC32F_DMA_PREP 0 /* DMA chnls configured */
119 /* Current transfer state */
120 struct completion xfer_done
;
121 /* PIO transfer specific */
127 void (*rx_fifo
)(struct pic32_spi
*);
128 void (*tx_fifo
)(struct pic32_spi
*);
131 static inline void pic32_spi_enable(struct pic32_spi
*pic32s
)
133 writel(CTRL_ON
| CTRL_SIDL
, &pic32s
->regs
->ctrl_set
);
136 static inline void pic32_spi_disable(struct pic32_spi
*pic32s
)
138 writel(CTRL_ON
| CTRL_SIDL
, &pic32s
->regs
->ctrl_clr
);
140 /* avoid SPI registers read/write at immediate next CPU clock */
144 static void pic32_spi_set_clk_rate(struct pic32_spi
*pic32s
, u32 spi_ck
)
148 /* div = (clk_in / 2 * spi_ck) - 1 */
149 div
= DIV_ROUND_CLOSEST(clk_get_rate(pic32s
->clk
), 2 * spi_ck
) - 1;
151 writel(div
& BAUD_MASK
, &pic32s
->regs
->baud
);
154 static inline u32
pic32_rx_fifo_level(struct pic32_spi
*pic32s
)
156 u32 sr
= readl(&pic32s
->regs
->status
);
158 return (sr
>> STAT_RF_LVL_SHIFT
) & STAT_RF_LVL_MASK
;
161 static inline u32
pic32_tx_fifo_level(struct pic32_spi
*pic32s
)
163 u32 sr
= readl(&pic32s
->regs
->status
);
165 return (sr
>> STAT_TF_LVL_SHIFT
) & STAT_TF_LVL_MASK
;
168 /* Return the max entries we can fill into tx fifo */
169 static u32
pic32_tx_max(struct pic32_spi
*pic32s
, int n_bytes
)
171 u32 tx_left
, tx_room
, rxtx_gap
;
173 tx_left
= (pic32s
->tx_end
- pic32s
->tx
) / n_bytes
;
174 tx_room
= pic32s
->fifo_n_elm
- pic32_tx_fifo_level(pic32s
);
177 * Another concern is about the tx/rx mismatch, we
178 * though to use (pic32s->fifo_n_byte - rxfl - txfl) as
179 * one maximum value for tx, but it doesn't cover the
180 * data which is out of tx/rx fifo and inside the
181 * shift registers. So a ctrl from sw point of
184 rxtx_gap
= ((pic32s
->rx_end
- pic32s
->rx
) -
185 (pic32s
->tx_end
- pic32s
->tx
)) / n_bytes
;
186 return min3(tx_left
, tx_room
, (u32
)(pic32s
->fifo_n_elm
- rxtx_gap
));
189 /* Return the max entries we should read out of rx fifo */
190 static u32
pic32_rx_max(struct pic32_spi
*pic32s
, int n_bytes
)
192 u32 rx_left
= (pic32s
->rx_end
- pic32s
->rx
) / n_bytes
;
194 return min_t(u32
, rx_left
, pic32_rx_fifo_level(pic32s
));
197 #define BUILD_SPI_FIFO_RW(__name, __type, __bwl) \
198 static void pic32_spi_rx_##__name(struct pic32_spi *pic32s) \
201 u32 mx = pic32_rx_max(pic32s, sizeof(__type)); \
203 v = read##__bwl(&pic32s->regs->buf); \
204 if (pic32s->rx_end - pic32s->len) \
205 *(__type *)(pic32s->rx) = v; \
206 pic32s->rx += sizeof(__type); \
210 static void pic32_spi_tx_##__name(struct pic32_spi *pic32s) \
213 u32 mx = pic32_tx_max(pic32s, sizeof(__type)); \
214 for (; mx ; mx--) { \
216 if (pic32s->tx_end - pic32s->len) \
217 v = *(__type *)(pic32s->tx); \
218 write##__bwl(v, &pic32s->regs->buf); \
219 pic32s->tx += sizeof(__type); \
223 BUILD_SPI_FIFO_RW(byte
, u8
, b
);
224 BUILD_SPI_FIFO_RW(word
, u16
, w
);
225 BUILD_SPI_FIFO_RW(dword
, u32
, l
);
227 static void pic32_err_stop(struct pic32_spi
*pic32s
, const char *msg
)
229 /* disable all interrupts */
230 disable_irq_nosync(pic32s
->fault_irq
);
231 disable_irq_nosync(pic32s
->rx_irq
);
232 disable_irq_nosync(pic32s
->tx_irq
);
234 /* Show err message and abort xfer with err */
235 dev_err(&pic32s
->master
->dev
, "%s\n", msg
);
236 if (pic32s
->master
->cur_msg
)
237 pic32s
->master
->cur_msg
->status
= -EIO
;
238 complete(&pic32s
->xfer_done
);
241 static irqreturn_t
pic32_spi_fault_irq(int irq
, void *dev_id
)
243 struct pic32_spi
*pic32s
= dev_id
;
246 status
= readl(&pic32s
->regs
->status
);
249 if (status
& (STAT_RX_OV
| STAT_TX_UR
)) {
250 writel(STAT_RX_OV
, &pic32s
->regs
->status_clr
);
251 writel(STAT_TX_UR
, &pic32s
->regs
->status_clr
);
252 pic32_err_stop(pic32s
, "err_irq: fifo ov/ur-run\n");
256 if (status
& STAT_FRM_ERR
) {
257 pic32_err_stop(pic32s
, "err_irq: frame error");
261 if (!pic32s
->master
->cur_msg
) {
262 pic32_err_stop(pic32s
, "err_irq: no mesg");
269 static irqreturn_t
pic32_spi_rx_irq(int irq
, void *dev_id
)
271 struct pic32_spi
*pic32s
= dev_id
;
273 pic32s
->rx_fifo(pic32s
);
276 if (pic32s
->rx_end
== pic32s
->rx
) {
277 /* disable all interrupts */
278 disable_irq_nosync(pic32s
->fault_irq
);
279 disable_irq_nosync(pic32s
->rx_irq
);
281 /* complete current xfer */
282 complete(&pic32s
->xfer_done
);
288 static irqreturn_t
pic32_spi_tx_irq(int irq
, void *dev_id
)
290 struct pic32_spi
*pic32s
= dev_id
;
292 pic32s
->tx_fifo(pic32s
);
294 /* tx complete? disable tx interrupt */
295 if (pic32s
->tx_end
== pic32s
->tx
)
296 disable_irq_nosync(pic32s
->tx_irq
);
301 static void pic32_spi_dma_rx_notify(void *data
)
303 struct pic32_spi
*pic32s
= data
;
305 complete(&pic32s
->xfer_done
);
308 static int pic32_spi_dma_transfer(struct pic32_spi
*pic32s
,
309 struct spi_transfer
*xfer
)
311 struct spi_master
*master
= pic32s
->master
;
312 struct dma_async_tx_descriptor
*desc_rx
;
313 struct dma_async_tx_descriptor
*desc_tx
;
317 if (!master
->dma_rx
|| !master
->dma_tx
)
320 desc_rx
= dmaengine_prep_slave_sg(master
->dma_rx
,
324 DMA_PREP_INTERRUPT
| DMA_CTRL_ACK
);
330 desc_tx
= dmaengine_prep_slave_sg(master
->dma_tx
,
334 DMA_PREP_INTERRUPT
| DMA_CTRL_ACK
);
340 /* Put callback on the RX transfer, that should finish last */
341 desc_rx
->callback
= pic32_spi_dma_rx_notify
;
342 desc_rx
->callback_param
= pic32s
;
344 cookie
= dmaengine_submit(desc_rx
);
345 ret
= dma_submit_error(cookie
);
349 cookie
= dmaengine_submit(desc_tx
);
350 ret
= dma_submit_error(cookie
);
354 dma_async_issue_pending(master
->dma_rx
);
355 dma_async_issue_pending(master
->dma_tx
);
360 dmaengine_terminate_all(master
->dma_rx
);
365 static int pic32_spi_dma_config(struct pic32_spi
*pic32s
, u32 dma_width
)
367 int buf_offset
= offsetof(struct pic32_spi_regs
, buf
);
368 struct spi_master
*master
= pic32s
->master
;
369 struct dma_slave_config cfg
;
372 cfg
.device_fc
= true;
373 cfg
.src_addr
= pic32s
->dma_base
+ buf_offset
;
374 cfg
.dst_addr
= pic32s
->dma_base
+ buf_offset
;
375 cfg
.src_maxburst
= pic32s
->fifo_n_elm
/ 2; /* fill one-half */
376 cfg
.dst_maxburst
= pic32s
->fifo_n_elm
/ 2; /* drain one-half */
377 cfg
.src_addr_width
= dma_width
;
378 cfg
.dst_addr_width
= dma_width
;
380 cfg
.slave_id
= pic32s
->tx_irq
;
381 cfg
.direction
= DMA_MEM_TO_DEV
;
382 ret
= dmaengine_slave_config(master
->dma_tx
, &cfg
);
384 dev_err(&master
->dev
, "tx channel setup failed\n");
388 cfg
.slave_id
= pic32s
->rx_irq
;
389 cfg
.direction
= DMA_DEV_TO_MEM
;
390 ret
= dmaengine_slave_config(master
->dma_rx
, &cfg
);
392 dev_err(&master
->dev
, "rx channel setup failed\n");
397 static int pic32_spi_set_word_size(struct pic32_spi
*pic32s
, u8 bits_per_word
)
399 enum dma_slave_buswidth dmawidth
;
402 switch (bits_per_word
) {
404 pic32s
->rx_fifo
= pic32_spi_rx_byte
;
405 pic32s
->tx_fifo
= pic32_spi_tx_byte
;
406 buswidth
= PIC32_BPW_8
;
407 dmawidth
= DMA_SLAVE_BUSWIDTH_1_BYTE
;
410 pic32s
->rx_fifo
= pic32_spi_rx_word
;
411 pic32s
->tx_fifo
= pic32_spi_tx_word
;
412 buswidth
= PIC32_BPW_16
;
413 dmawidth
= DMA_SLAVE_BUSWIDTH_2_BYTES
;
416 pic32s
->rx_fifo
= pic32_spi_rx_dword
;
417 pic32s
->tx_fifo
= pic32_spi_tx_dword
;
418 buswidth
= PIC32_BPW_32
;
419 dmawidth
= DMA_SLAVE_BUSWIDTH_4_BYTES
;
426 /* calculate maximum number of words fifos can hold */
427 pic32s
->fifo_n_elm
= DIV_ROUND_UP(pic32s
->fifo_n_byte
,
430 v
= readl(&pic32s
->regs
->ctrl
);
431 v
&= ~(CTRL_BPW_MASK
<< CTRL_BPW_SHIFT
);
432 v
|= buswidth
<< CTRL_BPW_SHIFT
;
433 writel(v
, &pic32s
->regs
->ctrl
);
435 /* re-configure dma width, if required */
436 if (test_bit(PIC32F_DMA_PREP
, &pic32s
->flags
))
437 pic32_spi_dma_config(pic32s
, dmawidth
);
442 static int pic32_spi_prepare_hardware(struct spi_master
*master
)
444 struct pic32_spi
*pic32s
= spi_master_get_devdata(master
);
446 pic32_spi_enable(pic32s
);
451 static int pic32_spi_prepare_message(struct spi_master
*master
,
452 struct spi_message
*msg
)
454 struct pic32_spi
*pic32s
= spi_master_get_devdata(master
);
455 struct spi_device
*spi
= msg
->spi
;
458 /* set device specific bits_per_word */
459 if (pic32s
->bits_per_word
!= spi
->bits_per_word
) {
460 pic32_spi_set_word_size(pic32s
, spi
->bits_per_word
);
461 pic32s
->bits_per_word
= spi
->bits_per_word
;
464 /* device specific speed change */
465 if (pic32s
->speed_hz
!= spi
->max_speed_hz
) {
466 pic32_spi_set_clk_rate(pic32s
, spi
->max_speed_hz
);
467 pic32s
->speed_hz
= spi
->max_speed_hz
;
470 /* device specific mode change */
471 if (pic32s
->mode
!= spi
->mode
) {
472 val
= readl(&pic32s
->regs
->ctrl
);
474 if (spi
->mode
& SPI_CPOL
)
478 /* tx on rising edge */
479 if (spi
->mode
& SPI_CPHA
)
484 /* rx at end of tx */
486 writel(val
, &pic32s
->regs
->ctrl
);
487 pic32s
->mode
= spi
->mode
;
493 static bool pic32_spi_can_dma(struct spi_master
*master
,
494 struct spi_device
*spi
,
495 struct spi_transfer
*xfer
)
497 struct pic32_spi
*pic32s
= spi_master_get_devdata(master
);
499 /* skip using DMA on small size transfer to avoid overhead.*/
500 return (xfer
->len
>= PIC32_DMA_LEN_MIN
) &&
501 test_bit(PIC32F_DMA_PREP
, &pic32s
->flags
);
504 static int pic32_spi_one_transfer(struct spi_master
*master
,
505 struct spi_device
*spi
,
506 struct spi_transfer
*transfer
)
508 struct pic32_spi
*pic32s
;
509 bool dma_issued
= false;
510 unsigned long timeout
;
513 pic32s
= spi_master_get_devdata(master
);
515 /* handle transfer specific word size change */
516 if (transfer
->bits_per_word
&&
517 (transfer
->bits_per_word
!= pic32s
->bits_per_word
)) {
518 ret
= pic32_spi_set_word_size(pic32s
, transfer
->bits_per_word
);
521 pic32s
->bits_per_word
= transfer
->bits_per_word
;
524 /* handle transfer specific speed change */
525 if (transfer
->speed_hz
&& (transfer
->speed_hz
!= pic32s
->speed_hz
)) {
526 pic32_spi_set_clk_rate(pic32s
, transfer
->speed_hz
);
527 pic32s
->speed_hz
= transfer
->speed_hz
;
530 reinit_completion(&pic32s
->xfer_done
);
532 /* transact by DMA mode */
533 if (transfer
->rx_sg
.nents
&& transfer
->tx_sg
.nents
) {
534 ret
= pic32_spi_dma_transfer(pic32s
, transfer
);
536 dev_err(&spi
->dev
, "dma submit error\n");
543 /* set current transfer information */
544 pic32s
->tx
= (const void *)transfer
->tx_buf
;
545 pic32s
->rx
= (const void *)transfer
->rx_buf
;
546 pic32s
->tx_end
= pic32s
->tx
+ transfer
->len
;
547 pic32s
->rx_end
= pic32s
->rx
+ transfer
->len
;
548 pic32s
->len
= transfer
->len
;
550 /* transact by interrupt driven PIO */
551 enable_irq(pic32s
->fault_irq
);
552 enable_irq(pic32s
->rx_irq
);
553 enable_irq(pic32s
->tx_irq
);
556 /* wait for completion */
557 timeout
= wait_for_completion_timeout(&pic32s
->xfer_done
, 2 * HZ
);
559 dev_err(&spi
->dev
, "wait error/timedout\n");
561 dmaengine_terminate_all(master
->dma_rx
);
562 dmaengine_terminate_all(master
->dma_rx
);
572 static int pic32_spi_unprepare_message(struct spi_master
*master
,
573 struct spi_message
*msg
)
579 static int pic32_spi_unprepare_hardware(struct spi_master
*master
)
581 struct pic32_spi
*pic32s
= spi_master_get_devdata(master
);
583 pic32_spi_disable(pic32s
);
588 /* This may be called multiple times by same spi dev */
589 static int pic32_spi_setup(struct spi_device
*spi
)
591 if (!spi
->max_speed_hz
) {
592 dev_err(&spi
->dev
, "No max speed HZ parameter\n");
596 /* PIC32 spi controller can drive /CS during transfer depending
597 * on tx fifo fill-level. /CS will stay asserted as long as TX
598 * fifo is non-empty, else will be deasserted indicating
599 * completion of the ongoing transfer. This might result into
600 * unreliable/erroneous SPI transactions.
601 * To avoid that we will always handle /CS by toggling GPIO.
603 if (!gpio_is_valid(spi
->cs_gpio
))
606 gpio_direction_output(spi
->cs_gpio
, !(spi
->mode
& SPI_CS_HIGH
));
611 static void pic32_spi_cleanup(struct spi_device
*spi
)
613 /* de-activate cs-gpio */
614 gpio_direction_output(spi
->cs_gpio
, !(spi
->mode
& SPI_CS_HIGH
));
617 static void pic32_spi_dma_prep(struct pic32_spi
*pic32s
, struct device
*dev
)
619 struct spi_master
*master
= pic32s
->master
;
623 dma_cap_set(DMA_SLAVE
, mask
);
625 master
->dma_rx
= dma_request_slave_channel_compat(mask
, NULL
, NULL
,
627 if (!master
->dma_rx
) {
628 dev_warn(dev
, "RX channel not found.\n");
632 master
->dma_tx
= dma_request_slave_channel_compat(mask
, NULL
, NULL
,
634 if (!master
->dma_tx
) {
635 dev_warn(dev
, "TX channel not found.\n");
639 if (pic32_spi_dma_config(pic32s
, DMA_SLAVE_BUSWIDTH_1_BYTE
))
642 /* DMA chnls allocated and prepared */
643 set_bit(PIC32F_DMA_PREP
, &pic32s
->flags
);
649 dma_release_channel(master
->dma_rx
);
652 dma_release_channel(master
->dma_tx
);
655 static void pic32_spi_dma_unprep(struct pic32_spi
*pic32s
)
657 if (!test_bit(PIC32F_DMA_PREP
, &pic32s
->flags
))
660 clear_bit(PIC32F_DMA_PREP
, &pic32s
->flags
);
661 if (pic32s
->master
->dma_rx
)
662 dma_release_channel(pic32s
->master
->dma_rx
);
664 if (pic32s
->master
->dma_tx
)
665 dma_release_channel(pic32s
->master
->dma_tx
);
668 static void pic32_spi_hw_init(struct pic32_spi
*pic32s
)
672 /* disable hardware */
673 pic32_spi_disable(pic32s
);
675 ctrl
= readl(&pic32s
->regs
->ctrl
);
676 /* enable enhanced fifo of 128bit deep */
678 pic32s
->fifo_n_byte
= 16;
680 /* disable framing mode */
683 /* enable master mode while disabled */
686 /* set tx fifo threshold interrupt */
687 ctrl
&= ~(0x3 << CTRL_TX_INT_SHIFT
);
688 ctrl
|= (TX_FIFO_HALF_EMPTY
<< CTRL_TX_INT_SHIFT
);
690 /* set rx fifo threshold interrupt */
691 ctrl
&= ~(0x3 << CTRL_RX_INT_SHIFT
);
692 ctrl
|= (RX_FIFO_NOT_EMPTY
<< CTRL_RX_INT_SHIFT
);
694 /* select clk source */
695 ctrl
&= ~CTRL_MCLKSEL
;
697 /* set manual /CS mode */
700 writel(ctrl
, &pic32s
->regs
->ctrl
);
702 /* enable error reporting */
703 ctrl
= CTRL2_TX_UR_EN
| CTRL2_RX_OV_EN
| CTRL2_FRM_ERR_EN
;
704 writel(ctrl
, &pic32s
->regs
->ctrl2_set
);
707 static int pic32_spi_hw_probe(struct platform_device
*pdev
,
708 struct pic32_spi
*pic32s
)
710 struct resource
*mem
;
713 mem
= platform_get_resource(pdev
, IORESOURCE_MEM
, 0);
714 pic32s
->regs
= devm_ioremap_resource(&pdev
->dev
, mem
);
715 if (IS_ERR(pic32s
->regs
))
716 return PTR_ERR(pic32s
->regs
);
718 pic32s
->dma_base
= mem
->start
;
720 /* get irq resources: err-irq, rx-irq, tx-irq */
721 pic32s
->fault_irq
= platform_get_irq_byname(pdev
, "fault");
722 if (pic32s
->fault_irq
< 0) {
723 dev_err(&pdev
->dev
, "fault-irq not found\n");
724 return pic32s
->fault_irq
;
727 pic32s
->rx_irq
= platform_get_irq_byname(pdev
, "rx");
728 if (pic32s
->rx_irq
< 0) {
729 dev_err(&pdev
->dev
, "rx-irq not found\n");
730 return pic32s
->rx_irq
;
733 pic32s
->tx_irq
= platform_get_irq_byname(pdev
, "tx");
734 if (pic32s
->tx_irq
< 0) {
735 dev_err(&pdev
->dev
, "tx-irq not found\n");
736 return pic32s
->tx_irq
;
740 pic32s
->clk
= devm_clk_get(&pdev
->dev
, "mck0");
741 if (IS_ERR(pic32s
->clk
)) {
742 dev_err(&pdev
->dev
, "clk not found\n");
743 ret
= PTR_ERR(pic32s
->clk
);
747 ret
= clk_prepare_enable(pic32s
->clk
);
751 pic32_spi_hw_init(pic32s
);
756 dev_err(&pdev
->dev
, "%s failed, err %d\n", __func__
, ret
);
760 static int pic32_spi_probe(struct platform_device
*pdev
)
762 struct spi_master
*master
;
763 struct pic32_spi
*pic32s
;
766 master
= spi_alloc_master(&pdev
->dev
, sizeof(*pic32s
));
770 pic32s
= spi_master_get_devdata(master
);
771 pic32s
->master
= master
;
773 ret
= pic32_spi_hw_probe(pdev
, pic32s
);
777 master
->dev
.of_node
= pdev
->dev
.of_node
;
778 master
->mode_bits
= SPI_MODE_3
| SPI_MODE_0
| SPI_CS_HIGH
;
779 master
->num_chipselect
= 1; /* single chip-select */
780 master
->max_speed_hz
= clk_get_rate(pic32s
->clk
);
781 master
->setup
= pic32_spi_setup
;
782 master
->cleanup
= pic32_spi_cleanup
;
783 master
->flags
= SPI_MASTER_MUST_TX
| SPI_MASTER_MUST_RX
;
784 master
->bits_per_word_mask
= SPI_BPW_MASK(8) | SPI_BPW_MASK(16) |
786 master
->transfer_one
= pic32_spi_one_transfer
;
787 master
->prepare_message
= pic32_spi_prepare_message
;
788 master
->unprepare_message
= pic32_spi_unprepare_message
;
789 master
->prepare_transfer_hardware
= pic32_spi_prepare_hardware
;
790 master
->unprepare_transfer_hardware
= pic32_spi_unprepare_hardware
;
792 /* optional DMA support */
793 pic32_spi_dma_prep(pic32s
, &pdev
->dev
);
794 if (test_bit(PIC32F_DMA_PREP
, &pic32s
->flags
))
795 master
->can_dma
= pic32_spi_can_dma
;
797 init_completion(&pic32s
->xfer_done
);
800 /* install irq handlers (with irq-disabled) */
801 irq_set_status_flags(pic32s
->fault_irq
, IRQ_NOAUTOEN
);
802 ret
= devm_request_irq(&pdev
->dev
, pic32s
->fault_irq
,
803 pic32_spi_fault_irq
, IRQF_NO_THREAD
,
804 dev_name(&pdev
->dev
), pic32s
);
806 dev_err(&pdev
->dev
, "request fault-irq %d\n", pic32s
->rx_irq
);
810 /* receive interrupt handler */
811 irq_set_status_flags(pic32s
->rx_irq
, IRQ_NOAUTOEN
);
812 ret
= devm_request_irq(&pdev
->dev
, pic32s
->rx_irq
,
813 pic32_spi_rx_irq
, IRQF_NO_THREAD
,
814 dev_name(&pdev
->dev
), pic32s
);
816 dev_err(&pdev
->dev
, "request rx-irq %d\n", pic32s
->rx_irq
);
820 /* transmit interrupt handler */
821 irq_set_status_flags(pic32s
->tx_irq
, IRQ_NOAUTOEN
);
822 ret
= devm_request_irq(&pdev
->dev
, pic32s
->tx_irq
,
823 pic32_spi_tx_irq
, IRQF_NO_THREAD
,
824 dev_name(&pdev
->dev
), pic32s
);
826 dev_err(&pdev
->dev
, "request tx-irq %d\n", pic32s
->tx_irq
);
830 /* register master */
831 ret
= devm_spi_register_master(&pdev
->dev
, master
);
833 dev_err(&master
->dev
, "failed registering spi master\n");
837 platform_set_drvdata(pdev
, pic32s
);
842 clk_disable_unprepare(pic32s
->clk
);
844 spi_master_put(master
);
848 static int pic32_spi_remove(struct platform_device
*pdev
)
850 struct pic32_spi
*pic32s
;
852 pic32s
= platform_get_drvdata(pdev
);
853 pic32_spi_disable(pic32s
);
854 clk_disable_unprepare(pic32s
->clk
);
855 pic32_spi_dma_unprep(pic32s
);
860 static const struct of_device_id pic32_spi_of_match
[] = {
861 {.compatible
= "microchip,pic32mzda-spi",},
864 MODULE_DEVICE_TABLE(of
, pic32_spi_of_match
);
866 static struct platform_driver pic32_spi_driver
= {
869 .of_match_table
= of_match_ptr(pic32_spi_of_match
),
871 .probe
= pic32_spi_probe
,
872 .remove
= pic32_spi_remove
,
875 module_platform_driver(pic32_spi_driver
);
877 MODULE_AUTHOR("Purna Chandra Mandal <purna.mandal@microchip.com>");
878 MODULE_DESCRIPTION("Microchip SPI driver for PIC32 SPI controller.");
879 MODULE_LICENSE("GPL v2");