md-cluster/raid10: resize all the bitmaps before start reshape
[linux/fpc-iii.git] / drivers / spi / spi-lantiq-ssc.c
blobd5976615d924b7b1d5c4a930aeca4765cf01ea47
1 /*
2 * Copyright (C) 2011-2015 Daniel Schwierzeck <daniel.schwierzeck@gmail.com>
3 * Copyright (C) 2016 Hauke Mehrtens <hauke@hauke-m.de>
5 * This program is free software; you can distribute it and/or modify it
6 * under the terms of the GNU General Public License (Version 2) as
7 * published by the Free Software Foundation.
8 */
10 #include <linux/kernel.h>
11 #include <linux/module.h>
12 #include <linux/of_device.h>
13 #include <linux/clk.h>
14 #include <linux/io.h>
15 #include <linux/delay.h>
16 #include <linux/interrupt.h>
17 #include <linux/sched.h>
18 #include <linux/completion.h>
19 #include <linux/spinlock.h>
20 #include <linux/err.h>
21 #include <linux/gpio.h>
22 #include <linux/pm_runtime.h>
23 #include <linux/spi/spi.h>
25 #ifdef CONFIG_LANTIQ
26 #include <lantiq_soc.h>
27 #endif
29 #define LTQ_SPI_RX_IRQ_NAME "spi_rx"
30 #define LTQ_SPI_TX_IRQ_NAME "spi_tx"
31 #define LTQ_SPI_ERR_IRQ_NAME "spi_err"
32 #define LTQ_SPI_FRM_IRQ_NAME "spi_frm"
34 #define LTQ_SPI_CLC 0x00
35 #define LTQ_SPI_PISEL 0x04
36 #define LTQ_SPI_ID 0x08
37 #define LTQ_SPI_CON 0x10
38 #define LTQ_SPI_STAT 0x14
39 #define LTQ_SPI_WHBSTATE 0x18
40 #define LTQ_SPI_TB 0x20
41 #define LTQ_SPI_RB 0x24
42 #define LTQ_SPI_RXFCON 0x30
43 #define LTQ_SPI_TXFCON 0x34
44 #define LTQ_SPI_FSTAT 0x38
45 #define LTQ_SPI_BRT 0x40
46 #define LTQ_SPI_BRSTAT 0x44
47 #define LTQ_SPI_SFCON 0x60
48 #define LTQ_SPI_SFSTAT 0x64
49 #define LTQ_SPI_GPOCON 0x70
50 #define LTQ_SPI_GPOSTAT 0x74
51 #define LTQ_SPI_FPGO 0x78
52 #define LTQ_SPI_RXREQ 0x80
53 #define LTQ_SPI_RXCNT 0x84
54 #define LTQ_SPI_DMACON 0xec
55 #define LTQ_SPI_IRNEN 0xf4
56 #define LTQ_SPI_IRNICR 0xf8
57 #define LTQ_SPI_IRNCR 0xfc
59 #define LTQ_SPI_CLC_SMC_S 16 /* Clock divider for sleep mode */
60 #define LTQ_SPI_CLC_SMC_M (0xFF << LTQ_SPI_CLC_SMC_S)
61 #define LTQ_SPI_CLC_RMC_S 8 /* Clock divider for normal run mode */
62 #define LTQ_SPI_CLC_RMC_M (0xFF << LTQ_SPI_CLC_RMC_S)
63 #define LTQ_SPI_CLC_DISS BIT(1) /* Disable status bit */
64 #define LTQ_SPI_CLC_DISR BIT(0) /* Disable request bit */
66 #define LTQ_SPI_ID_TXFS_S 24 /* Implemented TX FIFO size */
67 #define LTQ_SPI_ID_TXFS_M (0x3F << LTQ_SPI_ID_TXFS_S)
68 #define LTQ_SPI_ID_RXFS_S 16 /* Implemented RX FIFO size */
69 #define LTQ_SPI_ID_RXFS_M (0x3F << LTQ_SPI_ID_RXFS_S)
70 #define LTQ_SPI_ID_MOD_S 8 /* Module ID */
71 #define LTQ_SPI_ID_MOD_M (0xff << LTQ_SPI_ID_MOD_S)
72 #define LTQ_SPI_ID_CFG_S 5 /* DMA interface support */
73 #define LTQ_SPI_ID_CFG_M (1 << LTQ_SPI_ID_CFG_S)
74 #define LTQ_SPI_ID_REV_M 0x1F /* Hardware revision number */
76 #define LTQ_SPI_CON_BM_S 16 /* Data width selection */
77 #define LTQ_SPI_CON_BM_M (0x1F << LTQ_SPI_CON_BM_S)
78 #define LTQ_SPI_CON_EM BIT(24) /* Echo mode */
79 #define LTQ_SPI_CON_IDLE BIT(23) /* Idle bit value */
80 #define LTQ_SPI_CON_ENBV BIT(22) /* Enable byte valid control */
81 #define LTQ_SPI_CON_RUEN BIT(12) /* Receive underflow error enable */
82 #define LTQ_SPI_CON_TUEN BIT(11) /* Transmit underflow error enable */
83 #define LTQ_SPI_CON_AEN BIT(10) /* Abort error enable */
84 #define LTQ_SPI_CON_REN BIT(9) /* Receive overflow error enable */
85 #define LTQ_SPI_CON_TEN BIT(8) /* Transmit overflow error enable */
86 #define LTQ_SPI_CON_LB BIT(7) /* Loopback control */
87 #define LTQ_SPI_CON_PO BIT(6) /* Clock polarity control */
88 #define LTQ_SPI_CON_PH BIT(5) /* Clock phase control */
89 #define LTQ_SPI_CON_HB BIT(4) /* Heading control */
90 #define LTQ_SPI_CON_RXOFF BIT(1) /* Switch receiver off */
91 #define LTQ_SPI_CON_TXOFF BIT(0) /* Switch transmitter off */
93 #define LTQ_SPI_STAT_RXBV_S 28
94 #define LTQ_SPI_STAT_RXBV_M (0x7 << LTQ_SPI_STAT_RXBV_S)
95 #define LTQ_SPI_STAT_BSY BIT(13) /* Busy flag */
96 #define LTQ_SPI_STAT_RUE BIT(12) /* Receive underflow error flag */
97 #define LTQ_SPI_STAT_TUE BIT(11) /* Transmit underflow error flag */
98 #define LTQ_SPI_STAT_AE BIT(10) /* Abort error flag */
99 #define LTQ_SPI_STAT_RE BIT(9) /* Receive error flag */
100 #define LTQ_SPI_STAT_TE BIT(8) /* Transmit error flag */
101 #define LTQ_SPI_STAT_ME BIT(7) /* Mode error flag */
102 #define LTQ_SPI_STAT_MS BIT(1) /* Master/slave select bit */
103 #define LTQ_SPI_STAT_EN BIT(0) /* Enable bit */
104 #define LTQ_SPI_STAT_ERRORS (LTQ_SPI_STAT_ME | LTQ_SPI_STAT_TE | \
105 LTQ_SPI_STAT_RE | LTQ_SPI_STAT_AE | \
106 LTQ_SPI_STAT_TUE | LTQ_SPI_STAT_RUE)
108 #define LTQ_SPI_WHBSTATE_SETTUE BIT(15) /* Set transmit underflow error flag */
109 #define LTQ_SPI_WHBSTATE_SETAE BIT(14) /* Set abort error flag */
110 #define LTQ_SPI_WHBSTATE_SETRE BIT(13) /* Set receive error flag */
111 #define LTQ_SPI_WHBSTATE_SETTE BIT(12) /* Set transmit error flag */
112 #define LTQ_SPI_WHBSTATE_CLRTUE BIT(11) /* Clear transmit underflow error flag */
113 #define LTQ_SPI_WHBSTATE_CLRAE BIT(10) /* Clear abort error flag */
114 #define LTQ_SPI_WHBSTATE_CLRRE BIT(9) /* Clear receive error flag */
115 #define LTQ_SPI_WHBSTATE_CLRTE BIT(8) /* Clear transmit error flag */
116 #define LTQ_SPI_WHBSTATE_SETME BIT(7) /* Set mode error flag */
117 #define LTQ_SPI_WHBSTATE_CLRME BIT(6) /* Clear mode error flag */
118 #define LTQ_SPI_WHBSTATE_SETRUE BIT(5) /* Set receive underflow error flag */
119 #define LTQ_SPI_WHBSTATE_CLRRUE BIT(4) /* Clear receive underflow error flag */
120 #define LTQ_SPI_WHBSTATE_SETMS BIT(3) /* Set master select bit */
121 #define LTQ_SPI_WHBSTATE_CLRMS BIT(2) /* Clear master select bit */
122 #define LTQ_SPI_WHBSTATE_SETEN BIT(1) /* Set enable bit (operational mode) */
123 #define LTQ_SPI_WHBSTATE_CLREN BIT(0) /* Clear enable bit (config mode */
124 #define LTQ_SPI_WHBSTATE_CLR_ERRORS (LTQ_SPI_WHBSTATE_CLRRUE | \
125 LTQ_SPI_WHBSTATE_CLRME | \
126 LTQ_SPI_WHBSTATE_CLRTE | \
127 LTQ_SPI_WHBSTATE_CLRRE | \
128 LTQ_SPI_WHBSTATE_CLRAE | \
129 LTQ_SPI_WHBSTATE_CLRTUE)
131 #define LTQ_SPI_RXFCON_RXFITL_S 8 /* FIFO interrupt trigger level */
132 #define LTQ_SPI_RXFCON_RXFITL_M (0x3F << LTQ_SPI_RXFCON_RXFITL_S)
133 #define LTQ_SPI_RXFCON_RXFLU BIT(1) /* FIFO flush */
134 #define LTQ_SPI_RXFCON_RXFEN BIT(0) /* FIFO enable */
136 #define LTQ_SPI_TXFCON_TXFITL_S 8 /* FIFO interrupt trigger level */
137 #define LTQ_SPI_TXFCON_TXFITL_M (0x3F << LTQ_SPI_TXFCON_TXFITL_S)
138 #define LTQ_SPI_TXFCON_TXFLU BIT(1) /* FIFO flush */
139 #define LTQ_SPI_TXFCON_TXFEN BIT(0) /* FIFO enable */
141 #define LTQ_SPI_FSTAT_RXFFL_S 0
142 #define LTQ_SPI_FSTAT_RXFFL_M (0x3f << LTQ_SPI_FSTAT_RXFFL_S)
143 #define LTQ_SPI_FSTAT_TXFFL_S 8
144 #define LTQ_SPI_FSTAT_TXFFL_M (0x3f << LTQ_SPI_FSTAT_TXFFL_S)
146 #define LTQ_SPI_GPOCON_ISCSBN_S 8
147 #define LTQ_SPI_GPOCON_INVOUTN_S 0
149 #define LTQ_SPI_FGPO_SETOUTN_S 8
150 #define LTQ_SPI_FGPO_CLROUTN_S 0
152 #define LTQ_SPI_RXREQ_RXCNT_M 0xFFFF /* Receive count value */
153 #define LTQ_SPI_RXCNT_TODO_M 0xFFFF /* Recevie to-do value */
155 #define LTQ_SPI_IRNEN_TFI BIT(4) /* TX finished interrupt */
156 #define LTQ_SPI_IRNEN_F BIT(3) /* Frame end interrupt request */
157 #define LTQ_SPI_IRNEN_E BIT(2) /* Error end interrupt request */
158 #define LTQ_SPI_IRNEN_T_XWAY BIT(1) /* Transmit end interrupt request */
159 #define LTQ_SPI_IRNEN_R_XWAY BIT(0) /* Receive end interrupt request */
160 #define LTQ_SPI_IRNEN_R_XRX BIT(1) /* Transmit end interrupt request */
161 #define LTQ_SPI_IRNEN_T_XRX BIT(0) /* Receive end interrupt request */
162 #define LTQ_SPI_IRNEN_ALL 0x1F
164 struct lantiq_ssc_hwcfg {
165 unsigned int irnen_r;
166 unsigned int irnen_t;
169 struct lantiq_ssc_spi {
170 struct spi_master *master;
171 struct device *dev;
172 void __iomem *regbase;
173 struct clk *spi_clk;
174 struct clk *fpi_clk;
175 const struct lantiq_ssc_hwcfg *hwcfg;
177 spinlock_t lock;
178 struct workqueue_struct *wq;
179 struct work_struct work;
181 const u8 *tx;
182 u8 *rx;
183 unsigned int tx_todo;
184 unsigned int rx_todo;
185 unsigned int bits_per_word;
186 unsigned int speed_hz;
187 unsigned int tx_fifo_size;
188 unsigned int rx_fifo_size;
189 unsigned int base_cs;
192 static u32 lantiq_ssc_readl(const struct lantiq_ssc_spi *spi, u32 reg)
194 return __raw_readl(spi->regbase + reg);
197 static void lantiq_ssc_writel(const struct lantiq_ssc_spi *spi, u32 val,
198 u32 reg)
200 __raw_writel(val, spi->regbase + reg);
203 static void lantiq_ssc_maskl(const struct lantiq_ssc_spi *spi, u32 clr,
204 u32 set, u32 reg)
206 u32 val = __raw_readl(spi->regbase + reg);
208 val &= ~clr;
209 val |= set;
210 __raw_writel(val, spi->regbase + reg);
213 static unsigned int tx_fifo_level(const struct lantiq_ssc_spi *spi)
215 u32 fstat = lantiq_ssc_readl(spi, LTQ_SPI_FSTAT);
217 return (fstat & LTQ_SPI_FSTAT_TXFFL_M) >> LTQ_SPI_FSTAT_TXFFL_S;
220 static unsigned int rx_fifo_level(const struct lantiq_ssc_spi *spi)
222 u32 fstat = lantiq_ssc_readl(spi, LTQ_SPI_FSTAT);
224 return fstat & LTQ_SPI_FSTAT_RXFFL_M;
227 static unsigned int tx_fifo_free(const struct lantiq_ssc_spi *spi)
229 return spi->tx_fifo_size - tx_fifo_level(spi);
232 static void rx_fifo_reset(const struct lantiq_ssc_spi *spi)
234 u32 val = spi->rx_fifo_size << LTQ_SPI_RXFCON_RXFITL_S;
236 val |= LTQ_SPI_RXFCON_RXFEN | LTQ_SPI_RXFCON_RXFLU;
237 lantiq_ssc_writel(spi, val, LTQ_SPI_RXFCON);
240 static void tx_fifo_reset(const struct lantiq_ssc_spi *spi)
242 u32 val = 1 << LTQ_SPI_TXFCON_TXFITL_S;
244 val |= LTQ_SPI_TXFCON_TXFEN | LTQ_SPI_TXFCON_TXFLU;
245 lantiq_ssc_writel(spi, val, LTQ_SPI_TXFCON);
248 static void rx_fifo_flush(const struct lantiq_ssc_spi *spi)
250 lantiq_ssc_maskl(spi, 0, LTQ_SPI_RXFCON_RXFLU, LTQ_SPI_RXFCON);
253 static void tx_fifo_flush(const struct lantiq_ssc_spi *spi)
255 lantiq_ssc_maskl(spi, 0, LTQ_SPI_TXFCON_TXFLU, LTQ_SPI_TXFCON);
258 static void hw_enter_config_mode(const struct lantiq_ssc_spi *spi)
260 lantiq_ssc_writel(spi, LTQ_SPI_WHBSTATE_CLREN, LTQ_SPI_WHBSTATE);
263 static void hw_enter_active_mode(const struct lantiq_ssc_spi *spi)
265 lantiq_ssc_writel(spi, LTQ_SPI_WHBSTATE_SETEN, LTQ_SPI_WHBSTATE);
268 static void hw_setup_speed_hz(const struct lantiq_ssc_spi *spi,
269 unsigned int max_speed_hz)
271 u32 spi_clk, brt;
274 * SPI module clock is derived from FPI bus clock dependent on
275 * divider value in CLC.RMS which is always set to 1.
277 * f_SPI
278 * baudrate = --------------
279 * 2 * (BR + 1)
281 spi_clk = clk_get_rate(spi->fpi_clk) / 2;
283 if (max_speed_hz > spi_clk)
284 brt = 0;
285 else
286 brt = spi_clk / max_speed_hz - 1;
288 if (brt > 0xFFFF)
289 brt = 0xFFFF;
291 dev_dbg(spi->dev, "spi_clk %u, max_speed_hz %u, brt %u\n",
292 spi_clk, max_speed_hz, brt);
294 lantiq_ssc_writel(spi, brt, LTQ_SPI_BRT);
297 static void hw_setup_bits_per_word(const struct lantiq_ssc_spi *spi,
298 unsigned int bits_per_word)
300 u32 bm;
302 /* CON.BM value = bits_per_word - 1 */
303 bm = (bits_per_word - 1) << LTQ_SPI_CON_BM_S;
305 lantiq_ssc_maskl(spi, LTQ_SPI_CON_BM_M, bm, LTQ_SPI_CON);
308 static void hw_setup_clock_mode(const struct lantiq_ssc_spi *spi,
309 unsigned int mode)
311 u32 con_set = 0, con_clr = 0;
314 * SPI mode mapping in CON register:
315 * Mode CPOL CPHA CON.PO CON.PH
316 * 0 0 0 0 1
317 * 1 0 1 0 0
318 * 2 1 0 1 1
319 * 3 1 1 1 0
321 if (mode & SPI_CPHA)
322 con_clr |= LTQ_SPI_CON_PH;
323 else
324 con_set |= LTQ_SPI_CON_PH;
326 if (mode & SPI_CPOL)
327 con_set |= LTQ_SPI_CON_PO | LTQ_SPI_CON_IDLE;
328 else
329 con_clr |= LTQ_SPI_CON_PO | LTQ_SPI_CON_IDLE;
331 /* Set heading control */
332 if (mode & SPI_LSB_FIRST)
333 con_clr |= LTQ_SPI_CON_HB;
334 else
335 con_set |= LTQ_SPI_CON_HB;
337 /* Set loopback mode */
338 if (mode & SPI_LOOP)
339 con_set |= LTQ_SPI_CON_LB;
340 else
341 con_clr |= LTQ_SPI_CON_LB;
343 lantiq_ssc_maskl(spi, con_clr, con_set, LTQ_SPI_CON);
346 static void lantiq_ssc_hw_init(const struct lantiq_ssc_spi *spi)
348 const struct lantiq_ssc_hwcfg *hwcfg = spi->hwcfg;
351 * Set clock divider for run mode to 1 to
352 * run at same frequency as FPI bus
354 lantiq_ssc_writel(spi, 1 << LTQ_SPI_CLC_RMC_S, LTQ_SPI_CLC);
356 /* Put controller into config mode */
357 hw_enter_config_mode(spi);
359 /* Clear error flags */
360 lantiq_ssc_maskl(spi, 0, LTQ_SPI_WHBSTATE_CLR_ERRORS, LTQ_SPI_WHBSTATE);
362 /* Enable error checking, disable TX/RX */
363 lantiq_ssc_writel(spi, LTQ_SPI_CON_RUEN | LTQ_SPI_CON_AEN |
364 LTQ_SPI_CON_TEN | LTQ_SPI_CON_REN | LTQ_SPI_CON_TXOFF |
365 LTQ_SPI_CON_RXOFF, LTQ_SPI_CON);
367 /* Setup default SPI mode */
368 hw_setup_bits_per_word(spi, spi->bits_per_word);
369 hw_setup_clock_mode(spi, SPI_MODE_0);
371 /* Enable master mode and clear error flags */
372 lantiq_ssc_writel(spi, LTQ_SPI_WHBSTATE_SETMS |
373 LTQ_SPI_WHBSTATE_CLR_ERRORS,
374 LTQ_SPI_WHBSTATE);
376 /* Reset GPIO/CS registers */
377 lantiq_ssc_writel(spi, 0, LTQ_SPI_GPOCON);
378 lantiq_ssc_writel(spi, 0xFF00, LTQ_SPI_FPGO);
380 /* Enable and flush FIFOs */
381 rx_fifo_reset(spi);
382 tx_fifo_reset(spi);
384 /* Enable interrupts */
385 lantiq_ssc_writel(spi, hwcfg->irnen_t | hwcfg->irnen_r |
386 LTQ_SPI_IRNEN_E, LTQ_SPI_IRNEN);
389 static int lantiq_ssc_setup(struct spi_device *spidev)
391 struct spi_master *master = spidev->master;
392 struct lantiq_ssc_spi *spi = spi_master_get_devdata(master);
393 unsigned int cs = spidev->chip_select;
394 u32 gpocon;
396 /* GPIOs are used for CS */
397 if (gpio_is_valid(spidev->cs_gpio))
398 return 0;
400 dev_dbg(spi->dev, "using internal chipselect %u\n", cs);
402 if (cs < spi->base_cs) {
403 dev_err(spi->dev,
404 "chipselect %i too small (min %i)\n", cs, spi->base_cs);
405 return -EINVAL;
408 /* set GPO pin to CS mode */
409 gpocon = 1 << ((cs - spi->base_cs) + LTQ_SPI_GPOCON_ISCSBN_S);
411 /* invert GPO pin */
412 if (spidev->mode & SPI_CS_HIGH)
413 gpocon |= 1 << (cs - spi->base_cs);
415 lantiq_ssc_maskl(spi, 0, gpocon, LTQ_SPI_GPOCON);
417 return 0;
420 static int lantiq_ssc_prepare_message(struct spi_master *master,
421 struct spi_message *message)
423 struct lantiq_ssc_spi *spi = spi_master_get_devdata(master);
425 hw_enter_config_mode(spi);
426 hw_setup_clock_mode(spi, message->spi->mode);
427 hw_enter_active_mode(spi);
429 return 0;
432 static void hw_setup_transfer(struct lantiq_ssc_spi *spi,
433 struct spi_device *spidev, struct spi_transfer *t)
435 unsigned int speed_hz = t->speed_hz;
436 unsigned int bits_per_word = t->bits_per_word;
437 u32 con;
439 if (bits_per_word != spi->bits_per_word ||
440 speed_hz != spi->speed_hz) {
441 hw_enter_config_mode(spi);
442 hw_setup_speed_hz(spi, speed_hz);
443 hw_setup_bits_per_word(spi, bits_per_word);
444 hw_enter_active_mode(spi);
446 spi->speed_hz = speed_hz;
447 spi->bits_per_word = bits_per_word;
450 /* Configure transmitter and receiver */
451 con = lantiq_ssc_readl(spi, LTQ_SPI_CON);
452 if (t->tx_buf)
453 con &= ~LTQ_SPI_CON_TXOFF;
454 else
455 con |= LTQ_SPI_CON_TXOFF;
457 if (t->rx_buf)
458 con &= ~LTQ_SPI_CON_RXOFF;
459 else
460 con |= LTQ_SPI_CON_RXOFF;
462 lantiq_ssc_writel(spi, con, LTQ_SPI_CON);
465 static int lantiq_ssc_unprepare_message(struct spi_master *master,
466 struct spi_message *message)
468 struct lantiq_ssc_spi *spi = spi_master_get_devdata(master);
470 flush_workqueue(spi->wq);
472 /* Disable transmitter and receiver while idle */
473 lantiq_ssc_maskl(spi, 0, LTQ_SPI_CON_TXOFF | LTQ_SPI_CON_RXOFF,
474 LTQ_SPI_CON);
476 return 0;
479 static void tx_fifo_write(struct lantiq_ssc_spi *spi)
481 const u8 *tx8;
482 const u16 *tx16;
483 const u32 *tx32;
484 u32 data;
485 unsigned int tx_free = tx_fifo_free(spi);
487 while (spi->tx_todo && tx_free) {
488 switch (spi->bits_per_word) {
489 case 2 ... 8:
490 tx8 = spi->tx;
491 data = *tx8;
492 spi->tx_todo--;
493 spi->tx++;
494 break;
495 case 16:
496 tx16 = (u16 *) spi->tx;
497 data = *tx16;
498 spi->tx_todo -= 2;
499 spi->tx += 2;
500 break;
501 case 32:
502 tx32 = (u32 *) spi->tx;
503 data = *tx32;
504 spi->tx_todo -= 4;
505 spi->tx += 4;
506 break;
507 default:
508 WARN_ON(1);
509 data = 0;
510 break;
513 lantiq_ssc_writel(spi, data, LTQ_SPI_TB);
514 tx_free--;
518 static void rx_fifo_read_full_duplex(struct lantiq_ssc_spi *spi)
520 u8 *rx8;
521 u16 *rx16;
522 u32 *rx32;
523 u32 data;
524 unsigned int rx_fill = rx_fifo_level(spi);
526 while (rx_fill) {
527 data = lantiq_ssc_readl(spi, LTQ_SPI_RB);
529 switch (spi->bits_per_word) {
530 case 2 ... 8:
531 rx8 = spi->rx;
532 *rx8 = data;
533 spi->rx_todo--;
534 spi->rx++;
535 break;
536 case 16:
537 rx16 = (u16 *) spi->rx;
538 *rx16 = data;
539 spi->rx_todo -= 2;
540 spi->rx += 2;
541 break;
542 case 32:
543 rx32 = (u32 *) spi->rx;
544 *rx32 = data;
545 spi->rx_todo -= 4;
546 spi->rx += 4;
547 break;
548 default:
549 WARN_ON(1);
550 break;
553 rx_fill--;
557 static void rx_fifo_read_half_duplex(struct lantiq_ssc_spi *spi)
559 u32 data, *rx32;
560 u8 *rx8;
561 unsigned int rxbv, shift;
562 unsigned int rx_fill = rx_fifo_level(spi);
565 * In RX-only mode the bits per word value is ignored by HW. A value
566 * of 32 is used instead. Thus all 4 bytes per FIFO must be read.
567 * If remaining RX bytes are less than 4, the FIFO must be read
568 * differently. The amount of received and valid bytes is indicated
569 * by STAT.RXBV register value.
571 while (rx_fill) {
572 if (spi->rx_todo < 4) {
573 rxbv = (lantiq_ssc_readl(spi, LTQ_SPI_STAT) &
574 LTQ_SPI_STAT_RXBV_M) >> LTQ_SPI_STAT_RXBV_S;
575 data = lantiq_ssc_readl(spi, LTQ_SPI_RB);
577 shift = (rxbv - 1) * 8;
578 rx8 = spi->rx;
580 while (rxbv) {
581 *rx8++ = (data >> shift) & 0xFF;
582 rxbv--;
583 shift -= 8;
584 spi->rx_todo--;
585 spi->rx++;
587 } else {
588 data = lantiq_ssc_readl(spi, LTQ_SPI_RB);
589 rx32 = (u32 *) spi->rx;
591 *rx32++ = data;
592 spi->rx_todo -= 4;
593 spi->rx += 4;
595 rx_fill--;
599 static void rx_request(struct lantiq_ssc_spi *spi)
601 unsigned int rxreq, rxreq_max;
604 * To avoid receive overflows at high clocks it is better to request
605 * only the amount of bytes that fits into all FIFOs. This value
606 * depends on the FIFO size implemented in hardware.
608 rxreq = spi->rx_todo;
609 rxreq_max = spi->rx_fifo_size * 4;
610 if (rxreq > rxreq_max)
611 rxreq = rxreq_max;
613 lantiq_ssc_writel(spi, rxreq, LTQ_SPI_RXREQ);
616 static irqreturn_t lantiq_ssc_xmit_interrupt(int irq, void *data)
618 struct lantiq_ssc_spi *spi = data;
620 if (spi->tx) {
621 if (spi->rx && spi->rx_todo)
622 rx_fifo_read_full_duplex(spi);
624 if (spi->tx_todo)
625 tx_fifo_write(spi);
626 else if (!tx_fifo_level(spi))
627 goto completed;
628 } else if (spi->rx) {
629 if (spi->rx_todo) {
630 rx_fifo_read_half_duplex(spi);
632 if (spi->rx_todo)
633 rx_request(spi);
634 else
635 goto completed;
636 } else {
637 goto completed;
641 return IRQ_HANDLED;
643 completed:
644 queue_work(spi->wq, &spi->work);
646 return IRQ_HANDLED;
649 static irqreturn_t lantiq_ssc_err_interrupt(int irq, void *data)
651 struct lantiq_ssc_spi *spi = data;
652 u32 stat = lantiq_ssc_readl(spi, LTQ_SPI_STAT);
654 if (!(stat & LTQ_SPI_STAT_ERRORS))
655 return IRQ_NONE;
657 if (stat & LTQ_SPI_STAT_RUE)
658 dev_err(spi->dev, "receive underflow error\n");
659 if (stat & LTQ_SPI_STAT_TUE)
660 dev_err(spi->dev, "transmit underflow error\n");
661 if (stat & LTQ_SPI_STAT_AE)
662 dev_err(spi->dev, "abort error\n");
663 if (stat & LTQ_SPI_STAT_RE)
664 dev_err(spi->dev, "receive overflow error\n");
665 if (stat & LTQ_SPI_STAT_TE)
666 dev_err(spi->dev, "transmit overflow error\n");
667 if (stat & LTQ_SPI_STAT_ME)
668 dev_err(spi->dev, "mode error\n");
670 /* Clear error flags */
671 lantiq_ssc_maskl(spi, 0, LTQ_SPI_WHBSTATE_CLR_ERRORS, LTQ_SPI_WHBSTATE);
673 /* set bad status so it can be retried */
674 if (spi->master->cur_msg)
675 spi->master->cur_msg->status = -EIO;
676 queue_work(spi->wq, &spi->work);
678 return IRQ_HANDLED;
681 static int transfer_start(struct lantiq_ssc_spi *spi, struct spi_device *spidev,
682 struct spi_transfer *t)
684 unsigned long flags;
686 spin_lock_irqsave(&spi->lock, flags);
688 spi->tx = t->tx_buf;
689 spi->rx = t->rx_buf;
691 if (t->tx_buf) {
692 spi->tx_todo = t->len;
694 /* initially fill TX FIFO */
695 tx_fifo_write(spi);
698 if (spi->rx) {
699 spi->rx_todo = t->len;
701 /* start shift clock in RX-only mode */
702 if (!spi->tx)
703 rx_request(spi);
706 spin_unlock_irqrestore(&spi->lock, flags);
708 return t->len;
712 * The driver only gets an interrupt when the FIFO is empty, but there
713 * is an additional shift register from which the data is written to
714 * the wire. We get the last interrupt when the controller starts to
715 * write the last word to the wire, not when it is finished. Do busy
716 * waiting till it finishes.
718 static void lantiq_ssc_bussy_work(struct work_struct *work)
720 struct lantiq_ssc_spi *spi;
721 unsigned long long timeout = 8LL * 1000LL;
722 unsigned long end;
724 spi = container_of(work, typeof(*spi), work);
726 do_div(timeout, spi->speed_hz);
727 timeout += timeout + 100; /* some tolerance */
729 end = jiffies + msecs_to_jiffies(timeout);
730 do {
731 u32 stat = lantiq_ssc_readl(spi, LTQ_SPI_STAT);
733 if (!(stat & LTQ_SPI_STAT_BSY)) {
734 spi_finalize_current_transfer(spi->master);
735 return;
738 cond_resched();
739 } while (!time_after_eq(jiffies, end));
741 if (spi->master->cur_msg)
742 spi->master->cur_msg->status = -EIO;
743 spi_finalize_current_transfer(spi->master);
746 static void lantiq_ssc_handle_err(struct spi_master *master,
747 struct spi_message *message)
749 struct lantiq_ssc_spi *spi = spi_master_get_devdata(master);
751 /* flush FIFOs on timeout */
752 rx_fifo_flush(spi);
753 tx_fifo_flush(spi);
756 static void lantiq_ssc_set_cs(struct spi_device *spidev, bool enable)
758 struct lantiq_ssc_spi *spi = spi_master_get_devdata(spidev->master);
759 unsigned int cs = spidev->chip_select;
760 u32 fgpo;
762 if (!!(spidev->mode & SPI_CS_HIGH) == enable)
763 fgpo = (1 << (cs - spi->base_cs));
764 else
765 fgpo = (1 << (cs - spi->base_cs + LTQ_SPI_FGPO_SETOUTN_S));
767 lantiq_ssc_writel(spi, fgpo, LTQ_SPI_FPGO);
770 static int lantiq_ssc_transfer_one(struct spi_master *master,
771 struct spi_device *spidev,
772 struct spi_transfer *t)
774 struct lantiq_ssc_spi *spi = spi_master_get_devdata(master);
776 hw_setup_transfer(spi, spidev, t);
778 return transfer_start(spi, spidev, t);
781 static const struct lantiq_ssc_hwcfg lantiq_ssc_xway = {
782 .irnen_r = LTQ_SPI_IRNEN_R_XWAY,
783 .irnen_t = LTQ_SPI_IRNEN_T_XWAY,
786 static const struct lantiq_ssc_hwcfg lantiq_ssc_xrx = {
787 .irnen_r = LTQ_SPI_IRNEN_R_XRX,
788 .irnen_t = LTQ_SPI_IRNEN_T_XRX,
791 static const struct of_device_id lantiq_ssc_match[] = {
792 { .compatible = "lantiq,ase-spi", .data = &lantiq_ssc_xway, },
793 { .compatible = "lantiq,falcon-spi", .data = &lantiq_ssc_xrx, },
794 { .compatible = "lantiq,xrx100-spi", .data = &lantiq_ssc_xrx, },
797 MODULE_DEVICE_TABLE(of, lantiq_ssc_match);
799 static int lantiq_ssc_probe(struct platform_device *pdev)
801 struct device *dev = &pdev->dev;
802 struct spi_master *master;
803 struct resource *res;
804 struct lantiq_ssc_spi *spi;
805 const struct lantiq_ssc_hwcfg *hwcfg;
806 const struct of_device_id *match;
807 int err, rx_irq, tx_irq, err_irq;
808 u32 id, supports_dma, revision;
809 unsigned int num_cs;
811 match = of_match_device(lantiq_ssc_match, dev);
812 if (!match) {
813 dev_err(dev, "no device match\n");
814 return -EINVAL;
816 hwcfg = match->data;
818 res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
819 if (!res) {
820 dev_err(dev, "failed to get resources\n");
821 return -ENXIO;
824 rx_irq = platform_get_irq_byname(pdev, LTQ_SPI_RX_IRQ_NAME);
825 if (rx_irq < 0) {
826 dev_err(dev, "failed to get %s\n", LTQ_SPI_RX_IRQ_NAME);
827 return -ENXIO;
830 tx_irq = platform_get_irq_byname(pdev, LTQ_SPI_TX_IRQ_NAME);
831 if (tx_irq < 0) {
832 dev_err(dev, "failed to get %s\n", LTQ_SPI_TX_IRQ_NAME);
833 return -ENXIO;
836 err_irq = platform_get_irq_byname(pdev, LTQ_SPI_ERR_IRQ_NAME);
837 if (err_irq < 0) {
838 dev_err(dev, "failed to get %s\n", LTQ_SPI_ERR_IRQ_NAME);
839 return -ENXIO;
842 master = spi_alloc_master(dev, sizeof(struct lantiq_ssc_spi));
843 if (!master)
844 return -ENOMEM;
846 spi = spi_master_get_devdata(master);
847 spi->master = master;
848 spi->dev = dev;
849 spi->hwcfg = hwcfg;
850 platform_set_drvdata(pdev, spi);
852 spi->regbase = devm_ioremap_resource(dev, res);
853 if (IS_ERR(spi->regbase)) {
854 err = PTR_ERR(spi->regbase);
855 goto err_master_put;
858 err = devm_request_irq(dev, rx_irq, lantiq_ssc_xmit_interrupt,
859 0, LTQ_SPI_RX_IRQ_NAME, spi);
860 if (err)
861 goto err_master_put;
863 err = devm_request_irq(dev, tx_irq, lantiq_ssc_xmit_interrupt,
864 0, LTQ_SPI_TX_IRQ_NAME, spi);
865 if (err)
866 goto err_master_put;
868 err = devm_request_irq(dev, err_irq, lantiq_ssc_err_interrupt,
869 0, LTQ_SPI_ERR_IRQ_NAME, spi);
870 if (err)
871 goto err_master_put;
873 spi->spi_clk = devm_clk_get(dev, "gate");
874 if (IS_ERR(spi->spi_clk)) {
875 err = PTR_ERR(spi->spi_clk);
876 goto err_master_put;
878 err = clk_prepare_enable(spi->spi_clk);
879 if (err)
880 goto err_master_put;
883 * Use the old clk_get_fpi() function on Lantiq platform, till it
884 * supports common clk.
886 #if defined(CONFIG_LANTIQ) && !defined(CONFIG_COMMON_CLK)
887 spi->fpi_clk = clk_get_fpi();
888 #else
889 spi->fpi_clk = clk_get(dev, "freq");
890 #endif
891 if (IS_ERR(spi->fpi_clk)) {
892 err = PTR_ERR(spi->fpi_clk);
893 goto err_clk_disable;
896 num_cs = 8;
897 of_property_read_u32(pdev->dev.of_node, "num-cs", &num_cs);
899 spi->base_cs = 1;
900 of_property_read_u32(pdev->dev.of_node, "base-cs", &spi->base_cs);
902 spin_lock_init(&spi->lock);
903 spi->bits_per_word = 8;
904 spi->speed_hz = 0;
906 master->dev.of_node = pdev->dev.of_node;
907 master->num_chipselect = num_cs;
908 master->setup = lantiq_ssc_setup;
909 master->set_cs = lantiq_ssc_set_cs;
910 master->handle_err = lantiq_ssc_handle_err;
911 master->prepare_message = lantiq_ssc_prepare_message;
912 master->unprepare_message = lantiq_ssc_unprepare_message;
913 master->transfer_one = lantiq_ssc_transfer_one;
914 master->mode_bits = SPI_CPOL | SPI_CPHA | SPI_LSB_FIRST | SPI_CS_HIGH |
915 SPI_LOOP;
916 master->bits_per_word_mask = SPI_BPW_RANGE_MASK(2, 8) |
917 SPI_BPW_MASK(16) | SPI_BPW_MASK(32);
919 spi->wq = alloc_ordered_workqueue(dev_name(dev), 0);
920 if (!spi->wq) {
921 err = -ENOMEM;
922 goto err_clk_put;
924 INIT_WORK(&spi->work, lantiq_ssc_bussy_work);
926 id = lantiq_ssc_readl(spi, LTQ_SPI_ID);
927 spi->tx_fifo_size = (id & LTQ_SPI_ID_TXFS_M) >> LTQ_SPI_ID_TXFS_S;
928 spi->rx_fifo_size = (id & LTQ_SPI_ID_RXFS_M) >> LTQ_SPI_ID_RXFS_S;
929 supports_dma = (id & LTQ_SPI_ID_CFG_M) >> LTQ_SPI_ID_CFG_S;
930 revision = id & LTQ_SPI_ID_REV_M;
932 lantiq_ssc_hw_init(spi);
934 dev_info(dev,
935 "Lantiq SSC SPI controller (Rev %i, TXFS %u, RXFS %u, DMA %u)\n",
936 revision, spi->tx_fifo_size, spi->rx_fifo_size, supports_dma);
938 err = devm_spi_register_master(dev, master);
939 if (err) {
940 dev_err(dev, "failed to register spi_master\n");
941 goto err_wq_destroy;
944 return 0;
946 err_wq_destroy:
947 destroy_workqueue(spi->wq);
948 err_clk_put:
949 clk_put(spi->fpi_clk);
950 err_clk_disable:
951 clk_disable_unprepare(spi->spi_clk);
952 err_master_put:
953 spi_master_put(master);
955 return err;
958 static int lantiq_ssc_remove(struct platform_device *pdev)
960 struct lantiq_ssc_spi *spi = platform_get_drvdata(pdev);
962 lantiq_ssc_writel(spi, 0, LTQ_SPI_IRNEN);
963 lantiq_ssc_writel(spi, 0, LTQ_SPI_CLC);
964 rx_fifo_flush(spi);
965 tx_fifo_flush(spi);
966 hw_enter_config_mode(spi);
968 destroy_workqueue(spi->wq);
969 clk_disable_unprepare(spi->spi_clk);
970 clk_put(spi->fpi_clk);
972 return 0;
975 static struct platform_driver lantiq_ssc_driver = {
976 .probe = lantiq_ssc_probe,
977 .remove = lantiq_ssc_remove,
978 .driver = {
979 .name = "spi-lantiq-ssc",
980 .of_match_table = lantiq_ssc_match,
983 module_platform_driver(lantiq_ssc_driver);
985 MODULE_DESCRIPTION("Lantiq SSC SPI controller driver");
986 MODULE_AUTHOR("Daniel Schwierzeck <daniel.schwierzeck@gmail.com>");
987 MODULE_AUTHOR("Hauke Mehrtens <hauke@hauke-m.de>");
988 MODULE_LICENSE("GPL");
989 MODULE_ALIAS("platform:spi-lantiq-ssc");