rtc: stm32: fix misspelling and misalignment issues
[linux/fpc-iii.git] / drivers / net / ethernet / apm / xgene / xgene_enet_xgmac.c
blobb7d75d067c7af3252adb45e6762afefdba9b00ab
1 /* Applied Micro X-Gene SoC Ethernet Driver
3 * Copyright (c) 2014, Applied Micro Circuits Corporation
4 * Authors: Iyappan Subramanian <isubramanian@apm.com>
5 * Keyur Chudgar <kchudgar@apm.com>
7 * This program is free software; you can redistribute it and/or modify it
8 * under the terms of the GNU General Public License as published by the
9 * Free Software Foundation; either version 2 of the License, or (at your
10 * option) any later version.
12 * This program is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
15 * GNU General Public License for more details.
17 * You should have received a copy of the GNU General Public License
18 * along with this program. If not, see <http://www.gnu.org/licenses/>.
21 #include <linux/of_gpio.h>
22 #include <linux/gpio.h>
23 #include "xgene_enet_main.h"
24 #include "xgene_enet_hw.h"
25 #include "xgene_enet_xgmac.h"
27 static void xgene_enet_wr_csr(struct xgene_enet_pdata *pdata,
28 u32 offset, u32 val)
30 void __iomem *addr = pdata->eth_csr_addr + offset;
32 iowrite32(val, addr);
35 static void xgene_enet_wr_ring_if(struct xgene_enet_pdata *pdata,
36 u32 offset, u32 val)
38 void __iomem *addr = pdata->eth_ring_if_addr + offset;
40 iowrite32(val, addr);
43 static void xgene_enet_wr_diag_csr(struct xgene_enet_pdata *pdata,
44 u32 offset, u32 val)
46 void __iomem *addr = pdata->eth_diag_csr_addr + offset;
48 iowrite32(val, addr);
51 static bool xgene_enet_wr_indirect(void __iomem *addr, void __iomem *wr,
52 void __iomem *cmd, void __iomem *cmd_done,
53 u32 wr_addr, u32 wr_data)
55 u32 done;
56 u8 wait = 10;
58 iowrite32(wr_addr, addr);
59 iowrite32(wr_data, wr);
60 iowrite32(XGENE_ENET_WR_CMD, cmd);
62 /* wait for write command to complete */
63 while (!(done = ioread32(cmd_done)) && wait--)
64 udelay(1);
66 if (!done)
67 return false;
69 iowrite32(0, cmd);
71 return true;
74 static void xgene_enet_wr_pcs(struct xgene_enet_pdata *pdata,
75 u32 wr_addr, u32 wr_data)
77 void __iomem *addr, *wr, *cmd, *cmd_done;
79 addr = pdata->pcs_addr + PCS_ADDR_REG_OFFSET;
80 wr = pdata->pcs_addr + PCS_WRITE_REG_OFFSET;
81 cmd = pdata->pcs_addr + PCS_COMMAND_REG_OFFSET;
82 cmd_done = pdata->pcs_addr + PCS_COMMAND_DONE_REG_OFFSET;
84 if (!xgene_enet_wr_indirect(addr, wr, cmd, cmd_done, wr_addr, wr_data))
85 netdev_err(pdata->ndev, "PCS write failed, addr: %04x\n",
86 wr_addr);
89 static void xgene_enet_wr_axg_csr(struct xgene_enet_pdata *pdata,
90 u32 offset, u32 val)
92 void __iomem *addr = pdata->mcx_mac_csr_addr + offset;
94 iowrite32(val, addr);
97 static void xgene_enet_rd_csr(struct xgene_enet_pdata *pdata,
98 u32 offset, u32 *val)
100 void __iomem *addr = pdata->eth_csr_addr + offset;
102 *val = ioread32(addr);
105 static void xgene_enet_rd_diag_csr(struct xgene_enet_pdata *pdata,
106 u32 offset, u32 *val)
108 void __iomem *addr = pdata->eth_diag_csr_addr + offset;
110 *val = ioread32(addr);
113 static bool xgene_enet_rd_indirect(void __iomem *addr, void __iomem *rd,
114 void __iomem *cmd, void __iomem *cmd_done,
115 u32 rd_addr, u32 *rd_data)
117 u32 done;
118 u8 wait = 10;
120 iowrite32(rd_addr, addr);
121 iowrite32(XGENE_ENET_RD_CMD, cmd);
123 /* wait for read command to complete */
124 while (!(done = ioread32(cmd_done)) && wait--)
125 udelay(1);
127 if (!done)
128 return false;
130 *rd_data = ioread32(rd);
131 iowrite32(0, cmd);
133 return true;
136 static bool xgene_enet_rd_pcs(struct xgene_enet_pdata *pdata,
137 u32 rd_addr, u32 *rd_data)
139 void __iomem *addr, *rd, *cmd, *cmd_done;
140 bool success;
142 addr = pdata->pcs_addr + PCS_ADDR_REG_OFFSET;
143 rd = pdata->pcs_addr + PCS_READ_REG_OFFSET;
144 cmd = pdata->pcs_addr + PCS_COMMAND_REG_OFFSET;
145 cmd_done = pdata->pcs_addr + PCS_COMMAND_DONE_REG_OFFSET;
147 success = xgene_enet_rd_indirect(addr, rd, cmd, cmd_done, rd_addr, rd_data);
148 if (!success)
149 netdev_err(pdata->ndev, "PCS read failed, addr: %04x\n",
150 rd_addr);
152 return success;
155 static void xgene_enet_rd_axg_csr(struct xgene_enet_pdata *pdata,
156 u32 offset, u32 *val)
158 void __iomem *addr = pdata->mcx_mac_csr_addr + offset;
160 *val = ioread32(addr);
163 static int xgene_enet_ecc_init(struct xgene_enet_pdata *pdata)
165 struct net_device *ndev = pdata->ndev;
166 u32 data;
167 u8 wait = 10;
169 xgene_enet_wr_diag_csr(pdata, ENET_CFG_MEM_RAM_SHUTDOWN_ADDR, 0x0);
170 do {
171 usleep_range(100, 110);
172 xgene_enet_rd_diag_csr(pdata, ENET_BLOCK_MEM_RDY_ADDR, &data);
173 } while ((data != 0xffffffff) && wait--);
175 if (data != 0xffffffff) {
176 netdev_err(ndev, "Failed to release memory from shutdown\n");
177 return -ENODEV;
180 return 0;
183 static void xgene_xgmac_get_drop_cnt(struct xgene_enet_pdata *pdata,
184 u32 *rx, u32 *tx)
186 u32 count;
188 xgene_enet_rd_axg_csr(pdata, XGENET_ICM_ECM_DROP_COUNT_REG0, &count);
189 *rx = ICM_DROP_COUNT(count);
190 *tx = ECM_DROP_COUNT(count);
191 /* Errata: 10GE_4 - ICM_ECM_DROP_COUNT not clear-on-read */
192 xgene_enet_rd_axg_csr(pdata, XGENET_ECM_CONFIG0_REG_0, &count);
195 static void xgene_enet_config_ring_if_assoc(struct xgene_enet_pdata *pdata)
197 xgene_enet_wr_ring_if(pdata, ENET_CFGSSQMIWQASSOC_ADDR, 0);
198 xgene_enet_wr_ring_if(pdata, ENET_CFGSSQMIFPQASSOC_ADDR, 0);
199 xgene_enet_wr_ring_if(pdata, ENET_CFGSSQMIQMLITEWQASSOC_ADDR, 0);
200 xgene_enet_wr_ring_if(pdata, ENET_CFGSSQMIQMLITEFPQASSOC_ADDR, 0);
203 static void xgene_xgmac_reset(struct xgene_enet_pdata *pdata)
205 xgene_enet_wr_mac(pdata, AXGMAC_CONFIG_0, HSTMACRST);
206 xgene_enet_wr_mac(pdata, AXGMAC_CONFIG_0, 0);
209 static void xgene_pcs_reset(struct xgene_enet_pdata *pdata)
211 u32 data;
213 if (!xgene_enet_rd_pcs(pdata, PCS_CONTROL_1, &data))
214 return;
216 xgene_enet_wr_pcs(pdata, PCS_CONTROL_1, data | PCS_CTRL_PCS_RST);
217 xgene_enet_wr_pcs(pdata, PCS_CONTROL_1, data & ~PCS_CTRL_PCS_RST);
220 static void xgene_xgmac_set_mac_addr(struct xgene_enet_pdata *pdata)
222 u32 addr0, addr1;
223 u8 *dev_addr = pdata->ndev->dev_addr;
225 addr0 = (dev_addr[3] << 24) | (dev_addr[2] << 16) |
226 (dev_addr[1] << 8) | dev_addr[0];
227 addr1 = (dev_addr[5] << 24) | (dev_addr[4] << 16);
229 xgene_enet_wr_mac(pdata, HSTMACADR_LSW_ADDR, addr0);
230 xgene_enet_wr_mac(pdata, HSTMACADR_MSW_ADDR, addr1);
233 static void xgene_xgmac_set_mss(struct xgene_enet_pdata *pdata,
234 u16 mss, u8 index)
236 u8 offset;
237 u32 data;
239 offset = (index < 2) ? 0 : 4;
240 xgene_enet_rd_csr(pdata, XG_TSIF_MSS_REG0_ADDR + offset, &data);
242 if (!(index & 0x1))
243 data = SET_VAL(TSO_MSS1, data >> TSO_MSS1_POS) |
244 SET_VAL(TSO_MSS0, mss);
245 else
246 data = SET_VAL(TSO_MSS1, mss) | SET_VAL(TSO_MSS0, data);
248 xgene_enet_wr_csr(pdata, XG_TSIF_MSS_REG0_ADDR + offset, data);
251 static void xgene_xgmac_set_frame_size(struct xgene_enet_pdata *pdata, int size)
253 xgene_enet_wr_mac(pdata, HSTMAXFRAME_LENGTH_ADDR,
254 ((((size + 2) >> 2) << 16) | size));
257 static u32 xgene_enet_link_status(struct xgene_enet_pdata *pdata)
259 u32 data;
261 xgene_enet_rd_csr(pdata, XG_LINK_STATUS_ADDR, &data);
263 return data;
266 static void xgene_xgmac_enable_tx_pause(struct xgene_enet_pdata *pdata,
267 bool enable)
269 u32 data;
271 xgene_enet_rd_axg_csr(pdata, XGENET_CSR_ECM_CFG_0_ADDR, &data);
273 if (enable)
274 data |= MULTI_DPF_AUTOCTRL | PAUSE_XON_EN;
275 else
276 data &= ~(MULTI_DPF_AUTOCTRL | PAUSE_XON_EN);
278 xgene_enet_wr_axg_csr(pdata, XGENET_CSR_ECM_CFG_0_ADDR, data);
281 static void xgene_xgmac_flowctl_tx(struct xgene_enet_pdata *pdata, bool enable)
283 u32 data;
285 data = xgene_enet_rd_mac(pdata, AXGMAC_CONFIG_1);
287 if (enable)
288 data |= HSTTCTLEN;
289 else
290 data &= ~HSTTCTLEN;
292 xgene_enet_wr_mac(pdata, AXGMAC_CONFIG_1, data);
294 pdata->mac_ops->enable_tx_pause(pdata, enable);
297 static void xgene_xgmac_flowctl_rx(struct xgene_enet_pdata *pdata, bool enable)
299 u32 data;
301 data = xgene_enet_rd_mac(pdata, AXGMAC_CONFIG_1);
303 if (enable)
304 data |= HSTRCTLEN;
305 else
306 data &= ~HSTRCTLEN;
308 xgene_enet_wr_mac(pdata, AXGMAC_CONFIG_1, data);
311 static void xgene_xgmac_init(struct xgene_enet_pdata *pdata)
313 u32 data;
315 xgene_xgmac_reset(pdata);
317 data = xgene_enet_rd_mac(pdata, AXGMAC_CONFIG_1);
318 data |= HSTPPEN;
319 data &= ~HSTLENCHK;
320 xgene_enet_wr_mac(pdata, AXGMAC_CONFIG_1, data);
322 xgene_xgmac_set_mac_addr(pdata);
324 xgene_enet_rd_csr(pdata, XG_RSIF_CONFIG_REG_ADDR, &data);
325 data |= CFG_RSIF_FPBUFF_TIMEOUT_EN;
326 /* Errata 10GE_1 - FIFO threshold default value incorrect */
327 RSIF_CLE_BUFF_THRESH_SET(&data, XG_RSIF_CLE_BUFF_THRESH);
328 xgene_enet_wr_csr(pdata, XG_RSIF_CONFIG_REG_ADDR, data);
330 /* Errata 10GE_1 - FIFO threshold default value incorrect */
331 xgene_enet_rd_csr(pdata, XG_RSIF_CONFIG1_REG_ADDR, &data);
332 RSIF_PLC_CLE_BUFF_THRESH_SET(&data, XG_RSIF_PLC_CLE_BUFF_THRESH);
333 xgene_enet_wr_csr(pdata, XG_RSIF_CONFIG1_REG_ADDR, data);
335 xgene_enet_rd_csr(pdata, XG_ENET_SPARE_CFG_REG_ADDR, &data);
336 data |= BIT(12);
337 xgene_enet_wr_csr(pdata, XG_ENET_SPARE_CFG_REG_ADDR, data);
338 xgene_enet_wr_csr(pdata, XG_ENET_SPARE_CFG_REG_1_ADDR, 0x82);
339 xgene_enet_wr_csr(pdata, XGENET_RX_DV_GATE_REG_0_ADDR, 0);
340 xgene_enet_wr_csr(pdata, XG_CFG_BYPASS_ADDR, RESUME_TX);
342 /* Configure HW pause frame generation */
343 xgene_enet_rd_axg_csr(pdata, XGENET_CSR_MULTI_DPF0_ADDR, &data);
344 data = (DEF_QUANTA << 16) | (data & 0xFFFF);
345 xgene_enet_wr_axg_csr(pdata, XGENET_CSR_MULTI_DPF0_ADDR, data);
347 if (pdata->enet_id != XGENE_ENET1) {
348 xgene_enet_rd_axg_csr(pdata, XGENET_CSR_MULTI_DPF1_ADDR, &data);
349 data = (NORM_PAUSE_OPCODE << 16) | (data & 0xFFFF);
350 xgene_enet_wr_axg_csr(pdata, XGENET_CSR_MULTI_DPF1_ADDR, data);
353 data = (XG_DEF_PAUSE_OFF_THRES << 16) | XG_DEF_PAUSE_THRES;
354 xgene_enet_wr_csr(pdata, XG_RXBUF_PAUSE_THRESH, data);
356 xgene_xgmac_flowctl_tx(pdata, pdata->tx_pause);
357 xgene_xgmac_flowctl_rx(pdata, pdata->rx_pause);
360 static void xgene_xgmac_rx_enable(struct xgene_enet_pdata *pdata)
362 u32 data;
364 data = xgene_enet_rd_mac(pdata, AXGMAC_CONFIG_1);
365 xgene_enet_wr_mac(pdata, AXGMAC_CONFIG_1, data | HSTRFEN);
368 static void xgene_xgmac_tx_enable(struct xgene_enet_pdata *pdata)
370 u32 data;
372 data = xgene_enet_rd_mac(pdata, AXGMAC_CONFIG_1);
373 xgene_enet_wr_mac(pdata, AXGMAC_CONFIG_1, data | HSTTFEN);
376 static void xgene_xgmac_rx_disable(struct xgene_enet_pdata *pdata)
378 u32 data;
380 data = xgene_enet_rd_mac(pdata, AXGMAC_CONFIG_1);
381 xgene_enet_wr_mac(pdata, AXGMAC_CONFIG_1, data & ~HSTRFEN);
384 static void xgene_xgmac_tx_disable(struct xgene_enet_pdata *pdata)
386 u32 data;
388 data = xgene_enet_rd_mac(pdata, AXGMAC_CONFIG_1);
389 xgene_enet_wr_mac(pdata, AXGMAC_CONFIG_1, data & ~HSTTFEN);
392 static int xgene_enet_reset(struct xgene_enet_pdata *pdata)
394 struct device *dev = &pdata->pdev->dev;
396 if (!xgene_ring_mgr_init(pdata))
397 return -ENODEV;
399 if (dev->of_node) {
400 clk_prepare_enable(pdata->clk);
401 udelay(5);
402 clk_disable_unprepare(pdata->clk);
403 udelay(5);
404 clk_prepare_enable(pdata->clk);
405 udelay(5);
406 } else {
407 #ifdef CONFIG_ACPI
408 if (acpi_has_method(ACPI_HANDLE(&pdata->pdev->dev), "_RST")) {
409 acpi_evaluate_object(ACPI_HANDLE(&pdata->pdev->dev),
410 "_RST", NULL, NULL);
411 } else if (acpi_has_method(ACPI_HANDLE(&pdata->pdev->dev),
412 "_INI")) {
413 acpi_evaluate_object(ACPI_HANDLE(&pdata->pdev->dev),
414 "_INI", NULL, NULL);
416 #endif
419 xgene_enet_ecc_init(pdata);
420 xgene_enet_config_ring_if_assoc(pdata);
422 return 0;
425 static void xgene_enet_xgcle_bypass(struct xgene_enet_pdata *pdata,
426 u32 dst_ring_num, u16 bufpool_id,
427 u16 nxtbufpool_id)
429 u32 cb, fpsel, nxtfpsel;
431 xgene_enet_rd_csr(pdata, XCLE_BYPASS_REG0_ADDR, &cb);
432 cb |= CFG_CLE_BYPASS_EN0;
433 CFG_CLE_IP_PROTOCOL0_SET(&cb, 3);
434 xgene_enet_wr_csr(pdata, XCLE_BYPASS_REG0_ADDR, cb);
436 fpsel = xgene_enet_get_fpsel(bufpool_id);
437 nxtfpsel = xgene_enet_get_fpsel(nxtbufpool_id);
438 xgene_enet_rd_csr(pdata, XCLE_BYPASS_REG1_ADDR, &cb);
439 CFG_CLE_DSTQID0_SET(&cb, dst_ring_num);
440 CFG_CLE_FPSEL0_SET(&cb, fpsel);
441 CFG_CLE_NXTFPSEL0_SET(&cb, nxtfpsel);
442 xgene_enet_wr_csr(pdata, XCLE_BYPASS_REG1_ADDR, cb);
443 pr_info("+ cle_bypass: fpsel: %d nxtfpsel: %d\n", fpsel, nxtfpsel);
446 static void xgene_enet_shutdown(struct xgene_enet_pdata *pdata)
448 struct device *dev = &pdata->pdev->dev;
450 if (dev->of_node) {
451 if (!IS_ERR(pdata->clk))
452 clk_disable_unprepare(pdata->clk);
456 static void xgene_enet_clear(struct xgene_enet_pdata *pdata,
457 struct xgene_enet_desc_ring *ring)
459 u32 addr, data;
461 if (xgene_enet_is_bufpool(ring->id)) {
462 addr = ENET_CFGSSQMIFPRESET_ADDR;
463 data = BIT(xgene_enet_get_fpsel(ring->id));
464 } else {
465 addr = ENET_CFGSSQMIWQRESET_ADDR;
466 data = BIT(xgene_enet_ring_bufnum(ring->id));
469 xgene_enet_wr_ring_if(pdata, addr, data);
472 static int xgene_enet_gpio_lookup(struct xgene_enet_pdata *pdata)
474 struct device *dev = &pdata->pdev->dev;
476 pdata->sfp_rdy = gpiod_get(dev, "rxlos", GPIOD_IN);
477 if (IS_ERR(pdata->sfp_rdy))
478 pdata->sfp_rdy = gpiod_get(dev, "sfp", GPIOD_IN);
480 if (IS_ERR(pdata->sfp_rdy))
481 return -ENODEV;
483 return 0;
486 static void xgene_enet_link_state(struct work_struct *work)
488 struct xgene_enet_pdata *pdata = container_of(to_delayed_work(work),
489 struct xgene_enet_pdata, link_work);
490 struct net_device *ndev = pdata->ndev;
491 u32 link_status, poll_interval;
493 link_status = xgene_enet_link_status(pdata);
494 if (pdata->sfp_gpio_en && link_status &&
495 (!IS_ERR(pdata->sfp_rdy) || !xgene_enet_gpio_lookup(pdata)) &&
496 !gpiod_get_value(pdata->sfp_rdy))
497 link_status = 0;
499 if (link_status) {
500 if (!netif_carrier_ok(ndev)) {
501 netif_carrier_on(ndev);
502 xgene_xgmac_rx_enable(pdata);
503 xgene_xgmac_tx_enable(pdata);
504 netdev_info(ndev, "Link is Up - 10Gbps\n");
506 poll_interval = PHY_POLL_LINK_ON;
507 } else {
508 if (netif_carrier_ok(ndev)) {
509 xgene_xgmac_rx_disable(pdata);
510 xgene_xgmac_tx_disable(pdata);
511 netif_carrier_off(ndev);
512 netdev_info(ndev, "Link is Down\n");
514 poll_interval = PHY_POLL_LINK_OFF;
516 xgene_pcs_reset(pdata);
519 schedule_delayed_work(&pdata->link_work, poll_interval);
522 const struct xgene_mac_ops xgene_xgmac_ops = {
523 .init = xgene_xgmac_init,
524 .reset = xgene_xgmac_reset,
525 .rx_enable = xgene_xgmac_rx_enable,
526 .tx_enable = xgene_xgmac_tx_enable,
527 .rx_disable = xgene_xgmac_rx_disable,
528 .tx_disable = xgene_xgmac_tx_disable,
529 .set_mac_addr = xgene_xgmac_set_mac_addr,
530 .set_framesize = xgene_xgmac_set_frame_size,
531 .set_mss = xgene_xgmac_set_mss,
532 .get_drop_cnt = xgene_xgmac_get_drop_cnt,
533 .link_state = xgene_enet_link_state,
534 .enable_tx_pause = xgene_xgmac_enable_tx_pause,
535 .flowctl_rx = xgene_xgmac_flowctl_rx,
536 .flowctl_tx = xgene_xgmac_flowctl_tx
539 const struct xgene_port_ops xgene_xgport_ops = {
540 .reset = xgene_enet_reset,
541 .clear = xgene_enet_clear,
542 .cle_bypass = xgene_enet_xgcle_bypass,
543 .shutdown = xgene_enet_shutdown,