1 /* MOXA ART Ethernet (RTL8201CP) driver.
3 * Copyright (C) 2013 Jonas Jensen
5 * Jonas Jensen <jonas.jensen@gmail.com>
8 * Moxa Technology Co., Ltd. <www.moxa.com>
10 * This file is licensed under the terms of the GNU General Public
11 * License version 2. This program is licensed "as is" without any
12 * warranty of any kind, whether express or implied.
15 #include <linux/module.h>
16 #include <linux/netdevice.h>
17 #include <linux/etherdevice.h>
18 #include <linux/skbuff.h>
19 #include <linux/dma-mapping.h>
20 #include <linux/ethtool.h>
21 #include <linux/platform_device.h>
22 #include <linux/interrupt.h>
23 #include <linux/irq.h>
24 #include <linux/of_address.h>
25 #include <linux/of_irq.h>
26 #include <linux/crc32.h>
27 #include <linux/crc32c.h>
28 #include <linux/circ_buf.h>
30 #include "moxart_ether.h"
32 static inline void moxart_desc_write(u32 data
, u32
*desc
)
34 *desc
= cpu_to_le32(data
);
37 static inline u32
moxart_desc_read(u32
*desc
)
39 return le32_to_cpu(*desc
);
42 static inline void moxart_emac_write(struct net_device
*ndev
,
43 unsigned int reg
, unsigned long value
)
45 struct moxart_mac_priv_t
*priv
= netdev_priv(ndev
);
47 writel(value
, priv
->base
+ reg
);
50 static void moxart_update_mac_address(struct net_device
*ndev
)
52 moxart_emac_write(ndev
, REG_MAC_MS_ADDRESS
,
53 ((ndev
->dev_addr
[0] << 8) | (ndev
->dev_addr
[1])));
54 moxart_emac_write(ndev
, REG_MAC_MS_ADDRESS
+ 4,
55 ((ndev
->dev_addr
[2] << 24) |
56 (ndev
->dev_addr
[3] << 16) |
57 (ndev
->dev_addr
[4] << 8) |
58 (ndev
->dev_addr
[5])));
61 static int moxart_set_mac_address(struct net_device
*ndev
, void *addr
)
63 struct sockaddr
*address
= addr
;
65 if (!is_valid_ether_addr(address
->sa_data
))
66 return -EADDRNOTAVAIL
;
68 memcpy(ndev
->dev_addr
, address
->sa_data
, ndev
->addr_len
);
69 moxart_update_mac_address(ndev
);
74 static void moxart_mac_free_memory(struct net_device
*ndev
)
76 struct moxart_mac_priv_t
*priv
= netdev_priv(ndev
);
79 for (i
= 0; i
< RX_DESC_NUM
; i
++)
80 dma_unmap_single(&ndev
->dev
, priv
->rx_mapping
[i
],
81 priv
->rx_buf_size
, DMA_FROM_DEVICE
);
83 if (priv
->tx_desc_base
)
84 dma_free_coherent(&priv
->pdev
->dev
,
85 TX_REG_DESC_SIZE
* TX_DESC_NUM
,
86 priv
->tx_desc_base
, priv
->tx_base
);
88 if (priv
->rx_desc_base
)
89 dma_free_coherent(&priv
->pdev
->dev
,
90 RX_REG_DESC_SIZE
* RX_DESC_NUM
,
91 priv
->rx_desc_base
, priv
->rx_base
);
93 kfree(priv
->tx_buf_base
);
94 kfree(priv
->rx_buf_base
);
97 static void moxart_mac_reset(struct net_device
*ndev
)
99 struct moxart_mac_priv_t
*priv
= netdev_priv(ndev
);
101 writel(SW_RST
, priv
->base
+ REG_MAC_CTRL
);
102 while (readl(priv
->base
+ REG_MAC_CTRL
) & SW_RST
)
105 writel(0, priv
->base
+ REG_INTERRUPT_MASK
);
107 priv
->reg_maccr
= RX_BROADPKT
| FULLDUP
| CRC_APD
| RX_FTL
;
110 static void moxart_mac_enable(struct net_device
*ndev
)
112 struct moxart_mac_priv_t
*priv
= netdev_priv(ndev
);
114 writel(0x00001010, priv
->base
+ REG_INT_TIMER_CTRL
);
115 writel(0x00000001, priv
->base
+ REG_APOLL_TIMER_CTRL
);
116 writel(0x00000390, priv
->base
+ REG_DMA_BLEN_CTRL
);
118 priv
->reg_imr
|= (RPKT_FINISH_M
| XPKT_FINISH_M
);
119 writel(priv
->reg_imr
, priv
->base
+ REG_INTERRUPT_MASK
);
121 priv
->reg_maccr
|= (RCV_EN
| XMT_EN
| RDMA_EN
| XDMA_EN
);
122 writel(priv
->reg_maccr
, priv
->base
+ REG_MAC_CTRL
);
125 static void moxart_mac_setup_desc_ring(struct net_device
*ndev
)
127 struct moxart_mac_priv_t
*priv
= netdev_priv(ndev
);
131 for (i
= 0; i
< TX_DESC_NUM
; i
++) {
132 desc
= priv
->tx_desc_base
+ i
* TX_REG_DESC_SIZE
;
133 memset(desc
, 0, TX_REG_DESC_SIZE
);
135 priv
->tx_buf
[i
] = priv
->tx_buf_base
+ priv
->tx_buf_size
* i
;
137 moxart_desc_write(TX_DESC1_END
, desc
+ TX_REG_OFFSET_DESC1
);
142 for (i
= 0; i
< RX_DESC_NUM
; i
++) {
143 desc
= priv
->rx_desc_base
+ i
* RX_REG_DESC_SIZE
;
144 memset(desc
, 0, RX_REG_DESC_SIZE
);
145 moxart_desc_write(RX_DESC0_DMA_OWN
, desc
+ RX_REG_OFFSET_DESC0
);
146 moxart_desc_write(RX_BUF_SIZE
& RX_DESC1_BUF_SIZE_MASK
,
147 desc
+ RX_REG_OFFSET_DESC1
);
149 priv
->rx_buf
[i
] = priv
->rx_buf_base
+ priv
->rx_buf_size
* i
;
150 priv
->rx_mapping
[i
] = dma_map_single(&ndev
->dev
,
154 if (dma_mapping_error(&ndev
->dev
, priv
->rx_mapping
[i
]))
155 netdev_err(ndev
, "DMA mapping error\n");
157 moxart_desc_write(priv
->rx_mapping
[i
],
158 desc
+ RX_REG_OFFSET_DESC2
+ RX_DESC2_ADDRESS_PHYS
);
159 moxart_desc_write((uintptr_t)priv
->rx_buf
[i
],
160 desc
+ RX_REG_OFFSET_DESC2
+ RX_DESC2_ADDRESS_VIRT
);
162 moxart_desc_write(RX_DESC1_END
, desc
+ RX_REG_OFFSET_DESC1
);
166 /* reset the MAC controller TX/RX descriptor base address */
167 writel(priv
->tx_base
, priv
->base
+ REG_TXR_BASE_ADDRESS
);
168 writel(priv
->rx_base
, priv
->base
+ REG_RXR_BASE_ADDRESS
);
171 static int moxart_mac_open(struct net_device
*ndev
)
173 struct moxart_mac_priv_t
*priv
= netdev_priv(ndev
);
175 if (!is_valid_ether_addr(ndev
->dev_addr
))
176 return -EADDRNOTAVAIL
;
178 napi_enable(&priv
->napi
);
180 moxart_mac_reset(ndev
);
181 moxart_update_mac_address(ndev
);
182 moxart_mac_setup_desc_ring(ndev
);
183 moxart_mac_enable(ndev
);
184 netif_start_queue(ndev
);
186 netdev_dbg(ndev
, "%s: IMR=0x%x, MACCR=0x%x\n",
187 __func__
, readl(priv
->base
+ REG_INTERRUPT_MASK
),
188 readl(priv
->base
+ REG_MAC_CTRL
));
193 static int moxart_mac_stop(struct net_device
*ndev
)
195 struct moxart_mac_priv_t
*priv
= netdev_priv(ndev
);
197 napi_disable(&priv
->napi
);
199 netif_stop_queue(ndev
);
201 /* disable all interrupts */
202 writel(0, priv
->base
+ REG_INTERRUPT_MASK
);
204 /* disable all functions */
205 writel(0, priv
->base
+ REG_MAC_CTRL
);
210 static int moxart_rx_poll(struct napi_struct
*napi
, int budget
)
212 struct moxart_mac_priv_t
*priv
= container_of(napi
,
213 struct moxart_mac_priv_t
,
215 struct net_device
*ndev
= priv
->ndev
;
218 unsigned int desc0
, len
;
219 int rx_head
= priv
->rx_head
;
222 while (rx
< budget
) {
223 desc
= priv
->rx_desc_base
+ (RX_REG_DESC_SIZE
* rx_head
);
224 desc0
= moxart_desc_read(desc
+ RX_REG_OFFSET_DESC0
);
225 rmb(); /* ensure desc0 is up to date */
227 if (desc0
& RX_DESC0_DMA_OWN
)
230 if (desc0
& (RX_DESC0_ERR
| RX_DESC0_CRC_ERR
| RX_DESC0_FTL
|
231 RX_DESC0_RUNT
| RX_DESC0_ODD_NB
)) {
232 net_dbg_ratelimited("packet error\n");
233 ndev
->stats
.rx_dropped
++;
234 ndev
->stats
.rx_errors
++;
238 len
= desc0
& RX_DESC0_FRAME_LEN_MASK
;
240 if (len
> RX_BUF_SIZE
)
243 dma_sync_single_for_cpu(&ndev
->dev
,
244 priv
->rx_mapping
[rx_head
],
245 priv
->rx_buf_size
, DMA_FROM_DEVICE
);
246 skb
= netdev_alloc_skb_ip_align(ndev
, len
);
248 if (unlikely(!skb
)) {
249 net_dbg_ratelimited("netdev_alloc_skb_ip_align failed\n");
250 ndev
->stats
.rx_dropped
++;
251 ndev
->stats
.rx_errors
++;
255 memcpy(skb
->data
, priv
->rx_buf
[rx_head
], len
);
257 skb
->protocol
= eth_type_trans(skb
, ndev
);
258 napi_gro_receive(&priv
->napi
, skb
);
261 ndev
->stats
.rx_packets
++;
262 ndev
->stats
.rx_bytes
+= len
;
263 if (desc0
& RX_DESC0_MULTICAST
)
264 ndev
->stats
.multicast
++;
267 wmb(); /* prevent setting ownership back too early */
268 moxart_desc_write(RX_DESC0_DMA_OWN
, desc
+ RX_REG_OFFSET_DESC0
);
270 rx_head
= RX_NEXT(rx_head
);
271 priv
->rx_head
= rx_head
;
275 napi_complete_done(napi
, rx
);
277 priv
->reg_imr
|= RPKT_FINISH_M
;
278 writel(priv
->reg_imr
, priv
->base
+ REG_INTERRUPT_MASK
);
283 static int moxart_tx_queue_space(struct net_device
*ndev
)
285 struct moxart_mac_priv_t
*priv
= netdev_priv(ndev
);
287 return CIRC_SPACE(priv
->tx_head
, priv
->tx_tail
, TX_DESC_NUM
);
290 static void moxart_tx_finished(struct net_device
*ndev
)
292 struct moxart_mac_priv_t
*priv
= netdev_priv(ndev
);
293 unsigned int tx_head
= priv
->tx_head
;
294 unsigned int tx_tail
= priv
->tx_tail
;
296 while (tx_tail
!= tx_head
) {
297 dma_unmap_single(&ndev
->dev
, priv
->tx_mapping
[tx_tail
],
298 priv
->tx_len
[tx_tail
], DMA_TO_DEVICE
);
300 ndev
->stats
.tx_packets
++;
301 ndev
->stats
.tx_bytes
+= priv
->tx_skb
[tx_tail
]->len
;
303 dev_consume_skb_irq(priv
->tx_skb
[tx_tail
]);
304 priv
->tx_skb
[tx_tail
] = NULL
;
306 tx_tail
= TX_NEXT(tx_tail
);
308 priv
->tx_tail
= tx_tail
;
309 if (netif_queue_stopped(ndev
) &&
310 moxart_tx_queue_space(ndev
) >= TX_WAKE_THRESHOLD
)
311 netif_wake_queue(ndev
);
314 static irqreturn_t
moxart_mac_interrupt(int irq
, void *dev_id
)
316 struct net_device
*ndev
= (struct net_device
*)dev_id
;
317 struct moxart_mac_priv_t
*priv
= netdev_priv(ndev
);
318 unsigned int ists
= readl(priv
->base
+ REG_INTERRUPT_STATUS
);
320 if (ists
& XPKT_OK_INT_STS
)
321 moxart_tx_finished(ndev
);
323 if (ists
& RPKT_FINISH
) {
324 if (napi_schedule_prep(&priv
->napi
)) {
325 priv
->reg_imr
&= ~RPKT_FINISH_M
;
326 writel(priv
->reg_imr
, priv
->base
+ REG_INTERRUPT_MASK
);
327 __napi_schedule(&priv
->napi
);
334 static int moxart_mac_start_xmit(struct sk_buff
*skb
, struct net_device
*ndev
)
336 struct moxart_mac_priv_t
*priv
= netdev_priv(ndev
);
339 unsigned int tx_head
;
341 int ret
= NETDEV_TX_BUSY
;
343 spin_lock_irq(&priv
->txlock
);
345 tx_head
= priv
->tx_head
;
346 desc
= priv
->tx_desc_base
+ (TX_REG_DESC_SIZE
* tx_head
);
348 if (moxart_tx_queue_space(ndev
) == 1)
349 netif_stop_queue(ndev
);
351 if (moxart_desc_read(desc
+ TX_REG_OFFSET_DESC0
) & TX_DESC0_DMA_OWN
) {
352 net_dbg_ratelimited("no TX space for packet\n");
353 ndev
->stats
.tx_dropped
++;
356 rmb(); /* ensure data is only read that had TX_DESC0_DMA_OWN cleared */
358 len
= skb
->len
> TX_BUF_SIZE
? TX_BUF_SIZE
: skb
->len
;
360 priv
->tx_mapping
[tx_head
] = dma_map_single(&ndev
->dev
, skb
->data
,
362 if (dma_mapping_error(&ndev
->dev
, priv
->tx_mapping
[tx_head
])) {
363 netdev_err(ndev
, "DMA mapping error\n");
367 priv
->tx_len
[tx_head
] = len
;
368 priv
->tx_skb
[tx_head
] = skb
;
370 moxart_desc_write(priv
->tx_mapping
[tx_head
],
371 desc
+ TX_REG_OFFSET_DESC2
+ TX_DESC2_ADDRESS_PHYS
);
372 moxart_desc_write((uintptr_t)skb
->data
,
373 desc
+ TX_REG_OFFSET_DESC2
+ TX_DESC2_ADDRESS_VIRT
);
375 if (skb
->len
< ETH_ZLEN
) {
376 memset(&skb
->data
[skb
->len
],
377 0, ETH_ZLEN
- skb
->len
);
381 dma_sync_single_for_device(&ndev
->dev
, priv
->tx_mapping
[tx_head
],
382 priv
->tx_buf_size
, DMA_TO_DEVICE
);
384 txdes1
= TX_DESC1_LTS
| TX_DESC1_FTS
| (len
& TX_DESC1_BUF_SIZE_MASK
);
385 if (tx_head
== TX_DESC_NUM_MASK
)
386 txdes1
|= TX_DESC1_END
;
387 moxart_desc_write(txdes1
, desc
+ TX_REG_OFFSET_DESC1
);
388 wmb(); /* flush descriptor before transferring ownership */
389 moxart_desc_write(TX_DESC0_DMA_OWN
, desc
+ TX_REG_OFFSET_DESC0
);
391 /* start to send packet */
392 writel(0xffffffff, priv
->base
+ REG_TX_POLL_DEMAND
);
394 priv
->tx_head
= TX_NEXT(tx_head
);
396 netif_trans_update(ndev
);
399 spin_unlock_irq(&priv
->txlock
);
404 static void moxart_mac_setmulticast(struct net_device
*ndev
)
406 struct moxart_mac_priv_t
*priv
= netdev_priv(ndev
);
407 struct netdev_hw_addr
*ha
;
410 netdev_for_each_mc_addr(ha
, ndev
) {
411 crc_val
= crc32_le(~0, ha
->addr
, ETH_ALEN
);
412 crc_val
= (crc_val
>> 26) & 0x3f;
414 writel(readl(priv
->base
+ REG_MCAST_HASH_TABLE1
) |
415 (1UL << (crc_val
- 32)),
416 priv
->base
+ REG_MCAST_HASH_TABLE1
);
418 writel(readl(priv
->base
+ REG_MCAST_HASH_TABLE0
) |
420 priv
->base
+ REG_MCAST_HASH_TABLE0
);
425 static void moxart_mac_set_rx_mode(struct net_device
*ndev
)
427 struct moxart_mac_priv_t
*priv
= netdev_priv(ndev
);
429 spin_lock_irq(&priv
->txlock
);
431 (ndev
->flags
& IFF_PROMISC
) ? (priv
->reg_maccr
|= RCV_ALL
) :
432 (priv
->reg_maccr
&= ~RCV_ALL
);
434 (ndev
->flags
& IFF_ALLMULTI
) ? (priv
->reg_maccr
|= RX_MULTIPKT
) :
435 (priv
->reg_maccr
&= ~RX_MULTIPKT
);
437 if ((ndev
->flags
& IFF_MULTICAST
) && netdev_mc_count(ndev
)) {
438 priv
->reg_maccr
|= HT_MULTI_EN
;
439 moxart_mac_setmulticast(ndev
);
441 priv
->reg_maccr
&= ~HT_MULTI_EN
;
444 writel(priv
->reg_maccr
, priv
->base
+ REG_MAC_CTRL
);
446 spin_unlock_irq(&priv
->txlock
);
449 static const struct net_device_ops moxart_netdev_ops
= {
450 .ndo_open
= moxart_mac_open
,
451 .ndo_stop
= moxart_mac_stop
,
452 .ndo_start_xmit
= moxart_mac_start_xmit
,
453 .ndo_set_rx_mode
= moxart_mac_set_rx_mode
,
454 .ndo_set_mac_address
= moxart_set_mac_address
,
455 .ndo_validate_addr
= eth_validate_addr
,
458 static int moxart_mac_probe(struct platform_device
*pdev
)
460 struct device
*p_dev
= &pdev
->dev
;
461 struct device_node
*node
= p_dev
->of_node
;
462 struct net_device
*ndev
;
463 struct moxart_mac_priv_t
*priv
;
464 struct resource
*res
;
468 ndev
= alloc_etherdev(sizeof(struct moxart_mac_priv_t
));
472 irq
= irq_of_parse_and_map(node
, 0);
474 netdev_err(ndev
, "irq_of_parse_and_map failed\n");
479 priv
= netdev_priv(ndev
);
483 res
= platform_get_resource(pdev
, IORESOURCE_MEM
, 0);
484 ndev
->base_addr
= res
->start
;
485 priv
->base
= devm_ioremap_resource(p_dev
, res
);
486 if (IS_ERR(priv
->base
)) {
487 dev_err(p_dev
, "devm_ioremap_resource failed\n");
488 ret
= PTR_ERR(priv
->base
);
492 spin_lock_init(&priv
->txlock
);
494 priv
->tx_buf_size
= TX_BUF_SIZE
;
495 priv
->rx_buf_size
= RX_BUF_SIZE
;
497 priv
->tx_desc_base
= dma_alloc_coherent(&pdev
->dev
, TX_REG_DESC_SIZE
*
498 TX_DESC_NUM
, &priv
->tx_base
,
499 GFP_DMA
| GFP_KERNEL
);
500 if (!priv
->tx_desc_base
) {
505 priv
->rx_desc_base
= dma_alloc_coherent(&pdev
->dev
, RX_REG_DESC_SIZE
*
506 RX_DESC_NUM
, &priv
->rx_base
,
507 GFP_DMA
| GFP_KERNEL
);
508 if (!priv
->rx_desc_base
) {
513 priv
->tx_buf_base
= kmalloc_array(priv
->tx_buf_size
, TX_DESC_NUM
,
515 if (!priv
->tx_buf_base
) {
520 priv
->rx_buf_base
= kmalloc_array(priv
->rx_buf_size
, RX_DESC_NUM
,
522 if (!priv
->rx_buf_base
) {
527 platform_set_drvdata(pdev
, ndev
);
529 ret
= devm_request_irq(p_dev
, irq
, moxart_mac_interrupt
, 0,
532 netdev_err(ndev
, "devm_request_irq failed\n");
536 ndev
->netdev_ops
= &moxart_netdev_ops
;
537 netif_napi_add(ndev
, &priv
->napi
, moxart_rx_poll
, RX_DESC_NUM
);
538 ndev
->priv_flags
|= IFF_UNICAST_FLT
;
541 SET_NETDEV_DEV(ndev
, &pdev
->dev
);
543 ret
= register_netdev(ndev
);
549 netdev_dbg(ndev
, "%s: IRQ=%d address=%pM\n",
550 __func__
, ndev
->irq
, ndev
->dev_addr
);
555 netdev_err(ndev
, "init failed\n");
556 moxart_mac_free_memory(ndev
);
562 static int moxart_remove(struct platform_device
*pdev
)
564 struct net_device
*ndev
= platform_get_drvdata(pdev
);
566 unregister_netdev(ndev
);
567 free_irq(ndev
->irq
, ndev
);
568 moxart_mac_free_memory(ndev
);
574 static const struct of_device_id moxart_mac_match
[] = {
575 { .compatible
= "moxa,moxart-mac" },
578 MODULE_DEVICE_TABLE(of
, moxart_mac_match
);
580 static struct platform_driver moxart_mac_driver
= {
581 .probe
= moxart_mac_probe
,
582 .remove
= moxart_remove
,
584 .name
= "moxart-ethernet",
585 .of_match_table
= moxart_mac_match
,
588 module_platform_driver(moxart_mac_driver
);
590 MODULE_DESCRIPTION("MOXART RTL8201CP Ethernet driver");
591 MODULE_LICENSE("GPL v2");
592 MODULE_AUTHOR("Jonas Jensen <jonas.jensen@gmail.com>");