2 * Copyright (C) 2004-2013 Synopsys, Inc. (www.synopsys.com)
4 * This program is free software; you can redistribute it and/or modify
5 * it under the terms of the GNU General Public License version 2 as
6 * published by the Free Software Foundation.
8 * Driver for the ARC EMAC 10100 (hardware revision 5)
16 #include <linux/crc32.h>
17 #include <linux/etherdevice.h>
18 #include <linux/interrupt.h>
20 #include <linux/module.h>
21 #include <linux/of_address.h>
22 #include <linux/of_irq.h>
23 #include <linux/of_mdio.h>
24 #include <linux/of_net.h>
25 #include <linux/of_platform.h>
29 static void arc_emac_restart(struct net_device
*ndev
);
32 * arc_emac_tx_avail - Return the number of available slots in the tx ring.
33 * @priv: Pointer to ARC EMAC private data structure.
35 * returns: the number of slots available for transmission in tx the ring.
37 static inline int arc_emac_tx_avail(struct arc_emac_priv
*priv
)
39 return (priv
->txbd_dirty
+ TX_BD_NUM
- priv
->txbd_curr
- 1) % TX_BD_NUM
;
43 * arc_emac_adjust_link - Adjust the PHY link duplex.
44 * @ndev: Pointer to the net_device structure.
46 * This function is called to change the duplex setting after auto negotiation
49 static void arc_emac_adjust_link(struct net_device
*ndev
)
51 struct arc_emac_priv
*priv
= netdev_priv(ndev
);
52 struct phy_device
*phy_dev
= ndev
->phydev
;
53 unsigned int reg
, state_changed
= 0;
55 if (priv
->link
!= phy_dev
->link
) {
56 priv
->link
= phy_dev
->link
;
60 if (priv
->speed
!= phy_dev
->speed
) {
61 priv
->speed
= phy_dev
->speed
;
63 if (priv
->set_mac_speed
)
64 priv
->set_mac_speed(priv
, priv
->speed
);
67 if (priv
->duplex
!= phy_dev
->duplex
) {
68 reg
= arc_reg_get(priv
, R_CTRL
);
70 if (phy_dev
->duplex
== DUPLEX_FULL
)
75 arc_reg_set(priv
, R_CTRL
, reg
);
76 priv
->duplex
= phy_dev
->duplex
;
81 phy_print_status(phy_dev
);
85 * arc_emac_get_drvinfo - Get EMAC driver information.
86 * @ndev: Pointer to net_device structure.
87 * @info: Pointer to ethtool_drvinfo structure.
89 * This implements ethtool command for getting the driver information.
90 * Issue "ethtool -i ethX" under linux prompt to execute this function.
92 static void arc_emac_get_drvinfo(struct net_device
*ndev
,
93 struct ethtool_drvinfo
*info
)
95 struct arc_emac_priv
*priv
= netdev_priv(ndev
);
97 strlcpy(info
->driver
, priv
->drv_name
, sizeof(info
->driver
));
98 strlcpy(info
->version
, priv
->drv_version
, sizeof(info
->version
));
101 static const struct ethtool_ops arc_emac_ethtool_ops
= {
102 .get_drvinfo
= arc_emac_get_drvinfo
,
103 .get_link
= ethtool_op_get_link
,
104 .get_link_ksettings
= phy_ethtool_get_link_ksettings
,
105 .set_link_ksettings
= phy_ethtool_set_link_ksettings
,
108 #define FIRST_OR_LAST_MASK (FIRST_MASK | LAST_MASK)
111 * arc_emac_tx_clean - clears processed by EMAC Tx BDs.
112 * @ndev: Pointer to the network device.
114 static void arc_emac_tx_clean(struct net_device
*ndev
)
116 struct arc_emac_priv
*priv
= netdev_priv(ndev
);
117 struct net_device_stats
*stats
= &ndev
->stats
;
120 for (i
= 0; i
< TX_BD_NUM
; i
++) {
121 unsigned int *txbd_dirty
= &priv
->txbd_dirty
;
122 struct arc_emac_bd
*txbd
= &priv
->txbd
[*txbd_dirty
];
123 struct buffer_state
*tx_buff
= &priv
->tx_buff
[*txbd_dirty
];
124 struct sk_buff
*skb
= tx_buff
->skb
;
125 unsigned int info
= le32_to_cpu(txbd
->info
);
127 if ((info
& FOR_EMAC
) || !txbd
->data
|| !skb
)
130 if (unlikely(info
& (DROP
| DEFR
| LTCL
| UFLO
))) {
135 stats
->tx_carrier_errors
++;
141 stats
->tx_fifo_errors
++;
142 } else if (likely(info
& FIRST_OR_LAST_MASK
)) {
144 stats
->tx_bytes
+= skb
->len
;
147 dma_unmap_single(&ndev
->dev
, dma_unmap_addr(tx_buff
, addr
),
148 dma_unmap_len(tx_buff
, len
), DMA_TO_DEVICE
);
150 /* return the sk_buff to system */
151 dev_kfree_skb_irq(skb
);
157 *txbd_dirty
= (*txbd_dirty
+ 1) % TX_BD_NUM
;
160 /* Ensure that txbd_dirty is visible to tx() before checking
165 if (netif_queue_stopped(ndev
) && arc_emac_tx_avail(priv
))
166 netif_wake_queue(ndev
);
170 * arc_emac_rx - processing of Rx packets.
171 * @ndev: Pointer to the network device.
172 * @budget: How many BDs to process on 1 call.
174 * returns: Number of processed BDs
176 * Iterate through Rx BDs and deliver received packages to upper layer.
178 static int arc_emac_rx(struct net_device
*ndev
, int budget
)
180 struct arc_emac_priv
*priv
= netdev_priv(ndev
);
181 unsigned int work_done
;
183 for (work_done
= 0; work_done
< budget
; work_done
++) {
184 unsigned int *last_rx_bd
= &priv
->last_rx_bd
;
185 struct net_device_stats
*stats
= &ndev
->stats
;
186 struct buffer_state
*rx_buff
= &priv
->rx_buff
[*last_rx_bd
];
187 struct arc_emac_bd
*rxbd
= &priv
->rxbd
[*last_rx_bd
];
188 unsigned int pktlen
, info
= le32_to_cpu(rxbd
->info
);
192 if (unlikely((info
& OWN_MASK
) == FOR_EMAC
))
195 /* Make a note that we saw a packet at this BD.
196 * So next time, driver starts from this + 1
198 *last_rx_bd
= (*last_rx_bd
+ 1) % RX_BD_NUM
;
200 if (unlikely((info
& FIRST_OR_LAST_MASK
) !=
201 FIRST_OR_LAST_MASK
)) {
202 /* We pre-allocate buffers of MTU size so incoming
203 * packets won't be split/chained.
206 netdev_err(ndev
, "incomplete packet received\n");
208 /* Return ownership to EMAC */
209 rxbd
->info
= cpu_to_le32(FOR_EMAC
| EMAC_BUFFER_SIZE
);
211 stats
->rx_length_errors
++;
215 /* Prepare the BD for next cycle. netif_receive_skb()
216 * only if new skb was allocated and mapped to avoid holes
219 skb
= netdev_alloc_skb_ip_align(ndev
, EMAC_BUFFER_SIZE
);
220 if (unlikely(!skb
)) {
222 netdev_err(ndev
, "cannot allocate skb\n");
223 /* Return ownership to EMAC */
224 rxbd
->info
= cpu_to_le32(FOR_EMAC
| EMAC_BUFFER_SIZE
);
230 addr
= dma_map_single(&ndev
->dev
, (void *)skb
->data
,
231 EMAC_BUFFER_SIZE
, DMA_FROM_DEVICE
);
232 if (dma_mapping_error(&ndev
->dev
, addr
)) {
234 netdev_err(ndev
, "cannot map dma buffer\n");
236 /* Return ownership to EMAC */
237 rxbd
->info
= cpu_to_le32(FOR_EMAC
| EMAC_BUFFER_SIZE
);
243 /* unmap previosly mapped skb */
244 dma_unmap_single(&ndev
->dev
, dma_unmap_addr(rx_buff
, addr
),
245 dma_unmap_len(rx_buff
, len
), DMA_FROM_DEVICE
);
247 pktlen
= info
& LEN_MASK
;
249 stats
->rx_bytes
+= pktlen
;
250 skb_put(rx_buff
->skb
, pktlen
);
251 rx_buff
->skb
->dev
= ndev
;
252 rx_buff
->skb
->protocol
= eth_type_trans(rx_buff
->skb
, ndev
);
254 netif_receive_skb(rx_buff
->skb
);
257 dma_unmap_addr_set(rx_buff
, addr
, addr
);
258 dma_unmap_len_set(rx_buff
, len
, EMAC_BUFFER_SIZE
);
260 rxbd
->data
= cpu_to_le32(addr
);
262 /* Make sure pointer to data buffer is set */
265 /* Return ownership to EMAC */
266 rxbd
->info
= cpu_to_le32(FOR_EMAC
| EMAC_BUFFER_SIZE
);
273 * arc_emac_rx_miss_handle - handle R_MISS register
274 * @ndev: Pointer to the net_device structure.
276 static void arc_emac_rx_miss_handle(struct net_device
*ndev
)
278 struct arc_emac_priv
*priv
= netdev_priv(ndev
);
279 struct net_device_stats
*stats
= &ndev
->stats
;
282 miss
= arc_reg_get(priv
, R_MISS
);
284 stats
->rx_errors
+= miss
;
285 stats
->rx_missed_errors
+= miss
;
286 priv
->rx_missed_errors
+= miss
;
291 * arc_emac_rx_stall_check - check RX stall
292 * @ndev: Pointer to the net_device structure.
293 * @budget: How many BDs requested to process on 1 call.
294 * @work_done: How many BDs processed
296 * Under certain conditions EMAC stop reception of incoming packets and
297 * continuously increment R_MISS register instead of saving data into
298 * provided buffer. This function detect that condition and restart
301 static void arc_emac_rx_stall_check(struct net_device
*ndev
,
302 int budget
, unsigned int work_done
)
304 struct arc_emac_priv
*priv
= netdev_priv(ndev
);
305 struct arc_emac_bd
*rxbd
;
308 priv
->rx_missed_errors
= 0;
310 if (priv
->rx_missed_errors
&& budget
) {
311 rxbd
= &priv
->rxbd
[priv
->last_rx_bd
];
312 if (le32_to_cpu(rxbd
->info
) & FOR_EMAC
) {
313 arc_emac_restart(ndev
);
314 priv
->rx_missed_errors
= 0;
320 * arc_emac_poll - NAPI poll handler.
321 * @napi: Pointer to napi_struct structure.
322 * @budget: How many BDs to process on 1 call.
324 * returns: Number of processed BDs
326 static int arc_emac_poll(struct napi_struct
*napi
, int budget
)
328 struct net_device
*ndev
= napi
->dev
;
329 struct arc_emac_priv
*priv
= netdev_priv(ndev
);
330 unsigned int work_done
;
332 arc_emac_tx_clean(ndev
);
333 arc_emac_rx_miss_handle(ndev
);
335 work_done
= arc_emac_rx(ndev
, budget
);
336 if (work_done
< budget
) {
337 napi_complete_done(napi
, work_done
);
338 arc_reg_or(priv
, R_ENABLE
, RXINT_MASK
| TXINT_MASK
);
341 arc_emac_rx_stall_check(ndev
, budget
, work_done
);
347 * arc_emac_intr - Global interrupt handler for EMAC.
349 * @dev_instance: device instance.
351 * returns: IRQ_HANDLED for all cases.
353 * ARC EMAC has only 1 interrupt line, and depending on bits raised in
354 * STATUS register we may tell what is a reason for interrupt to fire.
356 static irqreturn_t
arc_emac_intr(int irq
, void *dev_instance
)
358 struct net_device
*ndev
= dev_instance
;
359 struct arc_emac_priv
*priv
= netdev_priv(ndev
);
360 struct net_device_stats
*stats
= &ndev
->stats
;
363 status
= arc_reg_get(priv
, R_STATUS
);
364 status
&= ~MDIO_MASK
;
366 /* Reset all flags except "MDIO complete" */
367 arc_reg_set(priv
, R_STATUS
, status
);
369 if (status
& (RXINT_MASK
| TXINT_MASK
)) {
370 if (likely(napi_schedule_prep(&priv
->napi
))) {
371 arc_reg_clr(priv
, R_ENABLE
, RXINT_MASK
| TXINT_MASK
);
372 __napi_schedule(&priv
->napi
);
376 if (status
& ERR_MASK
) {
377 /* MSER/RXCR/RXFR/RXFL interrupt fires on corresponding
378 * 8-bit error counter overrun.
381 if (status
& MSER_MASK
) {
382 stats
->rx_missed_errors
+= 0x100;
383 stats
->rx_errors
+= 0x100;
384 priv
->rx_missed_errors
+= 0x100;
385 napi_schedule(&priv
->napi
);
388 if (status
& RXCR_MASK
) {
389 stats
->rx_crc_errors
+= 0x100;
390 stats
->rx_errors
+= 0x100;
393 if (status
& RXFR_MASK
) {
394 stats
->rx_frame_errors
+= 0x100;
395 stats
->rx_errors
+= 0x100;
398 if (status
& RXFL_MASK
) {
399 stats
->rx_over_errors
+= 0x100;
400 stats
->rx_errors
+= 0x100;
407 #ifdef CONFIG_NET_POLL_CONTROLLER
408 static void arc_emac_poll_controller(struct net_device
*dev
)
410 disable_irq(dev
->irq
);
411 arc_emac_intr(dev
->irq
, dev
);
412 enable_irq(dev
->irq
);
417 * arc_emac_open - Open the network device.
418 * @ndev: Pointer to the network device.
420 * returns: 0, on success or non-zero error value on failure.
422 * This function sets the MAC address, requests and enables an IRQ
423 * for the EMAC device and starts the Tx queue.
424 * It also connects to the phy device.
426 static int arc_emac_open(struct net_device
*ndev
)
428 struct arc_emac_priv
*priv
= netdev_priv(ndev
);
429 struct phy_device
*phy_dev
= ndev
->phydev
;
432 phy_dev
->autoneg
= AUTONEG_ENABLE
;
435 phy_dev
->advertising
&= phy_dev
->supported
;
437 priv
->last_rx_bd
= 0;
439 /* Allocate and set buffers for Rx BD's */
440 for (i
= 0; i
< RX_BD_NUM
; i
++) {
442 unsigned int *last_rx_bd
= &priv
->last_rx_bd
;
443 struct arc_emac_bd
*rxbd
= &priv
->rxbd
[*last_rx_bd
];
444 struct buffer_state
*rx_buff
= &priv
->rx_buff
[*last_rx_bd
];
446 rx_buff
->skb
= netdev_alloc_skb_ip_align(ndev
,
448 if (unlikely(!rx_buff
->skb
))
451 addr
= dma_map_single(&ndev
->dev
, (void *)rx_buff
->skb
->data
,
452 EMAC_BUFFER_SIZE
, DMA_FROM_DEVICE
);
453 if (dma_mapping_error(&ndev
->dev
, addr
)) {
454 netdev_err(ndev
, "cannot dma map\n");
455 dev_kfree_skb(rx_buff
->skb
);
458 dma_unmap_addr_set(rx_buff
, addr
, addr
);
459 dma_unmap_len_set(rx_buff
, len
, EMAC_BUFFER_SIZE
);
461 rxbd
->data
= cpu_to_le32(addr
);
463 /* Make sure pointer to data buffer is set */
466 /* Return ownership to EMAC */
467 rxbd
->info
= cpu_to_le32(FOR_EMAC
| EMAC_BUFFER_SIZE
);
469 *last_rx_bd
= (*last_rx_bd
+ 1) % RX_BD_NUM
;
473 priv
->txbd_dirty
= 0;
476 memset(priv
->txbd
, 0, TX_RING_SZ
);
478 /* Initialize logical address filter */
479 arc_reg_set(priv
, R_LAFL
, 0);
480 arc_reg_set(priv
, R_LAFH
, 0);
482 /* Set BD ring pointers for device side */
483 arc_reg_set(priv
, R_RX_RING
, (unsigned int)priv
->rxbd_dma
);
484 arc_reg_set(priv
, R_TX_RING
, (unsigned int)priv
->txbd_dma
);
486 /* Enable interrupts */
487 arc_reg_set(priv
, R_ENABLE
, RXINT_MASK
| TXINT_MASK
| ERR_MASK
);
490 arc_reg_set(priv
, R_CTRL
,
491 (RX_BD_NUM
<< 24) | /* RX BD table length */
492 (TX_BD_NUM
<< 16) | /* TX BD table length */
493 TXRN_MASK
| RXRN_MASK
);
495 napi_enable(&priv
->napi
);
498 arc_reg_or(priv
, R_CTRL
, EN_MASK
);
500 phy_start(ndev
->phydev
);
502 netif_start_queue(ndev
);
508 * arc_emac_set_rx_mode - Change the receive filtering mode.
509 * @ndev: Pointer to the network device.
511 * This function enables/disables promiscuous or all-multicast mode
512 * and updates the multicast filtering list of the network device.
514 static void arc_emac_set_rx_mode(struct net_device
*ndev
)
516 struct arc_emac_priv
*priv
= netdev_priv(ndev
);
518 if (ndev
->flags
& IFF_PROMISC
) {
519 arc_reg_or(priv
, R_CTRL
, PROM_MASK
);
521 arc_reg_clr(priv
, R_CTRL
, PROM_MASK
);
523 if (ndev
->flags
& IFF_ALLMULTI
) {
524 arc_reg_set(priv
, R_LAFL
, ~0);
525 arc_reg_set(priv
, R_LAFH
, ~0);
526 } else if (ndev
->flags
& IFF_MULTICAST
) {
527 struct netdev_hw_addr
*ha
;
528 unsigned int filter
[2] = { 0, 0 };
531 netdev_for_each_mc_addr(ha
, ndev
) {
532 bit
= ether_crc_le(ETH_ALEN
, ha
->addr
) >> 26;
533 filter
[bit
>> 5] |= 1 << (bit
& 31);
536 arc_reg_set(priv
, R_LAFL
, filter
[0]);
537 arc_reg_set(priv
, R_LAFH
, filter
[1]);
539 arc_reg_set(priv
, R_LAFL
, 0);
540 arc_reg_set(priv
, R_LAFH
, 0);
546 * arc_free_tx_queue - free skb from tx queue
547 * @ndev: Pointer to the network device.
549 * This function must be called while EMAC disable
551 static void arc_free_tx_queue(struct net_device
*ndev
)
553 struct arc_emac_priv
*priv
= netdev_priv(ndev
);
556 for (i
= 0; i
< TX_BD_NUM
; i
++) {
557 struct arc_emac_bd
*txbd
= &priv
->txbd
[i
];
558 struct buffer_state
*tx_buff
= &priv
->tx_buff
[i
];
561 dma_unmap_single(&ndev
->dev
,
562 dma_unmap_addr(tx_buff
, addr
),
563 dma_unmap_len(tx_buff
, len
),
566 /* return the sk_buff to system */
567 dev_kfree_skb_irq(tx_buff
->skb
);
577 * arc_free_rx_queue - free skb from rx queue
578 * @ndev: Pointer to the network device.
580 * This function must be called while EMAC disable
582 static void arc_free_rx_queue(struct net_device
*ndev
)
584 struct arc_emac_priv
*priv
= netdev_priv(ndev
);
587 for (i
= 0; i
< RX_BD_NUM
; i
++) {
588 struct arc_emac_bd
*rxbd
= &priv
->rxbd
[i
];
589 struct buffer_state
*rx_buff
= &priv
->rx_buff
[i
];
592 dma_unmap_single(&ndev
->dev
,
593 dma_unmap_addr(rx_buff
, addr
),
594 dma_unmap_len(rx_buff
, len
),
597 /* return the sk_buff to system */
598 dev_kfree_skb_irq(rx_buff
->skb
);
608 * arc_emac_stop - Close the network device.
609 * @ndev: Pointer to the network device.
611 * This function stops the Tx queue, disables interrupts and frees the IRQ for
613 * It also disconnects the PHY device associated with the EMAC device.
615 static int arc_emac_stop(struct net_device
*ndev
)
617 struct arc_emac_priv
*priv
= netdev_priv(ndev
);
619 napi_disable(&priv
->napi
);
620 netif_stop_queue(ndev
);
622 phy_stop(ndev
->phydev
);
624 /* Disable interrupts */
625 arc_reg_clr(priv
, R_ENABLE
, RXINT_MASK
| TXINT_MASK
| ERR_MASK
);
628 arc_reg_clr(priv
, R_CTRL
, EN_MASK
);
630 /* Return the sk_buff to system */
631 arc_free_tx_queue(ndev
);
632 arc_free_rx_queue(ndev
);
638 * arc_emac_stats - Get system network statistics.
639 * @ndev: Pointer to net_device structure.
641 * Returns the address of the device statistics structure.
642 * Statistics are updated in interrupt handler.
644 static struct net_device_stats
*arc_emac_stats(struct net_device
*ndev
)
646 struct arc_emac_priv
*priv
= netdev_priv(ndev
);
647 struct net_device_stats
*stats
= &ndev
->stats
;
648 unsigned long miss
, rxerr
;
649 u8 rxcrc
, rxfram
, rxoflow
;
651 rxerr
= arc_reg_get(priv
, R_RXERR
);
652 miss
= arc_reg_get(priv
, R_MISS
);
656 rxoflow
= rxerr
>> 16;
658 stats
->rx_errors
+= miss
;
659 stats
->rx_errors
+= rxcrc
+ rxfram
+ rxoflow
;
661 stats
->rx_over_errors
+= rxoflow
;
662 stats
->rx_frame_errors
+= rxfram
;
663 stats
->rx_crc_errors
+= rxcrc
;
664 stats
->rx_missed_errors
+= miss
;
670 * arc_emac_tx - Starts the data transmission.
671 * @skb: sk_buff pointer that contains data to be Transmitted.
672 * @ndev: Pointer to net_device structure.
674 * returns: NETDEV_TX_OK, on success
675 * NETDEV_TX_BUSY, if any of the descriptors are not free.
677 * This function is invoked from upper layers to initiate transmission.
679 static int arc_emac_tx(struct sk_buff
*skb
, struct net_device
*ndev
)
681 struct arc_emac_priv
*priv
= netdev_priv(ndev
);
682 unsigned int len
, *txbd_curr
= &priv
->txbd_curr
;
683 struct net_device_stats
*stats
= &ndev
->stats
;
684 __le32
*info
= &priv
->txbd
[*txbd_curr
].info
;
687 if (skb_padto(skb
, ETH_ZLEN
))
690 len
= max_t(unsigned int, ETH_ZLEN
, skb
->len
);
692 if (unlikely(!arc_emac_tx_avail(priv
))) {
693 netif_stop_queue(ndev
);
694 netdev_err(ndev
, "BUG! Tx Ring full when queue awake!\n");
695 return NETDEV_TX_BUSY
;
698 addr
= dma_map_single(&ndev
->dev
, (void *)skb
->data
, len
,
701 if (unlikely(dma_mapping_error(&ndev
->dev
, addr
))) {
704 dev_kfree_skb_any(skb
);
707 dma_unmap_addr_set(&priv
->tx_buff
[*txbd_curr
], addr
, addr
);
708 dma_unmap_len_set(&priv
->tx_buff
[*txbd_curr
], len
, len
);
710 priv
->txbd
[*txbd_curr
].data
= cpu_to_le32(addr
);
712 /* Make sure pointer to data buffer is set */
715 skb_tx_timestamp(skb
);
717 *info
= cpu_to_le32(FOR_EMAC
| FIRST_OR_LAST_MASK
| len
);
719 /* Make sure info word is set */
722 priv
->tx_buff
[*txbd_curr
].skb
= skb
;
724 /* Increment index to point to the next BD */
725 *txbd_curr
= (*txbd_curr
+ 1) % TX_BD_NUM
;
727 /* Ensure that tx_clean() sees the new txbd_curr before
728 * checking the queue status. This prevents an unneeded wake
729 * of the queue in tx_clean().
733 if (!arc_emac_tx_avail(priv
)) {
734 netif_stop_queue(ndev
);
735 /* Refresh tx_dirty */
737 if (arc_emac_tx_avail(priv
))
738 netif_start_queue(ndev
);
741 arc_reg_set(priv
, R_STATUS
, TXPL_MASK
);
746 static void arc_emac_set_address_internal(struct net_device
*ndev
)
748 struct arc_emac_priv
*priv
= netdev_priv(ndev
);
749 unsigned int addr_low
, addr_hi
;
751 addr_low
= le32_to_cpu(*(__le32
*)&ndev
->dev_addr
[0]);
752 addr_hi
= le16_to_cpu(*(__le16
*)&ndev
->dev_addr
[4]);
754 arc_reg_set(priv
, R_ADDRL
, addr_low
);
755 arc_reg_set(priv
, R_ADDRH
, addr_hi
);
759 * arc_emac_set_address - Set the MAC address for this device.
760 * @ndev: Pointer to net_device structure.
761 * @p: 6 byte Address to be written as MAC address.
763 * This function copies the HW address from the sockaddr structure to the
764 * net_device structure and updates the address in HW.
766 * returns: -EBUSY if the net device is busy or 0 if the address is set
769 static int arc_emac_set_address(struct net_device
*ndev
, void *p
)
771 struct sockaddr
*addr
= p
;
773 if (netif_running(ndev
))
776 if (!is_valid_ether_addr(addr
->sa_data
))
777 return -EADDRNOTAVAIL
;
779 memcpy(ndev
->dev_addr
, addr
->sa_data
, ndev
->addr_len
);
781 arc_emac_set_address_internal(ndev
);
786 static int arc_emac_ioctl(struct net_device
*dev
, struct ifreq
*rq
, int cmd
)
788 if (!netif_running(dev
))
794 return phy_mii_ioctl(dev
->phydev
, rq
, cmd
);
799 * arc_emac_restart - Restart EMAC
800 * @ndev: Pointer to net_device structure.
802 * This function do hardware reset of EMAC in order to restore
803 * network packets reception.
805 static void arc_emac_restart(struct net_device
*ndev
)
807 struct arc_emac_priv
*priv
= netdev_priv(ndev
);
808 struct net_device_stats
*stats
= &ndev
->stats
;
812 netdev_warn(ndev
, "restarting stalled EMAC\n");
814 netif_stop_queue(ndev
);
816 /* Disable interrupts */
817 arc_reg_clr(priv
, R_ENABLE
, RXINT_MASK
| TXINT_MASK
| ERR_MASK
);
820 arc_reg_clr(priv
, R_CTRL
, EN_MASK
);
822 /* Return the sk_buff to system */
823 arc_free_tx_queue(ndev
);
827 priv
->txbd_dirty
= 0;
828 memset(priv
->txbd
, 0, TX_RING_SZ
);
830 for (i
= 0; i
< RX_BD_NUM
; i
++) {
831 struct arc_emac_bd
*rxbd
= &priv
->rxbd
[i
];
832 unsigned int info
= le32_to_cpu(rxbd
->info
);
834 if (!(info
& FOR_EMAC
)) {
838 /* Return ownership to EMAC */
839 rxbd
->info
= cpu_to_le32(FOR_EMAC
| EMAC_BUFFER_SIZE
);
841 priv
->last_rx_bd
= 0;
843 /* Make sure info is visible to EMAC before enable */
846 /* Enable interrupts */
847 arc_reg_set(priv
, R_ENABLE
, RXINT_MASK
| TXINT_MASK
| ERR_MASK
);
850 arc_reg_or(priv
, R_CTRL
, EN_MASK
);
852 netif_start_queue(ndev
);
855 static const struct net_device_ops arc_emac_netdev_ops
= {
856 .ndo_open
= arc_emac_open
,
857 .ndo_stop
= arc_emac_stop
,
858 .ndo_start_xmit
= arc_emac_tx
,
859 .ndo_set_mac_address
= arc_emac_set_address
,
860 .ndo_get_stats
= arc_emac_stats
,
861 .ndo_set_rx_mode
= arc_emac_set_rx_mode
,
862 .ndo_do_ioctl
= arc_emac_ioctl
,
863 #ifdef CONFIG_NET_POLL_CONTROLLER
864 .ndo_poll_controller
= arc_emac_poll_controller
,
868 int arc_emac_probe(struct net_device
*ndev
, int interface
)
870 struct device
*dev
= ndev
->dev
.parent
;
871 struct resource res_regs
;
872 struct device_node
*phy_node
;
873 struct phy_device
*phydev
= NULL
;
874 struct arc_emac_priv
*priv
;
875 const char *mac_addr
;
876 unsigned int id
, clock_frequency
, irq
;
879 /* Get PHY from device tree */
880 phy_node
= of_parse_phandle(dev
->of_node
, "phy", 0);
882 dev_err(dev
, "failed to retrieve phy description from device tree\n");
886 /* Get EMAC registers base address from device tree */
887 err
= of_address_to_resource(dev
->of_node
, 0, &res_regs
);
889 dev_err(dev
, "failed to retrieve registers base from device tree\n");
894 /* Get IRQ from device tree */
895 irq
= irq_of_parse_and_map(dev
->of_node
, 0);
897 dev_err(dev
, "failed to retrieve <irq> value from device tree\n");
902 ndev
->netdev_ops
= &arc_emac_netdev_ops
;
903 ndev
->ethtool_ops
= &arc_emac_ethtool_ops
;
904 ndev
->watchdog_timeo
= TX_TIMEOUT
;
906 priv
= netdev_priv(ndev
);
909 priv
->regs
= devm_ioremap_resource(dev
, &res_regs
);
910 if (IS_ERR(priv
->regs
)) {
911 err
= PTR_ERR(priv
->regs
);
915 dev_dbg(dev
, "Registers base address is 0x%p\n", priv
->regs
);
918 err
= clk_prepare_enable(priv
->clk
);
920 dev_err(dev
, "failed to enable clock\n");
924 clock_frequency
= clk_get_rate(priv
->clk
);
926 /* Get CPU clock frequency from device tree */
927 if (of_property_read_u32(dev
->of_node
, "clock-frequency",
929 dev_err(dev
, "failed to retrieve <clock-frequency> from device tree\n");
935 id
= arc_reg_get(priv
, R_ID
);
937 /* Check for EMAC revision 5 or 7, magic number */
938 if (!(id
== 0x0005fd02 || id
== 0x0007fd02)) {
939 dev_err(dev
, "ARC EMAC not detected, id=0x%x\n", id
);
943 dev_info(dev
, "ARC EMAC detected with id: 0x%x\n", id
);
945 /* Set poll rate so that it polls every 1 ms */
946 arc_reg_set(priv
, R_POLLRATE
, clock_frequency
/ 1000000);
949 dev_info(dev
, "IRQ is %d\n", ndev
->irq
);
951 /* Register interrupt handler for device */
952 err
= devm_request_irq(dev
, ndev
->irq
, arc_emac_intr
, 0,
955 dev_err(dev
, "could not allocate IRQ\n");
959 /* Get MAC address from device tree */
960 mac_addr
= of_get_mac_address(dev
->of_node
);
963 memcpy(ndev
->dev_addr
, mac_addr
, ETH_ALEN
);
965 eth_hw_addr_random(ndev
);
967 arc_emac_set_address_internal(ndev
);
968 dev_info(dev
, "MAC address is now %pM\n", ndev
->dev_addr
);
970 /* Do 1 allocation instead of 2 separate ones for Rx and Tx BD rings */
971 priv
->rxbd
= dmam_alloc_coherent(dev
, RX_RING_SZ
+ TX_RING_SZ
,
972 &priv
->rxbd_dma
, GFP_KERNEL
);
975 dev_err(dev
, "failed to allocate data buffers\n");
980 priv
->txbd
= priv
->rxbd
+ RX_BD_NUM
;
982 priv
->txbd_dma
= priv
->rxbd_dma
+ RX_RING_SZ
;
983 dev_dbg(dev
, "EMAC Device addr: Rx Ring [0x%x], Tx Ring[%x]\n",
984 (unsigned int)priv
->rxbd_dma
, (unsigned int)priv
->txbd_dma
);
986 err
= arc_mdio_probe(priv
);
988 dev_err(dev
, "failed to probe MII bus\n");
992 phydev
= of_phy_connect(ndev
, phy_node
, arc_emac_adjust_link
, 0,
995 dev_err(dev
, "of_phy_connect() failed\n");
1000 dev_info(dev
, "connected to %s phy with id 0x%x\n",
1001 phydev
->drv
->name
, phydev
->phy_id
);
1003 netif_napi_add(ndev
, &priv
->napi
, arc_emac_poll
, ARC_EMAC_NAPI_WEIGHT
);
1005 err
= register_netdev(ndev
);
1007 dev_err(dev
, "failed to register network device\n");
1011 of_node_put(phy_node
);
1015 netif_napi_del(&priv
->napi
);
1016 phy_disconnect(phydev
);
1018 arc_mdio_remove(priv
);
1021 clk_disable_unprepare(priv
->clk
);
1023 of_node_put(phy_node
);
1027 EXPORT_SYMBOL_GPL(arc_emac_probe
);
1029 int arc_emac_remove(struct net_device
*ndev
)
1031 struct arc_emac_priv
*priv
= netdev_priv(ndev
);
1033 phy_disconnect(ndev
->phydev
);
1034 arc_mdio_remove(priv
);
1035 unregister_netdev(ndev
);
1036 netif_napi_del(&priv
->napi
);
1038 if (!IS_ERR(priv
->clk
))
1039 clk_disable_unprepare(priv
->clk
);
1043 EXPORT_SYMBOL_GPL(arc_emac_remove
);
1045 MODULE_AUTHOR("Alexey Brodkin <abrodkin@synopsys.com>");
1046 MODULE_DESCRIPTION("ARC EMAC driver");
1047 MODULE_LICENSE("GPL");