1 // SPDX-License-Identifier: GPL-2.0-only
3 * FUJITSU Extended Socket Network Device driver
4 * Copyright (c) 2015 FUJITSU LIMITED
7 #include <linux/module.h>
8 #include <linux/types.h>
10 #include <linux/platform_device.h>
11 #include <linux/netdevice.h>
12 #include <linux/interrupt.h>
15 #include "fjes_trace.h"
19 #define DRV_VERSION __stringify(MAJ) "." __stringify(MIN)
20 #define DRV_NAME "fjes"
21 char fjes_driver_name
[] = DRV_NAME
;
22 char fjes_driver_version
[] = DRV_VERSION
;
23 static const char fjes_driver_string
[] =
24 "FUJITSU Extended Socket Network Device Driver";
25 static const char fjes_copyright
[] =
26 "Copyright (c) 2015 FUJITSU LIMITED";
28 MODULE_AUTHOR("Taku Izumi <izumi.taku@jp.fujitsu.com>");
29 MODULE_DESCRIPTION("FUJITSU Extended Socket Network Device Driver");
30 MODULE_LICENSE("GPL");
31 MODULE_VERSION(DRV_VERSION
);
33 #define ACPI_MOTHERBOARD_RESOURCE_HID "PNP0C02"
35 static int fjes_request_irq(struct fjes_adapter
*);
36 static void fjes_free_irq(struct fjes_adapter
*);
38 static int fjes_open(struct net_device
*);
39 static int fjes_close(struct net_device
*);
40 static int fjes_setup_resources(struct fjes_adapter
*);
41 static void fjes_free_resources(struct fjes_adapter
*);
42 static netdev_tx_t
fjes_xmit_frame(struct sk_buff
*, struct net_device
*);
43 static void fjes_raise_intr_rxdata_task(struct work_struct
*);
44 static void fjes_tx_stall_task(struct work_struct
*);
45 static void fjes_force_close_task(struct work_struct
*);
46 static irqreturn_t
fjes_intr(int, void*);
47 static void fjes_get_stats64(struct net_device
*, struct rtnl_link_stats64
*);
48 static int fjes_change_mtu(struct net_device
*, int);
49 static int fjes_vlan_rx_add_vid(struct net_device
*, __be16 proto
, u16
);
50 static int fjes_vlan_rx_kill_vid(struct net_device
*, __be16 proto
, u16
);
51 static void fjes_tx_retry(struct net_device
*);
53 static int fjes_acpi_add(struct acpi_device
*);
54 static int fjes_acpi_remove(struct acpi_device
*);
55 static acpi_status
fjes_get_acpi_resource(struct acpi_resource
*, void*);
57 static int fjes_probe(struct platform_device
*);
58 static int fjes_remove(struct platform_device
*);
60 static int fjes_sw_init(struct fjes_adapter
*);
61 static void fjes_netdev_setup(struct net_device
*);
62 static void fjes_irq_watch_task(struct work_struct
*);
63 static void fjes_watch_unshare_task(struct work_struct
*);
64 static void fjes_rx_irq(struct fjes_adapter
*, int);
65 static int fjes_poll(struct napi_struct
*, int);
67 static const struct acpi_device_id fjes_acpi_ids
[] = {
68 {ACPI_MOTHERBOARD_RESOURCE_HID
, 0},
71 MODULE_DEVICE_TABLE(acpi
, fjes_acpi_ids
);
73 static struct acpi_driver fjes_acpi_driver
= {
80 .remove
= fjes_acpi_remove
,
84 static struct platform_driver fjes_driver
= {
89 .remove
= fjes_remove
,
92 static struct resource fjes_resource
[] = {
94 .flags
= IORESOURCE_MEM
,
99 .flags
= IORESOURCE_IRQ
,
105 static bool is_extended_socket_device(struct acpi_device
*device
)
107 struct acpi_buffer buffer
= { ACPI_ALLOCATE_BUFFER
, NULL
};
108 char str_buf
[sizeof(FJES_ACPI_SYMBOL
) + 1];
109 union acpi_object
*str
;
113 status
= acpi_evaluate_object(device
->handle
, "_STR", NULL
, &buffer
);
114 if (ACPI_FAILURE(status
))
117 str
= buffer
.pointer
;
118 result
= utf16s_to_utf8s((wchar_t *)str
->string
.pointer
,
119 str
->string
.length
, UTF16_LITTLE_ENDIAN
,
120 str_buf
, sizeof(str_buf
) - 1);
123 if (strncmp(FJES_ACPI_SYMBOL
, str_buf
, strlen(FJES_ACPI_SYMBOL
)) != 0) {
124 kfree(buffer
.pointer
);
127 kfree(buffer
.pointer
);
132 static int acpi_check_extended_socket_status(struct acpi_device
*device
)
134 unsigned long long sta
;
137 status
= acpi_evaluate_integer(device
->handle
, "_STA", NULL
, &sta
);
138 if (ACPI_FAILURE(status
))
141 if (!((sta
& ACPI_STA_DEVICE_PRESENT
) &&
142 (sta
& ACPI_STA_DEVICE_ENABLED
) &&
143 (sta
& ACPI_STA_DEVICE_UI
) &&
144 (sta
& ACPI_STA_DEVICE_FUNCTIONING
)))
150 static int fjes_acpi_add(struct acpi_device
*device
)
152 struct platform_device
*plat_dev
;
155 if (!is_extended_socket_device(device
))
158 if (acpi_check_extended_socket_status(device
))
161 status
= acpi_walk_resources(device
->handle
, METHOD_NAME__CRS
,
162 fjes_get_acpi_resource
, fjes_resource
);
163 if (ACPI_FAILURE(status
))
166 /* create platform_device */
167 plat_dev
= platform_device_register_simple(DRV_NAME
, 0, fjes_resource
,
168 ARRAY_SIZE(fjes_resource
));
169 device
->driver_data
= plat_dev
;
174 static int fjes_acpi_remove(struct acpi_device
*device
)
176 struct platform_device
*plat_dev
;
178 plat_dev
= (struct platform_device
*)acpi_driver_data(device
);
179 platform_device_unregister(plat_dev
);
185 fjes_get_acpi_resource(struct acpi_resource
*acpi_res
, void *data
)
187 struct acpi_resource_address32
*addr
;
188 struct acpi_resource_irq
*irq
;
189 struct resource
*res
= data
;
191 switch (acpi_res
->type
) {
192 case ACPI_RESOURCE_TYPE_ADDRESS32
:
193 addr
= &acpi_res
->data
.address32
;
194 res
[0].start
= addr
->address
.minimum
;
195 res
[0].end
= addr
->address
.minimum
+
196 addr
->address
.address_length
- 1;
199 case ACPI_RESOURCE_TYPE_IRQ
:
200 irq
= &acpi_res
->data
.irq
;
201 if (irq
->interrupt_count
!= 1)
203 res
[1].start
= irq
->interrupts
[0];
204 res
[1].end
= irq
->interrupts
[0];
214 static int fjes_request_irq(struct fjes_adapter
*adapter
)
216 struct net_device
*netdev
= adapter
->netdev
;
219 adapter
->interrupt_watch_enable
= true;
220 if (!delayed_work_pending(&adapter
->interrupt_watch_task
)) {
221 queue_delayed_work(adapter
->control_wq
,
222 &adapter
->interrupt_watch_task
,
223 FJES_IRQ_WATCH_DELAY
);
226 if (!adapter
->irq_registered
) {
227 result
= request_irq(adapter
->hw
.hw_res
.irq
, fjes_intr
,
228 IRQF_SHARED
, netdev
->name
, adapter
);
230 adapter
->irq_registered
= false;
232 adapter
->irq_registered
= true;
238 static void fjes_free_irq(struct fjes_adapter
*adapter
)
240 struct fjes_hw
*hw
= &adapter
->hw
;
242 adapter
->interrupt_watch_enable
= false;
243 cancel_delayed_work_sync(&adapter
->interrupt_watch_task
);
245 fjes_hw_set_irqmask(hw
, REG_ICTL_MASK_ALL
, true);
247 if (adapter
->irq_registered
) {
248 free_irq(adapter
->hw
.hw_res
.irq
, adapter
);
249 adapter
->irq_registered
= false;
253 static const struct net_device_ops fjes_netdev_ops
= {
254 .ndo_open
= fjes_open
,
255 .ndo_stop
= fjes_close
,
256 .ndo_start_xmit
= fjes_xmit_frame
,
257 .ndo_get_stats64
= fjes_get_stats64
,
258 .ndo_change_mtu
= fjes_change_mtu
,
259 .ndo_tx_timeout
= fjes_tx_retry
,
260 .ndo_vlan_rx_add_vid
= fjes_vlan_rx_add_vid
,
261 .ndo_vlan_rx_kill_vid
= fjes_vlan_rx_kill_vid
,
264 /* fjes_open - Called when a network interface is made active */
265 static int fjes_open(struct net_device
*netdev
)
267 struct fjes_adapter
*adapter
= netdev_priv(netdev
);
268 struct fjes_hw
*hw
= &adapter
->hw
;
271 if (adapter
->open_guard
)
274 result
= fjes_setup_resources(adapter
);
278 hw
->txrx_stop_req_bit
= 0;
279 hw
->epstop_req_bit
= 0;
281 napi_enable(&adapter
->napi
);
283 fjes_hw_capture_interrupt_status(hw
);
285 result
= fjes_request_irq(adapter
);
289 fjes_hw_set_irqmask(hw
, REG_ICTL_MASK_ALL
, false);
291 netif_tx_start_all_queues(netdev
);
292 netif_carrier_on(netdev
);
297 fjes_free_irq(adapter
);
298 napi_disable(&adapter
->napi
);
301 fjes_free_resources(adapter
);
305 /* fjes_close - Disables a network interface */
306 static int fjes_close(struct net_device
*netdev
)
308 struct fjes_adapter
*adapter
= netdev_priv(netdev
);
309 struct fjes_hw
*hw
= &adapter
->hw
;
313 netif_tx_stop_all_queues(netdev
);
314 netif_carrier_off(netdev
);
316 fjes_hw_raise_epstop(hw
);
318 napi_disable(&adapter
->napi
);
320 spin_lock_irqsave(&hw
->rx_status_lock
, flags
);
321 for (epidx
= 0; epidx
< hw
->max_epid
; epidx
++) {
322 if (epidx
== hw
->my_epid
)
325 if (fjes_hw_get_partner_ep_status(hw
, epidx
) ==
327 adapter
->hw
.ep_shm_info
[epidx
]
328 .tx
.info
->v1i
.rx_status
&=
331 spin_unlock_irqrestore(&hw
->rx_status_lock
, flags
);
333 fjes_free_irq(adapter
);
335 cancel_delayed_work_sync(&adapter
->interrupt_watch_task
);
336 cancel_work_sync(&adapter
->unshare_watch_task
);
337 adapter
->unshare_watch_bitmask
= 0;
338 cancel_work_sync(&adapter
->raise_intr_rxdata_task
);
339 cancel_work_sync(&adapter
->tx_stall_task
);
341 cancel_work_sync(&hw
->update_zone_task
);
342 cancel_work_sync(&hw
->epstop_task
);
344 fjes_hw_wait_epstop(hw
);
346 fjes_free_resources(adapter
);
351 static int fjes_setup_resources(struct fjes_adapter
*adapter
)
353 struct net_device
*netdev
= adapter
->netdev
;
354 struct ep_share_mem_info
*buf_pair
;
355 struct fjes_hw
*hw
= &adapter
->hw
;
360 mutex_lock(&hw
->hw_info
.lock
);
361 result
= fjes_hw_request_info(hw
);
364 for (epidx
= 0; epidx
< hw
->max_epid
; epidx
++) {
365 hw
->ep_shm_info
[epidx
].es_status
=
366 hw
->hw_info
.res_buf
->info
.info
[epidx
].es_status
;
367 hw
->ep_shm_info
[epidx
].zone
=
368 hw
->hw_info
.res_buf
->info
.info
[epidx
].zone
;
374 adapter
->force_reset
= true;
376 mutex_unlock(&hw
->hw_info
.lock
);
379 mutex_unlock(&hw
->hw_info
.lock
);
381 for (epidx
= 0; epidx
< (hw
->max_epid
); epidx
++) {
382 if ((epidx
!= hw
->my_epid
) &&
383 (hw
->ep_shm_info
[epidx
].es_status
==
384 FJES_ZONING_STATUS_ENABLE
)) {
385 fjes_hw_raise_interrupt(hw
, epidx
,
386 REG_ICTL_MASK_INFO_UPDATE
);
387 hw
->ep_shm_info
[epidx
].ep_stats
388 .send_intr_zoneupdate
+= 1;
392 msleep(FJES_OPEN_ZONE_UPDATE_WAIT
* hw
->max_epid
);
394 for (epidx
= 0; epidx
< (hw
->max_epid
); epidx
++) {
395 if (epidx
== hw
->my_epid
)
398 buf_pair
= &hw
->ep_shm_info
[epidx
];
400 spin_lock_irqsave(&hw
->rx_status_lock
, flags
);
401 fjes_hw_setup_epbuf(&buf_pair
->tx
, netdev
->dev_addr
,
403 spin_unlock_irqrestore(&hw
->rx_status_lock
, flags
);
405 if (fjes_hw_epid_is_same_zone(hw
, epidx
)) {
406 mutex_lock(&hw
->hw_info
.lock
);
408 fjes_hw_register_buff_addr(hw
, epidx
, buf_pair
);
409 mutex_unlock(&hw
->hw_info
.lock
);
417 adapter
->force_reset
= true;
421 hw
->ep_shm_info
[epidx
].ep_stats
422 .com_regist_buf_exec
+= 1;
429 static void fjes_free_resources(struct fjes_adapter
*adapter
)
431 struct net_device
*netdev
= adapter
->netdev
;
432 struct fjes_device_command_param param
;
433 struct ep_share_mem_info
*buf_pair
;
434 struct fjes_hw
*hw
= &adapter
->hw
;
435 bool reset_flag
= false;
440 for (epidx
= 0; epidx
< hw
->max_epid
; epidx
++) {
441 if (epidx
== hw
->my_epid
)
444 mutex_lock(&hw
->hw_info
.lock
);
445 result
= fjes_hw_unregister_buff_addr(hw
, epidx
);
446 mutex_unlock(&hw
->hw_info
.lock
);
448 hw
->ep_shm_info
[epidx
].ep_stats
.com_unregist_buf_exec
+= 1;
453 buf_pair
= &hw
->ep_shm_info
[epidx
];
455 spin_lock_irqsave(&hw
->rx_status_lock
, flags
);
456 fjes_hw_setup_epbuf(&buf_pair
->tx
,
457 netdev
->dev_addr
, netdev
->mtu
);
458 spin_unlock_irqrestore(&hw
->rx_status_lock
, flags
);
460 clear_bit(epidx
, &hw
->txrx_stop_req_bit
);
463 if (reset_flag
|| adapter
->force_reset
) {
464 result
= fjes_hw_reset(hw
);
466 adapter
->force_reset
= false;
469 adapter
->open_guard
= true;
471 hw
->hw_info
.buffer_share_bit
= 0;
473 memset((void *)¶m
, 0, sizeof(param
));
475 param
.req_len
= hw
->hw_info
.req_buf_size
;
476 param
.req_start
= __pa(hw
->hw_info
.req_buf
);
477 param
.res_len
= hw
->hw_info
.res_buf_size
;
478 param
.res_start
= __pa(hw
->hw_info
.res_buf
);
479 param
.share_start
= __pa(hw
->hw_info
.share
->ep_status
);
481 fjes_hw_init_command_registers(hw
, ¶m
);
485 static void fjes_tx_stall_task(struct work_struct
*work
)
487 struct fjes_adapter
*adapter
= container_of(work
,
488 struct fjes_adapter
, tx_stall_task
);
489 struct net_device
*netdev
= adapter
->netdev
;
490 struct fjes_hw
*hw
= &adapter
->hw
;
491 int all_queue_available
, sendable
;
492 enum ep_partner_status pstatus
;
493 int max_epid
, my_epid
, epid
;
494 union ep_buffer_info
*info
;
498 dev_trans_start(netdev
)) > FJES_TX_TX_STALL_TIMEOUT
) {
499 netif_wake_queue(netdev
);
503 my_epid
= hw
->my_epid
;
504 max_epid
= hw
->max_epid
;
506 for (i
= 0; i
< 5; i
++) {
507 all_queue_available
= 1;
509 for (epid
= 0; epid
< max_epid
; epid
++) {
513 pstatus
= fjes_hw_get_partner_ep_status(hw
, epid
);
514 sendable
= (pstatus
== EP_PARTNER_SHARED
);
518 info
= adapter
->hw
.ep_shm_info
[epid
].tx
.info
;
520 if (!(info
->v1i
.rx_status
& FJES_RX_MTU_CHANGING_DONE
))
523 if (EP_RING_FULL(info
->v1i
.head
, info
->v1i
.tail
,
524 info
->v1i
.count_max
)) {
525 all_queue_available
= 0;
530 if (all_queue_available
) {
531 netif_wake_queue(netdev
);
536 usleep_range(50, 100);
538 queue_work(adapter
->txrx_wq
, &adapter
->tx_stall_task
);
541 static void fjes_force_close_task(struct work_struct
*work
)
543 struct fjes_adapter
*adapter
= container_of(work
,
544 struct fjes_adapter
, force_close_task
);
545 struct net_device
*netdev
= adapter
->netdev
;
552 static void fjes_raise_intr_rxdata_task(struct work_struct
*work
)
554 struct fjes_adapter
*adapter
= container_of(work
,
555 struct fjes_adapter
, raise_intr_rxdata_task
);
556 struct fjes_hw
*hw
= &adapter
->hw
;
557 enum ep_partner_status pstatus
;
558 int max_epid
, my_epid
, epid
;
560 my_epid
= hw
->my_epid
;
561 max_epid
= hw
->max_epid
;
563 for (epid
= 0; epid
< max_epid
; epid
++)
564 hw
->ep_shm_info
[epid
].tx_status_work
= 0;
566 for (epid
= 0; epid
< max_epid
; epid
++) {
570 pstatus
= fjes_hw_get_partner_ep_status(hw
, epid
);
571 if (pstatus
== EP_PARTNER_SHARED
) {
572 hw
->ep_shm_info
[epid
].tx_status_work
=
573 hw
->ep_shm_info
[epid
].tx
.info
->v1i
.tx_status
;
575 if (hw
->ep_shm_info
[epid
].tx_status_work
==
576 FJES_TX_DELAY_SEND_PENDING
) {
577 hw
->ep_shm_info
[epid
].tx
.info
->v1i
.tx_status
=
578 FJES_TX_DELAY_SEND_NONE
;
583 for (epid
= 0; epid
< max_epid
; epid
++) {
587 pstatus
= fjes_hw_get_partner_ep_status(hw
, epid
);
588 if ((hw
->ep_shm_info
[epid
].tx_status_work
==
589 FJES_TX_DELAY_SEND_PENDING
) &&
590 (pstatus
== EP_PARTNER_SHARED
) &&
591 !(hw
->ep_shm_info
[epid
].rx
.info
->v1i
.rx_status
&
592 FJES_RX_POLL_WORK
)) {
593 fjes_hw_raise_interrupt(hw
, epid
,
594 REG_ICTL_MASK_RX_DATA
);
595 hw
->ep_shm_info
[epid
].ep_stats
.send_intr_rx
+= 1;
599 usleep_range(500, 1000);
602 static int fjes_tx_send(struct fjes_adapter
*adapter
, int dest
,
603 void *data
, size_t len
)
607 retval
= fjes_hw_epbuf_tx_pkt_send(&adapter
->hw
.ep_shm_info
[dest
].tx
,
612 adapter
->hw
.ep_shm_info
[dest
].tx
.info
->v1i
.tx_status
=
613 FJES_TX_DELAY_SEND_PENDING
;
614 if (!work_pending(&adapter
->raise_intr_rxdata_task
))
615 queue_work(adapter
->txrx_wq
,
616 &adapter
->raise_intr_rxdata_task
);
623 fjes_xmit_frame(struct sk_buff
*skb
, struct net_device
*netdev
)
625 struct fjes_adapter
*adapter
= netdev_priv(netdev
);
626 struct fjes_hw
*hw
= &adapter
->hw
;
628 int max_epid
, my_epid
, dest_epid
;
629 enum ep_partner_status pstatus
;
630 struct netdev_queue
*cur_queue
;
631 char shortpkt
[VLAN_ETH_HLEN
];
642 cur_queue
= netdev_get_tx_queue(netdev
, queue_no
);
644 eth
= (struct ethhdr
*)skb
->data
;
645 my_epid
= hw
->my_epid
;
647 vlan
= (vlan_get_tag(skb
, &vlan_id
) == 0) ? true : false;
652 if (is_multicast_ether_addr(eth
->h_dest
)) {
654 max_epid
= hw
->max_epid
;
656 } else if (is_local_ether_addr(eth
->h_dest
)) {
657 dest_epid
= eth
->h_dest
[ETH_ALEN
- 1];
658 max_epid
= dest_epid
+ 1;
660 if ((eth
->h_dest
[0] == 0x02) &&
661 (0x00 == (eth
->h_dest
[1] | eth
->h_dest
[2] |
662 eth
->h_dest
[3] | eth
->h_dest
[4])) &&
663 (dest_epid
< hw
->max_epid
)) {
670 adapter
->stats64
.tx_packets
+= 1;
671 hw
->ep_shm_info
[my_epid
].net_stats
.tx_packets
+= 1;
672 adapter
->stats64
.tx_bytes
+= len
;
673 hw
->ep_shm_info
[my_epid
].net_stats
.tx_bytes
+= len
;
680 adapter
->stats64
.tx_packets
+= 1;
681 hw
->ep_shm_info
[my_epid
].net_stats
.tx_packets
+= 1;
682 adapter
->stats64
.tx_bytes
+= len
;
683 hw
->ep_shm_info
[my_epid
].net_stats
.tx_bytes
+= len
;
686 for (; dest_epid
< max_epid
; dest_epid
++) {
687 if (my_epid
== dest_epid
)
690 pstatus
= fjes_hw_get_partner_ep_status(hw
, dest_epid
);
691 if (pstatus
!= EP_PARTNER_SHARED
) {
693 hw
->ep_shm_info
[dest_epid
].ep_stats
694 .tx_dropped_not_shared
+= 1;
696 } else if (!fjes_hw_check_epbuf_version(
697 &adapter
->hw
.ep_shm_info
[dest_epid
].rx
, 0)) {
698 /* version is NOT 0 */
699 adapter
->stats64
.tx_carrier_errors
+= 1;
700 hw
->ep_shm_info
[dest_epid
].net_stats
701 .tx_carrier_errors
+= 1;
702 hw
->ep_shm_info
[dest_epid
].ep_stats
703 .tx_dropped_ver_mismatch
+= 1;
706 } else if (!fjes_hw_check_mtu(
707 &adapter
->hw
.ep_shm_info
[dest_epid
].rx
,
709 adapter
->stats64
.tx_dropped
+= 1;
710 hw
->ep_shm_info
[dest_epid
].net_stats
.tx_dropped
+= 1;
711 adapter
->stats64
.tx_errors
+= 1;
712 hw
->ep_shm_info
[dest_epid
].net_stats
.tx_errors
+= 1;
713 hw
->ep_shm_info
[dest_epid
].ep_stats
714 .tx_dropped_buf_size_mismatch
+= 1;
718 !fjes_hw_check_vlan_id(
719 &adapter
->hw
.ep_shm_info
[dest_epid
].rx
,
721 hw
->ep_shm_info
[dest_epid
].ep_stats
722 .tx_dropped_vlanid_mismatch
+= 1;
725 if (len
< VLAN_ETH_HLEN
) {
726 memset(shortpkt
, 0, VLAN_ETH_HLEN
);
727 memcpy(shortpkt
, skb
->data
, skb
->len
);
732 if (adapter
->tx_retry_count
== 0) {
733 adapter
->tx_start_jiffies
= jiffies
;
734 adapter
->tx_retry_count
= 1;
736 adapter
->tx_retry_count
++;
739 if (fjes_tx_send(adapter
, dest_epid
, data
, len
)) {
744 (long)adapter
->tx_start_jiffies
) >=
745 FJES_TX_RETRY_TIMEOUT
) {
746 adapter
->stats64
.tx_fifo_errors
+= 1;
747 hw
->ep_shm_info
[dest_epid
].net_stats
748 .tx_fifo_errors
+= 1;
749 adapter
->stats64
.tx_errors
+= 1;
750 hw
->ep_shm_info
[dest_epid
].net_stats
755 netif_trans_update(netdev
);
756 hw
->ep_shm_info
[dest_epid
].ep_stats
757 .tx_buffer_full
+= 1;
758 netif_tx_stop_queue(cur_queue
);
760 if (!work_pending(&adapter
->tx_stall_task
))
761 queue_work(adapter
->txrx_wq
,
762 &adapter
->tx_stall_task
);
764 ret
= NETDEV_TX_BUSY
;
768 adapter
->stats64
.tx_packets
+= 1;
769 hw
->ep_shm_info
[dest_epid
].net_stats
771 adapter
->stats64
.tx_bytes
+= len
;
772 hw
->ep_shm_info
[dest_epid
].net_stats
776 adapter
->tx_retry_count
= 0;
782 if (ret
== NETDEV_TX_OK
) {
785 adapter
->stats64
.tx_packets
+= 1;
786 hw
->ep_shm_info
[my_epid
].net_stats
.tx_packets
+= 1;
787 adapter
->stats64
.tx_bytes
+= 1;
788 hw
->ep_shm_info
[my_epid
].net_stats
.tx_bytes
+= len
;
795 static void fjes_tx_retry(struct net_device
*netdev
)
797 struct netdev_queue
*queue
= netdev_get_tx_queue(netdev
, 0);
799 netif_tx_wake_queue(queue
);
803 fjes_get_stats64(struct net_device
*netdev
, struct rtnl_link_stats64
*stats
)
805 struct fjes_adapter
*adapter
= netdev_priv(netdev
);
807 memcpy(stats
, &adapter
->stats64
, sizeof(struct rtnl_link_stats64
));
810 static int fjes_change_mtu(struct net_device
*netdev
, int new_mtu
)
812 struct fjes_adapter
*adapter
= netdev_priv(netdev
);
813 bool running
= netif_running(netdev
);
814 struct fjes_hw
*hw
= &adapter
->hw
;
819 for (idx
= 0; fjes_support_mtu
[idx
] != 0; idx
++) {
820 if (new_mtu
<= fjes_support_mtu
[idx
]) {
821 new_mtu
= fjes_support_mtu
[idx
];
822 if (new_mtu
== netdev
->mtu
)
834 spin_lock_irqsave(&hw
->rx_status_lock
, flags
);
835 for (epidx
= 0; epidx
< hw
->max_epid
; epidx
++) {
836 if (epidx
== hw
->my_epid
)
838 hw
->ep_shm_info
[epidx
].tx
.info
->v1i
.rx_status
&=
839 ~FJES_RX_MTU_CHANGING_DONE
;
841 spin_unlock_irqrestore(&hw
->rx_status_lock
, flags
);
843 netif_tx_stop_all_queues(netdev
);
844 netif_carrier_off(netdev
);
845 cancel_work_sync(&adapter
->tx_stall_task
);
846 napi_disable(&adapter
->napi
);
850 netif_tx_stop_all_queues(netdev
);
853 netdev
->mtu
= new_mtu
;
856 for (epidx
= 0; epidx
< hw
->max_epid
; epidx
++) {
857 if (epidx
== hw
->my_epid
)
860 spin_lock_irqsave(&hw
->rx_status_lock
, flags
);
861 fjes_hw_setup_epbuf(&hw
->ep_shm_info
[epidx
].tx
,
865 hw
->ep_shm_info
[epidx
].tx
.info
->v1i
.rx_status
|=
866 FJES_RX_MTU_CHANGING_DONE
;
867 spin_unlock_irqrestore(&hw
->rx_status_lock
, flags
);
870 netif_tx_wake_all_queues(netdev
);
871 netif_carrier_on(netdev
);
872 napi_enable(&adapter
->napi
);
873 napi_schedule(&adapter
->napi
);
879 static int fjes_vlan_rx_add_vid(struct net_device
*netdev
,
880 __be16 proto
, u16 vid
)
882 struct fjes_adapter
*adapter
= netdev_priv(netdev
);
886 for (epid
= 0; epid
< adapter
->hw
.max_epid
; epid
++) {
887 if (epid
== adapter
->hw
.my_epid
)
890 if (!fjes_hw_check_vlan_id(
891 &adapter
->hw
.ep_shm_info
[epid
].tx
, vid
))
892 ret
= fjes_hw_set_vlan_id(
893 &adapter
->hw
.ep_shm_info
[epid
].tx
, vid
);
896 return ret
? 0 : -ENOSPC
;
899 static int fjes_vlan_rx_kill_vid(struct net_device
*netdev
,
900 __be16 proto
, u16 vid
)
902 struct fjes_adapter
*adapter
= netdev_priv(netdev
);
905 for (epid
= 0; epid
< adapter
->hw
.max_epid
; epid
++) {
906 if (epid
== adapter
->hw
.my_epid
)
909 fjes_hw_del_vlan_id(&adapter
->hw
.ep_shm_info
[epid
].tx
, vid
);
915 static void fjes_txrx_stop_req_irq(struct fjes_adapter
*adapter
,
918 struct fjes_hw
*hw
= &adapter
->hw
;
919 enum ep_partner_status status
;
922 status
= fjes_hw_get_partner_ep_status(hw
, src_epid
);
923 trace_fjes_txrx_stop_req_irq_pre(hw
, src_epid
, status
);
925 case EP_PARTNER_UNSHARE
:
926 case EP_PARTNER_COMPLETE
:
929 case EP_PARTNER_WAITING
:
930 if (src_epid
< hw
->my_epid
) {
931 spin_lock_irqsave(&hw
->rx_status_lock
, flags
);
932 hw
->ep_shm_info
[src_epid
].tx
.info
->v1i
.rx_status
|=
933 FJES_RX_STOP_REQ_DONE
;
934 spin_unlock_irqrestore(&hw
->rx_status_lock
, flags
);
936 clear_bit(src_epid
, &hw
->txrx_stop_req_bit
);
937 set_bit(src_epid
, &adapter
->unshare_watch_bitmask
);
939 if (!work_pending(&adapter
->unshare_watch_task
))
940 queue_work(adapter
->control_wq
,
941 &adapter
->unshare_watch_task
);
944 case EP_PARTNER_SHARED
:
945 if (hw
->ep_shm_info
[src_epid
].rx
.info
->v1i
.rx_status
&
946 FJES_RX_STOP_REQ_REQUEST
) {
947 set_bit(src_epid
, &hw
->epstop_req_bit
);
948 if (!work_pending(&hw
->epstop_task
))
949 queue_work(adapter
->control_wq
,
954 trace_fjes_txrx_stop_req_irq_post(hw
, src_epid
);
957 static void fjes_stop_req_irq(struct fjes_adapter
*adapter
, int src_epid
)
959 struct fjes_hw
*hw
= &adapter
->hw
;
960 enum ep_partner_status status
;
963 set_bit(src_epid
, &hw
->hw_info
.buffer_unshare_reserve_bit
);
965 status
= fjes_hw_get_partner_ep_status(hw
, src_epid
);
966 trace_fjes_stop_req_irq_pre(hw
, src_epid
, status
);
968 case EP_PARTNER_WAITING
:
969 spin_lock_irqsave(&hw
->rx_status_lock
, flags
);
970 hw
->ep_shm_info
[src_epid
].tx
.info
->v1i
.rx_status
|=
971 FJES_RX_STOP_REQ_DONE
;
972 spin_unlock_irqrestore(&hw
->rx_status_lock
, flags
);
973 clear_bit(src_epid
, &hw
->txrx_stop_req_bit
);
975 case EP_PARTNER_UNSHARE
:
976 case EP_PARTNER_COMPLETE
:
978 set_bit(src_epid
, &adapter
->unshare_watch_bitmask
);
979 if (!work_pending(&adapter
->unshare_watch_task
))
980 queue_work(adapter
->control_wq
,
981 &adapter
->unshare_watch_task
);
983 case EP_PARTNER_SHARED
:
984 set_bit(src_epid
, &hw
->epstop_req_bit
);
986 if (!work_pending(&hw
->epstop_task
))
987 queue_work(adapter
->control_wq
, &hw
->epstop_task
);
990 trace_fjes_stop_req_irq_post(hw
, src_epid
);
993 static void fjes_update_zone_irq(struct fjes_adapter
*adapter
,
996 struct fjes_hw
*hw
= &adapter
->hw
;
998 if (!work_pending(&hw
->update_zone_task
))
999 queue_work(adapter
->control_wq
, &hw
->update_zone_task
);
1002 static irqreturn_t
fjes_intr(int irq
, void *data
)
1004 struct fjes_adapter
*adapter
= data
;
1005 struct fjes_hw
*hw
= &adapter
->hw
;
1009 icr
= fjes_hw_capture_interrupt_status(hw
);
1011 if (icr
& REG_IS_MASK_IS_ASSERT
) {
1012 if (icr
& REG_ICTL_MASK_RX_DATA
) {
1013 fjes_rx_irq(adapter
, icr
& REG_IS_MASK_EPID
);
1014 hw
->ep_shm_info
[icr
& REG_IS_MASK_EPID
].ep_stats
1018 if (icr
& REG_ICTL_MASK_DEV_STOP_REQ
) {
1019 fjes_stop_req_irq(adapter
, icr
& REG_IS_MASK_EPID
);
1020 hw
->ep_shm_info
[icr
& REG_IS_MASK_EPID
].ep_stats
1021 .recv_intr_stop
+= 1;
1024 if (icr
& REG_ICTL_MASK_TXRX_STOP_REQ
) {
1025 fjes_txrx_stop_req_irq(adapter
, icr
& REG_IS_MASK_EPID
);
1026 hw
->ep_shm_info
[icr
& REG_IS_MASK_EPID
].ep_stats
1027 .recv_intr_unshare
+= 1;
1030 if (icr
& REG_ICTL_MASK_TXRX_STOP_DONE
)
1031 fjes_hw_set_irqmask(hw
,
1032 REG_ICTL_MASK_TXRX_STOP_DONE
, true);
1034 if (icr
& REG_ICTL_MASK_INFO_UPDATE
) {
1035 fjes_update_zone_irq(adapter
, icr
& REG_IS_MASK_EPID
);
1036 hw
->ep_shm_info
[icr
& REG_IS_MASK_EPID
].ep_stats
1037 .recv_intr_zoneupdate
+= 1;
1048 static int fjes_rxframe_search_exist(struct fjes_adapter
*adapter
,
1051 struct fjes_hw
*hw
= &adapter
->hw
;
1052 enum ep_partner_status pstatus
;
1053 int max_epid
, cur_epid
;
1056 max_epid
= hw
->max_epid
;
1057 start_epid
= (start_epid
+ 1 + max_epid
) % max_epid
;
1059 for (i
= 0; i
< max_epid
; i
++) {
1060 cur_epid
= (start_epid
+ i
) % max_epid
;
1061 if (cur_epid
== hw
->my_epid
)
1064 pstatus
= fjes_hw_get_partner_ep_status(hw
, cur_epid
);
1065 if (pstatus
== EP_PARTNER_SHARED
) {
1066 if (!fjes_hw_epbuf_rx_is_empty(
1067 &hw
->ep_shm_info
[cur_epid
].rx
))
1074 static void *fjes_rxframe_get(struct fjes_adapter
*adapter
, size_t *psize
,
1079 *cur_epid
= fjes_rxframe_search_exist(adapter
, *cur_epid
);
1084 fjes_hw_epbuf_rx_curpkt_get_addr(
1085 &adapter
->hw
.ep_shm_info
[*cur_epid
].rx
, psize
);
1090 static void fjes_rxframe_release(struct fjes_adapter
*adapter
, int cur_epid
)
1092 fjes_hw_epbuf_rx_curpkt_drop(&adapter
->hw
.ep_shm_info
[cur_epid
].rx
);
1095 static void fjes_rx_irq(struct fjes_adapter
*adapter
, int src_epid
)
1097 struct fjes_hw
*hw
= &adapter
->hw
;
1099 fjes_hw_set_irqmask(hw
, REG_ICTL_MASK_RX_DATA
, true);
1101 adapter
->unset_rx_last
= true;
1102 napi_schedule(&adapter
->napi
);
1105 static int fjes_poll(struct napi_struct
*napi
, int budget
)
1107 struct fjes_adapter
*adapter
=
1108 container_of(napi
, struct fjes_adapter
, napi
);
1109 struct net_device
*netdev
= napi
->dev
;
1110 struct fjes_hw
*hw
= &adapter
->hw
;
1111 struct sk_buff
*skb
;
1118 spin_lock(&hw
->rx_status_lock
);
1119 for (epidx
= 0; epidx
< hw
->max_epid
; epidx
++) {
1120 if (epidx
== hw
->my_epid
)
1123 if (fjes_hw_get_partner_ep_status(hw
, epidx
) ==
1125 adapter
->hw
.ep_shm_info
[epidx
]
1126 .tx
.info
->v1i
.rx_status
|= FJES_RX_POLL_WORK
;
1128 spin_unlock(&hw
->rx_status_lock
);
1130 while (work_done
< budget
) {
1131 prefetch(&adapter
->hw
);
1132 frame
= fjes_rxframe_get(adapter
, &frame_len
, &cur_epid
);
1135 skb
= napi_alloc_skb(napi
, frame_len
);
1137 adapter
->stats64
.rx_dropped
+= 1;
1138 hw
->ep_shm_info
[cur_epid
].net_stats
1140 adapter
->stats64
.rx_errors
+= 1;
1141 hw
->ep_shm_info
[cur_epid
].net_stats
1144 skb_put_data(skb
, frame
, frame_len
);
1145 skb
->protocol
= eth_type_trans(skb
, netdev
);
1146 skb
->ip_summed
= CHECKSUM_UNNECESSARY
;
1148 netif_receive_skb(skb
);
1152 adapter
->stats64
.rx_packets
+= 1;
1153 hw
->ep_shm_info
[cur_epid
].net_stats
1155 adapter
->stats64
.rx_bytes
+= frame_len
;
1156 hw
->ep_shm_info
[cur_epid
].net_stats
1157 .rx_bytes
+= frame_len
;
1159 if (is_multicast_ether_addr(
1160 ((struct ethhdr
*)frame
)->h_dest
)) {
1161 adapter
->stats64
.multicast
+= 1;
1162 hw
->ep_shm_info
[cur_epid
].net_stats
1167 fjes_rxframe_release(adapter
, cur_epid
);
1168 adapter
->unset_rx_last
= true;
1174 if (work_done
< budget
) {
1175 napi_complete_done(napi
, work_done
);
1177 if (adapter
->unset_rx_last
) {
1178 adapter
->rx_last_jiffies
= jiffies
;
1179 adapter
->unset_rx_last
= false;
1182 if (((long)jiffies
- (long)adapter
->rx_last_jiffies
) < 3) {
1183 napi_reschedule(napi
);
1185 spin_lock(&hw
->rx_status_lock
);
1186 for (epidx
= 0; epidx
< hw
->max_epid
; epidx
++) {
1187 if (epidx
== hw
->my_epid
)
1189 if (fjes_hw_get_partner_ep_status(hw
, epidx
) ==
1191 adapter
->hw
.ep_shm_info
[epidx
].tx
1192 .info
->v1i
.rx_status
&=
1195 spin_unlock(&hw
->rx_status_lock
);
1197 fjes_hw_set_irqmask(hw
, REG_ICTL_MASK_RX_DATA
, false);
1204 /* fjes_probe - Device Initialization Routine */
1205 static int fjes_probe(struct platform_device
*plat_dev
)
1207 struct fjes_adapter
*adapter
;
1208 struct net_device
*netdev
;
1209 struct resource
*res
;
1214 netdev
= alloc_netdev_mq(sizeof(struct fjes_adapter
), "es%d",
1215 NET_NAME_UNKNOWN
, fjes_netdev_setup
,
1221 SET_NETDEV_DEV(netdev
, &plat_dev
->dev
);
1223 dev_set_drvdata(&plat_dev
->dev
, netdev
);
1224 adapter
= netdev_priv(netdev
);
1225 adapter
->netdev
= netdev
;
1226 adapter
->plat_dev
= plat_dev
;
1230 /* setup the private structure */
1231 err
= fjes_sw_init(adapter
);
1233 goto err_free_netdev
;
1235 INIT_WORK(&adapter
->force_close_task
, fjes_force_close_task
);
1236 adapter
->force_reset
= false;
1237 adapter
->open_guard
= false;
1239 adapter
->txrx_wq
= alloc_workqueue(DRV_NAME
"/txrx", WQ_MEM_RECLAIM
, 0);
1240 adapter
->control_wq
= alloc_workqueue(DRV_NAME
"/control",
1243 INIT_WORK(&adapter
->tx_stall_task
, fjes_tx_stall_task
);
1244 INIT_WORK(&adapter
->raise_intr_rxdata_task
,
1245 fjes_raise_intr_rxdata_task
);
1246 INIT_WORK(&adapter
->unshare_watch_task
, fjes_watch_unshare_task
);
1247 adapter
->unshare_watch_bitmask
= 0;
1249 INIT_DELAYED_WORK(&adapter
->interrupt_watch_task
, fjes_irq_watch_task
);
1250 adapter
->interrupt_watch_enable
= false;
1252 res
= platform_get_resource(plat_dev
, IORESOURCE_MEM
, 0);
1253 hw
->hw_res
.start
= res
->start
;
1254 hw
->hw_res
.size
= resource_size(res
);
1255 hw
->hw_res
.irq
= platform_get_irq(plat_dev
, 0);
1256 err
= fjes_hw_init(&adapter
->hw
);
1258 goto err_free_netdev
;
1260 /* setup MAC address (02:00:00:00:00:[epid])*/
1261 netdev
->dev_addr
[0] = 2;
1262 netdev
->dev_addr
[1] = 0;
1263 netdev
->dev_addr
[2] = 0;
1264 netdev
->dev_addr
[3] = 0;
1265 netdev
->dev_addr
[4] = 0;
1266 netdev
->dev_addr
[5] = hw
->my_epid
; /* EPID */
1268 err
= register_netdev(netdev
);
1272 netif_carrier_off(netdev
);
1274 fjes_dbg_adapter_init(adapter
);
1279 fjes_hw_exit(&adapter
->hw
);
1281 free_netdev(netdev
);
1286 /* fjes_remove - Device Removal Routine */
1287 static int fjes_remove(struct platform_device
*plat_dev
)
1289 struct net_device
*netdev
= dev_get_drvdata(&plat_dev
->dev
);
1290 struct fjes_adapter
*adapter
= netdev_priv(netdev
);
1291 struct fjes_hw
*hw
= &adapter
->hw
;
1293 fjes_dbg_adapter_exit(adapter
);
1295 cancel_delayed_work_sync(&adapter
->interrupt_watch_task
);
1296 cancel_work_sync(&adapter
->unshare_watch_task
);
1297 cancel_work_sync(&adapter
->raise_intr_rxdata_task
);
1298 cancel_work_sync(&adapter
->tx_stall_task
);
1299 if (adapter
->control_wq
)
1300 destroy_workqueue(adapter
->control_wq
);
1301 if (adapter
->txrx_wq
)
1302 destroy_workqueue(adapter
->txrx_wq
);
1304 unregister_netdev(netdev
);
1308 netif_napi_del(&adapter
->napi
);
1310 free_netdev(netdev
);
1315 static int fjes_sw_init(struct fjes_adapter
*adapter
)
1317 struct net_device
*netdev
= adapter
->netdev
;
1319 netif_napi_add(netdev
, &adapter
->napi
, fjes_poll
, 64);
1324 /* fjes_netdev_setup - netdevice initialization routine */
1325 static void fjes_netdev_setup(struct net_device
*netdev
)
1327 ether_setup(netdev
);
1329 netdev
->watchdog_timeo
= FJES_TX_RETRY_INTERVAL
;
1330 netdev
->netdev_ops
= &fjes_netdev_ops
;
1331 fjes_set_ethtool_ops(netdev
);
1332 netdev
->mtu
= fjes_support_mtu
[3];
1333 netdev
->min_mtu
= fjes_support_mtu
[0];
1334 netdev
->max_mtu
= fjes_support_mtu
[3];
1335 netdev
->features
|= NETIF_F_HW_VLAN_CTAG_FILTER
;
1338 static void fjes_irq_watch_task(struct work_struct
*work
)
1340 struct fjes_adapter
*adapter
= container_of(to_delayed_work(work
),
1341 struct fjes_adapter
, interrupt_watch_task
);
1343 local_irq_disable();
1344 fjes_intr(adapter
->hw
.hw_res
.irq
, adapter
);
1347 if (fjes_rxframe_search_exist(adapter
, 0) >= 0)
1348 napi_schedule(&adapter
->napi
);
1350 if (adapter
->interrupt_watch_enable
) {
1351 if (!delayed_work_pending(&adapter
->interrupt_watch_task
))
1352 queue_delayed_work(adapter
->control_wq
,
1353 &adapter
->interrupt_watch_task
,
1354 FJES_IRQ_WATCH_DELAY
);
1358 static void fjes_watch_unshare_task(struct work_struct
*work
)
1360 struct fjes_adapter
*adapter
=
1361 container_of(work
, struct fjes_adapter
, unshare_watch_task
);
1363 struct net_device
*netdev
= adapter
->netdev
;
1364 struct fjes_hw
*hw
= &adapter
->hw
;
1366 int unshare_watch
, unshare_reserve
;
1367 int max_epid
, my_epid
, epidx
;
1368 int stop_req
, stop_req_done
;
1369 ulong unshare_watch_bitmask
;
1370 unsigned long flags
;
1375 my_epid
= hw
->my_epid
;
1376 max_epid
= hw
->max_epid
;
1378 unshare_watch_bitmask
= adapter
->unshare_watch_bitmask
;
1379 adapter
->unshare_watch_bitmask
= 0;
1381 while ((unshare_watch_bitmask
|| hw
->txrx_stop_req_bit
) &&
1382 (wait_time
< 3000)) {
1383 for (epidx
= 0; epidx
< max_epid
; epidx
++) {
1384 if (epidx
== my_epid
)
1387 is_shared
= fjes_hw_epid_is_shared(hw
->hw_info
.share
,
1390 stop_req
= test_bit(epidx
, &hw
->txrx_stop_req_bit
);
1392 stop_req_done
= hw
->ep_shm_info
[epidx
].rx
.info
->v1i
.rx_status
&
1393 FJES_RX_STOP_REQ_DONE
;
1395 unshare_watch
= test_bit(epidx
, &unshare_watch_bitmask
);
1397 unshare_reserve
= test_bit(epidx
,
1398 &hw
->hw_info
.buffer_unshare_reserve_bit
);
1401 (is_shared
&& (!is_shared
|| !stop_req_done
))) &&
1402 (is_shared
|| !unshare_watch
|| !unshare_reserve
))
1405 mutex_lock(&hw
->hw_info
.lock
);
1406 ret
= fjes_hw_unregister_buff_addr(hw
, epidx
);
1414 &adapter
->force_close_task
)) {
1415 adapter
->force_reset
= true;
1417 &adapter
->force_close_task
);
1421 mutex_unlock(&hw
->hw_info
.lock
);
1422 hw
->ep_shm_info
[epidx
].ep_stats
1423 .com_unregist_buf_exec
+= 1;
1425 spin_lock_irqsave(&hw
->rx_status_lock
, flags
);
1426 fjes_hw_setup_epbuf(&hw
->ep_shm_info
[epidx
].tx
,
1427 netdev
->dev_addr
, netdev
->mtu
);
1428 spin_unlock_irqrestore(&hw
->rx_status_lock
, flags
);
1430 clear_bit(epidx
, &hw
->txrx_stop_req_bit
);
1431 clear_bit(epidx
, &unshare_watch_bitmask
);
1433 &hw
->hw_info
.buffer_unshare_reserve_bit
);
1440 if (hw
->hw_info
.buffer_unshare_reserve_bit
) {
1441 for (epidx
= 0; epidx
< max_epid
; epidx
++) {
1442 if (epidx
== my_epid
)
1446 &hw
->hw_info
.buffer_unshare_reserve_bit
)) {
1447 mutex_lock(&hw
->hw_info
.lock
);
1449 ret
= fjes_hw_unregister_buff_addr(hw
, epidx
);
1457 &adapter
->force_close_task
)) {
1458 adapter
->force_reset
= true;
1460 &adapter
->force_close_task
);
1464 mutex_unlock(&hw
->hw_info
.lock
);
1466 hw
->ep_shm_info
[epidx
].ep_stats
1467 .com_unregist_buf_exec
+= 1;
1469 spin_lock_irqsave(&hw
->rx_status_lock
, flags
);
1470 fjes_hw_setup_epbuf(
1471 &hw
->ep_shm_info
[epidx
].tx
,
1472 netdev
->dev_addr
, netdev
->mtu
);
1473 spin_unlock_irqrestore(&hw
->rx_status_lock
,
1476 clear_bit(epidx
, &hw
->txrx_stop_req_bit
);
1477 clear_bit(epidx
, &unshare_watch_bitmask
);
1478 clear_bit(epidx
, &hw
->hw_info
.buffer_unshare_reserve_bit
);
1481 if (test_bit(epidx
, &unshare_watch_bitmask
)) {
1482 spin_lock_irqsave(&hw
->rx_status_lock
, flags
);
1483 hw
->ep_shm_info
[epidx
].tx
.info
->v1i
.rx_status
&=
1484 ~FJES_RX_STOP_REQ_DONE
;
1485 spin_unlock_irqrestore(&hw
->rx_status_lock
,
1493 acpi_find_extended_socket_device(acpi_handle obj_handle
, u32 level
,
1494 void *context
, void **return_value
)
1496 struct acpi_device
*device
;
1497 bool *found
= context
;
1500 result
= acpi_bus_get_device(obj_handle
, &device
);
1504 if (strcmp(acpi_device_hid(device
), ACPI_MOTHERBOARD_RESOURCE_HID
))
1507 if (!is_extended_socket_device(device
))
1510 if (acpi_check_extended_socket_status(device
))
1514 return AE_CTRL_TERMINATE
;
1517 /* fjes_init_module - Driver Registration Routine */
1518 static int __init
fjes_init_module(void)
1523 acpi_walk_namespace(ACPI_TYPE_DEVICE
, ACPI_ROOT_OBJECT
, ACPI_UINT32_MAX
,
1524 acpi_find_extended_socket_device
, NULL
, &found
,
1530 pr_info("%s - version %s - %s\n",
1531 fjes_driver_string
, fjes_driver_version
, fjes_copyright
);
1535 result
= platform_driver_register(&fjes_driver
);
1541 result
= acpi_bus_register_driver(&fjes_acpi_driver
);
1543 goto fail_acpi_driver
;
1548 platform_driver_unregister(&fjes_driver
);
1553 module_init(fjes_init_module
);
1555 /* fjes_exit_module - Driver Exit Cleanup Routine */
1556 static void __exit
fjes_exit_module(void)
1558 acpi_bus_unregister_driver(&fjes_acpi_driver
);
1559 platform_driver_unregister(&fjes_driver
);
1563 module_exit(fjes_exit_module
);