perf annotate: Pass 'struct annotation_options' to map_symbol__annotation_dump()
[linux/fpc-iii.git] / drivers / net / wan / wanxl.c
blobd573a57bc3018f4b737d4da0480f2ae2ba6fa8ae
1 /*
2 * wanXL serial card driver for Linux
3 * host part
5 * Copyright (C) 2003 Krzysztof Halasa <khc@pm.waw.pl>
7 * This program is free software; you can redistribute it and/or modify it
8 * under the terms of version 2 of the GNU General Public License
9 * as published by the Free Software Foundation.
11 * Status:
12 * - Only DTE (external clock) support with NRZ and NRZI encodings
13 * - wanXL100 will require minor driver modifications, no access to hw
16 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
18 #include <linux/module.h>
19 #include <linux/kernel.h>
20 #include <linux/slab.h>
21 #include <linux/sched.h>
22 #include <linux/types.h>
23 #include <linux/fcntl.h>
24 #include <linux/string.h>
25 #include <linux/errno.h>
26 #include <linux/init.h>
27 #include <linux/interrupt.h>
28 #include <linux/ioport.h>
29 #include <linux/netdevice.h>
30 #include <linux/hdlc.h>
31 #include <linux/pci.h>
32 #include <linux/dma-mapping.h>
33 #include <linux/delay.h>
34 #include <asm/io.h>
36 #include "wanxl.h"
38 static const char* version = "wanXL serial card driver version: 0.48";
40 #define PLX_CTL_RESET 0x40000000 /* adapter reset */
42 #undef DEBUG_PKT
43 #undef DEBUG_PCI
45 /* MAILBOX #1 - PUTS COMMANDS */
46 #define MBX1_CMD_ABORTJ 0x85000000 /* Abort and Jump */
47 #ifdef __LITTLE_ENDIAN
48 #define MBX1_CMD_BSWAP 0x8C000001 /* little-endian Byte Swap Mode */
49 #else
50 #define MBX1_CMD_BSWAP 0x8C000000 /* big-endian Byte Swap Mode */
51 #endif
53 /* MAILBOX #2 - DRAM SIZE */
54 #define MBX2_MEMSZ_MASK 0xFFFF0000 /* PUTS Memory Size Register mask */
57 struct port {
58 struct net_device *dev;
59 struct card *card;
60 spinlock_t lock; /* for wanxl_xmit */
61 int node; /* physical port #0 - 3 */
62 unsigned int clock_type;
63 int tx_in, tx_out;
64 struct sk_buff *tx_skbs[TX_BUFFERS];
68 struct card_status {
69 desc_t rx_descs[RX_QUEUE_LENGTH];
70 port_status_t port_status[4];
74 struct card {
75 int n_ports; /* 1, 2 or 4 ports */
76 u8 irq;
78 u8 __iomem *plx; /* PLX PCI9060 virtual base address */
79 struct pci_dev *pdev; /* for pci_name(pdev) */
80 int rx_in;
81 struct sk_buff *rx_skbs[RX_QUEUE_LENGTH];
82 struct card_status *status; /* shared between host and card */
83 dma_addr_t status_address;
84 struct port ports[0]; /* 1 - 4 port structures follow */
89 static inline struct port *dev_to_port(struct net_device *dev)
91 return (struct port *)dev_to_hdlc(dev)->priv;
95 static inline port_status_t *get_status(struct port *port)
97 return &port->card->status->port_status[port->node];
101 #ifdef DEBUG_PCI
102 static inline dma_addr_t pci_map_single_debug(struct pci_dev *pdev, void *ptr,
103 size_t size, int direction)
105 dma_addr_t addr = pci_map_single(pdev, ptr, size, direction);
106 if (addr + size > 0x100000000LL)
107 pr_crit("%s: pci_map_single() returned memory at 0x%llx!\n",
108 pci_name(pdev), (unsigned long long)addr);
109 return addr;
112 #undef pci_map_single
113 #define pci_map_single pci_map_single_debug
114 #endif
117 /* Cable and/or personality module change interrupt service */
118 static inline void wanxl_cable_intr(struct port *port)
120 u32 value = get_status(port)->cable;
121 int valid = 1;
122 const char *cable, *pm, *dte = "", *dsr = "", *dcd = "";
124 switch(value & 0x7) {
125 case STATUS_CABLE_V35: cable = "V.35"; break;
126 case STATUS_CABLE_X21: cable = "X.21"; break;
127 case STATUS_CABLE_V24: cable = "V.24"; break;
128 case STATUS_CABLE_EIA530: cable = "EIA530"; break;
129 case STATUS_CABLE_NONE: cable = "no"; break;
130 default: cable = "invalid";
133 switch((value >> STATUS_CABLE_PM_SHIFT) & 0x7) {
134 case STATUS_CABLE_V35: pm = "V.35"; break;
135 case STATUS_CABLE_X21: pm = "X.21"; break;
136 case STATUS_CABLE_V24: pm = "V.24"; break;
137 case STATUS_CABLE_EIA530: pm = "EIA530"; break;
138 case STATUS_CABLE_NONE: pm = "no personality"; valid = 0; break;
139 default: pm = "invalid personality"; valid = 0;
142 if (valid) {
143 if ((value & 7) == ((value >> STATUS_CABLE_PM_SHIFT) & 7)) {
144 dsr = (value & STATUS_CABLE_DSR) ? ", DSR ON" :
145 ", DSR off";
146 dcd = (value & STATUS_CABLE_DCD) ? ", carrier ON" :
147 ", carrier off";
149 dte = (value & STATUS_CABLE_DCE) ? " DCE" : " DTE";
151 netdev_info(port->dev, "%s%s module, %s cable%s%s\n",
152 pm, dte, cable, dsr, dcd);
154 if (value & STATUS_CABLE_DCD)
155 netif_carrier_on(port->dev);
156 else
157 netif_carrier_off(port->dev);
162 /* Transmit complete interrupt service */
163 static inline void wanxl_tx_intr(struct port *port)
165 struct net_device *dev = port->dev;
166 while (1) {
167 desc_t *desc = &get_status(port)->tx_descs[port->tx_in];
168 struct sk_buff *skb = port->tx_skbs[port->tx_in];
170 switch (desc->stat) {
171 case PACKET_FULL:
172 case PACKET_EMPTY:
173 netif_wake_queue(dev);
174 return;
176 case PACKET_UNDERRUN:
177 dev->stats.tx_errors++;
178 dev->stats.tx_fifo_errors++;
179 break;
181 default:
182 dev->stats.tx_packets++;
183 dev->stats.tx_bytes += skb->len;
185 desc->stat = PACKET_EMPTY; /* Free descriptor */
186 pci_unmap_single(port->card->pdev, desc->address, skb->len,
187 PCI_DMA_TODEVICE);
188 dev_kfree_skb_irq(skb);
189 port->tx_in = (port->tx_in + 1) % TX_BUFFERS;
195 /* Receive complete interrupt service */
196 static inline void wanxl_rx_intr(struct card *card)
198 desc_t *desc;
199 while (desc = &card->status->rx_descs[card->rx_in],
200 desc->stat != PACKET_EMPTY) {
201 if ((desc->stat & PACKET_PORT_MASK) > card->n_ports)
202 pr_crit("%s: received packet for nonexistent port\n",
203 pci_name(card->pdev));
204 else {
205 struct sk_buff *skb = card->rx_skbs[card->rx_in];
206 struct port *port = &card->ports[desc->stat &
207 PACKET_PORT_MASK];
208 struct net_device *dev = port->dev;
210 if (!skb)
211 dev->stats.rx_dropped++;
212 else {
213 pci_unmap_single(card->pdev, desc->address,
214 BUFFER_LENGTH,
215 PCI_DMA_FROMDEVICE);
216 skb_put(skb, desc->length);
218 #ifdef DEBUG_PKT
219 printk(KERN_DEBUG "%s RX(%i):", dev->name,
220 skb->len);
221 debug_frame(skb);
222 #endif
223 dev->stats.rx_packets++;
224 dev->stats.rx_bytes += skb->len;
225 skb->protocol = hdlc_type_trans(skb, dev);
226 netif_rx(skb);
227 skb = NULL;
230 if (!skb) {
231 skb = dev_alloc_skb(BUFFER_LENGTH);
232 desc->address = skb ?
233 pci_map_single(card->pdev, skb->data,
234 BUFFER_LENGTH,
235 PCI_DMA_FROMDEVICE) : 0;
236 card->rx_skbs[card->rx_in] = skb;
239 desc->stat = PACKET_EMPTY; /* Free descriptor */
240 card->rx_in = (card->rx_in + 1) % RX_QUEUE_LENGTH;
246 static irqreturn_t wanxl_intr(int irq, void* dev_id)
248 struct card *card = dev_id;
249 int i;
250 u32 stat;
251 int handled = 0;
254 while((stat = readl(card->plx + PLX_DOORBELL_FROM_CARD)) != 0) {
255 handled = 1;
256 writel(stat, card->plx + PLX_DOORBELL_FROM_CARD);
258 for (i = 0; i < card->n_ports; i++) {
259 if (stat & (1 << (DOORBELL_FROM_CARD_TX_0 + i)))
260 wanxl_tx_intr(&card->ports[i]);
261 if (stat & (1 << (DOORBELL_FROM_CARD_CABLE_0 + i)))
262 wanxl_cable_intr(&card->ports[i]);
264 if (stat & (1 << DOORBELL_FROM_CARD_RX))
265 wanxl_rx_intr(card);
268 return IRQ_RETVAL(handled);
273 static netdev_tx_t wanxl_xmit(struct sk_buff *skb, struct net_device *dev)
275 struct port *port = dev_to_port(dev);
276 desc_t *desc;
278 spin_lock(&port->lock);
280 desc = &get_status(port)->tx_descs[port->tx_out];
281 if (desc->stat != PACKET_EMPTY) {
282 /* should never happen - previous xmit should stop queue */
283 #ifdef DEBUG_PKT
284 printk(KERN_DEBUG "%s: transmitter buffer full\n", dev->name);
285 #endif
286 netif_stop_queue(dev);
287 spin_unlock(&port->lock);
288 return NETDEV_TX_BUSY; /* request packet to be queued */
291 #ifdef DEBUG_PKT
292 printk(KERN_DEBUG "%s TX(%i):", dev->name, skb->len);
293 debug_frame(skb);
294 #endif
296 port->tx_skbs[port->tx_out] = skb;
297 desc->address = pci_map_single(port->card->pdev, skb->data, skb->len,
298 PCI_DMA_TODEVICE);
299 desc->length = skb->len;
300 desc->stat = PACKET_FULL;
301 writel(1 << (DOORBELL_TO_CARD_TX_0 + port->node),
302 port->card->plx + PLX_DOORBELL_TO_CARD);
304 port->tx_out = (port->tx_out + 1) % TX_BUFFERS;
306 if (get_status(port)->tx_descs[port->tx_out].stat != PACKET_EMPTY) {
307 netif_stop_queue(dev);
308 #ifdef DEBUG_PKT
309 printk(KERN_DEBUG "%s: transmitter buffer full\n", dev->name);
310 #endif
313 spin_unlock(&port->lock);
314 return NETDEV_TX_OK;
319 static int wanxl_attach(struct net_device *dev, unsigned short encoding,
320 unsigned short parity)
322 struct port *port = dev_to_port(dev);
324 if (encoding != ENCODING_NRZ &&
325 encoding != ENCODING_NRZI)
326 return -EINVAL;
328 if (parity != PARITY_NONE &&
329 parity != PARITY_CRC32_PR1_CCITT &&
330 parity != PARITY_CRC16_PR1_CCITT &&
331 parity != PARITY_CRC32_PR0_CCITT &&
332 parity != PARITY_CRC16_PR0_CCITT)
333 return -EINVAL;
335 get_status(port)->encoding = encoding;
336 get_status(port)->parity = parity;
337 return 0;
342 static int wanxl_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd)
344 const size_t size = sizeof(sync_serial_settings);
345 sync_serial_settings line;
346 struct port *port = dev_to_port(dev);
348 if (cmd != SIOCWANDEV)
349 return hdlc_ioctl(dev, ifr, cmd);
351 switch (ifr->ifr_settings.type) {
352 case IF_GET_IFACE:
353 ifr->ifr_settings.type = IF_IFACE_SYNC_SERIAL;
354 if (ifr->ifr_settings.size < size) {
355 ifr->ifr_settings.size = size; /* data size wanted */
356 return -ENOBUFS;
358 memset(&line, 0, sizeof(line));
359 line.clock_type = get_status(port)->clocking;
360 line.clock_rate = 0;
361 line.loopback = 0;
363 if (copy_to_user(ifr->ifr_settings.ifs_ifsu.sync, &line, size))
364 return -EFAULT;
365 return 0;
367 case IF_IFACE_SYNC_SERIAL:
368 if (!capable(CAP_NET_ADMIN))
369 return -EPERM;
370 if (dev->flags & IFF_UP)
371 return -EBUSY;
373 if (copy_from_user(&line, ifr->ifr_settings.ifs_ifsu.sync,
374 size))
375 return -EFAULT;
377 if (line.clock_type != CLOCK_EXT &&
378 line.clock_type != CLOCK_TXFROMRX)
379 return -EINVAL; /* No such clock setting */
381 if (line.loopback != 0)
382 return -EINVAL;
384 get_status(port)->clocking = line.clock_type;
385 return 0;
387 default:
388 return hdlc_ioctl(dev, ifr, cmd);
394 static int wanxl_open(struct net_device *dev)
396 struct port *port = dev_to_port(dev);
397 u8 __iomem *dbr = port->card->plx + PLX_DOORBELL_TO_CARD;
398 unsigned long timeout;
399 int i;
401 if (get_status(port)->open) {
402 netdev_err(dev, "port already open\n");
403 return -EIO;
405 if ((i = hdlc_open(dev)) != 0)
406 return i;
408 port->tx_in = port->tx_out = 0;
409 for (i = 0; i < TX_BUFFERS; i++)
410 get_status(port)->tx_descs[i].stat = PACKET_EMPTY;
411 /* signal the card */
412 writel(1 << (DOORBELL_TO_CARD_OPEN_0 + port->node), dbr);
414 timeout = jiffies + HZ;
415 do {
416 if (get_status(port)->open) {
417 netif_start_queue(dev);
418 return 0;
420 } while (time_after(timeout, jiffies));
422 netdev_err(dev, "unable to open port\n");
423 /* ask the card to close the port, should it be still alive */
424 writel(1 << (DOORBELL_TO_CARD_CLOSE_0 + port->node), dbr);
425 return -EFAULT;
430 static int wanxl_close(struct net_device *dev)
432 struct port *port = dev_to_port(dev);
433 unsigned long timeout;
434 int i;
436 hdlc_close(dev);
437 /* signal the card */
438 writel(1 << (DOORBELL_TO_CARD_CLOSE_0 + port->node),
439 port->card->plx + PLX_DOORBELL_TO_CARD);
441 timeout = jiffies + HZ;
442 do {
443 if (!get_status(port)->open)
444 break;
445 } while (time_after(timeout, jiffies));
447 if (get_status(port)->open)
448 netdev_err(dev, "unable to close port\n");
450 netif_stop_queue(dev);
452 for (i = 0; i < TX_BUFFERS; i++) {
453 desc_t *desc = &get_status(port)->tx_descs[i];
455 if (desc->stat != PACKET_EMPTY) {
456 desc->stat = PACKET_EMPTY;
457 pci_unmap_single(port->card->pdev, desc->address,
458 port->tx_skbs[i]->len,
459 PCI_DMA_TODEVICE);
460 dev_kfree_skb(port->tx_skbs[i]);
463 return 0;
468 static struct net_device_stats *wanxl_get_stats(struct net_device *dev)
470 struct port *port = dev_to_port(dev);
472 dev->stats.rx_over_errors = get_status(port)->rx_overruns;
473 dev->stats.rx_frame_errors = get_status(port)->rx_frame_errors;
474 dev->stats.rx_errors = dev->stats.rx_over_errors +
475 dev->stats.rx_frame_errors;
476 return &dev->stats;
481 static int wanxl_puts_command(struct card *card, u32 cmd)
483 unsigned long timeout = jiffies + 5 * HZ;
485 writel(cmd, card->plx + PLX_MAILBOX_1);
486 do {
487 if (readl(card->plx + PLX_MAILBOX_1) == 0)
488 return 0;
490 schedule();
491 }while (time_after(timeout, jiffies));
493 return -1;
498 static void wanxl_reset(struct card *card)
500 u32 old_value = readl(card->plx + PLX_CONTROL) & ~PLX_CTL_RESET;
502 writel(0x80, card->plx + PLX_MAILBOX_0);
503 writel(old_value | PLX_CTL_RESET, card->plx + PLX_CONTROL);
504 readl(card->plx + PLX_CONTROL); /* wait for posted write */
505 udelay(1);
506 writel(old_value, card->plx + PLX_CONTROL);
507 readl(card->plx + PLX_CONTROL); /* wait for posted write */
512 static void wanxl_pci_remove_one(struct pci_dev *pdev)
514 struct card *card = pci_get_drvdata(pdev);
515 int i;
517 for (i = 0; i < card->n_ports; i++) {
518 unregister_hdlc_device(card->ports[i].dev);
519 free_netdev(card->ports[i].dev);
522 /* unregister and free all host resources */
523 if (card->irq)
524 free_irq(card->irq, card);
526 wanxl_reset(card);
528 for (i = 0; i < RX_QUEUE_LENGTH; i++)
529 if (card->rx_skbs[i]) {
530 pci_unmap_single(card->pdev,
531 card->status->rx_descs[i].address,
532 BUFFER_LENGTH, PCI_DMA_FROMDEVICE);
533 dev_kfree_skb(card->rx_skbs[i]);
536 if (card->plx)
537 iounmap(card->plx);
539 if (card->status)
540 pci_free_consistent(pdev, sizeof(struct card_status),
541 card->status, card->status_address);
543 pci_release_regions(pdev);
544 pci_disable_device(pdev);
545 kfree(card);
549 #include "wanxlfw.inc"
551 static const struct net_device_ops wanxl_ops = {
552 .ndo_open = wanxl_open,
553 .ndo_stop = wanxl_close,
554 .ndo_start_xmit = hdlc_start_xmit,
555 .ndo_do_ioctl = wanxl_ioctl,
556 .ndo_get_stats = wanxl_get_stats,
559 static int wanxl_pci_init_one(struct pci_dev *pdev,
560 const struct pci_device_id *ent)
562 struct card *card;
563 u32 ramsize, stat;
564 unsigned long timeout;
565 u32 plx_phy; /* PLX PCI base address */
566 u32 mem_phy; /* memory PCI base addr */
567 u8 __iomem *mem; /* memory virtual base addr */
568 int i, ports, alloc_size;
570 #ifndef MODULE
571 pr_info_once("%s\n", version);
572 #endif
574 i = pci_enable_device(pdev);
575 if (i)
576 return i;
578 /* QUICC can only access first 256 MB of host RAM directly,
579 but PLX9060 DMA does 32-bits for actual packet data transfers */
581 /* FIXME when PCI/DMA subsystems are fixed.
582 We set both dma_mask and consistent_dma_mask to 28 bits
583 and pray pci_alloc_consistent() will use this info. It should
584 work on most platforms */
585 if (pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(28)) ||
586 pci_set_dma_mask(pdev, DMA_BIT_MASK(28))) {
587 pr_err("No usable DMA configuration\n");
588 pci_disable_device(pdev);
589 return -EIO;
592 i = pci_request_regions(pdev, "wanXL");
593 if (i) {
594 pci_disable_device(pdev);
595 return i;
598 switch (pdev->device) {
599 case PCI_DEVICE_ID_SBE_WANXL100: ports = 1; break;
600 case PCI_DEVICE_ID_SBE_WANXL200: ports = 2; break;
601 default: ports = 4;
604 alloc_size = sizeof(struct card) + ports * sizeof(struct port);
605 card = kzalloc(alloc_size, GFP_KERNEL);
606 if (card == NULL) {
607 pci_release_regions(pdev);
608 pci_disable_device(pdev);
609 return -ENOBUFS;
612 pci_set_drvdata(pdev, card);
613 card->pdev = pdev;
615 card->status = pci_alloc_consistent(pdev,
616 sizeof(struct card_status),
617 &card->status_address);
618 if (card->status == NULL) {
619 wanxl_pci_remove_one(pdev);
620 return -ENOBUFS;
623 #ifdef DEBUG_PCI
624 printk(KERN_DEBUG "wanXL %s: pci_alloc_consistent() returned memory"
625 " at 0x%LX\n", pci_name(pdev),
626 (unsigned long long)card->status_address);
627 #endif
629 /* FIXME when PCI/DMA subsystems are fixed.
630 We set both dma_mask and consistent_dma_mask back to 32 bits
631 to indicate the card can do 32-bit DMA addressing */
632 if (pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(32)) ||
633 pci_set_dma_mask(pdev, DMA_BIT_MASK(32))) {
634 pr_err("No usable DMA configuration\n");
635 wanxl_pci_remove_one(pdev);
636 return -EIO;
639 /* set up PLX mapping */
640 plx_phy = pci_resource_start(pdev, 0);
642 card->plx = ioremap_nocache(plx_phy, 0x70);
643 if (!card->plx) {
644 pr_err("ioremap() failed\n");
645 wanxl_pci_remove_one(pdev);
646 return -EFAULT;
649 #if RESET_WHILE_LOADING
650 wanxl_reset(card);
651 #endif
653 timeout = jiffies + 20 * HZ;
654 while ((stat = readl(card->plx + PLX_MAILBOX_0)) != 0) {
655 if (time_before(timeout, jiffies)) {
656 pr_warn("%s: timeout waiting for PUTS to complete\n",
657 pci_name(pdev));
658 wanxl_pci_remove_one(pdev);
659 return -ENODEV;
662 switch(stat & 0xC0) {
663 case 0x00: /* hmm - PUTS completed with non-zero code? */
664 case 0x80: /* PUTS still testing the hardware */
665 break;
667 default:
668 pr_warn("%s: PUTS test 0x%X failed\n",
669 pci_name(pdev), stat & 0x30);
670 wanxl_pci_remove_one(pdev);
671 return -ENODEV;
674 schedule();
677 /* get on-board memory size (PUTS detects no more than 4 MB) */
678 ramsize = readl(card->plx + PLX_MAILBOX_2) & MBX2_MEMSZ_MASK;
680 /* set up on-board RAM mapping */
681 mem_phy = pci_resource_start(pdev, 2);
684 /* sanity check the board's reported memory size */
685 if (ramsize < BUFFERS_ADDR +
686 (TX_BUFFERS + RX_BUFFERS) * BUFFER_LENGTH * ports) {
687 pr_warn("%s: no enough on-board RAM (%u bytes detected, %u bytes required)\n",
688 pci_name(pdev), ramsize,
689 BUFFERS_ADDR +
690 (TX_BUFFERS + RX_BUFFERS) * BUFFER_LENGTH * ports);
691 wanxl_pci_remove_one(pdev);
692 return -ENODEV;
695 if (wanxl_puts_command(card, MBX1_CMD_BSWAP)) {
696 pr_warn("%s: unable to Set Byte Swap Mode\n", pci_name(pdev));
697 wanxl_pci_remove_one(pdev);
698 return -ENODEV;
701 for (i = 0; i < RX_QUEUE_LENGTH; i++) {
702 struct sk_buff *skb = dev_alloc_skb(BUFFER_LENGTH);
703 card->rx_skbs[i] = skb;
704 if (skb)
705 card->status->rx_descs[i].address =
706 pci_map_single(card->pdev, skb->data,
707 BUFFER_LENGTH,
708 PCI_DMA_FROMDEVICE);
711 mem = ioremap_nocache(mem_phy, PDM_OFFSET + sizeof(firmware));
712 if (!mem) {
713 pr_err("ioremap() failed\n");
714 wanxl_pci_remove_one(pdev);
715 return -EFAULT;
718 for (i = 0; i < sizeof(firmware); i += 4)
719 writel(ntohl(*(__be32*)(firmware + i)), mem + PDM_OFFSET + i);
721 for (i = 0; i < ports; i++)
722 writel(card->status_address +
723 (void *)&card->status->port_status[i] -
724 (void *)card->status, mem + PDM_OFFSET + 4 + i * 4);
725 writel(card->status_address, mem + PDM_OFFSET + 20);
726 writel(PDM_OFFSET, mem);
727 iounmap(mem);
729 writel(0, card->plx + PLX_MAILBOX_5);
731 if (wanxl_puts_command(card, MBX1_CMD_ABORTJ)) {
732 pr_warn("%s: unable to Abort and Jump\n", pci_name(pdev));
733 wanxl_pci_remove_one(pdev);
734 return -ENODEV;
737 timeout = jiffies + 5 * HZ;
738 do {
739 if ((stat = readl(card->plx + PLX_MAILBOX_5)) != 0)
740 break;
741 schedule();
742 }while (time_after(timeout, jiffies));
744 if (!stat) {
745 pr_warn("%s: timeout while initializing card firmware\n",
746 pci_name(pdev));
747 wanxl_pci_remove_one(pdev);
748 return -ENODEV;
751 #if DETECT_RAM
752 ramsize = stat;
753 #endif
755 pr_info("%s: at 0x%X, %u KB of RAM at 0x%X, irq %u\n",
756 pci_name(pdev), plx_phy, ramsize / 1024, mem_phy, pdev->irq);
758 /* Allocate IRQ */
759 if (request_irq(pdev->irq, wanxl_intr, IRQF_SHARED, "wanXL", card)) {
760 pr_warn("%s: could not allocate IRQ%i\n",
761 pci_name(pdev), pdev->irq);
762 wanxl_pci_remove_one(pdev);
763 return -EBUSY;
765 card->irq = pdev->irq;
767 for (i = 0; i < ports; i++) {
768 hdlc_device *hdlc;
769 struct port *port = &card->ports[i];
770 struct net_device *dev = alloc_hdlcdev(port);
771 if (!dev) {
772 pr_err("%s: unable to allocate memory\n",
773 pci_name(pdev));
774 wanxl_pci_remove_one(pdev);
775 return -ENOMEM;
778 port->dev = dev;
779 hdlc = dev_to_hdlc(dev);
780 spin_lock_init(&port->lock);
781 dev->tx_queue_len = 50;
782 dev->netdev_ops = &wanxl_ops;
783 hdlc->attach = wanxl_attach;
784 hdlc->xmit = wanxl_xmit;
785 port->card = card;
786 port->node = i;
787 get_status(port)->clocking = CLOCK_EXT;
788 if (register_hdlc_device(dev)) {
789 pr_err("%s: unable to register hdlc device\n",
790 pci_name(pdev));
791 free_netdev(dev);
792 wanxl_pci_remove_one(pdev);
793 return -ENOBUFS;
795 card->n_ports++;
798 pr_info("%s: port", pci_name(pdev));
799 for (i = 0; i < ports; i++)
800 pr_cont("%s #%i: %s",
801 i ? "," : "", i, card->ports[i].dev->name);
802 pr_cont("\n");
804 for (i = 0; i < ports; i++)
805 wanxl_cable_intr(&card->ports[i]); /* get carrier status etc.*/
807 return 0;
810 static const struct pci_device_id wanxl_pci_tbl[] = {
811 { PCI_VENDOR_ID_SBE, PCI_DEVICE_ID_SBE_WANXL100, PCI_ANY_ID,
812 PCI_ANY_ID, 0, 0, 0 },
813 { PCI_VENDOR_ID_SBE, PCI_DEVICE_ID_SBE_WANXL200, PCI_ANY_ID,
814 PCI_ANY_ID, 0, 0, 0 },
815 { PCI_VENDOR_ID_SBE, PCI_DEVICE_ID_SBE_WANXL400, PCI_ANY_ID,
816 PCI_ANY_ID, 0, 0, 0 },
817 { 0, }
821 static struct pci_driver wanxl_pci_driver = {
822 .name = "wanXL",
823 .id_table = wanxl_pci_tbl,
824 .probe = wanxl_pci_init_one,
825 .remove = wanxl_pci_remove_one,
829 static int __init wanxl_init_module(void)
831 #ifdef MODULE
832 pr_info("%s\n", version);
833 #endif
834 return pci_register_driver(&wanxl_pci_driver);
837 static void __exit wanxl_cleanup_module(void)
839 pci_unregister_driver(&wanxl_pci_driver);
843 MODULE_AUTHOR("Krzysztof Halasa <khc@pm.waw.pl>");
844 MODULE_DESCRIPTION("SBE Inc. wanXL serial port driver");
845 MODULE_LICENSE("GPL v2");
846 MODULE_DEVICE_TABLE(pci, wanxl_pci_tbl);
848 module_init(wanxl_init_module);
849 module_exit(wanxl_cleanup_module);