1 /* $Id: plip.c,v 1.3.6.2 1997/04/16 15:07:56 phil Exp $ */
2 /* PLIP: A parallel port "network" driver for Linux. */
3 /* This driver is for parallel port with 5-bit cable (LapLink (R) cable). */
5 * Authors: Donald Becker <becker@scyld.com>
6 * Tommy Thorn <thorn@daimi.aau.dk>
7 * Tanabe Hiroyasu <hiro@sanpo.t.u-tokyo.ac.jp>
8 * Alan Cox <gw4pts@gw4pts.ampr.org>
9 * Peter Bauer <100136.3530@compuserve.com>
10 * Niibe Yutaka <gniibe@mri.co.jp>
11 * Nimrod Zimerman <zimerman@mailandnews.com>
14 * Modularization and ifreq/ifmap support by Alan Cox.
15 * Rewritten by Niibe Yutaka.
16 * parport-sharing awareness code by Philip Blundell.
17 * SMP locking by Niibe Yutaka.
18 * Support for parallel ports with no IRQ (poll mode),
19 * Modifications to use the parallel port API
24 * - Module initialization.
26 * - Make sure other end is OK, before sending a packet.
27 * - Fix immediate timer problem.
30 * - Changed {enable,disable}_irq handling to make it work
31 * with new ("stack") semantics.
33 * This program is free software; you can redistribute it and/or
34 * modify it under the terms of the GNU General Public License
35 * as published by the Free Software Foundation; either version
36 * 2 of the License, or (at your option) any later version.
40 * Original version and the name 'PLIP' from Donald Becker <becker@scyld.com>
41 * inspired by Russ Nelson's parallel port packet driver.
44 * Tanabe Hiroyasu had changed the protocol, and it was in Linux v1.0.
45 * Because of the necessity to communicate to DOS machines with the
46 * Crynwr packet driver, Peter Bauer changed the protocol again
47 * back to original protocol.
49 * This version follows original PLIP protocol.
50 * So, this PLIP can't communicate the PLIP of Linux v1.0.
54 * To use with DOS box, please do (Turn on ARP switch):
55 * # ifconfig plip[0-2] arp
57 static const char version
[] = "NET3 PLIP version 2.4-parport gniibe@mri.co.jp\n";
61 Ideas and protocols came from Russ Nelson's <nelson@crynwr.com>
62 "parallel.asm" parallel port packet driver.
64 The "Crynwr" parallel port standard specifies the following protocol:
65 Trigger by sending nibble '0x8' (this causes interrupt on other end)
70 Each octet is sent as <wait for rx. '0x1?'> <send 0x10+(octet&0x0F)>
71 <wait for rx. '0x0?'> <send 0x00+((octet>>4)&0x0F)>
73 The packet is encapsulated as if it were ethernet.
75 The cable used is a de facto standard parallel null cable -- sold as
76 a "LapLink" cable by various places. You'll need a 12-conductor cable to
77 make one yourself. The wiring is:
80 D0->ERROR 2 - 15 15 - 2
81 D1->SLCT 3 - 13 13 - 3
82 D2->PAPOUT 4 - 12 12 - 4
84 D4->BUSY 6 - 11 11 - 6
85 Do not connect the other pins. They are
87 STROBE is 1, FEED is 14, INIT is 16
88 extra grounds are 18,19,20,21,22,23,24
91 #include <linux/module.h>
92 #include <linux/kernel.h>
93 #include <linux/types.h>
94 #include <linux/fcntl.h>
95 #include <linux/interrupt.h>
96 #include <linux/string.h>
97 #include <linux/if_ether.h>
99 #include <linux/errno.h>
100 #include <linux/delay.h>
101 #include <linux/lp.h>
102 #include <linux/init.h>
103 #include <linux/netdevice.h>
104 #include <linux/etherdevice.h>
105 #include <linux/inetdevice.h>
106 #include <linux/skbuff.h>
107 #include <linux/if_plip.h>
108 #include <linux/workqueue.h>
109 #include <linux/ioport.h>
110 #include <linux/spinlock.h>
111 #include <linux/parport.h>
112 #include <linux/bitops.h>
114 #include <net/neighbour.h>
116 #include <asm/system.h>
118 #include <asm/byteorder.h>
119 #include <asm/semaphore.h>
121 /* Maximum number of devices to support. */
124 /* Use 0 for production, 1 for verification, >2 for debug */
128 static unsigned int net_debug
= NET_DEBUG
;
130 #define ENABLE(irq) if (irq != -1) enable_irq(irq)
131 #define DISABLE(irq) if (irq != -1) disable_irq(irq)
133 /* In micro second */
134 #define PLIP_DELAY_UNIT 1
136 /* Connection time out = PLIP_TRIGGER_WAIT * PLIP_DELAY_UNIT usec */
137 #define PLIP_TRIGGER_WAIT 500
139 /* Nibble time out = PLIP_NIBBLE_WAIT * PLIP_DELAY_UNIT usec */
140 #define PLIP_NIBBLE_WAIT 3000
143 static void plip_kick_bh(struct net_device
*dev
);
144 static void plip_bh(struct net_device
*dev
);
145 static void plip_timer_bh(struct net_device
*dev
);
147 /* Interrupt handler */
148 static void plip_interrupt(int irq
, void *dev_id
, struct pt_regs
*regs
);
150 /* Functions for DEV methods */
151 static int plip_tx_packet(struct sk_buff
*skb
, struct net_device
*dev
);
152 static int plip_hard_header(struct sk_buff
*skb
, struct net_device
*dev
,
153 unsigned short type
, void *daddr
,
154 void *saddr
, unsigned len
);
155 static int plip_hard_header_cache(struct neighbour
*neigh
,
156 struct hh_cache
*hh
);
157 static int plip_open(struct net_device
*dev
);
158 static int plip_close(struct net_device
*dev
);
159 static struct net_device_stats
*plip_get_stats(struct net_device
*dev
);
160 static int plip_ioctl(struct net_device
*dev
, struct ifreq
*ifr
, int cmd
);
161 static int plip_preempt(void *handle
);
162 static void plip_wakeup(void *handle
);
164 enum plip_connection_state
{
172 enum plip_packet_state
{
181 enum plip_nibble_state
{
188 enum plip_packet_state state
;
189 enum plip_nibble_state nibble
;
192 #if defined(__LITTLE_ENDIAN)
195 #elif defined(__BIG_ENDIAN)
199 #error "Please fix the endianness defines in <asm/byteorder.h>"
205 unsigned char checksum
;
211 struct net_device_stats enet_stats
;
212 struct work_struct immediate
;
213 struct work_struct deferred
;
214 struct work_struct timer
;
215 struct plip_local snd_data
;
216 struct plip_local rcv_data
;
217 struct pardevice
*pardev
;
218 unsigned long trigger
;
219 unsigned long nibble
;
220 enum plip_connection_state connection
;
221 unsigned short timeout_count
;
224 int should_relinquish
;
225 int (*orig_hard_header
)(struct sk_buff
*skb
, struct net_device
*dev
,
226 unsigned short type
, void *daddr
,
227 void *saddr
, unsigned len
);
228 int (*orig_hard_header_cache
)(struct neighbour
*neigh
,
229 struct hh_cache
*hh
);
232 struct semaphore killed_timer_sem
;
235 inline static void enable_parport_interrupts (struct net_device
*dev
)
239 struct parport
*port
=
240 ((struct net_local
*)dev
->priv
)->pardev
->port
;
241 port
->ops
->enable_irq (port
);
245 inline static void disable_parport_interrupts (struct net_device
*dev
)
249 struct parport
*port
=
250 ((struct net_local
*)dev
->priv
)->pardev
->port
;
251 port
->ops
->disable_irq (port
);
255 inline static void write_data (struct net_device
*dev
, unsigned char data
)
257 struct parport
*port
=
258 ((struct net_local
*)dev
->priv
)->pardev
->port
;
260 port
->ops
->write_data (port
, data
);
263 inline static unsigned char read_status (struct net_device
*dev
)
265 struct parport
*port
=
266 ((struct net_local
*)dev
->priv
)->pardev
->port
;
268 return port
->ops
->read_status (port
);
271 /* Entry point of PLIP driver.
272 Probe the hardware, and register/initialize the driver.
274 PLIP is rather weird, because of the way it interacts with the parport
275 system. It is _not_ initialised from Space.c. Instead, plip_init()
276 is called, and that function makes up a "struct net_device" for each port, and
281 plip_init_netdev(struct net_device
*dev
)
283 struct net_local
*nl
= netdev_priv(dev
);
285 /* Then, override parts of it */
286 dev
->hard_start_xmit
= plip_tx_packet
;
287 dev
->open
= plip_open
;
288 dev
->stop
= plip_close
;
289 dev
->get_stats
= plip_get_stats
;
290 dev
->do_ioctl
= plip_ioctl
;
291 dev
->header_cache_update
= NULL
;
292 dev
->tx_queue_len
= 10;
293 dev
->flags
= IFF_POINTOPOINT
|IFF_NOARP
;
294 memset(dev
->dev_addr
, 0xfc, ETH_ALEN
);
296 /* Set the private structure */
297 nl
->orig_hard_header
= dev
->hard_header
;
298 dev
->hard_header
= plip_hard_header
;
300 nl
->orig_hard_header_cache
= dev
->hard_header_cache
;
301 dev
->hard_header_cache
= plip_hard_header_cache
;
306 /* Initialize constants */
307 nl
->trigger
= PLIP_TRIGGER_WAIT
;
308 nl
->nibble
= PLIP_NIBBLE_WAIT
;
310 /* Initialize task queue structures */
311 INIT_WORK(&nl
->immediate
, (void (*)(void *))plip_bh
, dev
);
312 INIT_WORK(&nl
->deferred
, (void (*)(void *))plip_kick_bh
, dev
);
315 INIT_WORK(&nl
->timer
, (void (*)(void *))plip_timer_bh
, dev
);
317 spin_lock_init(&nl
->lock
);
320 /* Bottom half handler for the delayed request.
321 This routine is kicked by do_timer().
322 Request `plip_bh' to be invoked. */
324 plip_kick_bh(struct net_device
*dev
)
326 struct net_local
*nl
= netdev_priv(dev
);
329 schedule_work(&nl
->immediate
);
332 /* Forward declarations of internal routines */
333 static int plip_none(struct net_device
*, struct net_local
*,
334 struct plip_local
*, struct plip_local
*);
335 static int plip_receive_packet(struct net_device
*, struct net_local
*,
336 struct plip_local
*, struct plip_local
*);
337 static int plip_send_packet(struct net_device
*, struct net_local
*,
338 struct plip_local
*, struct plip_local
*);
339 static int plip_connection_close(struct net_device
*, struct net_local
*,
340 struct plip_local
*, struct plip_local
*);
341 static int plip_error(struct net_device
*, struct net_local
*,
342 struct plip_local
*, struct plip_local
*);
343 static int plip_bh_timeout_error(struct net_device
*dev
, struct net_local
*nl
,
344 struct plip_local
*snd
,
345 struct plip_local
*rcv
,
353 typedef int (*plip_func
)(struct net_device
*dev
, struct net_local
*nl
,
354 struct plip_local
*snd
, struct plip_local
*rcv
);
356 static plip_func connection_state_table
[] =
361 plip_connection_close
,
365 /* Bottom half handler of PLIP. */
367 plip_bh(struct net_device
*dev
)
369 struct net_local
*nl
= netdev_priv(dev
);
370 struct plip_local
*snd
= &nl
->snd_data
;
371 struct plip_local
*rcv
= &nl
->rcv_data
;
376 f
= connection_state_table
[nl
->connection
];
377 if ((r
= (*f
)(dev
, nl
, snd
, rcv
)) != OK
378 && (r
= plip_bh_timeout_error(dev
, nl
, snd
, rcv
, r
)) != OK
) {
380 schedule_delayed_work(&nl
->deferred
, 1);
385 plip_timer_bh(struct net_device
*dev
)
387 struct net_local
*nl
= netdev_priv(dev
);
389 if (!(atomic_read (&nl
->kill_timer
))) {
390 plip_interrupt (-1, dev
, NULL
);
392 schedule_delayed_work(&nl
->timer
, 1);
395 up (&nl
->killed_timer_sem
);
400 plip_bh_timeout_error(struct net_device
*dev
, struct net_local
*nl
,
401 struct plip_local
*snd
, struct plip_local
*rcv
,
406 * This is tricky. If we got here from the beginning of send (either
407 * with ERROR or HS_TIMEOUT) we have IRQ enabled. Otherwise it's
408 * already disabled. With the old variant of {enable,disable}_irq()
409 * extra disable_irq() was a no-op. Now it became mortal - it's
410 * unbalanced and thus we'll never re-enable IRQ (until rmmod plip,
411 * that is). So we have to treat HS_TIMEOUT and ERROR from send
415 spin_lock_irq(&nl
->lock
);
416 if (nl
->connection
== PLIP_CN_SEND
) {
418 if (error
!= ERROR
) { /* Timeout */
420 if ((error
== HS_TIMEOUT
421 && nl
->timeout_count
<= 10)
422 || nl
->timeout_count
<= 3) {
423 spin_unlock_irq(&nl
->lock
);
424 /* Try again later */
427 c0
= read_status(dev
);
428 printk(KERN_WARNING
"%s: transmit timeout(%d,%02x)\n",
429 dev
->name
, snd
->state
, c0
);
432 nl
->enet_stats
.tx_errors
++;
433 nl
->enet_stats
.tx_aborted_errors
++;
434 } else if (nl
->connection
== PLIP_CN_RECEIVE
) {
435 if (rcv
->state
== PLIP_PK_TRIGGER
) {
436 /* Transmission was interrupted. */
437 spin_unlock_irq(&nl
->lock
);
440 if (error
!= ERROR
) { /* Timeout */
441 if (++nl
->timeout_count
<= 3) {
442 spin_unlock_irq(&nl
->lock
);
443 /* Try again later */
446 c0
= read_status(dev
);
447 printk(KERN_WARNING
"%s: receive timeout(%d,%02x)\n",
448 dev
->name
, rcv
->state
, c0
);
450 nl
->enet_stats
.rx_dropped
++;
452 rcv
->state
= PLIP_PK_DONE
;
457 snd
->state
= PLIP_PK_DONE
;
459 dev_kfree_skb(snd
->skb
);
462 spin_unlock_irq(&nl
->lock
);
463 if (error
== HS_TIMEOUT
) {
465 synchronize_irq(dev
->irq
);
467 disable_parport_interrupts (dev
);
468 netif_stop_queue (dev
);
469 nl
->connection
= PLIP_CN_ERROR
;
470 write_data (dev
, 0x00);
476 plip_none(struct net_device
*dev
, struct net_local
*nl
,
477 struct plip_local
*snd
, struct plip_local
*rcv
)
482 /* PLIP_RECEIVE --- receive a byte(two nibbles)
483 Returns OK on success, TIMEOUT on timeout */
485 plip_receive(unsigned short nibble_timeout
, struct net_device
*dev
,
486 enum plip_nibble_state
*ns_p
, unsigned char *data_p
)
488 unsigned char c0
, c1
;
495 c0
= read_status(dev
);
496 udelay(PLIP_DELAY_UNIT
);
497 if ((c0
& 0x80) == 0) {
498 c1
= read_status(dev
);
505 *data_p
= (c0
>> 3) & 0x0f;
506 write_data (dev
, 0x10); /* send ACK */
512 c0
= read_status(dev
);
513 udelay(PLIP_DELAY_UNIT
);
515 c1
= read_status(dev
);
522 *data_p
|= (c0
<< 1) & 0xf0;
523 write_data (dev
, 0x00); /* send ACK */
524 *ns_p
= PLIP_NB_BEGIN
;
532 * Determine the packet's protocol ID. The rule here is that we
533 * assume 802.3 if the type field is short enough to be a length.
534 * This is normal practice and works for any 'now in use' protocol.
536 * PLIP is ethernet ish but the daddr might not be valid if unicast.
537 * PLIP fortunately has no bus architecture (its Point-to-point).
539 * We can't fix the daddr thing as that quirk (more bug) is embedded
540 * in far too many old systems not all even running Linux.
543 static unsigned short plip_type_trans(struct sk_buff
*skb
, struct net_device
*dev
)
548 skb
->mac
.raw
=skb
->data
;
549 skb_pull(skb
,dev
->hard_header_len
);
554 if(memcmp(eth
->h_dest
,dev
->broadcast
, ETH_ALEN
)==0)
555 skb
->pkt_type
=PACKET_BROADCAST
;
557 skb
->pkt_type
=PACKET_MULTICAST
;
561 * This ALLMULTI check should be redundant by 1.4
562 * so don't forget to remove it.
565 if (ntohs(eth
->h_proto
) >= 1536)
571 * This is a magic hack to spot IPX packets. Older Novell breaks
572 * the protocol design and runs IPX over 802.3 without an 802.2 LLC
573 * layer. We look for FFFF which isn't a used 802.2 SSAP/DSAP. This
574 * won't work for fault tolerant netware but does for the rest.
576 if (*(unsigned short *)rawp
== 0xFFFF)
577 return htons(ETH_P_802_3
);
582 return htons(ETH_P_802_2
);
586 /* PLIP_RECEIVE_PACKET --- receive a packet */
588 plip_receive_packet(struct net_device
*dev
, struct net_local
*nl
,
589 struct plip_local
*snd
, struct plip_local
*rcv
)
591 unsigned short nibble_timeout
= nl
->nibble
;
594 switch (rcv
->state
) {
595 case PLIP_PK_TRIGGER
:
597 /* Don't need to synchronize irq, as we can safely ignore it */
598 disable_parport_interrupts (dev
);
599 write_data (dev
, 0x01); /* send ACK */
601 printk(KERN_DEBUG
"%s: receive start\n", dev
->name
);
602 rcv
->state
= PLIP_PK_LENGTH_LSB
;
603 rcv
->nibble
= PLIP_NB_BEGIN
;
605 case PLIP_PK_LENGTH_LSB
:
606 if (snd
->state
!= PLIP_PK_DONE
) {
607 if (plip_receive(nl
->trigger
, dev
,
608 &rcv
->nibble
, &rcv
->length
.b
.lsb
)) {
609 /* collision, here dev->tbusy == 1 */
610 rcv
->state
= PLIP_PK_DONE
;
612 nl
->connection
= PLIP_CN_SEND
;
613 schedule_delayed_work(&nl
->deferred
, 1);
614 enable_parport_interrupts (dev
);
619 if (plip_receive(nibble_timeout
, dev
,
620 &rcv
->nibble
, &rcv
->length
.b
.lsb
))
623 rcv
->state
= PLIP_PK_LENGTH_MSB
;
625 case PLIP_PK_LENGTH_MSB
:
626 if (plip_receive(nibble_timeout
, dev
,
627 &rcv
->nibble
, &rcv
->length
.b
.msb
))
629 if (rcv
->length
.h
> dev
->mtu
+ dev
->hard_header_len
630 || rcv
->length
.h
< 8) {
631 printk(KERN_WARNING
"%s: bogus packet size %d.\n", dev
->name
, rcv
->length
.h
);
634 /* Malloc up new buffer. */
635 rcv
->skb
= dev_alloc_skb(rcv
->length
.h
+ 2);
636 if (rcv
->skb
== NULL
) {
637 printk(KERN_ERR
"%s: Memory squeeze.\n", dev
->name
);
640 skb_reserve(rcv
->skb
, 2); /* Align IP on 16 byte boundaries */
641 skb_put(rcv
->skb
,rcv
->length
.h
);
643 rcv
->state
= PLIP_PK_DATA
;
648 lbuf
= rcv
->skb
->data
;
650 if (plip_receive(nibble_timeout
, dev
,
651 &rcv
->nibble
, &lbuf
[rcv
->byte
]))
653 while (++rcv
->byte
< rcv
->length
.h
);
655 rcv
->checksum
+= lbuf
[--rcv
->byte
];
657 rcv
->state
= PLIP_PK_CHECKSUM
;
659 case PLIP_PK_CHECKSUM
:
660 if (plip_receive(nibble_timeout
, dev
,
661 &rcv
->nibble
, &rcv
->data
))
663 if (rcv
->data
!= rcv
->checksum
) {
664 nl
->enet_stats
.rx_crc_errors
++;
666 printk(KERN_DEBUG
"%s: checksum error\n", dev
->name
);
669 rcv
->state
= PLIP_PK_DONE
;
672 /* Inform the upper layer for the arrival of a packet. */
673 rcv
->skb
->protocol
=plip_type_trans(rcv
->skb
, dev
);
675 dev
->last_rx
= jiffies
;
676 nl
->enet_stats
.rx_bytes
+= rcv
->length
.h
;
677 nl
->enet_stats
.rx_packets
++;
680 printk(KERN_DEBUG
"%s: receive end\n", dev
->name
);
682 /* Close the connection. */
683 write_data (dev
, 0x00);
684 spin_lock_irq(&nl
->lock
);
685 if (snd
->state
!= PLIP_PK_DONE
) {
686 nl
->connection
= PLIP_CN_SEND
;
687 spin_unlock_irq(&nl
->lock
);
688 schedule_work(&nl
->immediate
);
689 enable_parport_interrupts (dev
);
693 nl
->connection
= PLIP_CN_NONE
;
694 spin_unlock_irq(&nl
->lock
);
695 enable_parport_interrupts (dev
);
703 /* PLIP_SEND --- send a byte (two nibbles)
704 Returns OK on success, TIMEOUT when timeout */
706 plip_send(unsigned short nibble_timeout
, struct net_device
*dev
,
707 enum plip_nibble_state
*ns_p
, unsigned char data
)
714 write_data (dev
, data
& 0x0f);
718 write_data (dev
, 0x10 | (data
& 0x0f));
721 c0
= read_status(dev
);
722 if ((c0
& 0x80) == 0)
726 udelay(PLIP_DELAY_UNIT
);
728 write_data (dev
, 0x10 | (data
>> 4));
732 write_data (dev
, (data
>> 4));
735 c0
= read_status(dev
);
740 udelay(PLIP_DELAY_UNIT
);
742 *ns_p
= PLIP_NB_BEGIN
;
748 /* PLIP_SEND_PACKET --- send a packet */
750 plip_send_packet(struct net_device
*dev
, struct net_local
*nl
,
751 struct plip_local
*snd
, struct plip_local
*rcv
)
753 unsigned short nibble_timeout
= nl
->nibble
;
758 if (snd
->skb
== NULL
|| (lbuf
= snd
->skb
->data
) == NULL
) {
759 printk(KERN_DEBUG
"%s: send skb lost\n", dev
->name
);
760 snd
->state
= PLIP_PK_DONE
;
765 switch (snd
->state
) {
766 case PLIP_PK_TRIGGER
:
767 if ((read_status(dev
) & 0xf8) != 0x80)
770 /* Trigger remote rx interrupt. */
771 write_data (dev
, 0x08);
774 udelay(PLIP_DELAY_UNIT
);
775 spin_lock_irq(&nl
->lock
);
776 if (nl
->connection
== PLIP_CN_RECEIVE
) {
777 spin_unlock_irq(&nl
->lock
);
779 nl
->enet_stats
.collisions
++;
782 c0
= read_status(dev
);
784 spin_unlock_irq(&nl
->lock
);
786 synchronize_irq(dev
->irq
);
787 if (nl
->connection
== PLIP_CN_RECEIVE
) {
789 We don't need to enable irq,
790 as it is soon disabled. */
791 /* Yes, we do. New variant of
792 {enable,disable}_irq *counts*
795 nl
->enet_stats
.collisions
++;
798 disable_parport_interrupts (dev
);
800 printk(KERN_DEBUG
"%s: send start\n", dev
->name
);
801 snd
->state
= PLIP_PK_LENGTH_LSB
;
802 snd
->nibble
= PLIP_NB_BEGIN
;
803 nl
->timeout_count
= 0;
806 spin_unlock_irq(&nl
->lock
);
808 write_data (dev
, 0x00);
813 case PLIP_PK_LENGTH_LSB
:
814 if (plip_send(nibble_timeout
, dev
,
815 &snd
->nibble
, snd
->length
.b
.lsb
))
817 snd
->state
= PLIP_PK_LENGTH_MSB
;
819 case PLIP_PK_LENGTH_MSB
:
820 if (plip_send(nibble_timeout
, dev
,
821 &snd
->nibble
, snd
->length
.b
.msb
))
823 snd
->state
= PLIP_PK_DATA
;
829 if (plip_send(nibble_timeout
, dev
,
830 &snd
->nibble
, lbuf
[snd
->byte
]))
832 while (++snd
->byte
< snd
->length
.h
);
834 snd
->checksum
+= lbuf
[--snd
->byte
];
836 snd
->state
= PLIP_PK_CHECKSUM
;
838 case PLIP_PK_CHECKSUM
:
839 if (plip_send(nibble_timeout
, dev
,
840 &snd
->nibble
, snd
->checksum
))
843 nl
->enet_stats
.tx_bytes
+= snd
->skb
->len
;
844 dev_kfree_skb(snd
->skb
);
845 nl
->enet_stats
.tx_packets
++;
846 snd
->state
= PLIP_PK_DONE
;
849 /* Close the connection */
850 write_data (dev
, 0x00);
853 printk(KERN_DEBUG
"%s: send end\n", dev
->name
);
854 nl
->connection
= PLIP_CN_CLOSING
;
856 schedule_delayed_work(&nl
->deferred
, 1);
857 enable_parport_interrupts (dev
);
865 plip_connection_close(struct net_device
*dev
, struct net_local
*nl
,
866 struct plip_local
*snd
, struct plip_local
*rcv
)
868 spin_lock_irq(&nl
->lock
);
869 if (nl
->connection
== PLIP_CN_CLOSING
) {
870 nl
->connection
= PLIP_CN_NONE
;
871 netif_wake_queue (dev
);
873 spin_unlock_irq(&nl
->lock
);
874 if (nl
->should_relinquish
) {
875 nl
->should_relinquish
= nl
->port_owner
= 0;
876 parport_release(nl
->pardev
);
881 /* PLIP_ERROR --- wait till other end settled */
883 plip_error(struct net_device
*dev
, struct net_local
*nl
,
884 struct plip_local
*snd
, struct plip_local
*rcv
)
886 unsigned char status
;
888 status
= read_status(dev
);
889 if ((status
& 0xf8) == 0x80) {
891 printk(KERN_DEBUG
"%s: reset interface.\n", dev
->name
);
892 nl
->connection
= PLIP_CN_NONE
;
893 nl
->should_relinquish
= 0;
894 netif_start_queue (dev
);
895 enable_parport_interrupts (dev
);
897 netif_wake_queue (dev
);
900 schedule_delayed_work(&nl
->deferred
, 1);
906 /* Handle the parallel port interrupts. */
908 plip_interrupt(int irq
, void *dev_id
, struct pt_regs
* regs
)
910 struct net_device
*dev
= dev_id
;
911 struct net_local
*nl
;
912 struct plip_local
*rcv
;
916 printk(KERN_DEBUG
"plip_interrupt: irq %d for unknown device.\n", irq
);
920 nl
= netdev_priv(dev
);
923 spin_lock_irq (&nl
->lock
);
925 c0
= read_status(dev
);
926 if ((c0
& 0xf8) != 0xc0) {
927 if ((dev
->irq
!= -1) && (net_debug
> 1))
928 printk(KERN_DEBUG
"%s: spurious interrupt\n", dev
->name
);
929 spin_unlock_irq (&nl
->lock
);
934 printk(KERN_DEBUG
"%s: interrupt.\n", dev
->name
);
936 switch (nl
->connection
) {
937 case PLIP_CN_CLOSING
:
938 netif_wake_queue (dev
);
941 rcv
->state
= PLIP_PK_TRIGGER
;
942 nl
->connection
= PLIP_CN_RECEIVE
;
943 nl
->timeout_count
= 0;
944 schedule_work(&nl
->immediate
);
947 case PLIP_CN_RECEIVE
:
948 /* May occur because there is race condition
949 around test and set of dev->interrupt.
950 Ignore this interrupt. */
954 printk(KERN_ERR
"%s: receive interrupt in error state\n", dev
->name
);
958 spin_unlock_irq(&nl
->lock
);
962 plip_tx_packet(struct sk_buff
*skb
, struct net_device
*dev
)
964 struct net_local
*nl
= netdev_priv(dev
);
965 struct plip_local
*snd
= &nl
->snd_data
;
967 if (netif_queue_stopped(dev
))
970 /* We may need to grab the bus */
971 if (!nl
->port_owner
) {
972 if (parport_claim(nl
->pardev
))
977 netif_stop_queue (dev
);
979 if (skb
->len
> dev
->mtu
+ dev
->hard_header_len
) {
980 printk(KERN_WARNING
"%s: packet too big, %d.\n", dev
->name
, (int)skb
->len
);
981 netif_start_queue (dev
);
986 printk(KERN_DEBUG
"%s: send request\n", dev
->name
);
988 spin_lock_irq(&nl
->lock
);
989 dev
->trans_start
= jiffies
;
991 snd
->length
.h
= skb
->len
;
992 snd
->state
= PLIP_PK_TRIGGER
;
993 if (nl
->connection
== PLIP_CN_NONE
) {
994 nl
->connection
= PLIP_CN_SEND
;
995 nl
->timeout_count
= 0;
997 schedule_work(&nl
->immediate
);
998 spin_unlock_irq(&nl
->lock
);
1004 plip_rewrite_address(struct net_device
*dev
, struct ethhdr
*eth
)
1006 struct in_device
*in_dev
;
1008 if ((in_dev
=dev
->ip_ptr
) != NULL
) {
1009 /* Any address will do - we take the first */
1010 struct in_ifaddr
*ifa
=in_dev
->ifa_list
;
1012 memcpy(eth
->h_source
, dev
->dev_addr
, 6);
1013 memset(eth
->h_dest
, 0xfc, 2);
1014 memcpy(eth
->h_dest
+2, &ifa
->ifa_address
, 4);
1020 plip_hard_header(struct sk_buff
*skb
, struct net_device
*dev
,
1021 unsigned short type
, void *daddr
,
1022 void *saddr
, unsigned len
)
1024 struct net_local
*nl
= netdev_priv(dev
);
1027 if ((ret
= nl
->orig_hard_header(skb
, dev
, type
, daddr
, saddr
, len
)) >= 0)
1028 plip_rewrite_address (dev
, (struct ethhdr
*)skb
->data
);
1033 int plip_hard_header_cache(struct neighbour
*neigh
,
1034 struct hh_cache
*hh
)
1036 struct net_local
*nl
= neigh
->dev
->priv
;
1039 if ((ret
= nl
->orig_hard_header_cache(neigh
, hh
)) == 0)
1043 eth
= (struct ethhdr
*)(((u8
*)hh
->hh_data
) +
1044 HH_DATA_OFF(sizeof(*eth
)));
1045 plip_rewrite_address (neigh
->dev
, eth
);
1051 /* Open/initialize the board. This is called (in the current kernel)
1052 sometime after booting when the 'ifconfig' program is run.
1054 This routine gets exclusive access to the parallel port by allocating
1058 plip_open(struct net_device
*dev
)
1060 struct net_local
*nl
= netdev_priv(dev
);
1061 struct in_device
*in_dev
;
1064 if (!nl
->port_owner
) {
1065 if (parport_claim(nl
->pardev
)) return -EAGAIN
;
1069 nl
->should_relinquish
= 0;
1071 /* Clear the data port. */
1072 write_data (dev
, 0x00);
1074 /* Enable rx interrupt. */
1075 enable_parport_interrupts (dev
);
1078 atomic_set (&nl
->kill_timer
, 0);
1079 schedule_delayed_work(&nl
->timer
, 1);
1082 /* Initialize the state machine. */
1083 nl
->rcv_data
.state
= nl
->snd_data
.state
= PLIP_PK_DONE
;
1084 nl
->rcv_data
.skb
= nl
->snd_data
.skb
= NULL
;
1085 nl
->connection
= PLIP_CN_NONE
;
1086 nl
->is_deferred
= 0;
1088 /* Fill in the MAC-level header.
1089 We used to abuse dev->broadcast to store the point-to-point
1090 MAC address, but we no longer do it. Instead, we fetch the
1091 interface address whenever it is needed, which is cheap enough
1092 because we use the hh_cache. Actually, abusing dev->broadcast
1093 didn't work, because when using plip_open the point-to-point
1094 address isn't yet known.
1095 PLIP doesn't have a real MAC address, but we need it to be
1096 DOS compatible, and to properly support taps (otherwise,
1097 when the device address isn't identical to the address of a
1098 received frame, the kernel incorrectly drops it). */
1100 if ((in_dev
=dev
->ip_ptr
) != NULL
) {
1101 /* Any address will do - we take the first. We already
1102 have the first two bytes filled with 0xfc, from
1104 struct in_ifaddr
*ifa
=in_dev
->ifa_list
;
1106 memcpy(dev
->dev_addr
+2, &ifa
->ifa_local
, 4);
1110 netif_start_queue (dev
);
1115 /* The inverse routine to plip_open (). */
1117 plip_close(struct net_device
*dev
)
1119 struct net_local
*nl
= netdev_priv(dev
);
1120 struct plip_local
*snd
= &nl
->snd_data
;
1121 struct plip_local
*rcv
= &nl
->rcv_data
;
1123 netif_stop_queue (dev
);
1125 synchronize_irq(dev
->irq
);
1129 init_MUTEX_LOCKED (&nl
->killed_timer_sem
);
1130 atomic_set (&nl
->kill_timer
, 1);
1131 down (&nl
->killed_timer_sem
);
1135 outb(0x00, PAR_DATA(dev
));
1137 nl
->is_deferred
= 0;
1138 nl
->connection
= PLIP_CN_NONE
;
1139 if (nl
->port_owner
) {
1140 parport_release(nl
->pardev
);
1144 snd
->state
= PLIP_PK_DONE
;
1146 dev_kfree_skb(snd
->skb
);
1149 rcv
->state
= PLIP_PK_DONE
;
1151 kfree_skb(rcv
->skb
);
1157 outb(0x00, PAR_CONTROL(dev
));
1163 plip_preempt(void *handle
)
1165 struct net_device
*dev
= (struct net_device
*)handle
;
1166 struct net_local
*nl
= netdev_priv(dev
);
1168 /* Stand our ground if a datagram is on the wire */
1169 if (nl
->connection
!= PLIP_CN_NONE
) {
1170 nl
->should_relinquish
= 1;
1174 nl
->port_owner
= 0; /* Remember that we released the bus */
1179 plip_wakeup(void *handle
)
1181 struct net_device
*dev
= (struct net_device
*)handle
;
1182 struct net_local
*nl
= netdev_priv(dev
);
1184 if (nl
->port_owner
) {
1185 /* Why are we being woken up? */
1186 printk(KERN_DEBUG
"%s: why am I being woken up?\n", dev
->name
);
1187 if (!parport_claim(nl
->pardev
))
1188 /* bus_owner is already set (but why?) */
1189 printk(KERN_DEBUG
"%s: I'm broken.\n", dev
->name
);
1194 if (!(dev
->flags
& IFF_UP
))
1195 /* Don't need the port when the interface is down */
1198 if (!parport_claim(nl
->pardev
)) {
1200 /* Clear the data port. */
1201 write_data (dev
, 0x00);
1207 static struct net_device_stats
*
1208 plip_get_stats(struct net_device
*dev
)
1210 struct net_local
*nl
= netdev_priv(dev
);
1211 struct net_device_stats
*r
= &nl
->enet_stats
;
1217 plip_ioctl(struct net_device
*dev
, struct ifreq
*rq
, int cmd
)
1219 struct net_local
*nl
= netdev_priv(dev
);
1220 struct plipconf
*pc
= (struct plipconf
*) &rq
->ifr_ifru
;
1222 if (cmd
!= SIOCDEVPLIP
)
1226 case PLIP_GET_TIMEOUT
:
1227 pc
->trigger
= nl
->trigger
;
1228 pc
->nibble
= nl
->nibble
;
1230 case PLIP_SET_TIMEOUT
:
1231 if(!capable(CAP_NET_ADMIN
))
1233 nl
->trigger
= pc
->trigger
;
1234 nl
->nibble
= pc
->nibble
;
1242 static int parport
[PLIP_MAX
] = { [0 ... PLIP_MAX
-1] = -1 };
1245 module_param_array(parport
, int, NULL
, 0);
1246 module_param(timid
, int, 0);
1247 MODULE_PARM_DESC(parport
, "List of parport device numbers to use by plip");
1249 static struct net_device
*dev_plip
[PLIP_MAX
] = { NULL
, };
1252 plip_searchfor(int list
[], int a
)
1255 for (i
= 0; i
< PLIP_MAX
&& list
[i
] != -1; i
++) {
1256 if (list
[i
] == a
) return 1;
1261 /* plip_attach() is called (by the parport code) when a port is
1262 * available to use. */
1263 static void plip_attach (struct parport
*port
)
1266 struct net_device
*dev
;
1267 struct net_local
*nl
;
1268 char name
[IFNAMSIZ
];
1270 if ((parport
[0] == -1 && (!timid
|| !port
->devices
)) ||
1271 plip_searchfor(parport
, port
->number
)) {
1272 if (unit
== PLIP_MAX
) {
1273 printk(KERN_ERR
"plip: too many devices\n");
1277 sprintf(name
, "plip%d", unit
);
1278 dev
= alloc_etherdev(sizeof(struct net_local
));
1280 printk(KERN_ERR
"plip: memory squeeze\n");
1284 strcpy(dev
->name
, name
);
1286 SET_MODULE_OWNER(dev
);
1287 dev
->irq
= port
->irq
;
1288 dev
->base_addr
= port
->base
;
1289 if (port
->irq
== -1) {
1290 printk(KERN_INFO
"plip: %s has no IRQ. Using IRQ-less mode,"
1291 "which is fairly inefficient!\n", port
->name
);
1294 nl
= netdev_priv(dev
);
1295 nl
->pardev
= parport_register_device(port
, name
, plip_preempt
,
1296 plip_wakeup
, plip_interrupt
,
1300 printk(KERN_ERR
"%s: parport_register failed\n", name
);
1305 plip_init_netdev(dev
);
1307 if (register_netdev(dev
)) {
1308 printk(KERN_ERR
"%s: network register failed\n", name
);
1309 goto err_parport_unregister
;
1312 printk(KERN_INFO
"%s", version
);
1314 printk(KERN_INFO
"%s: Parallel port at %#3lx, "
1316 dev
->name
, dev
->base_addr
, dev
->irq
);
1318 printk(KERN_INFO
"%s: Parallel port at %#3lx, "
1320 dev
->name
, dev
->base_addr
);
1321 dev_plip
[unit
++] = dev
;
1325 err_parport_unregister
:
1326 parport_unregister_device(nl
->pardev
);
1332 /* plip_detach() is called (by the parport code) when a port is
1333 * no longer available to use. */
1334 static void plip_detach (struct parport
*port
)
1339 static struct parport_driver plip_driver
= {
1341 .attach
= plip_attach
,
1342 .detach
= plip_detach
1345 static void __exit
plip_cleanup_module (void)
1347 struct net_device
*dev
;
1350 parport_unregister_driver (&plip_driver
);
1352 for (i
=0; i
< PLIP_MAX
; i
++) {
1353 if ((dev
= dev_plip
[i
])) {
1354 struct net_local
*nl
= netdev_priv(dev
);
1355 unregister_netdev(dev
);
1357 parport_release(nl
->pardev
);
1358 parport_unregister_device(nl
->pardev
);
1367 static int parport_ptr
;
1369 static int __init
plip_setup(char *str
)
1373 str
= get_options(str
, ARRAY_SIZE(ints
), ints
);
1376 if (!strncmp(str
, "parport", 7)) {
1377 int n
= simple_strtoul(str
+7, NULL
, 10);
1378 if (parport_ptr
< PLIP_MAX
)
1379 parport
[parport_ptr
++] = n
;
1381 printk(KERN_INFO
"plip: too many ports, %s ignored.\n",
1383 } else if (!strcmp(str
, "timid")) {
1386 if (ints
[0] == 0 || ints
[1] == 0) {
1387 /* disable driver on "plip=" or "plip=0" */
1390 printk(KERN_WARNING
"warning: 'plip=0x%x' ignored\n",
1397 __setup("plip=", plip_setup
);
1399 #endif /* !MODULE */
1401 static int __init
plip_init (void)
1403 if (parport
[0] == -2)
1406 if (parport
[0] != -1 && timid
) {
1407 printk(KERN_WARNING
"plip: warning, ignoring `timid' since specific ports given.\n");
1411 if (parport_register_driver (&plip_driver
)) {
1412 printk (KERN_WARNING
"plip: couldn't register driver\n");
1419 module_init(plip_init
);
1420 module_exit(plip_cleanup_module
);
1421 MODULE_LICENSE("GPL");
1425 * compile-command: "gcc -DMODULE -DMODVERSIONS -D__KERNEL__ -Wall -Wstrict-prototypes -O2 -g -fomit-frame-pointer -pipe -c plip.c"