writeback: split writeback_inodes_wb
[linux-2.6/next.git] / net / irda / irlan / irlan_eth.c
blob9616c32d1076dda982fff4c6da5c6e1057cf39d1
1 /*********************************************************************
3 * Filename: irlan_eth.c
4 * Version:
5 * Description:
6 * Status: Experimental.
7 * Author: Dag Brattli <dagb@cs.uit.no>
8 * Created at: Thu Oct 15 08:37:58 1998
9 * Modified at: Tue Mar 21 09:06:41 2000
10 * Modified by: Dag Brattli <dagb@cs.uit.no>
11 * Sources: skeleton.c by Donald Becker <becker@CESDIS.gsfc.nasa.gov>
12 * slip.c by Laurence Culhane, <loz@holmes.demon.co.uk>
13 * Fred N. van Kempen, <waltje@uwalt.nl.mugnet.org>
15 * Copyright (c) 1998-2000 Dag Brattli, All Rights Reserved.
17 * This program is free software; you can redistribute it and/or
18 * modify it under the terms of the GNU General Public License as
19 * published by the Free Software Foundation; either version 2 of
20 * the License, or (at your option) any later version.
22 * Neither Dag Brattli nor University of Tromsø admit liability nor
23 * provide warranty for any of this software. This material is
24 * provided "AS-IS" and at no charge.
26 ********************************************************************/
28 #include <linux/netdevice.h>
29 #include <linux/etherdevice.h>
30 #include <linux/inetdevice.h>
31 #include <linux/if_arp.h>
32 #include <linux/module.h>
33 #include <linux/sched.h>
34 #include <net/arp.h>
36 #include <net/irda/irda.h>
37 #include <net/irda/irmod.h>
38 #include <net/irda/irlan_common.h>
39 #include <net/irda/irlan_client.h>
40 #include <net/irda/irlan_event.h>
41 #include <net/irda/irlan_eth.h>
43 static int irlan_eth_open(struct net_device *dev);
44 static int irlan_eth_close(struct net_device *dev);
45 static netdev_tx_t irlan_eth_xmit(struct sk_buff *skb,
46 struct net_device *dev);
47 static void irlan_eth_set_multicast_list( struct net_device *dev);
48 static struct net_device_stats *irlan_eth_get_stats(struct net_device *dev);
50 static const struct net_device_ops irlan_eth_netdev_ops = {
51 .ndo_open = irlan_eth_open,
52 .ndo_stop = irlan_eth_close,
53 .ndo_start_xmit = irlan_eth_xmit,
54 .ndo_get_stats = irlan_eth_get_stats,
55 .ndo_set_multicast_list = irlan_eth_set_multicast_list,
56 .ndo_change_mtu = eth_change_mtu,
57 .ndo_validate_addr = eth_validate_addr,
61 * Function irlan_eth_setup (dev)
63 * The network device initialization function.
66 static void irlan_eth_setup(struct net_device *dev)
68 ether_setup(dev);
70 dev->netdev_ops = &irlan_eth_netdev_ops;
71 dev->destructor = free_netdev;
75 * Lets do all queueing in IrTTP instead of this device driver.
76 * Queueing here as well can introduce some strange latency
77 * problems, which we will avoid by setting the queue size to 0.
80 * The bugs in IrTTP and IrLAN that created this latency issue
81 * have now been fixed, and we can propagate flow control properly
82 * to the network layer. However, this requires a minimal queue of
83 * packets for the device.
84 * Without flow control, the Tx Queue is 14 (ttp) + 0 (dev) = 14
85 * With flow control, the Tx Queue is 7 (ttp) + 4 (dev) = 11
86 * See irlan_eth_flow_indication()...
87 * Note : this number was randomly selected and would need to
88 * be adjusted.
89 * Jean II */
90 dev->tx_queue_len = 4;
94 * Function alloc_irlandev
96 * Allocate network device and control block
99 struct net_device *alloc_irlandev(const char *name)
101 return alloc_netdev(sizeof(struct irlan_cb), name,
102 irlan_eth_setup);
106 * Function irlan_eth_open (dev)
108 * Network device has been opened by user
111 static int irlan_eth_open(struct net_device *dev)
113 struct irlan_cb *self = netdev_priv(dev);
115 IRDA_DEBUG(2, "%s()\n", __func__ );
117 /* Ready to play! */
118 netif_stop_queue(dev); /* Wait until data link is ready */
120 /* We are now open, so time to do some work */
121 self->disconnect_reason = 0;
122 irlan_client_wakeup(self, self->saddr, self->daddr);
124 /* Make sure we have a hardware address before we return,
125 so DHCP clients gets happy */
126 return wait_event_interruptible(self->open_wait,
127 !self->tsap_data->connected);
131 * Function irlan_eth_close (dev)
133 * Stop the ether network device, his function will usually be called by
134 * ifconfig down. We should now disconnect the link, We start the
135 * close timer, so that the instance will be removed if we are unable
136 * to discover the remote device after the disconnect.
138 static int irlan_eth_close(struct net_device *dev)
140 struct irlan_cb *self = netdev_priv(dev);
142 IRDA_DEBUG(2, "%s()\n", __func__ );
144 /* Stop device */
145 netif_stop_queue(dev);
147 irlan_close_data_channel(self);
148 irlan_close_tsaps(self);
150 irlan_do_client_event(self, IRLAN_LMP_DISCONNECT, NULL);
151 irlan_do_provider_event(self, IRLAN_LMP_DISCONNECT, NULL);
153 /* Remove frames queued on the control channel */
154 skb_queue_purge(&self->client.txq);
156 self->client.tx_busy = 0;
158 return 0;
162 * Function irlan_eth_tx (skb)
164 * Transmits ethernet frames over IrDA link.
167 static netdev_tx_t irlan_eth_xmit(struct sk_buff *skb,
168 struct net_device *dev)
170 struct irlan_cb *self = netdev_priv(dev);
171 int ret;
173 /* skb headroom large enough to contain all IrDA-headers? */
174 if ((skb_headroom(skb) < self->max_header_size) || (skb_shared(skb))) {
175 struct sk_buff *new_skb =
176 skb_realloc_headroom(skb, self->max_header_size);
178 /* We have to free the original skb anyway */
179 dev_kfree_skb(skb);
181 /* Did the realloc succeed? */
182 if (new_skb == NULL)
183 return NETDEV_TX_OK;
185 /* Use the new skb instead */
186 skb = new_skb;
189 dev->trans_start = jiffies;
191 /* Now queue the packet in the transport layer */
192 if (self->use_udata)
193 ret = irttp_udata_request(self->tsap_data, skb);
194 else
195 ret = irttp_data_request(self->tsap_data, skb);
197 if (ret < 0) {
199 * IrTTPs tx queue is full, so we just have to
200 * drop the frame! You might think that we should
201 * just return -1 and don't deallocate the frame,
202 * but that is dangerous since it's possible that
203 * we have replaced the original skb with a new
204 * one with larger headroom, and that would really
205 * confuse do_dev_queue_xmit() in dev.c! I have
206 * tried :-) DB
208 /* irttp_data_request already free the packet */
209 self->stats.tx_dropped++;
210 } else {
211 self->stats.tx_packets++;
212 self->stats.tx_bytes += skb->len;
215 return NETDEV_TX_OK;
219 * Function irlan_eth_receive (handle, skb)
221 * This function gets the data that is received on the data channel
224 int irlan_eth_receive(void *instance, void *sap, struct sk_buff *skb)
226 struct irlan_cb *self = instance;
228 if (skb == NULL) {
229 ++self->stats.rx_dropped;
230 return 0;
232 if (skb->len < ETH_HLEN) {
233 IRDA_DEBUG(0, "%s() : IrLAN frame too short (%d)\n",
234 __func__, skb->len);
235 ++self->stats.rx_dropped;
236 dev_kfree_skb(skb);
237 return 0;
241 * Adopt this frame! Important to set all these fields since they
242 * might have been previously set by the low level IrDA network
243 * device driver
245 skb->protocol = eth_type_trans(skb, self->dev); /* Remove eth header */
247 self->stats.rx_packets++;
248 self->stats.rx_bytes += skb->len;
250 netif_rx(skb); /* Eat it! */
252 return 0;
256 * Function irlan_eth_flow (status)
258 * Do flow control between IP/Ethernet and IrLAN/IrTTP. This is done by
259 * controlling the queue stop/start.
261 * The IrDA link layer has the advantage to have flow control, and
262 * IrTTP now properly handles that. Flow controlling the higher layers
263 * prevent us to drop Tx packets in here (up to 15% for a TCP socket,
264 * more for UDP socket).
265 * Also, this allow us to reduce the overall transmit queue, which means
266 * less latency in case of mixed traffic.
267 * Jean II
269 void irlan_eth_flow_indication(void *instance, void *sap, LOCAL_FLOW flow)
271 struct irlan_cb *self;
272 struct net_device *dev;
274 self = (struct irlan_cb *) instance;
276 IRDA_ASSERT(self != NULL, return;);
277 IRDA_ASSERT(self->magic == IRLAN_MAGIC, return;);
279 dev = self->dev;
281 IRDA_ASSERT(dev != NULL, return;);
283 IRDA_DEBUG(0, "%s() : flow %s ; running %d\n", __func__,
284 flow == FLOW_STOP ? "FLOW_STOP" : "FLOW_START",
285 netif_running(dev));
287 switch (flow) {
288 case FLOW_STOP:
289 /* IrTTP is full, stop higher layers */
290 netif_stop_queue(dev);
291 break;
292 case FLOW_START:
293 default:
294 /* Tell upper layers that its time to transmit frames again */
295 /* Schedule network layer */
296 netif_wake_queue(dev);
297 break;
302 * Function set_multicast_list (dev)
304 * Configure the filtering of the device
307 #define HW_MAX_ADDRS 4 /* Must query to get it! */
308 static void irlan_eth_set_multicast_list(struct net_device *dev)
310 struct irlan_cb *self = netdev_priv(dev);
312 IRDA_DEBUG(2, "%s()\n", __func__ );
314 /* Check if data channel has been connected yet */
315 if (self->client.state != IRLAN_DATA) {
316 IRDA_DEBUG(1, "%s(), delaying!\n", __func__ );
317 return;
320 if (dev->flags & IFF_PROMISC) {
321 /* Enable promiscuous mode */
322 IRDA_WARNING("Promiscuous mode not implemented by IrLAN!\n");
324 else if ((dev->flags & IFF_ALLMULTI) ||
325 netdev_mc_count(dev) > HW_MAX_ADDRS) {
326 /* Disable promiscuous mode, use normal mode. */
327 IRDA_DEBUG(4, "%s(), Setting multicast filter\n", __func__ );
328 /* hardware_set_filter(NULL); */
330 irlan_set_multicast_filter(self, TRUE);
332 else if (!netdev_mc_empty(dev)) {
333 IRDA_DEBUG(4, "%s(), Setting multicast filter\n", __func__ );
334 /* Walk the address list, and load the filter */
335 /* hardware_set_filter(dev->mc_list); */
337 irlan_set_multicast_filter(self, TRUE);
339 else {
340 IRDA_DEBUG(4, "%s(), Clearing multicast filter\n", __func__ );
341 irlan_set_multicast_filter(self, FALSE);
344 if (dev->flags & IFF_BROADCAST)
345 irlan_set_broadcast_filter(self, TRUE);
346 else
347 irlan_set_broadcast_filter(self, FALSE);
351 * Function irlan_get_stats (dev)
353 * Get the current statistics for this device
356 static struct net_device_stats *irlan_eth_get_stats(struct net_device *dev)
358 struct irlan_cb *self = netdev_priv(dev);
360 return &self->stats;