1 /* -----------------------------------------------------------------------------
2 * Copyright (c) 2011 Ozmo Inc
3 * Released under the GNU General Public License Version 2 (GPLv2).
4 * -----------------------------------------------------------------------------
6 #include <linux/init.h>
7 #include <linux/module.h>
8 #include <linux/timer.h>
9 #include <linux/sched.h>
10 #include <linux/netdevice.h>
11 #include <linux/errno.h>
12 #include <linux/ieee80211.h>
14 #include "ozprotocol.h"
22 #include <asm/unaligned.h>
23 #include <linux/uaccess.h>
24 #include <net/psnap.h>
25 /*------------------------------------------------------------------------------
27 #define OZ_CF_CONN_SUCCESS 1
28 #define OZ_CF_CONN_FAILURE 2
33 /* States of the timer.
35 #define OZ_TIMER_IDLE 0
36 #define OZ_TIMER_SET 1
37 #define OZ_TIMER_IN_HANDLER 2
39 #define OZ_MAX_TIMER_POOL_SIZE 16
41 /*------------------------------------------------------------------------------
44 struct packet_type ptype
;
45 char name
[OZ_MAX_BINDING_LEN
];
46 struct oz_binding
*next
;
50 struct list_head link
;
52 unsigned long due_time
;
55 /*------------------------------------------------------------------------------
56 * Static external variables.
58 static DEFINE_SPINLOCK(g_polling_lock
);
59 static LIST_HEAD(g_pd_list
);
60 static struct oz_binding
*g_binding
;
61 static DEFINE_SPINLOCK(g_binding_lock
);
62 static struct sk_buff_head g_rx_queue
;
63 static u8 g_session_id
;
64 static u16 g_apps
= 0x1;
65 static int g_processing_rx
;
66 static struct timer_list g_timer
;
67 static struct oz_timer
*g_cur_timer
;
68 static struct list_head
*g_timer_pool
;
69 static int g_timer_pool_count
;
70 static int g_timer_state
= OZ_TIMER_IDLE
;
71 static LIST_HEAD(g_timer_list
);
72 /*------------------------------------------------------------------------------
74 static void oz_protocol_timer_start(void);
75 /*------------------------------------------------------------------------------
76 * Context: softirq-serialized
78 static u8
oz_get_new_session_id(u8 exclude
)
80 if (++g_session_id
== 0)
82 if (g_session_id
== exclude
) {
83 if (++g_session_id
== 0)
88 /*------------------------------------------------------------------------------
89 * Context: softirq-serialized
91 static void oz_send_conn_rsp(struct oz_pd
*pd
, u8 status
)
94 struct net_device
*dev
= pd
->net_dev
;
95 struct oz_hdr
*oz_hdr
;
97 struct oz_elt_connect_rsp
*body
;
98 int sz
= sizeof(struct oz_hdr
) + sizeof(struct oz_elt
) +
99 sizeof(struct oz_elt_connect_rsp
);
100 skb
= alloc_skb(sz
+ OZ_ALLOCATED_SPACE(dev
), GFP_ATOMIC
);
103 skb_reserve(skb
, LL_RESERVED_SPACE(dev
));
104 skb_reset_network_header(skb
);
105 oz_hdr
= (struct oz_hdr
*)skb_put(skb
, sz
);
106 elt
= (struct oz_elt
*)(oz_hdr
+1);
107 body
= (struct oz_elt_connect_rsp
*)(elt
+1);
109 skb
->protocol
= htons(OZ_ETHERTYPE
);
110 /* Fill in device header */
111 if (dev_hard_header(skb
, dev
, OZ_ETHERTYPE
, pd
->mac_addr
,
112 dev
->dev_addr
, skb
->len
) < 0) {
116 oz_hdr
->control
= (OZ_PROTOCOL_VERSION
<<OZ_VERSION_SHIFT
);
117 oz_hdr
->last_pkt_num
= 0;
118 put_unaligned(0, &oz_hdr
->pkt_num
);
119 oz_event_log(OZ_EVT_CONNECT_RSP
, 0, 0, NULL
, 0);
120 elt
->type
= OZ_ELT_CONNECT_RSP
;
121 elt
->length
= sizeof(struct oz_elt_connect_rsp
);
122 memset(body
, 0, sizeof(struct oz_elt_connect_rsp
));
123 body
->status
= status
;
125 body
->mode
= pd
->mode
;
126 body
->session_id
= pd
->session_id
;
127 put_unaligned(cpu_to_le16(pd
->total_apps
), &body
->apps
);
129 oz_trace("TX: OZ_ELT_CONNECT_RSP %d", status
);
133 /*------------------------------------------------------------------------------
134 * Context: softirq-serialized
136 static void pd_set_keepalive(struct oz_pd
*pd
, u8 kalive
)
138 unsigned long keep_alive
= kalive
& OZ_KALIVE_VALUE_MASK
;
140 switch (kalive
& OZ_KALIVE_TYPE_MASK
) {
141 case OZ_KALIVE_SPECIAL
:
143 oz_ms_to_jiffies(keep_alive
* 1000*60*60*24*20);
146 pd
->keep_alive_j
= oz_ms_to_jiffies(keep_alive
*1000);
149 pd
->keep_alive_j
= oz_ms_to_jiffies(keep_alive
*1000*60);
151 case OZ_KALIVE_HOURS
:
152 pd
->keep_alive_j
= oz_ms_to_jiffies(keep_alive
*1000*60*60);
155 pd
->keep_alive_j
= 0;
157 oz_trace("Keepalive = %lu jiffies\n", pd
->keep_alive_j
);
159 /*------------------------------------------------------------------------------
160 * Context: softirq-serialized
162 static void pd_set_presleep(struct oz_pd
*pd
, u8 presleep
)
165 pd
->presleep_j
= oz_ms_to_jiffies(presleep
*100);
167 pd
->presleep_j
= OZ_PRESLEEP_TOUT_J
;
168 oz_trace("Presleep time = %lu jiffies\n", pd
->presleep_j
);
170 /*------------------------------------------------------------------------------
171 * Context: softirq-serialized
173 static struct oz_pd
*oz_connect_req(struct oz_pd
*cur_pd
, struct oz_elt
*elt
,
174 const u8
*pd_addr
, struct net_device
*net_dev
)
177 struct oz_elt_connect_req
*body
=
178 (struct oz_elt_connect_req
*)(elt
+1);
179 u8 rsp_status
= OZ_STATUS_SUCCESS
;
181 u16 new_apps
= g_apps
;
182 struct net_device
*old_net_dev
= NULL
;
183 struct oz_pd
*free_pd
= NULL
;
186 spin_lock_bh(&g_polling_lock
);
188 struct oz_pd
*pd2
= NULL
;
190 pd
= oz_pd_alloc(pd_addr
);
193 pd
->last_rx_time_j
= jiffies
;
194 spin_lock_bh(&g_polling_lock
);
195 list_for_each(e
, &g_pd_list
) {
196 pd2
= container_of(e
, struct oz_pd
, link
);
197 if (memcmp(pd2
->mac_addr
, pd_addr
, ETH_ALEN
) == 0) {
204 list_add_tail(&pd
->link
, &g_pd_list
);
207 spin_unlock_bh(&g_polling_lock
);
210 if (pd
->net_dev
!= net_dev
) {
211 old_net_dev
= pd
->net_dev
;
213 pd
->net_dev
= net_dev
;
215 oz_trace("Host vendor: %d\n", body
->host_vendor
);
216 pd
->max_tx_size
= OZ_MAX_TX_SIZE
;
217 pd
->mode
= body
->mode
;
218 pd
->pd_info
= body
->pd_info
;
219 if (pd
->mode
& OZ_F_ISOC_NO_ELTS
) {
220 pd
->ms_per_isoc
= body
->ms_per_isoc
;
221 if (!pd
->ms_per_isoc
)
224 switch (body
->ms_isoc_latency
& OZ_LATENCY_MASK
) {
225 case OZ_ONE_MS_LATENCY
:
226 pd
->isoc_latency
= (body
->ms_isoc_latency
&
227 ~OZ_LATENCY_MASK
) / pd
->ms_per_isoc
;
229 case OZ_TEN_MS_LATENCY
:
230 pd
->isoc_latency
= ((body
->ms_isoc_latency
&
231 ~OZ_LATENCY_MASK
) * 10) / pd
->ms_per_isoc
;
234 pd
->isoc_latency
= OZ_MAX_TX_QUEUE_ISOC
;
237 if (body
->max_len_div16
)
238 pd
->max_tx_size
= ((u16
)body
->max_len_div16
)<<4;
239 oz_trace("Max frame:%u Ms per isoc:%u\n",
240 pd
->max_tx_size
, pd
->ms_per_isoc
);
241 pd
->max_stream_buffering
= 3*1024;
242 pd
->timeout_time_j
= jiffies
+ OZ_CONNECTION_TOUT_J
;
243 pd
->pulse_period_j
= OZ_QUANTUM_J
;
244 pd_set_presleep(pd
, body
->presleep
);
245 pd_set_keepalive(pd
, body
->keep_alive
);
247 new_apps
&= le16_to_cpu(get_unaligned(&body
->apps
));
248 if ((new_apps
& 0x1) && (body
->session_id
)) {
249 if (pd
->session_id
) {
250 if (pd
->session_id
!= body
->session_id
) {
251 rsp_status
= OZ_STATUS_SESSION_MISMATCH
;
255 new_apps
&= ~0x1; /* Resume not permitted */
257 oz_get_new_session_id(body
->session_id
);
260 if (pd
->session_id
&& !body
->session_id
) {
261 rsp_status
= OZ_STATUS_SESSION_TEARDOWN
;
264 new_apps
&= ~0x1; /* Resume not permitted */
266 oz_get_new_session_id(body
->session_id
);
270 if (rsp_status
== OZ_STATUS_SUCCESS
) {
271 u16 start_apps
= new_apps
& ~pd
->total_apps
& ~0x1;
272 u16 stop_apps
= pd
->total_apps
& ~new_apps
& ~0x1;
273 u16 resume_apps
= new_apps
& pd
->paused_apps
& ~0x1;
274 spin_unlock_bh(&g_polling_lock
);
275 oz_pd_set_state(pd
, OZ_PD_S_CONNECTED
);
276 oz_timer_delete(pd
, OZ_TIMER_STOP
);
277 oz_trace("new_apps=0x%x total_apps=0x%x paused_apps=0x%x\n",
278 new_apps
, pd
->total_apps
, pd
->paused_apps
);
280 if (oz_services_start(pd
, start_apps
, 0))
281 rsp_status
= OZ_STATUS_TOO_MANY_PDS
;
284 if (oz_services_start(pd
, resume_apps
, 1))
285 rsp_status
= OZ_STATUS_TOO_MANY_PDS
;
287 oz_services_stop(pd
, stop_apps
, 0);
288 oz_pd_request_heartbeat(pd
);
290 spin_unlock_bh(&g_polling_lock
);
292 oz_send_conn_rsp(pd
, rsp_status
);
293 if (rsp_status
!= OZ_STATUS_SUCCESS
) {
300 dev_put(old_net_dev
);
302 oz_pd_destroy(free_pd
);
305 /*------------------------------------------------------------------------------
306 * Context: softirq-serialized
308 static void oz_add_farewell(struct oz_pd
*pd
, u8 ep_num
, u8 index
,
309 const u8
*report
, u8 len
)
311 struct oz_farewell
*f
;
312 struct oz_farewell
*f2
;
314 f
= kmalloc(sizeof(struct oz_farewell
) + len
- 1, GFP_ATOMIC
);
319 memcpy(f
->report
, report
, len
);
320 oz_trace("RX: Adding farewell report\n");
321 spin_lock(&g_polling_lock
);
322 list_for_each_entry(f2
, &pd
->farewell_list
, link
) {
323 if ((f2
->ep_num
== ep_num
) && (f2
->index
== index
)) {
329 list_add_tail(&f
->link
, &pd
->farewell_list
);
330 spin_unlock(&g_polling_lock
);
334 /*------------------------------------------------------------------------------
335 * Context: softirq-serialized
337 static void oz_rx_frame(struct sk_buff
*skb
)
343 struct oz_pd
*pd
= NULL
;
344 struct oz_hdr
*oz_hdr
= (struct oz_hdr
*)skb_network_header(skb
);
348 oz_event_log(OZ_EVT_RX_PROCESS
, 0,
349 (((u16
)oz_hdr
->control
)<<8)|oz_hdr
->last_pkt_num
,
350 NULL
, oz_hdr
->pkt_num
);
351 oz_trace2(OZ_TRACE_RX_FRAMES
,
352 "RX frame PN=0x%x LPN=0x%x control=0x%x\n",
353 oz_hdr
->pkt_num
, oz_hdr
->last_pkt_num
, oz_hdr
->control
);
354 mac_hdr
= skb_mac_header(skb
);
355 src_addr
= &mac_hdr
[ETH_ALEN
] ;
358 /* Check the version field */
359 if (oz_get_prot_ver(oz_hdr
->control
) != OZ_PROTOCOL_VERSION
) {
360 oz_trace("Incorrect protocol version: %d\n",
361 oz_get_prot_ver(oz_hdr
->control
));
365 pkt_num
= le32_to_cpu(get_unaligned(&oz_hdr
->pkt_num
));
367 pd
= oz_pd_find(src_addr
);
369 pd
->last_rx_time_j
= jiffies
;
370 oz_timer_add(pd
, OZ_TIMER_TOUT
,
371 pd
->last_rx_time_j
+ pd
->presleep_j
, 1);
372 if (pkt_num
!= pd
->last_rx_pkt_num
) {
373 pd
->last_rx_pkt_num
= pkt_num
;
376 oz_trace("Duplicate frame\n");
380 if (pd
&& !dup
&& ((pd
->mode
& OZ_MODE_MASK
) == OZ_MODE_TRIGGERED
)) {
381 oz_trace2(OZ_TRACE_RX_FRAMES
, "Received TRIGGER Frame\n");
382 pd
->last_sent_frame
= &pd
->tx_queue
;
383 if (oz_hdr
->control
& OZ_F_ACK
) {
384 /* Retire completed frames */
385 oz_retire_tx_frames(pd
, oz_hdr
->last_pkt_num
);
387 if ((oz_hdr
->control
& OZ_F_ACK_REQUESTED
) &&
388 (pd
->state
== OZ_PD_S_CONNECTED
)) {
389 int backlog
= pd
->nb_queued_frames
;
390 pd
->trigger_pkt_num
= pkt_num
;
391 /* Send queued frames */
392 oz_send_queued_frames(pd
, backlog
);
396 length
-= sizeof(struct oz_hdr
);
397 elt
= (struct oz_elt
*)((u8
*)oz_hdr
+ sizeof(struct oz_hdr
));
399 while (length
>= sizeof(struct oz_elt
)) {
400 length
-= sizeof(struct oz_elt
) + elt
->length
;
404 case OZ_ELT_CONNECT_REQ
:
405 oz_event_log(OZ_EVT_CONNECT_REQ
, 0, 0, NULL
, 0);
406 oz_trace("RX: OZ_ELT_CONNECT_REQ\n");
407 pd
= oz_connect_req(pd
, elt
, src_addr
, skb
->dev
);
409 case OZ_ELT_DISCONNECT
:
410 oz_trace("RX: OZ_ELT_DISCONNECT\n");
414 case OZ_ELT_UPDATE_PARAM_REQ
: {
415 struct oz_elt_update_param
*body
=
416 (struct oz_elt_update_param
*)(elt
+ 1);
417 oz_trace("RX: OZ_ELT_UPDATE_PARAM_REQ\n");
418 if (pd
&& (pd
->state
& OZ_PD_S_CONNECTED
)) {
419 spin_lock(&g_polling_lock
);
420 pd_set_keepalive(pd
, body
->keepalive
);
421 pd_set_presleep(pd
, body
->presleep
);
422 spin_unlock(&g_polling_lock
);
426 case OZ_ELT_FAREWELL_REQ
: {
427 struct oz_elt_farewell
*body
=
428 (struct oz_elt_farewell
*)(elt
+ 1);
429 oz_trace("RX: OZ_ELT_FAREWELL_REQ\n");
430 oz_add_farewell(pd
, body
->ep_num
,
431 body
->index
, body
->report
,
432 elt
->length
+ 1 - sizeof(*body
));
435 case OZ_ELT_APP_DATA
:
436 if (pd
&& (pd
->state
& OZ_PD_S_CONNECTED
)) {
437 struct oz_app_hdr
*app_hdr
=
438 (struct oz_app_hdr
*)(elt
+1);
441 oz_handle_app_elt(pd
, app_hdr
->app_id
, elt
);
445 oz_trace("RX: Unknown elt %02x\n", elt
->type
);
447 elt
= oz_next_elt(elt
);
454 /*------------------------------------------------------------------------------
457 void oz_protocol_term(void)
459 struct list_head
*chain
;
460 del_timer_sync(&g_timer
);
461 /* Walk the list of bindings and remove each one.
463 spin_lock_bh(&g_binding_lock
);
465 struct oz_binding
*b
= g_binding
;
467 spin_unlock_bh(&g_binding_lock
);
468 dev_remove_pack(&b
->ptype
);
470 dev_put(b
->ptype
.dev
);
472 spin_lock_bh(&g_binding_lock
);
474 spin_unlock_bh(&g_binding_lock
);
475 /* Walk the list of PDs and stop each one. This causes the PD to be
476 * removed from the list so we can just pull each one from the head
479 spin_lock_bh(&g_polling_lock
);
480 while (!list_empty(&g_pd_list
)) {
482 list_first_entry(&g_pd_list
, struct oz_pd
, link
);
484 spin_unlock_bh(&g_polling_lock
);
487 spin_lock_bh(&g_polling_lock
);
489 chain
= g_timer_pool
;
491 spin_unlock_bh(&g_polling_lock
);
493 struct oz_timer
*t
= container_of(chain
, struct oz_timer
, link
);
497 oz_trace("Protocol stopped\n");
499 /*------------------------------------------------------------------------------
502 static void oz_pd_handle_timer(struct oz_pd
*pd
, int type
)
511 case OZ_TIMER_HEARTBEAT
: {
513 spin_lock_bh(&g_polling_lock
);
514 pd
->heartbeat_requested
= 0;
515 if (pd
->state
& OZ_PD_S_CONNECTED
)
516 apps
= pd
->total_apps
;
517 spin_unlock_bh(&g_polling_lock
);
519 oz_pd_heartbeat(pd
, apps
);
524 /*------------------------------------------------------------------------------
527 static void oz_protocol_timer(unsigned long arg
)
532 spin_lock_bh(&g_polling_lock
);
534 /* This happens if we remove the current timer but can't stop
535 * the timer from firing. In this case just get out.
537 oz_event_log(OZ_EVT_TIMER
, 0, 0, NULL
, 0);
538 spin_unlock_bh(&g_polling_lock
);
541 g_timer_state
= OZ_TIMER_IN_HANDLER
;
545 spin_unlock_bh(&g_polling_lock
);
548 oz_event_log(OZ_EVT_TIMER
, 0, t
->type
, NULL
, 0);
549 oz_pd_handle_timer(pd
, t
->type
);
550 spin_lock_bh(&g_polling_lock
);
551 if (g_timer_pool_count
< OZ_MAX_TIMER_POOL_SIZE
) {
552 t
->link
.next
= g_timer_pool
;
553 g_timer_pool
= &t
->link
;
554 g_timer_pool_count
++;
557 if (!list_empty(&g_timer_list
)) {
558 t2
= container_of(g_timer_list
.next
,
559 struct oz_timer
, link
);
560 if (time_before_eq(t2
->due_time
, jiffies
))
567 spin_unlock_bh(&g_polling_lock
);
572 g_timer_state
= OZ_TIMER_IDLE
;
573 oz_protocol_timer_start();
575 /*------------------------------------------------------------------------------
578 static void oz_protocol_timer_start(void)
580 spin_lock_bh(&g_polling_lock
);
581 if (!list_empty(&g_timer_list
)) {
583 container_of(g_timer_list
.next
, struct oz_timer
, link
);
584 if (g_timer_state
== OZ_TIMER_SET
) {
585 oz_event_log(OZ_EVT_TIMER_CTRL
, 3,
586 (u16
)g_cur_timer
->type
, NULL
,
587 (unsigned)g_cur_timer
->due_time
);
588 mod_timer(&g_timer
, g_cur_timer
->due_time
);
590 oz_event_log(OZ_EVT_TIMER_CTRL
, 4,
591 (u16
)g_cur_timer
->type
, NULL
,
592 (unsigned)g_cur_timer
->due_time
);
593 g_timer
.expires
= g_cur_timer
->due_time
;
594 g_timer
.function
= oz_protocol_timer
;
598 g_timer_state
= OZ_TIMER_SET
;
600 oz_trace("No queued timers\n");
602 spin_unlock_bh(&g_polling_lock
);
604 /*------------------------------------------------------------------------------
605 * Context: softirq or process
607 void oz_timer_add(struct oz_pd
*pd
, int type
, unsigned long due_time
,
611 struct oz_timer
*t
= NULL
;
612 int restart_needed
= 0;
613 oz_event_log(OZ_EVT_TIMER_CTRL
, 1, (u16
)type
, NULL
, (unsigned)due_time
);
614 spin_lock(&g_polling_lock
);
616 list_for_each(e
, &g_timer_list
) {
617 t
= container_of(e
, struct oz_timer
, link
);
618 if ((t
->pd
== pd
) && (t
->type
== type
)) {
619 if (g_cur_timer
== t
) {
631 t
= container_of(g_timer_pool
, struct oz_timer
, link
);
632 g_timer_pool
= g_timer_pool
->next
;
633 g_timer_pool_count
--;
635 t
= kmalloc(sizeof(struct oz_timer
), GFP_ATOMIC
);
645 t
->due_time
= due_time
;
646 list_for_each(e
, &g_timer_list
) {
647 t2
= container_of(e
, struct oz_timer
, link
);
648 if (time_before(due_time
, t2
->due_time
)) {
649 if (t2
== g_cur_timer
) {
656 list_add_tail(&t
->link
, e
);
658 if (g_timer_state
== OZ_TIMER_IDLE
)
660 else if (g_timer_state
== OZ_TIMER_IN_HANDLER
)
662 spin_unlock(&g_polling_lock
);
664 oz_protocol_timer_start();
666 /*------------------------------------------------------------------------------
667 * Context: softirq or process
669 void oz_timer_delete(struct oz_pd
*pd
, int type
)
671 struct list_head
*chain
= NULL
;
674 int restart_needed
= 0;
676 oz_event_log(OZ_EVT_TIMER_CTRL
, 2, (u16
)type
, NULL
, 0);
677 spin_lock(&g_polling_lock
);
678 list_for_each_entry_safe(t
, n
, &g_timer_list
, link
) {
679 if ((t
->pd
== pd
) && ((type
== 0) || (t
->type
== type
))) {
680 if (g_cur_timer
== t
) {
687 if (g_timer_pool_count
< OZ_MAX_TIMER_POOL_SIZE
) {
688 t
->link
.next
= g_timer_pool
;
689 g_timer_pool
= &t
->link
;
690 g_timer_pool_count
++;
692 t
->link
.next
= chain
;
699 if (g_timer_state
== OZ_TIMER_IN_HANDLER
)
701 else if (restart_needed
)
702 g_timer_state
= OZ_TIMER_IDLE
;
703 spin_unlock(&g_polling_lock
);
705 oz_protocol_timer_start();
709 t
= container_of(chain
, struct oz_timer
, link
);
714 /*------------------------------------------------------------------------------
715 * Context: softirq or process
717 void oz_pd_request_heartbeat(struct oz_pd
*pd
)
719 unsigned long now
= jiffies
;
721 spin_lock(&g_polling_lock
);
722 if (pd
->heartbeat_requested
) {
723 spin_unlock(&g_polling_lock
);
726 if (pd
->pulse_period_j
)
727 t
= ((now
/ pd
->pulse_period_j
) + 1) * pd
->pulse_period_j
;
730 pd
->heartbeat_requested
= 1;
731 spin_unlock(&g_polling_lock
);
732 oz_timer_add(pd
, OZ_TIMER_HEARTBEAT
, t
, 0);
734 /*------------------------------------------------------------------------------
735 * Context: softirq or process
737 struct oz_pd
*oz_pd_find(const u8
*mac_addr
)
741 spin_lock_bh(&g_polling_lock
);
742 list_for_each(e
, &g_pd_list
) {
743 pd
= container_of(e
, struct oz_pd
, link
);
744 if (memcmp(pd
->mac_addr
, mac_addr
, ETH_ALEN
) == 0) {
745 atomic_inc(&pd
->ref_count
);
746 spin_unlock_bh(&g_polling_lock
);
750 spin_unlock_bh(&g_polling_lock
);
753 /*------------------------------------------------------------------------------
756 void oz_app_enable(int app_id
, int enable
)
758 if (app_id
<= OZ_APPID_MAX
) {
759 spin_lock_bh(&g_polling_lock
);
761 g_apps
|= (1<<app_id
);
763 g_apps
&= ~(1<<app_id
);
764 spin_unlock_bh(&g_polling_lock
);
767 /*------------------------------------------------------------------------------
770 static int oz_pkt_recv(struct sk_buff
*skb
, struct net_device
*dev
,
771 struct packet_type
*pt
, struct net_device
*orig_dev
)
773 oz_event_log(OZ_EVT_RX_FRAME
, 0, 0, NULL
, 0);
774 skb
= skb_share_check(skb
, GFP_ATOMIC
);
777 spin_lock_bh(&g_rx_queue
.lock
);
778 if (g_processing_rx
) {
779 /* We already hold the lock so use __ variant.
781 __skb_queue_head(&g_rx_queue
, skb
);
782 spin_unlock_bh(&g_rx_queue
.lock
);
787 spin_unlock_bh(&g_rx_queue
.lock
);
789 spin_lock_bh(&g_rx_queue
.lock
);
790 if (skb_queue_empty(&g_rx_queue
)) {
792 spin_unlock_bh(&g_rx_queue
.lock
);
795 /* We already hold the lock so use __ variant.
797 skb
= __skb_dequeue(&g_rx_queue
);
802 /*------------------------------------------------------------------------------
805 void oz_binding_add(char *net_dev
)
807 struct oz_binding
*binding
;
809 binding
= kmalloc(sizeof(struct oz_binding
), GFP_KERNEL
);
811 binding
->ptype
.type
= __constant_htons(OZ_ETHERTYPE
);
812 binding
->ptype
.func
= oz_pkt_recv
;
813 memcpy(binding
->name
, net_dev
, OZ_MAX_BINDING_LEN
);
814 if (net_dev
&& *net_dev
) {
815 oz_trace("Adding binding: %s\n", net_dev
);
817 dev_get_by_name(&init_net
, net_dev
);
818 if (binding
->ptype
.dev
== NULL
) {
819 oz_trace("Netdev %s not found\n", net_dev
);
824 oz_trace("Binding to all netcards\n");
825 binding
->ptype
.dev
= NULL
;
828 dev_add_pack(&binding
->ptype
);
829 spin_lock_bh(&g_binding_lock
);
830 binding
->next
= g_binding
;
832 spin_unlock_bh(&g_binding_lock
);
836 /*------------------------------------------------------------------------------
839 static int compare_binding_name(char *s1
, char *s2
)
842 for (i
= 0; i
< OZ_MAX_BINDING_LEN
; i
++) {
851 /*------------------------------------------------------------------------------
854 static void pd_stop_all_for_device(struct net_device
*net_dev
)
860 spin_lock_bh(&g_polling_lock
);
861 list_for_each_entry_safe(pd
, n
, &g_pd_list
, link
) {
862 if (pd
->net_dev
== net_dev
) {
863 list_move(&pd
->link
, &h
);
867 spin_unlock_bh(&g_polling_lock
);
868 while (!list_empty(&h
)) {
869 pd
= list_first_entry(&h
, struct oz_pd
, link
);
874 /*------------------------------------------------------------------------------
877 void oz_binding_remove(char *net_dev
)
879 struct oz_binding
*binding
;
880 struct oz_binding
**link
;
881 oz_trace("Removing binding: %s\n", net_dev
);
882 spin_lock_bh(&g_binding_lock
);
886 if (compare_binding_name(binding
->name
, net_dev
)) {
887 oz_trace("Binding '%s' found\n", net_dev
);
888 *link
= binding
->next
;
892 binding
= binding
->next
;
895 spin_unlock_bh(&g_binding_lock
);
897 dev_remove_pack(&binding
->ptype
);
898 if (binding
->ptype
.dev
) {
899 dev_put(binding
->ptype
.dev
);
900 pd_stop_all_for_device(binding
->ptype
.dev
);
905 /*------------------------------------------------------------------------------
908 static char *oz_get_next_device_name(char *s
, char *dname
, int max_size
)
912 while (*s
&& (*s
!= ',') && max_size
> 1) {
919 /*------------------------------------------------------------------------------
922 int oz_protocol_init(char *devs
)
924 skb_queue_head_init(&g_rx_queue
);
925 if (devs
&& (devs
[0] == '*')) {
926 oz_binding_add(NULL
);
930 devs
= oz_get_next_device_name(devs
, d
, sizeof(d
));
935 init_timer(&g_timer
);
938 /*------------------------------------------------------------------------------
941 int oz_get_pd_list(struct oz_mac_addr
*addr
, int max_count
)
946 spin_lock_bh(&g_polling_lock
);
947 list_for_each(e
, &g_pd_list
) {
948 if (count
>= max_count
)
950 pd
= container_of(e
, struct oz_pd
, link
);
951 memcpy(&addr
[count
++], pd
->mac_addr
, ETH_ALEN
);
953 spin_unlock_bh(&g_polling_lock
);
956 /*------------------------------------------------------------------------------
958 void oz_polling_lock_bh(void)
960 spin_lock_bh(&g_polling_lock
);
962 /*------------------------------------------------------------------------------
964 void oz_polling_unlock_bh(void)
966 spin_unlock_bh(&g_polling_lock
);