Linux 2.6.31.6
[linux/fpc-iii.git] / net / mac80211 / mesh.c
blob11cf45bce38a36cf17c4592b5b0991a489a28c5e
1 /*
2 * Copyright (c) 2008 open80211s Ltd.
3 * Authors: Luis Carlos Cobo <luisca@cozybit.com>
4 * Javier Cardona <javier@cozybit.com>
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License version 2 as
8 * published by the Free Software Foundation.
9 */
11 #include <asm/unaligned.h>
12 #include "ieee80211_i.h"
13 #include "mesh.h"
15 #define IEEE80211_MESH_PEER_INACTIVITY_LIMIT (1800 * HZ)
16 #define IEEE80211_MESH_HOUSEKEEPING_INTERVAL (60 * HZ)
18 #define PP_OFFSET 1 /* Path Selection Protocol */
19 #define PM_OFFSET 5 /* Path Selection Metric */
20 #define CC_OFFSET 9 /* Congestion Control Mode */
21 #define CAPAB_OFFSET 17
22 #define ACCEPT_PLINKS 0x80
24 #define TMR_RUNNING_HK 0
25 #define TMR_RUNNING_MP 1
27 int mesh_allocated;
28 static struct kmem_cache *rm_cache;
30 void ieee80211s_init(void)
32 mesh_pathtbl_init();
33 mesh_allocated = 1;
34 rm_cache = kmem_cache_create("mesh_rmc", sizeof(struct rmc_entry),
35 0, 0, NULL);
38 void ieee80211s_stop(void)
40 mesh_pathtbl_unregister();
41 kmem_cache_destroy(rm_cache);
44 static void ieee80211_mesh_housekeeping_timer(unsigned long data)
46 struct ieee80211_sub_if_data *sdata = (void *) data;
47 struct ieee80211_local *local = sdata->local;
48 struct ieee80211_if_mesh *ifmsh = &sdata->u.mesh;
50 ifmsh->housekeeping = true;
52 if (local->quiescing) {
53 set_bit(TMR_RUNNING_HK, &ifmsh->timers_running);
54 return;
57 queue_work(local->hw.workqueue, &ifmsh->work);
60 /**
61 * mesh_matches_local - check if the config of a mesh point matches ours
63 * @ie: information elements of a management frame from the mesh peer
64 * @sdata: local mesh subif
66 * This function checks if the mesh configuration of a mesh point matches the
67 * local mesh configuration, i.e. if both nodes belong to the same mesh network.
69 bool mesh_matches_local(struct ieee802_11_elems *ie, struct ieee80211_sub_if_data *sdata)
71 struct ieee80211_if_mesh *ifmsh = &sdata->u.mesh;
74 * As support for each feature is added, check for matching
75 * - On mesh config capabilities
76 * - Power Save Support En
77 * - Sync support enabled
78 * - Sync support active
79 * - Sync support required from peer
80 * - MDA enabled
81 * - Power management control on fc
83 if (ifmsh->mesh_id_len == ie->mesh_id_len &&
84 memcmp(ifmsh->mesh_id, ie->mesh_id, ie->mesh_id_len) == 0 &&
85 memcmp(ifmsh->mesh_pp_id, ie->mesh_config + PP_OFFSET, 4) == 0 &&
86 memcmp(ifmsh->mesh_pm_id, ie->mesh_config + PM_OFFSET, 4) == 0 &&
87 memcmp(ifmsh->mesh_cc_id, ie->mesh_config + CC_OFFSET, 4) == 0)
88 return true;
90 return false;
93 /**
94 * mesh_peer_accepts_plinks - check if an mp is willing to establish peer links
96 * @ie: information elements of a management frame from the mesh peer
98 bool mesh_peer_accepts_plinks(struct ieee802_11_elems *ie)
100 return (*(ie->mesh_config + CAPAB_OFFSET) & ACCEPT_PLINKS) != 0;
104 * mesh_accept_plinks_update: update accepting_plink in local mesh beacons
106 * @sdata: mesh interface in which mesh beacons are going to be updated
108 void mesh_accept_plinks_update(struct ieee80211_sub_if_data *sdata)
110 bool free_plinks;
112 /* In case mesh_plink_free_count > 0 and mesh_plinktbl_capacity == 0,
113 * the mesh interface might be able to establish plinks with peers that
114 * are already on the table but are not on PLINK_ESTAB state. However,
115 * in general the mesh interface is not accepting peer link requests
116 * from new peers, and that must be reflected in the beacon
118 free_plinks = mesh_plink_availables(sdata);
120 if (free_plinks != sdata->u.mesh.accepting_plinks)
121 ieee80211_mesh_housekeeping_timer((unsigned long) sdata);
124 void mesh_ids_set_default(struct ieee80211_if_mesh *sta)
126 u8 def_id[4] = {0x00, 0x0F, 0xAC, 0xff};
128 memcpy(sta->mesh_pp_id, def_id, 4);
129 memcpy(sta->mesh_pm_id, def_id, 4);
130 memcpy(sta->mesh_cc_id, def_id, 4);
133 int mesh_rmc_init(struct ieee80211_sub_if_data *sdata)
135 int i;
137 sdata->u.mesh.rmc = kmalloc(sizeof(struct mesh_rmc), GFP_KERNEL);
138 if (!sdata->u.mesh.rmc)
139 return -ENOMEM;
140 sdata->u.mesh.rmc->idx_mask = RMC_BUCKETS - 1;
141 for (i = 0; i < RMC_BUCKETS; i++)
142 INIT_LIST_HEAD(&sdata->u.mesh.rmc->bucket[i].list);
143 return 0;
146 void mesh_rmc_free(struct ieee80211_sub_if_data *sdata)
148 struct mesh_rmc *rmc = sdata->u.mesh.rmc;
149 struct rmc_entry *p, *n;
150 int i;
152 if (!sdata->u.mesh.rmc)
153 return;
155 for (i = 0; i < RMC_BUCKETS; i++)
156 list_for_each_entry_safe(p, n, &rmc->bucket[i].list, list) {
157 list_del(&p->list);
158 kmem_cache_free(rm_cache, p);
161 kfree(rmc);
162 sdata->u.mesh.rmc = NULL;
166 * mesh_rmc_check - Check frame in recent multicast cache and add if absent.
168 * @sa: source address
169 * @mesh_hdr: mesh_header
171 * Returns: 0 if the frame is not in the cache, nonzero otherwise.
173 * Checks using the source address and the mesh sequence number if we have
174 * received this frame lately. If the frame is not in the cache, it is added to
175 * it.
177 int mesh_rmc_check(u8 *sa, struct ieee80211s_hdr *mesh_hdr,
178 struct ieee80211_sub_if_data *sdata)
180 struct mesh_rmc *rmc = sdata->u.mesh.rmc;
181 u32 seqnum = 0;
182 int entries = 0;
183 u8 idx;
184 struct rmc_entry *p, *n;
186 /* Don't care about endianness since only match matters */
187 memcpy(&seqnum, &mesh_hdr->seqnum, sizeof(mesh_hdr->seqnum));
188 idx = le32_to_cpu(mesh_hdr->seqnum) & rmc->idx_mask;
189 list_for_each_entry_safe(p, n, &rmc->bucket[idx].list, list) {
190 ++entries;
191 if (time_after(jiffies, p->exp_time) ||
192 (entries == RMC_QUEUE_MAX_LEN)) {
193 list_del(&p->list);
194 kmem_cache_free(rm_cache, p);
195 --entries;
196 } else if ((seqnum == p->seqnum)
197 && (memcmp(sa, p->sa, ETH_ALEN) == 0))
198 return -1;
201 p = kmem_cache_alloc(rm_cache, GFP_ATOMIC);
202 if (!p) {
203 printk(KERN_DEBUG "o11s: could not allocate RMC entry\n");
204 return 0;
206 p->seqnum = seqnum;
207 p->exp_time = jiffies + RMC_TIMEOUT;
208 memcpy(p->sa, sa, ETH_ALEN);
209 list_add(&p->list, &rmc->bucket[idx].list);
210 return 0;
213 void mesh_mgmt_ies_add(struct sk_buff *skb, struct ieee80211_sub_if_data *sdata)
215 struct ieee80211_local *local = sdata->local;
216 struct ieee80211_supported_band *sband;
217 u8 *pos;
218 int len, i, rate;
220 sband = local->hw.wiphy->bands[local->hw.conf.channel->band];
221 len = sband->n_bitrates;
222 if (len > 8)
223 len = 8;
224 pos = skb_put(skb, len + 2);
225 *pos++ = WLAN_EID_SUPP_RATES;
226 *pos++ = len;
227 for (i = 0; i < len; i++) {
228 rate = sband->bitrates[i].bitrate;
229 *pos++ = (u8) (rate / 5);
232 if (sband->n_bitrates > len) {
233 pos = skb_put(skb, sband->n_bitrates - len + 2);
234 *pos++ = WLAN_EID_EXT_SUPP_RATES;
235 *pos++ = sband->n_bitrates - len;
236 for (i = len; i < sband->n_bitrates; i++) {
237 rate = sband->bitrates[i].bitrate;
238 *pos++ = (u8) (rate / 5);
242 pos = skb_put(skb, 2 + sdata->u.mesh.mesh_id_len);
243 *pos++ = WLAN_EID_MESH_ID;
244 *pos++ = sdata->u.mesh.mesh_id_len;
245 if (sdata->u.mesh.mesh_id_len)
246 memcpy(pos, sdata->u.mesh.mesh_id, sdata->u.mesh.mesh_id_len);
248 pos = skb_put(skb, 21);
249 *pos++ = WLAN_EID_MESH_CONFIG;
250 *pos++ = IEEE80211_MESH_CONFIG_LEN;
251 /* Version */
252 *pos++ = 1;
254 /* Active path selection protocol ID */
255 memcpy(pos, sdata->u.mesh.mesh_pp_id, 4);
256 pos += 4;
258 /* Active path selection metric ID */
259 memcpy(pos, sdata->u.mesh.mesh_pm_id, 4);
260 pos += 4;
262 /* Congestion control mode identifier */
263 memcpy(pos, sdata->u.mesh.mesh_cc_id, 4);
264 pos += 4;
266 /* Channel precedence:
267 * Not running simple channel unification protocol
269 memset(pos, 0x00, 4);
270 pos += 4;
272 /* Mesh capability */
273 sdata->u.mesh.accepting_plinks = mesh_plink_availables(sdata);
274 *pos++ = sdata->u.mesh.accepting_plinks ? ACCEPT_PLINKS : 0x00;
275 *pos++ = 0x00;
277 return;
280 u32 mesh_table_hash(u8 *addr, struct ieee80211_sub_if_data *sdata, struct mesh_table *tbl)
282 /* Use last four bytes of hw addr and interface index as hash index */
283 return jhash_2words(*(u32 *)(addr+2), sdata->dev->ifindex, tbl->hash_rnd)
284 & tbl->hash_mask;
287 struct mesh_table *mesh_table_alloc(int size_order)
289 int i;
290 struct mesh_table *newtbl;
292 newtbl = kmalloc(sizeof(struct mesh_table), GFP_KERNEL);
293 if (!newtbl)
294 return NULL;
296 newtbl->hash_buckets = kzalloc(sizeof(struct hlist_head) *
297 (1 << size_order), GFP_KERNEL);
299 if (!newtbl->hash_buckets) {
300 kfree(newtbl);
301 return NULL;
304 newtbl->hashwlock = kmalloc(sizeof(spinlock_t) *
305 (1 << size_order), GFP_KERNEL);
306 if (!newtbl->hashwlock) {
307 kfree(newtbl->hash_buckets);
308 kfree(newtbl);
309 return NULL;
312 newtbl->size_order = size_order;
313 newtbl->hash_mask = (1 << size_order) - 1;
314 atomic_set(&newtbl->entries, 0);
315 get_random_bytes(&newtbl->hash_rnd,
316 sizeof(newtbl->hash_rnd));
317 for (i = 0; i <= newtbl->hash_mask; i++)
318 spin_lock_init(&newtbl->hashwlock[i]);
320 return newtbl;
323 static void __mesh_table_free(struct mesh_table *tbl)
325 kfree(tbl->hash_buckets);
326 kfree(tbl->hashwlock);
327 kfree(tbl);
330 void mesh_table_free(struct mesh_table *tbl, bool free_leafs)
332 struct hlist_head *mesh_hash;
333 struct hlist_node *p, *q;
334 int i;
336 mesh_hash = tbl->hash_buckets;
337 for (i = 0; i <= tbl->hash_mask; i++) {
338 spin_lock(&tbl->hashwlock[i]);
339 hlist_for_each_safe(p, q, &mesh_hash[i]) {
340 tbl->free_node(p, free_leafs);
341 atomic_dec(&tbl->entries);
343 spin_unlock(&tbl->hashwlock[i]);
345 __mesh_table_free(tbl);
348 static void ieee80211_mesh_path_timer(unsigned long data)
350 struct ieee80211_sub_if_data *sdata =
351 (struct ieee80211_sub_if_data *) data;
352 struct ieee80211_if_mesh *ifmsh = &sdata->u.mesh;
353 struct ieee80211_local *local = sdata->local;
355 if (local->quiescing) {
356 set_bit(TMR_RUNNING_MP, &ifmsh->timers_running);
357 return;
360 queue_work(local->hw.workqueue, &ifmsh->work);
363 struct mesh_table *mesh_table_grow(struct mesh_table *tbl)
365 struct mesh_table *newtbl;
366 struct hlist_head *oldhash;
367 struct hlist_node *p, *q;
368 int i;
370 if (atomic_read(&tbl->entries)
371 < tbl->mean_chain_len * (tbl->hash_mask + 1))
372 goto endgrow;
374 newtbl = mesh_table_alloc(tbl->size_order + 1);
375 if (!newtbl)
376 goto endgrow;
378 newtbl->free_node = tbl->free_node;
379 newtbl->mean_chain_len = tbl->mean_chain_len;
380 newtbl->copy_node = tbl->copy_node;
381 atomic_set(&newtbl->entries, atomic_read(&tbl->entries));
383 oldhash = tbl->hash_buckets;
384 for (i = 0; i <= tbl->hash_mask; i++)
385 hlist_for_each(p, &oldhash[i])
386 if (tbl->copy_node(p, newtbl) < 0)
387 goto errcopy;
389 return newtbl;
391 errcopy:
392 for (i = 0; i <= newtbl->hash_mask; i++) {
393 hlist_for_each_safe(p, q, &newtbl->hash_buckets[i])
394 tbl->free_node(p, 0);
396 __mesh_table_free(newtbl);
397 endgrow:
398 return NULL;
402 * ieee80211_new_mesh_header - create a new mesh header
403 * @meshhdr: uninitialized mesh header
404 * @sdata: mesh interface to be used
406 * Return the header length.
408 int ieee80211_new_mesh_header(struct ieee80211s_hdr *meshhdr,
409 struct ieee80211_sub_if_data *sdata)
411 meshhdr->flags = 0;
412 meshhdr->ttl = sdata->u.mesh.mshcfg.dot11MeshTTL;
413 put_unaligned(cpu_to_le32(sdata->u.mesh.mesh_seqnum), &meshhdr->seqnum);
414 sdata->u.mesh.mesh_seqnum++;
416 return 6;
419 static void ieee80211_mesh_housekeeping(struct ieee80211_sub_if_data *sdata,
420 struct ieee80211_if_mesh *ifmsh)
422 bool free_plinks;
424 #ifdef CONFIG_MAC80211_VERBOSE_DEBUG
425 printk(KERN_DEBUG "%s: running mesh housekeeping\n",
426 sdata->dev->name);
427 #endif
429 ieee80211_sta_expire(sdata, IEEE80211_MESH_PEER_INACTIVITY_LIMIT);
430 mesh_path_expire(sdata);
432 free_plinks = mesh_plink_availables(sdata);
433 if (free_plinks != sdata->u.mesh.accepting_plinks)
434 ieee80211_bss_info_change_notify(sdata, BSS_CHANGED_BEACON);
436 ifmsh->housekeeping = false;
437 mod_timer(&ifmsh->housekeeping_timer,
438 round_jiffies(jiffies + IEEE80211_MESH_HOUSEKEEPING_INTERVAL));
441 #ifdef CONFIG_PM
442 void ieee80211_mesh_quiesce(struct ieee80211_sub_if_data *sdata)
444 struct ieee80211_if_mesh *ifmsh = &sdata->u.mesh;
446 /* might restart the timer but that doesn't matter */
447 cancel_work_sync(&ifmsh->work);
449 /* use atomic bitops in case both timers fire at the same time */
451 if (del_timer_sync(&ifmsh->housekeeping_timer))
452 set_bit(TMR_RUNNING_HK, &ifmsh->timers_running);
453 if (del_timer_sync(&ifmsh->mesh_path_timer))
454 set_bit(TMR_RUNNING_MP, &ifmsh->timers_running);
457 void ieee80211_mesh_restart(struct ieee80211_sub_if_data *sdata)
459 struct ieee80211_if_mesh *ifmsh = &sdata->u.mesh;
461 if (test_and_clear_bit(TMR_RUNNING_HK, &ifmsh->timers_running))
462 add_timer(&ifmsh->housekeeping_timer);
463 if (test_and_clear_bit(TMR_RUNNING_MP, &ifmsh->timers_running))
464 add_timer(&ifmsh->mesh_path_timer);
466 #endif
468 void ieee80211_start_mesh(struct ieee80211_sub_if_data *sdata)
470 struct ieee80211_if_mesh *ifmsh = &sdata->u.mesh;
471 struct ieee80211_local *local = sdata->local;
473 ifmsh->housekeeping = true;
474 queue_work(local->hw.workqueue, &ifmsh->work);
475 ieee80211_bss_info_change_notify(sdata, BSS_CHANGED_BEACON |
476 BSS_CHANGED_BEACON_ENABLED);
479 void ieee80211_stop_mesh(struct ieee80211_sub_if_data *sdata)
481 del_timer_sync(&sdata->u.mesh.housekeeping_timer);
483 * If the timer fired while we waited for it, it will have
484 * requeued the work. Now the work will be running again
485 * but will not rearm the timer again because it checks
486 * whether the interface is running, which, at this point,
487 * it no longer is.
489 cancel_work_sync(&sdata->u.mesh.work);
492 * When we get here, the interface is marked down.
493 * Call synchronize_rcu() to wait for the RX path
494 * should it be using the interface and enqueuing
495 * frames at this very time on another CPU.
497 rcu_barrier(); /* Wait for RX path and call_rcu()'s */
498 skb_queue_purge(&sdata->u.mesh.skb_queue);
501 static void ieee80211_mesh_rx_bcn_presp(struct ieee80211_sub_if_data *sdata,
502 u16 stype,
503 struct ieee80211_mgmt *mgmt,
504 size_t len,
505 struct ieee80211_rx_status *rx_status)
507 struct ieee80211_local *local = sdata->local;
508 struct ieee802_11_elems elems;
509 struct ieee80211_channel *channel;
510 u32 supp_rates = 0;
511 size_t baselen;
512 int freq;
513 enum ieee80211_band band = rx_status->band;
515 /* ignore ProbeResp to foreign address */
516 if (stype == IEEE80211_STYPE_PROBE_RESP &&
517 compare_ether_addr(mgmt->da, sdata->dev->dev_addr))
518 return;
520 baselen = (u8 *) mgmt->u.probe_resp.variable - (u8 *) mgmt;
521 if (baselen > len)
522 return;
524 ieee802_11_parse_elems(mgmt->u.probe_resp.variable, len - baselen,
525 &elems);
527 if (elems.ds_params && elems.ds_params_len == 1)
528 freq = ieee80211_channel_to_frequency(elems.ds_params[0]);
529 else
530 freq = rx_status->freq;
532 channel = ieee80211_get_channel(local->hw.wiphy, freq);
534 if (!channel || channel->flags & IEEE80211_CHAN_DISABLED)
535 return;
537 if (elems.mesh_id && elems.mesh_config &&
538 mesh_matches_local(&elems, sdata)) {
539 supp_rates = ieee80211_sta_get_rates(local, &elems, band);
541 mesh_neighbour_update(mgmt->sa, supp_rates, sdata,
542 mesh_peer_accepts_plinks(&elems));
546 static void ieee80211_mesh_rx_mgmt_action(struct ieee80211_sub_if_data *sdata,
547 struct ieee80211_mgmt *mgmt,
548 size_t len,
549 struct ieee80211_rx_status *rx_status)
551 switch (mgmt->u.action.category) {
552 case PLINK_CATEGORY:
553 mesh_rx_plink_frame(sdata, mgmt, len, rx_status);
554 break;
555 case MESH_PATH_SEL_CATEGORY:
556 mesh_rx_path_sel_frame(sdata, mgmt, len);
557 break;
561 static void ieee80211_mesh_rx_queued_mgmt(struct ieee80211_sub_if_data *sdata,
562 struct sk_buff *skb)
564 struct ieee80211_rx_status *rx_status;
565 struct ieee80211_if_mesh *ifmsh;
566 struct ieee80211_mgmt *mgmt;
567 u16 stype;
569 ifmsh = &sdata->u.mesh;
571 rx_status = (struct ieee80211_rx_status *) skb->cb;
572 mgmt = (struct ieee80211_mgmt *) skb->data;
573 stype = le16_to_cpu(mgmt->frame_control) & IEEE80211_FCTL_STYPE;
575 switch (stype) {
576 case IEEE80211_STYPE_PROBE_RESP:
577 case IEEE80211_STYPE_BEACON:
578 ieee80211_mesh_rx_bcn_presp(sdata, stype, mgmt, skb->len,
579 rx_status);
580 break;
581 case IEEE80211_STYPE_ACTION:
582 ieee80211_mesh_rx_mgmt_action(sdata, mgmt, skb->len, rx_status);
583 break;
586 kfree_skb(skb);
589 static void ieee80211_mesh_work(struct work_struct *work)
591 struct ieee80211_sub_if_data *sdata =
592 container_of(work, struct ieee80211_sub_if_data, u.mesh.work);
593 struct ieee80211_local *local = sdata->local;
594 struct ieee80211_if_mesh *ifmsh = &sdata->u.mesh;
595 struct sk_buff *skb;
597 if (!netif_running(sdata->dev))
598 return;
600 if (local->sw_scanning || local->hw_scanning)
601 return;
603 while ((skb = skb_dequeue(&ifmsh->skb_queue)))
604 ieee80211_mesh_rx_queued_mgmt(sdata, skb);
606 if (ifmsh->preq_queue_len &&
607 time_after(jiffies,
608 ifmsh->last_preq + msecs_to_jiffies(ifmsh->mshcfg.dot11MeshHWMPpreqMinInterval)))
609 mesh_path_start_discovery(sdata);
611 if (ifmsh->housekeeping)
612 ieee80211_mesh_housekeeping(sdata, ifmsh);
615 void ieee80211_mesh_notify_scan_completed(struct ieee80211_local *local)
617 struct ieee80211_sub_if_data *sdata;
619 rcu_read_lock();
620 list_for_each_entry_rcu(sdata, &local->interfaces, list)
621 if (ieee80211_vif_is_mesh(&sdata->vif))
622 queue_work(local->hw.workqueue, &sdata->u.mesh.work);
623 rcu_read_unlock();
626 void ieee80211_mesh_init_sdata(struct ieee80211_sub_if_data *sdata)
628 struct ieee80211_if_mesh *ifmsh = &sdata->u.mesh;
630 INIT_WORK(&ifmsh->work, ieee80211_mesh_work);
631 setup_timer(&ifmsh->housekeeping_timer,
632 ieee80211_mesh_housekeeping_timer,
633 (unsigned long) sdata);
634 skb_queue_head_init(&sdata->u.mesh.skb_queue);
636 ifmsh->mshcfg.dot11MeshRetryTimeout = MESH_RET_T;
637 ifmsh->mshcfg.dot11MeshConfirmTimeout = MESH_CONF_T;
638 ifmsh->mshcfg.dot11MeshHoldingTimeout = MESH_HOLD_T;
639 ifmsh->mshcfg.dot11MeshMaxRetries = MESH_MAX_RETR;
640 ifmsh->mshcfg.dot11MeshTTL = MESH_TTL;
641 ifmsh->mshcfg.auto_open_plinks = true;
642 ifmsh->mshcfg.dot11MeshMaxPeerLinks =
643 MESH_MAX_ESTAB_PLINKS;
644 ifmsh->mshcfg.dot11MeshHWMPactivePathTimeout =
645 MESH_PATH_TIMEOUT;
646 ifmsh->mshcfg.dot11MeshHWMPpreqMinInterval =
647 MESH_PREQ_MIN_INT;
648 ifmsh->mshcfg.dot11MeshHWMPnetDiameterTraversalTime =
649 MESH_DIAM_TRAVERSAL_TIME;
650 ifmsh->mshcfg.dot11MeshHWMPmaxPREQretries =
651 MESH_MAX_PREQ_RETRIES;
652 ifmsh->mshcfg.path_refresh_time =
653 MESH_PATH_REFRESH_TIME;
654 ifmsh->mshcfg.min_discovery_timeout =
655 MESH_MIN_DISCOVERY_TIMEOUT;
656 ifmsh->accepting_plinks = true;
657 ifmsh->preq_id = 0;
658 ifmsh->dsn = 0;
659 atomic_set(&ifmsh->mpaths, 0);
660 mesh_rmc_init(sdata);
661 ifmsh->last_preq = jiffies;
662 /* Allocate all mesh structures when creating the first mesh interface. */
663 if (!mesh_allocated)
664 ieee80211s_init();
665 mesh_ids_set_default(ifmsh);
666 setup_timer(&ifmsh->mesh_path_timer,
667 ieee80211_mesh_path_timer,
668 (unsigned long) sdata);
669 INIT_LIST_HEAD(&ifmsh->preq_queue.list);
670 spin_lock_init(&ifmsh->mesh_preq_queue_lock);
673 ieee80211_rx_result
674 ieee80211_mesh_rx_mgmt(struct ieee80211_sub_if_data *sdata, struct sk_buff *skb,
675 struct ieee80211_rx_status *rx_status)
677 struct ieee80211_local *local = sdata->local;
678 struct ieee80211_if_mesh *ifmsh = &sdata->u.mesh;
679 struct ieee80211_mgmt *mgmt;
680 u16 fc;
682 if (skb->len < 24)
683 return RX_DROP_MONITOR;
685 mgmt = (struct ieee80211_mgmt *) skb->data;
686 fc = le16_to_cpu(mgmt->frame_control);
688 switch (fc & IEEE80211_FCTL_STYPE) {
689 case IEEE80211_STYPE_PROBE_RESP:
690 case IEEE80211_STYPE_BEACON:
691 case IEEE80211_STYPE_ACTION:
692 memcpy(skb->cb, rx_status, sizeof(*rx_status));
693 skb_queue_tail(&ifmsh->skb_queue, skb);
694 queue_work(local->hw.workqueue, &ifmsh->work);
695 return RX_QUEUED;
698 return RX_CONTINUE;