2 * net/sched/ife.c Inter-FE action based on ForCES WG InterFE LFB
5 * draft-ietf-forces-interfelfb-03
8 * "Distributing Linux Traffic Control Classifier-Action
10 * Authors: Jamal Hadi Salim and Damascene M. Joachimpillai
12 * This program is free software; you can redistribute it and/or
13 * modify it under the terms of the GNU General Public License
14 * as published by the Free Software Foundation; either version
15 * 2 of the License, or (at your option) any later version.
17 * copyright Jamal Hadi Salim (2015)
21 #include <linux/types.h>
22 #include <linux/kernel.h>
23 #include <linux/string.h>
24 #include <linux/errno.h>
25 #include <linux/skbuff.h>
26 #include <linux/rtnetlink.h>
27 #include <linux/module.h>
28 #include <linux/init.h>
29 #include <net/net_namespace.h>
30 #include <net/netlink.h>
31 #include <net/pkt_sched.h>
32 #include <uapi/linux/tc_act/tc_ife.h>
33 #include <net/tc_act/tc_ife.h>
34 #include <linux/etherdevice.h>
36 #define IFE_TAB_MASK 15
38 static int ife_net_id
;
39 static int max_metacnt
= IFE_META_MAX
+ 1;
41 static const struct nla_policy ife_policy
[TCA_IFE_MAX
+ 1] = {
42 [TCA_IFE_PARMS
] = { .len
= sizeof(struct tc_ife
)},
43 [TCA_IFE_DMAC
] = { .len
= ETH_ALEN
},
44 [TCA_IFE_SMAC
] = { .len
= ETH_ALEN
},
45 [TCA_IFE_TYPE
] = { .type
= NLA_U16
},
48 /* Caller takes care of presenting data in network order
50 int ife_tlv_meta_encode(void *skbdata
, u16 attrtype
, u16 dlen
, const void *dval
)
52 u32
*tlv
= (u32
*)(skbdata
);
53 u16 totlen
= nla_total_size(dlen
); /*alignment + hdr */
54 char *dptr
= (char *)tlv
+ NLA_HDRLEN
;
55 u32 htlv
= attrtype
<< 16 | totlen
;
58 memset(dptr
, 0, totlen
- NLA_HDRLEN
);
59 memcpy(dptr
, dval
, dlen
);
63 EXPORT_SYMBOL_GPL(ife_tlv_meta_encode
);
65 int ife_get_meta_u32(struct sk_buff
*skb
, struct tcf_meta_info
*mi
)
68 return nla_put_u32(skb
, mi
->metaid
, *(u32
*)mi
->metaval
);
70 return nla_put(skb
, mi
->metaid
, 0, NULL
);
72 EXPORT_SYMBOL_GPL(ife_get_meta_u32
);
74 int ife_check_meta_u32(u32 metaval
, struct tcf_meta_info
*mi
)
76 if (metaval
|| mi
->metaval
)
77 return 8; /* T+L+V == 2+2+4 */
81 EXPORT_SYMBOL_GPL(ife_check_meta_u32
);
83 int ife_encode_meta_u32(u32 metaval
, void *skbdata
, struct tcf_meta_info
*mi
)
88 edata
= *(u32
*)mi
->metaval
;
92 if (!edata
) /* will not encode */
96 return ife_tlv_meta_encode(skbdata
, mi
->metaid
, 4, &edata
);
98 EXPORT_SYMBOL_GPL(ife_encode_meta_u32
);
100 int ife_get_meta_u16(struct sk_buff
*skb
, struct tcf_meta_info
*mi
)
103 return nla_put_u16(skb
, mi
->metaid
, *(u16
*)mi
->metaval
);
105 return nla_put(skb
, mi
->metaid
, 0, NULL
);
107 EXPORT_SYMBOL_GPL(ife_get_meta_u16
);
109 int ife_alloc_meta_u32(struct tcf_meta_info
*mi
, void *metaval
)
111 mi
->metaval
= kmemdup(metaval
, sizeof(u32
), GFP_KERNEL
);
117 EXPORT_SYMBOL_GPL(ife_alloc_meta_u32
);
119 int ife_alloc_meta_u16(struct tcf_meta_info
*mi
, void *metaval
)
121 mi
->metaval
= kmemdup(metaval
, sizeof(u16
), GFP_KERNEL
);
127 EXPORT_SYMBOL_GPL(ife_alloc_meta_u16
);
129 void ife_release_meta_gen(struct tcf_meta_info
*mi
)
133 EXPORT_SYMBOL_GPL(ife_release_meta_gen
);
135 int ife_validate_meta_u32(void *val
, int len
)
142 EXPORT_SYMBOL_GPL(ife_validate_meta_u32
);
144 int ife_validate_meta_u16(void *val
, int len
)
146 /* length will include padding */
147 if (len
== NLA_ALIGN(2))
152 EXPORT_SYMBOL_GPL(ife_validate_meta_u16
);
154 static LIST_HEAD(ifeoplist
);
155 static DEFINE_RWLOCK(ife_mod_lock
);
157 static struct tcf_meta_ops
*find_ife_oplist(u16 metaid
)
159 struct tcf_meta_ops
*o
;
161 read_lock(&ife_mod_lock
);
162 list_for_each_entry(o
, &ifeoplist
, list
) {
163 if (o
->metaid
== metaid
) {
164 if (!try_module_get(o
->owner
))
166 read_unlock(&ife_mod_lock
);
170 read_unlock(&ife_mod_lock
);
175 int register_ife_op(struct tcf_meta_ops
*mops
)
177 struct tcf_meta_ops
*m
;
179 if (!mops
->metaid
|| !mops
->metatype
|| !mops
->name
||
180 !mops
->check_presence
|| !mops
->encode
|| !mops
->decode
||
181 !mops
->get
|| !mops
->alloc
)
184 write_lock(&ife_mod_lock
);
186 list_for_each_entry(m
, &ifeoplist
, list
) {
187 if (m
->metaid
== mops
->metaid
||
188 (strcmp(mops
->name
, m
->name
) == 0)) {
189 write_unlock(&ife_mod_lock
);
195 mops
->release
= ife_release_meta_gen
;
197 list_add_tail(&mops
->list
, &ifeoplist
);
198 write_unlock(&ife_mod_lock
);
201 EXPORT_SYMBOL_GPL(unregister_ife_op
);
203 int unregister_ife_op(struct tcf_meta_ops
*mops
)
205 struct tcf_meta_ops
*m
;
208 write_lock(&ife_mod_lock
);
209 list_for_each_entry(m
, &ifeoplist
, list
) {
210 if (m
->metaid
== mops
->metaid
) {
211 list_del(&mops
->list
);
216 write_unlock(&ife_mod_lock
);
220 EXPORT_SYMBOL_GPL(register_ife_op
);
222 static int ife_validate_metatype(struct tcf_meta_ops
*ops
, void *val
, int len
)
225 /* XXX: unfortunately cant use nla_policy at this point
226 * because a length of 0 is valid in the case of
227 * "allow". "use" semantics do enforce for proper
228 * length and i couldve use nla_policy but it makes it hard
229 * to use it just for that..
232 return ops
->validate(val
, len
);
234 if (ops
->metatype
== NLA_U32
)
235 ret
= ife_validate_meta_u32(val
, len
);
236 else if (ops
->metatype
== NLA_U16
)
237 ret
= ife_validate_meta_u16(val
, len
);
242 /* called when adding new meta information
243 * under ife->tcf_lock
245 static int load_metaops_and_vet(struct tcf_ife_info
*ife
, u32 metaid
,
248 struct tcf_meta_ops
*ops
= find_ife_oplist(metaid
);
253 #ifdef CONFIG_MODULES
254 spin_unlock_bh(&ife
->tcf_lock
);
256 request_module("ifemeta%u", metaid
);
258 spin_lock_bh(&ife
->tcf_lock
);
259 ops
= find_ife_oplist(metaid
);
266 ret
= ife_validate_metatype(ops
, val
, len
);
268 module_put(ops
->owner
);
274 /* called when adding new meta information
275 * under ife->tcf_lock
277 static int add_metainfo(struct tcf_ife_info
*ife
, u32 metaid
, void *metaval
,
280 struct tcf_meta_info
*mi
= NULL
;
281 struct tcf_meta_ops
*ops
= find_ife_oplist(metaid
);
287 mi
= kzalloc(sizeof(*mi
), GFP_KERNEL
);
289 /*put back what find_ife_oplist took */
290 module_put(ops
->owner
);
297 ret
= ops
->alloc(mi
, metaval
);
300 module_put(ops
->owner
);
305 list_add_tail(&mi
->metalist
, &ife
->metalist
);
310 static int use_all_metadata(struct tcf_ife_info
*ife
)
312 struct tcf_meta_ops
*o
;
316 list_for_each_entry(o
, &ifeoplist
, list
) {
317 rc
= add_metainfo(ife
, o
->metaid
, NULL
, 0);
328 static int dump_metalist(struct sk_buff
*skb
, struct tcf_ife_info
*ife
)
330 struct tcf_meta_info
*e
;
332 unsigned char *b
= skb_tail_pointer(skb
);
333 int total_encoded
= 0;
335 /*can only happen on decode */
336 if (list_empty(&ife
->metalist
))
339 nest
= nla_nest_start(skb
, TCA_IFE_METALST
);
343 list_for_each_entry(e
, &ife
->metalist
, metalist
) {
344 if (!e
->ops
->get(skb
, e
))
351 nla_nest_end(skb
, nest
);
360 /* under ife->tcf_lock */
361 static void _tcf_ife_cleanup(struct tc_action
*a
, int bind
)
363 struct tcf_ife_info
*ife
= a
->priv
;
364 struct tcf_meta_info
*e
, *n
;
366 list_for_each_entry_safe(e
, n
, &ife
->metalist
, metalist
) {
367 module_put(e
->ops
->owner
);
368 list_del(&e
->metalist
);
379 static void tcf_ife_cleanup(struct tc_action
*a
, int bind
)
381 struct tcf_ife_info
*ife
= a
->priv
;
383 spin_lock_bh(&ife
->tcf_lock
);
384 _tcf_ife_cleanup(a
, bind
);
385 spin_unlock_bh(&ife
->tcf_lock
);
388 /* under ife->tcf_lock */
389 static int populate_metalist(struct tcf_ife_info
*ife
, struct nlattr
**tb
)
396 for (i
= 1; i
< max_metacnt
; i
++) {
398 val
= nla_data(tb
[i
]);
399 len
= nla_len(tb
[i
]);
401 rc
= load_metaops_and_vet(ife
, i
, val
, len
);
405 rc
= add_metainfo(ife
, i
, val
, len
);
414 static int tcf_ife_init(struct net
*net
, struct nlattr
*nla
,
415 struct nlattr
*est
, struct tc_action
*a
,
418 struct tc_action_net
*tn
= net_generic(net
, ife_net_id
);
419 struct nlattr
*tb
[TCA_IFE_MAX
+ 1];
420 struct nlattr
*tb2
[IFE_META_MAX
+ 1];
421 struct tcf_ife_info
*ife
;
426 int ret
= 0, exists
= 0;
429 err
= nla_parse_nested(tb
, TCA_IFE_MAX
, nla
, ife_policy
);
433 if (!tb
[TCA_IFE_PARMS
])
436 parm
= nla_data(tb
[TCA_IFE_PARMS
]);
438 exists
= tcf_hash_check(tn
, parm
->index
, a
, bind
);
442 if (parm
->flags
& IFE_ENCODE
) {
443 /* Until we get issued the ethertype, we cant have
446 if (!tb
[TCA_IFE_TYPE
]) {
448 tcf_hash_release(a
, bind
);
449 pr_info("You MUST pass etherype for encoding\n");
455 ret
= tcf_hash_create(tn
, parm
->index
, est
, a
, sizeof(*ife
),
461 tcf_hash_release(a
, bind
);
467 ife
->flags
= parm
->flags
;
469 if (parm
->flags
& IFE_ENCODE
) {
470 ife_type
= nla_get_u16(tb
[TCA_IFE_TYPE
]);
471 if (tb
[TCA_IFE_DMAC
])
472 daddr
= nla_data(tb
[TCA_IFE_DMAC
]);
473 if (tb
[TCA_IFE_SMAC
])
474 saddr
= nla_data(tb
[TCA_IFE_SMAC
]);
477 spin_lock_bh(&ife
->tcf_lock
);
478 ife
->tcf_action
= parm
->action
;
480 if (parm
->flags
& IFE_ENCODE
) {
482 ether_addr_copy(ife
->eth_dst
, daddr
);
484 eth_zero_addr(ife
->eth_dst
);
487 ether_addr_copy(ife
->eth_src
, saddr
);
489 eth_zero_addr(ife
->eth_src
);
491 ife
->eth_type
= ife_type
;
494 if (ret
== ACT_P_CREATED
)
495 INIT_LIST_HEAD(&ife
->metalist
);
497 if (tb
[TCA_IFE_METALST
]) {
498 err
= nla_parse_nested(tb2
, IFE_META_MAX
, tb
[TCA_IFE_METALST
],
503 tcf_hash_release(a
, bind
);
504 if (ret
== ACT_P_CREATED
)
505 _tcf_ife_cleanup(a
, bind
);
507 spin_unlock_bh(&ife
->tcf_lock
);
511 err
= populate_metalist(ife
, tb2
);
513 goto metadata_parse_err
;
516 /* if no passed metadata allow list or passed allow-all
517 * then here we process by adding as many supported metadatum
518 * as we can. You better have at least one else we are
521 err
= use_all_metadata(ife
);
523 if (ret
== ACT_P_CREATED
)
524 _tcf_ife_cleanup(a
, bind
);
526 spin_unlock_bh(&ife
->tcf_lock
);
531 spin_unlock_bh(&ife
->tcf_lock
);
533 if (ret
== ACT_P_CREATED
)
534 tcf_hash_insert(tn
, a
);
539 static int tcf_ife_dump(struct sk_buff
*skb
, struct tc_action
*a
, int bind
,
542 unsigned char *b
= skb_tail_pointer(skb
);
543 struct tcf_ife_info
*ife
= a
->priv
;
544 struct tc_ife opt
= {
545 .index
= ife
->tcf_index
,
546 .refcnt
= ife
->tcf_refcnt
- ref
,
547 .bindcnt
= ife
->tcf_bindcnt
- bind
,
548 .action
= ife
->tcf_action
,
553 if (nla_put(skb
, TCA_IFE_PARMS
, sizeof(opt
), &opt
))
554 goto nla_put_failure
;
556 t
.install
= jiffies_to_clock_t(jiffies
- ife
->tcf_tm
.install
);
557 t
.lastuse
= jiffies_to_clock_t(jiffies
- ife
->tcf_tm
.lastuse
);
558 t
.expires
= jiffies_to_clock_t(ife
->tcf_tm
.expires
);
559 if (nla_put_64bit(skb
, TCA_IFE_TM
, sizeof(t
), &t
, TCA_IFE_PAD
))
560 goto nla_put_failure
;
562 if (!is_zero_ether_addr(ife
->eth_dst
)) {
563 if (nla_put(skb
, TCA_IFE_DMAC
, ETH_ALEN
, ife
->eth_dst
))
564 goto nla_put_failure
;
567 if (!is_zero_ether_addr(ife
->eth_src
)) {
568 if (nla_put(skb
, TCA_IFE_SMAC
, ETH_ALEN
, ife
->eth_src
))
569 goto nla_put_failure
;
572 if (nla_put(skb
, TCA_IFE_TYPE
, 2, &ife
->eth_type
))
573 goto nla_put_failure
;
575 if (dump_metalist(skb
, ife
)) {
576 /*ignore failure to dump metalist */
577 pr_info("Failed to dump metalist\n");
587 int find_decode_metaid(struct sk_buff
*skb
, struct tcf_ife_info
*ife
,
588 u16 metaid
, u16 mlen
, void *mdata
)
590 struct tcf_meta_info
*e
;
592 /* XXX: use hash to speed up */
593 list_for_each_entry(e
, &ife
->metalist
, metalist
) {
594 if (metaid
== e
->metaid
) {
596 /* We check for decode presence already */
597 return e
->ops
->decode(skb
, mdata
, mlen
);
615 static int tcf_ife_decode(struct sk_buff
*skb
, const struct tc_action
*a
,
616 struct tcf_result
*res
)
618 struct tcf_ife_info
*ife
= a
->priv
;
619 int action
= ife
->tcf_action
;
620 struct ifeheadr
*ifehdr
= (struct ifeheadr
*)skb
->data
;
621 u16 ifehdrln
= ifehdr
->metalen
;
622 struct meta_tlvhdr
*tlv
= (struct meta_tlvhdr
*)(ifehdr
->tlv_data
);
624 spin_lock(&ife
->tcf_lock
);
625 bstats_update(&ife
->tcf_bstats
, skb
);
626 ife
->tcf_tm
.lastuse
= jiffies
;
627 spin_unlock(&ife
->tcf_lock
);
629 ifehdrln
= ntohs(ifehdrln
);
630 if (unlikely(!pskb_may_pull(skb
, ifehdrln
))) {
631 spin_lock(&ife
->tcf_lock
);
632 ife
->tcf_qstats
.drops
++;
633 spin_unlock(&ife
->tcf_lock
);
637 skb_set_mac_header(skb
, ifehdrln
);
638 __skb_pull(skb
, ifehdrln
);
639 skb
->protocol
= eth_type_trans(skb
, skb
->dev
);
640 ifehdrln
-= IFE_METAHDRLEN
;
642 while (ifehdrln
> 0) {
643 u8
*tlvdata
= (u8
*)tlv
;
644 u16 mtype
= tlv
->type
;
647 mtype
= ntohs(mtype
);
650 if (find_decode_metaid(skb
, ife
, mtype
, (mlen
- 4),
651 (void *)(tlvdata
+ 4))) {
652 /* abuse overlimits to count when we receive metadata
653 * but dont have an ops for it
655 pr_info_ratelimited("Unknown metaid %d alnlen %d\n",
657 ife
->tcf_qstats
.overlimits
++;
662 tlv
= (struct meta_tlvhdr
*)tlvdata
;
665 skb_reset_network_header(skb
);
669 /*XXX: check if we can do this at install time instead of current
672 static int ife_get_sz(struct sk_buff
*skb
, struct tcf_ife_info
*ife
)
674 struct tcf_meta_info
*e
, *n
;
675 int tot_run_sz
= 0, run_sz
= 0;
677 list_for_each_entry_safe(e
, n
, &ife
->metalist
, metalist
) {
678 if (e
->ops
->check_presence
) {
679 run_sz
= e
->ops
->check_presence(skb
, e
);
680 tot_run_sz
+= run_sz
;
687 static int tcf_ife_encode(struct sk_buff
*skb
, const struct tc_action
*a
,
688 struct tcf_result
*res
)
690 struct tcf_ife_info
*ife
= a
->priv
;
691 int action
= ife
->tcf_action
;
692 struct ethhdr
*oethh
; /* outer ether header */
693 struct ethhdr
*iethh
; /* inner eth header */
694 struct tcf_meta_info
*e
;
696 OUTERHDR:TOTMETALEN:{TLVHDR:Metadatum:TLVHDR..}:ORIGDATA
697 where ORIGDATA = original ethernet header ...
699 u16 metalen
= ife_get_sz(skb
, ife
);
700 int hdrm
= metalen
+ skb
->dev
->hard_header_len
+ IFE_METAHDRLEN
;
701 unsigned int skboff
= skb
->dev
->hard_header_len
;
702 u32 at
= G_TC_AT(skb
->tc_verd
);
703 int new_len
= skb
->len
+ hdrm
;
704 bool exceed_mtu
= false;
707 if (at
& AT_EGRESS
) {
708 if (new_len
> skb
->dev
->mtu
)
712 spin_lock(&ife
->tcf_lock
);
713 bstats_update(&ife
->tcf_bstats
, skb
);
714 ife
->tcf_tm
.lastuse
= jiffies
;
716 if (!metalen
) { /* no metadata to send */
717 /* abuse overlimits to count when we allow packet
720 ife
->tcf_qstats
.overlimits
++;
721 spin_unlock(&ife
->tcf_lock
);
724 /* could be stupid policy setup or mtu config
725 * so lets be conservative.. */
726 if ((action
== TC_ACT_SHOT
) || exceed_mtu
) {
727 ife
->tcf_qstats
.drops
++;
728 spin_unlock(&ife
->tcf_lock
);
732 iethh
= eth_hdr(skb
);
734 err
= skb_cow_head(skb
, hdrm
);
736 ife
->tcf_qstats
.drops
++;
737 spin_unlock(&ife
->tcf_lock
);
741 if (!(at
& AT_EGRESS
))
742 skb_push(skb
, skb
->dev
->hard_header_len
);
744 __skb_push(skb
, hdrm
);
745 memcpy(skb
->data
, iethh
, skb
->mac_len
);
746 skb_reset_mac_header(skb
);
747 oethh
= eth_hdr(skb
);
749 /*total metadata length */
750 metalen
+= IFE_METAHDRLEN
;
751 metalen
= htons(metalen
);
752 memcpy((skb
->data
+ skboff
), &metalen
, IFE_METAHDRLEN
);
753 skboff
+= IFE_METAHDRLEN
;
755 /* XXX: we dont have a clever way of telling encode to
756 * not repeat some of the computations that are done by
757 * ops->presence_check...
759 list_for_each_entry(e
, &ife
->metalist
, metalist
) {
760 if (e
->ops
->encode
) {
761 err
= e
->ops
->encode(skb
, (void *)(skb
->data
+ skboff
),
765 /* too corrupt to keep around if overwritten */
766 ife
->tcf_qstats
.drops
++;
767 spin_unlock(&ife
->tcf_lock
);
773 if (!is_zero_ether_addr(ife
->eth_src
))
774 ether_addr_copy(oethh
->h_source
, ife
->eth_src
);
776 ether_addr_copy(oethh
->h_source
, iethh
->h_source
);
777 if (!is_zero_ether_addr(ife
->eth_dst
))
778 ether_addr_copy(oethh
->h_dest
, ife
->eth_dst
);
780 ether_addr_copy(oethh
->h_dest
, iethh
->h_dest
);
781 oethh
->h_proto
= htons(ife
->eth_type
);
783 if (!(at
& AT_EGRESS
))
784 skb_pull(skb
, skb
->dev
->hard_header_len
);
786 spin_unlock(&ife
->tcf_lock
);
791 static int tcf_ife_act(struct sk_buff
*skb
, const struct tc_action
*a
,
792 struct tcf_result
*res
)
794 struct tcf_ife_info
*ife
= a
->priv
;
796 if (ife
->flags
& IFE_ENCODE
)
797 return tcf_ife_encode(skb
, a
, res
);
799 if (!(ife
->flags
& IFE_ENCODE
))
800 return tcf_ife_decode(skb
, a
, res
);
802 pr_info_ratelimited("unknown failure(policy neither de/encode\n");
803 spin_lock(&ife
->tcf_lock
);
804 bstats_update(&ife
->tcf_bstats
, skb
);
805 ife
->tcf_tm
.lastuse
= jiffies
;
806 ife
->tcf_qstats
.drops
++;
807 spin_unlock(&ife
->tcf_lock
);
812 static int tcf_ife_walker(struct net
*net
, struct sk_buff
*skb
,
813 struct netlink_callback
*cb
, int type
,
816 struct tc_action_net
*tn
= net_generic(net
, ife_net_id
);
818 return tcf_generic_walker(tn
, skb
, cb
, type
, a
);
821 static int tcf_ife_search(struct net
*net
, struct tc_action
*a
, u32 index
)
823 struct tc_action_net
*tn
= net_generic(net
, ife_net_id
);
825 return tcf_hash_search(tn
, a
, index
);
828 static struct tc_action_ops act_ife_ops
= {
831 .owner
= THIS_MODULE
,
833 .dump
= tcf_ife_dump
,
834 .cleanup
= tcf_ife_cleanup
,
835 .init
= tcf_ife_init
,
836 .walk
= tcf_ife_walker
,
837 .lookup
= tcf_ife_search
,
840 static __net_init
int ife_init_net(struct net
*net
)
842 struct tc_action_net
*tn
= net_generic(net
, ife_net_id
);
844 return tc_action_net_init(tn
, &act_ife_ops
, IFE_TAB_MASK
);
847 static void __net_exit
ife_exit_net(struct net
*net
)
849 struct tc_action_net
*tn
= net_generic(net
, ife_net_id
);
851 tc_action_net_exit(tn
);
854 static struct pernet_operations ife_net_ops
= {
855 .init
= ife_init_net
,
856 .exit
= ife_exit_net
,
858 .size
= sizeof(struct tc_action_net
),
861 static int __init
ife_init_module(void)
863 return tcf_register_action(&act_ife_ops
, &ife_net_ops
);
866 static void __exit
ife_cleanup_module(void)
868 tcf_unregister_action(&act_ife_ops
, &ife_net_ops
);
871 module_init(ife_init_module
);
872 module_exit(ife_cleanup_module
);
874 MODULE_AUTHOR("Jamal Hadi Salim(2015)");
875 MODULE_DESCRIPTION("Inter-FE LFB action");
876 MODULE_LICENSE("GPL");