2 * net/sched/em_meta.c Metadata ematch
4 * This program is free software; you can redistribute it and/or
5 * modify it under the terms of the GNU General Public License
6 * as published by the Free Software Foundation; either version
7 * 2 of the License, or (at your option) any later version.
9 * Authors: Thomas Graf <tgraf@suug.ch>
11 * ==========================================================================
13 * The metadata ematch compares two meta objects where each object
14 * represents either a meta value stored in the kernel or a static
15 * value provided by userspace. The objects are not provided by
16 * userspace itself but rather a definition providing the information
17 * to build them. Every object is of a certain type which must be
18 * equal to the object it is being compared to.
20 * The definition of a objects conists of the type (meta type), a
21 * identifier (meta id) and additional type specific information.
22 * The meta id is either TCF_META_TYPE_VALUE for values provided by
23 * userspace or a index to the meta operations table consisting of
24 * function pointers to type specific meta data collectors returning
25 * the value of the requested meta value.
28 * +-----------+ +-----------+
29 * | type: INT | | type: INT |
30 * def | id: DEV | | id: VALUE |
31 * | data: | | data: 3 |
32 * +-----------+ +-----------+
34 * ---> meta_ops[INT][DEV](...) |
38 * +-----------+ +-----------+
39 * | type: INT | | type: INT |
40 * obj | id: DEV | | id: VALUE |
41 * | data: 2 |<--data got filled out | data: 3 |
42 * +-----------+ +-----------+
44 * --------------> 2 equals 3 <--------------
46 * This is a simplified schema, the complexity varies depending
47 * on the meta type. Obviously, the length of the data must also
48 * be provided for non-numeric types.
50 * Additionally, type dependent modifiers such as shift operators
51 * or mask may be applied to extend the functionaliy. As of now,
52 * the variable length type supports shifting the byte string to
53 * the right, eating up any number of octets and thus supporting
54 * wildcard interface name comparisons such as "ppp%" matching
57 * NOTE: Certain meta values depend on other subsystems and are
58 * only available if that subsystem is enabled in the kernel.
61 #include <linux/slab.h>
62 #include <linux/module.h>
63 #include <linux/types.h>
64 #include <linux/kernel.h>
65 #include <linux/sched.h>
66 #include <linux/sched/loadavg.h>
67 #include <linux/string.h>
68 #include <linux/skbuff.h>
69 #include <linux/random.h>
70 #include <linux/if_vlan.h>
71 #include <linux/tc_ematch/tc_em_meta.h>
73 #include <net/route.h>
74 #include <net/pkt_cls.h>
83 struct tcf_meta_val hdr
;
89 struct meta_value lvalue
;
90 struct meta_value rvalue
;
93 static inline int meta_id(struct meta_value
*v
)
95 return TCF_META_ID(v
->hdr
.kind
);
98 static inline int meta_type(struct meta_value
*v
)
100 return TCF_META_TYPE(v
->hdr
.kind
);
103 #define META_COLLECTOR(FUNC) static void meta_##FUNC(struct sk_buff *skb, \
104 struct tcf_pkt_info *info, struct meta_value *v, \
105 struct meta_obj *dst, int *err)
107 /**************************************************************************
108 * System status & misc
109 **************************************************************************/
111 META_COLLECTOR(int_random
)
113 get_random_bytes(&dst
->value
, sizeof(dst
->value
));
116 static inline unsigned long fixed_loadavg(int load
)
118 int rnd_load
= load
+ (FIXED_1
/200);
119 int rnd_frac
= ((rnd_load
& (FIXED_1
-1)) * 100) >> FSHIFT
;
121 return ((rnd_load
>> FSHIFT
) * 100) + rnd_frac
;
124 META_COLLECTOR(int_loadavg_0
)
126 dst
->value
= fixed_loadavg(avenrun
[0]);
129 META_COLLECTOR(int_loadavg_1
)
131 dst
->value
= fixed_loadavg(avenrun
[1]);
134 META_COLLECTOR(int_loadavg_2
)
136 dst
->value
= fixed_loadavg(avenrun
[2]);
139 /**************************************************************************
140 * Device names & indices
141 **************************************************************************/
143 static inline int int_dev(struct net_device
*dev
, struct meta_obj
*dst
)
145 if (unlikely(dev
== NULL
))
148 dst
->value
= dev
->ifindex
;
152 static inline int var_dev(struct net_device
*dev
, struct meta_obj
*dst
)
154 if (unlikely(dev
== NULL
))
157 dst
->value
= (unsigned long) dev
->name
;
158 dst
->len
= strlen(dev
->name
);
162 META_COLLECTOR(int_dev
)
164 *err
= int_dev(skb
->dev
, dst
);
167 META_COLLECTOR(var_dev
)
169 *err
= var_dev(skb
->dev
, dst
);
172 /**************************************************************************
174 **************************************************************************/
176 META_COLLECTOR(int_vlan_tag
)
180 if (skb_vlan_tag_present(skb
))
181 dst
->value
= skb_vlan_tag_get(skb
);
182 else if (!__vlan_get_tag(skb
, &tag
))
190 /**************************************************************************
192 **************************************************************************/
194 META_COLLECTOR(int_priority
)
196 dst
->value
= skb
->priority
;
199 META_COLLECTOR(int_protocol
)
201 /* Let userspace take care of the byte ordering */
202 dst
->value
= tc_skb_protocol(skb
);
205 META_COLLECTOR(int_pkttype
)
207 dst
->value
= skb
->pkt_type
;
210 META_COLLECTOR(int_pktlen
)
212 dst
->value
= skb
->len
;
215 META_COLLECTOR(int_datalen
)
217 dst
->value
= skb
->data_len
;
220 META_COLLECTOR(int_maclen
)
222 dst
->value
= skb
->mac_len
;
225 META_COLLECTOR(int_rxhash
)
227 dst
->value
= skb_get_hash(skb
);
230 /**************************************************************************
232 **************************************************************************/
234 META_COLLECTOR(int_mark
)
236 dst
->value
= skb
->mark
;
239 /**************************************************************************
241 **************************************************************************/
243 META_COLLECTOR(int_tcindex
)
245 dst
->value
= skb
->tc_index
;
248 /**************************************************************************
250 **************************************************************************/
252 META_COLLECTOR(int_rtclassid
)
254 if (unlikely(skb_dst(skb
) == NULL
))
257 #ifdef CONFIG_IP_ROUTE_CLASSID
258 dst
->value
= skb_dst(skb
)->tclassid
;
264 META_COLLECTOR(int_rtiif
)
266 if (unlikely(skb_rtable(skb
) == NULL
))
269 dst
->value
= inet_iif(skb
);
272 /**************************************************************************
274 **************************************************************************/
276 #define skip_nonlocal(skb) \
277 (unlikely(skb->sk == NULL))
279 META_COLLECTOR(int_sk_family
)
281 if (skip_nonlocal(skb
)) {
285 dst
->value
= skb
->sk
->sk_family
;
288 META_COLLECTOR(int_sk_state
)
290 if (skip_nonlocal(skb
)) {
294 dst
->value
= skb
->sk
->sk_state
;
297 META_COLLECTOR(int_sk_reuse
)
299 if (skip_nonlocal(skb
)) {
303 dst
->value
= skb
->sk
->sk_reuse
;
306 META_COLLECTOR(int_sk_bound_if
)
308 if (skip_nonlocal(skb
)) {
312 /* No error if bound_dev_if is 0, legal userspace check */
313 dst
->value
= skb
->sk
->sk_bound_dev_if
;
316 META_COLLECTOR(var_sk_bound_if
)
318 if (skip_nonlocal(skb
)) {
323 if (skb
->sk
->sk_bound_dev_if
== 0) {
324 dst
->value
= (unsigned long) "any";
327 struct net_device
*dev
;
330 dev
= dev_get_by_index_rcu(sock_net(skb
->sk
),
331 skb
->sk
->sk_bound_dev_if
);
332 *err
= var_dev(dev
, dst
);
337 META_COLLECTOR(int_sk_refcnt
)
339 if (skip_nonlocal(skb
)) {
343 dst
->value
= refcount_read(&skb
->sk
->sk_refcnt
);
346 META_COLLECTOR(int_sk_rcvbuf
)
348 const struct sock
*sk
= skb_to_full_sk(skb
);
354 dst
->value
= sk
->sk_rcvbuf
;
357 META_COLLECTOR(int_sk_shutdown
)
359 const struct sock
*sk
= skb_to_full_sk(skb
);
365 dst
->value
= sk
->sk_shutdown
;
368 META_COLLECTOR(int_sk_proto
)
370 const struct sock
*sk
= skb_to_full_sk(skb
);
376 dst
->value
= sk
->sk_protocol
;
379 META_COLLECTOR(int_sk_type
)
381 const struct sock
*sk
= skb_to_full_sk(skb
);
387 dst
->value
= sk
->sk_type
;
390 META_COLLECTOR(int_sk_rmem_alloc
)
392 const struct sock
*sk
= skb_to_full_sk(skb
);
398 dst
->value
= sk_rmem_alloc_get(sk
);
401 META_COLLECTOR(int_sk_wmem_alloc
)
403 const struct sock
*sk
= skb_to_full_sk(skb
);
409 dst
->value
= sk_wmem_alloc_get(sk
);
412 META_COLLECTOR(int_sk_omem_alloc
)
414 const struct sock
*sk
= skb_to_full_sk(skb
);
420 dst
->value
= atomic_read(&sk
->sk_omem_alloc
);
423 META_COLLECTOR(int_sk_rcv_qlen
)
425 const struct sock
*sk
= skb_to_full_sk(skb
);
431 dst
->value
= sk
->sk_receive_queue
.qlen
;
434 META_COLLECTOR(int_sk_snd_qlen
)
436 const struct sock
*sk
= skb_to_full_sk(skb
);
442 dst
->value
= sk
->sk_write_queue
.qlen
;
445 META_COLLECTOR(int_sk_wmem_queued
)
447 const struct sock
*sk
= skb_to_full_sk(skb
);
453 dst
->value
= sk
->sk_wmem_queued
;
456 META_COLLECTOR(int_sk_fwd_alloc
)
458 const struct sock
*sk
= skb_to_full_sk(skb
);
464 dst
->value
= sk
->sk_forward_alloc
;
467 META_COLLECTOR(int_sk_sndbuf
)
469 const struct sock
*sk
= skb_to_full_sk(skb
);
475 dst
->value
= sk
->sk_sndbuf
;
478 META_COLLECTOR(int_sk_alloc
)
480 const struct sock
*sk
= skb_to_full_sk(skb
);
486 dst
->value
= (__force
int) sk
->sk_allocation
;
489 META_COLLECTOR(int_sk_hash
)
491 if (skip_nonlocal(skb
)) {
495 dst
->value
= skb
->sk
->sk_hash
;
498 META_COLLECTOR(int_sk_lingertime
)
500 const struct sock
*sk
= skb_to_full_sk(skb
);
506 dst
->value
= sk
->sk_lingertime
/ HZ
;
509 META_COLLECTOR(int_sk_err_qlen
)
511 const struct sock
*sk
= skb_to_full_sk(skb
);
517 dst
->value
= sk
->sk_error_queue
.qlen
;
520 META_COLLECTOR(int_sk_ack_bl
)
522 const struct sock
*sk
= skb_to_full_sk(skb
);
528 dst
->value
= sk
->sk_ack_backlog
;
531 META_COLLECTOR(int_sk_max_ack_bl
)
533 const struct sock
*sk
= skb_to_full_sk(skb
);
539 dst
->value
= sk
->sk_max_ack_backlog
;
542 META_COLLECTOR(int_sk_prio
)
544 const struct sock
*sk
= skb_to_full_sk(skb
);
550 dst
->value
= sk
->sk_priority
;
553 META_COLLECTOR(int_sk_rcvlowat
)
555 const struct sock
*sk
= skb_to_full_sk(skb
);
561 dst
->value
= sk
->sk_rcvlowat
;
564 META_COLLECTOR(int_sk_rcvtimeo
)
566 const struct sock
*sk
= skb_to_full_sk(skb
);
572 dst
->value
= sk
->sk_rcvtimeo
/ HZ
;
575 META_COLLECTOR(int_sk_sndtimeo
)
577 const struct sock
*sk
= skb_to_full_sk(skb
);
583 dst
->value
= sk
->sk_sndtimeo
/ HZ
;
586 META_COLLECTOR(int_sk_sendmsg_off
)
588 const struct sock
*sk
= skb_to_full_sk(skb
);
594 dst
->value
= sk
->sk_frag
.offset
;
597 META_COLLECTOR(int_sk_write_pend
)
599 const struct sock
*sk
= skb_to_full_sk(skb
);
605 dst
->value
= sk
->sk_write_pending
;
608 /**************************************************************************
609 * Meta value collectors assignment table
610 **************************************************************************/
613 void (*get
)(struct sk_buff
*, struct tcf_pkt_info
*,
614 struct meta_value
*, struct meta_obj
*, int *);
617 #define META_ID(name) TCF_META_ID_##name
618 #define META_FUNC(name) { .get = meta_##name }
620 /* Meta value operations table listing all meta value collectors and
621 * assigns them to a type and meta id. */
622 static struct meta_ops __meta_ops
[TCF_META_TYPE_MAX
+ 1][TCF_META_ID_MAX
+ 1] = {
623 [TCF_META_TYPE_VAR
] = {
624 [META_ID(DEV
)] = META_FUNC(var_dev
),
625 [META_ID(SK_BOUND_IF
)] = META_FUNC(var_sk_bound_if
),
627 [TCF_META_TYPE_INT
] = {
628 [META_ID(RANDOM
)] = META_FUNC(int_random
),
629 [META_ID(LOADAVG_0
)] = META_FUNC(int_loadavg_0
),
630 [META_ID(LOADAVG_1
)] = META_FUNC(int_loadavg_1
),
631 [META_ID(LOADAVG_2
)] = META_FUNC(int_loadavg_2
),
632 [META_ID(DEV
)] = META_FUNC(int_dev
),
633 [META_ID(PRIORITY
)] = META_FUNC(int_priority
),
634 [META_ID(PROTOCOL
)] = META_FUNC(int_protocol
),
635 [META_ID(PKTTYPE
)] = META_FUNC(int_pkttype
),
636 [META_ID(PKTLEN
)] = META_FUNC(int_pktlen
),
637 [META_ID(DATALEN
)] = META_FUNC(int_datalen
),
638 [META_ID(MACLEN
)] = META_FUNC(int_maclen
),
639 [META_ID(NFMARK
)] = META_FUNC(int_mark
),
640 [META_ID(TCINDEX
)] = META_FUNC(int_tcindex
),
641 [META_ID(RTCLASSID
)] = META_FUNC(int_rtclassid
),
642 [META_ID(RTIIF
)] = META_FUNC(int_rtiif
),
643 [META_ID(SK_FAMILY
)] = META_FUNC(int_sk_family
),
644 [META_ID(SK_STATE
)] = META_FUNC(int_sk_state
),
645 [META_ID(SK_REUSE
)] = META_FUNC(int_sk_reuse
),
646 [META_ID(SK_BOUND_IF
)] = META_FUNC(int_sk_bound_if
),
647 [META_ID(SK_REFCNT
)] = META_FUNC(int_sk_refcnt
),
648 [META_ID(SK_RCVBUF
)] = META_FUNC(int_sk_rcvbuf
),
649 [META_ID(SK_SNDBUF
)] = META_FUNC(int_sk_sndbuf
),
650 [META_ID(SK_SHUTDOWN
)] = META_FUNC(int_sk_shutdown
),
651 [META_ID(SK_PROTO
)] = META_FUNC(int_sk_proto
),
652 [META_ID(SK_TYPE
)] = META_FUNC(int_sk_type
),
653 [META_ID(SK_RMEM_ALLOC
)] = META_FUNC(int_sk_rmem_alloc
),
654 [META_ID(SK_WMEM_ALLOC
)] = META_FUNC(int_sk_wmem_alloc
),
655 [META_ID(SK_OMEM_ALLOC
)] = META_FUNC(int_sk_omem_alloc
),
656 [META_ID(SK_WMEM_QUEUED
)] = META_FUNC(int_sk_wmem_queued
),
657 [META_ID(SK_RCV_QLEN
)] = META_FUNC(int_sk_rcv_qlen
),
658 [META_ID(SK_SND_QLEN
)] = META_FUNC(int_sk_snd_qlen
),
659 [META_ID(SK_ERR_QLEN
)] = META_FUNC(int_sk_err_qlen
),
660 [META_ID(SK_FORWARD_ALLOCS
)] = META_FUNC(int_sk_fwd_alloc
),
661 [META_ID(SK_ALLOCS
)] = META_FUNC(int_sk_alloc
),
662 [META_ID(SK_HASH
)] = META_FUNC(int_sk_hash
),
663 [META_ID(SK_LINGERTIME
)] = META_FUNC(int_sk_lingertime
),
664 [META_ID(SK_ACK_BACKLOG
)] = META_FUNC(int_sk_ack_bl
),
665 [META_ID(SK_MAX_ACK_BACKLOG
)] = META_FUNC(int_sk_max_ack_bl
),
666 [META_ID(SK_PRIO
)] = META_FUNC(int_sk_prio
),
667 [META_ID(SK_RCVLOWAT
)] = META_FUNC(int_sk_rcvlowat
),
668 [META_ID(SK_RCVTIMEO
)] = META_FUNC(int_sk_rcvtimeo
),
669 [META_ID(SK_SNDTIMEO
)] = META_FUNC(int_sk_sndtimeo
),
670 [META_ID(SK_SENDMSG_OFF
)] = META_FUNC(int_sk_sendmsg_off
),
671 [META_ID(SK_WRITE_PENDING
)] = META_FUNC(int_sk_write_pend
),
672 [META_ID(VLAN_TAG
)] = META_FUNC(int_vlan_tag
),
673 [META_ID(RXHASH
)] = META_FUNC(int_rxhash
),
677 static inline struct meta_ops
*meta_ops(struct meta_value
*val
)
679 return &__meta_ops
[meta_type(val
)][meta_id(val
)];
682 /**************************************************************************
683 * Type specific operations for TCF_META_TYPE_VAR
684 **************************************************************************/
686 static int meta_var_compare(struct meta_obj
*a
, struct meta_obj
*b
)
688 int r
= a
->len
- b
->len
;
691 r
= memcmp((void *) a
->value
, (void *) b
->value
, a
->len
);
696 static int meta_var_change(struct meta_value
*dst
, struct nlattr
*nla
)
698 int len
= nla_len(nla
);
700 dst
->val
= (unsigned long)kmemdup(nla_data(nla
), len
, GFP_KERNEL
);
707 static void meta_var_destroy(struct meta_value
*v
)
709 kfree((void *) v
->val
);
712 static void meta_var_apply_extras(struct meta_value
*v
,
713 struct meta_obj
*dst
)
715 int shift
= v
->hdr
.shift
;
717 if (shift
&& shift
< dst
->len
)
721 static int meta_var_dump(struct sk_buff
*skb
, struct meta_value
*v
, int tlv
)
723 if (v
->val
&& v
->len
&&
724 nla_put(skb
, tlv
, v
->len
, (void *) v
->val
))
725 goto nla_put_failure
;
732 /**************************************************************************
733 * Type specific operations for TCF_META_TYPE_INT
734 **************************************************************************/
736 static int meta_int_compare(struct meta_obj
*a
, struct meta_obj
*b
)
738 /* Let gcc optimize it, the unlikely is not really based on
739 * some numbers but jump free code for mismatches seems
741 if (unlikely(a
->value
== b
->value
))
743 else if (a
->value
< b
->value
)
749 static int meta_int_change(struct meta_value
*dst
, struct nlattr
*nla
)
751 if (nla_len(nla
) >= sizeof(unsigned long)) {
752 dst
->val
= *(unsigned long *) nla_data(nla
);
753 dst
->len
= sizeof(unsigned long);
754 } else if (nla_len(nla
) == sizeof(u32
)) {
755 dst
->val
= nla_get_u32(nla
);
756 dst
->len
= sizeof(u32
);
763 static void meta_int_apply_extras(struct meta_value
*v
,
764 struct meta_obj
*dst
)
767 dst
->value
>>= v
->hdr
.shift
;
770 dst
->value
&= v
->val
;
773 static int meta_int_dump(struct sk_buff
*skb
, struct meta_value
*v
, int tlv
)
775 if (v
->len
== sizeof(unsigned long)) {
776 if (nla_put(skb
, tlv
, sizeof(unsigned long), &v
->val
))
777 goto nla_put_failure
;
778 } else if (v
->len
== sizeof(u32
)) {
779 if (nla_put_u32(skb
, tlv
, v
->val
))
780 goto nla_put_failure
;
789 /**************************************************************************
790 * Type specific operations table
791 **************************************************************************/
793 struct meta_type_ops
{
794 void (*destroy
)(struct meta_value
*);
795 int (*compare
)(struct meta_obj
*, struct meta_obj
*);
796 int (*change
)(struct meta_value
*, struct nlattr
*);
797 void (*apply_extras
)(struct meta_value
*, struct meta_obj
*);
798 int (*dump
)(struct sk_buff
*, struct meta_value
*, int);
801 static const struct meta_type_ops __meta_type_ops
[TCF_META_TYPE_MAX
+ 1] = {
802 [TCF_META_TYPE_VAR
] = {
803 .destroy
= meta_var_destroy
,
804 .compare
= meta_var_compare
,
805 .change
= meta_var_change
,
806 .apply_extras
= meta_var_apply_extras
,
807 .dump
= meta_var_dump
809 [TCF_META_TYPE_INT
] = {
810 .compare
= meta_int_compare
,
811 .change
= meta_int_change
,
812 .apply_extras
= meta_int_apply_extras
,
813 .dump
= meta_int_dump
817 static inline const struct meta_type_ops
*meta_type_ops(struct meta_value
*v
)
819 return &__meta_type_ops
[meta_type(v
)];
822 /**************************************************************************
824 **************************************************************************/
826 static int meta_get(struct sk_buff
*skb
, struct tcf_pkt_info
*info
,
827 struct meta_value
*v
, struct meta_obj
*dst
)
831 if (meta_id(v
) == TCF_META_ID_VALUE
) {
837 meta_ops(v
)->get(skb
, info
, v
, dst
, &err
);
841 if (meta_type_ops(v
)->apply_extras
)
842 meta_type_ops(v
)->apply_extras(v
, dst
);
847 static int em_meta_match(struct sk_buff
*skb
, struct tcf_ematch
*m
,
848 struct tcf_pkt_info
*info
)
851 struct meta_match
*meta
= (struct meta_match
*) m
->data
;
852 struct meta_obj l_value
, r_value
;
854 if (meta_get(skb
, info
, &meta
->lvalue
, &l_value
) < 0 ||
855 meta_get(skb
, info
, &meta
->rvalue
, &r_value
) < 0)
858 r
= meta_type_ops(&meta
->lvalue
)->compare(&l_value
, &r_value
);
860 switch (meta
->lvalue
.hdr
.op
) {
872 static void meta_delete(struct meta_match
*meta
)
875 const struct meta_type_ops
*ops
= meta_type_ops(&meta
->lvalue
);
877 if (ops
&& ops
->destroy
) {
878 ops
->destroy(&meta
->lvalue
);
879 ops
->destroy(&meta
->rvalue
);
886 static inline int meta_change_data(struct meta_value
*dst
, struct nlattr
*nla
)
889 if (nla_len(nla
) == 0)
892 return meta_type_ops(dst
)->change(dst
, nla
);
898 static inline int meta_is_supported(struct meta_value
*val
)
900 return !meta_id(val
) || meta_ops(val
)->get
;
903 static const struct nla_policy meta_policy
[TCA_EM_META_MAX
+ 1] = {
904 [TCA_EM_META_HDR
] = { .len
= sizeof(struct tcf_meta_hdr
) },
907 static int em_meta_change(struct net
*net
, void *data
, int len
,
908 struct tcf_ematch
*m
)
911 struct nlattr
*tb
[TCA_EM_META_MAX
+ 1];
912 struct tcf_meta_hdr
*hdr
;
913 struct meta_match
*meta
= NULL
;
915 err
= nla_parse(tb
, TCA_EM_META_MAX
, data
, len
, meta_policy
, NULL
);
920 if (tb
[TCA_EM_META_HDR
] == NULL
)
922 hdr
= nla_data(tb
[TCA_EM_META_HDR
]);
924 if (TCF_META_TYPE(hdr
->left
.kind
) != TCF_META_TYPE(hdr
->right
.kind
) ||
925 TCF_META_TYPE(hdr
->left
.kind
) > TCF_META_TYPE_MAX
||
926 TCF_META_ID(hdr
->left
.kind
) > TCF_META_ID_MAX
||
927 TCF_META_ID(hdr
->right
.kind
) > TCF_META_ID_MAX
)
930 meta
= kzalloc(sizeof(*meta
), GFP_KERNEL
);
936 memcpy(&meta
->lvalue
.hdr
, &hdr
->left
, sizeof(hdr
->left
));
937 memcpy(&meta
->rvalue
.hdr
, &hdr
->right
, sizeof(hdr
->right
));
939 if (!meta_is_supported(&meta
->lvalue
) ||
940 !meta_is_supported(&meta
->rvalue
)) {
945 if (meta_change_data(&meta
->lvalue
, tb
[TCA_EM_META_LVALUE
]) < 0 ||
946 meta_change_data(&meta
->rvalue
, tb
[TCA_EM_META_RVALUE
]) < 0)
949 m
->datalen
= sizeof(*meta
);
950 m
->data
= (unsigned long) meta
;
959 static void em_meta_destroy(struct tcf_ematch
*m
)
962 meta_delete((struct meta_match
*) m
->data
);
965 static int em_meta_dump(struct sk_buff
*skb
, struct tcf_ematch
*em
)
967 struct meta_match
*meta
= (struct meta_match
*) em
->data
;
968 struct tcf_meta_hdr hdr
;
969 const struct meta_type_ops
*ops
;
971 memset(&hdr
, 0, sizeof(hdr
));
972 memcpy(&hdr
.left
, &meta
->lvalue
.hdr
, sizeof(hdr
.left
));
973 memcpy(&hdr
.right
, &meta
->rvalue
.hdr
, sizeof(hdr
.right
));
975 if (nla_put(skb
, TCA_EM_META_HDR
, sizeof(hdr
), &hdr
))
976 goto nla_put_failure
;
978 ops
= meta_type_ops(&meta
->lvalue
);
979 if (ops
->dump(skb
, &meta
->lvalue
, TCA_EM_META_LVALUE
) < 0 ||
980 ops
->dump(skb
, &meta
->rvalue
, TCA_EM_META_RVALUE
) < 0)
981 goto nla_put_failure
;
989 static struct tcf_ematch_ops em_meta_ops
= {
991 .change
= em_meta_change
,
992 .match
= em_meta_match
,
993 .destroy
= em_meta_destroy
,
994 .dump
= em_meta_dump
,
995 .owner
= THIS_MODULE
,
996 .link
= LIST_HEAD_INIT(em_meta_ops
.link
)
999 static int __init
init_em_meta(void)
1001 return tcf_em_register(&em_meta_ops
);
1004 static void __exit
exit_em_meta(void)
1006 tcf_em_unregister(&em_meta_ops
);
1009 MODULE_LICENSE("GPL");
1011 module_init(init_em_meta
);
1012 module_exit(exit_em_meta
);
1014 MODULE_ALIAS_TCF_EMATCH(TCF_EM_META
);