2 * gw.c - CAN frame Gateway/Router/Bridge with netlink interface
4 * Copyright (c) 2011 Volkswagen Group Electronic Research
7 * Redistribution and use in source and binary forms, with or without
8 * modification, are permitted provided that the following conditions
10 * 1. Redistributions of source code must retain the above copyright
11 * notice, this list of conditions and the following disclaimer.
12 * 2. Redistributions in binary form must reproduce the above copyright
13 * notice, this list of conditions and the following disclaimer in the
14 * documentation and/or other materials provided with the distribution.
15 * 3. Neither the name of Volkswagen nor the names of its contributors
16 * may be used to endorse or promote products derived from this software
17 * without specific prior written permission.
19 * Alternatively, provided that this notice is retained in full, this
20 * software may be distributed under the terms of the GNU General
21 * Public License ("GPL") version 2, in which case the provisions of the
22 * GPL apply INSTEAD OF those given above.
24 * The provided data structures and external interfaces from this code
25 * are not restricted to be used by modules with a GPL compatible license.
27 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
28 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
29 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
30 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
31 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
32 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
33 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
34 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
35 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
36 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
37 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH
42 #include <linux/module.h>
43 #include <linux/init.h>
44 #include <linux/types.h>
45 #include <linux/kernel.h>
46 #include <linux/list.h>
47 #include <linux/spinlock.h>
48 #include <linux/rcupdate.h>
49 #include <linux/rculist.h>
50 #include <linux/net.h>
51 #include <linux/netdevice.h>
52 #include <linux/if_arp.h>
53 #include <linux/skbuff.h>
54 #include <linux/can.h>
55 #include <linux/can/core.h>
56 #include <linux/can/skb.h>
57 #include <linux/can/gw.h>
58 #include <net/rtnetlink.h>
59 #include <net/net_namespace.h>
62 #define CAN_GW_VERSION "20130117"
63 #define CAN_GW_NAME "can-gw"
65 MODULE_DESCRIPTION("PF_CAN netlink gateway");
66 MODULE_LICENSE("Dual BSD/GPL");
67 MODULE_AUTHOR("Oliver Hartkopp <oliver.hartkopp@volkswagen.de>");
68 MODULE_ALIAS(CAN_GW_NAME
);
70 #define CGW_MIN_HOPS 1
71 #define CGW_MAX_HOPS 6
72 #define CGW_DEFAULT_HOPS 1
74 static unsigned int max_hops __read_mostly
= CGW_DEFAULT_HOPS
;
75 module_param(max_hops
, uint
, S_IRUGO
);
76 MODULE_PARM_DESC(max_hops
,
77 "maximum " CAN_GW_NAME
" routing hops for CAN frames "
78 "(valid values: " __stringify(CGW_MIN_HOPS
) "-"
79 __stringify(CGW_MAX_HOPS
) " hops, "
80 "default: " __stringify(CGW_DEFAULT_HOPS
) ")");
82 static HLIST_HEAD(cgw_list
);
83 static struct notifier_block notifier
;
85 static struct kmem_cache
*cgw_cache __read_mostly
;
87 /* structure that contains the (on-the-fly) CAN frame modifications */
101 void (*modfunc
[MAX_MODFUNCTIONS
])(struct can_frame
*cf
,
104 /* CAN frame checksum calculation after CAN frame modifications */
106 struct cgw_csum_xor
xor;
107 struct cgw_csum_crc8 crc8
;
110 void (*xor)(struct can_frame
*cf
, struct cgw_csum_xor
*xor);
111 void (*crc8
)(struct can_frame
*cf
, struct cgw_csum_crc8
*crc8
);
117 * So far we just support CAN -> CAN routing and frame modifications.
119 * The internal can_can_gw structure contains data and attributes for
120 * a CAN -> CAN gateway job.
123 struct can_filter filter
;
128 /* list entry for CAN gateways jobs */
130 struct hlist_node list
;
137 /* CAN frame data source */
138 struct net_device
*dev
;
141 /* CAN frame data destination */
142 struct net_device
*dev
;
145 struct can_can_gw ccgw
;
153 /* modification functions that are invoked in the hot path in can_can_gw_rcv */
155 #define MODFUNC(func, op) static void func(struct can_frame *cf, \
156 struct cf_mod *mod) { op ; }
158 MODFUNC(mod_and_id
, cf
->can_id
&= mod
->modframe
.and.can_id
)
159 MODFUNC(mod_and_dlc
, cf
->can_dlc
&= mod
->modframe
.and.can_dlc
)
160 MODFUNC(mod_and_data
, *(u64
*)cf
->data
&= *(u64
*)mod
->modframe
.and.data
)
161 MODFUNC(mod_or_id
, cf
->can_id
|= mod
->modframe
.or.can_id
)
162 MODFUNC(mod_or_dlc
, cf
->can_dlc
|= mod
->modframe
.or.can_dlc
)
163 MODFUNC(mod_or_data
, *(u64
*)cf
->data
|= *(u64
*)mod
->modframe
.or.data
)
164 MODFUNC(mod_xor_id
, cf
->can_id
^= mod
->modframe
.xor.can_id
)
165 MODFUNC(mod_xor_dlc
, cf
->can_dlc
^= mod
->modframe
.xor.can_dlc
)
166 MODFUNC(mod_xor_data
, *(u64
*)cf
->data
^= *(u64
*)mod
->modframe
.xor.data
)
167 MODFUNC(mod_set_id
, cf
->can_id
= mod
->modframe
.set
.can_id
)
168 MODFUNC(mod_set_dlc
, cf
->can_dlc
= mod
->modframe
.set
.can_dlc
)
169 MODFUNC(mod_set_data
, *(u64
*)cf
->data
= *(u64
*)mod
->modframe
.set
.data
)
171 static inline void canframecpy(struct can_frame
*dst
, struct can_frame
*src
)
174 * Copy the struct members separately to ensure that no uninitialized
175 * data are copied in the 3 bytes hole of the struct. This is needed
176 * to make easy compares of the data in the struct cf_mod.
179 dst
->can_id
= src
->can_id
;
180 dst
->can_dlc
= src
->can_dlc
;
181 *(u64
*)dst
->data
= *(u64
*)src
->data
;
184 static int cgw_chk_csum_parms(s8 fr
, s8 to
, s8 re
)
187 * absolute dlc values 0 .. 7 => 0 .. 7, e.g. data [0]
188 * relative to received dlc -1 .. -8 :
189 * e.g. for received dlc = 8
190 * -1 => index = 7 (data[7])
191 * -3 => index = 5 (data[5])
192 * -8 => index = 0 (data[0])
195 if (fr
> -9 && fr
< 8 &&
203 static inline int calc_idx(int idx
, int rx_dlc
)
211 static void cgw_csum_xor_rel(struct can_frame
*cf
, struct cgw_csum_xor
*xor)
213 int from
= calc_idx(xor->from_idx
, cf
->can_dlc
);
214 int to
= calc_idx(xor->to_idx
, cf
->can_dlc
);
215 int res
= calc_idx(xor->result_idx
, cf
->can_dlc
);
216 u8 val
= xor->init_xor_val
;
219 if (from
< 0 || to
< 0 || res
< 0)
223 for (i
= from
; i
<= to
; i
++)
226 for (i
= from
; i
>= to
; i
--)
233 static void cgw_csum_xor_pos(struct can_frame
*cf
, struct cgw_csum_xor
*xor)
235 u8 val
= xor->init_xor_val
;
238 for (i
= xor->from_idx
; i
<= xor->to_idx
; i
++)
241 cf
->data
[xor->result_idx
] = val
;
244 static void cgw_csum_xor_neg(struct can_frame
*cf
, struct cgw_csum_xor
*xor)
246 u8 val
= xor->init_xor_val
;
249 for (i
= xor->from_idx
; i
>= xor->to_idx
; i
--)
252 cf
->data
[xor->result_idx
] = val
;
255 static void cgw_csum_crc8_rel(struct can_frame
*cf
, struct cgw_csum_crc8
*crc8
)
257 int from
= calc_idx(crc8
->from_idx
, cf
->can_dlc
);
258 int to
= calc_idx(crc8
->to_idx
, cf
->can_dlc
);
259 int res
= calc_idx(crc8
->result_idx
, cf
->can_dlc
);
260 u8 crc
= crc8
->init_crc_val
;
263 if (from
< 0 || to
< 0 || res
< 0)
267 for (i
= crc8
->from_idx
; i
<= crc8
->to_idx
; i
++)
268 crc
= crc8
->crctab
[crc
^cf
->data
[i
]];
270 for (i
= crc8
->from_idx
; i
>= crc8
->to_idx
; i
--)
271 crc
= crc8
->crctab
[crc
^cf
->data
[i
]];
274 switch (crc8
->profile
) {
276 case CGW_CRC8PRF_1U8
:
277 crc
= crc8
->crctab
[crc
^crc8
->profile_data
[0]];
280 case CGW_CRC8PRF_16U8
:
281 crc
= crc8
->crctab
[crc
^crc8
->profile_data
[cf
->data
[1] & 0xF]];
284 case CGW_CRC8PRF_SFFID_XOR
:
285 crc
= crc8
->crctab
[crc
^(cf
->can_id
& 0xFF)^
286 (cf
->can_id
>> 8 & 0xFF)];
291 cf
->data
[crc8
->result_idx
] = crc
^crc8
->final_xor_val
;
294 static void cgw_csum_crc8_pos(struct can_frame
*cf
, struct cgw_csum_crc8
*crc8
)
296 u8 crc
= crc8
->init_crc_val
;
299 for (i
= crc8
->from_idx
; i
<= crc8
->to_idx
; i
++)
300 crc
= crc8
->crctab
[crc
^cf
->data
[i
]];
302 switch (crc8
->profile
) {
304 case CGW_CRC8PRF_1U8
:
305 crc
= crc8
->crctab
[crc
^crc8
->profile_data
[0]];
308 case CGW_CRC8PRF_16U8
:
309 crc
= crc8
->crctab
[crc
^crc8
->profile_data
[cf
->data
[1] & 0xF]];
312 case CGW_CRC8PRF_SFFID_XOR
:
313 crc
= crc8
->crctab
[crc
^(cf
->can_id
& 0xFF)^
314 (cf
->can_id
>> 8 & 0xFF)];
318 cf
->data
[crc8
->result_idx
] = crc
^crc8
->final_xor_val
;
321 static void cgw_csum_crc8_neg(struct can_frame
*cf
, struct cgw_csum_crc8
*crc8
)
323 u8 crc
= crc8
->init_crc_val
;
326 for (i
= crc8
->from_idx
; i
>= crc8
->to_idx
; i
--)
327 crc
= crc8
->crctab
[crc
^cf
->data
[i
]];
329 switch (crc8
->profile
) {
331 case CGW_CRC8PRF_1U8
:
332 crc
= crc8
->crctab
[crc
^crc8
->profile_data
[0]];
335 case CGW_CRC8PRF_16U8
:
336 crc
= crc8
->crctab
[crc
^crc8
->profile_data
[cf
->data
[1] & 0xF]];
339 case CGW_CRC8PRF_SFFID_XOR
:
340 crc
= crc8
->crctab
[crc
^(cf
->can_id
& 0xFF)^
341 (cf
->can_id
>> 8 & 0xFF)];
345 cf
->data
[crc8
->result_idx
] = crc
^crc8
->final_xor_val
;
348 /* the receive & process & send function */
349 static void can_can_gw_rcv(struct sk_buff
*skb
, void *data
)
351 struct cgw_job
*gwj
= (struct cgw_job
*)data
;
352 struct can_frame
*cf
;
353 struct sk_buff
*nskb
;
357 * Do not handle CAN frames routed more than 'max_hops' times.
358 * In general we should never catch this delimiter which is intended
359 * to cover a misconfiguration protection (e.g. circular CAN routes).
361 * The Controller Area Network controllers only accept CAN frames with
362 * correct CRCs - which are not visible in the controller registers.
363 * According to skbuff.h documentation the csum_start element for IP
364 * checksums is undefined/unsued when ip_summed == CHECKSUM_UNNECESSARY.
365 * Only CAN skbs can be processed here which already have this property.
368 #define cgw_hops(skb) ((skb)->csum_start)
370 BUG_ON(skb
->ip_summed
!= CHECKSUM_UNNECESSARY
);
372 if (cgw_hops(skb
) >= max_hops
) {
373 /* indicate deleted frames due to misconfiguration */
374 gwj
->deleted_frames
++;
378 if (!(gwj
->dst
.dev
->flags
& IFF_UP
)) {
379 gwj
->dropped_frames
++;
383 /* is sending the skb back to the incoming interface not allowed? */
384 if (!(gwj
->flags
& CGW_FLAGS_CAN_IIF_TX_OK
) &&
385 can_skb_prv(skb
)->ifindex
== gwj
->dst
.dev
->ifindex
)
389 * clone the given skb, which has not been done in can_rcv()
391 * When there is at least one modification function activated,
392 * we need to copy the skb as we want to modify skb->data.
394 if (gwj
->mod
.modfunc
[0])
395 nskb
= skb_copy(skb
, GFP_ATOMIC
);
397 nskb
= skb_clone(skb
, GFP_ATOMIC
);
400 gwj
->dropped_frames
++;
404 /* put the incremented hop counter in the cloned skb */
405 cgw_hops(nskb
) = cgw_hops(skb
) + 1;
407 /* first processing of this CAN frame -> adjust to private hop limit */
408 if (gwj
->limit_hops
&& cgw_hops(nskb
) == 1)
409 cgw_hops(nskb
) = max_hops
- gwj
->limit_hops
+ 1;
411 nskb
->dev
= gwj
->dst
.dev
;
413 /* pointer to modifiable CAN frame */
414 cf
= (struct can_frame
*)nskb
->data
;
416 /* perform preprocessed modification functions if there are any */
417 while (modidx
< MAX_MODFUNCTIONS
&& gwj
->mod
.modfunc
[modidx
])
418 (*gwj
->mod
.modfunc
[modidx
++])(cf
, &gwj
->mod
);
420 /* check for checksum updates when the CAN frame has been modified */
422 if (gwj
->mod
.csumfunc
.crc8
)
423 (*gwj
->mod
.csumfunc
.crc8
)(cf
, &gwj
->mod
.csum
.crc8
);
425 if (gwj
->mod
.csumfunc
.xor)
426 (*gwj
->mod
.csumfunc
.xor)(cf
, &gwj
->mod
.csum
.xor);
429 /* clear the skb timestamp if not configured the other way */
430 if (!(gwj
->flags
& CGW_FLAGS_CAN_SRC_TSTAMP
))
431 nskb
->tstamp
.tv64
= 0;
433 /* send to netdevice */
434 if (can_send(nskb
, gwj
->flags
& CGW_FLAGS_CAN_ECHO
))
435 gwj
->dropped_frames
++;
437 gwj
->handled_frames
++;
440 static inline int cgw_register_filter(struct cgw_job
*gwj
)
442 return can_rx_register(gwj
->src
.dev
, gwj
->ccgw
.filter
.can_id
,
443 gwj
->ccgw
.filter
.can_mask
, can_can_gw_rcv
,
447 static inline void cgw_unregister_filter(struct cgw_job
*gwj
)
449 can_rx_unregister(gwj
->src
.dev
, gwj
->ccgw
.filter
.can_id
,
450 gwj
->ccgw
.filter
.can_mask
, can_can_gw_rcv
, gwj
);
453 static int cgw_notifier(struct notifier_block
*nb
,
454 unsigned long msg
, void *ptr
)
456 struct net_device
*dev
= netdev_notifier_info_to_dev(ptr
);
458 if (!net_eq(dev_net(dev
), &init_net
))
460 if (dev
->type
!= ARPHRD_CAN
)
463 if (msg
== NETDEV_UNREGISTER
) {
465 struct cgw_job
*gwj
= NULL
;
466 struct hlist_node
*nx
;
470 hlist_for_each_entry_safe(gwj
, nx
, &cgw_list
, list
) {
472 if (gwj
->src
.dev
== dev
|| gwj
->dst
.dev
== dev
) {
473 hlist_del(&gwj
->list
);
474 cgw_unregister_filter(gwj
);
475 kmem_cache_free(cgw_cache
, gwj
);
483 static int cgw_put_job(struct sk_buff
*skb
, struct cgw_job
*gwj
, int type
,
484 u32 pid
, u32 seq
, int flags
)
486 struct cgw_frame_mod mb
;
487 struct rtcanmsg
*rtcan
;
488 struct nlmsghdr
*nlh
;
490 nlh
= nlmsg_put(skb
, pid
, seq
, type
, sizeof(*rtcan
), flags
);
494 rtcan
= nlmsg_data(nlh
);
495 rtcan
->can_family
= AF_CAN
;
496 rtcan
->gwtype
= gwj
->gwtype
;
497 rtcan
->flags
= gwj
->flags
;
499 /* add statistics if available */
501 if (gwj
->handled_frames
) {
502 if (nla_put_u32(skb
, CGW_HANDLED
, gwj
->handled_frames
) < 0)
506 if (gwj
->dropped_frames
) {
507 if (nla_put_u32(skb
, CGW_DROPPED
, gwj
->dropped_frames
) < 0)
511 if (gwj
->deleted_frames
) {
512 if (nla_put_u32(skb
, CGW_DELETED
, gwj
->deleted_frames
) < 0)
516 /* check non default settings of attributes */
518 if (gwj
->limit_hops
) {
519 if (nla_put_u8(skb
, CGW_LIM_HOPS
, gwj
->limit_hops
) < 0)
523 if (gwj
->mod
.modtype
.and) {
524 memcpy(&mb
.cf
, &gwj
->mod
.modframe
.and, sizeof(mb
.cf
));
525 mb
.modtype
= gwj
->mod
.modtype
.and;
526 if (nla_put(skb
, CGW_MOD_AND
, sizeof(mb
), &mb
) < 0)
530 if (gwj
->mod
.modtype
.or) {
531 memcpy(&mb
.cf
, &gwj
->mod
.modframe
.or, sizeof(mb
.cf
));
532 mb
.modtype
= gwj
->mod
.modtype
.or;
533 if (nla_put(skb
, CGW_MOD_OR
, sizeof(mb
), &mb
) < 0)
537 if (gwj
->mod
.modtype
.xor) {
538 memcpy(&mb
.cf
, &gwj
->mod
.modframe
.xor, sizeof(mb
.cf
));
539 mb
.modtype
= gwj
->mod
.modtype
.xor;
540 if (nla_put(skb
, CGW_MOD_XOR
, sizeof(mb
), &mb
) < 0)
544 if (gwj
->mod
.modtype
.set
) {
545 memcpy(&mb
.cf
, &gwj
->mod
.modframe
.set
, sizeof(mb
.cf
));
546 mb
.modtype
= gwj
->mod
.modtype
.set
;
547 if (nla_put(skb
, CGW_MOD_SET
, sizeof(mb
), &mb
) < 0)
551 if (gwj
->mod
.csumfunc
.crc8
) {
552 if (nla_put(skb
, CGW_CS_CRC8
, CGW_CS_CRC8_LEN
,
553 &gwj
->mod
.csum
.crc8
) < 0)
557 if (gwj
->mod
.csumfunc
.xor) {
558 if (nla_put(skb
, CGW_CS_XOR
, CGW_CS_XOR_LEN
,
559 &gwj
->mod
.csum
.xor) < 0)
563 if (gwj
->gwtype
== CGW_TYPE_CAN_CAN
) {
565 if (gwj
->ccgw
.filter
.can_id
|| gwj
->ccgw
.filter
.can_mask
) {
566 if (nla_put(skb
, CGW_FILTER
, sizeof(struct can_filter
),
567 &gwj
->ccgw
.filter
) < 0)
571 if (nla_put_u32(skb
, CGW_SRC_IF
, gwj
->ccgw
.src_idx
) < 0)
574 if (nla_put_u32(skb
, CGW_DST_IF
, gwj
->ccgw
.dst_idx
) < 0)
578 return nlmsg_end(skb
, nlh
);
581 nlmsg_cancel(skb
, nlh
);
585 /* Dump information about all CAN gateway jobs, in response to RTM_GETROUTE */
586 static int cgw_dump_jobs(struct sk_buff
*skb
, struct netlink_callback
*cb
)
588 struct cgw_job
*gwj
= NULL
;
590 int s_idx
= cb
->args
[0];
593 hlist_for_each_entry_rcu(gwj
, &cgw_list
, list
) {
597 if (cgw_put_job(skb
, gwj
, RTM_NEWROUTE
, NETLINK_CB(cb
->skb
).portid
,
598 cb
->nlh
->nlmsg_seq
, NLM_F_MULTI
) < 0)
610 static const struct nla_policy cgw_policy
[CGW_MAX
+1] = {
611 [CGW_MOD_AND
] = { .len
= sizeof(struct cgw_frame_mod
) },
612 [CGW_MOD_OR
] = { .len
= sizeof(struct cgw_frame_mod
) },
613 [CGW_MOD_XOR
] = { .len
= sizeof(struct cgw_frame_mod
) },
614 [CGW_MOD_SET
] = { .len
= sizeof(struct cgw_frame_mod
) },
615 [CGW_CS_XOR
] = { .len
= sizeof(struct cgw_csum_xor
) },
616 [CGW_CS_CRC8
] = { .len
= sizeof(struct cgw_csum_crc8
) },
617 [CGW_SRC_IF
] = { .type
= NLA_U32
},
618 [CGW_DST_IF
] = { .type
= NLA_U32
},
619 [CGW_FILTER
] = { .len
= sizeof(struct can_filter
) },
620 [CGW_LIM_HOPS
] = { .type
= NLA_U8
},
623 /* check for common and gwtype specific attributes */
624 static int cgw_parse_attr(struct nlmsghdr
*nlh
, struct cf_mod
*mod
,
625 u8 gwtype
, void *gwtypeattr
, u8
*limhops
)
627 struct nlattr
*tb
[CGW_MAX
+1];
628 struct cgw_frame_mod mb
;
632 /* initialize modification & checksum data space */
633 memset(mod
, 0, sizeof(*mod
));
635 err
= nlmsg_parse(nlh
, sizeof(struct rtcanmsg
), tb
, CGW_MAX
,
640 if (tb
[CGW_LIM_HOPS
]) {
641 *limhops
= nla_get_u8(tb
[CGW_LIM_HOPS
]);
643 if (*limhops
< 1 || *limhops
> max_hops
)
647 /* check for AND/OR/XOR/SET modifications */
649 if (tb
[CGW_MOD_AND
]) {
650 nla_memcpy(&mb
, tb
[CGW_MOD_AND
], CGW_MODATTR_LEN
);
652 canframecpy(&mod
->modframe
.and, &mb
.cf
);
653 mod
->modtype
.and = mb
.modtype
;
655 if (mb
.modtype
& CGW_MOD_ID
)
656 mod
->modfunc
[modidx
++] = mod_and_id
;
658 if (mb
.modtype
& CGW_MOD_DLC
)
659 mod
->modfunc
[modidx
++] = mod_and_dlc
;
661 if (mb
.modtype
& CGW_MOD_DATA
)
662 mod
->modfunc
[modidx
++] = mod_and_data
;
665 if (tb
[CGW_MOD_OR
]) {
666 nla_memcpy(&mb
, tb
[CGW_MOD_OR
], CGW_MODATTR_LEN
);
668 canframecpy(&mod
->modframe
.or, &mb
.cf
);
669 mod
->modtype
.or = mb
.modtype
;
671 if (mb
.modtype
& CGW_MOD_ID
)
672 mod
->modfunc
[modidx
++] = mod_or_id
;
674 if (mb
.modtype
& CGW_MOD_DLC
)
675 mod
->modfunc
[modidx
++] = mod_or_dlc
;
677 if (mb
.modtype
& CGW_MOD_DATA
)
678 mod
->modfunc
[modidx
++] = mod_or_data
;
681 if (tb
[CGW_MOD_XOR
]) {
682 nla_memcpy(&mb
, tb
[CGW_MOD_XOR
], CGW_MODATTR_LEN
);
684 canframecpy(&mod
->modframe
.xor, &mb
.cf
);
685 mod
->modtype
.xor = mb
.modtype
;
687 if (mb
.modtype
& CGW_MOD_ID
)
688 mod
->modfunc
[modidx
++] = mod_xor_id
;
690 if (mb
.modtype
& CGW_MOD_DLC
)
691 mod
->modfunc
[modidx
++] = mod_xor_dlc
;
693 if (mb
.modtype
& CGW_MOD_DATA
)
694 mod
->modfunc
[modidx
++] = mod_xor_data
;
697 if (tb
[CGW_MOD_SET
]) {
698 nla_memcpy(&mb
, tb
[CGW_MOD_SET
], CGW_MODATTR_LEN
);
700 canframecpy(&mod
->modframe
.set
, &mb
.cf
);
701 mod
->modtype
.set
= mb
.modtype
;
703 if (mb
.modtype
& CGW_MOD_ID
)
704 mod
->modfunc
[modidx
++] = mod_set_id
;
706 if (mb
.modtype
& CGW_MOD_DLC
)
707 mod
->modfunc
[modidx
++] = mod_set_dlc
;
709 if (mb
.modtype
& CGW_MOD_DATA
)
710 mod
->modfunc
[modidx
++] = mod_set_data
;
713 /* check for checksum operations after CAN frame modifications */
716 if (tb
[CGW_CS_CRC8
]) {
717 struct cgw_csum_crc8
*c
= nla_data(tb
[CGW_CS_CRC8
]);
719 err
= cgw_chk_csum_parms(c
->from_idx
, c
->to_idx
,
724 nla_memcpy(&mod
->csum
.crc8
, tb
[CGW_CS_CRC8
],
728 * select dedicated processing function to reduce
729 * runtime operations in receive hot path.
731 if (c
->from_idx
< 0 || c
->to_idx
< 0 ||
733 mod
->csumfunc
.crc8
= cgw_csum_crc8_rel
;
734 else if (c
->from_idx
<= c
->to_idx
)
735 mod
->csumfunc
.crc8
= cgw_csum_crc8_pos
;
737 mod
->csumfunc
.crc8
= cgw_csum_crc8_neg
;
740 if (tb
[CGW_CS_XOR
]) {
741 struct cgw_csum_xor
*c
= nla_data(tb
[CGW_CS_XOR
]);
743 err
= cgw_chk_csum_parms(c
->from_idx
, c
->to_idx
,
748 nla_memcpy(&mod
->csum
.xor, tb
[CGW_CS_XOR
],
752 * select dedicated processing function to reduce
753 * runtime operations in receive hot path.
755 if (c
->from_idx
< 0 || c
->to_idx
< 0 ||
757 mod
->csumfunc
.xor = cgw_csum_xor_rel
;
758 else if (c
->from_idx
<= c
->to_idx
)
759 mod
->csumfunc
.xor = cgw_csum_xor_pos
;
761 mod
->csumfunc
.xor = cgw_csum_xor_neg
;
765 if (gwtype
== CGW_TYPE_CAN_CAN
) {
767 /* check CGW_TYPE_CAN_CAN specific attributes */
769 struct can_can_gw
*ccgw
= (struct can_can_gw
*)gwtypeattr
;
770 memset(ccgw
, 0, sizeof(*ccgw
));
772 /* check for can_filter in attributes */
774 nla_memcpy(&ccgw
->filter
, tb
[CGW_FILTER
],
775 sizeof(struct can_filter
));
779 /* specifying two interfaces is mandatory */
780 if (!tb
[CGW_SRC_IF
] || !tb
[CGW_DST_IF
])
783 ccgw
->src_idx
= nla_get_u32(tb
[CGW_SRC_IF
]);
784 ccgw
->dst_idx
= nla_get_u32(tb
[CGW_DST_IF
]);
786 /* both indices set to 0 for flushing all routing entries */
787 if (!ccgw
->src_idx
&& !ccgw
->dst_idx
)
790 /* only one index set to 0 is an error */
791 if (!ccgw
->src_idx
|| !ccgw
->dst_idx
)
795 /* add the checks for other gwtypes here */
800 static int cgw_create_job(struct sk_buff
*skb
, struct nlmsghdr
*nlh
)
807 if (!netlink_capable(skb
, CAP_NET_ADMIN
))
810 if (nlmsg_len(nlh
) < sizeof(*r
))
814 if (r
->can_family
!= AF_CAN
)
815 return -EPFNOSUPPORT
;
817 /* so far we only support CAN -> CAN routings */
818 if (r
->gwtype
!= CGW_TYPE_CAN_CAN
)
821 gwj
= kmem_cache_alloc(cgw_cache
, GFP_KERNEL
);
825 gwj
->handled_frames
= 0;
826 gwj
->dropped_frames
= 0;
827 gwj
->deleted_frames
= 0;
828 gwj
->flags
= r
->flags
;
829 gwj
->gwtype
= r
->gwtype
;
831 err
= cgw_parse_attr(nlh
, &gwj
->mod
, CGW_TYPE_CAN_CAN
, &gwj
->ccgw
,
838 /* ifindex == 0 is not allowed for job creation */
839 if (!gwj
->ccgw
.src_idx
|| !gwj
->ccgw
.dst_idx
)
842 gwj
->src
.dev
= __dev_get_by_index(&init_net
, gwj
->ccgw
.src_idx
);
847 if (gwj
->src
.dev
->type
!= ARPHRD_CAN
)
850 gwj
->dst
.dev
= __dev_get_by_index(&init_net
, gwj
->ccgw
.dst_idx
);
855 if (gwj
->dst
.dev
->type
!= ARPHRD_CAN
)
858 gwj
->limit_hops
= limhops
;
862 err
= cgw_register_filter(gwj
);
864 hlist_add_head_rcu(&gwj
->list
, &cgw_list
);
867 kmem_cache_free(cgw_cache
, gwj
);
872 static void cgw_remove_all_jobs(void)
874 struct cgw_job
*gwj
= NULL
;
875 struct hlist_node
*nx
;
879 hlist_for_each_entry_safe(gwj
, nx
, &cgw_list
, list
) {
880 hlist_del(&gwj
->list
);
881 cgw_unregister_filter(gwj
);
882 kmem_cache_free(cgw_cache
, gwj
);
886 static int cgw_remove_job(struct sk_buff
*skb
, struct nlmsghdr
*nlh
)
888 struct cgw_job
*gwj
= NULL
;
889 struct hlist_node
*nx
;
892 struct can_can_gw ccgw
;
896 if (!netlink_capable(skb
, CAP_NET_ADMIN
))
899 if (nlmsg_len(nlh
) < sizeof(*r
))
903 if (r
->can_family
!= AF_CAN
)
904 return -EPFNOSUPPORT
;
906 /* so far we only support CAN -> CAN routings */
907 if (r
->gwtype
!= CGW_TYPE_CAN_CAN
)
910 err
= cgw_parse_attr(nlh
, &mod
, CGW_TYPE_CAN_CAN
, &ccgw
, &limhops
);
914 /* two interface indices both set to 0 => remove all entries */
915 if (!ccgw
.src_idx
&& !ccgw
.dst_idx
) {
916 cgw_remove_all_jobs();
924 /* remove only the first matching entry */
925 hlist_for_each_entry_safe(gwj
, nx
, &cgw_list
, list
) {
927 if (gwj
->flags
!= r
->flags
)
930 if (gwj
->limit_hops
!= limhops
)
933 if (memcmp(&gwj
->mod
, &mod
, sizeof(mod
)))
936 /* if (r->gwtype == CGW_TYPE_CAN_CAN) - is made sure here */
937 if (memcmp(&gwj
->ccgw
, &ccgw
, sizeof(ccgw
)))
940 hlist_del(&gwj
->list
);
941 cgw_unregister_filter(gwj
);
942 kmem_cache_free(cgw_cache
, gwj
);
950 static __init
int cgw_module_init(void)
952 /* sanitize given module parameter */
953 max_hops
= clamp_t(unsigned int, max_hops
, CGW_MIN_HOPS
, CGW_MAX_HOPS
);
955 pr_info("can: netlink gateway (rev " CAN_GW_VERSION
") max_hops=%d\n",
958 cgw_cache
= kmem_cache_create("can_gw", sizeof(struct cgw_job
),
965 notifier
.notifier_call
= cgw_notifier
;
966 register_netdevice_notifier(¬ifier
);
968 if (__rtnl_register(PF_CAN
, RTM_GETROUTE
, NULL
, cgw_dump_jobs
, NULL
)) {
969 unregister_netdevice_notifier(¬ifier
);
970 kmem_cache_destroy(cgw_cache
);
974 /* Only the first call to __rtnl_register can fail */
975 __rtnl_register(PF_CAN
, RTM_NEWROUTE
, cgw_create_job
, NULL
, NULL
);
976 __rtnl_register(PF_CAN
, RTM_DELROUTE
, cgw_remove_job
, NULL
, NULL
);
981 static __exit
void cgw_module_exit(void)
983 rtnl_unregister_all(PF_CAN
);
985 unregister_netdevice_notifier(¬ifier
);
988 cgw_remove_all_jobs();
991 rcu_barrier(); /* Wait for completion of call_rcu()'s */
993 kmem_cache_destroy(cgw_cache
);
996 module_init(cgw_module_init
);
997 module_exit(cgw_module_exit
);