3 #include <linux/mutex.h>
4 #include <linux/socket.h>
5 #include <linux/skbuff.h>
6 #include <net/netlink.h>
7 #include <net/net_namespace.h>
8 #include <linux/module.h>
10 #include <linux/kernel.h>
11 #include <linux/tcp.h>
12 #include <linux/workqueue.h>
14 #include <linux/inet_diag.h>
15 #include <linux/sock_diag.h>
17 static const struct sock_diag_handler
*sock_diag_handlers
[AF_MAX
];
18 static int (*inet_rcv_compat
)(struct sk_buff
*skb
, struct nlmsghdr
*nlh
);
19 static DEFINE_MUTEX(sock_diag_table_mutex
);
20 static struct workqueue_struct
*broadcast_wq
;
22 static u64
sock_gen_cookie(struct sock
*sk
)
25 u64 res
= atomic64_read(&sk
->sk_cookie
);
29 res
= atomic64_inc_return(&sock_net(sk
)->cookie_gen
);
30 atomic64_cmpxchg(&sk
->sk_cookie
, 0, res
);
34 int sock_diag_check_cookie(struct sock
*sk
, const __u32
*cookie
)
38 if (cookie
[0] == INET_DIAG_NOCOOKIE
&& cookie
[1] == INET_DIAG_NOCOOKIE
)
41 res
= sock_gen_cookie(sk
);
42 if ((u32
)res
!= cookie
[0] || (u32
)(res
>> 32) != cookie
[1])
47 EXPORT_SYMBOL_GPL(sock_diag_check_cookie
);
49 void sock_diag_save_cookie(struct sock
*sk
, __u32
*cookie
)
51 u64 res
= sock_gen_cookie(sk
);
54 cookie
[1] = (u32
)(res
>> 32);
56 EXPORT_SYMBOL_GPL(sock_diag_save_cookie
);
58 int sock_diag_put_meminfo(struct sock
*sk
, struct sk_buff
*skb
, int attrtype
)
60 u32 mem
[SK_MEMINFO_VARS
];
62 mem
[SK_MEMINFO_RMEM_ALLOC
] = sk_rmem_alloc_get(sk
);
63 mem
[SK_MEMINFO_RCVBUF
] = sk
->sk_rcvbuf
;
64 mem
[SK_MEMINFO_WMEM_ALLOC
] = sk_wmem_alloc_get(sk
);
65 mem
[SK_MEMINFO_SNDBUF
] = sk
->sk_sndbuf
;
66 mem
[SK_MEMINFO_FWD_ALLOC
] = sk
->sk_forward_alloc
;
67 mem
[SK_MEMINFO_WMEM_QUEUED
] = sk
->sk_wmem_queued
;
68 mem
[SK_MEMINFO_OPTMEM
] = atomic_read(&sk
->sk_omem_alloc
);
69 mem
[SK_MEMINFO_BACKLOG
] = sk
->sk_backlog
.len
;
70 mem
[SK_MEMINFO_DROPS
] = atomic_read(&sk
->sk_drops
);
72 return nla_put(skb
, attrtype
, sizeof(mem
), &mem
);
74 EXPORT_SYMBOL_GPL(sock_diag_put_meminfo
);
76 int sock_diag_put_filterinfo(bool may_report_filterinfo
, struct sock
*sk
,
77 struct sk_buff
*skb
, int attrtype
)
79 struct sock_fprog_kern
*fprog
;
80 struct sk_filter
*filter
;
85 if (!may_report_filterinfo
) {
86 nla_reserve(skb
, attrtype
, 0);
91 filter
= rcu_dereference(sk
->sk_filter
);
95 fprog
= filter
->prog
->orig_prog
;
99 flen
= bpf_classic_proglen(fprog
);
101 attr
= nla_reserve(skb
, attrtype
, flen
);
107 memcpy(nla_data(attr
), fprog
->filter
, flen
);
112 EXPORT_SYMBOL(sock_diag_put_filterinfo
);
114 struct broadcast_sk
{
116 struct work_struct work
;
119 static size_t sock_diag_nlmsg_size(void)
121 return NLMSG_ALIGN(sizeof(struct inet_diag_msg
)
122 + nla_total_size(sizeof(u8
)) /* INET_DIAG_PROTOCOL */
123 + nla_total_size_64bit(sizeof(struct tcp_info
))); /* INET_DIAG_INFO */
126 static void sock_diag_broadcast_destroy_work(struct work_struct
*work
)
128 struct broadcast_sk
*bsk
=
129 container_of(work
, struct broadcast_sk
, work
);
130 struct sock
*sk
= bsk
->sk
;
131 const struct sock_diag_handler
*hndl
;
133 const enum sknetlink_groups group
= sock_diag_destroy_group(sk
);
136 WARN_ON(group
== SKNLGRP_NONE
);
138 skb
= nlmsg_new(sock_diag_nlmsg_size(), GFP_KERNEL
);
142 mutex_lock(&sock_diag_table_mutex
);
143 hndl
= sock_diag_handlers
[sk
->sk_family
];
144 if (hndl
&& hndl
->get_info
)
145 err
= hndl
->get_info(skb
, sk
);
146 mutex_unlock(&sock_diag_table_mutex
);
149 nlmsg_multicast(sock_net(sk
)->diag_nlsk
, skb
, 0, group
,
158 void sock_diag_broadcast_destroy(struct sock
*sk
)
160 /* Note, this function is often called from an interrupt context. */
161 struct broadcast_sk
*bsk
=
162 kmalloc(sizeof(struct broadcast_sk
), GFP_ATOMIC
);
164 return sk_destruct(sk
);
166 INIT_WORK(&bsk
->work
, sock_diag_broadcast_destroy_work
);
167 queue_work(broadcast_wq
, &bsk
->work
);
170 void sock_diag_register_inet_compat(int (*fn
)(struct sk_buff
*skb
, struct nlmsghdr
*nlh
))
172 mutex_lock(&sock_diag_table_mutex
);
173 inet_rcv_compat
= fn
;
174 mutex_unlock(&sock_diag_table_mutex
);
176 EXPORT_SYMBOL_GPL(sock_diag_register_inet_compat
);
178 void sock_diag_unregister_inet_compat(int (*fn
)(struct sk_buff
*skb
, struct nlmsghdr
*nlh
))
180 mutex_lock(&sock_diag_table_mutex
);
181 inet_rcv_compat
= NULL
;
182 mutex_unlock(&sock_diag_table_mutex
);
184 EXPORT_SYMBOL_GPL(sock_diag_unregister_inet_compat
);
186 int sock_diag_register(const struct sock_diag_handler
*hndl
)
190 if (hndl
->family
>= AF_MAX
)
193 mutex_lock(&sock_diag_table_mutex
);
194 if (sock_diag_handlers
[hndl
->family
])
197 sock_diag_handlers
[hndl
->family
] = hndl
;
198 mutex_unlock(&sock_diag_table_mutex
);
202 EXPORT_SYMBOL_GPL(sock_diag_register
);
204 void sock_diag_unregister(const struct sock_diag_handler
*hnld
)
206 int family
= hnld
->family
;
208 if (family
>= AF_MAX
)
211 mutex_lock(&sock_diag_table_mutex
);
212 BUG_ON(sock_diag_handlers
[family
] != hnld
);
213 sock_diag_handlers
[family
] = NULL
;
214 mutex_unlock(&sock_diag_table_mutex
);
216 EXPORT_SYMBOL_GPL(sock_diag_unregister
);
218 static int __sock_diag_cmd(struct sk_buff
*skb
, struct nlmsghdr
*nlh
)
221 struct sock_diag_req
*req
= nlmsg_data(nlh
);
222 const struct sock_diag_handler
*hndl
;
224 if (nlmsg_len(nlh
) < sizeof(*req
))
227 if (req
->sdiag_family
>= AF_MAX
)
230 if (sock_diag_handlers
[req
->sdiag_family
] == NULL
)
231 request_module("net-pf-%d-proto-%d-type-%d", PF_NETLINK
,
232 NETLINK_SOCK_DIAG
, req
->sdiag_family
);
234 mutex_lock(&sock_diag_table_mutex
);
235 hndl
= sock_diag_handlers
[req
->sdiag_family
];
238 else if (nlh
->nlmsg_type
== SOCK_DIAG_BY_FAMILY
)
239 err
= hndl
->dump(skb
, nlh
);
240 else if (nlh
->nlmsg_type
== SOCK_DESTROY
&& hndl
->destroy
)
241 err
= hndl
->destroy(skb
, nlh
);
244 mutex_unlock(&sock_diag_table_mutex
);
249 static int sock_diag_rcv_msg(struct sk_buff
*skb
, struct nlmsghdr
*nlh
)
253 switch (nlh
->nlmsg_type
) {
254 case TCPDIAG_GETSOCK
:
255 case DCCPDIAG_GETSOCK
:
256 if (inet_rcv_compat
== NULL
)
257 request_module("net-pf-%d-proto-%d-type-%d", PF_NETLINK
,
258 NETLINK_SOCK_DIAG
, AF_INET
);
260 mutex_lock(&sock_diag_table_mutex
);
261 if (inet_rcv_compat
!= NULL
)
262 ret
= inet_rcv_compat(skb
, nlh
);
265 mutex_unlock(&sock_diag_table_mutex
);
268 case SOCK_DIAG_BY_FAMILY
:
270 return __sock_diag_cmd(skb
, nlh
);
276 static DEFINE_MUTEX(sock_diag_mutex
);
278 static void sock_diag_rcv(struct sk_buff
*skb
)
280 mutex_lock(&sock_diag_mutex
);
281 netlink_rcv_skb(skb
, &sock_diag_rcv_msg
);
282 mutex_unlock(&sock_diag_mutex
);
285 static int sock_diag_bind(struct net
*net
, int group
)
288 case SKNLGRP_INET_TCP_DESTROY
:
289 case SKNLGRP_INET_UDP_DESTROY
:
290 if (!sock_diag_handlers
[AF_INET
])
291 request_module("net-pf-%d-proto-%d-type-%d", PF_NETLINK
,
292 NETLINK_SOCK_DIAG
, AF_INET
);
294 case SKNLGRP_INET6_TCP_DESTROY
:
295 case SKNLGRP_INET6_UDP_DESTROY
:
296 if (!sock_diag_handlers
[AF_INET6
])
297 request_module("net-pf-%d-proto-%d-type-%d", PF_NETLINK
,
298 NETLINK_SOCK_DIAG
, AF_INET
);
304 int sock_diag_destroy(struct sock
*sk
, int err
)
306 if (!ns_capable(sock_net(sk
)->user_ns
, CAP_NET_ADMIN
))
309 if (!sk
->sk_prot
->diag_destroy
)
312 return sk
->sk_prot
->diag_destroy(sk
, err
);
314 EXPORT_SYMBOL_GPL(sock_diag_destroy
);
316 static int __net_init
diag_net_init(struct net
*net
)
318 struct netlink_kernel_cfg cfg
= {
319 .groups
= SKNLGRP_MAX
,
320 .input
= sock_diag_rcv
,
321 .bind
= sock_diag_bind
,
322 .flags
= NL_CFG_F_NONROOT_RECV
,
325 net
->diag_nlsk
= netlink_kernel_create(net
, NETLINK_SOCK_DIAG
, &cfg
);
326 return net
->diag_nlsk
== NULL
? -ENOMEM
: 0;
329 static void __net_exit
diag_net_exit(struct net
*net
)
331 netlink_kernel_release(net
->diag_nlsk
);
332 net
->diag_nlsk
= NULL
;
335 static struct pernet_operations diag_net_ops
= {
336 .init
= diag_net_init
,
337 .exit
= diag_net_exit
,
340 static int __init
sock_diag_init(void)
342 broadcast_wq
= alloc_workqueue("sock_diag_events", 0, 0);
343 BUG_ON(!broadcast_wq
);
344 return register_pernet_subsys(&diag_net_ops
);
346 device_initcall(sock_diag_init
);