2 * net/sched/sch_cbs.c Credit Based Shaper
4 * This program is free software; you can redistribute it and/or
5 * modify it under the terms of the GNU General Public License
6 * as published by the Free Software Foundation; either version
7 * 2 of the License, or (at your option) any later version.
9 * Authors: Vinicius Costa Gomes <vinicius.gomes@intel.com>
13 /* Credit Based Shaper (CBS)
14 * =========================
16 * This is a simple rate-limiting shaper aimed at TSN applications on
17 * systems with known traffic workloads.
19 * Its algorithm is defined by the IEEE 802.1Q-2014 Specification,
20 * Section 8.6.8.2, and explained in more detail in the Annex L of the
23 * There are four tunables to be considered:
25 * 'idleslope': Idleslope is the rate of credits that is
26 * accumulated (in kilobits per second) when there is at least
27 * one packet waiting for transmission. Packets are transmitted
28 * when the current value of credits is equal or greater than
29 * zero. When there is no packet to be transmitted the amount of
30 * credits is set to zero. This is the main tunable of the CBS
34 * Sendslope is the rate of credits that is depleted (it should be a
35 * negative number of kilobits per second) when a transmission is
36 * ocurring. It can be calculated as follows, (IEEE 802.1Q-2014 Section
39 * sendslope = idleslope - port_transmit_rate
41 * 'hicredit': Hicredit defines the maximum amount of credits (in
42 * bytes) that can be accumulated. Hicredit depends on the
43 * characteristics of interfering traffic,
44 * 'max_interference_size' is the maximum size of any burst of
45 * traffic that can delay the transmission of a frame that is
46 * available for transmission for this traffic class, (IEEE
47 * 802.1Q-2014 Annex L, Equation L-3):
49 * hicredit = max_interference_size * (idleslope / port_transmit_rate)
51 * 'locredit': Locredit is the minimum amount of credits that can
52 * be reached. It is a function of the traffic flowing through
53 * this qdisc (IEEE 802.1Q-2014 Annex L, Equation L-2):
55 * locredit = max_frame_size * (sendslope / port_transmit_rate)
58 #include <linux/module.h>
59 #include <linux/types.h>
60 #include <linux/kernel.h>
61 #include <linux/string.h>
62 #include <linux/errno.h>
63 #include <linux/skbuff.h>
64 #include <net/netevent.h>
65 #include <net/netlink.h>
66 #include <net/sch_generic.h>
67 #include <net/pkt_sched.h>
69 static LIST_HEAD(cbs_list
);
70 static DEFINE_SPINLOCK(cbs_list_lock
);
72 #define BYTES_PER_KBIT (1000LL / 8)
74 struct cbs_sched_data
{
77 atomic64_t port_rate
; /* in bytes/s */
78 s64 last
; /* timestamp in ns */
79 s64 credits
; /* in bytes */
80 s32 locredit
; /* in bytes */
81 s32 hicredit
; /* in bytes */
82 s64 sendslope
; /* in bytes/s */
83 s64 idleslope
; /* in bytes/s */
84 struct qdisc_watchdog watchdog
;
85 int (*enqueue
)(struct sk_buff
*skb
, struct Qdisc
*sch
,
86 struct sk_buff
**to_free
);
87 struct sk_buff
*(*dequeue
)(struct Qdisc
*sch
);
89 struct list_head cbs_list
;
92 static int cbs_child_enqueue(struct sk_buff
*skb
, struct Qdisc
*sch
,
94 struct sk_buff
**to_free
)
98 err
= child
->ops
->enqueue(skb
, child
, to_free
);
99 if (err
!= NET_XMIT_SUCCESS
)
102 qdisc_qstats_backlog_inc(sch
, skb
);
105 return NET_XMIT_SUCCESS
;
108 static int cbs_enqueue_offload(struct sk_buff
*skb
, struct Qdisc
*sch
,
109 struct sk_buff
**to_free
)
111 struct cbs_sched_data
*q
= qdisc_priv(sch
);
112 struct Qdisc
*qdisc
= q
->qdisc
;
114 return cbs_child_enqueue(skb
, sch
, qdisc
, to_free
);
117 static int cbs_enqueue_soft(struct sk_buff
*skb
, struct Qdisc
*sch
,
118 struct sk_buff
**to_free
)
120 struct cbs_sched_data
*q
= qdisc_priv(sch
);
121 struct Qdisc
*qdisc
= q
->qdisc
;
123 if (sch
->q
.qlen
== 0 && q
->credits
> 0) {
124 /* We need to stop accumulating credits when there's
125 * no enqueued packets and q->credits is positive.
128 q
->last
= ktime_get_ns();
131 return cbs_child_enqueue(skb
, sch
, qdisc
, to_free
);
134 static int cbs_enqueue(struct sk_buff
*skb
, struct Qdisc
*sch
,
135 struct sk_buff
**to_free
)
137 struct cbs_sched_data
*q
= qdisc_priv(sch
);
139 return q
->enqueue(skb
, sch
, to_free
);
142 /* timediff is in ns, slope is in bytes/s */
143 static s64
timediff_to_credits(s64 timediff
, s64 slope
)
145 return div64_s64(timediff
* slope
, NSEC_PER_SEC
);
148 static s64
delay_from_credits(s64 credits
, s64 slope
)
150 if (unlikely(slope
== 0))
153 return div64_s64(-credits
* NSEC_PER_SEC
, slope
);
156 static s64
credits_from_len(unsigned int len
, s64 slope
, s64 port_rate
)
158 if (unlikely(port_rate
== 0))
161 return div64_s64(len
* slope
, port_rate
);
164 static struct sk_buff
*cbs_child_dequeue(struct Qdisc
*sch
, struct Qdisc
*child
)
168 skb
= child
->ops
->dequeue(child
);
172 qdisc_qstats_backlog_dec(sch
, skb
);
173 qdisc_bstats_update(sch
, skb
);
179 static struct sk_buff
*cbs_dequeue_soft(struct Qdisc
*sch
)
181 struct cbs_sched_data
*q
= qdisc_priv(sch
);
182 struct Qdisc
*qdisc
= q
->qdisc
;
183 s64 now
= ktime_get_ns();
188 /* The previous packet is still being sent */
190 qdisc_watchdog_schedule_ns(&q
->watchdog
, q
->last
);
193 if (q
->credits
< 0) {
194 credits
= timediff_to_credits(now
- q
->last
, q
->idleslope
);
196 credits
= q
->credits
+ credits
;
197 q
->credits
= min_t(s64
, credits
, q
->hicredit
);
199 if (q
->credits
< 0) {
202 delay
= delay_from_credits(q
->credits
, q
->idleslope
);
203 qdisc_watchdog_schedule_ns(&q
->watchdog
, now
+ delay
);
210 skb
= cbs_child_dequeue(sch
, qdisc
);
214 len
= qdisc_pkt_len(skb
);
216 /* As sendslope is a negative number, this will decrease the
217 * amount of q->credits.
219 credits
= credits_from_len(len
, q
->sendslope
,
220 atomic64_read(&q
->port_rate
));
221 credits
+= q
->credits
;
223 q
->credits
= max_t(s64
, credits
, q
->locredit
);
224 /* Estimate of the transmission of the last byte of the packet in ns */
225 if (unlikely(atomic64_read(&q
->port_rate
) == 0))
228 q
->last
= now
+ div64_s64(len
* NSEC_PER_SEC
,
229 atomic64_read(&q
->port_rate
));
234 static struct sk_buff
*cbs_dequeue_offload(struct Qdisc
*sch
)
236 struct cbs_sched_data
*q
= qdisc_priv(sch
);
237 struct Qdisc
*qdisc
= q
->qdisc
;
239 return cbs_child_dequeue(sch
, qdisc
);
242 static struct sk_buff
*cbs_dequeue(struct Qdisc
*sch
)
244 struct cbs_sched_data
*q
= qdisc_priv(sch
);
246 return q
->dequeue(sch
);
249 static const struct nla_policy cbs_policy
[TCA_CBS_MAX
+ 1] = {
250 [TCA_CBS_PARMS
] = { .len
= sizeof(struct tc_cbs_qopt
) },
253 static void cbs_disable_offload(struct net_device
*dev
,
254 struct cbs_sched_data
*q
)
256 struct tc_cbs_qopt_offload cbs
= { };
257 const struct net_device_ops
*ops
;
263 q
->enqueue
= cbs_enqueue_soft
;
264 q
->dequeue
= cbs_dequeue_soft
;
266 ops
= dev
->netdev_ops
;
267 if (!ops
->ndo_setup_tc
)
270 cbs
.queue
= q
->queue
;
273 err
= ops
->ndo_setup_tc(dev
, TC_SETUP_QDISC_CBS
, &cbs
);
275 pr_warn("Couldn't disable CBS offload for queue %d\n",
279 static int cbs_enable_offload(struct net_device
*dev
, struct cbs_sched_data
*q
,
280 const struct tc_cbs_qopt
*opt
,
281 struct netlink_ext_ack
*extack
)
283 const struct net_device_ops
*ops
= dev
->netdev_ops
;
284 struct tc_cbs_qopt_offload cbs
= { };
287 if (!ops
->ndo_setup_tc
) {
288 NL_SET_ERR_MSG(extack
, "Specified device does not support cbs offload");
292 cbs
.queue
= q
->queue
;
295 cbs
.hicredit
= opt
->hicredit
;
296 cbs
.locredit
= opt
->locredit
;
297 cbs
.idleslope
= opt
->idleslope
;
298 cbs
.sendslope
= opt
->sendslope
;
300 err
= ops
->ndo_setup_tc(dev
, TC_SETUP_QDISC_CBS
, &cbs
);
302 NL_SET_ERR_MSG(extack
, "Specified device failed to setup cbs hardware offload");
306 q
->enqueue
= cbs_enqueue_offload
;
307 q
->dequeue
= cbs_dequeue_offload
;
312 static void cbs_set_port_rate(struct net_device
*dev
, struct cbs_sched_data
*q
)
314 struct ethtool_link_ksettings ecmd
;
315 int speed
= SPEED_10
;
319 err
= __ethtool_get_link_ksettings(dev
, &ecmd
);
323 if (ecmd
.base
.speed
&& ecmd
.base
.speed
!= SPEED_UNKNOWN
)
324 speed
= ecmd
.base
.speed
;
327 port_rate
= speed
* 1000 * BYTES_PER_KBIT
;
329 atomic64_set(&q
->port_rate
, port_rate
);
330 netdev_dbg(dev
, "cbs: set %s's port_rate to: %lld, linkspeed: %d\n",
331 dev
->name
, (long long)atomic64_read(&q
->port_rate
),
335 static int cbs_dev_notifier(struct notifier_block
*nb
, unsigned long event
,
338 struct net_device
*dev
= netdev_notifier_info_to_dev(ptr
);
339 struct cbs_sched_data
*q
;
340 struct net_device
*qdev
;
345 if (event
!= NETDEV_UP
&& event
!= NETDEV_CHANGE
)
348 spin_lock(&cbs_list_lock
);
349 list_for_each_entry(q
, &cbs_list
, cbs_list
) {
350 qdev
= qdisc_dev(q
->qdisc
);
356 spin_unlock(&cbs_list_lock
);
359 cbs_set_port_rate(dev
, q
);
364 static int cbs_change(struct Qdisc
*sch
, struct nlattr
*opt
,
365 struct netlink_ext_ack
*extack
)
367 struct cbs_sched_data
*q
= qdisc_priv(sch
);
368 struct net_device
*dev
= qdisc_dev(sch
);
369 struct nlattr
*tb
[TCA_CBS_MAX
+ 1];
370 struct tc_cbs_qopt
*qopt
;
373 err
= nla_parse_nested(tb
, TCA_CBS_MAX
, opt
, cbs_policy
, extack
);
377 if (!tb
[TCA_CBS_PARMS
]) {
378 NL_SET_ERR_MSG(extack
, "Missing CBS parameter which are mandatory");
382 qopt
= nla_data(tb
[TCA_CBS_PARMS
]);
384 if (!qopt
->offload
) {
385 cbs_set_port_rate(dev
, q
);
386 cbs_disable_offload(dev
, q
);
388 err
= cbs_enable_offload(dev
, q
, qopt
, extack
);
393 /* Everything went OK, save the parameters used. */
394 q
->hicredit
= qopt
->hicredit
;
395 q
->locredit
= qopt
->locredit
;
396 q
->idleslope
= qopt
->idleslope
* BYTES_PER_KBIT
;
397 q
->sendslope
= qopt
->sendslope
* BYTES_PER_KBIT
;
398 q
->offload
= qopt
->offload
;
403 static int cbs_init(struct Qdisc
*sch
, struct nlattr
*opt
,
404 struct netlink_ext_ack
*extack
)
406 struct cbs_sched_data
*q
= qdisc_priv(sch
);
407 struct net_device
*dev
= qdisc_dev(sch
);
411 NL_SET_ERR_MSG(extack
, "Missing CBS qdisc options which are mandatory");
415 q
->qdisc
= qdisc_create_dflt(sch
->dev_queue
, &pfifo_qdisc_ops
,
416 sch
->handle
, extack
);
420 qdisc_hash_add(q
->qdisc
, false);
422 q
->queue
= sch
->dev_queue
- netdev_get_tx_queue(dev
, 0);
424 q
->enqueue
= cbs_enqueue_soft
;
425 q
->dequeue
= cbs_dequeue_soft
;
427 qdisc_watchdog_init(&q
->watchdog
, sch
);
429 err
= cbs_change(sch
, opt
, extack
);
434 spin_lock(&cbs_list_lock
);
435 list_add(&q
->cbs_list
, &cbs_list
);
436 spin_unlock(&cbs_list_lock
);
442 static void cbs_destroy(struct Qdisc
*sch
)
444 struct cbs_sched_data
*q
= qdisc_priv(sch
);
445 struct net_device
*dev
= qdisc_dev(sch
);
447 spin_lock(&cbs_list_lock
);
448 list_del(&q
->cbs_list
);
449 spin_unlock(&cbs_list_lock
);
451 qdisc_watchdog_cancel(&q
->watchdog
);
452 cbs_disable_offload(dev
, q
);
455 qdisc_destroy(q
->qdisc
);
458 static int cbs_dump(struct Qdisc
*sch
, struct sk_buff
*skb
)
460 struct cbs_sched_data
*q
= qdisc_priv(sch
);
461 struct tc_cbs_qopt opt
= { };
464 nest
= nla_nest_start(skb
, TCA_OPTIONS
);
466 goto nla_put_failure
;
468 opt
.hicredit
= q
->hicredit
;
469 opt
.locredit
= q
->locredit
;
470 opt
.sendslope
= div64_s64(q
->sendslope
, BYTES_PER_KBIT
);
471 opt
.idleslope
= div64_s64(q
->idleslope
, BYTES_PER_KBIT
);
472 opt
.offload
= q
->offload
;
474 if (nla_put(skb
, TCA_CBS_PARMS
, sizeof(opt
), &opt
))
475 goto nla_put_failure
;
477 return nla_nest_end(skb
, nest
);
480 nla_nest_cancel(skb
, nest
);
484 static int cbs_dump_class(struct Qdisc
*sch
, unsigned long cl
,
485 struct sk_buff
*skb
, struct tcmsg
*tcm
)
487 struct cbs_sched_data
*q
= qdisc_priv(sch
);
489 if (cl
!= 1 || !q
->qdisc
) /* only one class */
492 tcm
->tcm_handle
|= TC_H_MIN(1);
493 tcm
->tcm_info
= q
->qdisc
->handle
;
498 static int cbs_graft(struct Qdisc
*sch
, unsigned long arg
, struct Qdisc
*new,
499 struct Qdisc
**old
, struct netlink_ext_ack
*extack
)
501 struct cbs_sched_data
*q
= qdisc_priv(sch
);
504 new = qdisc_create_dflt(sch
->dev_queue
, &pfifo_qdisc_ops
,
510 *old
= qdisc_replace(sch
, new, &q
->qdisc
);
514 static struct Qdisc
*cbs_leaf(struct Qdisc
*sch
, unsigned long arg
)
516 struct cbs_sched_data
*q
= qdisc_priv(sch
);
521 static unsigned long cbs_find(struct Qdisc
*sch
, u32 classid
)
526 static void cbs_walk(struct Qdisc
*sch
, struct qdisc_walker
*walker
)
529 if (walker
->count
>= walker
->skip
) {
530 if (walker
->fn(sch
, 1, walker
) < 0) {
539 static const struct Qdisc_class_ops cbs_class_ops
= {
544 .dump
= cbs_dump_class
,
547 static struct Qdisc_ops cbs_qdisc_ops __read_mostly
= {
549 .cl_ops
= &cbs_class_ops
,
550 .priv_size
= sizeof(struct cbs_sched_data
),
551 .enqueue
= cbs_enqueue
,
552 .dequeue
= cbs_dequeue
,
553 .peek
= qdisc_peek_dequeued
,
555 .reset
= qdisc_reset_queue
,
556 .destroy
= cbs_destroy
,
557 .change
= cbs_change
,
559 .owner
= THIS_MODULE
,
562 static struct notifier_block cbs_device_notifier
= {
563 .notifier_call
= cbs_dev_notifier
,
566 static int __init
cbs_module_init(void)
570 err
= register_netdevice_notifier(&cbs_device_notifier
);
574 err
= register_qdisc(&cbs_qdisc_ops
);
576 unregister_netdevice_notifier(&cbs_device_notifier
);
581 static void __exit
cbs_module_exit(void)
583 unregister_qdisc(&cbs_qdisc_ops
);
584 unregister_netdevice_notifier(&cbs_device_notifier
);
586 module_init(cbs_module_init
)
587 module_exit(cbs_module_exit
)
588 MODULE_LICENSE("GPL");