.. | .. |
---|
| 1 | +// SPDX-License-Identifier: GPL-2.0-or-later |
---|
1 | 2 | /* |
---|
2 | 3 | * net/sched/sch_red.c Random Early Detection queue. |
---|
3 | | - * |
---|
4 | | - * This program is free software; you can redistribute it and/or |
---|
5 | | - * modify it under the terms of the GNU General Public License |
---|
6 | | - * as published by the Free Software Foundation; either version |
---|
7 | | - * 2 of the License, or (at your option) any later version. |
---|
8 | 4 | * |
---|
9 | 5 | * Authors: Alexey Kuznetsov, <kuznet@ms2.inr.ac.ru> |
---|
10 | 6 | * |
---|
.. | .. |
---|
39 | 35 | |
---|
40 | 36 | struct red_sched_data { |
---|
41 | 37 | u32 limit; /* HARD maximal queue length */ |
---|
| 38 | + |
---|
42 | 39 | unsigned char flags; |
---|
| 40 | + /* Non-flags in tc_red_qopt.flags. */ |
---|
| 41 | + unsigned char userbits; |
---|
| 42 | + |
---|
43 | 43 | struct timer_list adapt_timer; |
---|
44 | 44 | struct Qdisc *sch; |
---|
45 | 45 | struct red_parms parms; |
---|
46 | 46 | struct red_vars vars; |
---|
47 | 47 | struct red_stats stats; |
---|
48 | 48 | struct Qdisc *qdisc; |
---|
| 49 | + struct tcf_qevent qe_early_drop; |
---|
| 50 | + struct tcf_qevent qe_mark; |
---|
49 | 51 | }; |
---|
| 52 | + |
---|
| 53 | +#define TC_RED_SUPPORTED_FLAGS (TC_RED_HISTORIC_FLAGS | TC_RED_NODROP) |
---|
50 | 54 | |
---|
51 | 55 | static inline int red_use_ecn(struct red_sched_data *q) |
---|
52 | 56 | { |
---|
.. | .. |
---|
58 | 62 | return q->flags & TC_RED_HARDDROP; |
---|
59 | 63 | } |
---|
60 | 64 | |
---|
| 65 | +static int red_use_nodrop(struct red_sched_data *q) |
---|
| 66 | +{ |
---|
| 67 | + return q->flags & TC_RED_NODROP; |
---|
| 68 | +} |
---|
| 69 | + |
---|
61 | 70 | static int red_enqueue(struct sk_buff *skb, struct Qdisc *sch, |
---|
62 | 71 | struct sk_buff **to_free) |
---|
63 | 72 | { |
---|
64 | 73 | struct red_sched_data *q = qdisc_priv(sch); |
---|
65 | 74 | struct Qdisc *child = q->qdisc; |
---|
| 75 | + unsigned int len; |
---|
66 | 76 | int ret; |
---|
67 | 77 | |
---|
68 | 78 | q->vars.qavg = red_calc_qavg(&q->parms, |
---|
.. | .. |
---|
78 | 88 | |
---|
79 | 89 | case RED_PROB_MARK: |
---|
80 | 90 | qdisc_qstats_overlimit(sch); |
---|
81 | | - if (!red_use_ecn(q) || !INET_ECN_set_ce(skb)) { |
---|
| 91 | + if (!red_use_ecn(q)) { |
---|
82 | 92 | q->stats.prob_drop++; |
---|
83 | 93 | goto congestion_drop; |
---|
84 | 94 | } |
---|
85 | 95 | |
---|
86 | | - q->stats.prob_mark++; |
---|
| 96 | + if (INET_ECN_set_ce(skb)) { |
---|
| 97 | + q->stats.prob_mark++; |
---|
| 98 | + skb = tcf_qevent_handle(&q->qe_mark, sch, skb, to_free, &ret); |
---|
| 99 | + if (!skb) |
---|
| 100 | + return NET_XMIT_CN | ret; |
---|
| 101 | + } else if (!red_use_nodrop(q)) { |
---|
| 102 | + q->stats.prob_drop++; |
---|
| 103 | + goto congestion_drop; |
---|
| 104 | + } |
---|
| 105 | + |
---|
| 106 | + /* Non-ECT packet in ECN nodrop mode: queue it. */ |
---|
87 | 107 | break; |
---|
88 | 108 | |
---|
89 | 109 | case RED_HARD_MARK: |
---|
90 | 110 | qdisc_qstats_overlimit(sch); |
---|
91 | | - if (red_use_harddrop(q) || !red_use_ecn(q) || |
---|
92 | | - !INET_ECN_set_ce(skb)) { |
---|
| 111 | + if (red_use_harddrop(q) || !red_use_ecn(q)) { |
---|
93 | 112 | q->stats.forced_drop++; |
---|
94 | 113 | goto congestion_drop; |
---|
95 | 114 | } |
---|
96 | 115 | |
---|
97 | | - q->stats.forced_mark++; |
---|
| 116 | + if (INET_ECN_set_ce(skb)) { |
---|
| 117 | + q->stats.forced_mark++; |
---|
| 118 | + skb = tcf_qevent_handle(&q->qe_mark, sch, skb, to_free, &ret); |
---|
| 119 | + if (!skb) |
---|
| 120 | + return NET_XMIT_CN | ret; |
---|
| 121 | + } else if (!red_use_nodrop(q)) { |
---|
| 122 | + q->stats.forced_drop++; |
---|
| 123 | + goto congestion_drop; |
---|
| 124 | + } |
---|
| 125 | + |
---|
| 126 | + /* Non-ECT packet in ECN nodrop mode: queue it. */ |
---|
98 | 127 | break; |
---|
99 | 128 | } |
---|
100 | 129 | |
---|
| 130 | + len = qdisc_pkt_len(skb); |
---|
101 | 131 | ret = qdisc_enqueue(skb, child, to_free); |
---|
102 | 132 | if (likely(ret == NET_XMIT_SUCCESS)) { |
---|
103 | | - qdisc_qstats_backlog_inc(sch, skb); |
---|
| 133 | + sch->qstats.backlog += len; |
---|
104 | 134 | sch->q.qlen++; |
---|
105 | 135 | } else if (net_xmit_drop_count(ret)) { |
---|
106 | 136 | q->stats.pdrop++; |
---|
.. | .. |
---|
109 | 139 | return ret; |
---|
110 | 140 | |
---|
111 | 141 | congestion_drop: |
---|
| 142 | + skb = tcf_qevent_handle(&q->qe_early_drop, sch, skb, to_free, &ret); |
---|
| 143 | + if (!skb) |
---|
| 144 | + return NET_XMIT_CN | ret; |
---|
| 145 | + |
---|
112 | 146 | qdisc_drop(skb, sch, to_free); |
---|
113 | 147 | return NET_XMIT_CN; |
---|
114 | 148 | } |
---|
.. | .. |
---|
144 | 178 | struct red_sched_data *q = qdisc_priv(sch); |
---|
145 | 179 | |
---|
146 | 180 | qdisc_reset(q->qdisc); |
---|
147 | | - sch->qstats.backlog = 0; |
---|
148 | | - sch->q.qlen = 0; |
---|
149 | 181 | red_restart(&q->vars); |
---|
150 | 182 | } |
---|
151 | 183 | |
---|
.. | .. |
---|
166 | 198 | opt.set.min = q->parms.qth_min >> q->parms.Wlog; |
---|
167 | 199 | opt.set.max = q->parms.qth_max >> q->parms.Wlog; |
---|
168 | 200 | opt.set.probability = q->parms.max_P; |
---|
| 201 | + opt.set.limit = q->limit; |
---|
169 | 202 | opt.set.is_ecn = red_use_ecn(q); |
---|
| 203 | + opt.set.is_harddrop = red_use_harddrop(q); |
---|
| 204 | + opt.set.is_nodrop = red_use_nodrop(q); |
---|
170 | 205 | opt.set.qstats = &sch->qstats; |
---|
171 | 206 | } else { |
---|
172 | 207 | opt.command = TC_RED_DESTROY; |
---|
.. | .. |
---|
179 | 214 | { |
---|
180 | 215 | struct red_sched_data *q = qdisc_priv(sch); |
---|
181 | 216 | |
---|
| 217 | + tcf_qevent_destroy(&q->qe_mark, sch); |
---|
| 218 | + tcf_qevent_destroy(&q->qe_early_drop, sch); |
---|
182 | 219 | del_timer_sync(&q->adapt_timer); |
---|
183 | 220 | red_offload(sch, false); |
---|
184 | 221 | qdisc_put(q->qdisc); |
---|
185 | 222 | } |
---|
186 | 223 | |
---|
187 | 224 | static const struct nla_policy red_policy[TCA_RED_MAX + 1] = { |
---|
| 225 | + [TCA_RED_UNSPEC] = { .strict_start_type = TCA_RED_FLAGS }, |
---|
188 | 226 | [TCA_RED_PARMS] = { .len = sizeof(struct tc_red_qopt) }, |
---|
189 | 227 | [TCA_RED_STAB] = { .len = RED_STAB_SIZE }, |
---|
190 | 228 | [TCA_RED_MAX_P] = { .type = NLA_U32 }, |
---|
| 229 | + [TCA_RED_FLAGS] = NLA_POLICY_BITFIELD32(TC_RED_SUPPORTED_FLAGS), |
---|
| 230 | + [TCA_RED_EARLY_DROP_BLOCK] = { .type = NLA_U32 }, |
---|
| 231 | + [TCA_RED_MARK_BLOCK] = { .type = NLA_U32 }, |
---|
191 | 232 | }; |
---|
192 | 233 | |
---|
193 | | -static int red_change(struct Qdisc *sch, struct nlattr *opt, |
---|
194 | | - struct netlink_ext_ack *extack) |
---|
| 234 | +static int __red_change(struct Qdisc *sch, struct nlattr **tb, |
---|
| 235 | + struct netlink_ext_ack *extack) |
---|
195 | 236 | { |
---|
| 237 | + struct Qdisc *old_child = NULL, *child = NULL; |
---|
196 | 238 | struct red_sched_data *q = qdisc_priv(sch); |
---|
197 | | - struct nlattr *tb[TCA_RED_MAX + 1]; |
---|
| 239 | + struct nla_bitfield32 flags_bf; |
---|
198 | 240 | struct tc_red_qopt *ctl; |
---|
199 | | - struct Qdisc *child = NULL; |
---|
| 241 | + unsigned char userbits; |
---|
| 242 | + unsigned char flags; |
---|
200 | 243 | int err; |
---|
201 | 244 | u32 max_P; |
---|
202 | 245 | u8 *stab; |
---|
203 | | - |
---|
204 | | - if (opt == NULL) |
---|
205 | | - return -EINVAL; |
---|
206 | | - |
---|
207 | | - err = nla_parse_nested(tb, TCA_RED_MAX, opt, red_policy, NULL); |
---|
208 | | - if (err < 0) |
---|
209 | | - return err; |
---|
210 | 246 | |
---|
211 | 247 | if (tb[TCA_RED_PARMS] == NULL || |
---|
212 | 248 | tb[TCA_RED_STAB] == NULL) |
---|
.. | .. |
---|
220 | 256 | ctl->Scell_log, stab)) |
---|
221 | 257 | return -EINVAL; |
---|
222 | 258 | |
---|
| 259 | + err = red_get_flags(ctl->flags, TC_RED_HISTORIC_FLAGS, |
---|
| 260 | + tb[TCA_RED_FLAGS], TC_RED_SUPPORTED_FLAGS, |
---|
| 261 | + &flags_bf, &userbits, extack); |
---|
| 262 | + if (err) |
---|
| 263 | + return err; |
---|
| 264 | + |
---|
223 | 265 | if (ctl->limit > 0) { |
---|
224 | 266 | child = fifo_create_dflt(sch, &bfifo_qdisc_ops, ctl->limit, |
---|
225 | 267 | extack); |
---|
.. | .. |
---|
231 | 273 | } |
---|
232 | 274 | |
---|
233 | 275 | sch_tree_lock(sch); |
---|
234 | | - q->flags = ctl->flags; |
---|
| 276 | + |
---|
| 277 | + flags = (q->flags & ~flags_bf.selector) | flags_bf.value; |
---|
| 278 | + err = red_validate_flags(flags, extack); |
---|
| 279 | + if (err) |
---|
| 280 | + goto unlock_out; |
---|
| 281 | + |
---|
| 282 | + q->flags = flags; |
---|
| 283 | + q->userbits = userbits; |
---|
235 | 284 | q->limit = ctl->limit; |
---|
236 | 285 | if (child) { |
---|
237 | | - qdisc_tree_reduce_backlog(q->qdisc, q->qdisc->q.qlen, |
---|
238 | | - q->qdisc->qstats.backlog); |
---|
239 | | - qdisc_put(q->qdisc); |
---|
| 286 | + qdisc_tree_flush_backlog(q->qdisc); |
---|
| 287 | + old_child = q->qdisc; |
---|
240 | 288 | q->qdisc = child; |
---|
241 | 289 | } |
---|
242 | 290 | |
---|
.. | .. |
---|
255 | 303 | red_start_of_idle_period(&q->vars); |
---|
256 | 304 | |
---|
257 | 305 | sch_tree_unlock(sch); |
---|
| 306 | + |
---|
258 | 307 | red_offload(sch, true); |
---|
| 308 | + |
---|
| 309 | + if (old_child) |
---|
| 310 | + qdisc_put(old_child); |
---|
259 | 311 | return 0; |
---|
| 312 | + |
---|
| 313 | +unlock_out: |
---|
| 314 | + sch_tree_unlock(sch); |
---|
| 315 | + if (child) |
---|
| 316 | + qdisc_put(child); |
---|
| 317 | + return err; |
---|
260 | 318 | } |
---|
261 | 319 | |
---|
262 | 320 | static inline void red_adaptative_timer(struct timer_list *t) |
---|
.. | .. |
---|
275 | 333 | struct netlink_ext_ack *extack) |
---|
276 | 334 | { |
---|
277 | 335 | struct red_sched_data *q = qdisc_priv(sch); |
---|
| 336 | + struct nlattr *tb[TCA_RED_MAX + 1]; |
---|
| 337 | + int err; |
---|
278 | 338 | |
---|
279 | 339 | q->qdisc = &noop_qdisc; |
---|
280 | 340 | q->sch = sch; |
---|
281 | 341 | timer_setup(&q->adapt_timer, red_adaptative_timer, 0); |
---|
282 | | - return red_change(sch, opt, extack); |
---|
| 342 | + |
---|
| 343 | + if (!opt) |
---|
| 344 | + return -EINVAL; |
---|
| 345 | + |
---|
| 346 | + err = nla_parse_nested_deprecated(tb, TCA_RED_MAX, opt, red_policy, |
---|
| 347 | + extack); |
---|
| 348 | + if (err < 0) |
---|
| 349 | + return err; |
---|
| 350 | + |
---|
| 351 | + err = __red_change(sch, tb, extack); |
---|
| 352 | + if (err) |
---|
| 353 | + return err; |
---|
| 354 | + |
---|
| 355 | + err = tcf_qevent_init(&q->qe_early_drop, sch, |
---|
| 356 | + FLOW_BLOCK_BINDER_TYPE_RED_EARLY_DROP, |
---|
| 357 | + tb[TCA_RED_EARLY_DROP_BLOCK], extack); |
---|
| 358 | + if (err) |
---|
| 359 | + return err; |
---|
| 360 | + |
---|
| 361 | + return tcf_qevent_init(&q->qe_mark, sch, |
---|
| 362 | + FLOW_BLOCK_BINDER_TYPE_RED_MARK, |
---|
| 363 | + tb[TCA_RED_MARK_BLOCK], extack); |
---|
283 | 364 | } |
---|
284 | 365 | |
---|
285 | | -static int red_dump_offload_stats(struct Qdisc *sch, struct tc_red_qopt *opt) |
---|
| 366 | +static int red_change(struct Qdisc *sch, struct nlattr *opt, |
---|
| 367 | + struct netlink_ext_ack *extack) |
---|
286 | 368 | { |
---|
287 | | - struct net_device *dev = qdisc_dev(sch); |
---|
| 369 | + struct red_sched_data *q = qdisc_priv(sch); |
---|
| 370 | + struct nlattr *tb[TCA_RED_MAX + 1]; |
---|
| 371 | + int err; |
---|
| 372 | + |
---|
| 373 | + if (!opt) |
---|
| 374 | + return -EINVAL; |
---|
| 375 | + |
---|
| 376 | + err = nla_parse_nested_deprecated(tb, TCA_RED_MAX, opt, red_policy, |
---|
| 377 | + extack); |
---|
| 378 | + if (err < 0) |
---|
| 379 | + return err; |
---|
| 380 | + |
---|
| 381 | + err = tcf_qevent_validate_change(&q->qe_early_drop, |
---|
| 382 | + tb[TCA_RED_EARLY_DROP_BLOCK], extack); |
---|
| 383 | + if (err) |
---|
| 384 | + return err; |
---|
| 385 | + |
---|
| 386 | + err = tcf_qevent_validate_change(&q->qe_mark, |
---|
| 387 | + tb[TCA_RED_MARK_BLOCK], extack); |
---|
| 388 | + if (err) |
---|
| 389 | + return err; |
---|
| 390 | + |
---|
| 391 | + return __red_change(sch, tb, extack); |
---|
| 392 | +} |
---|
| 393 | + |
---|
| 394 | +static int red_dump_offload_stats(struct Qdisc *sch) |
---|
| 395 | +{ |
---|
288 | 396 | struct tc_red_qopt_offload hw_stats = { |
---|
289 | 397 | .command = TC_RED_STATS, |
---|
290 | 398 | .handle = sch->handle, |
---|
.. | .. |
---|
294 | 402 | .stats.qstats = &sch->qstats, |
---|
295 | 403 | }, |
---|
296 | 404 | }; |
---|
297 | | - int err; |
---|
298 | 405 | |
---|
299 | | - sch->flags &= ~TCQ_F_OFFLOADED; |
---|
300 | | - |
---|
301 | | - if (!tc_can_offload(dev) || !dev->netdev_ops->ndo_setup_tc) |
---|
302 | | - return 0; |
---|
303 | | - |
---|
304 | | - err = dev->netdev_ops->ndo_setup_tc(dev, TC_SETUP_QDISC_RED, |
---|
305 | | - &hw_stats); |
---|
306 | | - if (err == -EOPNOTSUPP) |
---|
307 | | - return 0; |
---|
308 | | - |
---|
309 | | - if (!err) |
---|
310 | | - sch->flags |= TCQ_F_OFFLOADED; |
---|
311 | | - |
---|
312 | | - return err; |
---|
| 406 | + return qdisc_offload_dump_helper(sch, TC_SETUP_QDISC_RED, &hw_stats); |
---|
313 | 407 | } |
---|
314 | 408 | |
---|
315 | 409 | static int red_dump(struct Qdisc *sch, struct sk_buff *skb) |
---|
.. | .. |
---|
318 | 412 | struct nlattr *opts = NULL; |
---|
319 | 413 | struct tc_red_qopt opt = { |
---|
320 | 414 | .limit = q->limit, |
---|
321 | | - .flags = q->flags, |
---|
| 415 | + .flags = (q->flags & TC_RED_HISTORIC_FLAGS) | |
---|
| 416 | + q->userbits, |
---|
322 | 417 | .qth_min = q->parms.qth_min >> q->parms.Wlog, |
---|
323 | 418 | .qth_max = q->parms.qth_max >> q->parms.Wlog, |
---|
324 | 419 | .Wlog = q->parms.Wlog, |
---|
.. | .. |
---|
327 | 422 | }; |
---|
328 | 423 | int err; |
---|
329 | 424 | |
---|
330 | | - err = red_dump_offload_stats(sch, &opt); |
---|
| 425 | + err = red_dump_offload_stats(sch); |
---|
331 | 426 | if (err) |
---|
332 | 427 | goto nla_put_failure; |
---|
333 | 428 | |
---|
334 | | - opts = nla_nest_start(skb, TCA_OPTIONS); |
---|
| 429 | + opts = nla_nest_start_noflag(skb, TCA_OPTIONS); |
---|
335 | 430 | if (opts == NULL) |
---|
336 | 431 | goto nla_put_failure; |
---|
337 | 432 | if (nla_put(skb, TCA_RED_PARMS, sizeof(opt), &opt) || |
---|
338 | | - nla_put_u32(skb, TCA_RED_MAX_P, q->parms.max_P)) |
---|
| 433 | + nla_put_u32(skb, TCA_RED_MAX_P, q->parms.max_P) || |
---|
| 434 | + nla_put_bitfield32(skb, TCA_RED_FLAGS, |
---|
| 435 | + q->flags, TC_RED_SUPPORTED_FLAGS) || |
---|
| 436 | + tcf_qevent_dump(skb, TCA_RED_MARK_BLOCK, &q->qe_mark) || |
---|
| 437 | + tcf_qevent_dump(skb, TCA_RED_EARLY_DROP_BLOCK, &q->qe_early_drop)) |
---|
339 | 438 | goto nla_put_failure; |
---|
340 | 439 | return nla_nest_end(skb, opts); |
---|
341 | 440 | |
---|
.. | .. |
---|
380 | 479 | return 0; |
---|
381 | 480 | } |
---|
382 | 481 | |
---|
| 482 | +static void red_graft_offload(struct Qdisc *sch, |
---|
| 483 | + struct Qdisc *new, struct Qdisc *old, |
---|
| 484 | + struct netlink_ext_ack *extack) |
---|
| 485 | +{ |
---|
| 486 | + struct tc_red_qopt_offload graft_offload = { |
---|
| 487 | + .handle = sch->handle, |
---|
| 488 | + .parent = sch->parent, |
---|
| 489 | + .child_handle = new->handle, |
---|
| 490 | + .command = TC_RED_GRAFT, |
---|
| 491 | + }; |
---|
| 492 | + |
---|
| 493 | + qdisc_offload_graft_helper(qdisc_dev(sch), sch, new, old, |
---|
| 494 | + TC_SETUP_QDISC_RED, &graft_offload, extack); |
---|
| 495 | +} |
---|
| 496 | + |
---|
383 | 497 | static int red_graft(struct Qdisc *sch, unsigned long arg, struct Qdisc *new, |
---|
384 | 498 | struct Qdisc **old, struct netlink_ext_ack *extack) |
---|
385 | 499 | { |
---|
.. | .. |
---|
389 | 503 | new = &noop_qdisc; |
---|
390 | 504 | |
---|
391 | 505 | *old = qdisc_replace(sch, new, &q->qdisc); |
---|
| 506 | + |
---|
| 507 | + red_graft_offload(sch, new, *old, extack); |
---|
392 | 508 | return 0; |
---|
393 | 509 | } |
---|
394 | 510 | |
---|