mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-12-27 10:01:39 -05:00
net/sched: Fix backlog accounting in qdisc_dequeue_internal
This issue applies for the following qdiscs: hhf, fq, fq_codel, and fq_pie, and occurs in their change handlers when adjusting to the new limit. The problem is the following in the values passed to the subsequent qdisc_tree_reduce_backlog call given a tbf parent: When the tbf parent runs out of tokens, skbs of these qdiscs will be placed in gso_skb. Their peek handlers are qdisc_peek_dequeued, which accounts for both qlen and backlog. However, in the case of qdisc_dequeue_internal, ONLY qlen is accounted for when pulling from gso_skb. This means that these qdiscs are missing a qdisc_qstats_backlog_dec when dropping packets to satisfy the new limit in their change handlers. One can observe this issue with the following (with tc patched to support a limit of 0): export TARGET=fq tc qdisc del dev lo root tc qdisc add dev lo root handle 1: tbf rate 8bit burst 100b latency 1ms tc qdisc replace dev lo handle 3: parent 1:1 $TARGET limit 1000 echo ''; echo 'add child'; tc -s -d qdisc show dev lo ping -I lo -f -c2 -s32 -W0.001 127.0.0.1 2>&1 >/dev/null echo ''; echo 'after ping'; tc -s -d qdisc show dev lo tc qdisc change dev lo handle 3: parent 1:1 $TARGET limit 0 echo ''; echo 'after limit drop'; tc -s -d qdisc show dev lo tc qdisc replace dev lo handle 2: parent 1:1 sfq echo ''; echo 'post graft'; tc -s -d qdisc show dev lo The second to last show command shows 0 packets but a positive number (74) of backlog bytes. The problem becomes clearer in the last show command, where qdisc_purge_queue triggers qdisc_tree_reduce_backlog with the positive backlog and causes an underflow in the tbf parent's backlog (4096 Mb instead of 0). To fix this issue, the codepath for all clients of qdisc_dequeue_internal has been simplified: codel, pie, hhf, fq, fq_pie, and fq_codel. qdisc_dequeue_internal handles the backlog adjustments for all cases that do not directly use the dequeue handler. The old fq_codel_change limit adjustment loop accumulated the arguments to the subsequent qdisc_tree_reduce_backlog call through the cstats field. However, this is confusing and error prone as fq_codel_dequeue could also potentially mutate this field (which qdisc_dequeue_internal calls in the non gso_skb case), so we have unified the code here with other qdiscs. Fixes:2d3cbfd6d5("net_sched: Flush gso_skb list too during ->change()") Fixes:4b549a2ef4("fq_codel: Fair Queue Codel AQM") Fixes:10239edf86("net-qdisc-hhf: Heavy-Hitter Filter (HHF) qdisc") Signed-off-by: William Liu <will@willsroot.io> Reviewed-by: Savino Dicanosa <savy@syst3mfailure.io> Link: https://patch.msgid.link/20250812235725.45243-1-will@willsroot.io Signed-off-by: Jakub Kicinski <kuba@kernel.org>
This commit is contained in:
committed by
Jakub Kicinski
parent
d1547bf460
commit
52bf272636
@@ -1038,12 +1038,17 @@ static inline struct sk_buff *qdisc_dequeue_internal(struct Qdisc *sch, bool dir
|
||||
skb = __skb_dequeue(&sch->gso_skb);
|
||||
if (skb) {
|
||||
sch->q.qlen--;
|
||||
qdisc_qstats_backlog_dec(sch, skb);
|
||||
return skb;
|
||||
}
|
||||
if (direct)
|
||||
return __qdisc_dequeue_head(&sch->q);
|
||||
else
|
||||
if (direct) {
|
||||
skb = __qdisc_dequeue_head(&sch->q);
|
||||
if (skb)
|
||||
qdisc_qstats_backlog_dec(sch, skb);
|
||||
return skb;
|
||||
} else {
|
||||
return sch->dequeue(sch);
|
||||
}
|
||||
}
|
||||
|
||||
static inline struct sk_buff *qdisc_dequeue_head(struct Qdisc *sch)
|
||||
|
||||
@@ -101,9 +101,9 @@ static const struct nla_policy codel_policy[TCA_CODEL_MAX + 1] = {
|
||||
static int codel_change(struct Qdisc *sch, struct nlattr *opt,
|
||||
struct netlink_ext_ack *extack)
|
||||
{
|
||||
unsigned int dropped_pkts = 0, dropped_bytes = 0;
|
||||
struct codel_sched_data *q = qdisc_priv(sch);
|
||||
struct nlattr *tb[TCA_CODEL_MAX + 1];
|
||||
unsigned int qlen, dropped = 0;
|
||||
int err;
|
||||
|
||||
err = nla_parse_nested_deprecated(tb, TCA_CODEL_MAX, opt,
|
||||
@@ -142,15 +142,17 @@ static int codel_change(struct Qdisc *sch, struct nlattr *opt,
|
||||
WRITE_ONCE(q->params.ecn,
|
||||
!!nla_get_u32(tb[TCA_CODEL_ECN]));
|
||||
|
||||
qlen = sch->q.qlen;
|
||||
while (sch->q.qlen > sch->limit) {
|
||||
struct sk_buff *skb = qdisc_dequeue_internal(sch, true);
|
||||
|
||||
dropped += qdisc_pkt_len(skb);
|
||||
qdisc_qstats_backlog_dec(sch, skb);
|
||||
if (!skb)
|
||||
break;
|
||||
|
||||
dropped_pkts++;
|
||||
dropped_bytes += qdisc_pkt_len(skb);
|
||||
rtnl_qdisc_drop(skb, sch);
|
||||
}
|
||||
qdisc_tree_reduce_backlog(sch, qlen - sch->q.qlen, dropped);
|
||||
qdisc_tree_reduce_backlog(sch, dropped_pkts, dropped_bytes);
|
||||
|
||||
sch_tree_unlock(sch);
|
||||
return 0;
|
||||
|
||||
@@ -1013,11 +1013,11 @@ static int fq_load_priomap(struct fq_sched_data *q,
|
||||
static int fq_change(struct Qdisc *sch, struct nlattr *opt,
|
||||
struct netlink_ext_ack *extack)
|
||||
{
|
||||
unsigned int dropped_pkts = 0, dropped_bytes = 0;
|
||||
struct fq_sched_data *q = qdisc_priv(sch);
|
||||
struct nlattr *tb[TCA_FQ_MAX + 1];
|
||||
int err, drop_count = 0;
|
||||
unsigned drop_len = 0;
|
||||
u32 fq_log;
|
||||
int err;
|
||||
|
||||
err = nla_parse_nested_deprecated(tb, TCA_FQ_MAX, opt, fq_policy,
|
||||
NULL);
|
||||
@@ -1135,16 +1135,18 @@ static int fq_change(struct Qdisc *sch, struct nlattr *opt,
|
||||
err = fq_resize(sch, fq_log);
|
||||
sch_tree_lock(sch);
|
||||
}
|
||||
|
||||
while (sch->q.qlen > sch->limit) {
|
||||
struct sk_buff *skb = qdisc_dequeue_internal(sch, false);
|
||||
|
||||
if (!skb)
|
||||
break;
|
||||
drop_len += qdisc_pkt_len(skb);
|
||||
|
||||
dropped_pkts++;
|
||||
dropped_bytes += qdisc_pkt_len(skb);
|
||||
rtnl_kfree_skbs(skb, skb);
|
||||
drop_count++;
|
||||
}
|
||||
qdisc_tree_reduce_backlog(sch, drop_count, drop_len);
|
||||
qdisc_tree_reduce_backlog(sch, dropped_pkts, dropped_bytes);
|
||||
|
||||
sch_tree_unlock(sch);
|
||||
return err;
|
||||
|
||||
@@ -366,6 +366,7 @@ static const struct nla_policy fq_codel_policy[TCA_FQ_CODEL_MAX + 1] = {
|
||||
static int fq_codel_change(struct Qdisc *sch, struct nlattr *opt,
|
||||
struct netlink_ext_ack *extack)
|
||||
{
|
||||
unsigned int dropped_pkts = 0, dropped_bytes = 0;
|
||||
struct fq_codel_sched_data *q = qdisc_priv(sch);
|
||||
struct nlattr *tb[TCA_FQ_CODEL_MAX + 1];
|
||||
u32 quantum = 0;
|
||||
@@ -443,13 +444,14 @@ static int fq_codel_change(struct Qdisc *sch, struct nlattr *opt,
|
||||
q->memory_usage > q->memory_limit) {
|
||||
struct sk_buff *skb = qdisc_dequeue_internal(sch, false);
|
||||
|
||||
q->cstats.drop_len += qdisc_pkt_len(skb);
|
||||
if (!skb)
|
||||
break;
|
||||
|
||||
dropped_pkts++;
|
||||
dropped_bytes += qdisc_pkt_len(skb);
|
||||
rtnl_kfree_skbs(skb, skb);
|
||||
q->cstats.drop_count++;
|
||||
}
|
||||
qdisc_tree_reduce_backlog(sch, q->cstats.drop_count, q->cstats.drop_len);
|
||||
q->cstats.drop_count = 0;
|
||||
q->cstats.drop_len = 0;
|
||||
qdisc_tree_reduce_backlog(sch, dropped_pkts, dropped_bytes);
|
||||
|
||||
sch_tree_unlock(sch);
|
||||
return 0;
|
||||
|
||||
@@ -287,10 +287,9 @@ static struct sk_buff *fq_pie_qdisc_dequeue(struct Qdisc *sch)
|
||||
static int fq_pie_change(struct Qdisc *sch, struct nlattr *opt,
|
||||
struct netlink_ext_ack *extack)
|
||||
{
|
||||
unsigned int dropped_pkts = 0, dropped_bytes = 0;
|
||||
struct fq_pie_sched_data *q = qdisc_priv(sch);
|
||||
struct nlattr *tb[TCA_FQ_PIE_MAX + 1];
|
||||
unsigned int len_dropped = 0;
|
||||
unsigned int num_dropped = 0;
|
||||
int err;
|
||||
|
||||
err = nla_parse_nested(tb, TCA_FQ_PIE_MAX, opt, fq_pie_policy, extack);
|
||||
@@ -368,11 +367,14 @@ static int fq_pie_change(struct Qdisc *sch, struct nlattr *opt,
|
||||
while (sch->q.qlen > sch->limit) {
|
||||
struct sk_buff *skb = qdisc_dequeue_internal(sch, false);
|
||||
|
||||
len_dropped += qdisc_pkt_len(skb);
|
||||
num_dropped += 1;
|
||||
if (!skb)
|
||||
break;
|
||||
|
||||
dropped_pkts++;
|
||||
dropped_bytes += qdisc_pkt_len(skb);
|
||||
rtnl_kfree_skbs(skb, skb);
|
||||
}
|
||||
qdisc_tree_reduce_backlog(sch, num_dropped, len_dropped);
|
||||
qdisc_tree_reduce_backlog(sch, dropped_pkts, dropped_bytes);
|
||||
|
||||
sch_tree_unlock(sch);
|
||||
return 0;
|
||||
|
||||
@@ -508,9 +508,9 @@ static const struct nla_policy hhf_policy[TCA_HHF_MAX + 1] = {
|
||||
static int hhf_change(struct Qdisc *sch, struct nlattr *opt,
|
||||
struct netlink_ext_ack *extack)
|
||||
{
|
||||
unsigned int dropped_pkts = 0, dropped_bytes = 0;
|
||||
struct hhf_sched_data *q = qdisc_priv(sch);
|
||||
struct nlattr *tb[TCA_HHF_MAX + 1];
|
||||
unsigned int qlen, prev_backlog;
|
||||
int err;
|
||||
u64 non_hh_quantum;
|
||||
u32 new_quantum = q->quantum;
|
||||
@@ -561,15 +561,17 @@ static int hhf_change(struct Qdisc *sch, struct nlattr *opt,
|
||||
usecs_to_jiffies(us));
|
||||
}
|
||||
|
||||
qlen = sch->q.qlen;
|
||||
prev_backlog = sch->qstats.backlog;
|
||||
while (sch->q.qlen > sch->limit) {
|
||||
struct sk_buff *skb = qdisc_dequeue_internal(sch, false);
|
||||
|
||||
if (!skb)
|
||||
break;
|
||||
|
||||
dropped_pkts++;
|
||||
dropped_bytes += qdisc_pkt_len(skb);
|
||||
rtnl_kfree_skbs(skb, skb);
|
||||
}
|
||||
qdisc_tree_reduce_backlog(sch, qlen - sch->q.qlen,
|
||||
prev_backlog - sch->qstats.backlog);
|
||||
qdisc_tree_reduce_backlog(sch, dropped_pkts, dropped_bytes);
|
||||
|
||||
sch_tree_unlock(sch);
|
||||
return 0;
|
||||
|
||||
@@ -141,9 +141,9 @@ static const struct nla_policy pie_policy[TCA_PIE_MAX + 1] = {
|
||||
static int pie_change(struct Qdisc *sch, struct nlattr *opt,
|
||||
struct netlink_ext_ack *extack)
|
||||
{
|
||||
unsigned int dropped_pkts = 0, dropped_bytes = 0;
|
||||
struct pie_sched_data *q = qdisc_priv(sch);
|
||||
struct nlattr *tb[TCA_PIE_MAX + 1];
|
||||
unsigned int qlen, dropped = 0;
|
||||
int err;
|
||||
|
||||
err = nla_parse_nested_deprecated(tb, TCA_PIE_MAX, opt, pie_policy,
|
||||
@@ -193,15 +193,17 @@ static int pie_change(struct Qdisc *sch, struct nlattr *opt,
|
||||
nla_get_u32(tb[TCA_PIE_DQ_RATE_ESTIMATOR]));
|
||||
|
||||
/* Drop excess packets if new limit is lower */
|
||||
qlen = sch->q.qlen;
|
||||
while (sch->q.qlen > sch->limit) {
|
||||
struct sk_buff *skb = qdisc_dequeue_internal(sch, true);
|
||||
|
||||
dropped += qdisc_pkt_len(skb);
|
||||
qdisc_qstats_backlog_dec(sch, skb);
|
||||
if (!skb)
|
||||
break;
|
||||
|
||||
dropped_pkts++;
|
||||
dropped_bytes += qdisc_pkt_len(skb);
|
||||
rtnl_qdisc_drop(skb, sch);
|
||||
}
|
||||
qdisc_tree_reduce_backlog(sch, qlen - sch->q.qlen, dropped);
|
||||
qdisc_tree_reduce_backlog(sch, dropped_pkts, dropped_bytes);
|
||||
|
||||
sch_tree_unlock(sch);
|
||||
return 0;
|
||||
|
||||
Reference in New Issue
Block a user