| // SPDX-License-Identifier: GPL-2.0-only |
| /* |
| * net/sched/sch_ets.c Enhanced Transmission Selection scheduler |
| * |
| * Description |
| * ----------- |
| * |
| * The Enhanced Transmission Selection scheduler is a classful queuing |
| * discipline that merges functionality of PRIO and DRR qdiscs in one scheduler. |
| * ETS makes it easy to configure a set of strict and bandwidth-sharing bands to |
| * implement the transmission selection described in 802.1Qaz. |
| * |
| * Although ETS is technically classful, it's not possible to add and remove |
| * classes at will. Instead one specifies number of classes, how many are |
| * PRIO-like and how many DRR-like, and quanta for the latter. |
| * |
| * Algorithm |
| * --------- |
| * |
| * The strict classes, if any, are tried for traffic first: first band 0, if it |
| * has no traffic then band 1, etc. |
| * |
| * When there is no traffic in any of the strict queues, the bandwidth-sharing |
| * ones are tried next. Each band is assigned a deficit counter, initialized to |
| * "quantum" of that band. ETS maintains a list of active bandwidth-sharing |
| * bands whose qdiscs are non-empty. A packet is dequeued from the band at the |
| * head of the list if the packet size is smaller or equal to the deficit |
| * counter. If the counter is too small, it is increased by "quantum" and the |
| * scheduler moves on to the next band in the active list. |
| */ |
| |
| #include <linux/module.h> |
| #include <net/gen_stats.h> |
| #include <net/netlink.h> |
| #include <net/pkt_cls.h> |
| #include <net/pkt_sched.h> |
| #include <net/sch_generic.h> |
| |
| struct ets_class { |
| struct list_head alist; /* In struct ets_sched.active. */ |
| struct Qdisc *qdisc; |
| u32 quantum; |
| u32 deficit; |
| struct gnet_stats_basic_packed bstats; |
| struct gnet_stats_queue qstats; |
| }; |
| |
| struct ets_sched { |
| struct list_head active; |
| struct tcf_proto __rcu *filter_list; |
| struct tcf_block *block; |
| unsigned int nbands; |
| unsigned int nstrict; |
| u8 prio2band[TC_PRIO_MAX + 1]; |
| struct ets_class classes[TCQ_ETS_MAX_BANDS]; |
| }; |
| |
| static const struct nla_policy ets_policy[TCA_ETS_MAX + 1] = { |
| [TCA_ETS_NBANDS] = { .type = NLA_U8 }, |
| [TCA_ETS_NSTRICT] = { .type = NLA_U8 }, |
| [TCA_ETS_QUANTA] = { .type = NLA_NESTED }, |
| [TCA_ETS_PRIOMAP] = { .type = NLA_NESTED }, |
| }; |
| |
| static const struct nla_policy ets_priomap_policy[TCA_ETS_MAX + 1] = { |
| [TCA_ETS_PRIOMAP_BAND] = { .type = NLA_U8 }, |
| }; |
| |
| static const struct nla_policy ets_quanta_policy[TCA_ETS_MAX + 1] = { |
| [TCA_ETS_QUANTA_BAND] = { .type = NLA_U32 }, |
| }; |
| |
| static const struct nla_policy ets_class_policy[TCA_ETS_MAX + 1] = { |
| [TCA_ETS_QUANTA_BAND] = { .type = NLA_U32 }, |
| }; |
| |
| static int ets_quantum_parse(struct Qdisc *sch, const struct nlattr *attr, |
| unsigned int *quantum, |
| struct netlink_ext_ack *extack) |
| { |
| *quantum = nla_get_u32(attr); |
| if (!*quantum) { |
| NL_SET_ERR_MSG(extack, "ETS quantum cannot be zero"); |
| return -EINVAL; |
| } |
| return 0; |
| } |
| |
| static struct ets_class * |
| ets_class_from_arg(struct Qdisc *sch, unsigned long arg) |
| { |
| struct ets_sched *q = qdisc_priv(sch); |
| |
| return &q->classes[arg - 1]; |
| } |
| |
| static u32 ets_class_id(struct Qdisc *sch, const struct ets_class *cl) |
| { |
| struct ets_sched *q = qdisc_priv(sch); |
| int band = cl - q->classes; |
| |
| return TC_H_MAKE(sch->handle, band + 1); |
| } |
| |
| static bool ets_class_is_strict(struct ets_sched *q, const struct ets_class *cl) |
| { |
| unsigned int band = cl - q->classes; |
| |
| return band < q->nstrict; |
| } |
| |
| static int ets_class_change(struct Qdisc *sch, u32 classid, u32 parentid, |
| struct nlattr **tca, unsigned long *arg, |
| struct netlink_ext_ack *extack) |
| { |
| struct ets_class *cl = ets_class_from_arg(sch, *arg); |
| struct ets_sched *q = qdisc_priv(sch); |
| struct nlattr *opt = tca[TCA_OPTIONS]; |
| struct nlattr *tb[TCA_ETS_MAX + 1]; |
| unsigned int quantum; |
| int err; |
| |
| /* Classes can be added and removed only through Qdisc_ops.change |
| * interface. |
| */ |
| if (!cl) { |
| NL_SET_ERR_MSG(extack, "Fine-grained class addition and removal is not supported"); |
| return -EOPNOTSUPP; |
| } |
| |
| if (!opt) { |
| NL_SET_ERR_MSG(extack, "ETS options are required for this operation"); |
| return -EINVAL; |
| } |
| |
| err = nla_parse_nested(tb, TCA_ETS_MAX, opt, ets_class_policy, extack); |
| if (err < 0) |
| return err; |
| |
| if (!tb[TCA_ETS_QUANTA_BAND]) |
| /* Nothing to configure. */ |
| return 0; |
| |
| if (ets_class_is_strict(q, cl)) { |
| NL_SET_ERR_MSG(extack, "Strict bands do not have a configurable quantum"); |
| return -EINVAL; |
| } |
| |
| err = ets_quantum_parse(sch, tb[TCA_ETS_QUANTA_BAND], &quantum, |
| extack); |
| if (err) |
| return err; |
| |
| sch_tree_lock(sch); |
| cl->quantum = quantum; |
| sch_tree_unlock(sch); |
| return 0; |
| } |
| |
| static int ets_class_graft(struct Qdisc *sch, unsigned long arg, |
| struct Qdisc *new, struct Qdisc **old, |
| struct netlink_ext_ack *extack) |
| { |
| struct ets_class *cl = ets_class_from_arg(sch, arg); |
| |
| if (!new) { |
| new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, |
| ets_class_id(sch, cl), NULL); |
| if (!new) |
| new = &noop_qdisc; |
| else |
| qdisc_hash_add(new, true); |
| } |
| |
| *old = qdisc_replace(sch, new, &cl->qdisc); |
| return 0; |
| } |
| |
| static struct Qdisc *ets_class_leaf(struct Qdisc *sch, unsigned long arg) |
| { |
| struct ets_class *cl = ets_class_from_arg(sch, arg); |
| |
| return cl->qdisc; |
| } |
| |
| static unsigned long ets_class_find(struct Qdisc *sch, u32 classid) |
| { |
| unsigned long band = TC_H_MIN(classid); |
| struct ets_sched *q = qdisc_priv(sch); |
| |
| if (band - 1 >= q->nbands) |
| return 0; |
| return band; |
| } |
| |
| static void ets_class_qlen_notify(struct Qdisc *sch, unsigned long arg) |
| { |
| struct ets_class *cl = ets_class_from_arg(sch, arg); |
| struct ets_sched *q = qdisc_priv(sch); |
| |
| /* We get notified about zero-length child Qdiscs as well if they are |
| * offloaded. Those aren't on the active list though, so don't attempt |
| * to remove them. |
| */ |
| if (!ets_class_is_strict(q, cl) && sch->q.qlen) |
| list_del(&cl->alist); |
| } |
| |
| static int ets_class_dump(struct Qdisc *sch, unsigned long arg, |
| struct sk_buff *skb, struct tcmsg *tcm) |
| { |
| struct ets_class *cl = ets_class_from_arg(sch, arg); |
| struct ets_sched *q = qdisc_priv(sch); |
| struct nlattr *nest; |
| |
| tcm->tcm_parent = TC_H_ROOT; |
| tcm->tcm_handle = ets_class_id(sch, cl); |
| tcm->tcm_info = cl->qdisc->handle; |
| |
| nest = nla_nest_start_noflag(skb, TCA_OPTIONS); |
| if (!nest) |
| goto nla_put_failure; |
| if (!ets_class_is_strict(q, cl)) { |
| if (nla_put_u32(skb, TCA_ETS_QUANTA_BAND, cl->quantum)) |
| goto nla_put_failure; |
| } |
| return nla_nest_end(skb, nest); |
| |
| nla_put_failure: |
| nla_nest_cancel(skb, nest); |
| return -EMSGSIZE; |
| } |
| |
| static int ets_class_dump_stats(struct Qdisc *sch, unsigned long arg, |
| struct gnet_dump *d) |
| { |
| struct ets_class *cl = ets_class_from_arg(sch, arg); |
| struct Qdisc *cl_q = cl->qdisc; |
| |
| if (gnet_stats_copy_basic(qdisc_root_sleeping_running(sch), |
| d, NULL, &cl_q->bstats) < 0 || |
| qdisc_qstats_copy(d, cl_q) < 0) |
| return -1; |
| |
| return 0; |
| } |
| |
| static void ets_qdisc_walk(struct Qdisc *sch, struct qdisc_walker *arg) |
| { |
| struct ets_sched *q = qdisc_priv(sch); |
| int i; |
| |
| if (arg->stop) |
| return; |
| |
| for (i = 0; i < q->nbands; i++) { |
| if (arg->count < arg->skip) { |
| arg->count++; |
| continue; |
| } |
| if (arg->fn(sch, i + 1, arg) < 0) { |
| arg->stop = 1; |
| break; |
| } |
| arg->count++; |
| } |
| } |
| |
| static struct tcf_block * |
| ets_qdisc_tcf_block(struct Qdisc *sch, unsigned long cl, |
| struct netlink_ext_ack *extack) |
| { |
| struct ets_sched *q = qdisc_priv(sch); |
| |
| if (cl) { |
| NL_SET_ERR_MSG(extack, "ETS classid must be zero"); |
| return NULL; |
| } |
| |
| return q->block; |
| } |
| |
| static unsigned long ets_qdisc_bind_tcf(struct Qdisc *sch, unsigned long parent, |
| u32 classid) |
| { |
| return ets_class_find(sch, classid); |
| } |
| |
| static void ets_qdisc_unbind_tcf(struct Qdisc *sch, unsigned long arg) |
| { |
| } |
| |
| static struct ets_class *ets_classify(struct sk_buff *skb, struct Qdisc *sch, |
| int *qerr) |
| { |
| struct ets_sched *q = qdisc_priv(sch); |
| u32 band = skb->priority; |
| struct tcf_result res; |
| struct tcf_proto *fl; |
| int err; |
| |
| *qerr = NET_XMIT_SUCCESS | __NET_XMIT_BYPASS; |
| if (TC_H_MAJ(skb->priority) != sch->handle) { |
| fl = rcu_dereference_bh(q->filter_list); |
| err = tcf_classify(skb, fl, &res, false); |
| #ifdef CONFIG_NET_CLS_ACT |
| switch (err) { |
| case TC_ACT_STOLEN: |
| case TC_ACT_QUEUED: |
| case TC_ACT_TRAP: |
| *qerr = NET_XMIT_SUCCESS | __NET_XMIT_STOLEN; |
| /* fall through */ |
| case TC_ACT_SHOT: |
| return NULL; |
| } |
| #endif |
| if (!fl || err < 0) { |
| if (TC_H_MAJ(band)) |
| band = 0; |
| return &q->classes[q->prio2band[band & TC_PRIO_MAX]]; |
| } |
| band = res.classid; |
| } |
| band = TC_H_MIN(band) - 1; |
| if (band >= q->nbands) |
| return &q->classes[q->prio2band[0]]; |
| return &q->classes[band]; |
| } |
| |
| static int ets_qdisc_enqueue(struct sk_buff *skb, struct Qdisc *sch, |
| struct sk_buff **to_free) |
| { |
| unsigned int len = qdisc_pkt_len(skb); |
| struct ets_sched *q = qdisc_priv(sch); |
| struct ets_class *cl; |
| int err = 0; |
| bool first; |
| |
| cl = ets_classify(skb, sch, &err); |
| if (!cl) { |
| if (err & __NET_XMIT_BYPASS) |
| qdisc_qstats_drop(sch); |
| __qdisc_drop(skb, to_free); |
| return err; |
| } |
| |
| first = !cl->qdisc->q.qlen; |
| err = qdisc_enqueue(skb, cl->qdisc, to_free); |
| if (unlikely(err != NET_XMIT_SUCCESS)) { |
| if (net_xmit_drop_count(err)) { |
| cl->qstats.drops++; |
| qdisc_qstats_drop(sch); |
| } |
| return err; |
| } |
| |
| if (first && !ets_class_is_strict(q, cl)) { |
| list_add_tail(&cl->alist, &q->active); |
| cl->deficit = cl->quantum; |
| } |
| |
| sch->qstats.backlog += len; |
| sch->q.qlen++; |
| return err; |
| } |
| |
| static struct sk_buff * |
| ets_qdisc_dequeue_skb(struct Qdisc *sch, struct sk_buff *skb) |
| { |
| qdisc_bstats_update(sch, skb); |
| qdisc_qstats_backlog_dec(sch, skb); |
| sch->q.qlen--; |
| return skb; |
| } |
| |
| static struct sk_buff *ets_qdisc_dequeue(struct Qdisc *sch) |
| { |
| struct ets_sched *q = qdisc_priv(sch); |
| struct ets_class *cl; |
| struct sk_buff *skb; |
| unsigned int band; |
| unsigned int len; |
| |
| while (1) { |
| for (band = 0; band < q->nstrict; band++) { |
| cl = &q->classes[band]; |
| skb = qdisc_dequeue_peeked(cl->qdisc); |
| if (skb) |
| return ets_qdisc_dequeue_skb(sch, skb); |
| } |
| |
| if (list_empty(&q->active)) |
| goto out; |
| |
| cl = list_first_entry(&q->active, struct ets_class, alist); |
| skb = cl->qdisc->ops->peek(cl->qdisc); |
| if (!skb) { |
| qdisc_warn_nonwc(__func__, cl->qdisc); |
| goto out; |
| } |
| |
| len = qdisc_pkt_len(skb); |
| if (len <= cl->deficit) { |
| cl->deficit -= len; |
| skb = qdisc_dequeue_peeked(cl->qdisc); |
| if (unlikely(!skb)) |
| goto out; |
| if (cl->qdisc->q.qlen == 0) |
| list_del(&cl->alist); |
| return ets_qdisc_dequeue_skb(sch, skb); |
| } |
| |
| cl->deficit += cl->quantum; |
| list_move_tail(&cl->alist, &q->active); |
| } |
| out: |
| return NULL; |
| } |
| |
| static int ets_qdisc_priomap_parse(struct nlattr *priomap_attr, |
| unsigned int nbands, u8 *priomap, |
| struct netlink_ext_ack *extack) |
| { |
| const struct nlattr *attr; |
| int prio = 0; |
| u8 band; |
| int rem; |
| int err; |
| |
| err = __nla_validate_nested(priomap_attr, TCA_ETS_MAX, |
| ets_priomap_policy, NL_VALIDATE_STRICT, |
| extack); |
| if (err) |
| return err; |
| |
| nla_for_each_nested(attr, priomap_attr, rem) { |
| switch (nla_type(attr)) { |
| case TCA_ETS_PRIOMAP_BAND: |
| if (prio > TC_PRIO_MAX) { |
| NL_SET_ERR_MSG_MOD(extack, "Too many priorities in ETS priomap"); |
| return -EINVAL; |
| } |
| band = nla_get_u8(attr); |
| if (band >= nbands) { |
| NL_SET_ERR_MSG_MOD(extack, "Invalid band number in ETS priomap"); |
| return -EINVAL; |
| } |
| priomap[prio++] = band; |
| break; |
| default: |
| WARN_ON_ONCE(1); /* Validate should have caught this. */ |
| return -EINVAL; |
| } |
| } |
| |
| return 0; |
| } |
| |
| static int ets_qdisc_quanta_parse(struct Qdisc *sch, struct nlattr *quanta_attr, |
| unsigned int nbands, unsigned int nstrict, |
| unsigned int *quanta, |
| struct netlink_ext_ack *extack) |
| { |
| const struct nlattr *attr; |
| int band = nstrict; |
| int rem; |
| int err; |
| |
| err = __nla_validate_nested(quanta_attr, TCA_ETS_MAX, |
| ets_quanta_policy, NL_VALIDATE_STRICT, |
| extack); |
| if (err < 0) |
| return err; |
| |
| nla_for_each_nested(attr, quanta_attr, rem) { |
| switch (nla_type(attr)) { |
| case TCA_ETS_QUANTA_BAND: |
| if (band >= nbands) { |
| NL_SET_ERR_MSG_MOD(extack, "ETS quanta has more values than bands"); |
| return -EINVAL; |
| } |
| err = ets_quantum_parse(sch, attr, &quanta[band++], |
| extack); |
| if (err) |
| return err; |
| break; |
| default: |
| WARN_ON_ONCE(1); /* Validate should have caught this. */ |
| return -EINVAL; |
| } |
| } |
| |
| return 0; |
| } |
| |
| static int ets_qdisc_change(struct Qdisc *sch, struct nlattr *opt, |
| struct netlink_ext_ack *extack) |
| { |
| unsigned int quanta[TCQ_ETS_MAX_BANDS] = {0}; |
| struct Qdisc *queues[TCQ_ETS_MAX_BANDS]; |
| struct ets_sched *q = qdisc_priv(sch); |
| struct nlattr *tb[TCA_ETS_MAX + 1]; |
| unsigned int oldbands = q->nbands; |
| u8 priomap[TC_PRIO_MAX + 1]; |
| unsigned int nstrict = 0; |
| unsigned int nbands; |
| unsigned int i; |
| int err; |
| |
| if (!opt) { |
| NL_SET_ERR_MSG(extack, "ETS options are required for this operation"); |
| return -EINVAL; |
| } |
| |
| err = nla_parse_nested(tb, TCA_ETS_MAX, opt, ets_policy, extack); |
| if (err < 0) |
| return err; |
| |
| if (!tb[TCA_ETS_NBANDS]) { |
| NL_SET_ERR_MSG_MOD(extack, "Number of bands is a required argument"); |
| return -EINVAL; |
| } |
| nbands = nla_get_u8(tb[TCA_ETS_NBANDS]); |
| if (nbands < 1 || nbands > TCQ_ETS_MAX_BANDS) { |
| NL_SET_ERR_MSG_MOD(extack, "Invalid number of bands"); |
| return -EINVAL; |
| } |
| /* Unless overridden, traffic goes to the last band. */ |
| memset(priomap, nbands - 1, sizeof(priomap)); |
| |
| if (tb[TCA_ETS_NSTRICT]) { |
| nstrict = nla_get_u8(tb[TCA_ETS_NSTRICT]); |
| if (nstrict > nbands) { |
| NL_SET_ERR_MSG_MOD(extack, "Invalid number of strict bands"); |
| return -EINVAL; |
| } |
| } |
| |
| if (tb[TCA_ETS_PRIOMAP]) { |
| err = ets_qdisc_priomap_parse(tb[TCA_ETS_PRIOMAP], |
| nbands, priomap, extack); |
| if (err) |
| return err; |
| } |
| |
| if (tb[TCA_ETS_QUANTA]) { |
| err = ets_qdisc_quanta_parse(sch, tb[TCA_ETS_QUANTA], |
| nbands, nstrict, quanta, extack); |
| if (err) |
| return err; |
| } |
| /* If there are more bands than strict + quanta provided, the remaining |
| * ones are ETS with quantum of MTU. Initialize the missing values here. |
| */ |
| for (i = nstrict; i < nbands; i++) { |
| if (!quanta[i]) |
| quanta[i] = psched_mtu(qdisc_dev(sch)); |
| } |
| |
| /* Before commit, make sure we can allocate all new qdiscs */ |
| for (i = oldbands; i < nbands; i++) { |
| queues[i] = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, |
| ets_class_id(sch, &q->classes[i]), |
| extack); |
| if (!queues[i]) { |
| while (i > oldbands) |
| qdisc_put(queues[--i]); |
| return -ENOMEM; |
| } |
| } |
| |
| sch_tree_lock(sch); |
| |
| q->nbands = nbands; |
| q->nstrict = nstrict; |
| memcpy(q->prio2band, priomap, sizeof(priomap)); |
| |
| for (i = q->nbands; i < oldbands; i++) |
| qdisc_tree_flush_backlog(q->classes[i].qdisc); |
| |
| for (i = 0; i < q->nbands; i++) |
| q->classes[i].quantum = quanta[i]; |
| |
| for (i = oldbands; i < q->nbands; i++) { |
| q->classes[i].qdisc = queues[i]; |
| if (q->classes[i].qdisc != &noop_qdisc) |
| qdisc_hash_add(q->classes[i].qdisc, true); |
| } |
| |
| sch_tree_unlock(sch); |
| |
| for (i = q->nbands; i < oldbands; i++) { |
| qdisc_put(q->classes[i].qdisc); |
| memset(&q->classes[i], 0, sizeof(q->classes[i])); |
| } |
| return 0; |
| } |
| |
| static int ets_qdisc_init(struct Qdisc *sch, struct nlattr *opt, |
| struct netlink_ext_ack *extack) |
| { |
| struct ets_sched *q = qdisc_priv(sch); |
| int err; |
| |
| if (!opt) |
| return -EINVAL; |
| |
| err = tcf_block_get(&q->block, &q->filter_list, sch, extack); |
| if (err) |
| return err; |
| |
| INIT_LIST_HEAD(&q->active); |
| return ets_qdisc_change(sch, opt, extack); |
| } |
| |
| static void ets_qdisc_reset(struct Qdisc *sch) |
| { |
| struct ets_sched *q = qdisc_priv(sch); |
| int band; |
| |
| for (band = q->nstrict; band < q->nbands; band++) { |
| if (q->classes[band].qdisc->q.qlen) |
| list_del(&q->classes[band].alist); |
| } |
| for (band = 0; band < q->nbands; band++) |
| qdisc_reset(q->classes[band].qdisc); |
| sch->qstats.backlog = 0; |
| sch->q.qlen = 0; |
| } |
| |
| static void ets_qdisc_destroy(struct Qdisc *sch) |
| { |
| struct ets_sched *q = qdisc_priv(sch); |
| int band; |
| |
| tcf_block_put(q->block); |
| for (band = 0; band < q->nbands; band++) |
| qdisc_put(q->classes[band].qdisc); |
| } |
| |
| static int ets_qdisc_dump(struct Qdisc *sch, struct sk_buff *skb) |
| { |
| struct ets_sched *q = qdisc_priv(sch); |
| struct nlattr *opts; |
| struct nlattr *nest; |
| int band; |
| int prio; |
| |
| opts = nla_nest_start_noflag(skb, TCA_OPTIONS); |
| if (!opts) |
| goto nla_err; |
| |
| if (nla_put_u8(skb, TCA_ETS_NBANDS, q->nbands)) |
| goto nla_err; |
| |
| if (q->nstrict && |
| nla_put_u8(skb, TCA_ETS_NSTRICT, q->nstrict)) |
| goto nla_err; |
| |
| if (q->nbands > q->nstrict) { |
| nest = nla_nest_start(skb, TCA_ETS_QUANTA); |
| if (!nest) |
| goto nla_err; |
| |
| for (band = q->nstrict; band < q->nbands; band++) { |
| if (nla_put_u32(skb, TCA_ETS_QUANTA_BAND, |
| q->classes[band].quantum)) |
| goto nla_err; |
| } |
| |
| nla_nest_end(skb, nest); |
| } |
| |
| nest = nla_nest_start(skb, TCA_ETS_PRIOMAP); |
| if (!nest) |
| goto nla_err; |
| |
| for (prio = 0; prio <= TC_PRIO_MAX; prio++) { |
| if (nla_put_u8(skb, TCA_ETS_PRIOMAP_BAND, q->prio2band[prio])) |
| goto nla_err; |
| } |
| |
| nla_nest_end(skb, nest); |
| |
| return nla_nest_end(skb, opts); |
| |
| nla_err: |
| nla_nest_cancel(skb, opts); |
| return -EMSGSIZE; |
| } |
| |
| static const struct Qdisc_class_ops ets_class_ops = { |
| .change = ets_class_change, |
| .graft = ets_class_graft, |
| .leaf = ets_class_leaf, |
| .find = ets_class_find, |
| .qlen_notify = ets_class_qlen_notify, |
| .dump = ets_class_dump, |
| .dump_stats = ets_class_dump_stats, |
| .walk = ets_qdisc_walk, |
| .tcf_block = ets_qdisc_tcf_block, |
| .bind_tcf = ets_qdisc_bind_tcf, |
| .unbind_tcf = ets_qdisc_unbind_tcf, |
| }; |
| |
| static struct Qdisc_ops ets_qdisc_ops __read_mostly = { |
| .cl_ops = &ets_class_ops, |
| .id = "ets", |
| .priv_size = sizeof(struct ets_sched), |
| .enqueue = ets_qdisc_enqueue, |
| .dequeue = ets_qdisc_dequeue, |
| .peek = qdisc_peek_dequeued, |
| .change = ets_qdisc_change, |
| .init = ets_qdisc_init, |
| .reset = ets_qdisc_reset, |
| .destroy = ets_qdisc_destroy, |
| .dump = ets_qdisc_dump, |
| .owner = THIS_MODULE, |
| }; |
| |
| static int __init ets_init(void) |
| { |
| return register_qdisc(&ets_qdisc_ops); |
| } |
| |
| static void __exit ets_exit(void) |
| { |
| unregister_qdisc(&ets_qdisc_ops); |
| } |
| |
| module_init(ets_init); |
| module_exit(ets_exit); |
| MODULE_LICENSE("GPL"); |