| // SPDX-License-Identifier: (GPL-2.0-only OR BSD-2-Clause) |
| /* Copyright (C) 2019 Netronome Systems, Inc. */ |
| |
| #include <linux/hash.h> |
| #include <linux/hashtable.h> |
| #include <linux/jhash.h> |
| #include <linux/math64.h> |
| #include <linux/vmalloc.h> |
| #include <net/pkt_cls.h> |
| #include <net/pkt_sched.h> |
| |
| #include "cmsg.h" |
| #include "main.h" |
| #include "../nfp_port.h" |
| |
| #define NFP_FL_QOS_UPDATE msecs_to_jiffies(1000) |
| #define NFP_FL_QOS_PPS BIT(15) |
| #define NFP_FL_QOS_METER BIT(10) |
| |
| struct nfp_police_cfg_head { |
| __be32 flags_opts; |
| union { |
| __be32 meter_id; |
| __be32 port; |
| }; |
| }; |
| |
| enum NFP_FL_QOS_TYPES { |
| NFP_FL_QOS_TYPE_BPS, |
| NFP_FL_QOS_TYPE_PPS, |
| NFP_FL_QOS_TYPE_MAX, |
| }; |
| |
| /* Police cmsg for configuring a trTCM traffic conditioner (8W/32B) |
| * See RFC 2698 for more details. |
| * ---------------------------------------------------------------- |
| * 3 2 1 |
| * 1 0 9 8 7 6 5 4 3 2 1 0 9 8 7 6 5 4 3 2 1 0 9 8 7 6 5 4 3 2 1 0 |
| * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+ |
| * | Reserved |p| Reserved | |
| * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+ |
| * | Port Ingress | |
| * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+ |
| * | Token Bucket Peak | |
| * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+ |
| * | Token Bucket Committed | |
| * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+ |
| * | Peak Burst Size | |
| * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+ |
| * | Committed Burst Size | |
| * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+ |
| * | Peak Information Rate | |
| * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+ |
| * | Committed Information Rate | |
| * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+ |
| * Word[0](FLag options): |
| * [15] p(pps) 1 for pps, 0 for bps |
| * |
| * Meter control message |
| * 1 0 9 8 7 6 5 4 3 2 1 0 9 8 7 6 5 4 3 2 1 0 9 8 7 6 5 4 3 2 1 0 |
| * +-------------------------------+-+---+-----+-+---------+-+---+-+ |
| * | Reserved |p| Y |TYPE |E|TSHFV |P| PC|R| |
| * +-------------------------------+-+---+-----+-+---------+-+---+-+ |
| * | meter ID | |
| * +-------------------------------+-------------------------------+ |
| * |
| */ |
| struct nfp_police_config { |
| struct nfp_police_cfg_head head; |
| __be32 bkt_tkn_p; |
| __be32 bkt_tkn_c; |
| __be32 pbs; |
| __be32 cbs; |
| __be32 pir; |
| __be32 cir; |
| }; |
| |
| struct nfp_police_stats_reply { |
| struct nfp_police_cfg_head head; |
| __be64 pass_bytes; |
| __be64 pass_pkts; |
| __be64 drop_bytes; |
| __be64 drop_pkts; |
| }; |
| |
| int nfp_flower_offload_one_police(struct nfp_app *app, bool ingress, |
| bool pps, u32 id, u32 rate, u32 burst) |
| { |
| struct nfp_police_config *config; |
| struct sk_buff *skb; |
| |
| skb = nfp_flower_cmsg_alloc(app, sizeof(struct nfp_police_config), |
| NFP_FLOWER_CMSG_TYPE_QOS_MOD, GFP_KERNEL); |
| if (!skb) |
| return -ENOMEM; |
| |
| config = nfp_flower_cmsg_get_data(skb); |
| memset(config, 0, sizeof(struct nfp_police_config)); |
| if (pps) |
| config->head.flags_opts |= cpu_to_be32(NFP_FL_QOS_PPS); |
| if (!ingress) |
| config->head.flags_opts |= cpu_to_be32(NFP_FL_QOS_METER); |
| |
| if (ingress) |
| config->head.port = cpu_to_be32(id); |
| else |
| config->head.meter_id = cpu_to_be32(id); |
| |
| config->bkt_tkn_p = cpu_to_be32(burst); |
| config->bkt_tkn_c = cpu_to_be32(burst); |
| config->pbs = cpu_to_be32(burst); |
| config->cbs = cpu_to_be32(burst); |
| config->pir = cpu_to_be32(rate); |
| config->cir = cpu_to_be32(rate); |
| nfp_ctrl_tx(app->ctrl, skb); |
| |
| return 0; |
| } |
| |
| static int nfp_policer_validate(const struct flow_action *action, |
| const struct flow_action_entry *act, |
| struct netlink_ext_ack *extack, |
| bool ingress) |
| { |
| if (act->police.exceed.act_id != FLOW_ACTION_DROP) { |
| NL_SET_ERR_MSG_MOD(extack, |
| "Offload not supported when exceed action is not drop"); |
| return -EOPNOTSUPP; |
| } |
| |
| if (ingress) { |
| if (act->police.notexceed.act_id != FLOW_ACTION_CONTINUE && |
| act->police.notexceed.act_id != FLOW_ACTION_ACCEPT) { |
| NL_SET_ERR_MSG_MOD(extack, |
| "Offload not supported when conform action is not continue or ok"); |
| return -EOPNOTSUPP; |
| } |
| } else { |
| if (act->police.notexceed.act_id != FLOW_ACTION_PIPE && |
| act->police.notexceed.act_id != FLOW_ACTION_ACCEPT) { |
| NL_SET_ERR_MSG_MOD(extack, |
| "Offload not supported when conform action is not pipe or ok"); |
| return -EOPNOTSUPP; |
| } |
| } |
| |
| if (act->police.notexceed.act_id == FLOW_ACTION_ACCEPT && |
| !flow_action_is_last_entry(action, act)) { |
| NL_SET_ERR_MSG_MOD(extack, |
| "Offload not supported when conform action is ok, but action is not last"); |
| return -EOPNOTSUPP; |
| } |
| |
| if (act->police.peakrate_bytes_ps || |
| act->police.avrate || act->police.overhead) { |
| NL_SET_ERR_MSG_MOD(extack, |
| "Offload not supported when peakrate/avrate/overhead is configured"); |
| return -EOPNOTSUPP; |
| } |
| |
| return 0; |
| } |
| |
| static int |
| nfp_flower_install_rate_limiter(struct nfp_app *app, struct net_device *netdev, |
| struct tc_cls_matchall_offload *flow, |
| struct netlink_ext_ack *extack) |
| { |
| struct flow_action_entry *paction = &flow->rule->action.entries[0]; |
| u32 action_num = flow->rule->action.num_entries; |
| struct nfp_flower_priv *fl_priv = app->priv; |
| struct flow_action_entry *action = NULL; |
| struct nfp_flower_repr_priv *repr_priv; |
| u32 netdev_port_id, i; |
| struct nfp_repr *repr; |
| bool pps_support; |
| u32 bps_num = 0; |
| u32 pps_num = 0; |
| u32 burst; |
| bool pps; |
| u64 rate; |
| int err; |
| |
| if (!nfp_netdev_is_nfp_repr(netdev)) { |
| NL_SET_ERR_MSG_MOD(extack, "unsupported offload: qos rate limit offload not supported on higher level port"); |
| return -EOPNOTSUPP; |
| } |
| repr = netdev_priv(netdev); |
| repr_priv = repr->app_priv; |
| netdev_port_id = nfp_repr_get_port_id(netdev); |
| pps_support = !!(fl_priv->flower_ext_feats & NFP_FL_FEATS_QOS_PPS); |
| |
| if (repr_priv->block_shared) { |
| NL_SET_ERR_MSG_MOD(extack, "unsupported offload: qos rate limit offload not supported on shared blocks"); |
| return -EOPNOTSUPP; |
| } |
| |
| if (repr->port->type != NFP_PORT_VF_PORT) { |
| NL_SET_ERR_MSG_MOD(extack, "unsupported offload: qos rate limit offload not supported on non-VF ports"); |
| return -EOPNOTSUPP; |
| } |
| |
| if (pps_support) { |
| if (action_num > 2 || action_num == 0) { |
| NL_SET_ERR_MSG_MOD(extack, |
| "unsupported offload: qos rate limit offload only support action number 1 or 2"); |
| return -EOPNOTSUPP; |
| } |
| } else { |
| if (!flow_offload_has_one_action(&flow->rule->action)) { |
| NL_SET_ERR_MSG_MOD(extack, |
| "unsupported offload: qos rate limit offload requires a single action"); |
| return -EOPNOTSUPP; |
| } |
| } |
| |
| if (flow->common.prio != 1) { |
| NL_SET_ERR_MSG_MOD(extack, "unsupported offload: qos rate limit offload requires highest priority"); |
| return -EOPNOTSUPP; |
| } |
| |
| for (i = 0 ; i < action_num; i++) { |
| action = paction + i; |
| if (action->id != FLOW_ACTION_POLICE) { |
| NL_SET_ERR_MSG_MOD(extack, |
| "unsupported offload: qos rate limit offload requires police action"); |
| return -EOPNOTSUPP; |
| } |
| |
| err = nfp_policer_validate(&flow->rule->action, action, extack, true); |
| if (err) |
| return err; |
| |
| if (action->police.rate_bytes_ps > 0) { |
| if (bps_num++) { |
| NL_SET_ERR_MSG_MOD(extack, |
| "unsupported offload: qos rate limit offload only support one BPS action"); |
| return -EOPNOTSUPP; |
| } |
| } |
| if (action->police.rate_pkt_ps > 0) { |
| if (!pps_support) { |
| NL_SET_ERR_MSG_MOD(extack, |
| "unsupported offload: FW does not support PPS action"); |
| return -EOPNOTSUPP; |
| } |
| if (pps_num++) { |
| NL_SET_ERR_MSG_MOD(extack, |
| "unsupported offload: qos rate limit offload only support one PPS action"); |
| return -EOPNOTSUPP; |
| } |
| } |
| } |
| |
| for (i = 0 ; i < action_num; i++) { |
| /* Set QoS data for this interface */ |
| action = paction + i; |
| if (action->police.rate_bytes_ps > 0) { |
| rate = action->police.rate_bytes_ps; |
| burst = action->police.burst; |
| } else if (action->police.rate_pkt_ps > 0) { |
| rate = action->police.rate_pkt_ps; |
| burst = action->police.burst_pkt; |
| } else { |
| NL_SET_ERR_MSG_MOD(extack, |
| "unsupported offload: qos rate limit is not BPS or PPS"); |
| continue; |
| } |
| |
| if (rate != 0) { |
| pps = false; |
| if (action->police.rate_pkt_ps > 0) |
| pps = true; |
| nfp_flower_offload_one_police(repr->app, true, |
| pps, netdev_port_id, |
| rate, burst); |
| } |
| } |
| repr_priv->qos_table.netdev_port_id = netdev_port_id; |
| fl_priv->qos_rate_limiters++; |
| if (fl_priv->qos_rate_limiters == 1) |
| schedule_delayed_work(&fl_priv->qos_stats_work, |
| NFP_FL_QOS_UPDATE); |
| |
| return 0; |
| } |
| |
| static int |
| nfp_flower_remove_rate_limiter(struct nfp_app *app, struct net_device *netdev, |
| struct tc_cls_matchall_offload *flow, |
| struct netlink_ext_ack *extack) |
| { |
| struct nfp_flower_priv *fl_priv = app->priv; |
| struct nfp_flower_repr_priv *repr_priv; |
| struct nfp_police_config *config; |
| u32 netdev_port_id, i; |
| struct nfp_repr *repr; |
| struct sk_buff *skb; |
| bool pps_support; |
| |
| if (!nfp_netdev_is_nfp_repr(netdev)) { |
| NL_SET_ERR_MSG_MOD(extack, "unsupported offload: qos rate limit offload not supported on higher level port"); |
| return -EOPNOTSUPP; |
| } |
| repr = netdev_priv(netdev); |
| |
| netdev_port_id = nfp_repr_get_port_id(netdev); |
| repr_priv = repr->app_priv; |
| pps_support = !!(fl_priv->flower_ext_feats & NFP_FL_FEATS_QOS_PPS); |
| |
| if (!repr_priv->qos_table.netdev_port_id) { |
| NL_SET_ERR_MSG_MOD(extack, "unsupported offload: cannot remove qos entry that does not exist"); |
| return -EOPNOTSUPP; |
| } |
| |
| memset(&repr_priv->qos_table, 0, sizeof(struct nfp_fl_qos)); |
| fl_priv->qos_rate_limiters--; |
| if (!fl_priv->qos_rate_limiters) |
| cancel_delayed_work_sync(&fl_priv->qos_stats_work); |
| for (i = 0 ; i < NFP_FL_QOS_TYPE_MAX; i++) { |
| if (i == NFP_FL_QOS_TYPE_PPS && !pps_support) |
| break; |
| /* 0:bps 1:pps |
| * Clear QoS data for this interface. |
| * There is no need to check if a specific QOS_TYPE was |
| * configured as the firmware handles clearing a QoS entry |
| * safely, even if it wasn't explicitly added. |
| */ |
| skb = nfp_flower_cmsg_alloc(repr->app, sizeof(struct nfp_police_config), |
| NFP_FLOWER_CMSG_TYPE_QOS_DEL, GFP_KERNEL); |
| if (!skb) |
| return -ENOMEM; |
| |
| config = nfp_flower_cmsg_get_data(skb); |
| memset(config, 0, sizeof(struct nfp_police_config)); |
| if (i == NFP_FL_QOS_TYPE_PPS) |
| config->head.flags_opts = cpu_to_be32(NFP_FL_QOS_PPS); |
| config->head.port = cpu_to_be32(netdev_port_id); |
| nfp_ctrl_tx(repr->app->ctrl, skb); |
| } |
| |
| return 0; |
| } |
| |
| void nfp_flower_stats_rlim_reply(struct nfp_app *app, struct sk_buff *skb) |
| { |
| struct nfp_flower_priv *fl_priv = app->priv; |
| struct nfp_flower_repr_priv *repr_priv; |
| struct nfp_police_stats_reply *msg; |
| struct nfp_stat_pair *curr_stats; |
| struct nfp_stat_pair *prev_stats; |
| struct net_device *netdev; |
| struct nfp_repr *repr; |
| u32 netdev_port_id; |
| |
| msg = nfp_flower_cmsg_get_data(skb); |
| if (be32_to_cpu(msg->head.flags_opts) & NFP_FL_QOS_METER) |
| return nfp_act_stats_reply(app, msg); |
| |
| netdev_port_id = be32_to_cpu(msg->head.port); |
| rcu_read_lock(); |
| netdev = nfp_app_dev_get(app, netdev_port_id, NULL); |
| if (!netdev) |
| goto exit_unlock_rcu; |
| |
| repr = netdev_priv(netdev); |
| repr_priv = repr->app_priv; |
| curr_stats = &repr_priv->qos_table.curr_stats; |
| prev_stats = &repr_priv->qos_table.prev_stats; |
| |
| spin_lock_bh(&fl_priv->qos_stats_lock); |
| curr_stats->pkts = be64_to_cpu(msg->pass_pkts) + |
| be64_to_cpu(msg->drop_pkts); |
| curr_stats->bytes = be64_to_cpu(msg->pass_bytes) + |
| be64_to_cpu(msg->drop_bytes); |
| |
| if (!repr_priv->qos_table.last_update) { |
| prev_stats->pkts = curr_stats->pkts; |
| prev_stats->bytes = curr_stats->bytes; |
| } |
| |
| repr_priv->qos_table.last_update = jiffies; |
| spin_unlock_bh(&fl_priv->qos_stats_lock); |
| |
| exit_unlock_rcu: |
| rcu_read_unlock(); |
| } |
| |
| static void |
| nfp_flower_stats_rlim_request(struct nfp_flower_priv *fl_priv, |
| u32 id, bool ingress) |
| { |
| struct nfp_police_cfg_head *head; |
| struct sk_buff *skb; |
| |
| skb = nfp_flower_cmsg_alloc(fl_priv->app, |
| sizeof(struct nfp_police_cfg_head), |
| NFP_FLOWER_CMSG_TYPE_QOS_STATS, |
| GFP_ATOMIC); |
| if (!skb) |
| return; |
| head = nfp_flower_cmsg_get_data(skb); |
| |
| memset(head, 0, sizeof(struct nfp_police_cfg_head)); |
| if (ingress) { |
| head->port = cpu_to_be32(id); |
| } else { |
| head->flags_opts = cpu_to_be32(NFP_FL_QOS_METER); |
| head->meter_id = cpu_to_be32(id); |
| } |
| |
| nfp_ctrl_tx(fl_priv->app->ctrl, skb); |
| } |
| |
| static void |
| nfp_flower_stats_rlim_request_all(struct nfp_flower_priv *fl_priv) |
| { |
| struct nfp_reprs *repr_set; |
| int i; |
| |
| rcu_read_lock(); |
| repr_set = rcu_dereference(fl_priv->app->reprs[NFP_REPR_TYPE_VF]); |
| if (!repr_set) |
| goto exit_unlock_rcu; |
| |
| for (i = 0; i < repr_set->num_reprs; i++) { |
| struct net_device *netdev; |
| |
| netdev = rcu_dereference(repr_set->reprs[i]); |
| if (netdev) { |
| struct nfp_repr *priv = netdev_priv(netdev); |
| struct nfp_flower_repr_priv *repr_priv; |
| u32 netdev_port_id; |
| |
| repr_priv = priv->app_priv; |
| netdev_port_id = repr_priv->qos_table.netdev_port_id; |
| if (!netdev_port_id) |
| continue; |
| |
| nfp_flower_stats_rlim_request(fl_priv, |
| netdev_port_id, true); |
| } |
| } |
| |
| exit_unlock_rcu: |
| rcu_read_unlock(); |
| } |
| |
| static void update_stats_cache(struct work_struct *work) |
| { |
| struct delayed_work *delayed_work; |
| struct nfp_flower_priv *fl_priv; |
| |
| delayed_work = to_delayed_work(work); |
| fl_priv = container_of(delayed_work, struct nfp_flower_priv, |
| qos_stats_work); |
| |
| nfp_flower_stats_rlim_request_all(fl_priv); |
| nfp_flower_stats_meter_request_all(fl_priv); |
| |
| schedule_delayed_work(&fl_priv->qos_stats_work, NFP_FL_QOS_UPDATE); |
| } |
| |
| static int |
| nfp_flower_stats_rate_limiter(struct nfp_app *app, struct net_device *netdev, |
| struct tc_cls_matchall_offload *flow, |
| struct netlink_ext_ack *extack) |
| { |
| struct nfp_flower_priv *fl_priv = app->priv; |
| struct nfp_flower_repr_priv *repr_priv; |
| struct nfp_stat_pair *curr_stats; |
| struct nfp_stat_pair *prev_stats; |
| u64 diff_bytes, diff_pkts; |
| struct nfp_repr *repr; |
| |
| if (!nfp_netdev_is_nfp_repr(netdev)) { |
| NL_SET_ERR_MSG_MOD(extack, "unsupported offload: qos rate limit offload not supported on higher level port"); |
| return -EOPNOTSUPP; |
| } |
| repr = netdev_priv(netdev); |
| |
| repr_priv = repr->app_priv; |
| if (!repr_priv->qos_table.netdev_port_id) { |
| NL_SET_ERR_MSG_MOD(extack, "unsupported offload: cannot find qos entry for stats update"); |
| return -EOPNOTSUPP; |
| } |
| |
| spin_lock_bh(&fl_priv->qos_stats_lock); |
| curr_stats = &repr_priv->qos_table.curr_stats; |
| prev_stats = &repr_priv->qos_table.prev_stats; |
| diff_pkts = curr_stats->pkts - prev_stats->pkts; |
| diff_bytes = curr_stats->bytes - prev_stats->bytes; |
| prev_stats->pkts = curr_stats->pkts; |
| prev_stats->bytes = curr_stats->bytes; |
| spin_unlock_bh(&fl_priv->qos_stats_lock); |
| |
| flow_stats_update(&flow->stats, diff_bytes, diff_pkts, 0, |
| repr_priv->qos_table.last_update, |
| FLOW_ACTION_HW_STATS_DELAYED); |
| return 0; |
| } |
| |
| void nfp_flower_qos_init(struct nfp_app *app) |
| { |
| struct nfp_flower_priv *fl_priv = app->priv; |
| |
| spin_lock_init(&fl_priv->qos_stats_lock); |
| mutex_init(&fl_priv->meter_stats_lock); |
| nfp_init_meter_table(app); |
| |
| INIT_DELAYED_WORK(&fl_priv->qos_stats_work, &update_stats_cache); |
| } |
| |
| void nfp_flower_qos_cleanup(struct nfp_app *app) |
| { |
| struct nfp_flower_priv *fl_priv = app->priv; |
| |
| cancel_delayed_work_sync(&fl_priv->qos_stats_work); |
| } |
| |
| int nfp_flower_setup_qos_offload(struct nfp_app *app, struct net_device *netdev, |
| struct tc_cls_matchall_offload *flow) |
| { |
| struct netlink_ext_ack *extack = flow->common.extack; |
| struct nfp_flower_priv *fl_priv = app->priv; |
| int ret; |
| |
| if (!(fl_priv->flower_ext_feats & NFP_FL_FEATS_VF_RLIM)) { |
| NL_SET_ERR_MSG_MOD(extack, "unsupported offload: loaded firmware does not support qos rate limit offload"); |
| return -EOPNOTSUPP; |
| } |
| |
| mutex_lock(&fl_priv->nfp_fl_lock); |
| switch (flow->command) { |
| case TC_CLSMATCHALL_REPLACE: |
| ret = nfp_flower_install_rate_limiter(app, netdev, flow, extack); |
| break; |
| case TC_CLSMATCHALL_DESTROY: |
| ret = nfp_flower_remove_rate_limiter(app, netdev, flow, extack); |
| break; |
| case TC_CLSMATCHALL_STATS: |
| ret = nfp_flower_stats_rate_limiter(app, netdev, flow, extack); |
| break; |
| default: |
| ret = -EOPNOTSUPP; |
| break; |
| } |
| mutex_unlock(&fl_priv->nfp_fl_lock); |
| |
| return ret; |
| } |
| |
| /* Offload tc action, currently only for tc police */ |
| |
| static const struct rhashtable_params stats_meter_table_params = { |
| .key_offset = offsetof(struct nfp_meter_entry, meter_id), |
| .head_offset = offsetof(struct nfp_meter_entry, ht_node), |
| .key_len = sizeof(u32), |
| }; |
| |
| struct nfp_meter_entry * |
| nfp_flower_search_meter_entry(struct nfp_app *app, u32 meter_id) |
| { |
| struct nfp_flower_priv *priv = app->priv; |
| |
| return rhashtable_lookup_fast(&priv->meter_table, &meter_id, |
| stats_meter_table_params); |
| } |
| |
| static struct nfp_meter_entry * |
| nfp_flower_add_meter_entry(struct nfp_app *app, u32 meter_id) |
| { |
| struct nfp_meter_entry *meter_entry = NULL; |
| struct nfp_flower_priv *priv = app->priv; |
| |
| meter_entry = rhashtable_lookup_fast(&priv->meter_table, |
| &meter_id, |
| stats_meter_table_params); |
| if (meter_entry) |
| return meter_entry; |
| |
| meter_entry = kzalloc(sizeof(*meter_entry), GFP_KERNEL); |
| if (!meter_entry) |
| return NULL; |
| |
| meter_entry->meter_id = meter_id; |
| meter_entry->used = jiffies; |
| if (rhashtable_insert_fast(&priv->meter_table, &meter_entry->ht_node, |
| stats_meter_table_params)) { |
| kfree(meter_entry); |
| return NULL; |
| } |
| |
| priv->qos_rate_limiters++; |
| if (priv->qos_rate_limiters == 1) |
| schedule_delayed_work(&priv->qos_stats_work, |
| NFP_FL_QOS_UPDATE); |
| |
| return meter_entry; |
| } |
| |
| static void nfp_flower_del_meter_entry(struct nfp_app *app, u32 meter_id) |
| { |
| struct nfp_meter_entry *meter_entry = NULL; |
| struct nfp_flower_priv *priv = app->priv; |
| |
| meter_entry = rhashtable_lookup_fast(&priv->meter_table, &meter_id, |
| stats_meter_table_params); |
| if (!meter_entry) |
| return; |
| |
| rhashtable_remove_fast(&priv->meter_table, |
| &meter_entry->ht_node, |
| stats_meter_table_params); |
| kfree(meter_entry); |
| priv->qos_rate_limiters--; |
| if (!priv->qos_rate_limiters) |
| cancel_delayed_work_sync(&priv->qos_stats_work); |
| } |
| |
| int nfp_flower_setup_meter_entry(struct nfp_app *app, |
| const struct flow_action_entry *action, |
| enum nfp_meter_op op, |
| u32 meter_id) |
| { |
| struct nfp_flower_priv *fl_priv = app->priv; |
| struct nfp_meter_entry *meter_entry = NULL; |
| int err = 0; |
| |
| mutex_lock(&fl_priv->meter_stats_lock); |
| |
| switch (op) { |
| case NFP_METER_DEL: |
| nfp_flower_del_meter_entry(app, meter_id); |
| goto exit_unlock; |
| case NFP_METER_ADD: |
| meter_entry = nfp_flower_add_meter_entry(app, meter_id); |
| break; |
| default: |
| err = -EOPNOTSUPP; |
| goto exit_unlock; |
| } |
| |
| if (!meter_entry) { |
| err = -ENOMEM; |
| goto exit_unlock; |
| } |
| |
| if (action->police.rate_bytes_ps > 0) { |
| meter_entry->bps = true; |
| meter_entry->rate = action->police.rate_bytes_ps; |
| meter_entry->burst = action->police.burst; |
| } else { |
| meter_entry->bps = false; |
| meter_entry->rate = action->police.rate_pkt_ps; |
| meter_entry->burst = action->police.burst_pkt; |
| } |
| |
| exit_unlock: |
| mutex_unlock(&fl_priv->meter_stats_lock); |
| return err; |
| } |
| |
| int nfp_init_meter_table(struct nfp_app *app) |
| { |
| struct nfp_flower_priv *priv = app->priv; |
| |
| return rhashtable_init(&priv->meter_table, &stats_meter_table_params); |
| } |
| |
| void |
| nfp_flower_stats_meter_request_all(struct nfp_flower_priv *fl_priv) |
| { |
| struct nfp_meter_entry *meter_entry = NULL; |
| struct rhashtable_iter iter; |
| |
| mutex_lock(&fl_priv->meter_stats_lock); |
| rhashtable_walk_enter(&fl_priv->meter_table, &iter); |
| rhashtable_walk_start(&iter); |
| |
| while ((meter_entry = rhashtable_walk_next(&iter)) != NULL) { |
| if (IS_ERR(meter_entry)) |
| continue; |
| nfp_flower_stats_rlim_request(fl_priv, |
| meter_entry->meter_id, false); |
| } |
| |
| rhashtable_walk_stop(&iter); |
| rhashtable_walk_exit(&iter); |
| mutex_unlock(&fl_priv->meter_stats_lock); |
| } |
| |
| static int |
| nfp_act_install_actions(struct nfp_app *app, struct flow_offload_action *fl_act, |
| struct netlink_ext_ack *extack) |
| { |
| struct flow_action_entry *paction = &fl_act->action.entries[0]; |
| u32 action_num = fl_act->action.num_entries; |
| struct nfp_flower_priv *fl_priv = app->priv; |
| struct flow_action_entry *action = NULL; |
| u32 burst, i, meter_id; |
| bool pps_support, pps; |
| bool add = false; |
| u64 rate; |
| int err; |
| |
| pps_support = !!(fl_priv->flower_ext_feats & NFP_FL_FEATS_QOS_PPS); |
| |
| for (i = 0 ; i < action_num; i++) { |
| /* Set qos associate data for this interface */ |
| action = paction + i; |
| if (action->id != FLOW_ACTION_POLICE) { |
| NL_SET_ERR_MSG_MOD(extack, |
| "unsupported offload: qos rate limit offload requires police action"); |
| continue; |
| } |
| |
| err = nfp_policer_validate(&fl_act->action, action, extack, false); |
| if (err) |
| return err; |
| |
| if (action->police.rate_bytes_ps > 0) { |
| rate = action->police.rate_bytes_ps; |
| burst = action->police.burst; |
| } else if (action->police.rate_pkt_ps > 0 && pps_support) { |
| rate = action->police.rate_pkt_ps; |
| burst = action->police.burst_pkt; |
| } else { |
| NL_SET_ERR_MSG_MOD(extack, |
| "unsupported offload: unsupported qos rate limit"); |
| continue; |
| } |
| |
| if (rate != 0) { |
| meter_id = action->hw_index; |
| if (nfp_flower_setup_meter_entry(app, action, NFP_METER_ADD, meter_id)) |
| continue; |
| |
| pps = false; |
| if (action->police.rate_pkt_ps > 0) |
| pps = true; |
| nfp_flower_offload_one_police(app, false, pps, meter_id, |
| rate, burst); |
| add = true; |
| } |
| } |
| |
| return add ? 0 : -EOPNOTSUPP; |
| } |
| |
| static int |
| nfp_act_remove_actions(struct nfp_app *app, struct flow_offload_action *fl_act, |
| struct netlink_ext_ack *extack) |
| { |
| struct nfp_meter_entry *meter_entry = NULL; |
| struct nfp_police_config *config; |
| struct sk_buff *skb; |
| u32 meter_id; |
| bool pps; |
| |
| /* Delete qos associate data for this interface */ |
| if (fl_act->id != FLOW_ACTION_POLICE) { |
| NL_SET_ERR_MSG_MOD(extack, |
| "unsupported offload: qos rate limit offload requires police action"); |
| return -EOPNOTSUPP; |
| } |
| |
| meter_id = fl_act->index; |
| meter_entry = nfp_flower_search_meter_entry(app, meter_id); |
| if (!meter_entry) { |
| NL_SET_ERR_MSG_MOD(extack, |
| "no meter entry when delete the action index."); |
| return -ENOENT; |
| } |
| pps = !meter_entry->bps; |
| |
| skb = nfp_flower_cmsg_alloc(app, sizeof(struct nfp_police_config), |
| NFP_FLOWER_CMSG_TYPE_QOS_DEL, GFP_KERNEL); |
| if (!skb) |
| return -ENOMEM; |
| |
| config = nfp_flower_cmsg_get_data(skb); |
| memset(config, 0, sizeof(struct nfp_police_config)); |
| config->head.flags_opts = cpu_to_be32(NFP_FL_QOS_METER); |
| config->head.meter_id = cpu_to_be32(meter_id); |
| if (pps) |
| config->head.flags_opts |= cpu_to_be32(NFP_FL_QOS_PPS); |
| |
| nfp_ctrl_tx(app->ctrl, skb); |
| nfp_flower_setup_meter_entry(app, NULL, NFP_METER_DEL, meter_id); |
| |
| return 0; |
| } |
| |
| void |
| nfp_act_stats_reply(struct nfp_app *app, void *pmsg) |
| { |
| struct nfp_flower_priv *fl_priv = app->priv; |
| struct nfp_meter_entry *meter_entry = NULL; |
| struct nfp_police_stats_reply *msg = pmsg; |
| u32 meter_id; |
| |
| meter_id = be32_to_cpu(msg->head.meter_id); |
| mutex_lock(&fl_priv->meter_stats_lock); |
| |
| meter_entry = nfp_flower_search_meter_entry(app, meter_id); |
| if (!meter_entry) |
| goto exit_unlock; |
| |
| meter_entry->stats.curr.pkts = be64_to_cpu(msg->pass_pkts) + |
| be64_to_cpu(msg->drop_pkts); |
| meter_entry->stats.curr.bytes = be64_to_cpu(msg->pass_bytes) + |
| be64_to_cpu(msg->drop_bytes); |
| meter_entry->stats.curr.drops = be64_to_cpu(msg->drop_pkts); |
| if (!meter_entry->stats.update) { |
| meter_entry->stats.prev.pkts = meter_entry->stats.curr.pkts; |
| meter_entry->stats.prev.bytes = meter_entry->stats.curr.bytes; |
| meter_entry->stats.prev.drops = meter_entry->stats.curr.drops; |
| } |
| |
| meter_entry->stats.update = jiffies; |
| |
| exit_unlock: |
| mutex_unlock(&fl_priv->meter_stats_lock); |
| } |
| |
| static int |
| nfp_act_stats_actions(struct nfp_app *app, struct flow_offload_action *fl_act, |
| struct netlink_ext_ack *extack) |
| { |
| struct nfp_flower_priv *fl_priv = app->priv; |
| struct nfp_meter_entry *meter_entry = NULL; |
| u64 diff_bytes, diff_pkts, diff_drops; |
| int err = 0; |
| |
| if (fl_act->id != FLOW_ACTION_POLICE) { |
| NL_SET_ERR_MSG_MOD(extack, |
| "unsupported offload: qos rate limit offload requires police action"); |
| return -EOPNOTSUPP; |
| } |
| |
| mutex_lock(&fl_priv->meter_stats_lock); |
| meter_entry = nfp_flower_search_meter_entry(app, fl_act->index); |
| if (!meter_entry) { |
| err = -ENOENT; |
| goto exit_unlock; |
| } |
| diff_pkts = meter_entry->stats.curr.pkts > meter_entry->stats.prev.pkts ? |
| meter_entry->stats.curr.pkts - meter_entry->stats.prev.pkts : 0; |
| diff_bytes = meter_entry->stats.curr.bytes > meter_entry->stats.prev.bytes ? |
| meter_entry->stats.curr.bytes - meter_entry->stats.prev.bytes : 0; |
| diff_drops = meter_entry->stats.curr.drops > meter_entry->stats.prev.drops ? |
| meter_entry->stats.curr.drops - meter_entry->stats.prev.drops : 0; |
| |
| flow_stats_update(&fl_act->stats, diff_bytes, diff_pkts, diff_drops, |
| meter_entry->stats.update, |
| FLOW_ACTION_HW_STATS_DELAYED); |
| |
| meter_entry->stats.prev.pkts = meter_entry->stats.curr.pkts; |
| meter_entry->stats.prev.bytes = meter_entry->stats.curr.bytes; |
| meter_entry->stats.prev.drops = meter_entry->stats.curr.drops; |
| |
| exit_unlock: |
| mutex_unlock(&fl_priv->meter_stats_lock); |
| return err; |
| } |
| |
| int nfp_setup_tc_act_offload(struct nfp_app *app, |
| struct flow_offload_action *fl_act) |
| { |
| struct netlink_ext_ack *extack = fl_act->extack; |
| struct nfp_flower_priv *fl_priv = app->priv; |
| |
| if (!(fl_priv->flower_ext_feats & NFP_FL_FEATS_QOS_METER)) |
| return -EOPNOTSUPP; |
| |
| switch (fl_act->command) { |
| case FLOW_ACT_REPLACE: |
| return nfp_act_install_actions(app, fl_act, extack); |
| case FLOW_ACT_DESTROY: |
| return nfp_act_remove_actions(app, fl_act, extack); |
| case FLOW_ACT_STATS: |
| return nfp_act_stats_actions(app, fl_act, extack); |
| default: |
| return -EOPNOTSUPP; |
| } |
| } |