| #ifndef _NET_FLOW_OFFLOAD_H |
| #define _NET_FLOW_OFFLOAD_H |
| |
| #include <linux/kernel.h> |
| #include <linux/list.h> |
| #include <linux/netlink.h> |
| #include <net/flow_dissector.h> |
| |
| struct flow_match { |
| struct flow_dissector *dissector; |
| void *mask; |
| void *key; |
| }; |
| |
| struct flow_match_meta { |
| struct flow_dissector_key_meta *key, *mask; |
| }; |
| |
| struct flow_match_basic { |
| struct flow_dissector_key_basic *key, *mask; |
| }; |
| |
| struct flow_match_control { |
| struct flow_dissector_key_control *key, *mask; |
| }; |
| |
| struct flow_match_eth_addrs { |
| struct flow_dissector_key_eth_addrs *key, *mask; |
| }; |
| |
| struct flow_match_vlan { |
| struct flow_dissector_key_vlan *key, *mask; |
| }; |
| |
| struct flow_match_arp { |
| struct flow_dissector_key_arp *key, *mask; |
| }; |
| |
| struct flow_match_ipv4_addrs { |
| struct flow_dissector_key_ipv4_addrs *key, *mask; |
| }; |
| |
| struct flow_match_ipv6_addrs { |
| struct flow_dissector_key_ipv6_addrs *key, *mask; |
| }; |
| |
| struct flow_match_ip { |
| struct flow_dissector_key_ip *key, *mask; |
| }; |
| |
| struct flow_match_ports { |
| struct flow_dissector_key_ports *key, *mask; |
| }; |
| |
| struct flow_match_ports_range { |
| struct flow_dissector_key_ports_range *key, *mask; |
| }; |
| |
| struct flow_match_icmp { |
| struct flow_dissector_key_icmp *key, *mask; |
| }; |
| |
| struct flow_match_tcp { |
| struct flow_dissector_key_tcp *key, *mask; |
| }; |
| |
| struct flow_match_ipsec { |
| struct flow_dissector_key_ipsec *key, *mask; |
| }; |
| |
| struct flow_match_mpls { |
| struct flow_dissector_key_mpls *key, *mask; |
| }; |
| |
| struct flow_match_enc_keyid { |
| struct flow_dissector_key_keyid *key, *mask; |
| }; |
| |
| struct flow_match_enc_opts { |
| struct flow_dissector_key_enc_opts *key, *mask; |
| }; |
| |
| struct flow_match_ct { |
| struct flow_dissector_key_ct *key, *mask; |
| }; |
| |
| struct flow_match_pppoe { |
| struct flow_dissector_key_pppoe *key, *mask; |
| }; |
| |
| struct flow_match_l2tpv3 { |
| struct flow_dissector_key_l2tpv3 *key, *mask; |
| }; |
| |
| struct flow_rule; |
| |
| void flow_rule_match_meta(const struct flow_rule *rule, |
| struct flow_match_meta *out); |
| void flow_rule_match_basic(const struct flow_rule *rule, |
| struct flow_match_basic *out); |
| void flow_rule_match_control(const struct flow_rule *rule, |
| struct flow_match_control *out); |
| void flow_rule_match_eth_addrs(const struct flow_rule *rule, |
| struct flow_match_eth_addrs *out); |
| void flow_rule_match_vlan(const struct flow_rule *rule, |
| struct flow_match_vlan *out); |
| void flow_rule_match_cvlan(const struct flow_rule *rule, |
| struct flow_match_vlan *out); |
| void flow_rule_match_arp(const struct flow_rule *rule, |
| struct flow_match_arp *out); |
| void flow_rule_match_ipv4_addrs(const struct flow_rule *rule, |
| struct flow_match_ipv4_addrs *out); |
| void flow_rule_match_ipv6_addrs(const struct flow_rule *rule, |
| struct flow_match_ipv6_addrs *out); |
| void flow_rule_match_ip(const struct flow_rule *rule, |
| struct flow_match_ip *out); |
| void flow_rule_match_ports(const struct flow_rule *rule, |
| struct flow_match_ports *out); |
| void flow_rule_match_ports_range(const struct flow_rule *rule, |
| struct flow_match_ports_range *out); |
| void flow_rule_match_tcp(const struct flow_rule *rule, |
| struct flow_match_tcp *out); |
| void flow_rule_match_ipsec(const struct flow_rule *rule, |
| struct flow_match_ipsec *out); |
| void flow_rule_match_icmp(const struct flow_rule *rule, |
| struct flow_match_icmp *out); |
| void flow_rule_match_mpls(const struct flow_rule *rule, |
| struct flow_match_mpls *out); |
| void flow_rule_match_enc_control(const struct flow_rule *rule, |
| struct flow_match_control *out); |
| void flow_rule_match_enc_ipv4_addrs(const struct flow_rule *rule, |
| struct flow_match_ipv4_addrs *out); |
| void flow_rule_match_enc_ipv6_addrs(const struct flow_rule *rule, |
| struct flow_match_ipv6_addrs *out); |
| void flow_rule_match_enc_ip(const struct flow_rule *rule, |
| struct flow_match_ip *out); |
| void flow_rule_match_enc_ports(const struct flow_rule *rule, |
| struct flow_match_ports *out); |
| void flow_rule_match_enc_keyid(const struct flow_rule *rule, |
| struct flow_match_enc_keyid *out); |
| void flow_rule_match_enc_opts(const struct flow_rule *rule, |
| struct flow_match_enc_opts *out); |
| void flow_rule_match_ct(const struct flow_rule *rule, |
| struct flow_match_ct *out); |
| void flow_rule_match_pppoe(const struct flow_rule *rule, |
| struct flow_match_pppoe *out); |
| void flow_rule_match_l2tpv3(const struct flow_rule *rule, |
| struct flow_match_l2tpv3 *out); |
| |
| enum flow_action_id { |
| FLOW_ACTION_ACCEPT = 0, |
| FLOW_ACTION_DROP, |
| FLOW_ACTION_TRAP, |
| FLOW_ACTION_GOTO, |
| FLOW_ACTION_REDIRECT, |
| FLOW_ACTION_MIRRED, |
| FLOW_ACTION_REDIRECT_INGRESS, |
| FLOW_ACTION_MIRRED_INGRESS, |
| FLOW_ACTION_VLAN_PUSH, |
| FLOW_ACTION_VLAN_POP, |
| FLOW_ACTION_VLAN_MANGLE, |
| FLOW_ACTION_TUNNEL_ENCAP, |
| FLOW_ACTION_TUNNEL_DECAP, |
| FLOW_ACTION_MANGLE, |
| FLOW_ACTION_ADD, |
| FLOW_ACTION_CSUM, |
| FLOW_ACTION_MARK, |
| FLOW_ACTION_PTYPE, |
| FLOW_ACTION_PRIORITY, |
| FLOW_ACTION_RX_QUEUE_MAPPING, |
| FLOW_ACTION_WAKE, |
| FLOW_ACTION_QUEUE, |
| FLOW_ACTION_SAMPLE, |
| FLOW_ACTION_POLICE, |
| FLOW_ACTION_CT, |
| FLOW_ACTION_CT_METADATA, |
| FLOW_ACTION_MPLS_PUSH, |
| FLOW_ACTION_MPLS_POP, |
| FLOW_ACTION_MPLS_MANGLE, |
| FLOW_ACTION_GATE, |
| FLOW_ACTION_PPPOE_PUSH, |
| FLOW_ACTION_JUMP, |
| FLOW_ACTION_PIPE, |
| FLOW_ACTION_VLAN_PUSH_ETH, |
| FLOW_ACTION_VLAN_POP_ETH, |
| FLOW_ACTION_CONTINUE, |
| NUM_FLOW_ACTIONS, |
| }; |
| |
| /* This is mirroring enum pedit_header_type definition for easy mapping between |
| * tc pedit action. Legacy TCA_PEDIT_KEY_EX_HDR_TYPE_NETWORK is mapped to |
| * FLOW_ACT_MANGLE_UNSPEC, which is supported by no driver. |
| */ |
| enum flow_action_mangle_base { |
| FLOW_ACT_MANGLE_UNSPEC = 0, |
| FLOW_ACT_MANGLE_HDR_TYPE_ETH, |
| FLOW_ACT_MANGLE_HDR_TYPE_IP4, |
| FLOW_ACT_MANGLE_HDR_TYPE_IP6, |
| FLOW_ACT_MANGLE_HDR_TYPE_TCP, |
| FLOW_ACT_MANGLE_HDR_TYPE_UDP, |
| }; |
| |
| enum flow_action_hw_stats_bit { |
| FLOW_ACTION_HW_STATS_IMMEDIATE_BIT, |
| FLOW_ACTION_HW_STATS_DELAYED_BIT, |
| FLOW_ACTION_HW_STATS_DISABLED_BIT, |
| |
| FLOW_ACTION_HW_STATS_NUM_BITS |
| }; |
| |
| enum flow_action_hw_stats { |
| FLOW_ACTION_HW_STATS_IMMEDIATE = |
| BIT(FLOW_ACTION_HW_STATS_IMMEDIATE_BIT), |
| FLOW_ACTION_HW_STATS_DELAYED = BIT(FLOW_ACTION_HW_STATS_DELAYED_BIT), |
| FLOW_ACTION_HW_STATS_ANY = FLOW_ACTION_HW_STATS_IMMEDIATE | |
| FLOW_ACTION_HW_STATS_DELAYED, |
| FLOW_ACTION_HW_STATS_DISABLED = |
| BIT(FLOW_ACTION_HW_STATS_DISABLED_BIT), |
| FLOW_ACTION_HW_STATS_DONT_CARE = BIT(FLOW_ACTION_HW_STATS_NUM_BITS) - 1, |
| }; |
| |
| typedef void (*action_destr)(void *priv); |
| |
| struct flow_action_cookie { |
| u32 cookie_len; |
| u8 cookie[]; |
| }; |
| |
| struct flow_action_cookie *flow_action_cookie_create(void *data, |
| unsigned int len, |
| gfp_t gfp); |
| void flow_action_cookie_destroy(struct flow_action_cookie *cookie); |
| |
| struct flow_action_entry { |
| enum flow_action_id id; |
| u32 hw_index; |
| unsigned long cookie; |
| u64 miss_cookie; |
| enum flow_action_hw_stats hw_stats; |
| action_destr destructor; |
| void *destructor_priv; |
| union { |
| u32 chain_index; /* FLOW_ACTION_GOTO */ |
| struct net_device *dev; /* FLOW_ACTION_REDIRECT */ |
| struct { /* FLOW_ACTION_VLAN */ |
| u16 vid; |
| __be16 proto; |
| u8 prio; |
| } vlan; |
| struct { /* FLOW_ACTION_VLAN_PUSH_ETH */ |
| unsigned char dst[ETH_ALEN]; |
| unsigned char src[ETH_ALEN]; |
| } vlan_push_eth; |
| struct { /* FLOW_ACTION_MANGLE */ |
| /* FLOW_ACTION_ADD */ |
| enum flow_action_mangle_base htype; |
| u32 offset; |
| u32 mask; |
| u32 val; |
| } mangle; |
| struct ip_tunnel_info *tunnel; /* FLOW_ACTION_TUNNEL_ENCAP */ |
| u32 csum_flags; /* FLOW_ACTION_CSUM */ |
| u32 mark; /* FLOW_ACTION_MARK */ |
| u16 ptype; /* FLOW_ACTION_PTYPE */ |
| u16 rx_queue; /* FLOW_ACTION_RX_QUEUE_MAPPING */ |
| u32 priority; /* FLOW_ACTION_PRIORITY */ |
| struct { /* FLOW_ACTION_QUEUE */ |
| u32 ctx; |
| u32 index; |
| u8 vf; |
| } queue; |
| struct { /* FLOW_ACTION_SAMPLE */ |
| struct psample_group *psample_group; |
| u32 rate; |
| u32 trunc_size; |
| bool truncate; |
| } sample; |
| struct { /* FLOW_ACTION_POLICE */ |
| u32 burst; |
| u64 rate_bytes_ps; |
| u64 peakrate_bytes_ps; |
| u32 avrate; |
| u16 overhead; |
| u64 burst_pkt; |
| u64 rate_pkt_ps; |
| u32 mtu; |
| struct { |
| enum flow_action_id act_id; |
| u32 extval; |
| } exceed, notexceed; |
| } police; |
| struct { /* FLOW_ACTION_CT */ |
| int action; |
| u16 zone; |
| struct nf_flowtable *flow_table; |
| } ct; |
| struct { |
| unsigned long cookie; |
| u32 mark; |
| u32 labels[4]; |
| bool orig_dir; |
| } ct_metadata; |
| struct { /* FLOW_ACTION_MPLS_PUSH */ |
| u32 label; |
| __be16 proto; |
| u8 tc; |
| u8 bos; |
| u8 ttl; |
| } mpls_push; |
| struct { /* FLOW_ACTION_MPLS_POP */ |
| __be16 proto; |
| } mpls_pop; |
| struct { /* FLOW_ACTION_MPLS_MANGLE */ |
| u32 label; |
| u8 tc; |
| u8 bos; |
| u8 ttl; |
| } mpls_mangle; |
| struct { |
| s32 prio; |
| u64 basetime; |
| u64 cycletime; |
| u64 cycletimeext; |
| u32 num_entries; |
| struct action_gate_entry *entries; |
| } gate; |
| struct { /* FLOW_ACTION_PPPOE_PUSH */ |
| u16 sid; |
| } pppoe; |
| }; |
| struct flow_action_cookie *user_cookie; /* user defined action cookie */ |
| }; |
| |
| struct flow_action { |
| unsigned int num_entries; |
| struct flow_action_entry entries[] __counted_by(num_entries); |
| }; |
| |
| static inline bool flow_action_has_entries(const struct flow_action *action) |
| { |
| return action->num_entries; |
| } |
| |
| /** |
| * flow_offload_has_one_action() - check if exactly one action is present |
| * @action: tc filter flow offload action |
| * |
| * Return: true if exactly one action is present. |
| */ |
| static inline bool flow_offload_has_one_action(const struct flow_action *action) |
| { |
| return action->num_entries == 1; |
| } |
| |
| static inline bool flow_action_is_last_entry(const struct flow_action *action, |
| const struct flow_action_entry *entry) |
| { |
| return entry == &action->entries[action->num_entries - 1]; |
| } |
| |
| #define flow_action_for_each(__i, __act, __actions) \ |
| for (__i = 0, __act = &(__actions)->entries[0]; \ |
| __i < (__actions)->num_entries; \ |
| __act = &(__actions)->entries[++__i]) |
| |
| static inline bool |
| flow_action_mixed_hw_stats_check(const struct flow_action *action, |
| struct netlink_ext_ack *extack) |
| { |
| const struct flow_action_entry *action_entry; |
| u8 last_hw_stats; |
| int i; |
| |
| if (flow_offload_has_one_action(action)) |
| return true; |
| |
| flow_action_for_each(i, action_entry, action) { |
| if (i && action_entry->hw_stats != last_hw_stats) { |
| NL_SET_ERR_MSG_MOD(extack, "Mixing HW stats types for actions is not supported"); |
| return false; |
| } |
| last_hw_stats = action_entry->hw_stats; |
| } |
| return true; |
| } |
| |
| static inline const struct flow_action_entry * |
| flow_action_first_entry_get(const struct flow_action *action) |
| { |
| WARN_ON(!flow_action_has_entries(action)); |
| return &action->entries[0]; |
| } |
| |
| static inline bool |
| __flow_action_hw_stats_check(const struct flow_action *action, |
| struct netlink_ext_ack *extack, |
| bool check_allow_bit, |
| enum flow_action_hw_stats_bit allow_bit) |
| { |
| const struct flow_action_entry *action_entry; |
| |
| if (!flow_action_has_entries(action)) |
| return true; |
| if (!flow_action_mixed_hw_stats_check(action, extack)) |
| return false; |
| |
| action_entry = flow_action_first_entry_get(action); |
| |
| /* Zero is not a legal value for hw_stats, catch anyone passing it */ |
| WARN_ON_ONCE(!action_entry->hw_stats); |
| |
| if (!check_allow_bit && |
| ~action_entry->hw_stats & FLOW_ACTION_HW_STATS_ANY) { |
| NL_SET_ERR_MSG_MOD(extack, "Driver supports only default HW stats type \"any\""); |
| return false; |
| } else if (check_allow_bit && |
| !(action_entry->hw_stats & BIT(allow_bit))) { |
| NL_SET_ERR_MSG_MOD(extack, "Driver does not support selected HW stats type"); |
| return false; |
| } |
| return true; |
| } |
| |
| static inline bool |
| flow_action_hw_stats_check(const struct flow_action *action, |
| struct netlink_ext_ack *extack, |
| enum flow_action_hw_stats_bit allow_bit) |
| { |
| return __flow_action_hw_stats_check(action, extack, true, allow_bit); |
| } |
| |
| static inline bool |
| flow_action_basic_hw_stats_check(const struct flow_action *action, |
| struct netlink_ext_ack *extack) |
| { |
| return __flow_action_hw_stats_check(action, extack, false, 0); |
| } |
| |
| struct flow_rule { |
| struct flow_match match; |
| struct flow_action action; |
| }; |
| |
| struct flow_rule *flow_rule_alloc(unsigned int num_actions); |
| |
| static inline bool flow_rule_match_key(const struct flow_rule *rule, |
| enum flow_dissector_key_id key) |
| { |
| return dissector_uses_key(rule->match.dissector, key); |
| } |
| |
| /** |
| * flow_rule_is_supp_control_flags() - check for supported control flags |
| * @supp_flags: control flags supported by driver |
| * @ctrl_flags: control flags present in rule |
| * @extack: The netlink extended ACK for reporting errors. |
| * |
| * Return: true if only supported control flags are set, false otherwise. |
| */ |
| static inline bool flow_rule_is_supp_control_flags(const u32 supp_flags, |
| const u32 ctrl_flags, |
| struct netlink_ext_ack *extack) |
| { |
| if (likely((ctrl_flags & ~supp_flags) == 0)) |
| return true; |
| |
| NL_SET_ERR_MSG_FMT_MOD(extack, |
| "Unsupported match on control.flags %#x", |
| ctrl_flags); |
| |
| return false; |
| } |
| |
| /** |
| * flow_rule_has_control_flags() - check for presence of any control flags |
| * @ctrl_flags: control flags present in rule |
| * @extack: The netlink extended ACK for reporting errors. |
| * |
| * Return: true if control flags are set, false otherwise. |
| */ |
| static inline bool flow_rule_has_control_flags(const u32 ctrl_flags, |
| struct netlink_ext_ack *extack) |
| { |
| return !flow_rule_is_supp_control_flags(0, ctrl_flags, extack); |
| } |
| |
| /** |
| * flow_rule_match_has_control_flags() - match and check for any control flags |
| * @rule: The flow_rule under evaluation. |
| * @extack: The netlink extended ACK for reporting errors. |
| * |
| * Return: true if control flags are set, false otherwise. |
| */ |
| static inline bool flow_rule_match_has_control_flags(struct flow_rule *rule, |
| struct netlink_ext_ack *extack) |
| { |
| struct flow_match_control match; |
| |
| if (!flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_CONTROL)) |
| return false; |
| |
| flow_rule_match_control(rule, &match); |
| |
| return flow_rule_has_control_flags(match.mask->flags, extack); |
| } |
| |
| struct flow_stats { |
| u64 pkts; |
| u64 bytes; |
| u64 drops; |
| u64 lastused; |
| enum flow_action_hw_stats used_hw_stats; |
| bool used_hw_stats_valid; |
| }; |
| |
| static inline void flow_stats_update(struct flow_stats *flow_stats, |
| u64 bytes, u64 pkts, |
| u64 drops, u64 lastused, |
| enum flow_action_hw_stats used_hw_stats) |
| { |
| flow_stats->pkts += pkts; |
| flow_stats->bytes += bytes; |
| flow_stats->drops += drops; |
| flow_stats->lastused = max_t(u64, flow_stats->lastused, lastused); |
| |
| /* The driver should pass value with a maximum of one bit set. |
| * Passing FLOW_ACTION_HW_STATS_ANY is invalid. |
| */ |
| WARN_ON(used_hw_stats == FLOW_ACTION_HW_STATS_ANY); |
| flow_stats->used_hw_stats |= used_hw_stats; |
| flow_stats->used_hw_stats_valid = true; |
| } |
| |
| enum flow_block_command { |
| FLOW_BLOCK_BIND, |
| FLOW_BLOCK_UNBIND, |
| }; |
| |
| enum flow_block_binder_type { |
| FLOW_BLOCK_BINDER_TYPE_UNSPEC, |
| FLOW_BLOCK_BINDER_TYPE_CLSACT_INGRESS, |
| FLOW_BLOCK_BINDER_TYPE_CLSACT_EGRESS, |
| FLOW_BLOCK_BINDER_TYPE_RED_EARLY_DROP, |
| FLOW_BLOCK_BINDER_TYPE_RED_MARK, |
| }; |
| |
| struct flow_block { |
| struct list_head cb_list; |
| }; |
| |
| struct netlink_ext_ack; |
| |
| struct flow_block_offload { |
| enum flow_block_command command; |
| enum flow_block_binder_type binder_type; |
| bool block_shared; |
| bool unlocked_driver_cb; |
| struct net *net; |
| struct flow_block *block; |
| struct list_head cb_list; |
| struct list_head *driver_block_list; |
| struct netlink_ext_ack *extack; |
| struct Qdisc *sch; |
| struct list_head *cb_list_head; |
| }; |
| |
| enum tc_setup_type; |
| typedef int flow_setup_cb_t(enum tc_setup_type type, void *type_data, |
| void *cb_priv); |
| |
| struct flow_block_cb; |
| |
| struct flow_block_indr { |
| struct list_head list; |
| struct net_device *dev; |
| struct Qdisc *sch; |
| enum flow_block_binder_type binder_type; |
| void *data; |
| void *cb_priv; |
| void (*cleanup)(struct flow_block_cb *block_cb); |
| }; |
| |
| struct flow_block_cb { |
| struct list_head driver_list; |
| struct list_head list; |
| flow_setup_cb_t *cb; |
| void *cb_ident; |
| void *cb_priv; |
| void (*release)(void *cb_priv); |
| struct flow_block_indr indr; |
| unsigned int refcnt; |
| }; |
| |
| struct flow_block_cb *flow_block_cb_alloc(flow_setup_cb_t *cb, |
| void *cb_ident, void *cb_priv, |
| void (*release)(void *cb_priv)); |
| struct flow_block_cb *flow_indr_block_cb_alloc(flow_setup_cb_t *cb, |
| void *cb_ident, void *cb_priv, |
| void (*release)(void *cb_priv), |
| struct flow_block_offload *bo, |
| struct net_device *dev, |
| struct Qdisc *sch, void *data, |
| void *indr_cb_priv, |
| void (*cleanup)(struct flow_block_cb *block_cb)); |
| void flow_block_cb_free(struct flow_block_cb *block_cb); |
| |
| struct flow_block_cb *flow_block_cb_lookup(struct flow_block *block, |
| flow_setup_cb_t *cb, void *cb_ident); |
| |
| void *flow_block_cb_priv(struct flow_block_cb *block_cb); |
| void flow_block_cb_incref(struct flow_block_cb *block_cb); |
| unsigned int flow_block_cb_decref(struct flow_block_cb *block_cb); |
| |
| static inline void flow_block_cb_add(struct flow_block_cb *block_cb, |
| struct flow_block_offload *offload) |
| { |
| list_add_tail(&block_cb->list, &offload->cb_list); |
| } |
| |
| static inline void flow_block_cb_remove(struct flow_block_cb *block_cb, |
| struct flow_block_offload *offload) |
| { |
| list_move(&block_cb->list, &offload->cb_list); |
| } |
| |
| static inline void flow_indr_block_cb_remove(struct flow_block_cb *block_cb, |
| struct flow_block_offload *offload) |
| { |
| list_del(&block_cb->indr.list); |
| list_move(&block_cb->list, &offload->cb_list); |
| } |
| |
| bool flow_block_cb_is_busy(flow_setup_cb_t *cb, void *cb_ident, |
| struct list_head *driver_block_list); |
| |
| int flow_block_cb_setup_simple(struct flow_block_offload *f, |
| struct list_head *driver_list, |
| flow_setup_cb_t *cb, |
| void *cb_ident, void *cb_priv, bool ingress_only); |
| |
| enum flow_cls_command { |
| FLOW_CLS_REPLACE, |
| FLOW_CLS_DESTROY, |
| FLOW_CLS_STATS, |
| FLOW_CLS_TMPLT_CREATE, |
| FLOW_CLS_TMPLT_DESTROY, |
| }; |
| |
| struct flow_cls_common_offload { |
| u32 chain_index; |
| __be16 protocol; |
| u32 prio; |
| struct netlink_ext_ack *extack; |
| }; |
| |
| struct flow_cls_offload { |
| struct flow_cls_common_offload common; |
| enum flow_cls_command command; |
| bool use_act_stats; |
| unsigned long cookie; |
| struct flow_rule *rule; |
| struct flow_stats stats; |
| u32 classid; |
| }; |
| |
| enum offload_act_command { |
| FLOW_ACT_REPLACE, |
| FLOW_ACT_DESTROY, |
| FLOW_ACT_STATS, |
| }; |
| |
| struct flow_offload_action { |
| struct netlink_ext_ack *extack; /* NULL in FLOW_ACT_STATS process*/ |
| enum offload_act_command command; |
| enum flow_action_id id; |
| u32 index; |
| unsigned long cookie; |
| struct flow_stats stats; |
| struct flow_action action; |
| }; |
| |
| struct flow_offload_action *offload_action_alloc(unsigned int num_actions); |
| |
| static inline struct flow_rule * |
| flow_cls_offload_flow_rule(struct flow_cls_offload *flow_cmd) |
| { |
| return flow_cmd->rule; |
| } |
| |
| static inline void flow_block_init(struct flow_block *flow_block) |
| { |
| INIT_LIST_HEAD(&flow_block->cb_list); |
| } |
| |
| typedef int flow_indr_block_bind_cb_t(struct net_device *dev, struct Qdisc *sch, void *cb_priv, |
| enum tc_setup_type type, void *type_data, |
| void *data, |
| void (*cleanup)(struct flow_block_cb *block_cb)); |
| |
| int flow_indr_dev_register(flow_indr_block_bind_cb_t *cb, void *cb_priv); |
| void flow_indr_dev_unregister(flow_indr_block_bind_cb_t *cb, void *cb_priv, |
| void (*release)(void *cb_priv)); |
| int flow_indr_dev_setup_offload(struct net_device *dev, struct Qdisc *sch, |
| enum tc_setup_type type, void *data, |
| struct flow_block_offload *bo, |
| void (*cleanup)(struct flow_block_cb *block_cb)); |
| bool flow_indr_dev_exists(void); |
| |
| #endif /* _NET_FLOW_OFFLOAD_H */ |