| // SPDX-License-Identifier: BSD-3-Clause OR GPL-2.0 |
| /* Copyright (c) 2017-2018 Mellanox Technologies. All rights reserved */ |
| |
| #include <linux/mutex.h> |
| #include <linux/rhashtable.h> |
| #include <net/ipv6.h> |
| |
| #include "spectrum_mr.h" |
| #include "spectrum_router.h" |
| |
| struct mlxsw_sp_mr { |
| const struct mlxsw_sp_mr_ops *mr_ops; |
| void *catchall_route_priv; |
| struct delayed_work stats_update_dw; |
| struct list_head table_list; |
| struct mutex table_list_lock; /* Protects table_list */ |
| #define MLXSW_SP_MR_ROUTES_COUNTER_UPDATE_INTERVAL 5000 /* ms */ |
| unsigned long priv[0]; |
| /* priv has to be always the last item */ |
| }; |
| |
| struct mlxsw_sp_mr_vif; |
| struct mlxsw_sp_mr_vif_ops { |
| bool (*is_regular)(const struct mlxsw_sp_mr_vif *vif); |
| }; |
| |
| struct mlxsw_sp_mr_vif { |
| struct net_device *dev; |
| const struct mlxsw_sp_rif *rif; |
| unsigned long vif_flags; |
| |
| /* A list of route_vif_entry structs that point to routes that the VIF |
| * instance is used as one of the egress VIFs |
| */ |
| struct list_head route_evif_list; |
| |
| /* A list of route_vif_entry structs that point to routes that the VIF |
| * instance is used as an ingress VIF |
| */ |
| struct list_head route_ivif_list; |
| |
| /* Protocol specific operations for a VIF */ |
| const struct mlxsw_sp_mr_vif_ops *ops; |
| }; |
| |
| struct mlxsw_sp_mr_route_vif_entry { |
| struct list_head vif_node; |
| struct list_head route_node; |
| struct mlxsw_sp_mr_vif *mr_vif; |
| struct mlxsw_sp_mr_route *mr_route; |
| }; |
| |
| struct mlxsw_sp_mr_table; |
| struct mlxsw_sp_mr_table_ops { |
| bool (*is_route_valid)(const struct mlxsw_sp_mr_table *mr_table, |
| const struct mr_mfc *mfc); |
| void (*key_create)(struct mlxsw_sp_mr_table *mr_table, |
| struct mlxsw_sp_mr_route_key *key, |
| struct mr_mfc *mfc); |
| bool (*is_route_starg)(const struct mlxsw_sp_mr_table *mr_table, |
| const struct mlxsw_sp_mr_route *mr_route); |
| }; |
| |
| struct mlxsw_sp_mr_table { |
| struct list_head node; |
| enum mlxsw_sp_l3proto proto; |
| struct mlxsw_sp *mlxsw_sp; |
| u32 vr_id; |
| struct mlxsw_sp_mr_vif vifs[MAXVIFS]; |
| struct list_head route_list; |
| struct mutex route_list_lock; /* Protects route_list */ |
| struct rhashtable route_ht; |
| const struct mlxsw_sp_mr_table_ops *ops; |
| char catchall_route_priv[]; |
| /* catchall_route_priv has to be always the last item */ |
| }; |
| |
| struct mlxsw_sp_mr_route { |
| struct list_head node; |
| struct rhash_head ht_node; |
| struct mlxsw_sp_mr_route_key key; |
| enum mlxsw_sp_mr_route_action route_action; |
| u16 min_mtu; |
| struct mr_mfc *mfc; |
| void *route_priv; |
| const struct mlxsw_sp_mr_table *mr_table; |
| /* A list of route_vif_entry structs that point to the egress VIFs */ |
| struct list_head evif_list; |
| /* A route_vif_entry struct that point to the ingress VIF */ |
| struct mlxsw_sp_mr_route_vif_entry ivif; |
| }; |
| |
| static const struct rhashtable_params mlxsw_sp_mr_route_ht_params = { |
| .key_len = sizeof(struct mlxsw_sp_mr_route_key), |
| .key_offset = offsetof(struct mlxsw_sp_mr_route, key), |
| .head_offset = offsetof(struct mlxsw_sp_mr_route, ht_node), |
| .automatic_shrinking = true, |
| }; |
| |
| static bool mlxsw_sp_mr_vif_valid(const struct mlxsw_sp_mr_vif *vif) |
| { |
| return vif->ops->is_regular(vif) && vif->dev && vif->rif; |
| } |
| |
| static bool mlxsw_sp_mr_vif_exists(const struct mlxsw_sp_mr_vif *vif) |
| { |
| return vif->dev; |
| } |
| |
| static bool |
| mlxsw_sp_mr_route_ivif_in_evifs(const struct mlxsw_sp_mr_route *mr_route) |
| { |
| vifi_t ivif = mr_route->mfc->mfc_parent; |
| |
| return mr_route->mfc->mfc_un.res.ttls[ivif] != 255; |
| } |
| |
| static int |
| mlxsw_sp_mr_route_valid_evifs_num(const struct mlxsw_sp_mr_route *mr_route) |
| { |
| struct mlxsw_sp_mr_route_vif_entry *rve; |
| int valid_evifs; |
| |
| valid_evifs = 0; |
| list_for_each_entry(rve, &mr_route->evif_list, route_node) |
| if (mlxsw_sp_mr_vif_valid(rve->mr_vif)) |
| valid_evifs++; |
| return valid_evifs; |
| } |
| |
| static enum mlxsw_sp_mr_route_action |
| mlxsw_sp_mr_route_action(const struct mlxsw_sp_mr_route *mr_route) |
| { |
| struct mlxsw_sp_mr_route_vif_entry *rve; |
| |
| /* If the ingress port is not regular and resolved, trap the route */ |
| if (!mlxsw_sp_mr_vif_valid(mr_route->ivif.mr_vif)) |
| return MLXSW_SP_MR_ROUTE_ACTION_TRAP; |
| |
| /* The kernel does not match a (*,G) route that the ingress interface is |
| * not one of the egress interfaces, so trap these kind of routes. |
| */ |
| if (mr_route->mr_table->ops->is_route_starg(mr_route->mr_table, |
| mr_route) && |
| !mlxsw_sp_mr_route_ivif_in_evifs(mr_route)) |
| return MLXSW_SP_MR_ROUTE_ACTION_TRAP; |
| |
| /* If the route has no valid eVIFs, trap it. */ |
| if (!mlxsw_sp_mr_route_valid_evifs_num(mr_route)) |
| return MLXSW_SP_MR_ROUTE_ACTION_TRAP; |
| |
| /* If one of the eVIFs has no RIF, trap-and-forward the route as there |
| * is some more routing to do in software too. |
| */ |
| list_for_each_entry(rve, &mr_route->evif_list, route_node) |
| if (mlxsw_sp_mr_vif_exists(rve->mr_vif) && !rve->mr_vif->rif) |
| return MLXSW_SP_MR_ROUTE_ACTION_TRAP_AND_FORWARD; |
| |
| return MLXSW_SP_MR_ROUTE_ACTION_FORWARD; |
| } |
| |
| static enum mlxsw_sp_mr_route_prio |
| mlxsw_sp_mr_route_prio(const struct mlxsw_sp_mr_route *mr_route) |
| { |
| return mr_route->mr_table->ops->is_route_starg(mr_route->mr_table, |
| mr_route) ? |
| MLXSW_SP_MR_ROUTE_PRIO_STARG : MLXSW_SP_MR_ROUTE_PRIO_SG; |
| } |
| |
| static int mlxsw_sp_mr_route_evif_link(struct mlxsw_sp_mr_route *mr_route, |
| struct mlxsw_sp_mr_vif *mr_vif) |
| { |
| struct mlxsw_sp_mr_route_vif_entry *rve; |
| |
| rve = kzalloc(sizeof(*rve), GFP_KERNEL); |
| if (!rve) |
| return -ENOMEM; |
| rve->mr_route = mr_route; |
| rve->mr_vif = mr_vif; |
| list_add_tail(&rve->route_node, &mr_route->evif_list); |
| list_add_tail(&rve->vif_node, &mr_vif->route_evif_list); |
| return 0; |
| } |
| |
| static void |
| mlxsw_sp_mr_route_evif_unlink(struct mlxsw_sp_mr_route_vif_entry *rve) |
| { |
| list_del(&rve->route_node); |
| list_del(&rve->vif_node); |
| kfree(rve); |
| } |
| |
| static void mlxsw_sp_mr_route_ivif_link(struct mlxsw_sp_mr_route *mr_route, |
| struct mlxsw_sp_mr_vif *mr_vif) |
| { |
| mr_route->ivif.mr_route = mr_route; |
| mr_route->ivif.mr_vif = mr_vif; |
| list_add_tail(&mr_route->ivif.vif_node, &mr_vif->route_ivif_list); |
| } |
| |
| static void mlxsw_sp_mr_route_ivif_unlink(struct mlxsw_sp_mr_route *mr_route) |
| { |
| list_del(&mr_route->ivif.vif_node); |
| } |
| |
| static int |
| mlxsw_sp_mr_route_info_create(struct mlxsw_sp_mr_table *mr_table, |
| struct mlxsw_sp_mr_route *mr_route, |
| struct mlxsw_sp_mr_route_info *route_info) |
| { |
| struct mlxsw_sp_mr_route_vif_entry *rve; |
| u16 *erif_indices; |
| u16 irif_index; |
| u16 erif = 0; |
| |
| erif_indices = kmalloc_array(MAXVIFS, sizeof(*erif_indices), |
| GFP_KERNEL); |
| if (!erif_indices) |
| return -ENOMEM; |
| |
| list_for_each_entry(rve, &mr_route->evif_list, route_node) { |
| if (mlxsw_sp_mr_vif_valid(rve->mr_vif)) { |
| u16 rifi = mlxsw_sp_rif_index(rve->mr_vif->rif); |
| |
| erif_indices[erif++] = rifi; |
| } |
| } |
| |
| if (mlxsw_sp_mr_vif_valid(mr_route->ivif.mr_vif)) |
| irif_index = mlxsw_sp_rif_index(mr_route->ivif.mr_vif->rif); |
| else |
| irif_index = 0; |
| |
| route_info->irif_index = irif_index; |
| route_info->erif_indices = erif_indices; |
| route_info->min_mtu = mr_route->min_mtu; |
| route_info->route_action = mr_route->route_action; |
| route_info->erif_num = erif; |
| return 0; |
| } |
| |
| static void |
| mlxsw_sp_mr_route_info_destroy(struct mlxsw_sp_mr_route_info *route_info) |
| { |
| kfree(route_info->erif_indices); |
| } |
| |
| static int mlxsw_sp_mr_route_write(struct mlxsw_sp_mr_table *mr_table, |
| struct mlxsw_sp_mr_route *mr_route, |
| bool replace) |
| { |
| struct mlxsw_sp *mlxsw_sp = mr_table->mlxsw_sp; |
| struct mlxsw_sp_mr_route_info route_info; |
| struct mlxsw_sp_mr *mr = mlxsw_sp->mr; |
| int err; |
| |
| err = mlxsw_sp_mr_route_info_create(mr_table, mr_route, &route_info); |
| if (err) |
| return err; |
| |
| if (!replace) { |
| struct mlxsw_sp_mr_route_params route_params; |
| |
| mr_route->route_priv = kzalloc(mr->mr_ops->route_priv_size, |
| GFP_KERNEL); |
| if (!mr_route->route_priv) { |
| err = -ENOMEM; |
| goto out; |
| } |
| |
| route_params.key = mr_route->key; |
| route_params.value = route_info; |
| route_params.prio = mlxsw_sp_mr_route_prio(mr_route); |
| err = mr->mr_ops->route_create(mlxsw_sp, mr->priv, |
| mr_route->route_priv, |
| &route_params); |
| if (err) |
| kfree(mr_route->route_priv); |
| } else { |
| err = mr->mr_ops->route_update(mlxsw_sp, mr_route->route_priv, |
| &route_info); |
| } |
| out: |
| mlxsw_sp_mr_route_info_destroy(&route_info); |
| return err; |
| } |
| |
| static void mlxsw_sp_mr_route_erase(struct mlxsw_sp_mr_table *mr_table, |
| struct mlxsw_sp_mr_route *mr_route) |
| { |
| struct mlxsw_sp *mlxsw_sp = mr_table->mlxsw_sp; |
| struct mlxsw_sp_mr *mr = mlxsw_sp->mr; |
| |
| mr->mr_ops->route_destroy(mlxsw_sp, mr->priv, mr_route->route_priv); |
| kfree(mr_route->route_priv); |
| } |
| |
| static struct mlxsw_sp_mr_route * |
| mlxsw_sp_mr_route_create(struct mlxsw_sp_mr_table *mr_table, |
| struct mr_mfc *mfc) |
| { |
| struct mlxsw_sp_mr_route_vif_entry *rve, *tmp; |
| struct mlxsw_sp_mr_route *mr_route; |
| int err = 0; |
| int i; |
| |
| /* Allocate and init a new route and fill it with parameters */ |
| mr_route = kzalloc(sizeof(*mr_route), GFP_KERNEL); |
| if (!mr_route) |
| return ERR_PTR(-ENOMEM); |
| INIT_LIST_HEAD(&mr_route->evif_list); |
| |
| /* Find min_mtu and link iVIF and eVIFs */ |
| mr_route->min_mtu = ETH_MAX_MTU; |
| mr_cache_hold(mfc); |
| mr_route->mfc = mfc; |
| mr_table->ops->key_create(mr_table, &mr_route->key, mr_route->mfc); |
| |
| mr_route->mr_table = mr_table; |
| for (i = 0; i < MAXVIFS; i++) { |
| if (mfc->mfc_un.res.ttls[i] != 255) { |
| err = mlxsw_sp_mr_route_evif_link(mr_route, |
| &mr_table->vifs[i]); |
| if (err) |
| goto err; |
| if (mr_table->vifs[i].dev && |
| mr_table->vifs[i].dev->mtu < mr_route->min_mtu) |
| mr_route->min_mtu = mr_table->vifs[i].dev->mtu; |
| } |
| } |
| mlxsw_sp_mr_route_ivif_link(mr_route, |
| &mr_table->vifs[mfc->mfc_parent]); |
| |
| mr_route->route_action = mlxsw_sp_mr_route_action(mr_route); |
| return mr_route; |
| err: |
| mr_cache_put(mfc); |
| list_for_each_entry_safe(rve, tmp, &mr_route->evif_list, route_node) |
| mlxsw_sp_mr_route_evif_unlink(rve); |
| kfree(mr_route); |
| return ERR_PTR(err); |
| } |
| |
| static void mlxsw_sp_mr_route_destroy(struct mlxsw_sp_mr_table *mr_table, |
| struct mlxsw_sp_mr_route *mr_route) |
| { |
| struct mlxsw_sp_mr_route_vif_entry *rve, *tmp; |
| |
| mlxsw_sp_mr_route_ivif_unlink(mr_route); |
| mr_cache_put(mr_route->mfc); |
| list_for_each_entry_safe(rve, tmp, &mr_route->evif_list, route_node) |
| mlxsw_sp_mr_route_evif_unlink(rve); |
| kfree(mr_route); |
| } |
| |
| static void mlxsw_sp_mr_mfc_offload_set(struct mlxsw_sp_mr_route *mr_route, |
| bool offload) |
| { |
| if (offload) |
| mr_route->mfc->mfc_flags |= MFC_OFFLOAD; |
| else |
| mr_route->mfc->mfc_flags &= ~MFC_OFFLOAD; |
| } |
| |
| static void mlxsw_sp_mr_mfc_offload_update(struct mlxsw_sp_mr_route *mr_route) |
| { |
| bool offload; |
| |
| offload = mr_route->route_action != MLXSW_SP_MR_ROUTE_ACTION_TRAP; |
| mlxsw_sp_mr_mfc_offload_set(mr_route, offload); |
| } |
| |
| static void __mlxsw_sp_mr_route_del(struct mlxsw_sp_mr_table *mr_table, |
| struct mlxsw_sp_mr_route *mr_route) |
| { |
| WARN_ON_ONCE(!mutex_is_locked(&mr_table->route_list_lock)); |
| |
| mlxsw_sp_mr_mfc_offload_set(mr_route, false); |
| rhashtable_remove_fast(&mr_table->route_ht, &mr_route->ht_node, |
| mlxsw_sp_mr_route_ht_params); |
| list_del(&mr_route->node); |
| mlxsw_sp_mr_route_erase(mr_table, mr_route); |
| mlxsw_sp_mr_route_destroy(mr_table, mr_route); |
| } |
| |
| int mlxsw_sp_mr_route_add(struct mlxsw_sp_mr_table *mr_table, |
| struct mr_mfc *mfc, bool replace) |
| { |
| struct mlxsw_sp_mr_route *mr_orig_route = NULL; |
| struct mlxsw_sp_mr_route *mr_route; |
| int err; |
| |
| if (!mr_table->ops->is_route_valid(mr_table, mfc)) |
| return -EINVAL; |
| |
| /* Create a new route */ |
| mr_route = mlxsw_sp_mr_route_create(mr_table, mfc); |
| if (IS_ERR(mr_route)) |
| return PTR_ERR(mr_route); |
| |
| /* Find any route with a matching key */ |
| mr_orig_route = rhashtable_lookup_fast(&mr_table->route_ht, |
| &mr_route->key, |
| mlxsw_sp_mr_route_ht_params); |
| if (replace) { |
| /* On replace case, make the route point to the new route_priv. |
| */ |
| if (WARN_ON(!mr_orig_route)) { |
| err = -ENOENT; |
| goto err_no_orig_route; |
| } |
| mr_route->route_priv = mr_orig_route->route_priv; |
| } else if (mr_orig_route) { |
| /* On non replace case, if another route with the same key was |
| * found, abort, as duplicate routes are used for proxy routes. |
| */ |
| dev_warn(mr_table->mlxsw_sp->bus_info->dev, |
| "Offloading proxy routes is not supported.\n"); |
| err = -EINVAL; |
| goto err_duplicate_route; |
| } |
| |
| /* Write the route to the hardware */ |
| err = mlxsw_sp_mr_route_write(mr_table, mr_route, replace); |
| if (err) |
| goto err_mr_route_write; |
| |
| /* Put it in the table data-structures */ |
| mutex_lock(&mr_table->route_list_lock); |
| list_add_tail(&mr_route->node, &mr_table->route_list); |
| mutex_unlock(&mr_table->route_list_lock); |
| err = rhashtable_insert_fast(&mr_table->route_ht, |
| &mr_route->ht_node, |
| mlxsw_sp_mr_route_ht_params); |
| if (err) |
| goto err_rhashtable_insert; |
| |
| /* Destroy the original route */ |
| if (replace) { |
| rhashtable_remove_fast(&mr_table->route_ht, |
| &mr_orig_route->ht_node, |
| mlxsw_sp_mr_route_ht_params); |
| list_del(&mr_orig_route->node); |
| mlxsw_sp_mr_route_destroy(mr_table, mr_orig_route); |
| } |
| |
| mlxsw_sp_mr_mfc_offload_update(mr_route); |
| return 0; |
| |
| err_rhashtable_insert: |
| mutex_lock(&mr_table->route_list_lock); |
| list_del(&mr_route->node); |
| mutex_unlock(&mr_table->route_list_lock); |
| mlxsw_sp_mr_route_erase(mr_table, mr_route); |
| err_mr_route_write: |
| err_no_orig_route: |
| err_duplicate_route: |
| mlxsw_sp_mr_route_destroy(mr_table, mr_route); |
| return err; |
| } |
| |
| void mlxsw_sp_mr_route_del(struct mlxsw_sp_mr_table *mr_table, |
| struct mr_mfc *mfc) |
| { |
| struct mlxsw_sp_mr_route *mr_route; |
| struct mlxsw_sp_mr_route_key key; |
| |
| mr_table->ops->key_create(mr_table, &key, mfc); |
| mr_route = rhashtable_lookup_fast(&mr_table->route_ht, &key, |
| mlxsw_sp_mr_route_ht_params); |
| if (mr_route) { |
| mutex_lock(&mr_table->route_list_lock); |
| __mlxsw_sp_mr_route_del(mr_table, mr_route); |
| mutex_unlock(&mr_table->route_list_lock); |
| } |
| } |
| |
| /* Should be called after the VIF struct is updated */ |
| static int |
| mlxsw_sp_mr_route_ivif_resolve(struct mlxsw_sp_mr_table *mr_table, |
| struct mlxsw_sp_mr_route_vif_entry *rve) |
| { |
| struct mlxsw_sp *mlxsw_sp = mr_table->mlxsw_sp; |
| enum mlxsw_sp_mr_route_action route_action; |
| struct mlxsw_sp_mr *mr = mlxsw_sp->mr; |
| u16 irif_index; |
| int err; |
| |
| route_action = mlxsw_sp_mr_route_action(rve->mr_route); |
| if (route_action == MLXSW_SP_MR_ROUTE_ACTION_TRAP) |
| return 0; |
| |
| /* rve->mr_vif->rif is guaranteed to be valid at this stage */ |
| irif_index = mlxsw_sp_rif_index(rve->mr_vif->rif); |
| err = mr->mr_ops->route_irif_update(mlxsw_sp, rve->mr_route->route_priv, |
| irif_index); |
| if (err) |
| return err; |
| |
| err = mr->mr_ops->route_action_update(mlxsw_sp, |
| rve->mr_route->route_priv, |
| route_action); |
| if (err) |
| /* No need to rollback here because the iRIF change only takes |
| * place after the action has been updated. |
| */ |
| return err; |
| |
| rve->mr_route->route_action = route_action; |
| mlxsw_sp_mr_mfc_offload_update(rve->mr_route); |
| return 0; |
| } |
| |
| static void |
| mlxsw_sp_mr_route_ivif_unresolve(struct mlxsw_sp_mr_table *mr_table, |
| struct mlxsw_sp_mr_route_vif_entry *rve) |
| { |
| struct mlxsw_sp *mlxsw_sp = mr_table->mlxsw_sp; |
| struct mlxsw_sp_mr *mr = mlxsw_sp->mr; |
| |
| mr->mr_ops->route_action_update(mlxsw_sp, rve->mr_route->route_priv, |
| MLXSW_SP_MR_ROUTE_ACTION_TRAP); |
| rve->mr_route->route_action = MLXSW_SP_MR_ROUTE_ACTION_TRAP; |
| mlxsw_sp_mr_mfc_offload_update(rve->mr_route); |
| } |
| |
| /* Should be called after the RIF struct is updated */ |
| static int |
| mlxsw_sp_mr_route_evif_resolve(struct mlxsw_sp_mr_table *mr_table, |
| struct mlxsw_sp_mr_route_vif_entry *rve) |
| { |
| struct mlxsw_sp *mlxsw_sp = mr_table->mlxsw_sp; |
| enum mlxsw_sp_mr_route_action route_action; |
| struct mlxsw_sp_mr *mr = mlxsw_sp->mr; |
| u16 erif_index = 0; |
| int err; |
| |
| /* Update the route action, as the new eVIF can be a tunnel or a pimreg |
| * device which will require updating the action. |
| */ |
| route_action = mlxsw_sp_mr_route_action(rve->mr_route); |
| if (route_action != rve->mr_route->route_action) { |
| err = mr->mr_ops->route_action_update(mlxsw_sp, |
| rve->mr_route->route_priv, |
| route_action); |
| if (err) |
| return err; |
| } |
| |
| /* Add the eRIF */ |
| if (mlxsw_sp_mr_vif_valid(rve->mr_vif)) { |
| erif_index = mlxsw_sp_rif_index(rve->mr_vif->rif); |
| err = mr->mr_ops->route_erif_add(mlxsw_sp, |
| rve->mr_route->route_priv, |
| erif_index); |
| if (err) |
| goto err_route_erif_add; |
| } |
| |
| /* Update the minimum MTU */ |
| if (rve->mr_vif->dev->mtu < rve->mr_route->min_mtu) { |
| rve->mr_route->min_mtu = rve->mr_vif->dev->mtu; |
| err = mr->mr_ops->route_min_mtu_update(mlxsw_sp, |
| rve->mr_route->route_priv, |
| rve->mr_route->min_mtu); |
| if (err) |
| goto err_route_min_mtu_update; |
| } |
| |
| rve->mr_route->route_action = route_action; |
| mlxsw_sp_mr_mfc_offload_update(rve->mr_route); |
| return 0; |
| |
| err_route_min_mtu_update: |
| if (mlxsw_sp_mr_vif_valid(rve->mr_vif)) |
| mr->mr_ops->route_erif_del(mlxsw_sp, rve->mr_route->route_priv, |
| erif_index); |
| err_route_erif_add: |
| if (route_action != rve->mr_route->route_action) |
| mr->mr_ops->route_action_update(mlxsw_sp, |
| rve->mr_route->route_priv, |
| rve->mr_route->route_action); |
| return err; |
| } |
| |
| /* Should be called before the RIF struct is updated */ |
| static void |
| mlxsw_sp_mr_route_evif_unresolve(struct mlxsw_sp_mr_table *mr_table, |
| struct mlxsw_sp_mr_route_vif_entry *rve) |
| { |
| struct mlxsw_sp *mlxsw_sp = mr_table->mlxsw_sp; |
| enum mlxsw_sp_mr_route_action route_action; |
| struct mlxsw_sp_mr *mr = mlxsw_sp->mr; |
| u16 rifi; |
| |
| /* If the unresolved RIF was not valid, no need to delete it */ |
| if (!mlxsw_sp_mr_vif_valid(rve->mr_vif)) |
| return; |
| |
| /* Update the route action: if there is only one valid eVIF in the |
| * route, set the action to trap as the VIF deletion will lead to zero |
| * valid eVIFs. On any other case, use the mlxsw_sp_mr_route_action to |
| * determine the route action. |
| */ |
| if (mlxsw_sp_mr_route_valid_evifs_num(rve->mr_route) == 1) |
| route_action = MLXSW_SP_MR_ROUTE_ACTION_TRAP; |
| else |
| route_action = mlxsw_sp_mr_route_action(rve->mr_route); |
| if (route_action != rve->mr_route->route_action) |
| mr->mr_ops->route_action_update(mlxsw_sp, |
| rve->mr_route->route_priv, |
| route_action); |
| |
| /* Delete the erif from the route */ |
| rifi = mlxsw_sp_rif_index(rve->mr_vif->rif); |
| mr->mr_ops->route_erif_del(mlxsw_sp, rve->mr_route->route_priv, rifi); |
| rve->mr_route->route_action = route_action; |
| mlxsw_sp_mr_mfc_offload_update(rve->mr_route); |
| } |
| |
| static int mlxsw_sp_mr_vif_resolve(struct mlxsw_sp_mr_table *mr_table, |
| struct net_device *dev, |
| struct mlxsw_sp_mr_vif *mr_vif, |
| unsigned long vif_flags, |
| const struct mlxsw_sp_rif *rif) |
| { |
| struct mlxsw_sp_mr_route_vif_entry *irve, *erve; |
| int err; |
| |
| /* Update the VIF */ |
| mr_vif->dev = dev; |
| mr_vif->rif = rif; |
| mr_vif->vif_flags = vif_flags; |
| |
| /* Update all routes where this VIF is used as an unresolved iRIF */ |
| list_for_each_entry(irve, &mr_vif->route_ivif_list, vif_node) { |
| err = mlxsw_sp_mr_route_ivif_resolve(mr_table, irve); |
| if (err) |
| goto err_irif_unresolve; |
| } |
| |
| /* Update all routes where this VIF is used as an unresolved eRIF */ |
| list_for_each_entry(erve, &mr_vif->route_evif_list, vif_node) { |
| err = mlxsw_sp_mr_route_evif_resolve(mr_table, erve); |
| if (err) |
| goto err_erif_unresolve; |
| } |
| return 0; |
| |
| err_erif_unresolve: |
| list_for_each_entry_continue_reverse(erve, &mr_vif->route_evif_list, |
| vif_node) |
| mlxsw_sp_mr_route_evif_unresolve(mr_table, erve); |
| err_irif_unresolve: |
| list_for_each_entry_continue_reverse(irve, &mr_vif->route_ivif_list, |
| vif_node) |
| mlxsw_sp_mr_route_ivif_unresolve(mr_table, irve); |
| mr_vif->rif = NULL; |
| return err; |
| } |
| |
| static void mlxsw_sp_mr_vif_unresolve(struct mlxsw_sp_mr_table *mr_table, |
| struct net_device *dev, |
| struct mlxsw_sp_mr_vif *mr_vif) |
| { |
| struct mlxsw_sp_mr_route_vif_entry *rve; |
| |
| /* Update all routes where this VIF is used as an unresolved eRIF */ |
| list_for_each_entry(rve, &mr_vif->route_evif_list, vif_node) |
| mlxsw_sp_mr_route_evif_unresolve(mr_table, rve); |
| |
| /* Update all routes where this VIF is used as an unresolved iRIF */ |
| list_for_each_entry(rve, &mr_vif->route_ivif_list, vif_node) |
| mlxsw_sp_mr_route_ivif_unresolve(mr_table, rve); |
| |
| /* Update the VIF */ |
| mr_vif->dev = dev; |
| mr_vif->rif = NULL; |
| } |
| |
| int mlxsw_sp_mr_vif_add(struct mlxsw_sp_mr_table *mr_table, |
| struct net_device *dev, vifi_t vif_index, |
| unsigned long vif_flags, const struct mlxsw_sp_rif *rif) |
| { |
| struct mlxsw_sp_mr_vif *mr_vif = &mr_table->vifs[vif_index]; |
| |
| if (WARN_ON(vif_index >= MAXVIFS)) |
| return -EINVAL; |
| if (mr_vif->dev) |
| return -EEXIST; |
| return mlxsw_sp_mr_vif_resolve(mr_table, dev, mr_vif, vif_flags, rif); |
| } |
| |
| void mlxsw_sp_mr_vif_del(struct mlxsw_sp_mr_table *mr_table, vifi_t vif_index) |
| { |
| struct mlxsw_sp_mr_vif *mr_vif = &mr_table->vifs[vif_index]; |
| |
| if (WARN_ON(vif_index >= MAXVIFS)) |
| return; |
| if (WARN_ON(!mr_vif->dev)) |
| return; |
| mlxsw_sp_mr_vif_unresolve(mr_table, NULL, mr_vif); |
| } |
| |
| static struct mlxsw_sp_mr_vif * |
| mlxsw_sp_mr_dev_vif_lookup(struct mlxsw_sp_mr_table *mr_table, |
| const struct net_device *dev) |
| { |
| vifi_t vif_index; |
| |
| for (vif_index = 0; vif_index < MAXVIFS; vif_index++) |
| if (mr_table->vifs[vif_index].dev == dev) |
| return &mr_table->vifs[vif_index]; |
| return NULL; |
| } |
| |
| int mlxsw_sp_mr_rif_add(struct mlxsw_sp_mr_table *mr_table, |
| const struct mlxsw_sp_rif *rif) |
| { |
| const struct net_device *rif_dev = mlxsw_sp_rif_dev(rif); |
| struct mlxsw_sp_mr_vif *mr_vif; |
| |
| if (!rif_dev) |
| return 0; |
| |
| mr_vif = mlxsw_sp_mr_dev_vif_lookup(mr_table, rif_dev); |
| if (!mr_vif) |
| return 0; |
| return mlxsw_sp_mr_vif_resolve(mr_table, mr_vif->dev, mr_vif, |
| mr_vif->vif_flags, rif); |
| } |
| |
| void mlxsw_sp_mr_rif_del(struct mlxsw_sp_mr_table *mr_table, |
| const struct mlxsw_sp_rif *rif) |
| { |
| const struct net_device *rif_dev = mlxsw_sp_rif_dev(rif); |
| struct mlxsw_sp_mr_vif *mr_vif; |
| |
| if (!rif_dev) |
| return; |
| |
| mr_vif = mlxsw_sp_mr_dev_vif_lookup(mr_table, rif_dev); |
| if (!mr_vif) |
| return; |
| mlxsw_sp_mr_vif_unresolve(mr_table, mr_vif->dev, mr_vif); |
| } |
| |
| void mlxsw_sp_mr_rif_mtu_update(struct mlxsw_sp_mr_table *mr_table, |
| const struct mlxsw_sp_rif *rif, int mtu) |
| { |
| const struct net_device *rif_dev = mlxsw_sp_rif_dev(rif); |
| struct mlxsw_sp *mlxsw_sp = mr_table->mlxsw_sp; |
| struct mlxsw_sp_mr_route_vif_entry *rve; |
| struct mlxsw_sp_mr *mr = mlxsw_sp->mr; |
| struct mlxsw_sp_mr_vif *mr_vif; |
| |
| if (!rif_dev) |
| return; |
| |
| /* Search for a VIF that use that RIF */ |
| mr_vif = mlxsw_sp_mr_dev_vif_lookup(mr_table, rif_dev); |
| if (!mr_vif) |
| return; |
| |
| /* Update all the routes that uses that VIF as eVIF */ |
| list_for_each_entry(rve, &mr_vif->route_evif_list, vif_node) { |
| if (mtu < rve->mr_route->min_mtu) { |
| rve->mr_route->min_mtu = mtu; |
| mr->mr_ops->route_min_mtu_update(mlxsw_sp, |
| rve->mr_route->route_priv, |
| mtu); |
| } |
| } |
| } |
| |
| /* Protocol specific functions */ |
| static bool |
| mlxsw_sp_mr_route4_validate(const struct mlxsw_sp_mr_table *mr_table, |
| const struct mr_mfc *c) |
| { |
| struct mfc_cache *mfc = (struct mfc_cache *) c; |
| |
| /* If the route is a (*,*) route, abort, as these kind of routes are |
| * used for proxy routes. |
| */ |
| if (mfc->mfc_origin == htonl(INADDR_ANY) && |
| mfc->mfc_mcastgrp == htonl(INADDR_ANY)) { |
| dev_warn(mr_table->mlxsw_sp->bus_info->dev, |
| "Offloading proxy routes is not supported.\n"); |
| return false; |
| } |
| return true; |
| } |
| |
| static void mlxsw_sp_mr_route4_key(struct mlxsw_sp_mr_table *mr_table, |
| struct mlxsw_sp_mr_route_key *key, |
| struct mr_mfc *c) |
| { |
| const struct mfc_cache *mfc = (struct mfc_cache *) c; |
| bool starg; |
| |
| starg = (mfc->mfc_origin == htonl(INADDR_ANY)); |
| |
| memset(key, 0, sizeof(*key)); |
| key->vrid = mr_table->vr_id; |
| key->proto = MLXSW_SP_L3_PROTO_IPV4; |
| key->group.addr4 = mfc->mfc_mcastgrp; |
| key->group_mask.addr4 = htonl(0xffffffff); |
| key->source.addr4 = mfc->mfc_origin; |
| key->source_mask.addr4 = htonl(starg ? 0 : 0xffffffff); |
| } |
| |
| static bool mlxsw_sp_mr_route4_starg(const struct mlxsw_sp_mr_table *mr_table, |
| const struct mlxsw_sp_mr_route *mr_route) |
| { |
| return mr_route->key.source_mask.addr4 == htonl(INADDR_ANY); |
| } |
| |
| static bool mlxsw_sp_mr_vif4_is_regular(const struct mlxsw_sp_mr_vif *vif) |
| { |
| return !(vif->vif_flags & (VIFF_TUNNEL | VIFF_REGISTER)); |
| } |
| |
| static bool |
| mlxsw_sp_mr_route6_validate(const struct mlxsw_sp_mr_table *mr_table, |
| const struct mr_mfc *c) |
| { |
| struct mfc6_cache *mfc = (struct mfc6_cache *) c; |
| |
| /* If the route is a (*,*) route, abort, as these kind of routes are |
| * used for proxy routes. |
| */ |
| if (ipv6_addr_any(&mfc->mf6c_origin) && |
| ipv6_addr_any(&mfc->mf6c_mcastgrp)) { |
| dev_warn(mr_table->mlxsw_sp->bus_info->dev, |
| "Offloading proxy routes is not supported.\n"); |
| return false; |
| } |
| return true; |
| } |
| |
| static void mlxsw_sp_mr_route6_key(struct mlxsw_sp_mr_table *mr_table, |
| struct mlxsw_sp_mr_route_key *key, |
| struct mr_mfc *c) |
| { |
| const struct mfc6_cache *mfc = (struct mfc6_cache *) c; |
| |
| memset(key, 0, sizeof(*key)); |
| key->vrid = mr_table->vr_id; |
| key->proto = MLXSW_SP_L3_PROTO_IPV6; |
| key->group.addr6 = mfc->mf6c_mcastgrp; |
| memset(&key->group_mask.addr6, 0xff, sizeof(key->group_mask.addr6)); |
| key->source.addr6 = mfc->mf6c_origin; |
| if (!ipv6_addr_any(&mfc->mf6c_origin)) |
| memset(&key->source_mask.addr6, 0xff, |
| sizeof(key->source_mask.addr6)); |
| } |
| |
| static bool mlxsw_sp_mr_route6_starg(const struct mlxsw_sp_mr_table *mr_table, |
| const struct mlxsw_sp_mr_route *mr_route) |
| { |
| return ipv6_addr_any(&mr_route->key.source_mask.addr6); |
| } |
| |
| static bool mlxsw_sp_mr_vif6_is_regular(const struct mlxsw_sp_mr_vif *vif) |
| { |
| return !(vif->vif_flags & MIFF_REGISTER); |
| } |
| |
| static struct |
| mlxsw_sp_mr_vif_ops mlxsw_sp_mr_vif_ops_arr[] = { |
| { |
| .is_regular = mlxsw_sp_mr_vif4_is_regular, |
| }, |
| { |
| .is_regular = mlxsw_sp_mr_vif6_is_regular, |
| }, |
| }; |
| |
| static struct |
| mlxsw_sp_mr_table_ops mlxsw_sp_mr_table_ops_arr[] = { |
| { |
| .is_route_valid = mlxsw_sp_mr_route4_validate, |
| .key_create = mlxsw_sp_mr_route4_key, |
| .is_route_starg = mlxsw_sp_mr_route4_starg, |
| }, |
| { |
| .is_route_valid = mlxsw_sp_mr_route6_validate, |
| .key_create = mlxsw_sp_mr_route6_key, |
| .is_route_starg = mlxsw_sp_mr_route6_starg, |
| }, |
| |
| }; |
| |
| struct mlxsw_sp_mr_table *mlxsw_sp_mr_table_create(struct mlxsw_sp *mlxsw_sp, |
| u32 vr_id, |
| enum mlxsw_sp_l3proto proto) |
| { |
| struct mlxsw_sp_mr_route_params catchall_route_params = { |
| .prio = MLXSW_SP_MR_ROUTE_PRIO_CATCHALL, |
| .key = { |
| .vrid = vr_id, |
| .proto = proto, |
| }, |
| .value = { |
| .route_action = MLXSW_SP_MR_ROUTE_ACTION_TRAP, |
| } |
| }; |
| struct mlxsw_sp_mr *mr = mlxsw_sp->mr; |
| struct mlxsw_sp_mr_table *mr_table; |
| int err; |
| int i; |
| |
| mr_table = kzalloc(sizeof(*mr_table) + mr->mr_ops->route_priv_size, |
| GFP_KERNEL); |
| if (!mr_table) |
| return ERR_PTR(-ENOMEM); |
| |
| mr_table->vr_id = vr_id; |
| mr_table->mlxsw_sp = mlxsw_sp; |
| mr_table->proto = proto; |
| mr_table->ops = &mlxsw_sp_mr_table_ops_arr[proto]; |
| INIT_LIST_HEAD(&mr_table->route_list); |
| mutex_init(&mr_table->route_list_lock); |
| |
| err = rhashtable_init(&mr_table->route_ht, |
| &mlxsw_sp_mr_route_ht_params); |
| if (err) |
| goto err_route_rhashtable_init; |
| |
| for (i = 0; i < MAXVIFS; i++) { |
| INIT_LIST_HEAD(&mr_table->vifs[i].route_evif_list); |
| INIT_LIST_HEAD(&mr_table->vifs[i].route_ivif_list); |
| mr_table->vifs[i].ops = &mlxsw_sp_mr_vif_ops_arr[proto]; |
| } |
| |
| err = mr->mr_ops->route_create(mlxsw_sp, mr->priv, |
| mr_table->catchall_route_priv, |
| &catchall_route_params); |
| if (err) |
| goto err_ops_route_create; |
| mutex_lock(&mr->table_list_lock); |
| list_add_tail(&mr_table->node, &mr->table_list); |
| mutex_unlock(&mr->table_list_lock); |
| return mr_table; |
| |
| err_ops_route_create: |
| rhashtable_destroy(&mr_table->route_ht); |
| err_route_rhashtable_init: |
| mutex_destroy(&mr_table->route_list_lock); |
| kfree(mr_table); |
| return ERR_PTR(err); |
| } |
| |
| void mlxsw_sp_mr_table_destroy(struct mlxsw_sp_mr_table *mr_table) |
| { |
| struct mlxsw_sp *mlxsw_sp = mr_table->mlxsw_sp; |
| struct mlxsw_sp_mr *mr = mlxsw_sp->mr; |
| |
| WARN_ON(!mlxsw_sp_mr_table_empty(mr_table)); |
| mutex_lock(&mr->table_list_lock); |
| list_del(&mr_table->node); |
| mutex_unlock(&mr->table_list_lock); |
| mr->mr_ops->route_destroy(mlxsw_sp, mr->priv, |
| &mr_table->catchall_route_priv); |
| rhashtable_destroy(&mr_table->route_ht); |
| mutex_destroy(&mr_table->route_list_lock); |
| kfree(mr_table); |
| } |
| |
| void mlxsw_sp_mr_table_flush(struct mlxsw_sp_mr_table *mr_table) |
| { |
| struct mlxsw_sp_mr_route *mr_route, *tmp; |
| int i; |
| |
| mutex_lock(&mr_table->route_list_lock); |
| list_for_each_entry_safe(mr_route, tmp, &mr_table->route_list, node) |
| __mlxsw_sp_mr_route_del(mr_table, mr_route); |
| mutex_unlock(&mr_table->route_list_lock); |
| |
| for (i = 0; i < MAXVIFS; i++) { |
| mr_table->vifs[i].dev = NULL; |
| mr_table->vifs[i].rif = NULL; |
| } |
| } |
| |
| bool mlxsw_sp_mr_table_empty(const struct mlxsw_sp_mr_table *mr_table) |
| { |
| int i; |
| |
| for (i = 0; i < MAXVIFS; i++) |
| if (mr_table->vifs[i].dev) |
| return false; |
| return list_empty(&mr_table->route_list); |
| } |
| |
| static void mlxsw_sp_mr_route_stats_update(struct mlxsw_sp *mlxsw_sp, |
| struct mlxsw_sp_mr_route *mr_route) |
| { |
| struct mlxsw_sp_mr *mr = mlxsw_sp->mr; |
| u64 packets, bytes; |
| |
| if (mr_route->route_action == MLXSW_SP_MR_ROUTE_ACTION_TRAP) |
| return; |
| |
| mr->mr_ops->route_stats(mlxsw_sp, mr_route->route_priv, &packets, |
| &bytes); |
| |
| if (mr_route->mfc->mfc_un.res.pkt != packets) |
| mr_route->mfc->mfc_un.res.lastuse = jiffies; |
| mr_route->mfc->mfc_un.res.pkt = packets; |
| mr_route->mfc->mfc_un.res.bytes = bytes; |
| } |
| |
| static void mlxsw_sp_mr_stats_update(struct work_struct *work) |
| { |
| struct mlxsw_sp_mr *mr = container_of(work, struct mlxsw_sp_mr, |
| stats_update_dw.work); |
| struct mlxsw_sp_mr_table *mr_table; |
| struct mlxsw_sp_mr_route *mr_route; |
| unsigned long interval; |
| |
| mutex_lock(&mr->table_list_lock); |
| list_for_each_entry(mr_table, &mr->table_list, node) { |
| mutex_lock(&mr_table->route_list_lock); |
| list_for_each_entry(mr_route, &mr_table->route_list, node) |
| mlxsw_sp_mr_route_stats_update(mr_table->mlxsw_sp, |
| mr_route); |
| mutex_unlock(&mr_table->route_list_lock); |
| } |
| mutex_unlock(&mr->table_list_lock); |
| |
| interval = msecs_to_jiffies(MLXSW_SP_MR_ROUTES_COUNTER_UPDATE_INTERVAL); |
| mlxsw_core_schedule_dw(&mr->stats_update_dw, interval); |
| } |
| |
| int mlxsw_sp_mr_init(struct mlxsw_sp *mlxsw_sp, |
| const struct mlxsw_sp_mr_ops *mr_ops) |
| { |
| struct mlxsw_sp_mr *mr; |
| unsigned long interval; |
| int err; |
| |
| mr = kzalloc(sizeof(*mr) + mr_ops->priv_size, GFP_KERNEL); |
| if (!mr) |
| return -ENOMEM; |
| mr->mr_ops = mr_ops; |
| mlxsw_sp->mr = mr; |
| INIT_LIST_HEAD(&mr->table_list); |
| mutex_init(&mr->table_list_lock); |
| |
| err = mr_ops->init(mlxsw_sp, mr->priv); |
| if (err) |
| goto err; |
| |
| /* Create the delayed work for counter updates */ |
| INIT_DELAYED_WORK(&mr->stats_update_dw, mlxsw_sp_mr_stats_update); |
| interval = msecs_to_jiffies(MLXSW_SP_MR_ROUTES_COUNTER_UPDATE_INTERVAL); |
| mlxsw_core_schedule_dw(&mr->stats_update_dw, interval); |
| return 0; |
| err: |
| mutex_destroy(&mr->table_list_lock); |
| kfree(mr); |
| return err; |
| } |
| |
| void mlxsw_sp_mr_fini(struct mlxsw_sp *mlxsw_sp) |
| { |
| struct mlxsw_sp_mr *mr = mlxsw_sp->mr; |
| |
| cancel_delayed_work_sync(&mr->stats_update_dw); |
| mr->mr_ops->fini(mlxsw_sp, mr->priv); |
| mutex_destroy(&mr->table_list_lock); |
| kfree(mr); |
| } |