| /* |
| * drivers/net/ethernet/mellanox/mlxsw/spectrum_mr_tcam.c |
| * Copyright (c) 2017 Mellanox Technologies. All rights reserved. |
| * Copyright (c) 2017 Yotam Gigi <yotamg@mellanox.com> |
| * |
| * Redistribution and use in source and binary forms, with or without |
| * modification, are permitted provided that the following conditions are met: |
| * |
| * 1. Redistributions of source code must retain the above copyright |
| * notice, this list of conditions and the following disclaimer. |
| * 2. Redistributions in binary form must reproduce the above copyright |
| * notice, this list of conditions and the following disclaimer in the |
| * documentation and/or other materials provided with the distribution. |
| * 3. Neither the names of the copyright holders nor the names of its |
| * contributors may be used to endorse or promote products derived from |
| * this software without specific prior written permission. |
| * |
| * Alternatively, this software may be distributed under the terms of the |
| * GNU General Public License ("GPL") version 2 as published by the Free |
| * Software Foundation. |
| * |
| * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" |
| * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
| * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE |
| * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE |
| * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR |
| * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF |
| * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS |
| * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN |
| * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) |
| * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE |
| * POSSIBILITY OF SUCH DAMAGE. |
| */ |
| |
| #include <linux/kernel.h> |
| #include <linux/list.h> |
| #include <linux/netdevice.h> |
| #include <linux/parman.h> |
| |
| #include "spectrum_mr_tcam.h" |
| #include "reg.h" |
| #include "spectrum.h" |
| #include "core_acl_flex_actions.h" |
| #include "spectrum_mr.h" |
| |
| struct mlxsw_sp_mr_tcam_region { |
| struct mlxsw_sp *mlxsw_sp; |
| enum mlxsw_reg_rtar_key_type rtar_key_type; |
| struct parman *parman; |
| struct parman_prio *parman_prios; |
| }; |
| |
| struct mlxsw_sp_mr_tcam { |
| struct mlxsw_sp_mr_tcam_region ipv4_tcam_region; |
| }; |
| |
| /* This struct maps to one RIGR2 register entry */ |
| struct mlxsw_sp_mr_erif_sublist { |
| struct list_head list; |
| u32 rigr2_kvdl_index; |
| int num_erifs; |
| u16 erif_indices[MLXSW_REG_RIGR2_MAX_ERIFS]; |
| bool synced; |
| }; |
| |
| struct mlxsw_sp_mr_tcam_erif_list { |
| struct list_head erif_sublists; |
| u32 kvdl_index; |
| }; |
| |
| static bool |
| mlxsw_sp_mr_erif_sublist_full(struct mlxsw_sp *mlxsw_sp, |
| struct mlxsw_sp_mr_erif_sublist *erif_sublist) |
| { |
| int erif_list_entries = MLXSW_CORE_RES_GET(mlxsw_sp->core, |
| MC_ERIF_LIST_ENTRIES); |
| |
| return erif_sublist->num_erifs == erif_list_entries; |
| } |
| |
| static void |
| mlxsw_sp_mr_erif_list_init(struct mlxsw_sp_mr_tcam_erif_list *erif_list) |
| { |
| INIT_LIST_HEAD(&erif_list->erif_sublists); |
| } |
| |
| #define MLXSW_SP_KVDL_RIGR2_SIZE 1 |
| |
| static struct mlxsw_sp_mr_erif_sublist * |
| mlxsw_sp_mr_erif_sublist_create(struct mlxsw_sp *mlxsw_sp, |
| struct mlxsw_sp_mr_tcam_erif_list *erif_list) |
| { |
| struct mlxsw_sp_mr_erif_sublist *erif_sublist; |
| int err; |
| |
| erif_sublist = kzalloc(sizeof(*erif_sublist), GFP_KERNEL); |
| if (!erif_sublist) |
| return ERR_PTR(-ENOMEM); |
| err = mlxsw_sp_kvdl_alloc(mlxsw_sp, MLXSW_SP_KVDL_RIGR2_SIZE, |
| &erif_sublist->rigr2_kvdl_index); |
| if (err) { |
| kfree(erif_sublist); |
| return ERR_PTR(err); |
| } |
| |
| list_add_tail(&erif_sublist->list, &erif_list->erif_sublists); |
| return erif_sublist; |
| } |
| |
| static void |
| mlxsw_sp_mr_erif_sublist_destroy(struct mlxsw_sp *mlxsw_sp, |
| struct mlxsw_sp_mr_erif_sublist *erif_sublist) |
| { |
| list_del(&erif_sublist->list); |
| mlxsw_sp_kvdl_free(mlxsw_sp, erif_sublist->rigr2_kvdl_index); |
| kfree(erif_sublist); |
| } |
| |
| static int |
| mlxsw_sp_mr_erif_list_add(struct mlxsw_sp *mlxsw_sp, |
| struct mlxsw_sp_mr_tcam_erif_list *erif_list, |
| u16 erif_index) |
| { |
| struct mlxsw_sp_mr_erif_sublist *sublist; |
| |
| /* If either there is no erif_entry or the last one is full, allocate a |
| * new one. |
| */ |
| if (list_empty(&erif_list->erif_sublists)) { |
| sublist = mlxsw_sp_mr_erif_sublist_create(mlxsw_sp, erif_list); |
| if (IS_ERR(sublist)) |
| return PTR_ERR(sublist); |
| erif_list->kvdl_index = sublist->rigr2_kvdl_index; |
| } else { |
| sublist = list_last_entry(&erif_list->erif_sublists, |
| struct mlxsw_sp_mr_erif_sublist, |
| list); |
| sublist->synced = false; |
| if (mlxsw_sp_mr_erif_sublist_full(mlxsw_sp, sublist)) { |
| sublist = mlxsw_sp_mr_erif_sublist_create(mlxsw_sp, |
| erif_list); |
| if (IS_ERR(sublist)) |
| return PTR_ERR(sublist); |
| } |
| } |
| |
| /* Add the eRIF to the last entry's last index */ |
| sublist->erif_indices[sublist->num_erifs++] = erif_index; |
| return 0; |
| } |
| |
| static void |
| mlxsw_sp_mr_erif_list_flush(struct mlxsw_sp *mlxsw_sp, |
| struct mlxsw_sp_mr_tcam_erif_list *erif_list) |
| { |
| struct mlxsw_sp_mr_erif_sublist *erif_sublist, *tmp; |
| |
| list_for_each_entry_safe(erif_sublist, tmp, &erif_list->erif_sublists, |
| list) |
| mlxsw_sp_mr_erif_sublist_destroy(mlxsw_sp, erif_sublist); |
| } |
| |
| static int |
| mlxsw_sp_mr_erif_list_commit(struct mlxsw_sp *mlxsw_sp, |
| struct mlxsw_sp_mr_tcam_erif_list *erif_list) |
| { |
| struct mlxsw_sp_mr_erif_sublist *curr_sublist; |
| char rigr2_pl[MLXSW_REG_RIGR2_LEN]; |
| int err; |
| int i; |
| |
| list_for_each_entry(curr_sublist, &erif_list->erif_sublists, list) { |
| if (curr_sublist->synced) |
| continue; |
| |
| /* If the sublist is not the last one, pack the next index */ |
| if (list_is_last(&curr_sublist->list, |
| &erif_list->erif_sublists)) { |
| mlxsw_reg_rigr2_pack(rigr2_pl, |
| curr_sublist->rigr2_kvdl_index, |
| false, 0); |
| } else { |
| struct mlxsw_sp_mr_erif_sublist *next_sublist; |
| |
| next_sublist = list_next_entry(curr_sublist, list); |
| mlxsw_reg_rigr2_pack(rigr2_pl, |
| curr_sublist->rigr2_kvdl_index, |
| true, |
| next_sublist->rigr2_kvdl_index); |
| } |
| |
| /* Pack all the erifs */ |
| for (i = 0; i < curr_sublist->num_erifs; i++) { |
| u16 erif_index = curr_sublist->erif_indices[i]; |
| |
| mlxsw_reg_rigr2_erif_entry_pack(rigr2_pl, i, true, |
| erif_index); |
| } |
| |
| /* Write the entry */ |
| err = mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rigr2), |
| rigr2_pl); |
| if (err) |
| /* No need of a rollback here because this |
| * hardware entry should not be pointed yet. |
| */ |
| return err; |
| curr_sublist->synced = true; |
| } |
| return 0; |
| } |
| |
| static void mlxsw_sp_mr_erif_list_move(struct mlxsw_sp_mr_tcam_erif_list *to, |
| struct mlxsw_sp_mr_tcam_erif_list *from) |
| { |
| list_splice(&from->erif_sublists, &to->erif_sublists); |
| to->kvdl_index = from->kvdl_index; |
| } |
| |
| struct mlxsw_sp_mr_tcam_route { |
| struct mlxsw_sp_mr_tcam_erif_list erif_list; |
| struct mlxsw_afa_block *afa_block; |
| u32 counter_index; |
| struct parman_item parman_item; |
| struct parman_prio *parman_prio; |
| enum mlxsw_sp_mr_route_action action; |
| struct mlxsw_sp_mr_route_key key; |
| u16 irif_index; |
| u16 min_mtu; |
| }; |
| |
| static struct mlxsw_afa_block * |
| mlxsw_sp_mr_tcam_afa_block_create(struct mlxsw_sp *mlxsw_sp, |
| enum mlxsw_sp_mr_route_action route_action, |
| u16 irif_index, u32 counter_index, |
| u16 min_mtu, |
| struct mlxsw_sp_mr_tcam_erif_list *erif_list) |
| { |
| struct mlxsw_afa_block *afa_block; |
| int err; |
| |
| afa_block = mlxsw_afa_block_create(mlxsw_sp->afa); |
| if (!afa_block) |
| return ERR_PTR(-ENOMEM); |
| |
| err = mlxsw_afa_block_append_allocated_counter(afa_block, |
| counter_index); |
| if (err) |
| goto err; |
| |
| switch (route_action) { |
| case MLXSW_SP_MR_ROUTE_ACTION_TRAP: |
| err = mlxsw_afa_block_append_trap(afa_block, |
| MLXSW_TRAP_ID_ACL1); |
| if (err) |
| goto err; |
| break; |
| case MLXSW_SP_MR_ROUTE_ACTION_TRAP_AND_FORWARD: |
| case MLXSW_SP_MR_ROUTE_ACTION_FORWARD: |
| /* If we are about to append a multicast router action, commit |
| * the erif_list. |
| */ |
| err = mlxsw_sp_mr_erif_list_commit(mlxsw_sp, erif_list); |
| if (err) |
| goto err; |
| |
| err = mlxsw_afa_block_append_mcrouter(afa_block, irif_index, |
| min_mtu, false, |
| erif_list->kvdl_index); |
| if (err) |
| goto err; |
| |
| if (route_action == MLXSW_SP_MR_ROUTE_ACTION_TRAP_AND_FORWARD) { |
| err = mlxsw_afa_block_append_trap_and_forward(afa_block, |
| MLXSW_TRAP_ID_ACL2); |
| if (err) |
| goto err; |
| } |
| break; |
| default: |
| err = -EINVAL; |
| goto err; |
| } |
| |
| err = mlxsw_afa_block_commit(afa_block); |
| if (err) |
| goto err; |
| return afa_block; |
| err: |
| mlxsw_afa_block_destroy(afa_block); |
| return ERR_PTR(err); |
| } |
| |
| static void |
| mlxsw_sp_mr_tcam_afa_block_destroy(struct mlxsw_afa_block *afa_block) |
| { |
| mlxsw_afa_block_destroy(afa_block); |
| } |
| |
| static int mlxsw_sp_mr_tcam_route_replace(struct mlxsw_sp *mlxsw_sp, |
| struct parman_item *parman_item, |
| struct mlxsw_sp_mr_route_key *key, |
| struct mlxsw_afa_block *afa_block) |
| { |
| char rmft2_pl[MLXSW_REG_RMFT2_LEN]; |
| |
| switch (key->proto) { |
| case MLXSW_SP_L3_PROTO_IPV4: |
| mlxsw_reg_rmft2_ipv4_pack(rmft2_pl, true, parman_item->index, |
| key->vrid, |
| MLXSW_REG_RMFT2_IRIF_MASK_IGNORE, 0, |
| ntohl(key->group.addr4), |
| ntohl(key->group_mask.addr4), |
| ntohl(key->source.addr4), |
| ntohl(key->source_mask.addr4), |
| mlxsw_afa_block_first_set(afa_block)); |
| break; |
| case MLXSW_SP_L3_PROTO_IPV6: |
| default: |
| WARN_ON_ONCE(1); |
| } |
| |
| return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rmft2), rmft2_pl); |
| } |
| |
| static int mlxsw_sp_mr_tcam_route_remove(struct mlxsw_sp *mlxsw_sp, int vrid, |
| struct parman_item *parman_item) |
| { |
| char rmft2_pl[MLXSW_REG_RMFT2_LEN]; |
| |
| mlxsw_reg_rmft2_ipv4_pack(rmft2_pl, false, parman_item->index, vrid, |
| 0, 0, 0, 0, 0, 0, NULL); |
| |
| return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rmft2), rmft2_pl); |
| } |
| |
| static int |
| mlxsw_sp_mr_tcam_erif_populate(struct mlxsw_sp *mlxsw_sp, |
| struct mlxsw_sp_mr_tcam_erif_list *erif_list, |
| struct mlxsw_sp_mr_route_info *route_info) |
| { |
| int err; |
| int i; |
| |
| for (i = 0; i < route_info->erif_num; i++) { |
| u16 erif_index = route_info->erif_indices[i]; |
| |
| err = mlxsw_sp_mr_erif_list_add(mlxsw_sp, erif_list, |
| erif_index); |
| if (err) |
| return err; |
| } |
| return 0; |
| } |
| |
| static int |
| mlxsw_sp_mr_tcam_route_parman_item_add(struct mlxsw_sp_mr_tcam *mr_tcam, |
| struct mlxsw_sp_mr_tcam_route *route, |
| enum mlxsw_sp_mr_route_prio prio) |
| { |
| struct parman_prio *parman_prio = NULL; |
| int err; |
| |
| switch (route->key.proto) { |
| case MLXSW_SP_L3_PROTO_IPV4: |
| parman_prio = &mr_tcam->ipv4_tcam_region.parman_prios[prio]; |
| err = parman_item_add(mr_tcam->ipv4_tcam_region.parman, |
| parman_prio, &route->parman_item); |
| if (err) |
| return err; |
| break; |
| case MLXSW_SP_L3_PROTO_IPV6: |
| default: |
| WARN_ON_ONCE(1); |
| } |
| route->parman_prio = parman_prio; |
| return 0; |
| } |
| |
| static void |
| mlxsw_sp_mr_tcam_route_parman_item_remove(struct mlxsw_sp_mr_tcam *mr_tcam, |
| struct mlxsw_sp_mr_tcam_route *route) |
| { |
| switch (route->key.proto) { |
| case MLXSW_SP_L3_PROTO_IPV4: |
| parman_item_remove(mr_tcam->ipv4_tcam_region.parman, |
| route->parman_prio, &route->parman_item); |
| break; |
| case MLXSW_SP_L3_PROTO_IPV6: |
| default: |
| WARN_ON_ONCE(1); |
| } |
| } |
| |
| static int |
| mlxsw_sp_mr_tcam_route_create(struct mlxsw_sp *mlxsw_sp, void *priv, |
| void *route_priv, |
| struct mlxsw_sp_mr_route_params *route_params) |
| { |
| struct mlxsw_sp_mr_tcam_route *route = route_priv; |
| struct mlxsw_sp_mr_tcam *mr_tcam = priv; |
| int err; |
| |
| route->key = route_params->key; |
| route->irif_index = route_params->value.irif_index; |
| route->min_mtu = route_params->value.min_mtu; |
| route->action = route_params->value.route_action; |
| |
| /* Create the egress RIFs list */ |
| mlxsw_sp_mr_erif_list_init(&route->erif_list); |
| err = mlxsw_sp_mr_tcam_erif_populate(mlxsw_sp, &route->erif_list, |
| &route_params->value); |
| if (err) |
| goto err_erif_populate; |
| |
| /* Create the flow counter */ |
| err = mlxsw_sp_flow_counter_alloc(mlxsw_sp, &route->counter_index); |
| if (err) |
| goto err_counter_alloc; |
| |
| /* Create the flexible action block */ |
| route->afa_block = mlxsw_sp_mr_tcam_afa_block_create(mlxsw_sp, |
| route->action, |
| route->irif_index, |
| route->counter_index, |
| route->min_mtu, |
| &route->erif_list); |
| if (IS_ERR(route->afa_block)) { |
| err = PTR_ERR(route->afa_block); |
| goto err_afa_block_create; |
| } |
| |
| /* Allocate place in the TCAM */ |
| err = mlxsw_sp_mr_tcam_route_parman_item_add(mr_tcam, route, |
| route_params->prio); |
| if (err) |
| goto err_parman_item_add; |
| |
| /* Write the route to the TCAM */ |
| err = mlxsw_sp_mr_tcam_route_replace(mlxsw_sp, &route->parman_item, |
| &route->key, route->afa_block); |
| if (err) |
| goto err_route_replace; |
| return 0; |
| |
| err_route_replace: |
| mlxsw_sp_mr_tcam_route_parman_item_remove(mr_tcam, route); |
| err_parman_item_add: |
| mlxsw_sp_mr_tcam_afa_block_destroy(route->afa_block); |
| err_afa_block_create: |
| mlxsw_sp_flow_counter_free(mlxsw_sp, route->counter_index); |
| err_erif_populate: |
| err_counter_alloc: |
| mlxsw_sp_mr_erif_list_flush(mlxsw_sp, &route->erif_list); |
| return err; |
| } |
| |
| static void mlxsw_sp_mr_tcam_route_destroy(struct mlxsw_sp *mlxsw_sp, |
| void *priv, void *route_priv) |
| { |
| struct mlxsw_sp_mr_tcam_route *route = route_priv; |
| struct mlxsw_sp_mr_tcam *mr_tcam = priv; |
| |
| mlxsw_sp_mr_tcam_route_remove(mlxsw_sp, route->key.vrid, |
| &route->parman_item); |
| mlxsw_sp_mr_tcam_route_parman_item_remove(mr_tcam, route); |
| mlxsw_sp_mr_tcam_afa_block_destroy(route->afa_block); |
| mlxsw_sp_flow_counter_free(mlxsw_sp, route->counter_index); |
| mlxsw_sp_mr_erif_list_flush(mlxsw_sp, &route->erif_list); |
| } |
| |
| static int mlxsw_sp_mr_tcam_route_stats(struct mlxsw_sp *mlxsw_sp, |
| void *route_priv, u64 *packets, |
| u64 *bytes) |
| { |
| struct mlxsw_sp_mr_tcam_route *route = route_priv; |
| |
| return mlxsw_sp_flow_counter_get(mlxsw_sp, route->counter_index, |
| packets, bytes); |
| } |
| |
| static int |
| mlxsw_sp_mr_tcam_route_action_update(struct mlxsw_sp *mlxsw_sp, |
| void *route_priv, |
| enum mlxsw_sp_mr_route_action route_action) |
| { |
| struct mlxsw_sp_mr_tcam_route *route = route_priv; |
| struct mlxsw_afa_block *afa_block; |
| int err; |
| |
| /* Create a new flexible action block */ |
| afa_block = mlxsw_sp_mr_tcam_afa_block_create(mlxsw_sp, route_action, |
| route->irif_index, |
| route->counter_index, |
| route->min_mtu, |
| &route->erif_list); |
| if (IS_ERR(afa_block)) |
| return PTR_ERR(afa_block); |
| |
| /* Update the TCAM route entry */ |
| err = mlxsw_sp_mr_tcam_route_replace(mlxsw_sp, &route->parman_item, |
| &route->key, afa_block); |
| if (err) |
| goto err; |
| |
| /* Delete the old one */ |
| mlxsw_sp_mr_tcam_afa_block_destroy(route->afa_block); |
| route->afa_block = afa_block; |
| route->action = route_action; |
| return 0; |
| err: |
| mlxsw_sp_mr_tcam_afa_block_destroy(afa_block); |
| return err; |
| } |
| |
| static int mlxsw_sp_mr_tcam_route_min_mtu_update(struct mlxsw_sp *mlxsw_sp, |
| void *route_priv, u16 min_mtu) |
| { |
| struct mlxsw_sp_mr_tcam_route *route = route_priv; |
| struct mlxsw_afa_block *afa_block; |
| int err; |
| |
| /* Create a new flexible action block */ |
| afa_block = mlxsw_sp_mr_tcam_afa_block_create(mlxsw_sp, |
| route->action, |
| route->irif_index, |
| route->counter_index, |
| min_mtu, |
| &route->erif_list); |
| if (IS_ERR(afa_block)) |
| return PTR_ERR(afa_block); |
| |
| /* Update the TCAM route entry */ |
| err = mlxsw_sp_mr_tcam_route_replace(mlxsw_sp, &route->parman_item, |
| &route->key, afa_block); |
| if (err) |
| goto err; |
| |
| /* Delete the old one */ |
| mlxsw_sp_mr_tcam_afa_block_destroy(route->afa_block); |
| route->afa_block = afa_block; |
| route->min_mtu = min_mtu; |
| return 0; |
| err: |
| mlxsw_sp_mr_tcam_afa_block_destroy(afa_block); |
| return err; |
| } |
| |
| static int mlxsw_sp_mr_tcam_route_irif_update(struct mlxsw_sp *mlxsw_sp, |
| void *route_priv, u16 irif_index) |
| { |
| struct mlxsw_sp_mr_tcam_route *route = route_priv; |
| |
| if (route->action != MLXSW_SP_MR_ROUTE_ACTION_TRAP) |
| return -EINVAL; |
| route->irif_index = irif_index; |
| return 0; |
| } |
| |
| static int mlxsw_sp_mr_tcam_route_erif_add(struct mlxsw_sp *mlxsw_sp, |
| void *route_priv, u16 erif_index) |
| { |
| struct mlxsw_sp_mr_tcam_route *route = route_priv; |
| int err; |
| |
| err = mlxsw_sp_mr_erif_list_add(mlxsw_sp, &route->erif_list, |
| erif_index); |
| if (err) |
| return err; |
| |
| /* Commit the action only if the route action is not TRAP */ |
| if (route->action != MLXSW_SP_MR_ROUTE_ACTION_TRAP) |
| return mlxsw_sp_mr_erif_list_commit(mlxsw_sp, |
| &route->erif_list); |
| return 0; |
| } |
| |
| static int mlxsw_sp_mr_tcam_route_erif_del(struct mlxsw_sp *mlxsw_sp, |
| void *route_priv, u16 erif_index) |
| { |
| struct mlxsw_sp_mr_tcam_route *route = route_priv; |
| struct mlxsw_sp_mr_erif_sublist *erif_sublist; |
| struct mlxsw_sp_mr_tcam_erif_list erif_list; |
| struct mlxsw_afa_block *afa_block; |
| int err; |
| int i; |
| |
| /* Create a copy of the original erif_list without the deleted entry */ |
| mlxsw_sp_mr_erif_list_init(&erif_list); |
| list_for_each_entry(erif_sublist, &route->erif_list.erif_sublists, list) { |
| for (i = 0; i < erif_sublist->num_erifs; i++) { |
| u16 curr_erif = erif_sublist->erif_indices[i]; |
| |
| if (curr_erif == erif_index) |
| continue; |
| err = mlxsw_sp_mr_erif_list_add(mlxsw_sp, &erif_list, |
| curr_erif); |
| if (err) |
| goto err_erif_list_add; |
| } |
| } |
| |
| /* Create the flexible action block pointing to the new erif_list */ |
| afa_block = mlxsw_sp_mr_tcam_afa_block_create(mlxsw_sp, route->action, |
| route->irif_index, |
| route->counter_index, |
| route->min_mtu, |
| &erif_list); |
| if (IS_ERR(afa_block)) { |
| err = PTR_ERR(afa_block); |
| goto err_afa_block_create; |
| } |
| |
| /* Update the TCAM route entry */ |
| err = mlxsw_sp_mr_tcam_route_replace(mlxsw_sp, &route->parman_item, |
| &route->key, afa_block); |
| if (err) |
| goto err_route_write; |
| |
| mlxsw_sp_mr_tcam_afa_block_destroy(route->afa_block); |
| mlxsw_sp_mr_erif_list_flush(mlxsw_sp, &route->erif_list); |
| route->afa_block = afa_block; |
| mlxsw_sp_mr_erif_list_move(&route->erif_list, &erif_list); |
| return 0; |
| |
| err_route_write: |
| mlxsw_sp_mr_tcam_afa_block_destroy(afa_block); |
| err_afa_block_create: |
| err_erif_list_add: |
| mlxsw_sp_mr_erif_list_flush(mlxsw_sp, &erif_list); |
| return err; |
| } |
| |
| static int |
| mlxsw_sp_mr_tcam_route_update(struct mlxsw_sp *mlxsw_sp, void *route_priv, |
| struct mlxsw_sp_mr_route_info *route_info) |
| { |
| struct mlxsw_sp_mr_tcam_route *route = route_priv; |
| struct mlxsw_sp_mr_tcam_erif_list erif_list; |
| struct mlxsw_afa_block *afa_block; |
| int err; |
| |
| /* Create a new erif_list */ |
| mlxsw_sp_mr_erif_list_init(&erif_list); |
| err = mlxsw_sp_mr_tcam_erif_populate(mlxsw_sp, &erif_list, route_info); |
| if (err) |
| goto err_erif_populate; |
| |
| /* Create the flexible action block pointing to the new erif_list */ |
| afa_block = mlxsw_sp_mr_tcam_afa_block_create(mlxsw_sp, |
| route_info->route_action, |
| route_info->irif_index, |
| route->counter_index, |
| route_info->min_mtu, |
| &erif_list); |
| if (IS_ERR(afa_block)) { |
| err = PTR_ERR(afa_block); |
| goto err_afa_block_create; |
| } |
| |
| /* Update the TCAM route entry */ |
| err = mlxsw_sp_mr_tcam_route_replace(mlxsw_sp, &route->parman_item, |
| &route->key, afa_block); |
| if (err) |
| goto err_route_write; |
| |
| mlxsw_sp_mr_tcam_afa_block_destroy(route->afa_block); |
| mlxsw_sp_mr_erif_list_flush(mlxsw_sp, &route->erif_list); |
| route->afa_block = afa_block; |
| mlxsw_sp_mr_erif_list_move(&route->erif_list, &erif_list); |
| route->action = route_info->route_action; |
| route->irif_index = route_info->irif_index; |
| route->min_mtu = route_info->min_mtu; |
| return 0; |
| |
| err_route_write: |
| mlxsw_sp_mr_tcam_afa_block_destroy(afa_block); |
| err_afa_block_create: |
| err_erif_populate: |
| mlxsw_sp_mr_erif_list_flush(mlxsw_sp, &erif_list); |
| return err; |
| } |
| |
| #define MLXSW_SP_MR_TCAM_REGION_BASE_COUNT 16 |
| #define MLXSW_SP_MR_TCAM_REGION_RESIZE_STEP 16 |
| |
| static int |
| mlxsw_sp_mr_tcam_region_alloc(struct mlxsw_sp_mr_tcam_region *mr_tcam_region) |
| { |
| struct mlxsw_sp *mlxsw_sp = mr_tcam_region->mlxsw_sp; |
| char rtar_pl[MLXSW_REG_RTAR_LEN]; |
| |
| mlxsw_reg_rtar_pack(rtar_pl, MLXSW_REG_RTAR_OP_ALLOCATE, |
| mr_tcam_region->rtar_key_type, |
| MLXSW_SP_MR_TCAM_REGION_BASE_COUNT); |
| return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rtar), rtar_pl); |
| } |
| |
| static void |
| mlxsw_sp_mr_tcam_region_free(struct mlxsw_sp_mr_tcam_region *mr_tcam_region) |
| { |
| struct mlxsw_sp *mlxsw_sp = mr_tcam_region->mlxsw_sp; |
| char rtar_pl[MLXSW_REG_RTAR_LEN]; |
| |
| mlxsw_reg_rtar_pack(rtar_pl, MLXSW_REG_RTAR_OP_DEALLOCATE, |
| mr_tcam_region->rtar_key_type, 0); |
| mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rtar), rtar_pl); |
| } |
| |
| static int mlxsw_sp_mr_tcam_region_parman_resize(void *priv, |
| unsigned long new_count) |
| { |
| struct mlxsw_sp_mr_tcam_region *mr_tcam_region = priv; |
| struct mlxsw_sp *mlxsw_sp = mr_tcam_region->mlxsw_sp; |
| char rtar_pl[MLXSW_REG_RTAR_LEN]; |
| u64 max_tcam_rules; |
| |
| max_tcam_rules = MLXSW_CORE_RES_GET(mlxsw_sp->core, ACL_MAX_TCAM_RULES); |
| if (new_count > max_tcam_rules) |
| return -EINVAL; |
| mlxsw_reg_rtar_pack(rtar_pl, MLXSW_REG_RTAR_OP_RESIZE, |
| mr_tcam_region->rtar_key_type, new_count); |
| return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rtar), rtar_pl); |
| } |
| |
| static void mlxsw_sp_mr_tcam_region_parman_move(void *priv, |
| unsigned long from_index, |
| unsigned long to_index, |
| unsigned long count) |
| { |
| struct mlxsw_sp_mr_tcam_region *mr_tcam_region = priv; |
| struct mlxsw_sp *mlxsw_sp = mr_tcam_region->mlxsw_sp; |
| char rrcr_pl[MLXSW_REG_RRCR_LEN]; |
| |
| mlxsw_reg_rrcr_pack(rrcr_pl, MLXSW_REG_RRCR_OP_MOVE, |
| from_index, count, |
| mr_tcam_region->rtar_key_type, to_index); |
| mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rrcr), rrcr_pl); |
| } |
| |
| static const struct parman_ops mlxsw_sp_mr_tcam_region_parman_ops = { |
| .base_count = MLXSW_SP_MR_TCAM_REGION_BASE_COUNT, |
| .resize_step = MLXSW_SP_MR_TCAM_REGION_RESIZE_STEP, |
| .resize = mlxsw_sp_mr_tcam_region_parman_resize, |
| .move = mlxsw_sp_mr_tcam_region_parman_move, |
| .algo = PARMAN_ALGO_TYPE_LSORT, |
| }; |
| |
| static int |
| mlxsw_sp_mr_tcam_region_init(struct mlxsw_sp *mlxsw_sp, |
| struct mlxsw_sp_mr_tcam_region *mr_tcam_region, |
| enum mlxsw_reg_rtar_key_type rtar_key_type) |
| { |
| struct parman_prio *parman_prios; |
| struct parman *parman; |
| int err; |
| int i; |
| |
| mr_tcam_region->rtar_key_type = rtar_key_type; |
| mr_tcam_region->mlxsw_sp = mlxsw_sp; |
| |
| err = mlxsw_sp_mr_tcam_region_alloc(mr_tcam_region); |
| if (err) |
| return err; |
| |
| parman = parman_create(&mlxsw_sp_mr_tcam_region_parman_ops, |
| mr_tcam_region); |
| if (!parman) { |
| err = -ENOMEM; |
| goto err_parman_create; |
| } |
| mr_tcam_region->parman = parman; |
| |
| parman_prios = kmalloc_array(MLXSW_SP_MR_ROUTE_PRIO_MAX + 1, |
| sizeof(*parman_prios), GFP_KERNEL); |
| if (!parman_prios) { |
| err = -ENOMEM; |
| goto err_parman_prios_alloc; |
| } |
| mr_tcam_region->parman_prios = parman_prios; |
| |
| for (i = 0; i < MLXSW_SP_MR_ROUTE_PRIO_MAX + 1; i++) |
| parman_prio_init(mr_tcam_region->parman, |
| &mr_tcam_region->parman_prios[i], i); |
| return 0; |
| |
| err_parman_prios_alloc: |
| parman_destroy(parman); |
| err_parman_create: |
| mlxsw_sp_mr_tcam_region_free(mr_tcam_region); |
| return err; |
| } |
| |
| static void |
| mlxsw_sp_mr_tcam_region_fini(struct mlxsw_sp_mr_tcam_region *mr_tcam_region) |
| { |
| int i; |
| |
| for (i = 0; i < MLXSW_SP_MR_ROUTE_PRIO_MAX + 1; i++) |
| parman_prio_fini(&mr_tcam_region->parman_prios[i]); |
| kfree(mr_tcam_region->parman_prios); |
| parman_destroy(mr_tcam_region->parman); |
| mlxsw_sp_mr_tcam_region_free(mr_tcam_region); |
| } |
| |
| static int mlxsw_sp_mr_tcam_init(struct mlxsw_sp *mlxsw_sp, void *priv) |
| { |
| struct mlxsw_sp_mr_tcam *mr_tcam = priv; |
| |
| if (!MLXSW_CORE_RES_VALID(mlxsw_sp->core, MC_ERIF_LIST_ENTRIES) || |
| !MLXSW_CORE_RES_VALID(mlxsw_sp->core, ACL_MAX_TCAM_RULES)) |
| return -EIO; |
| |
| return mlxsw_sp_mr_tcam_region_init(mlxsw_sp, |
| &mr_tcam->ipv4_tcam_region, |
| MLXSW_REG_RTAR_KEY_TYPE_IPV4_MULTICAST); |
| } |
| |
| static void mlxsw_sp_mr_tcam_fini(void *priv) |
| { |
| struct mlxsw_sp_mr_tcam *mr_tcam = priv; |
| |
| mlxsw_sp_mr_tcam_region_fini(&mr_tcam->ipv4_tcam_region); |
| } |
| |
| const struct mlxsw_sp_mr_ops mlxsw_sp_mr_tcam_ops = { |
| .priv_size = sizeof(struct mlxsw_sp_mr_tcam), |
| .route_priv_size = sizeof(struct mlxsw_sp_mr_tcam_route), |
| .init = mlxsw_sp_mr_tcam_init, |
| .route_create = mlxsw_sp_mr_tcam_route_create, |
| .route_update = mlxsw_sp_mr_tcam_route_update, |
| .route_stats = mlxsw_sp_mr_tcam_route_stats, |
| .route_action_update = mlxsw_sp_mr_tcam_route_action_update, |
| .route_min_mtu_update = mlxsw_sp_mr_tcam_route_min_mtu_update, |
| .route_irif_update = mlxsw_sp_mr_tcam_route_irif_update, |
| .route_erif_add = mlxsw_sp_mr_tcam_route_erif_add, |
| .route_erif_del = mlxsw_sp_mr_tcam_route_erif_del, |
| .route_destroy = mlxsw_sp_mr_tcam_route_destroy, |
| .fini = mlxsw_sp_mr_tcam_fini, |
| }; |