/* * drivers/net/ethernet/mellanox/mlxsw/spectrum_mr_tcam.c * Copyright (c) 2017 Mellanox Technologies. All rights reserved. * Copyright (c) 2017 Yotam Gigi * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are met: * * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. Neither the names of the copyright holders nor the names of its * contributors may be used to endorse or promote products derived from * this software without specific prior written permission. * * Alternatively, this software may be distributed under the terms of the * GNU General Public License ("GPL") version 2 as published by the Free * Software Foundation. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE * POSSIBILITY OF SUCH DAMAGE. */ #include #include #include #include #include "spectrum_mr_tcam.h" #include "reg.h" #include "spectrum.h" #include "core_acl_flex_actions.h" #include "spectrum_mr.h" struct mlxsw_sp_mr_tcam_region { struct mlxsw_sp *mlxsw_sp; enum mlxsw_reg_rtar_key_type rtar_key_type; struct parman *parman; struct parman_prio *parman_prios; }; struct mlxsw_sp_mr_tcam { struct mlxsw_sp_mr_tcam_region ipv4_tcam_region; }; /* This struct maps to one RIGR2 register entry */ struct mlxsw_sp_mr_erif_sublist { struct list_head list; u32 rigr2_kvdl_index; int num_erifs; u16 erif_indices[MLXSW_REG_RIGR2_MAX_ERIFS]; bool synced; }; struct mlxsw_sp_mr_tcam_erif_list { struct list_head erif_sublists; u32 kvdl_index; }; static bool mlxsw_sp_mr_erif_sublist_full(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_mr_erif_sublist *erif_sublist) { int erif_list_entries = MLXSW_CORE_RES_GET(mlxsw_sp->core, MC_ERIF_LIST_ENTRIES); return erif_sublist->num_erifs == erif_list_entries; } static void mlxsw_sp_mr_erif_list_init(struct mlxsw_sp_mr_tcam_erif_list *erif_list) { INIT_LIST_HEAD(&erif_list->erif_sublists); } #define MLXSW_SP_KVDL_RIGR2_SIZE 1 static struct mlxsw_sp_mr_erif_sublist * mlxsw_sp_mr_erif_sublist_create(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_mr_tcam_erif_list *erif_list) { struct mlxsw_sp_mr_erif_sublist *erif_sublist; int err; erif_sublist = kzalloc(sizeof(*erif_sublist), GFP_KERNEL); if (!erif_sublist) return ERR_PTR(-ENOMEM); err = mlxsw_sp_kvdl_alloc(mlxsw_sp, MLXSW_SP_KVDL_RIGR2_SIZE, &erif_sublist->rigr2_kvdl_index); if (err) { kfree(erif_sublist); return ERR_PTR(err); } list_add_tail(&erif_sublist->list, &erif_list->erif_sublists); return erif_sublist; } static void mlxsw_sp_mr_erif_sublist_destroy(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_mr_erif_sublist *erif_sublist) { list_del(&erif_sublist->list); mlxsw_sp_kvdl_free(mlxsw_sp, erif_sublist->rigr2_kvdl_index); kfree(erif_sublist); } static int mlxsw_sp_mr_erif_list_add(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_mr_tcam_erif_list *erif_list, u16 erif_index) { struct mlxsw_sp_mr_erif_sublist *sublist; /* If either there is no erif_entry or the last one is full, allocate a * new one. */ if (list_empty(&erif_list->erif_sublists)) { sublist = mlxsw_sp_mr_erif_sublist_create(mlxsw_sp, erif_list); if (IS_ERR(sublist)) return PTR_ERR(sublist); erif_list->kvdl_index = sublist->rigr2_kvdl_index; } else { sublist = list_last_entry(&erif_list->erif_sublists, struct mlxsw_sp_mr_erif_sublist, list); sublist->synced = false; if (mlxsw_sp_mr_erif_sublist_full(mlxsw_sp, sublist)) { sublist = mlxsw_sp_mr_erif_sublist_create(mlxsw_sp, erif_list); if (IS_ERR(sublist)) return PTR_ERR(sublist); } } /* Add the eRIF to the last entry's last index */ sublist->erif_indices[sublist->num_erifs++] = erif_index; return 0; } static void mlxsw_sp_mr_erif_list_flush(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_mr_tcam_erif_list *erif_list) { struct mlxsw_sp_mr_erif_sublist *erif_sublist, *tmp; list_for_each_entry_safe(erif_sublist, tmp, &erif_list->erif_sublists, list) mlxsw_sp_mr_erif_sublist_destroy(mlxsw_sp, erif_sublist); } static int mlxsw_sp_mr_erif_list_commit(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_mr_tcam_erif_list *erif_list) { struct mlxsw_sp_mr_erif_sublist *curr_sublist; char rigr2_pl[MLXSW_REG_RIGR2_LEN]; int err; int i; list_for_each_entry(curr_sublist, &erif_list->erif_sublists, list) { if (curr_sublist->synced) continue; /* If the sublist is not the last one, pack the next index */ if (list_is_last(&curr_sublist->list, &erif_list->erif_sublists)) { mlxsw_reg_rigr2_pack(rigr2_pl, curr_sublist->rigr2_kvdl_index, false, 0); } else { struct mlxsw_sp_mr_erif_sublist *next_sublist; next_sublist = list_next_entry(curr_sublist, list); mlxsw_reg_rigr2_pack(rigr2_pl, curr_sublist->rigr2_kvdl_index, true, next_sublist->rigr2_kvdl_index); } /* Pack all the erifs */ for (i = 0; i < curr_sublist->num_erifs; i++) { u16 erif_index = curr_sublist->erif_indices[i]; mlxsw_reg_rigr2_erif_entry_pack(rigr2_pl, i, true, erif_index); } /* Write the entry */ err = mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rigr2), rigr2_pl); if (err) /* No need of a rollback here because this * hardware entry should not be pointed yet. */ return err; curr_sublist->synced = true; } return 0; } static void mlxsw_sp_mr_erif_list_move(struct mlxsw_sp_mr_tcam_erif_list *to, struct mlxsw_sp_mr_tcam_erif_list *from) { list_splice(&from->erif_sublists, &to->erif_sublists); to->kvdl_index = from->kvdl_index; } struct mlxsw_sp_mr_tcam_route { struct mlxsw_sp_mr_tcam_erif_list erif_list; struct mlxsw_afa_block *afa_block; u32 counter_index; struct parman_item parman_item; struct parman_prio *parman_prio; enum mlxsw_sp_mr_route_action action; struct mlxsw_sp_mr_route_key key; u16 irif_index; u16 min_mtu; }; static struct mlxsw_afa_block * mlxsw_sp_mr_tcam_afa_block_create(struct mlxsw_sp *mlxsw_sp, enum mlxsw_sp_mr_route_action route_action, u16 irif_index, u32 counter_index, u16 min_mtu, struct mlxsw_sp_mr_tcam_erif_list *erif_list) { struct mlxsw_afa_block *afa_block; int err; afa_block = mlxsw_afa_block_create(mlxsw_sp->afa); if (!afa_block) return ERR_PTR(-ENOMEM); err = mlxsw_afa_block_append_counter(afa_block, counter_index); if (err) goto err; switch (route_action) { case MLXSW_SP_MR_ROUTE_ACTION_TRAP: err = mlxsw_afa_block_append_trap(afa_block, MLXSW_TRAP_ID_ACL1); if (err) goto err; break; case MLXSW_SP_MR_ROUTE_ACTION_TRAP_AND_FORWARD: case MLXSW_SP_MR_ROUTE_ACTION_FORWARD: /* If we are about to append a multicast router action, commit * the erif_list. */ err = mlxsw_sp_mr_erif_list_commit(mlxsw_sp, erif_list); if (err) goto err; err = mlxsw_afa_block_append_mcrouter(afa_block, irif_index, min_mtu, false, erif_list->kvdl_index); if (err) goto err; if (route_action == MLXSW_SP_MR_ROUTE_ACTION_TRAP_AND_FORWARD) { err = mlxsw_afa_block_append_trap_and_forward(afa_block, MLXSW_TRAP_ID_ACL2); if (err) goto err; } break; default: err = -EINVAL; goto err; } err = mlxsw_afa_block_commit(afa_block); if (err) goto err; return afa_block; err: mlxsw_afa_block_destroy(afa_block); return ERR_PTR(err); } static void mlxsw_sp_mr_tcam_afa_block_destroy(struct mlxsw_afa_block *afa_block) { mlxsw_afa_block_destroy(afa_block); } static int mlxsw_sp_mr_tcam_route_replace(struct mlxsw_sp *mlxsw_sp, struct parman_item *parman_item, struct mlxsw_sp_mr_route_key *key, struct mlxsw_afa_block *afa_block) { char rmft2_pl[MLXSW_REG_RMFT2_LEN]; switch (key->proto) { case MLXSW_SP_L3_PROTO_IPV4: mlxsw_reg_rmft2_ipv4_pack(rmft2_pl, true, parman_item->index, key->vrid, MLXSW_REG_RMFT2_IRIF_MASK_IGNORE, 0, ntohl(key->group.addr4), ntohl(key->group_mask.addr4), ntohl(key->source.addr4), ntohl(key->source_mask.addr4), mlxsw_afa_block_first_set(afa_block)); break; case MLXSW_SP_L3_PROTO_IPV6: default: WARN_ON_ONCE(1); } return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rmft2), rmft2_pl); } static int mlxsw_sp_mr_tcam_route_remove(struct mlxsw_sp *mlxsw_sp, int vrid, struct parman_item *parman_item) { char rmft2_pl[MLXSW_REG_RMFT2_LEN]; mlxsw_reg_rmft2_ipv4_pack(rmft2_pl, false, parman_item->index, vrid, 0, 0, 0, 0, 0, 0, NULL); return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rmft2), rmft2_pl); } static int mlxsw_sp_mr_tcam_erif_populate(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_mr_tcam_erif_list *erif_list, struct mlxsw_sp_mr_route_info *route_info) { int err; int i; for (i = 0; i < route_info->erif_num; i++) { u16 erif_index = route_info->erif_indices[i]; err = mlxsw_sp_mr_erif_list_add(mlxsw_sp, erif_list, erif_index); if (err) return err; } return 0; } static int mlxsw_sp_mr_tcam_route_parman_item_add(struct mlxsw_sp_mr_tcam *mr_tcam, struct mlxsw_sp_mr_tcam_route *route, enum mlxsw_sp_mr_route_prio prio) { struct parman_prio *parman_prio = NULL; int err; switch (route->key.proto) { case MLXSW_SP_L3_PROTO_IPV4: parman_prio = &mr_tcam->ipv4_tcam_region.parman_prios[prio]; err = parman_item_add(mr_tcam->ipv4_tcam_region.parman, parman_prio, &route->parman_item); if (err) return err; break; case MLXSW_SP_L3_PROTO_IPV6: default: WARN_ON_ONCE(1); } route->parman_prio = parman_prio; return 0; } static void mlxsw_sp_mr_tcam_route_parman_item_remove(struct mlxsw_sp_mr_tcam *mr_tcam, struct mlxsw_sp_mr_tcam_route *route) { switch (route->key.proto) { case MLXSW_SP_L3_PROTO_IPV4: parman_item_remove(mr_tcam->ipv4_tcam_region.parman, route->parman_prio, &route->parman_item); break; case MLXSW_SP_L3_PROTO_IPV6: default: WARN_ON_ONCE(1); } } static int mlxsw_sp_mr_tcam_route_create(struct mlxsw_sp *mlxsw_sp, void *priv, void *route_priv, struct mlxsw_sp_mr_route_params *route_params) { struct mlxsw_sp_mr_tcam_route *route = route_priv; struct mlxsw_sp_mr_tcam *mr_tcam = priv; int err; route->key = route_params->key; route->irif_index = route_params->value.irif_index; route->min_mtu = route_params->value.min_mtu; route->action = route_params->value.route_action; /* Create the egress RIFs list */ mlxsw_sp_mr_erif_list_init(&route->erif_list); err = mlxsw_sp_mr_tcam_erif_populate(mlxsw_sp, &route->erif_list, &route_params->value); if (err) goto err_erif_populate; /* Create the flow counter */ err = mlxsw_sp_flow_counter_alloc(mlxsw_sp, &route->counter_index); if (err) goto err_counter_alloc; /* Create the flexible action block */ route->afa_block = mlxsw_sp_mr_tcam_afa_block_create(mlxsw_sp, route->action, route->irif_index, route->counter_index, route->min_mtu, &route->erif_list); if (IS_ERR(route->afa_block)) { err = PTR_ERR(route->afa_block); goto err_afa_block_create; } /* Allocate place in the TCAM */ err = mlxsw_sp_mr_tcam_route_parman_item_add(mr_tcam, route, route_params->prio); if (err) goto err_parman_item_add; /* Write the route to the TCAM */ err = mlxsw_sp_mr_tcam_route_replace(mlxsw_sp, &route->parman_item, &route->key, route->afa_block); if (err) goto err_route_replace; return 0; err_route_replace: mlxsw_sp_mr_tcam_route_parman_item_remove(mr_tcam, route); err_parman_item_add: mlxsw_sp_mr_tcam_afa_block_destroy(route->afa_block); err_afa_block_create: mlxsw_sp_flow_counter_free(mlxsw_sp, route->counter_index); err_erif_populate: err_counter_alloc: mlxsw_sp_mr_erif_list_flush(mlxsw_sp, &route->erif_list); return err; } static void mlxsw_sp_mr_tcam_route_destroy(struct mlxsw_sp *mlxsw_sp, void *priv, void *route_priv) { struct mlxsw_sp_mr_tcam_route *route = route_priv; struct mlxsw_sp_mr_tcam *mr_tcam = priv; mlxsw_sp_mr_tcam_route_remove(mlxsw_sp, route->key.vrid, &route->parman_item); mlxsw_sp_mr_tcam_route_parman_item_remove(mr_tcam, route); mlxsw_sp_mr_tcam_afa_block_destroy(route->afa_block); mlxsw_sp_flow_counter_free(mlxsw_sp, route->counter_index); mlxsw_sp_mr_erif_list_flush(mlxsw_sp, &route->erif_list); } static int mlxsw_sp_mr_tcam_route_stats(struct mlxsw_sp *mlxsw_sp, void *route_priv, u64 *packets, u64 *bytes) { struct mlxsw_sp_mr_tcam_route *route = route_priv; return mlxsw_sp_flow_counter_get(mlxsw_sp, route->counter_index, packets, bytes); } static int mlxsw_sp_mr_tcam_route_action_update(struct mlxsw_sp *mlxsw_sp, void *route_priv, enum mlxsw_sp_mr_route_action route_action) { struct mlxsw_sp_mr_tcam_route *route = route_priv; struct mlxsw_afa_block *afa_block; int err; /* Create a new flexible action block */ afa_block = mlxsw_sp_mr_tcam_afa_block_create(mlxsw_sp, route_action, route->irif_index, route->counter_index, route->min_mtu, &route->erif_list); if (IS_ERR(afa_block)) return PTR_ERR(afa_block); /* Update the TCAM route entry */ err = mlxsw_sp_mr_tcam_route_replace(mlxsw_sp, &route->parman_item, &route->key, afa_block); if (err) goto err; /* Delete the old one */ mlxsw_sp_mr_tcam_afa_block_destroy(route->afa_block); route->afa_block = afa_block; route->action = route_action; return 0; err: mlxsw_sp_mr_tcam_afa_block_destroy(afa_block); return err; } static int mlxsw_sp_mr_tcam_route_min_mtu_update(struct mlxsw_sp *mlxsw_sp, void *route_priv, u16 min_mtu) { struct mlxsw_sp_mr_tcam_route *route = route_priv; struct mlxsw_afa_block *afa_block; int err; /* Create a new flexible action block */ afa_block = mlxsw_sp_mr_tcam_afa_block_create(mlxsw_sp, route->action, route->irif_index, route->counter_index, min_mtu, &route->erif_list); if (IS_ERR(afa_block)) return PTR_ERR(afa_block); /* Update the TCAM route entry */ err = mlxsw_sp_mr_tcam_route_replace(mlxsw_sp, &route->parman_item, &route->key, afa_block); if (err) goto err; /* Delete the old one */ mlxsw_sp_mr_tcam_afa_block_destroy(route->afa_block); route->afa_block = afa_block; route->min_mtu = min_mtu; return 0; err: mlxsw_sp_mr_tcam_afa_block_destroy(afa_block); return err; } static int mlxsw_sp_mr_tcam_route_irif_update(struct mlxsw_sp *mlxsw_sp, void *route_priv, u16 irif_index) { struct mlxsw_sp_mr_tcam_route *route = route_priv; if (route->action != MLXSW_SP_MR_ROUTE_ACTION_TRAP) return -EINVAL; route->irif_index = irif_index; return 0; } static int mlxsw_sp_mr_tcam_route_erif_add(struct mlxsw_sp *mlxsw_sp, void *route_priv, u16 erif_index) { struct mlxsw_sp_mr_tcam_route *route = route_priv; int err; err = mlxsw_sp_mr_erif_list_add(mlxsw_sp, &route->erif_list, erif_index); if (err) return err; /* Commit the action only if the route action is not TRAP */ if (route->action != MLXSW_SP_MR_ROUTE_ACTION_TRAP) return mlxsw_sp_mr_erif_list_commit(mlxsw_sp, &route->erif_list); return 0; } static int mlxsw_sp_mr_tcam_route_erif_del(struct mlxsw_sp *mlxsw_sp, void *route_priv, u16 erif_index) { struct mlxsw_sp_mr_tcam_route *route = route_priv; struct mlxsw_sp_mr_erif_sublist *erif_sublist; struct mlxsw_sp_mr_tcam_erif_list erif_list; struct mlxsw_afa_block *afa_block; int err; int i; /* Create a copy of the original erif_list without the deleted entry */ mlxsw_sp_mr_erif_list_init(&erif_list); list_for_each_entry(erif_sublist, &route->erif_list.erif_sublists, list) { for (i = 0; i < erif_sublist->num_erifs; i++) { u16 curr_erif = erif_sublist->erif_indices[i]; if (curr_erif == erif_index) continue; err = mlxsw_sp_mr_erif_list_add(mlxsw_sp, &erif_list, curr_erif); if (err) goto err_erif_list_add; } } /* Create the flexible action block pointing to the new erif_list */ afa_block = mlxsw_sp_mr_tcam_afa_block_create(mlxsw_sp, route->action, route->irif_index, route->counter_index, route->min_mtu, &erif_list); if (IS_ERR(afa_block)) { err = PTR_ERR(afa_block); goto err_afa_block_create; } /* Update the TCAM route entry */ err = mlxsw_sp_mr_tcam_route_replace(mlxsw_sp, &route->parman_item, &route->key, afa_block); if (err) goto err_route_write; mlxsw_sp_mr_tcam_afa_block_destroy(route->afa_block); mlxsw_sp_mr_erif_list_flush(mlxsw_sp, &route->erif_list); route->afa_block = afa_block; mlxsw_sp_mr_erif_list_move(&route->erif_list, &erif_list); return 0; err_route_write: mlxsw_sp_mr_tcam_afa_block_destroy(afa_block); err_afa_block_create: err_erif_list_add: mlxsw_sp_mr_erif_list_flush(mlxsw_sp, &erif_list); return err; } static int mlxsw_sp_mr_tcam_route_update(struct mlxsw_sp *mlxsw_sp, void *route_priv, struct mlxsw_sp_mr_route_info *route_info) { struct mlxsw_sp_mr_tcam_route *route = route_priv; struct mlxsw_sp_mr_tcam_erif_list erif_list; struct mlxsw_afa_block *afa_block; int err; /* Create a new erif_list */ mlxsw_sp_mr_erif_list_init(&erif_list); err = mlxsw_sp_mr_tcam_erif_populate(mlxsw_sp, &erif_list, route_info); if (err) goto err_erif_populate; /* Create the flexible action block pointing to the new erif_list */ afa_block = mlxsw_sp_mr_tcam_afa_block_create(mlxsw_sp, route_info->route_action, route_info->irif_index, route->counter_index, route_info->min_mtu, &erif_list); if (IS_ERR(afa_block)) { err = PTR_ERR(afa_block); goto err_afa_block_create; } /* Update the TCAM route entry */ err = mlxsw_sp_mr_tcam_route_replace(mlxsw_sp, &route->parman_item, &route->key, afa_block); if (err) goto err_route_write; mlxsw_sp_mr_tcam_afa_block_destroy(route->afa_block); mlxsw_sp_mr_erif_list_flush(mlxsw_sp, &route->erif_list); route->afa_block = afa_block; mlxsw_sp_mr_erif_list_move(&route->erif_list, &erif_list); route->action = route_info->route_action; route->irif_index = route_info->irif_index; route->min_mtu = route_info->min_mtu; return 0; err_route_write: mlxsw_sp_mr_tcam_afa_block_destroy(afa_block); err_afa_block_create: err_erif_populate: mlxsw_sp_mr_erif_list_flush(mlxsw_sp, &erif_list); return err; } #define MLXSW_SP_MR_TCAM_REGION_BASE_COUNT 16 #define MLXSW_SP_MR_TCAM_REGION_RESIZE_STEP 16 static int mlxsw_sp_mr_tcam_region_alloc(struct mlxsw_sp_mr_tcam_region *mr_tcam_region) { struct mlxsw_sp *mlxsw_sp = mr_tcam_region->mlxsw_sp; char rtar_pl[MLXSW_REG_RTAR_LEN]; mlxsw_reg_rtar_pack(rtar_pl, MLXSW_REG_RTAR_OP_ALLOCATE, mr_tcam_region->rtar_key_type, MLXSW_SP_MR_TCAM_REGION_BASE_COUNT); return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rtar), rtar_pl); } static void mlxsw_sp_mr_tcam_region_free(struct mlxsw_sp_mr_tcam_region *mr_tcam_region) { struct mlxsw_sp *mlxsw_sp = mr_tcam_region->mlxsw_sp; char rtar_pl[MLXSW_REG_RTAR_LEN]; mlxsw_reg_rtar_pack(rtar_pl, MLXSW_REG_RTAR_OP_DEALLOCATE, mr_tcam_region->rtar_key_type, 0); mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rtar), rtar_pl); } static int mlxsw_sp_mr_tcam_region_parman_resize(void *priv, unsigned long new_count) { struct mlxsw_sp_mr_tcam_region *mr_tcam_region = priv; struct mlxsw_sp *mlxsw_sp = mr_tcam_region->mlxsw_sp; char rtar_pl[MLXSW_REG_RTAR_LEN]; u64 max_tcam_rules; max_tcam_rules = MLXSW_CORE_RES_GET(mlxsw_sp->core, ACL_MAX_TCAM_RULES); if (new_count > max_tcam_rules) return -EINVAL; mlxsw_reg_rtar_pack(rtar_pl, MLXSW_REG_RTAR_OP_RESIZE, mr_tcam_region->rtar_key_type, new_count); return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rtar), rtar_pl); } static void mlxsw_sp_mr_tcam_region_parman_move(void *priv, unsigned long from_index, unsigned long to_index, unsigned long count) { struct mlxsw_sp_mr_tcam_region *mr_tcam_region = priv; struct mlxsw_sp *mlxsw_sp = mr_tcam_region->mlxsw_sp; char rrcr_pl[MLXSW_REG_RRCR_LEN]; mlxsw_reg_rrcr_pack(rrcr_pl, MLXSW_REG_RRCR_OP_MOVE, from_index, count, mr_tcam_region->rtar_key_type, to_index); mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(rrcr), rrcr_pl); } static const struct parman_ops mlxsw_sp_mr_tcam_region_parman_ops = { .base_count = MLXSW_SP_MR_TCAM_REGION_BASE_COUNT, .resize_step = MLXSW_SP_MR_TCAM_REGION_RESIZE_STEP, .resize = mlxsw_sp_mr_tcam_region_parman_resize, .move = mlxsw_sp_mr_tcam_region_parman_move, .algo = PARMAN_ALGO_TYPE_LSORT, }; static int mlxsw_sp_mr_tcam_region_init(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_mr_tcam_region *mr_tcam_region, enum mlxsw_reg_rtar_key_type rtar_key_type) { struct parman_prio *parman_prios; struct parman *parman; int err; int i; mr_tcam_region->rtar_key_type = rtar_key_type; mr_tcam_region->mlxsw_sp = mlxsw_sp; err = mlxsw_sp_mr_tcam_region_alloc(mr_tcam_region); if (err) return err; parman = parman_create(&mlxsw_sp_mr_tcam_region_parman_ops, mr_tcam_region); if (!parman) { err = -ENOMEM; goto err_parman_create; } mr_tcam_region->parman = parman; parman_prios = kmalloc_array(MLXSW_SP_MR_ROUTE_PRIO_MAX + 1, sizeof(*parman_prios), GFP_KERNEL); if (!parman_prios) { err = -ENOMEM; goto err_parman_prios_alloc; } mr_tcam_region->parman_prios = parman_prios; for (i = 0; i < MLXSW_SP_MR_ROUTE_PRIO_MAX + 1; i++) parman_prio_init(mr_tcam_region->parman, &mr_tcam_region->parman_prios[i], i); return 0; err_parman_prios_alloc: parman_destroy(parman); err_parman_create: mlxsw_sp_mr_tcam_region_free(mr_tcam_region); return err; } static void mlxsw_sp_mr_tcam_region_fini(struct mlxsw_sp_mr_tcam_region *mr_tcam_region) { int i; for (i = 0; i < MLXSW_SP_MR_ROUTE_PRIO_MAX + 1; i++) parman_prio_fini(&mr_tcam_region->parman_prios[i]); kfree(mr_tcam_region->parman_prios); parman_destroy(mr_tcam_region->parman); mlxsw_sp_mr_tcam_region_free(mr_tcam_region); } static int mlxsw_sp_mr_tcam_init(struct mlxsw_sp *mlxsw_sp, void *priv) { struct mlxsw_sp_mr_tcam *mr_tcam = priv; if (!MLXSW_CORE_RES_VALID(mlxsw_sp->core, MC_ERIF_LIST_ENTRIES) || !MLXSW_CORE_RES_VALID(mlxsw_sp->core, ACL_MAX_TCAM_RULES)) return -EIO; return mlxsw_sp_mr_tcam_region_init(mlxsw_sp, &mr_tcam->ipv4_tcam_region, MLXSW_REG_RTAR_KEY_TYPE_IPV4_MULTICAST); } static void mlxsw_sp_mr_tcam_fini(void *priv) { struct mlxsw_sp_mr_tcam *mr_tcam = priv; mlxsw_sp_mr_tcam_region_fini(&mr_tcam->ipv4_tcam_region); } const struct mlxsw_sp_mr_ops mlxsw_sp_mr_tcam_ops = { .priv_size = sizeof(struct mlxsw_sp_mr_tcam), .route_priv_size = sizeof(struct mlxsw_sp_mr_tcam_route), .init = mlxsw_sp_mr_tcam_init, .route_create = mlxsw_sp_mr_tcam_route_create, .route_update = mlxsw_sp_mr_tcam_route_update, .route_stats = mlxsw_sp_mr_tcam_route_stats, .route_action_update = mlxsw_sp_mr_tcam_route_action_update, .route_min_mtu_update = mlxsw_sp_mr_tcam_route_min_mtu_update, .route_irif_update = mlxsw_sp_mr_tcam_route_irif_update, .route_erif_add = mlxsw_sp_mr_tcam_route_erif_add, .route_erif_del = mlxsw_sp_mr_tcam_route_erif_del, .route_destroy = mlxsw_sp_mr_tcam_route_destroy, .fini = mlxsw_sp_mr_tcam_fini, };