net/mlx5: Reuse mlx5_esw_for_each_vf_vport macro in two files
Currently mlx5_esw_for_each_vf_vport iterates over mlx5_vport entries in eswitch.c Same macro in eswitch_offloads.c iterates over vport number in eswitch_offloads.c Instead of duplicate macro names, to avoid confusion and to reuse the same macro in both files, move it to eswitch.h. To iterate over vport numbers where there is no need to iterate over mlx5_vport, but only a vport number is needed, rename those macros in eswitch_offloads.c to mlx5_esw_for_each_vf_num_vport*. While at it, keep all vport and vport rep iterators together. Signed-off-by: Parav Pandit <parav@mellanox.com> Signed-off-by: Saeed Mahameed <saeedm@mellanox.com>
This commit is contained in:
parent
c9bbfb378b
commit
786ef904b4
|
@ -72,19 +72,6 @@ static void esw_cleanup_vepa_rules(struct mlx5_eswitch *esw);
|
|||
MC_ADDR_CHANGE | \
|
||||
PROMISC_CHANGE)
|
||||
|
||||
/* The vport getter/iterator are only valid after esw->total_vports
|
||||
* and vport->vport are initialized in mlx5_eswitch_init.
|
||||
*/
|
||||
#define mlx5_esw_for_all_vports(esw, i, vport) \
|
||||
for ((i) = MLX5_VPORT_PF; \
|
||||
(vport) = &(esw)->vports[i], \
|
||||
(i) < (esw)->total_vports; (i)++)
|
||||
|
||||
#define mlx5_esw_for_each_vf_vport(esw, i, vport, nvfs) \
|
||||
for ((i) = MLX5_VPORT_FIRST_VF; \
|
||||
(vport) = &(esw)->vports[i], \
|
||||
(i) <= (nvfs); (i)++)
|
||||
|
||||
static struct mlx5_vport *mlx5_eswitch_get_vport(struct mlx5_eswitch *esw,
|
||||
u16 vport_num)
|
||||
{
|
||||
|
|
|
@ -446,6 +446,48 @@ static inline int mlx5_eswitch_index_to_vport_num(struct mlx5_eswitch *esw,
|
|||
/* TODO: This mlx5e_tc function shouldn't be called by eswitch */
|
||||
void mlx5e_tc_clean_fdb_peer_flows(struct mlx5_eswitch *esw);
|
||||
|
||||
/* The vport getter/iterator are only valid after esw->total_vports
|
||||
* and vport->vport are initialized in mlx5_eswitch_init.
|
||||
*/
|
||||
#define mlx5_esw_for_all_vports(esw, i, vport) \
|
||||
for ((i) = MLX5_VPORT_PF; \
|
||||
(vport) = &(esw)->vports[i], \
|
||||
(i) < (esw)->total_vports; (i)++)
|
||||
|
||||
#define mlx5_esw_for_each_vf_vport(esw, i, vport, nvfs) \
|
||||
for ((i) = MLX5_VPORT_FIRST_VF; \
|
||||
(vport) = &(esw)->vports[(i)], \
|
||||
(i) <= (nvfs); (i)++)
|
||||
|
||||
#define mlx5_esw_for_each_vf_vport_reverse(esw, i, vport, nvfs) \
|
||||
for ((i) = (nvfs); \
|
||||
(vport) = &(esw)->vports[(i)], \
|
||||
(i) >= MLX5_VPORT_FIRST_VF; (i)--)
|
||||
|
||||
/* The rep getter/iterator are only valid after esw->total_vports
|
||||
* and vport->vport are initialized in mlx5_eswitch_init.
|
||||
*/
|
||||
#define mlx5_esw_for_all_reps(esw, i, rep) \
|
||||
for ((i) = MLX5_VPORT_PF; \
|
||||
(rep) = &(esw)->offloads.vport_reps[i], \
|
||||
(i) < (esw)->total_vports; (i)++)
|
||||
|
||||
#define mlx5_esw_for_each_vf_rep(esw, i, rep, nvfs) \
|
||||
for ((i) = MLX5_VPORT_FIRST_VF; \
|
||||
(rep) = &(esw)->offloads.vport_reps[i], \
|
||||
(i) <= (nvfs); (i)++)
|
||||
|
||||
#define mlx5_esw_for_each_vf_rep_reverse(esw, i, rep, nvfs) \
|
||||
for ((i) = (nvfs); \
|
||||
(rep) = &(esw)->offloads.vport_reps[i], \
|
||||
(i) >= MLX5_VPORT_FIRST_VF; (i)--)
|
||||
|
||||
#define mlx5_esw_for_each_vf_vport_num(esw, vport, nvfs) \
|
||||
for ((vport) = MLX5_VPORT_FIRST_VF; (vport) <= (nvfs); (vport)++)
|
||||
|
||||
#define mlx5_esw_for_each_vf_vport_num_reverse(esw, vport, nvfs) \
|
||||
for ((vport) = (nvfs); (vport) >= MLX5_VPORT_FIRST_VF; (vport)--)
|
||||
|
||||
#else /* CONFIG_MLX5_ESWITCH */
|
||||
/* eswitch API stubs */
|
||||
static inline int mlx5_eswitch_init(struct mlx5_core_dev *dev) { return 0; }
|
||||
|
|
|
@ -54,32 +54,6 @@
|
|||
|
||||
#define UPLINK_REP_INDEX 0
|
||||
|
||||
/* The rep getter/iterator are only valid after esw->total_vports
|
||||
* and vport->vport are initialized in mlx5_eswitch_init.
|
||||
*/
|
||||
#define mlx5_esw_for_all_reps(esw, i, rep) \
|
||||
for ((i) = MLX5_VPORT_PF; \
|
||||
(rep) = &(esw)->offloads.vport_reps[i], \
|
||||
(i) < (esw)->total_vports; (i)++)
|
||||
|
||||
#define mlx5_esw_for_each_vf_rep(esw, i, rep, nvfs) \
|
||||
for ((i) = MLX5_VPORT_FIRST_VF; \
|
||||
(rep) = &(esw)->offloads.vport_reps[i], \
|
||||
(i) <= (nvfs); (i)++)
|
||||
|
||||
#define mlx5_esw_for_each_vf_rep_reverse(esw, i, rep, nvfs) \
|
||||
for ((i) = (nvfs); \
|
||||
(rep) = &(esw)->offloads.vport_reps[i], \
|
||||
(i) >= MLX5_VPORT_FIRST_VF; (i)--)
|
||||
|
||||
#define mlx5_esw_for_each_vf_vport(esw, vport, nvfs) \
|
||||
for ((vport) = MLX5_VPORT_FIRST_VF; \
|
||||
(vport) <= (nvfs); (vport)++)
|
||||
|
||||
#define mlx5_esw_for_each_vf_vport_reverse(esw, vport, nvfs) \
|
||||
for ((vport) = (nvfs); \
|
||||
(vport) >= MLX5_VPORT_FIRST_VF; (vport)--)
|
||||
|
||||
static struct mlx5_eswitch_rep *mlx5_eswitch_get_rep(struct mlx5_eswitch *esw,
|
||||
u16 vport_num)
|
||||
{
|
||||
|
@ -659,7 +633,7 @@ static int esw_add_fdb_peer_miss_rules(struct mlx5_eswitch *esw,
|
|||
flows[mlx5_eswitch_ecpf_idx(esw)] = flow;
|
||||
}
|
||||
|
||||
mlx5_esw_for_each_vf_vport(esw, i, mlx5_core_max_vfs(esw->dev)) {
|
||||
mlx5_esw_for_each_vf_vport_num(esw, i, mlx5_core_max_vfs(esw->dev)) {
|
||||
MLX5_SET(fte_match_set_misc, misc, source_port, i);
|
||||
flow = mlx5_add_flow_rules(esw->fdb_table.offloads.slow_fdb,
|
||||
spec, &flow_act, &dest, 1);
|
||||
|
@ -677,7 +651,7 @@ static int esw_add_fdb_peer_miss_rules(struct mlx5_eswitch *esw,
|
|||
|
||||
add_vf_flow_err:
|
||||
nvports = --i;
|
||||
mlx5_esw_for_each_vf_vport_reverse(esw, i, nvports)
|
||||
mlx5_esw_for_each_vf_vport_num_reverse(esw, i, nvports)
|
||||
mlx5_del_flow_rules(flows[i]);
|
||||
|
||||
if (mlx5_ecpf_vport_exists(esw->dev))
|
||||
|
@ -700,7 +674,8 @@ static void esw_del_fdb_peer_miss_rules(struct mlx5_eswitch *esw)
|
|||
|
||||
flows = esw->fdb_table.offloads.peer_miss_rules;
|
||||
|
||||
mlx5_esw_for_each_vf_vport_reverse(esw, i, mlx5_core_max_vfs(esw->dev))
|
||||
mlx5_esw_for_each_vf_vport_num_reverse(esw, i,
|
||||
mlx5_core_max_vfs(esw->dev))
|
||||
mlx5_del_flow_rules(flows[i]);
|
||||
|
||||
if (mlx5_ecpf_vport_exists(esw->dev))
|
||||
|
@ -1727,14 +1702,15 @@ out_no_mem:
|
|||
|
||||
static int esw_prio_tag_acls_config(struct mlx5_eswitch *esw, int nvports)
|
||||
{
|
||||
struct mlx5_vport *vport = NULL;
|
||||
int i, j;
|
||||
int err;
|
||||
|
||||
mlx5_esw_for_each_vf_vport(esw, i, nvports) {
|
||||
err = esw_vport_ingress_prio_tag_config(esw, &esw->vports[i]);
|
||||
mlx5_esw_for_each_vf_vport(esw, i, vport, nvports) {
|
||||
err = esw_vport_ingress_prio_tag_config(esw, vport);
|
||||
if (err)
|
||||
goto err_ingress;
|
||||
err = esw_vport_egress_prio_tag_config(esw, &esw->vports[i]);
|
||||
err = esw_vport_egress_prio_tag_config(esw, vport);
|
||||
if (err)
|
||||
goto err_egress;
|
||||
}
|
||||
|
@ -1742,11 +1718,11 @@ static int esw_prio_tag_acls_config(struct mlx5_eswitch *esw, int nvports)
|
|||
return 0;
|
||||
|
||||
err_egress:
|
||||
esw_vport_disable_ingress_acl(esw, &esw->vports[i]);
|
||||
esw_vport_disable_ingress_acl(esw, vport);
|
||||
err_ingress:
|
||||
mlx5_esw_for_each_vf_vport_reverse(esw, j, i - 1) {
|
||||
esw_vport_disable_egress_acl(esw, &esw->vports[j]);
|
||||
esw_vport_disable_ingress_acl(esw, &esw->vports[j]);
|
||||
mlx5_esw_for_each_vf_vport_reverse(esw, j, vport, i - 1) {
|
||||
esw_vport_disable_egress_acl(esw, vport);
|
||||
esw_vport_disable_ingress_acl(esw, vport);
|
||||
}
|
||||
|
||||
return err;
|
||||
|
@ -1754,11 +1730,12 @@ err_ingress:
|
|||
|
||||
static void esw_prio_tag_acls_cleanup(struct mlx5_eswitch *esw)
|
||||
{
|
||||
struct mlx5_vport *vport;
|
||||
int i;
|
||||
|
||||
mlx5_esw_for_each_vf_vport(esw, i, esw->nvports) {
|
||||
esw_vport_disable_egress_acl(esw, &esw->vports[i]);
|
||||
esw_vport_disable_ingress_acl(esw, &esw->vports[i]);
|
||||
mlx5_esw_for_each_vf_vport(esw, i, vport, esw->nvports) {
|
||||
esw_vport_disable_egress_acl(esw, vport);
|
||||
esw_vport_disable_ingress_acl(esw, vport);
|
||||
}
|
||||
}
|
||||
|
||||
|
|
Loading…
Reference in New Issue