net/mlx4: Move the tunnel steering helper function to mlx4_core
Move the function which we use to set VXLAN DMFS (flow-steering) rules from mlx4_en to mlx4_core. This refactoring will allow the mlx4_ib driver to call the helper for the use case of user-space RAW Ethernet QPs, such that they can serve VXLAN traffic too. Signed-off-by: Or Gerlitz <ogerlitz@mellanox.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
362b37be01
commit
b95089d00c
|
@ -474,39 +474,12 @@ static int mlx4_en_tunnel_steer_add(struct mlx4_en_priv *priv, unsigned char *ad
|
||||||
int qpn, u64 *reg_id)
|
int qpn, u64 *reg_id)
|
||||||
{
|
{
|
||||||
int err;
|
int err;
|
||||||
struct mlx4_spec_list spec_eth_outer = { {NULL} };
|
|
||||||
struct mlx4_spec_list spec_vxlan = { {NULL} };
|
|
||||||
struct mlx4_spec_list spec_eth_inner = { {NULL} };
|
|
||||||
|
|
||||||
struct mlx4_net_trans_rule rule = {
|
|
||||||
.queue_mode = MLX4_NET_TRANS_Q_FIFO,
|
|
||||||
.exclusive = 0,
|
|
||||||
.allow_loopback = 1,
|
|
||||||
.promisc_mode = MLX4_FS_REGULAR,
|
|
||||||
.priority = MLX4_DOMAIN_NIC,
|
|
||||||
};
|
|
||||||
|
|
||||||
__be64 mac_mask = cpu_to_be64(MLX4_MAC_MASK << 16);
|
|
||||||
|
|
||||||
if (priv->mdev->dev->caps.tunnel_offload_mode != MLX4_TUNNEL_OFFLOAD_MODE_VXLAN)
|
if (priv->mdev->dev->caps.tunnel_offload_mode != MLX4_TUNNEL_OFFLOAD_MODE_VXLAN)
|
||||||
return 0; /* do nothing */
|
return 0; /* do nothing */
|
||||||
|
|
||||||
rule.port = priv->port;
|
err = mlx4_tunnel_steer_add(priv->mdev->dev, addr, priv->port, qpn,
|
||||||
rule.qpn = qpn;
|
MLX4_DOMAIN_NIC, reg_id);
|
||||||
INIT_LIST_HEAD(&rule.list);
|
|
||||||
|
|
||||||
spec_eth_outer.id = MLX4_NET_TRANS_RULE_ID_ETH;
|
|
||||||
memcpy(spec_eth_outer.eth.dst_mac, addr, ETH_ALEN);
|
|
||||||
memcpy(spec_eth_outer.eth.dst_mac_msk, &mac_mask, ETH_ALEN);
|
|
||||||
|
|
||||||
spec_vxlan.id = MLX4_NET_TRANS_RULE_ID_VXLAN; /* any vxlan header */
|
|
||||||
spec_eth_inner.id = MLX4_NET_TRANS_RULE_ID_ETH; /* any inner eth header */
|
|
||||||
|
|
||||||
list_add_tail(&spec_eth_outer.list, &rule.list);
|
|
||||||
list_add_tail(&spec_vxlan.list, &rule.list);
|
|
||||||
list_add_tail(&spec_eth_inner.list, &rule.list);
|
|
||||||
|
|
||||||
err = mlx4_flow_attach(priv->mdev->dev, &rule, reg_id);
|
|
||||||
if (err) {
|
if (err) {
|
||||||
en_err(priv, "failed to add vxlan steering rule, err %d\n", err);
|
en_err(priv, "failed to add vxlan steering rule, err %d\n", err);
|
||||||
return err;
|
return err;
|
||||||
|
|
|
@ -1020,6 +1020,44 @@ int mlx4_flow_detach(struct mlx4_dev *dev, u64 reg_id)
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(mlx4_flow_detach);
|
EXPORT_SYMBOL_GPL(mlx4_flow_detach);
|
||||||
|
|
||||||
|
int mlx4_tunnel_steer_add(struct mlx4_dev *dev, unsigned char *addr,
|
||||||
|
int port, int qpn, u16 prio, u64 *reg_id)
|
||||||
|
{
|
||||||
|
int err;
|
||||||
|
struct mlx4_spec_list spec_eth_outer = { {NULL} };
|
||||||
|
struct mlx4_spec_list spec_vxlan = { {NULL} };
|
||||||
|
struct mlx4_spec_list spec_eth_inner = { {NULL} };
|
||||||
|
|
||||||
|
struct mlx4_net_trans_rule rule = {
|
||||||
|
.queue_mode = MLX4_NET_TRANS_Q_FIFO,
|
||||||
|
.exclusive = 0,
|
||||||
|
.allow_loopback = 1,
|
||||||
|
.promisc_mode = MLX4_FS_REGULAR,
|
||||||
|
};
|
||||||
|
|
||||||
|
__be64 mac_mask = cpu_to_be64(MLX4_MAC_MASK << 16);
|
||||||
|
|
||||||
|
rule.port = port;
|
||||||
|
rule.qpn = qpn;
|
||||||
|
rule.priority = prio;
|
||||||
|
INIT_LIST_HEAD(&rule.list);
|
||||||
|
|
||||||
|
spec_eth_outer.id = MLX4_NET_TRANS_RULE_ID_ETH;
|
||||||
|
memcpy(spec_eth_outer.eth.dst_mac, addr, ETH_ALEN);
|
||||||
|
memcpy(spec_eth_outer.eth.dst_mac_msk, &mac_mask, ETH_ALEN);
|
||||||
|
|
||||||
|
spec_vxlan.id = MLX4_NET_TRANS_RULE_ID_VXLAN; /* any vxlan header */
|
||||||
|
spec_eth_inner.id = MLX4_NET_TRANS_RULE_ID_ETH; /* any inner eth header */
|
||||||
|
|
||||||
|
list_add_tail(&spec_eth_outer.list, &rule.list);
|
||||||
|
list_add_tail(&spec_vxlan.list, &rule.list);
|
||||||
|
list_add_tail(&spec_eth_inner.list, &rule.list);
|
||||||
|
|
||||||
|
err = mlx4_flow_attach(dev, &rule, reg_id);
|
||||||
|
return err;
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(mlx4_tunnel_steer_add);
|
||||||
|
|
||||||
int mlx4_FLOW_STEERING_IB_UC_QP_RANGE(struct mlx4_dev *dev, u32 min_range_qpn,
|
int mlx4_FLOW_STEERING_IB_UC_QP_RANGE(struct mlx4_dev *dev, u32 min_range_qpn,
|
||||||
u32 max_range_qpn)
|
u32 max_range_qpn)
|
||||||
{
|
{
|
||||||
|
|
|
@ -1196,6 +1196,9 @@ int mlx4_map_sw_to_hw_steering_id(struct mlx4_dev *dev,
|
||||||
enum mlx4_net_trans_rule_id id);
|
enum mlx4_net_trans_rule_id id);
|
||||||
int mlx4_hw_rule_sz(struct mlx4_dev *dev, enum mlx4_net_trans_rule_id id);
|
int mlx4_hw_rule_sz(struct mlx4_dev *dev, enum mlx4_net_trans_rule_id id);
|
||||||
|
|
||||||
|
int mlx4_tunnel_steer_add(struct mlx4_dev *dev, unsigned char *addr,
|
||||||
|
int port, int qpn, u16 prio, u64 *reg_id);
|
||||||
|
|
||||||
void mlx4_sync_pkey_table(struct mlx4_dev *dev, int slave, int port,
|
void mlx4_sync_pkey_table(struct mlx4_dev *dev, int slave, int port,
|
||||||
int i, int val);
|
int i, int val);
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue