Merge branch 'mlx4-misc-fixes'
Tariq Toukan says:
====================
mlx4 misc fixes
This patchset contains several bug fixes from the team to the
mlx4 Eth and Core drivers.
Series generated against net commit:
60133867f1
'net: wan: slic_ds26522: fix spelling mistake: "configurated" -> "configured"'
====================
Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
commit
e400b7977e
|
@ -1682,9 +1682,19 @@ static int __mlx4_ib_create_flow(struct ib_qp *qp, struct ib_flow_attr *flow_att
|
||||||
size += ret;
|
size += ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (mlx4_is_master(mdev->dev) && flow_type == MLX4_FS_REGULAR &&
|
||||||
|
flow_attr->num_of_specs == 1) {
|
||||||
|
struct _rule_hw *rule_header = (struct _rule_hw *)(ctrl + 1);
|
||||||
|
enum ib_flow_spec_type header_spec =
|
||||||
|
((union ib_flow_spec *)(flow_attr + 1))->type;
|
||||||
|
|
||||||
|
if (header_spec == IB_FLOW_SPEC_ETH)
|
||||||
|
mlx4_handle_eth_header_mcast_prio(ctrl, rule_header);
|
||||||
|
}
|
||||||
|
|
||||||
ret = mlx4_cmd_imm(mdev->dev, mailbox->dma, reg_id, size >> 2, 0,
|
ret = mlx4_cmd_imm(mdev->dev, mailbox->dma, reg_id, size >> 2, 0,
|
||||||
MLX4_QP_FLOW_STEERING_ATTACH, MLX4_CMD_TIME_CLASS_A,
|
MLX4_QP_FLOW_STEERING_ATTACH, MLX4_CMD_TIME_CLASS_A,
|
||||||
MLX4_CMD_WRAPPED);
|
MLX4_CMD_NATIVE);
|
||||||
if (ret == -ENOMEM)
|
if (ret == -ENOMEM)
|
||||||
pr_err("mcg table is full. Fail to register network rule.\n");
|
pr_err("mcg table is full. Fail to register network rule.\n");
|
||||||
else if (ret == -ENXIO)
|
else if (ret == -ENXIO)
|
||||||
|
@ -1701,7 +1711,7 @@ static int __mlx4_ib_destroy_flow(struct mlx4_dev *dev, u64 reg_id)
|
||||||
int err;
|
int err;
|
||||||
err = mlx4_cmd(dev, reg_id, 0, 0,
|
err = mlx4_cmd(dev, reg_id, 0, 0,
|
||||||
MLX4_QP_FLOW_STEERING_DETACH, MLX4_CMD_TIME_CLASS_A,
|
MLX4_QP_FLOW_STEERING_DETACH, MLX4_CMD_TIME_CLASS_A,
|
||||||
MLX4_CMD_WRAPPED);
|
MLX4_CMD_NATIVE);
|
||||||
if (err)
|
if (err)
|
||||||
pr_err("Fail to detach network rule. registration id = 0x%llx\n",
|
pr_err("Fail to detach network rule. registration id = 0x%llx\n",
|
||||||
reg_id);
|
reg_id);
|
||||||
|
|
|
@ -245,13 +245,9 @@ static u32 freq_to_shift(u16 freq)
|
||||||
{
|
{
|
||||||
u32 freq_khz = freq * 1000;
|
u32 freq_khz = freq * 1000;
|
||||||
u64 max_val_cycles = freq_khz * 1000 * MLX4_EN_WRAP_AROUND_SEC;
|
u64 max_val_cycles = freq_khz * 1000 * MLX4_EN_WRAP_AROUND_SEC;
|
||||||
u64 tmp_rounded =
|
u64 max_val_cycles_rounded = 1ULL << fls64(max_val_cycles - 1);
|
||||||
roundup_pow_of_two(max_val_cycles) > max_val_cycles ?
|
|
||||||
roundup_pow_of_two(max_val_cycles) - 1 : UINT_MAX;
|
|
||||||
u64 max_val_cycles_rounded = is_power_of_2(max_val_cycles + 1) ?
|
|
||||||
max_val_cycles : tmp_rounded;
|
|
||||||
/* calculate max possible multiplier in order to fit in 64bit */
|
/* calculate max possible multiplier in order to fit in 64bit */
|
||||||
u64 max_mul = div_u64(0xffffffffffffffffULL, max_val_cycles_rounded);
|
u64 max_mul = div64_u64(ULLONG_MAX, max_val_cycles_rounded);
|
||||||
|
|
||||||
/* This comes from the reverse of clocksource_khz2mult */
|
/* This comes from the reverse of clocksource_khz2mult */
|
||||||
return ilog2(div_u64(max_mul * freq_khz, 1000000));
|
return ilog2(div_u64(max_mul * freq_khz, 1000000));
|
||||||
|
|
|
@ -445,8 +445,14 @@ int mlx4_en_activate_rx_rings(struct mlx4_en_priv *priv)
|
||||||
ring->cqn = priv->rx_cq[ring_ind]->mcq.cqn;
|
ring->cqn = priv->rx_cq[ring_ind]->mcq.cqn;
|
||||||
|
|
||||||
ring->stride = stride;
|
ring->stride = stride;
|
||||||
if (ring->stride <= TXBB_SIZE)
|
if (ring->stride <= TXBB_SIZE) {
|
||||||
|
/* Stamp first unused send wqe */
|
||||||
|
__be32 *ptr = (__be32 *)ring->buf;
|
||||||
|
__be32 stamp = cpu_to_be32(1 << STAMP_SHIFT);
|
||||||
|
*ptr = stamp;
|
||||||
|
/* Move pointer to start of rx section */
|
||||||
ring->buf += TXBB_SIZE;
|
ring->buf += TXBB_SIZE;
|
||||||
|
}
|
||||||
|
|
||||||
ring->log_stride = ffs(ring->stride) - 1;
|
ring->log_stride = ffs(ring->stride) - 1;
|
||||||
ring->buf_size = ring->size * ring->stride;
|
ring->buf_size = ring->size * ring->stride;
|
||||||
|
|
|
@ -118,8 +118,13 @@ static int mlx4_alloc_icm_coherent(struct device *dev, struct scatterlist *mem,
|
||||||
if (!buf)
|
if (!buf)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
|
if (offset_in_page(buf)) {
|
||||||
|
dma_free_coherent(dev, PAGE_SIZE << order,
|
||||||
|
buf, sg_dma_address(mem));
|
||||||
|
return -ENOMEM;
|
||||||
|
}
|
||||||
|
|
||||||
sg_set_buf(mem, buf, PAGE_SIZE << order);
|
sg_set_buf(mem, buf, PAGE_SIZE << order);
|
||||||
BUG_ON(mem->offset);
|
|
||||||
sg_dma_len(mem) = PAGE_SIZE << order;
|
sg_dma_len(mem) = PAGE_SIZE << order;
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
|
@ -42,6 +42,7 @@
|
||||||
#include <linux/io-mapping.h>
|
#include <linux/io-mapping.h>
|
||||||
#include <linux/delay.h>
|
#include <linux/delay.h>
|
||||||
#include <linux/kmod.h>
|
#include <linux/kmod.h>
|
||||||
|
#include <linux/etherdevice.h>
|
||||||
#include <net/devlink.h>
|
#include <net/devlink.h>
|
||||||
|
|
||||||
#include <linux/mlx4/device.h>
|
#include <linux/mlx4/device.h>
|
||||||
|
@ -782,6 +783,23 @@ int mlx4_is_slave_active(struct mlx4_dev *dev, int slave)
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(mlx4_is_slave_active);
|
EXPORT_SYMBOL(mlx4_is_slave_active);
|
||||||
|
|
||||||
|
void mlx4_handle_eth_header_mcast_prio(struct mlx4_net_trans_rule_hw_ctrl *ctrl,
|
||||||
|
struct _rule_hw *eth_header)
|
||||||
|
{
|
||||||
|
if (is_multicast_ether_addr(eth_header->eth.dst_mac) ||
|
||||||
|
is_broadcast_ether_addr(eth_header->eth.dst_mac)) {
|
||||||
|
struct mlx4_net_trans_rule_hw_eth *eth =
|
||||||
|
(struct mlx4_net_trans_rule_hw_eth *)eth_header;
|
||||||
|
struct _rule_hw *next_rule = (struct _rule_hw *)(eth + 1);
|
||||||
|
bool last_rule = next_rule->size == 0 && next_rule->id == 0 &&
|
||||||
|
next_rule->rsvd == 0;
|
||||||
|
|
||||||
|
if (last_rule)
|
||||||
|
ctrl->prio = cpu_to_be16(MLX4_DOMAIN_NIC);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(mlx4_handle_eth_header_mcast_prio);
|
||||||
|
|
||||||
static void slave_adjust_steering_mode(struct mlx4_dev *dev,
|
static void slave_adjust_steering_mode(struct mlx4_dev *dev,
|
||||||
struct mlx4_dev_cap *dev_cap,
|
struct mlx4_dev_cap *dev_cap,
|
||||||
struct mlx4_init_hca_param *hca_param)
|
struct mlx4_init_hca_param *hca_param)
|
||||||
|
|
|
@ -4164,22 +4164,6 @@ static int validate_eth_header_mac(int slave, struct _rule_hw *eth_header,
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void handle_eth_header_mcast_prio(struct mlx4_net_trans_rule_hw_ctrl *ctrl,
|
|
||||||
struct _rule_hw *eth_header)
|
|
||||||
{
|
|
||||||
if (is_multicast_ether_addr(eth_header->eth.dst_mac) ||
|
|
||||||
is_broadcast_ether_addr(eth_header->eth.dst_mac)) {
|
|
||||||
struct mlx4_net_trans_rule_hw_eth *eth =
|
|
||||||
(struct mlx4_net_trans_rule_hw_eth *)eth_header;
|
|
||||||
struct _rule_hw *next_rule = (struct _rule_hw *)(eth + 1);
|
|
||||||
bool last_rule = next_rule->size == 0 && next_rule->id == 0 &&
|
|
||||||
next_rule->rsvd == 0;
|
|
||||||
|
|
||||||
if (last_rule)
|
|
||||||
ctrl->prio = cpu_to_be16(MLX4_DOMAIN_NIC);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* In case of missing eth header, append eth header with a MAC address
|
* In case of missing eth header, append eth header with a MAC address
|
||||||
* assigned to the VF.
|
* assigned to the VF.
|
||||||
|
@ -4363,10 +4347,7 @@ int mlx4_QP_FLOW_STEERING_ATTACH_wrapper(struct mlx4_dev *dev, int slave,
|
||||||
header_id = map_hw_to_sw_id(be16_to_cpu(rule_header->id));
|
header_id = map_hw_to_sw_id(be16_to_cpu(rule_header->id));
|
||||||
|
|
||||||
if (header_id == MLX4_NET_TRANS_RULE_ID_ETH)
|
if (header_id == MLX4_NET_TRANS_RULE_ID_ETH)
|
||||||
handle_eth_header_mcast_prio(ctrl, rule_header);
|
mlx4_handle_eth_header_mcast_prio(ctrl, rule_header);
|
||||||
|
|
||||||
if (slave == dev->caps.function)
|
|
||||||
goto execute;
|
|
||||||
|
|
||||||
switch (header_id) {
|
switch (header_id) {
|
||||||
case MLX4_NET_TRANS_RULE_ID_ETH:
|
case MLX4_NET_TRANS_RULE_ID_ETH:
|
||||||
|
@ -4394,7 +4375,6 @@ int mlx4_QP_FLOW_STEERING_ATTACH_wrapper(struct mlx4_dev *dev, int slave,
|
||||||
goto err_put_qp;
|
goto err_put_qp;
|
||||||
}
|
}
|
||||||
|
|
||||||
execute:
|
|
||||||
err = mlx4_cmd_imm(dev, inbox->dma, &vhcr->out_param,
|
err = mlx4_cmd_imm(dev, inbox->dma, &vhcr->out_param,
|
||||||
vhcr->in_modifier, 0,
|
vhcr->in_modifier, 0,
|
||||||
MLX4_QP_FLOW_STEERING_ATTACH, MLX4_CMD_TIME_CLASS_A,
|
MLX4_QP_FLOW_STEERING_ATTACH, MLX4_CMD_TIME_CLASS_A,
|
||||||
|
@ -4473,6 +4453,7 @@ int mlx4_QP_FLOW_STEERING_DETACH_wrapper(struct mlx4_dev *dev, int slave,
|
||||||
struct res_qp *rqp;
|
struct res_qp *rqp;
|
||||||
struct res_fs_rule *rrule;
|
struct res_fs_rule *rrule;
|
||||||
u64 mirr_reg_id;
|
u64 mirr_reg_id;
|
||||||
|
int qpn;
|
||||||
|
|
||||||
if (dev->caps.steering_mode !=
|
if (dev->caps.steering_mode !=
|
||||||
MLX4_STEERING_MODE_DEVICE_MANAGED)
|
MLX4_STEERING_MODE_DEVICE_MANAGED)
|
||||||
|
@ -4489,10 +4470,11 @@ int mlx4_QP_FLOW_STEERING_DETACH_wrapper(struct mlx4_dev *dev, int slave,
|
||||||
}
|
}
|
||||||
mirr_reg_id = rrule->mirr_rule_id;
|
mirr_reg_id = rrule->mirr_rule_id;
|
||||||
kfree(rrule->mirr_mbox);
|
kfree(rrule->mirr_mbox);
|
||||||
|
qpn = rrule->qpn;
|
||||||
|
|
||||||
/* Release the rule form busy state before removal */
|
/* Release the rule form busy state before removal */
|
||||||
put_res(dev, slave, vhcr->in_param, RES_FS_RULE);
|
put_res(dev, slave, vhcr->in_param, RES_FS_RULE);
|
||||||
err = get_res(dev, slave, rrule->qpn, RES_QP, &rqp);
|
err = get_res(dev, slave, qpn, RES_QP, &rqp);
|
||||||
if (err)
|
if (err)
|
||||||
return err;
|
return err;
|
||||||
|
|
||||||
|
@ -4517,7 +4499,7 @@ int mlx4_QP_FLOW_STEERING_DETACH_wrapper(struct mlx4_dev *dev, int slave,
|
||||||
if (!err)
|
if (!err)
|
||||||
atomic_dec(&rqp->ref_count);
|
atomic_dec(&rqp->ref_count);
|
||||||
out:
|
out:
|
||||||
put_res(dev, slave, rrule->qpn, RES_QP);
|
put_res(dev, slave, qpn, RES_QP);
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1384,6 +1384,8 @@ int set_phv_bit(struct mlx4_dev *dev, u8 port, int new_val);
|
||||||
int get_phv_bit(struct mlx4_dev *dev, u8 port, int *phv);
|
int get_phv_bit(struct mlx4_dev *dev, u8 port, int *phv);
|
||||||
int mlx4_get_is_vlan_offload_disabled(struct mlx4_dev *dev, u8 port,
|
int mlx4_get_is_vlan_offload_disabled(struct mlx4_dev *dev, u8 port,
|
||||||
bool *vlan_offload_disabled);
|
bool *vlan_offload_disabled);
|
||||||
|
void mlx4_handle_eth_header_mcast_prio(struct mlx4_net_trans_rule_hw_ctrl *ctrl,
|
||||||
|
struct _rule_hw *eth_header);
|
||||||
int mlx4_find_cached_mac(struct mlx4_dev *dev, u8 port, u64 mac, int *idx);
|
int mlx4_find_cached_mac(struct mlx4_dev *dev, u8 port, u64 mac, int *idx);
|
||||||
int mlx4_find_cached_vlan(struct mlx4_dev *dev, u8 port, u16 vid, int *idx);
|
int mlx4_find_cached_vlan(struct mlx4_dev *dev, u8 port, u16 vid, int *idx);
|
||||||
int mlx4_register_vlan(struct mlx4_dev *dev, u8 port, u16 vlan, int *index);
|
int mlx4_register_vlan(struct mlx4_dev *dev, u8 port, u16 vlan, int *index);
|
||||||
|
|
Loading…
Reference in New Issue