Merge branch 'mlx5-next'
Or Gerlitz says: ==================== Mellanox mlx5 driver update Bunch of changes from the team, while warming engines for the upcoming SRIOV support. ==================== Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
commit
366f02d873
|
@ -1136,6 +1136,7 @@ void mlx5_cmd_comp_handler(struct mlx5_core_dev *dev, unsigned long vector)
|
|||
mlx5_free_cmd_msg(dev, ent->out);
|
||||
free_msg(dev, ent->in);
|
||||
|
||||
err = err ? err : ent->status;
|
||||
free_cmd(ent);
|
||||
callback(err, context);
|
||||
} else {
|
||||
|
@ -1363,6 +1364,7 @@ int mlx5_cmd_init(struct mlx5_core_dev *dev)
|
|||
int err;
|
||||
int i;
|
||||
|
||||
memset(cmd, 0, sizeof(*cmd));
|
||||
cmd_if_rev = cmdif_rev(dev);
|
||||
if (cmd_if_rev != CMD_IF_REV) {
|
||||
dev_err(&dev->pdev->dev,
|
||||
|
|
|
@ -242,6 +242,7 @@ int mlx5_init_cq_table(struct mlx5_core_dev *dev)
|
|||
struct mlx5_cq_table *table = &dev->priv.cq_table;
|
||||
int err;
|
||||
|
||||
memset(table, 0, sizeof(*table));
|
||||
spin_lock_init(&table->lock);
|
||||
INIT_RADIX_TREE(&table->tree, GFP_ATOMIC);
|
||||
err = mlx5_cq_debugfs_init(dev);
|
||||
|
|
|
@ -1367,13 +1367,13 @@ int mlx5e_open_locked(struct net_device *netdev)
|
|||
|
||||
err = mlx5e_set_dev_port_mtu(netdev);
|
||||
if (err)
|
||||
return err;
|
||||
goto err_clear_state_opened_flag;
|
||||
|
||||
err = mlx5e_open_channels(priv);
|
||||
if (err) {
|
||||
netdev_err(netdev, "%s: mlx5e_open_channels failed, %d\n",
|
||||
__func__, err);
|
||||
return err;
|
||||
goto err_clear_state_opened_flag;
|
||||
}
|
||||
|
||||
mlx5e_update_carrier(priv);
|
||||
|
@ -1382,6 +1382,10 @@ int mlx5e_open_locked(struct net_device *netdev)
|
|||
schedule_delayed_work(&priv->update_stats_work, 0);
|
||||
|
||||
return 0;
|
||||
|
||||
err_clear_state_opened_flag:
|
||||
clear_bit(MLX5E_STATE_OPENED, &priv->state);
|
||||
return err;
|
||||
}
|
||||
|
||||
static int mlx5e_open(struct net_device *netdev)
|
||||
|
|
|
@ -346,6 +346,7 @@ int mlx5_create_map_eq(struct mlx5_core_dev *dev, struct mlx5_eq *eq, u8 vecidx,
|
|||
int inlen;
|
||||
|
||||
eq->nent = roundup_pow_of_two(nent + MLX5_NUM_SPARE_EQE);
|
||||
eq->cons_index = 0;
|
||||
err = mlx5_buf_alloc(dev, eq->nent * MLX5_EQE_SIZE, &eq->buf);
|
||||
if (err)
|
||||
return err;
|
||||
|
|
|
@ -46,12 +46,15 @@ enum {
|
|||
enum {
|
||||
MLX5_HEALTH_SYNDR_FW_ERR = 0x1,
|
||||
MLX5_HEALTH_SYNDR_IRISC_ERR = 0x7,
|
||||
MLX5_HEALTH_SYNDR_HW_UNRECOVERABLE_ERR = 0x8,
|
||||
MLX5_HEALTH_SYNDR_CRC_ERR = 0x9,
|
||||
MLX5_HEALTH_SYNDR_FETCH_PCI_ERR = 0xa,
|
||||
MLX5_HEALTH_SYNDR_HW_FTL_ERR = 0xb,
|
||||
MLX5_HEALTH_SYNDR_ASYNC_EQ_OVERRUN_ERR = 0xc,
|
||||
MLX5_HEALTH_SYNDR_EQ_ERR = 0xd,
|
||||
MLX5_HEALTH_SYNDR_EQ_INV = 0xe,
|
||||
MLX5_HEALTH_SYNDR_FFSER_ERR = 0xf,
|
||||
MLX5_HEALTH_SYNDR_HIGH_TEMP = 0x10
|
||||
};
|
||||
|
||||
static DEFINE_SPINLOCK(health_lock);
|
||||
|
@ -88,6 +91,8 @@ static const char *hsynd_str(u8 synd)
|
|||
return "firmware internal error";
|
||||
case MLX5_HEALTH_SYNDR_IRISC_ERR:
|
||||
return "irisc not responding";
|
||||
case MLX5_HEALTH_SYNDR_HW_UNRECOVERABLE_ERR:
|
||||
return "unrecoverable hardware error";
|
||||
case MLX5_HEALTH_SYNDR_CRC_ERR:
|
||||
return "firmware CRC error";
|
||||
case MLX5_HEALTH_SYNDR_FETCH_PCI_ERR:
|
||||
|
@ -98,8 +103,12 @@ static const char *hsynd_str(u8 synd)
|
|||
return "async EQ buffer overrun";
|
||||
case MLX5_HEALTH_SYNDR_EQ_ERR:
|
||||
return "EQ error";
|
||||
case MLX5_HEALTH_SYNDR_EQ_INV:
|
||||
return "Invalid EQ refrenced";
|
||||
case MLX5_HEALTH_SYNDR_FFSER_ERR:
|
||||
return "FFSER error";
|
||||
case MLX5_HEALTH_SYNDR_HIGH_TEMP:
|
||||
return "High temprature";
|
||||
default:
|
||||
return "unrecognized error";
|
||||
}
|
||||
|
@ -130,7 +139,7 @@ static void print_health_info(struct mlx5_core_dev *dev)
|
|||
pr_info("hw_id 0x%08x\n", read_be32(&h->hw_id));
|
||||
pr_info("irisc_index %d\n", readb(&h->irisc_index));
|
||||
pr_info("synd 0x%x: %s\n", readb(&h->synd), hsynd_str(readb(&h->synd)));
|
||||
pr_info("ext_sync 0x%04x\n", read_be16(&h->ext_sync));
|
||||
pr_info("ext_sync 0x%04x\n", read_be16(&h->ext_synd));
|
||||
}
|
||||
|
||||
static void poll_health(unsigned long data)
|
||||
|
|
|
@ -672,12 +672,126 @@ static void unmap_bf_area(struct mlx5_core_dev *dev)
|
|||
io_mapping_free(dev->priv.bf_mapping);
|
||||
}
|
||||
|
||||
static int mlx5_dev_init(struct mlx5_core_dev *dev, struct pci_dev *pdev)
|
||||
static void mlx5_add_device(struct mlx5_interface *intf, struct mlx5_priv *priv)
|
||||
{
|
||||
struct mlx5_device_context *dev_ctx;
|
||||
struct mlx5_core_dev *dev = container_of(priv, struct mlx5_core_dev, priv);
|
||||
|
||||
dev_ctx = kmalloc(sizeof(*dev_ctx), GFP_KERNEL);
|
||||
if (!dev_ctx)
|
||||
return;
|
||||
|
||||
dev_ctx->intf = intf;
|
||||
dev_ctx->context = intf->add(dev);
|
||||
|
||||
if (dev_ctx->context) {
|
||||
spin_lock_irq(&priv->ctx_lock);
|
||||
list_add_tail(&dev_ctx->list, &priv->ctx_list);
|
||||
spin_unlock_irq(&priv->ctx_lock);
|
||||
} else {
|
||||
kfree(dev_ctx);
|
||||
}
|
||||
}
|
||||
|
||||
static void mlx5_remove_device(struct mlx5_interface *intf, struct mlx5_priv *priv)
|
||||
{
|
||||
struct mlx5_device_context *dev_ctx;
|
||||
struct mlx5_core_dev *dev = container_of(priv, struct mlx5_core_dev, priv);
|
||||
|
||||
list_for_each_entry(dev_ctx, &priv->ctx_list, list)
|
||||
if (dev_ctx->intf == intf) {
|
||||
spin_lock_irq(&priv->ctx_lock);
|
||||
list_del(&dev_ctx->list);
|
||||
spin_unlock_irq(&priv->ctx_lock);
|
||||
|
||||
intf->remove(dev, dev_ctx->context);
|
||||
kfree(dev_ctx);
|
||||
return;
|
||||
}
|
||||
}
|
||||
|
||||
static int mlx5_register_device(struct mlx5_core_dev *dev)
|
||||
{
|
||||
struct mlx5_priv *priv = &dev->priv;
|
||||
int err;
|
||||
struct mlx5_interface *intf;
|
||||
|
||||
mutex_lock(&intf_mutex);
|
||||
list_add_tail(&priv->dev_list, &dev_list);
|
||||
list_for_each_entry(intf, &intf_list, list)
|
||||
mlx5_add_device(intf, priv);
|
||||
mutex_unlock(&intf_mutex);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void mlx5_unregister_device(struct mlx5_core_dev *dev)
|
||||
{
|
||||
struct mlx5_priv *priv = &dev->priv;
|
||||
struct mlx5_interface *intf;
|
||||
|
||||
mutex_lock(&intf_mutex);
|
||||
list_for_each_entry(intf, &intf_list, list)
|
||||
mlx5_remove_device(intf, priv);
|
||||
list_del(&priv->dev_list);
|
||||
mutex_unlock(&intf_mutex);
|
||||
}
|
||||
|
||||
int mlx5_register_interface(struct mlx5_interface *intf)
|
||||
{
|
||||
struct mlx5_priv *priv;
|
||||
|
||||
if (!intf->add || !intf->remove)
|
||||
return -EINVAL;
|
||||
|
||||
mutex_lock(&intf_mutex);
|
||||
list_add_tail(&intf->list, &intf_list);
|
||||
list_for_each_entry(priv, &dev_list, dev_list)
|
||||
mlx5_add_device(intf, priv);
|
||||
mutex_unlock(&intf_mutex);
|
||||
|
||||
return 0;
|
||||
}
|
||||
EXPORT_SYMBOL(mlx5_register_interface);
|
||||
|
||||
void mlx5_unregister_interface(struct mlx5_interface *intf)
|
||||
{
|
||||
struct mlx5_priv *priv;
|
||||
|
||||
mutex_lock(&intf_mutex);
|
||||
list_for_each_entry(priv, &dev_list, dev_list)
|
||||
mlx5_remove_device(intf, priv);
|
||||
list_del(&intf->list);
|
||||
mutex_unlock(&intf_mutex);
|
||||
}
|
||||
EXPORT_SYMBOL(mlx5_unregister_interface);
|
||||
|
||||
void *mlx5_get_protocol_dev(struct mlx5_core_dev *mdev, int protocol)
|
||||
{
|
||||
struct mlx5_priv *priv = &mdev->priv;
|
||||
struct mlx5_device_context *dev_ctx;
|
||||
unsigned long flags;
|
||||
void *result = NULL;
|
||||
|
||||
spin_lock_irqsave(&priv->ctx_lock, flags);
|
||||
|
||||
list_for_each_entry(dev_ctx, &mdev->priv.ctx_list, list)
|
||||
if ((dev_ctx->intf->protocol == protocol) &&
|
||||
dev_ctx->intf->get_dev) {
|
||||
result = dev_ctx->intf->get_dev(dev_ctx->context);
|
||||
break;
|
||||
}
|
||||
|
||||
spin_unlock_irqrestore(&priv->ctx_lock, flags);
|
||||
|
||||
return result;
|
||||
}
|
||||
EXPORT_SYMBOL(mlx5_get_protocol_dev);
|
||||
|
||||
static int mlx5_pci_init(struct mlx5_core_dev *dev, struct mlx5_priv *priv)
|
||||
{
|
||||
struct pci_dev *pdev = dev->pdev;
|
||||
int err = 0;
|
||||
|
||||
dev->pdev = pdev;
|
||||
pci_set_drvdata(dev->pdev, dev);
|
||||
strncpy(priv->name, dev_name(&pdev->dev), MLX5_MAX_NAME_LEN);
|
||||
priv->name[MLX5_MAX_NAME_LEN - 1] = 0;
|
||||
|
@ -721,13 +835,42 @@ static int mlx5_dev_init(struct mlx5_core_dev *dev, struct pci_dev *pdev)
|
|||
dev_err(&pdev->dev, "Failed mapping initialization segment, aborting\n");
|
||||
goto err_clr_master;
|
||||
}
|
||||
|
||||
return 0;
|
||||
|
||||
err_clr_master:
|
||||
pci_clear_master(dev->pdev);
|
||||
release_bar(dev->pdev);
|
||||
err_disable:
|
||||
pci_disable_device(dev->pdev);
|
||||
|
||||
err_dbg:
|
||||
debugfs_remove(priv->dbg_root);
|
||||
return err;
|
||||
}
|
||||
|
||||
static void mlx5_pci_close(struct mlx5_core_dev *dev, struct mlx5_priv *priv)
|
||||
{
|
||||
iounmap(dev->iseg);
|
||||
pci_clear_master(dev->pdev);
|
||||
release_bar(dev->pdev);
|
||||
pci_disable_device(dev->pdev);
|
||||
debugfs_remove(priv->dbg_root);
|
||||
}
|
||||
|
||||
#define MLX5_IB_MOD "mlx5_ib"
|
||||
static int mlx5_load_one(struct mlx5_core_dev *dev, struct mlx5_priv *priv)
|
||||
{
|
||||
struct pci_dev *pdev = dev->pdev;
|
||||
int err;
|
||||
|
||||
dev_info(&pdev->dev, "firmware version: %d.%d.%d\n", fw_rev_maj(dev),
|
||||
fw_rev_min(dev), fw_rev_sub(dev));
|
||||
|
||||
err = mlx5_cmd_init(dev);
|
||||
if (err) {
|
||||
dev_err(&pdev->dev, "Failed initializing command interface, aborting\n");
|
||||
goto err_unmap;
|
||||
return err;
|
||||
}
|
||||
|
||||
mlx5_pagealloc_init(dev);
|
||||
|
@ -842,8 +985,25 @@ static int mlx5_dev_init(struct mlx5_core_dev *dev, struct pci_dev *pdev)
|
|||
mlx5_init_srq_table(dev);
|
||||
mlx5_init_mr_table(dev);
|
||||
|
||||
err = mlx5_register_device(dev);
|
||||
if (err) {
|
||||
dev_err(&pdev->dev, "mlx5_register_device failed %d\n", err);
|
||||
goto err_reg_dev;
|
||||
}
|
||||
|
||||
err = request_module_nowait(MLX5_IB_MOD);
|
||||
if (err)
|
||||
pr_info("failed request module on %s\n", MLX5_IB_MOD);
|
||||
|
||||
return 0;
|
||||
|
||||
err_reg_dev:
|
||||
mlx5_cleanup_mr_table(dev);
|
||||
mlx5_cleanup_srq_table(dev);
|
||||
mlx5_cleanup_qp_table(dev);
|
||||
mlx5_cleanup_cq_table(dev);
|
||||
mlx5_irq_clear_affinity_hints(dev);
|
||||
|
||||
err_unmap_bf_area:
|
||||
unmap_bf_area(dev);
|
||||
|
||||
|
@ -881,25 +1041,14 @@ err_pagealloc_cleanup:
|
|||
mlx5_pagealloc_cleanup(dev);
|
||||
mlx5_cmd_cleanup(dev);
|
||||
|
||||
err_unmap:
|
||||
iounmap(dev->iseg);
|
||||
|
||||
err_clr_master:
|
||||
pci_clear_master(dev->pdev);
|
||||
release_bar(dev->pdev);
|
||||
|
||||
err_disable:
|
||||
pci_disable_device(dev->pdev);
|
||||
|
||||
err_dbg:
|
||||
debugfs_remove(priv->dbg_root);
|
||||
return err;
|
||||
}
|
||||
|
||||
static void mlx5_dev_cleanup(struct mlx5_core_dev *dev)
|
||||
static int mlx5_unload_one(struct mlx5_core_dev *dev, struct mlx5_priv *priv)
|
||||
{
|
||||
struct mlx5_priv *priv = &dev->priv;
|
||||
|
||||
mlx5_unregister_device(dev);
|
||||
mlx5_cleanup_mr_table(dev);
|
||||
mlx5_cleanup_srq_table(dev);
|
||||
mlx5_cleanup_qp_table(dev);
|
||||
mlx5_cleanup_cq_table(dev);
|
||||
|
@ -913,134 +1062,16 @@ static void mlx5_dev_cleanup(struct mlx5_core_dev *dev)
|
|||
mlx5_stop_health_poll(dev);
|
||||
if (mlx5_cmd_teardown_hca(dev)) {
|
||||
dev_err(&dev->pdev->dev, "tear_down_hca failed, skip cleanup\n");
|
||||
return;
|
||||
return 1;
|
||||
}
|
||||
mlx5_pagealloc_stop(dev);
|
||||
mlx5_reclaim_startup_pages(dev);
|
||||
mlx5_core_disable_hca(dev);
|
||||
mlx5_pagealloc_cleanup(dev);
|
||||
mlx5_cmd_cleanup(dev);
|
||||
iounmap(dev->iseg);
|
||||
pci_clear_master(dev->pdev);
|
||||
release_bar(dev->pdev);
|
||||
pci_disable_device(dev->pdev);
|
||||
debugfs_remove(priv->dbg_root);
|
||||
}
|
||||
|
||||
static void mlx5_add_device(struct mlx5_interface *intf, struct mlx5_priv *priv)
|
||||
{
|
||||
struct mlx5_device_context *dev_ctx;
|
||||
struct mlx5_core_dev *dev = container_of(priv, struct mlx5_core_dev, priv);
|
||||
|
||||
dev_ctx = kmalloc(sizeof(*dev_ctx), GFP_KERNEL);
|
||||
if (!dev_ctx) {
|
||||
pr_warn("mlx5_add_device: alloc context failed\n");
|
||||
return;
|
||||
}
|
||||
|
||||
dev_ctx->intf = intf;
|
||||
dev_ctx->context = intf->add(dev);
|
||||
|
||||
if (dev_ctx->context) {
|
||||
spin_lock_irq(&priv->ctx_lock);
|
||||
list_add_tail(&dev_ctx->list, &priv->ctx_list);
|
||||
spin_unlock_irq(&priv->ctx_lock);
|
||||
} else {
|
||||
kfree(dev_ctx);
|
||||
}
|
||||
}
|
||||
|
||||
static void mlx5_remove_device(struct mlx5_interface *intf, struct mlx5_priv *priv)
|
||||
{
|
||||
struct mlx5_device_context *dev_ctx;
|
||||
struct mlx5_core_dev *dev = container_of(priv, struct mlx5_core_dev, priv);
|
||||
|
||||
list_for_each_entry(dev_ctx, &priv->ctx_list, list)
|
||||
if (dev_ctx->intf == intf) {
|
||||
spin_lock_irq(&priv->ctx_lock);
|
||||
list_del(&dev_ctx->list);
|
||||
spin_unlock_irq(&priv->ctx_lock);
|
||||
|
||||
intf->remove(dev, dev_ctx->context);
|
||||
kfree(dev_ctx);
|
||||
return;
|
||||
}
|
||||
}
|
||||
static int mlx5_register_device(struct mlx5_core_dev *dev)
|
||||
{
|
||||
struct mlx5_priv *priv = &dev->priv;
|
||||
struct mlx5_interface *intf;
|
||||
|
||||
mutex_lock(&intf_mutex);
|
||||
list_add_tail(&priv->dev_list, &dev_list);
|
||||
list_for_each_entry(intf, &intf_list, list)
|
||||
mlx5_add_device(intf, priv);
|
||||
mutex_unlock(&intf_mutex);
|
||||
|
||||
return 0;
|
||||
}
|
||||
static void mlx5_unregister_device(struct mlx5_core_dev *dev)
|
||||
{
|
||||
struct mlx5_priv *priv = &dev->priv;
|
||||
struct mlx5_interface *intf;
|
||||
|
||||
mutex_lock(&intf_mutex);
|
||||
list_for_each_entry(intf, &intf_list, list)
|
||||
mlx5_remove_device(intf, priv);
|
||||
list_del(&priv->dev_list);
|
||||
mutex_unlock(&intf_mutex);
|
||||
}
|
||||
|
||||
int mlx5_register_interface(struct mlx5_interface *intf)
|
||||
{
|
||||
struct mlx5_priv *priv;
|
||||
|
||||
if (!intf->add || !intf->remove)
|
||||
return -EINVAL;
|
||||
|
||||
mutex_lock(&intf_mutex);
|
||||
list_add_tail(&intf->list, &intf_list);
|
||||
list_for_each_entry(priv, &dev_list, dev_list)
|
||||
mlx5_add_device(intf, priv);
|
||||
mutex_unlock(&intf_mutex);
|
||||
|
||||
return 0;
|
||||
}
|
||||
EXPORT_SYMBOL(mlx5_register_interface);
|
||||
|
||||
void mlx5_unregister_interface(struct mlx5_interface *intf)
|
||||
{
|
||||
struct mlx5_priv *priv;
|
||||
|
||||
mutex_lock(&intf_mutex);
|
||||
list_for_each_entry(priv, &dev_list, dev_list)
|
||||
mlx5_remove_device(intf, priv);
|
||||
list_del(&intf->list);
|
||||
mutex_unlock(&intf_mutex);
|
||||
}
|
||||
EXPORT_SYMBOL(mlx5_unregister_interface);
|
||||
|
||||
void *mlx5_get_protocol_dev(struct mlx5_core_dev *mdev, int protocol)
|
||||
{
|
||||
struct mlx5_priv *priv = &mdev->priv;
|
||||
struct mlx5_device_context *dev_ctx;
|
||||
unsigned long flags;
|
||||
void *result = NULL;
|
||||
|
||||
spin_lock_irqsave(&priv->ctx_lock, flags);
|
||||
|
||||
list_for_each_entry(dev_ctx, &mdev->priv.ctx_list, list)
|
||||
if ((dev_ctx->intf->protocol == protocol) &&
|
||||
dev_ctx->intf->get_dev) {
|
||||
result = dev_ctx->intf->get_dev(dev_ctx->context);
|
||||
break;
|
||||
}
|
||||
|
||||
spin_unlock_irqrestore(&priv->ctx_lock, flags);
|
||||
|
||||
return result;
|
||||
}
|
||||
EXPORT_SYMBOL(mlx5_get_protocol_dev);
|
||||
|
||||
static void mlx5_core_event(struct mlx5_core_dev *dev, enum mlx5_dev_event event,
|
||||
unsigned long param)
|
||||
|
@ -1064,7 +1095,6 @@ struct mlx5_core_event_handler {
|
|||
void *data);
|
||||
};
|
||||
|
||||
#define MLX5_IB_MOD "mlx5_ib"
|
||||
|
||||
static int init_one(struct pci_dev *pdev,
|
||||
const struct pci_device_id *id)
|
||||
|
@ -1088,40 +1118,45 @@ static int init_one(struct pci_dev *pdev,
|
|||
prof_sel = MLX5_DEFAULT_PROF;
|
||||
}
|
||||
dev->profile = &profile[prof_sel];
|
||||
dev->pdev = pdev;
|
||||
dev->event = mlx5_core_event;
|
||||
|
||||
INIT_LIST_HEAD(&priv->ctx_list);
|
||||
spin_lock_init(&priv->ctx_lock);
|
||||
err = mlx5_dev_init(dev, pdev);
|
||||
err = mlx5_pci_init(dev, priv);
|
||||
if (err) {
|
||||
dev_err(&pdev->dev, "mlx5_dev_init failed %d\n", err);
|
||||
goto out;
|
||||
dev_err(&pdev->dev, "mlx5_pci_init failed with error code %d\n", err);
|
||||
goto clean_dev;
|
||||
}
|
||||
|
||||
err = mlx5_register_device(dev);
|
||||
err = mlx5_load_one(dev, priv);
|
||||
if (err) {
|
||||
dev_err(&pdev->dev, "mlx5_register_device failed %d\n", err);
|
||||
goto out_init;
|
||||
dev_err(&pdev->dev, "mlx5_load_one failed with error code %d\n", err);
|
||||
goto close_pci;
|
||||
}
|
||||
|
||||
err = request_module_nowait(MLX5_IB_MOD);
|
||||
if (err)
|
||||
pr_info("failed request module on %s\n", MLX5_IB_MOD);
|
||||
|
||||
return 0;
|
||||
|
||||
out_init:
|
||||
mlx5_dev_cleanup(dev);
|
||||
out:
|
||||
close_pci:
|
||||
mlx5_pci_close(dev, priv);
|
||||
clean_dev:
|
||||
pci_set_drvdata(pdev, NULL);
|
||||
kfree(dev);
|
||||
|
||||
return err;
|
||||
}
|
||||
|
||||
static void remove_one(struct pci_dev *pdev)
|
||||
{
|
||||
struct mlx5_core_dev *dev = pci_get_drvdata(pdev);
|
||||
struct mlx5_priv *priv = &dev->priv;
|
||||
|
||||
mlx5_unregister_device(dev);
|
||||
mlx5_dev_cleanup(dev);
|
||||
if (mlx5_unload_one(dev, priv)) {
|
||||
dev_err(&dev->pdev->dev, "mlx5_unload_one failed\n");
|
||||
return;
|
||||
}
|
||||
mlx5_pci_close(dev, priv);
|
||||
pci_set_drvdata(pdev, NULL);
|
||||
kfree(dev);
|
||||
}
|
||||
|
||||
|
|
|
@ -40,6 +40,7 @@ void mlx5_init_mr_table(struct mlx5_core_dev *dev)
|
|||
{
|
||||
struct mlx5_mr_table *table = &dev->priv.mr_table;
|
||||
|
||||
memset(table, 0, sizeof(*table));
|
||||
rwlock_init(&table->lock);
|
||||
INIT_RADIX_TREE(&table->tree, GFP_ATOMIC);
|
||||
}
|
||||
|
|
|
@ -275,12 +275,36 @@ out_alloc:
|
|||
|
||||
return err;
|
||||
}
|
||||
|
||||
static void page_notify_fail(struct mlx5_core_dev *dev, u16 func_id)
|
||||
{
|
||||
struct mlx5_manage_pages_inbox *in;
|
||||
struct mlx5_manage_pages_outbox out;
|
||||
int err;
|
||||
|
||||
in = kzalloc(sizeof(*in), GFP_KERNEL);
|
||||
if (!in)
|
||||
return;
|
||||
|
||||
memset(&out, 0, sizeof(out));
|
||||
in->hdr.opcode = cpu_to_be16(MLX5_CMD_OP_MANAGE_PAGES);
|
||||
in->hdr.opmod = cpu_to_be16(MLX5_PAGES_CANT_GIVE);
|
||||
in->func_id = cpu_to_be16(func_id);
|
||||
err = mlx5_cmd_exec(dev, in, sizeof(*in), &out, sizeof(out));
|
||||
if (!err)
|
||||
err = mlx5_cmd_status_to_err(&out.hdr);
|
||||
|
||||
if (err)
|
||||
mlx5_core_warn(dev, "page notify failed\n");
|
||||
|
||||
kfree(in);
|
||||
}
|
||||
|
||||
static int give_pages(struct mlx5_core_dev *dev, u16 func_id, int npages,
|
||||
int notify_fail)
|
||||
{
|
||||
struct mlx5_manage_pages_inbox *in;
|
||||
struct mlx5_manage_pages_outbox out;
|
||||
struct mlx5_manage_pages_inbox *nin;
|
||||
int inlen;
|
||||
u64 addr;
|
||||
int err;
|
||||
|
@ -289,8 +313,9 @@ static int give_pages(struct mlx5_core_dev *dev, u16 func_id, int npages,
|
|||
inlen = sizeof(*in) + npages * sizeof(in->pas[0]);
|
||||
in = mlx5_vzalloc(inlen);
|
||||
if (!in) {
|
||||
err = -ENOMEM;
|
||||
mlx5_core_warn(dev, "vzalloc failed %d\n", inlen);
|
||||
return -ENOMEM;
|
||||
goto out_free;
|
||||
}
|
||||
memset(&out, 0, sizeof(out));
|
||||
|
||||
|
@ -316,43 +341,29 @@ retry:
|
|||
if (err) {
|
||||
mlx5_core_warn(dev, "func_id 0x%x, npages %d, err %d\n",
|
||||
func_id, npages, err);
|
||||
goto out_alloc;
|
||||
goto out_4k;
|
||||
}
|
||||
dev->priv.fw_pages += npages;
|
||||
|
||||
if (out.hdr.status) {
|
||||
err = mlx5_cmd_status_to_err(&out.hdr);
|
||||
if (err) {
|
||||
mlx5_core_warn(dev, "func_id 0x%x, npages %d, status %d\n",
|
||||
func_id, npages, out.hdr.status);
|
||||
goto out_alloc;
|
||||
}
|
||||
err = mlx5_cmd_status_to_err(&out.hdr);
|
||||
if (err) {
|
||||
mlx5_core_warn(dev, "func_id 0x%x, npages %d, status %d\n",
|
||||
func_id, npages, out.hdr.status);
|
||||
goto out_4k;
|
||||
}
|
||||
|
||||
mlx5_core_dbg(dev, "err %d\n", err);
|
||||
|
||||
goto out_free;
|
||||
|
||||
out_alloc:
|
||||
if (notify_fail) {
|
||||
nin = kzalloc(sizeof(*nin), GFP_KERNEL);
|
||||
if (!nin) {
|
||||
mlx5_core_warn(dev, "allocation failed\n");
|
||||
goto out_4k;
|
||||
}
|
||||
memset(&out, 0, sizeof(out));
|
||||
nin->hdr.opcode = cpu_to_be16(MLX5_CMD_OP_MANAGE_PAGES);
|
||||
nin->hdr.opmod = cpu_to_be16(MLX5_PAGES_CANT_GIVE);
|
||||
if (mlx5_cmd_exec(dev, nin, sizeof(*nin), &out, sizeof(out)))
|
||||
mlx5_core_warn(dev, "page notify failed\n");
|
||||
kfree(nin);
|
||||
}
|
||||
kvfree(in);
|
||||
return 0;
|
||||
|
||||
out_4k:
|
||||
for (i--; i >= 0; i--)
|
||||
free_4k(dev, be64_to_cpu(in->pas[i]));
|
||||
out_free:
|
||||
kvfree(in);
|
||||
if (notify_fail)
|
||||
page_notify_fail(dev, func_id);
|
||||
return err;
|
||||
}
|
||||
|
||||
|
|
|
@ -90,16 +90,13 @@ int mlx5_set_port_caps(struct mlx5_core_dev *dev, u8 port_num, u32 caps)
|
|||
{
|
||||
struct mlx5_reg_pcap in;
|
||||
struct mlx5_reg_pcap out;
|
||||
int err;
|
||||
|
||||
memset(&in, 0, sizeof(in));
|
||||
in.caps_127_96 = cpu_to_be32(caps);
|
||||
in.port_num = port_num;
|
||||
|
||||
err = mlx5_core_access_reg(dev, &in, sizeof(in), &out,
|
||||
sizeof(out), MLX5_REG_PCAP, 0, 1);
|
||||
|
||||
return err;
|
||||
return mlx5_core_access_reg(dev, &in, sizeof(in), &out,
|
||||
sizeof(out), MLX5_REG_PCAP, 0, 1);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(mlx5_set_port_caps);
|
||||
|
||||
|
@ -107,16 +104,13 @@ int mlx5_query_port_ptys(struct mlx5_core_dev *dev, u32 *ptys,
|
|||
int ptys_size, int proto_mask, u8 local_port)
|
||||
{
|
||||
u32 in[MLX5_ST_SZ_DW(ptys_reg)];
|
||||
int err;
|
||||
|
||||
memset(in, 0, sizeof(in));
|
||||
MLX5_SET(ptys_reg, in, local_port, local_port);
|
||||
MLX5_SET(ptys_reg, in, proto_mask, proto_mask);
|
||||
|
||||
err = mlx5_core_access_reg(dev, in, sizeof(in), ptys,
|
||||
ptys_size, MLX5_REG_PTYS, 0, 0);
|
||||
|
||||
return err;
|
||||
return mlx5_core_access_reg(dev, in, sizeof(in), ptys,
|
||||
ptys_size, MLX5_REG_PTYS, 0, 0);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(mlx5_query_port_ptys);
|
||||
|
||||
|
@ -199,7 +193,6 @@ int mlx5_set_port_proto(struct mlx5_core_dev *dev, u32 proto_admin,
|
|||
{
|
||||
u32 in[MLX5_ST_SZ_DW(ptys_reg)];
|
||||
u32 out[MLX5_ST_SZ_DW(ptys_reg)];
|
||||
int err;
|
||||
|
||||
memset(in, 0, sizeof(in));
|
||||
|
||||
|
@ -210,9 +203,8 @@ int mlx5_set_port_proto(struct mlx5_core_dev *dev, u32 proto_admin,
|
|||
else
|
||||
MLX5_SET(ptys_reg, in, ib_proto_admin, proto_admin);
|
||||
|
||||
err = mlx5_core_access_reg(dev, in, sizeof(in), out,
|
||||
sizeof(out), MLX5_REG_PTYS, 0, 1);
|
||||
return err;
|
||||
return mlx5_core_access_reg(dev, in, sizeof(in), out,
|
||||
sizeof(out), MLX5_REG_PTYS, 0, 1);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(mlx5_set_port_proto);
|
||||
|
||||
|
@ -250,7 +242,7 @@ int mlx5_query_port_admin_status(struct mlx5_core_dev *dev,
|
|||
return err;
|
||||
|
||||
*status = MLX5_GET(paos_reg, out, admin_status);
|
||||
return err;
|
||||
return 0;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(mlx5_query_port_admin_status);
|
||||
|
||||
|
@ -308,15 +300,12 @@ static int mlx5_query_port_pvlc(struct mlx5_core_dev *dev, u32 *pvlc,
|
|||
int pvlc_size, u8 local_port)
|
||||
{
|
||||
u32 in[MLX5_ST_SZ_DW(pvlc_reg)];
|
||||
int err;
|
||||
|
||||
memset(in, 0, sizeof(in));
|
||||
MLX5_SET(ptys_reg, in, local_port, local_port);
|
||||
|
||||
err = mlx5_core_access_reg(dev, in, sizeof(in), pvlc,
|
||||
pvlc_size, MLX5_REG_PVLC, 0, 0);
|
||||
|
||||
return err;
|
||||
return mlx5_core_access_reg(dev, in, sizeof(in), pvlc,
|
||||
pvlc_size, MLX5_REG_PVLC, 0, 0);
|
||||
}
|
||||
|
||||
int mlx5_query_port_vl_hw_cap(struct mlx5_core_dev *dev,
|
||||
|
@ -339,16 +328,14 @@ int mlx5_set_port_pause(struct mlx5_core_dev *dev, u32 rx_pause, u32 tx_pause)
|
|||
{
|
||||
u32 in[MLX5_ST_SZ_DW(pfcc_reg)];
|
||||
u32 out[MLX5_ST_SZ_DW(pfcc_reg)];
|
||||
int err;
|
||||
|
||||
memset(in, 0, sizeof(in));
|
||||
MLX5_SET(pfcc_reg, in, local_port, 1);
|
||||
MLX5_SET(pfcc_reg, in, pptx, tx_pause);
|
||||
MLX5_SET(pfcc_reg, in, pprx, rx_pause);
|
||||
|
||||
err = mlx5_core_access_reg(dev, in, sizeof(in), out,
|
||||
sizeof(out), MLX5_REG_PFCC, 0, 1);
|
||||
return err;
|
||||
return mlx5_core_access_reg(dev, in, sizeof(in), out,
|
||||
sizeof(out), MLX5_REG_PFCC, 0, 1);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(mlx5_set_port_pause);
|
||||
|
||||
|
|
|
@ -345,6 +345,7 @@ void mlx5_init_qp_table(struct mlx5_core_dev *dev)
|
|||
{
|
||||
struct mlx5_qp_table *table = &dev->priv.qp_table;
|
||||
|
||||
memset(table, 0, sizeof(*table));
|
||||
spin_lock_init(&table->lock);
|
||||
INIT_RADIX_TREE(&table->tree, GFP_ATOMIC);
|
||||
mlx5_qp_debugfs_init(dev);
|
||||
|
|
|
@ -531,6 +531,7 @@ void mlx5_init_srq_table(struct mlx5_core_dev *dev)
|
|||
{
|
||||
struct mlx5_srq_table *table = &dev->priv.srq_table;
|
||||
|
||||
memset(table, 0, sizeof(*table));
|
||||
spin_lock_init(&table->lock);
|
||||
INIT_RADIX_TREE(&table->tree, GFP_ATOMIC);
|
||||
}
|
||||
|
|
|
@ -177,7 +177,7 @@ int mlx5_core_modify_tir(struct mlx5_core_dev *dev, u32 tirn, u32 *in,
|
|||
|
||||
void mlx5_core_destroy_tir(struct mlx5_core_dev *dev, u32 tirn)
|
||||
{
|
||||
u32 in[MLX5_ST_SZ_DW(destroy_tir_out)];
|
||||
u32 in[MLX5_ST_SZ_DW(destroy_tir_in)];
|
||||
u32 out[MLX5_ST_SZ_DW(destroy_tir_out)];
|
||||
|
||||
memset(in, 0, sizeof(in));
|
||||
|
@ -206,7 +206,7 @@ int mlx5_core_create_tis(struct mlx5_core_dev *dev, u32 *in, int inlen,
|
|||
|
||||
void mlx5_core_destroy_tis(struct mlx5_core_dev *dev, u32 tisn)
|
||||
{
|
||||
u32 in[MLX5_ST_SZ_DW(destroy_tis_out)];
|
||||
u32 in[MLX5_ST_SZ_DW(destroy_tis_in)];
|
||||
u32 out[MLX5_ST_SZ_DW(destroy_tis_out)];
|
||||
|
||||
memset(in, 0, sizeof(in));
|
||||
|
|
|
@ -440,7 +440,7 @@ struct health_buffer {
|
|||
__be32 rsvd2;
|
||||
u8 irisc_index;
|
||||
u8 synd;
|
||||
__be16 ext_sync;
|
||||
__be16 ext_synd;
|
||||
};
|
||||
|
||||
struct mlx5_init_seg {
|
||||
|
|
Loading…
Reference in New Issue