RDMA/mlx5: Move xlt_emergency_page_mutex into mr.c
This is the only user, so remove the wrappers. Link: https://lore.kernel.org/r/20201026132314.1336717-3-leon@kernel.org Signed-off-by: Leon Romanovsky <leonro@nvidia.com> Signed-off-by: Jason Gunthorpe <jgg@nvidia.com>
This commit is contained in:
parent
aab8d3966d
commit
f22c30aa6d
|
@ -75,12 +75,6 @@ static LIST_HEAD(mlx5_ib_dev_list);
|
||||||
*/
|
*/
|
||||||
static DEFINE_MUTEX(mlx5_ib_multiport_mutex);
|
static DEFINE_MUTEX(mlx5_ib_multiport_mutex);
|
||||||
|
|
||||||
/* We can't use an array for xlt_emergency_page because dma_map_single
|
|
||||||
* doesn't work on kernel modules memory
|
|
||||||
*/
|
|
||||||
static unsigned long xlt_emergency_page;
|
|
||||||
static struct mutex xlt_emergency_page_mutex;
|
|
||||||
|
|
||||||
struct mlx5_ib_dev *mlx5_ib_get_ibdev_from_mpi(struct mlx5_ib_multiport_info *mpi)
|
struct mlx5_ib_dev *mlx5_ib_get_ibdev_from_mpi(struct mlx5_ib_multiport_info *mpi)
|
||||||
{
|
{
|
||||||
struct mlx5_ib_dev *dev;
|
struct mlx5_ib_dev *dev;
|
||||||
|
@ -4804,17 +4798,6 @@ static struct mlx5_interface mlx5_ib_interface = {
|
||||||
.protocol = MLX5_INTERFACE_PROTOCOL_IB,
|
.protocol = MLX5_INTERFACE_PROTOCOL_IB,
|
||||||
};
|
};
|
||||||
|
|
||||||
unsigned long mlx5_ib_get_xlt_emergency_page(void)
|
|
||||||
{
|
|
||||||
mutex_lock(&xlt_emergency_page_mutex);
|
|
||||||
return xlt_emergency_page;
|
|
||||||
}
|
|
||||||
|
|
||||||
void mlx5_ib_put_xlt_emergency_page(void)
|
|
||||||
{
|
|
||||||
mutex_unlock(&xlt_emergency_page_mutex);
|
|
||||||
}
|
|
||||||
|
|
||||||
static int __init mlx5_ib_init(void)
|
static int __init mlx5_ib_init(void)
|
||||||
{
|
{
|
||||||
int err;
|
int err;
|
||||||
|
@ -4823,8 +4806,6 @@ static int __init mlx5_ib_init(void)
|
||||||
if (!xlt_emergency_page)
|
if (!xlt_emergency_page)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
mutex_init(&xlt_emergency_page_mutex);
|
|
||||||
|
|
||||||
mlx5_ib_event_wq = alloc_ordered_workqueue("mlx5_ib_event_wq", 0);
|
mlx5_ib_event_wq = alloc_ordered_workqueue("mlx5_ib_event_wq", 0);
|
||||||
if (!mlx5_ib_event_wq) {
|
if (!mlx5_ib_event_wq) {
|
||||||
free_page(xlt_emergency_page);
|
free_page(xlt_emergency_page);
|
||||||
|
@ -4842,7 +4823,6 @@ static void __exit mlx5_ib_cleanup(void)
|
||||||
{
|
{
|
||||||
mlx5_unregister_interface(&mlx5_ib_interface);
|
mlx5_unregister_interface(&mlx5_ib_interface);
|
||||||
destroy_workqueue(mlx5_ib_event_wq);
|
destroy_workqueue(mlx5_ib_event_wq);
|
||||||
mutex_destroy(&xlt_emergency_page_mutex);
|
|
||||||
free_page(xlt_emergency_page);
|
free_page(xlt_emergency_page);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1454,8 +1454,7 @@ static inline int get_num_static_uars(struct mlx5_ib_dev *dev,
|
||||||
return get_uars_per_sys_page(dev, bfregi->lib_uar_4k) * bfregi->num_static_sys_pages;
|
return get_uars_per_sys_page(dev, bfregi->lib_uar_4k) * bfregi->num_static_sys_pages;
|
||||||
}
|
}
|
||||||
|
|
||||||
unsigned long mlx5_ib_get_xlt_emergency_page(void);
|
extern unsigned long xlt_emergency_page;
|
||||||
void mlx5_ib_put_xlt_emergency_page(void);
|
|
||||||
|
|
||||||
int bfregn_to_uar_index(struct mlx5_ib_dev *dev,
|
int bfregn_to_uar_index(struct mlx5_ib_dev *dev,
|
||||||
struct mlx5_bfreg_info *bfregi, u32 bfregn,
|
struct mlx5_bfreg_info *bfregi, u32 bfregn,
|
||||||
|
|
|
@ -41,6 +41,13 @@
|
||||||
#include <rdma/ib_verbs.h>
|
#include <rdma/ib_verbs.h>
|
||||||
#include "mlx5_ib.h"
|
#include "mlx5_ib.h"
|
||||||
|
|
||||||
|
/*
|
||||||
|
* We can't use an array for xlt_emergency_page because dma_map_single doesn't
|
||||||
|
* work on kernel modules memory
|
||||||
|
*/
|
||||||
|
unsigned long xlt_emergency_page;
|
||||||
|
static DEFINE_MUTEX(xlt_emergency_page_mutex);
|
||||||
|
|
||||||
enum {
|
enum {
|
||||||
MAX_PENDING_REG_MR = 8,
|
MAX_PENDING_REG_MR = 8,
|
||||||
};
|
};
|
||||||
|
@ -992,6 +999,17 @@ static struct mlx5_ib_mr *alloc_mr_from_cache(struct ib_pd *pd,
|
||||||
MLX5_UMR_MTT_ALIGNMENT)
|
MLX5_UMR_MTT_ALIGNMENT)
|
||||||
#define MLX5_SPARE_UMR_CHUNK 0x10000
|
#define MLX5_SPARE_UMR_CHUNK 0x10000
|
||||||
|
|
||||||
|
static unsigned long mlx5_ib_get_xlt_emergency_page(void)
|
||||||
|
{
|
||||||
|
mutex_lock(&xlt_emergency_page_mutex);
|
||||||
|
return xlt_emergency_page;
|
||||||
|
}
|
||||||
|
|
||||||
|
static void mlx5_ib_put_xlt_emergency_page(void)
|
||||||
|
{
|
||||||
|
mutex_unlock(&xlt_emergency_page_mutex);
|
||||||
|
}
|
||||||
|
|
||||||
int mlx5_ib_update_xlt(struct mlx5_ib_mr *mr, u64 idx, int npages,
|
int mlx5_ib_update_xlt(struct mlx5_ib_mr *mr, u64 idx, int npages,
|
||||||
int page_shift, int flags)
|
int page_shift, int flags)
|
||||||
{
|
{
|
||||||
|
|
Loading…
Reference in New Issue