[PATCH] IB uverbs: update kernel midlayer for new API
Update kernel InfiniBand midlayer to compile against the updated API for low-level drivers. This just amounts to passing NULL for all userspace-related parameters, and setting userspace-related structure members to NULL. Signed-off-by: Roland Dreier <rolandd@cisco.com> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
This commit is contained in:
parent
e2773c062e
commit
b5e81bf5e7
|
@ -4,6 +4,7 @@
|
||||||
* Copyright (c) 2004 Intel Corporation. All rights reserved.
|
* Copyright (c) 2004 Intel Corporation. All rights reserved.
|
||||||
* Copyright (c) 2004 Topspin Corporation. All rights reserved.
|
* Copyright (c) 2004 Topspin Corporation. All rights reserved.
|
||||||
* Copyright (c) 2004 Voltaire Corporation. All rights reserved.
|
* Copyright (c) 2004 Voltaire Corporation. All rights reserved.
|
||||||
|
* Copyright (c) 2005 Cisco Systems. All rights reserved.
|
||||||
*
|
*
|
||||||
* This software is available to you under a choice of one of two
|
* This software is available to you under a choice of one of two
|
||||||
* licenses. You may choose to be licensed under the terms of the GNU
|
* licenses. You may choose to be licensed under the terms of the GNU
|
||||||
|
@ -47,10 +48,11 @@ struct ib_pd *ib_alloc_pd(struct ib_device *device)
|
||||||
{
|
{
|
||||||
struct ib_pd *pd;
|
struct ib_pd *pd;
|
||||||
|
|
||||||
pd = device->alloc_pd(device);
|
pd = device->alloc_pd(device, NULL, NULL);
|
||||||
|
|
||||||
if (!IS_ERR(pd)) {
|
if (!IS_ERR(pd)) {
|
||||||
pd->device = device;
|
pd->device = device;
|
||||||
|
pd->uobject = NULL;
|
||||||
atomic_set(&pd->usecnt, 0);
|
atomic_set(&pd->usecnt, 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -76,8 +78,9 @@ struct ib_ah *ib_create_ah(struct ib_pd *pd, struct ib_ah_attr *ah_attr)
|
||||||
ah = pd->device->create_ah(pd, ah_attr);
|
ah = pd->device->create_ah(pd, ah_attr);
|
||||||
|
|
||||||
if (!IS_ERR(ah)) {
|
if (!IS_ERR(ah)) {
|
||||||
ah->device = pd->device;
|
ah->device = pd->device;
|
||||||
ah->pd = pd;
|
ah->pd = pd;
|
||||||
|
ah->uobject = NULL;
|
||||||
atomic_inc(&pd->usecnt);
|
atomic_inc(&pd->usecnt);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -122,7 +125,7 @@ struct ib_qp *ib_create_qp(struct ib_pd *pd,
|
||||||
{
|
{
|
||||||
struct ib_qp *qp;
|
struct ib_qp *qp;
|
||||||
|
|
||||||
qp = pd->device->create_qp(pd, qp_init_attr);
|
qp = pd->device->create_qp(pd, qp_init_attr, NULL);
|
||||||
|
|
||||||
if (!IS_ERR(qp)) {
|
if (!IS_ERR(qp)) {
|
||||||
qp->device = pd->device;
|
qp->device = pd->device;
|
||||||
|
@ -130,6 +133,7 @@ struct ib_qp *ib_create_qp(struct ib_pd *pd,
|
||||||
qp->send_cq = qp_init_attr->send_cq;
|
qp->send_cq = qp_init_attr->send_cq;
|
||||||
qp->recv_cq = qp_init_attr->recv_cq;
|
qp->recv_cq = qp_init_attr->recv_cq;
|
||||||
qp->srq = qp_init_attr->srq;
|
qp->srq = qp_init_attr->srq;
|
||||||
|
qp->uobject = NULL;
|
||||||
qp->event_handler = qp_init_attr->event_handler;
|
qp->event_handler = qp_init_attr->event_handler;
|
||||||
qp->qp_context = qp_init_attr->qp_context;
|
qp->qp_context = qp_init_attr->qp_context;
|
||||||
qp->qp_type = qp_init_attr->qp_type;
|
qp->qp_type = qp_init_attr->qp_type;
|
||||||
|
@ -197,10 +201,11 @@ struct ib_cq *ib_create_cq(struct ib_device *device,
|
||||||
{
|
{
|
||||||
struct ib_cq *cq;
|
struct ib_cq *cq;
|
||||||
|
|
||||||
cq = device->create_cq(device, cqe);
|
cq = device->create_cq(device, cqe, NULL, NULL);
|
||||||
|
|
||||||
if (!IS_ERR(cq)) {
|
if (!IS_ERR(cq)) {
|
||||||
cq->device = device;
|
cq->device = device;
|
||||||
|
cq->uobject = NULL;
|
||||||
cq->comp_handler = comp_handler;
|
cq->comp_handler = comp_handler;
|
||||||
cq->event_handler = event_handler;
|
cq->event_handler = event_handler;
|
||||||
cq->cq_context = cq_context;
|
cq->cq_context = cq_context;
|
||||||
|
@ -245,8 +250,9 @@ struct ib_mr *ib_get_dma_mr(struct ib_pd *pd, int mr_access_flags)
|
||||||
mr = pd->device->get_dma_mr(pd, mr_access_flags);
|
mr = pd->device->get_dma_mr(pd, mr_access_flags);
|
||||||
|
|
||||||
if (!IS_ERR(mr)) {
|
if (!IS_ERR(mr)) {
|
||||||
mr->device = pd->device;
|
mr->device = pd->device;
|
||||||
mr->pd = pd;
|
mr->pd = pd;
|
||||||
|
mr->uobject = NULL;
|
||||||
atomic_inc(&pd->usecnt);
|
atomic_inc(&pd->usecnt);
|
||||||
atomic_set(&mr->usecnt, 0);
|
atomic_set(&mr->usecnt, 0);
|
||||||
}
|
}
|
||||||
|
@ -267,8 +273,9 @@ struct ib_mr *ib_reg_phys_mr(struct ib_pd *pd,
|
||||||
mr_access_flags, iova_start);
|
mr_access_flags, iova_start);
|
||||||
|
|
||||||
if (!IS_ERR(mr)) {
|
if (!IS_ERR(mr)) {
|
||||||
mr->device = pd->device;
|
mr->device = pd->device;
|
||||||
mr->pd = pd;
|
mr->pd = pd;
|
||||||
|
mr->uobject = NULL;
|
||||||
atomic_inc(&pd->usecnt);
|
atomic_inc(&pd->usecnt);
|
||||||
atomic_set(&mr->usecnt, 0);
|
atomic_set(&mr->usecnt, 0);
|
||||||
}
|
}
|
||||||
|
@ -344,8 +351,9 @@ struct ib_mw *ib_alloc_mw(struct ib_pd *pd)
|
||||||
|
|
||||||
mw = pd->device->alloc_mw(pd);
|
mw = pd->device->alloc_mw(pd);
|
||||||
if (!IS_ERR(mw)) {
|
if (!IS_ERR(mw)) {
|
||||||
mw->device = pd->device;
|
mw->device = pd->device;
|
||||||
mw->pd = pd;
|
mw->pd = pd;
|
||||||
|
mw->uobject = NULL;
|
||||||
atomic_inc(&pd->usecnt);
|
atomic_inc(&pd->usecnt);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue