xprtrdma: Prevent loss of completion signals
Commit8301a2c047
("xprtrdma: Limit work done by completion handler") was supposed to prevent xprtrdma's upcall handlers from starving other softIRQ work by letting them return to the provider before all CQEs have been polled. The logic assumes the provider will call the upcall handler again immediately if the CQ is re-armed while there are still queued CQEs. This assumption is invalid. The IBTA spec says that after a CQ is armed, the hardware must interrupt only when a new CQE is inserted. xprtrdma can't rely on the provider calling again, even though some providers do. Therefore, leaving CQEs on queue makes sense only when there is another mechanism that ensures all remaining CQEs are consumed in a timely fashion. xprtrdma does not have such a mechanism. If a CQE remains queued, the transport can wait forever to send the next RPC. Finally, move the wcs array back onto the stack to ensure that the poll array is always local to the CPU where the completion upcall is running. Fixes:8301a2c047
("xprtrdma: Limit work done by completion ...") Signed-off-by: Chuck Lever <chuck.lever@oracle.com> Reviewed-by: Sagi Grimberg <sagig@mellanox.com> Reviewed-by: Devesh Sharma <devesh.sharma@avagotech.com> Tested-By: Devesh Sharma <devesh.sharma@avagotech.com> Signed-off-by: Anna Schumaker <Anna.Schumaker@Netapp.com>
This commit is contained in:
parent
7b3d770c67
commit
4220a07264
|
@ -158,25 +158,30 @@ rpcrdma_sendcq_process_wc(struct ib_wc *wc)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static int
|
/* The common case is a single send completion is waiting. By
|
||||||
rpcrdma_sendcq_poll(struct ib_cq *cq, struct rpcrdma_ep *ep)
|
* passing two WC entries to ib_poll_cq, a return code of 1
|
||||||
|
* means there is exactly one WC waiting and no more. We don't
|
||||||
|
* have to invoke ib_poll_cq again to know that the CQ has been
|
||||||
|
* properly drained.
|
||||||
|
*/
|
||||||
|
static void
|
||||||
|
rpcrdma_sendcq_poll(struct ib_cq *cq)
|
||||||
{
|
{
|
||||||
struct ib_wc *wcs;
|
struct ib_wc *pos, wcs[2];
|
||||||
int budget, count, rc;
|
int count, rc;
|
||||||
|
|
||||||
budget = RPCRDMA_WC_BUDGET / RPCRDMA_POLLSIZE;
|
|
||||||
do {
|
do {
|
||||||
wcs = ep->rep_send_wcs;
|
pos = wcs;
|
||||||
|
|
||||||
rc = ib_poll_cq(cq, RPCRDMA_POLLSIZE, wcs);
|
rc = ib_poll_cq(cq, ARRAY_SIZE(wcs), pos);
|
||||||
if (rc <= 0)
|
if (rc < 0)
|
||||||
return rc;
|
break;
|
||||||
|
|
||||||
count = rc;
|
count = rc;
|
||||||
while (count-- > 0)
|
while (count-- > 0)
|
||||||
rpcrdma_sendcq_process_wc(wcs++);
|
rpcrdma_sendcq_process_wc(pos++);
|
||||||
} while (rc == RPCRDMA_POLLSIZE && --budget);
|
} while (rc == ARRAY_SIZE(wcs));
|
||||||
return 0;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Handle provider send completion upcalls.
|
/* Handle provider send completion upcalls.
|
||||||
|
@ -184,10 +189,8 @@ rpcrdma_sendcq_poll(struct ib_cq *cq, struct rpcrdma_ep *ep)
|
||||||
static void
|
static void
|
||||||
rpcrdma_sendcq_upcall(struct ib_cq *cq, void *cq_context)
|
rpcrdma_sendcq_upcall(struct ib_cq *cq, void *cq_context)
|
||||||
{
|
{
|
||||||
struct rpcrdma_ep *ep = (struct rpcrdma_ep *)cq_context;
|
|
||||||
|
|
||||||
do {
|
do {
|
||||||
rpcrdma_sendcq_poll(cq, ep);
|
rpcrdma_sendcq_poll(cq);
|
||||||
} while (ib_req_notify_cq(cq, IB_CQ_NEXT_COMP |
|
} while (ib_req_notify_cq(cq, IB_CQ_NEXT_COMP |
|
||||||
IB_CQ_REPORT_MISSED_EVENTS) > 0);
|
IB_CQ_REPORT_MISSED_EVENTS) > 0);
|
||||||
}
|
}
|
||||||
|
@ -226,31 +229,32 @@ out_fail:
|
||||||
goto out_schedule;
|
goto out_schedule;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int
|
/* The wc array is on stack: automatic memory is always CPU-local.
|
||||||
rpcrdma_recvcq_poll(struct ib_cq *cq, struct rpcrdma_ep *ep)
|
*
|
||||||
|
* struct ib_wc is 64 bytes, making the poll array potentially
|
||||||
|
* large. But this is at the bottom of the call chain. Further
|
||||||
|
* substantial work is done in another thread.
|
||||||
|
*/
|
||||||
|
static void
|
||||||
|
rpcrdma_recvcq_poll(struct ib_cq *cq)
|
||||||
{
|
{
|
||||||
struct list_head sched_list;
|
struct ib_wc *pos, wcs[4];
|
||||||
struct ib_wc *wcs;
|
LIST_HEAD(sched_list);
|
||||||
int budget, count, rc;
|
int count, rc;
|
||||||
|
|
||||||
INIT_LIST_HEAD(&sched_list);
|
|
||||||
budget = RPCRDMA_WC_BUDGET / RPCRDMA_POLLSIZE;
|
|
||||||
do {
|
do {
|
||||||
wcs = ep->rep_recv_wcs;
|
pos = wcs;
|
||||||
|
|
||||||
rc = ib_poll_cq(cq, RPCRDMA_POLLSIZE, wcs);
|
rc = ib_poll_cq(cq, ARRAY_SIZE(wcs), pos);
|
||||||
if (rc <= 0)
|
if (rc < 0)
|
||||||
goto out_schedule;
|
break;
|
||||||
|
|
||||||
count = rc;
|
count = rc;
|
||||||
while (count-- > 0)
|
while (count-- > 0)
|
||||||
rpcrdma_recvcq_process_wc(wcs++, &sched_list);
|
rpcrdma_recvcq_process_wc(pos++, &sched_list);
|
||||||
} while (rc == RPCRDMA_POLLSIZE && --budget);
|
} while (rc == ARRAY_SIZE(wcs));
|
||||||
rc = 0;
|
|
||||||
|
|
||||||
out_schedule:
|
|
||||||
rpcrdma_schedule_tasklet(&sched_list);
|
rpcrdma_schedule_tasklet(&sched_list);
|
||||||
return rc;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Handle provider receive completion upcalls.
|
/* Handle provider receive completion upcalls.
|
||||||
|
@ -258,10 +262,8 @@ out_schedule:
|
||||||
static void
|
static void
|
||||||
rpcrdma_recvcq_upcall(struct ib_cq *cq, void *cq_context)
|
rpcrdma_recvcq_upcall(struct ib_cq *cq, void *cq_context)
|
||||||
{
|
{
|
||||||
struct rpcrdma_ep *ep = (struct rpcrdma_ep *)cq_context;
|
|
||||||
|
|
||||||
do {
|
do {
|
||||||
rpcrdma_recvcq_poll(cq, ep);
|
rpcrdma_recvcq_poll(cq);
|
||||||
} while (ib_req_notify_cq(cq, IB_CQ_NEXT_COMP |
|
} while (ib_req_notify_cq(cq, IB_CQ_NEXT_COMP |
|
||||||
IB_CQ_REPORT_MISSED_EVENTS) > 0);
|
IB_CQ_REPORT_MISSED_EVENTS) > 0);
|
||||||
}
|
}
|
||||||
|
@ -623,7 +625,7 @@ rpcrdma_ep_create(struct rpcrdma_ep *ep, struct rpcrdma_ia *ia,
|
||||||
|
|
||||||
cq_attr.cqe = ep->rep_attr.cap.max_send_wr + 1;
|
cq_attr.cqe = ep->rep_attr.cap.max_send_wr + 1;
|
||||||
sendcq = ib_create_cq(ia->ri_device, rpcrdma_sendcq_upcall,
|
sendcq = ib_create_cq(ia->ri_device, rpcrdma_sendcq_upcall,
|
||||||
rpcrdma_cq_async_error_upcall, ep, &cq_attr);
|
rpcrdma_cq_async_error_upcall, NULL, &cq_attr);
|
||||||
if (IS_ERR(sendcq)) {
|
if (IS_ERR(sendcq)) {
|
||||||
rc = PTR_ERR(sendcq);
|
rc = PTR_ERR(sendcq);
|
||||||
dprintk("RPC: %s: failed to create send CQ: %i\n",
|
dprintk("RPC: %s: failed to create send CQ: %i\n",
|
||||||
|
@ -640,7 +642,7 @@ rpcrdma_ep_create(struct rpcrdma_ep *ep, struct rpcrdma_ia *ia,
|
||||||
|
|
||||||
cq_attr.cqe = ep->rep_attr.cap.max_recv_wr + 1;
|
cq_attr.cqe = ep->rep_attr.cap.max_recv_wr + 1;
|
||||||
recvcq = ib_create_cq(ia->ri_device, rpcrdma_recvcq_upcall,
|
recvcq = ib_create_cq(ia->ri_device, rpcrdma_recvcq_upcall,
|
||||||
rpcrdma_cq_async_error_upcall, ep, &cq_attr);
|
rpcrdma_cq_async_error_upcall, NULL, &cq_attr);
|
||||||
if (IS_ERR(recvcq)) {
|
if (IS_ERR(recvcq)) {
|
||||||
rc = PTR_ERR(recvcq);
|
rc = PTR_ERR(recvcq);
|
||||||
dprintk("RPC: %s: failed to create recv CQ: %i\n",
|
dprintk("RPC: %s: failed to create recv CQ: %i\n",
|
||||||
|
|
|
@ -77,9 +77,6 @@ struct rpcrdma_ia {
|
||||||
* RDMA Endpoint -- one per transport instance
|
* RDMA Endpoint -- one per transport instance
|
||||||
*/
|
*/
|
||||||
|
|
||||||
#define RPCRDMA_WC_BUDGET (128)
|
|
||||||
#define RPCRDMA_POLLSIZE (16)
|
|
||||||
|
|
||||||
struct rpcrdma_ep {
|
struct rpcrdma_ep {
|
||||||
atomic_t rep_cqcount;
|
atomic_t rep_cqcount;
|
||||||
int rep_cqinit;
|
int rep_cqinit;
|
||||||
|
@ -89,8 +86,6 @@ struct rpcrdma_ep {
|
||||||
struct rdma_conn_param rep_remote_cma;
|
struct rdma_conn_param rep_remote_cma;
|
||||||
struct sockaddr_storage rep_remote_addr;
|
struct sockaddr_storage rep_remote_addr;
|
||||||
struct delayed_work rep_connect_worker;
|
struct delayed_work rep_connect_worker;
|
||||||
struct ib_wc rep_send_wcs[RPCRDMA_POLLSIZE];
|
|
||||||
struct ib_wc rep_recv_wcs[RPCRDMA_POLLSIZE];
|
|
||||||
};
|
};
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
|
Loading…
Reference in New Issue