RDMA/cma: Allow early transition to RTS to handle lost CM messages

During connection establishment, the passive side of a connection can
receive messages from the active side before the connection event has
been delivered to the user.  Allow the passive side to send messages
in response to received data before the event is delivered.  To handle
the case where the connection messages are lost, a new rdma_notify()
function is added that users may invoke to force a connection into the
established state.

Signed-off-by: Sean Hefty <sean.hefty@intel.com>
Signed-off-by: Roland Dreier <rolandd@cisco.com>
This commit is contained in:
Sean Hefty 2006-11-30 16:37:15 -08:00 committed by Roland Dreier
parent a1b1b61f80
commit 0fe313b000
3 changed files with 63 additions and 24 deletions

View File

@ -3289,6 +3289,10 @@ static int cm_init_qp_rts_attr(struct cm_id_private *cm_id_priv,
spin_lock_irqsave(&cm_id_priv->lock, flags); spin_lock_irqsave(&cm_id_priv->lock, flags);
switch (cm_id_priv->id.state) { switch (cm_id_priv->id.state) {
/* Allow transition to RTS before sending REP */
case IB_CM_REQ_RCVD:
case IB_CM_MRA_REQ_SENT:
case IB_CM_REP_RCVD: case IB_CM_REP_RCVD:
case IB_CM_MRA_REP_SENT: case IB_CM_MRA_REP_SENT:
case IB_CM_REP_SENT: case IB_CM_REP_SENT:

View File

@ -760,22 +760,6 @@ static int cma_verify_rep(struct rdma_id_private *id_priv, void *data)
return 0; return 0;
} }
static int cma_rtu_recv(struct rdma_id_private *id_priv)
{
int ret;
ret = cma_modify_qp_rts(&id_priv->id);
if (ret)
goto reject;
return 0;
reject:
cma_modify_qp_err(&id_priv->id);
ib_send_cm_rej(id_priv->cm_id.ib, IB_CM_REJ_CONSUMER_DEFINED,
NULL, 0, NULL, 0);
return ret;
}
static void cma_set_rep_event_data(struct rdma_cm_event *event, static void cma_set_rep_event_data(struct rdma_cm_event *event,
struct ib_cm_rep_event_param *rep_data, struct ib_cm_rep_event_param *rep_data,
void *private_data) void *private_data)
@ -821,9 +805,8 @@ static int cma_ib_handler(struct ib_cm_id *cm_id, struct ib_cm_event *ib_event)
ib_event->private_data); ib_event->private_data);
break; break;
case IB_CM_RTU_RECEIVED: case IB_CM_RTU_RECEIVED:
event.status = cma_rtu_recv(id_priv); case IB_CM_USER_ESTABLISHED:
event.event = event.status ? RDMA_CM_EVENT_CONNECT_ERROR : event.event = RDMA_CM_EVENT_ESTABLISHED;
RDMA_CM_EVENT_ESTABLISHED;
break; break;
case IB_CM_DREQ_ERROR: case IB_CM_DREQ_ERROR:
event.status = -ETIMEDOUT; /* fall through */ event.status = -ETIMEDOUT; /* fall through */
@ -1989,11 +1972,25 @@ static int cma_accept_ib(struct rdma_id_private *id_priv,
struct rdma_conn_param *conn_param) struct rdma_conn_param *conn_param)
{ {
struct ib_cm_rep_param rep; struct ib_cm_rep_param rep;
int ret; struct ib_qp_attr qp_attr;
int qp_attr_mask, ret;
ret = cma_modify_qp_rtr(&id_priv->id); if (id_priv->id.qp) {
if (ret) ret = cma_modify_qp_rtr(&id_priv->id);
return ret; if (ret)
goto out;
qp_attr.qp_state = IB_QPS_RTS;
ret = ib_cm_init_qp_attr(id_priv->cm_id.ib, &qp_attr,
&qp_attr_mask);
if (ret)
goto out;
qp_attr.max_rd_atomic = conn_param->initiator_depth;
ret = ib_modify_qp(id_priv->id.qp, &qp_attr, qp_attr_mask);
if (ret)
goto out;
}
memset(&rep, 0, sizeof rep); memset(&rep, 0, sizeof rep);
rep.qp_num = id_priv->qp_num; rep.qp_num = id_priv->qp_num;
@ -2008,7 +2005,9 @@ static int cma_accept_ib(struct rdma_id_private *id_priv,
rep.rnr_retry_count = conn_param->rnr_retry_count; rep.rnr_retry_count = conn_param->rnr_retry_count;
rep.srq = id_priv->srq ? 1 : 0; rep.srq = id_priv->srq ? 1 : 0;
return ib_send_cm_rep(id_priv->cm_id.ib, &rep); ret = ib_send_cm_rep(id_priv->cm_id.ib, &rep);
out:
return ret;
} }
static int cma_accept_iw(struct rdma_id_private *id_priv, static int cma_accept_iw(struct rdma_id_private *id_priv,
@ -2073,6 +2072,27 @@ reject:
} }
EXPORT_SYMBOL(rdma_accept); EXPORT_SYMBOL(rdma_accept);
int rdma_notify(struct rdma_cm_id *id, enum ib_event_type event)
{
struct rdma_id_private *id_priv;
int ret;
id_priv = container_of(id, struct rdma_id_private, id);
if (!cma_comp(id_priv, CMA_CONNECT))
return -EINVAL;
switch (id->device->node_type) {
case RDMA_NODE_IB_CA:
ret = ib_cm_notify(id_priv->cm_id.ib, event);
break;
default:
ret = 0;
break;
}
return ret;
}
EXPORT_SYMBOL(rdma_notify);
int rdma_reject(struct rdma_cm_id *id, const void *private_data, int rdma_reject(struct rdma_cm_id *id, const void *private_data,
u8 private_data_len) u8 private_data_len)
{ {

View File

@ -252,6 +252,21 @@ int rdma_listen(struct rdma_cm_id *id, int backlog);
*/ */
int rdma_accept(struct rdma_cm_id *id, struct rdma_conn_param *conn_param); int rdma_accept(struct rdma_cm_id *id, struct rdma_conn_param *conn_param);
/**
* rdma_notify - Notifies the RDMA CM of an asynchronous event that has
* occurred on the connection.
* @id: Connection identifier to transition to established.
* @event: Asynchronous event.
*
* This routine should be invoked by users to notify the CM of relevant
* communication events. Events that should be reported to the CM and
* when to report them are:
*
* IB_EVENT_COMM_EST - Used when a message is received on a connected
* QP before an RTU has been received.
*/
int rdma_notify(struct rdma_cm_id *id, enum ib_event_type event);
/** /**
* rdma_reject - Called to reject a connection request or response. * rdma_reject - Called to reject a connection request or response.
*/ */