mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-22 12:14:01 +08:00
xprtrdma: kmalloc rpcrdma_ep separate from rpcrdma_xprt
Change the rpcrdma_xprt_disconnect() function so that it no longer waits for the DISCONNECTED event. This prevents blocking if the remote is unresponsive. In rpcrdma_xprt_disconnect(), the transport's rpcrdma_ep is detached. Upon return from rpcrdma_xprt_disconnect(), the transport (r_xprt) is ready immediately for a new connection. The RDMA_CM_DEVICE_REMOVAL and RDMA_CM_DISCONNECTED events are now handled almost identically. However, because the lifetimes of rpcrdma_xprt structures and rpcrdma_ep structures are now independent, creating an rpcrdma_ep needs to take a module ref count. The ep now owns most of the hardware resources for a transport. Also, a kref is needed to ensure that rpcrdma_ep sticks around long enough for the cm_event_handler to finish. Signed-off-by: Chuck Lever <chuck.lever@oracle.com> Signed-off-by: Anna Schumaker <Anna.Schumaker@Netapp.com>
This commit is contained in:
parent
745b734c9b
commit
e28ce90083
@ -104,7 +104,7 @@ DECLARE_EVENT_CLASS(xprtrdma_connect_class,
|
||||
TP_fast_assign(
|
||||
__entry->r_xprt = r_xprt;
|
||||
__entry->rc = rc;
|
||||
__entry->connect_status = r_xprt->rx_ep.re_connect_status;
|
||||
__entry->connect_status = r_xprt->rx_ep->re_connect_status;
|
||||
__assign_str(addr, rpcrdma_addrstr(r_xprt));
|
||||
__assign_str(port, rpcrdma_portstr(r_xprt));
|
||||
),
|
||||
@ -342,37 +342,6 @@ DECLARE_EVENT_CLASS(xprtrdma_cb_event,
|
||||
** Connection events
|
||||
**/
|
||||
|
||||
TRACE_EVENT(xprtrdma_cm_event,
|
||||
TP_PROTO(
|
||||
const struct rpcrdma_xprt *r_xprt,
|
||||
struct rdma_cm_event *event
|
||||
),
|
||||
|
||||
TP_ARGS(r_xprt, event),
|
||||
|
||||
TP_STRUCT__entry(
|
||||
__field(const void *, r_xprt)
|
||||
__field(unsigned int, event)
|
||||
__field(int, status)
|
||||
__string(addr, rpcrdma_addrstr(r_xprt))
|
||||
__string(port, rpcrdma_portstr(r_xprt))
|
||||
),
|
||||
|
||||
TP_fast_assign(
|
||||
__entry->r_xprt = r_xprt;
|
||||
__entry->event = event->event;
|
||||
__entry->status = event->status;
|
||||
__assign_str(addr, rpcrdma_addrstr(r_xprt));
|
||||
__assign_str(port, rpcrdma_portstr(r_xprt));
|
||||
),
|
||||
|
||||
TP_printk("peer=[%s]:%s r_xprt=%p: %s (%u/%d)",
|
||||
__get_str(addr), __get_str(port),
|
||||
__entry->r_xprt, rdma_show_cm_event(__entry->event),
|
||||
__entry->event, __entry->status
|
||||
)
|
||||
);
|
||||
|
||||
TRACE_EVENT(xprtrdma_inline_thresh,
|
||||
TP_PROTO(
|
||||
const struct rpcrdma_ep *ep
|
||||
@ -409,34 +378,6 @@ TRACE_EVENT(xprtrdma_inline_thresh,
|
||||
)
|
||||
);
|
||||
|
||||
TRACE_EVENT(xprtrdma_remove,
|
||||
TP_PROTO(
|
||||
const struct rpcrdma_ep *ep
|
||||
),
|
||||
|
||||
TP_ARGS(ep),
|
||||
|
||||
TP_STRUCT__entry(
|
||||
__array(unsigned char, srcaddr, sizeof(struct sockaddr_in6))
|
||||
__array(unsigned char, dstaddr, sizeof(struct sockaddr_in6))
|
||||
__string(name, ep->re_id->device->name)
|
||||
),
|
||||
|
||||
TP_fast_assign(
|
||||
const struct rdma_cm_id *id = ep->re_id;
|
||||
|
||||
memcpy(__entry->srcaddr, &id->route.addr.src_addr,
|
||||
sizeof(struct sockaddr_in6));
|
||||
memcpy(__entry->dstaddr, &id->route.addr.dst_addr,
|
||||
sizeof(struct sockaddr_in6));
|
||||
__assign_str(name, id->device->name);
|
||||
),
|
||||
|
||||
TP_printk("%pISpc -> %pISpc device=%s",
|
||||
__entry->srcaddr, __entry->dstaddr, __get_str(name)
|
||||
)
|
||||
);
|
||||
|
||||
DEFINE_CONN_EVENT(connect);
|
||||
DEFINE_CONN_EVENT(disconnect);
|
||||
DEFINE_CONN_EVENT(flush_dct);
|
||||
@ -831,7 +772,7 @@ TRACE_EVENT(xprtrdma_post_recvs,
|
||||
__entry->r_xprt = r_xprt;
|
||||
__entry->count = count;
|
||||
__entry->status = status;
|
||||
__entry->posted = r_xprt->rx_ep.re_receive_count;
|
||||
__entry->posted = r_xprt->rx_ep->re_receive_count;
|
||||
__assign_str(addr, rpcrdma_addrstr(r_xprt));
|
||||
__assign_str(port, rpcrdma_portstr(r_xprt));
|
||||
),
|
||||
|
@ -44,7 +44,7 @@ int xprt_rdma_bc_setup(struct rpc_xprt *xprt, unsigned int reqs)
|
||||
size_t xprt_rdma_bc_maxpayload(struct rpc_xprt *xprt)
|
||||
{
|
||||
struct rpcrdma_xprt *r_xprt = rpcx_to_rdmax(xprt);
|
||||
struct rpcrdma_ep *ep = &r_xprt->rx_ep;
|
||||
struct rpcrdma_ep *ep = r_xprt->rx_ep;
|
||||
size_t maxmsg;
|
||||
|
||||
maxmsg = min_t(unsigned int, ep->re_inline_send, ep->re_inline_recv);
|
||||
@ -190,7 +190,7 @@ create_req:
|
||||
if (xprt->bc_alloc_count >= RPCRDMA_BACKWARD_WRS)
|
||||
return NULL;
|
||||
|
||||
size = min_t(size_t, r_xprt->rx_ep.re_inline_recv, PAGE_SIZE);
|
||||
size = min_t(size_t, r_xprt->rx_ep->re_inline_recv, PAGE_SIZE);
|
||||
req = rpcrdma_req_create(r_xprt, size, GFP_KERNEL);
|
||||
if (!req)
|
||||
return NULL;
|
||||
|
@ -74,7 +74,7 @@ static void frwr_mr_recycle(struct rpcrdma_mr *mr)
|
||||
|
||||
if (mr->mr_dir != DMA_NONE) {
|
||||
trace_xprtrdma_mr_unmap(mr);
|
||||
ib_dma_unmap_sg(r_xprt->rx_ep.re_id->device,
|
||||
ib_dma_unmap_sg(r_xprt->rx_ep->re_id->device,
|
||||
mr->mr_sg, mr->mr_nents, mr->mr_dir);
|
||||
mr->mr_dir = DMA_NONE;
|
||||
}
|
||||
@ -115,7 +115,7 @@ void frwr_reset(struct rpcrdma_req *req)
|
||||
*/
|
||||
int frwr_mr_init(struct rpcrdma_xprt *r_xprt, struct rpcrdma_mr *mr)
|
||||
{
|
||||
struct rpcrdma_ep *ep = &r_xprt->rx_ep;
|
||||
struct rpcrdma_ep *ep = r_xprt->rx_ep;
|
||||
unsigned int depth = ep->re_max_fr_depth;
|
||||
struct scatterlist *sg;
|
||||
struct ib_mr *frmr;
|
||||
@ -283,7 +283,7 @@ struct rpcrdma_mr_seg *frwr_map(struct rpcrdma_xprt *r_xprt,
|
||||
int nsegs, bool writing, __be32 xid,
|
||||
struct rpcrdma_mr *mr)
|
||||
{
|
||||
struct rpcrdma_ep *ep = &r_xprt->rx_ep;
|
||||
struct rpcrdma_ep *ep = r_xprt->rx_ep;
|
||||
struct ib_reg_wr *reg_wr;
|
||||
int i, n, dma_nents;
|
||||
struct ib_mr *ibmr;
|
||||
@ -405,7 +405,7 @@ int frwr_send(struct rpcrdma_xprt *r_xprt, struct rpcrdma_req *req)
|
||||
post_wr = &frwr->fr_regwr.wr;
|
||||
}
|
||||
|
||||
return ib_post_send(r_xprt->rx_ep.re_id->qp, post_wr, NULL);
|
||||
return ib_post_send(r_xprt->rx_ep->re_id->qp, post_wr, NULL);
|
||||
}
|
||||
|
||||
/**
|
||||
@ -535,7 +535,7 @@ void frwr_unmap_sync(struct rpcrdma_xprt *r_xprt, struct rpcrdma_req *req)
|
||||
* unless re_id->qp is a valid pointer.
|
||||
*/
|
||||
bad_wr = NULL;
|
||||
rc = ib_post_send(r_xprt->rx_ep.re_id->qp, first, &bad_wr);
|
||||
rc = ib_post_send(r_xprt->rx_ep->re_id->qp, first, &bad_wr);
|
||||
|
||||
/* The final LOCAL_INV WR in the chain is supposed to
|
||||
* do the wake. If it was never posted, the wake will
|
||||
@ -640,7 +640,7 @@ void frwr_unmap_async(struct rpcrdma_xprt *r_xprt, struct rpcrdma_req *req)
|
||||
* unless re_id->qp is a valid pointer.
|
||||
*/
|
||||
bad_wr = NULL;
|
||||
rc = ib_post_send(r_xprt->rx_ep.re_id->qp, first, &bad_wr);
|
||||
rc = ib_post_send(r_xprt->rx_ep->re_id->qp, first, &bad_wr);
|
||||
if (!rc)
|
||||
return;
|
||||
|
||||
|
@ -131,9 +131,10 @@ static bool rpcrdma_args_inline(struct rpcrdma_xprt *r_xprt,
|
||||
struct rpc_rqst *rqst)
|
||||
{
|
||||
struct xdr_buf *xdr = &rqst->rq_snd_buf;
|
||||
struct rpcrdma_ep *ep = r_xprt->rx_ep;
|
||||
unsigned int count, remaining, offset;
|
||||
|
||||
if (xdr->len > r_xprt->rx_ep.re_max_inline_send)
|
||||
if (xdr->len > ep->re_max_inline_send)
|
||||
return false;
|
||||
|
||||
if (xdr->page_len) {
|
||||
@ -144,7 +145,7 @@ static bool rpcrdma_args_inline(struct rpcrdma_xprt *r_xprt,
|
||||
remaining -= min_t(unsigned int,
|
||||
PAGE_SIZE - offset, remaining);
|
||||
offset = 0;
|
||||
if (++count > r_xprt->rx_ep.re_attr.cap.max_send_sge)
|
||||
if (++count > ep->re_attr.cap.max_send_sge)
|
||||
return false;
|
||||
}
|
||||
}
|
||||
@ -161,7 +162,7 @@ static bool rpcrdma_args_inline(struct rpcrdma_xprt *r_xprt,
|
||||
static bool rpcrdma_results_inline(struct rpcrdma_xprt *r_xprt,
|
||||
struct rpc_rqst *rqst)
|
||||
{
|
||||
return rqst->rq_rcv_buf.buflen <= r_xprt->rx_ep.re_max_inline_recv;
|
||||
return rqst->rq_rcv_buf.buflen <= r_xprt->rx_ep->re_max_inline_recv;
|
||||
}
|
||||
|
||||
/* The client is required to provide a Reply chunk if the maximum
|
||||
@ -175,7 +176,7 @@ rpcrdma_nonpayload_inline(const struct rpcrdma_xprt *r_xprt,
|
||||
const struct xdr_buf *buf = &rqst->rq_rcv_buf;
|
||||
|
||||
return (buf->head[0].iov_len + buf->tail[0].iov_len) <
|
||||
r_xprt->rx_ep.re_max_inline_recv;
|
||||
r_xprt->rx_ep->re_max_inline_recv;
|
||||
}
|
||||
|
||||
/* Split @vec on page boundaries into SGEs. FMR registers pages, not
|
||||
@ -254,7 +255,7 @@ rpcrdma_convert_iovs(struct rpcrdma_xprt *r_xprt, struct xdr_buf *xdrbuf,
|
||||
/* When encoding a Read chunk, the tail iovec contains an
|
||||
* XDR pad and may be omitted.
|
||||
*/
|
||||
if (type == rpcrdma_readch && r_xprt->rx_ep.re_implicit_roundup)
|
||||
if (type == rpcrdma_readch && r_xprt->rx_ep->re_implicit_roundup)
|
||||
goto out;
|
||||
|
||||
/* When encoding a Write chunk, some servers need to see an
|
||||
@ -262,7 +263,7 @@ rpcrdma_convert_iovs(struct rpcrdma_xprt *r_xprt, struct xdr_buf *xdrbuf,
|
||||
* layer provides space in the tail iovec that may be used
|
||||
* for this purpose.
|
||||
*/
|
||||
if (type == rpcrdma_writech && r_xprt->rx_ep.re_implicit_roundup)
|
||||
if (type == rpcrdma_writech && r_xprt->rx_ep->re_implicit_roundup)
|
||||
goto out;
|
||||
|
||||
if (xdrbuf->tail[0].iov_len)
|
||||
@ -1475,8 +1476,8 @@ void rpcrdma_reply_handler(struct rpcrdma_rep *rep)
|
||||
|
||||
if (credits == 0)
|
||||
credits = 1; /* don't deadlock */
|
||||
else if (credits > r_xprt->rx_ep.re_max_requests)
|
||||
credits = r_xprt->rx_ep.re_max_requests;
|
||||
else if (credits > r_xprt->rx_ep->re_max_requests)
|
||||
credits = r_xprt->rx_ep->re_max_requests;
|
||||
if (buf->rb_credits != credits)
|
||||
rpcrdma_update_cwnd(r_xprt, credits);
|
||||
rpcrdma_post_recvs(r_xprt, false);
|
||||
|
@ -238,12 +238,12 @@ xprt_rdma_connect_worker(struct work_struct *work)
|
||||
struct rpcrdma_xprt *r_xprt = container_of(work, struct rpcrdma_xprt,
|
||||
rx_connect_worker.work);
|
||||
struct rpc_xprt *xprt = &r_xprt->rx_xprt;
|
||||
struct rpcrdma_ep *ep = &r_xprt->rx_ep;
|
||||
int rc;
|
||||
|
||||
rc = rpcrdma_xprt_connect(r_xprt);
|
||||
xprt_clear_connecting(xprt);
|
||||
if (ep->re_connect_status > 0) {
|
||||
if (r_xprt->rx_ep && r_xprt->rx_ep->re_connect_status > 0) {
|
||||
xprt->connect_cookie++;
|
||||
xprt->stat.connect_count++;
|
||||
xprt->stat.connect_time += (long)jiffies -
|
||||
xprt->stat.connect_start;
|
||||
@ -266,7 +266,7 @@ xprt_rdma_inject_disconnect(struct rpc_xprt *xprt)
|
||||
struct rpcrdma_xprt *r_xprt = rpcx_to_rdmax(xprt);
|
||||
|
||||
trace_xprtrdma_op_inject_dsc(r_xprt);
|
||||
rdma_disconnect(r_xprt->rx_ep.re_id);
|
||||
rdma_disconnect(r_xprt->rx_ep->re_id);
|
||||
}
|
||||
|
||||
/**
|
||||
@ -316,10 +316,15 @@ xprt_setup_rdma(struct xprt_create *args)
|
||||
if (args->addrlen > sizeof(xprt->addr))
|
||||
return ERR_PTR(-EBADF);
|
||||
|
||||
if (!try_module_get(THIS_MODULE))
|
||||
return ERR_PTR(-EIO);
|
||||
|
||||
xprt = xprt_alloc(args->net, sizeof(struct rpcrdma_xprt), 0,
|
||||
xprt_rdma_slot_table_entries);
|
||||
if (!xprt)
|
||||
if (!xprt) {
|
||||
module_put(THIS_MODULE);
|
||||
return ERR_PTR(-ENOMEM);
|
||||
}
|
||||
|
||||
xprt->timeout = &xprt_rdma_default_timeout;
|
||||
xprt->connect_timeout = xprt->timeout->to_initval;
|
||||
@ -348,11 +353,12 @@ xprt_setup_rdma(struct xprt_create *args)
|
||||
|
||||
new_xprt = rpcx_to_rdmax(xprt);
|
||||
rc = rpcrdma_buffer_create(new_xprt);
|
||||
if (rc)
|
||||
goto out2;
|
||||
|
||||
if (!try_module_get(THIS_MODULE))
|
||||
goto out4;
|
||||
if (rc) {
|
||||
xprt_rdma_free_addresses(xprt);
|
||||
xprt_free(xprt);
|
||||
module_put(THIS_MODULE);
|
||||
return ERR_PTR(rc);
|
||||
}
|
||||
|
||||
INIT_DELAYED_WORK(&new_xprt->rx_connect_worker,
|
||||
xprt_rdma_connect_worker);
|
||||
@ -364,15 +370,6 @@ xprt_setup_rdma(struct xprt_create *args)
|
||||
xprt->address_strings[RPC_DISPLAY_PORT]);
|
||||
trace_xprtrdma_create(new_xprt);
|
||||
return xprt;
|
||||
|
||||
out4:
|
||||
rpcrdma_buffer_destroy(&new_xprt->rx_buf);
|
||||
rc = -ENODEV;
|
||||
out2:
|
||||
trace_xprtrdma_op_destroy(new_xprt);
|
||||
xprt_rdma_free_addresses(xprt);
|
||||
xprt_free(xprt);
|
||||
return ERR_PTR(rc);
|
||||
}
|
||||
|
||||
/**
|
||||
@ -491,11 +488,11 @@ static void
|
||||
xprt_rdma_connect(struct rpc_xprt *xprt, struct rpc_task *task)
|
||||
{
|
||||
struct rpcrdma_xprt *r_xprt = rpcx_to_rdmax(xprt);
|
||||
struct rpcrdma_ep *ep = &r_xprt->rx_ep;
|
||||
struct rpcrdma_ep *ep = r_xprt->rx_ep;
|
||||
unsigned long delay;
|
||||
|
||||
delay = 0;
|
||||
if (ep->re_connect_status != 0) {
|
||||
if (ep && ep->re_connect_status != 0) {
|
||||
delay = xprt_reconnect_delay(xprt);
|
||||
xprt_reconnect_backoff(xprt, RPCRDMA_INIT_REEST_TO);
|
||||
}
|
||||
|
@ -84,7 +84,7 @@ static void rpcrdma_rep_destroy(struct rpcrdma_rep *rep);
|
||||
static void rpcrdma_reps_unmap(struct rpcrdma_xprt *r_xprt);
|
||||
static void rpcrdma_mrs_create(struct rpcrdma_xprt *r_xprt);
|
||||
static void rpcrdma_mrs_destroy(struct rpcrdma_xprt *r_xprt);
|
||||
static void rpcrdma_ep_destroy(struct rpcrdma_xprt *r_xprt);
|
||||
static int rpcrdma_ep_destroy(struct rpcrdma_ep *ep);
|
||||
static struct rpcrdma_regbuf *
|
||||
rpcrdma_regbuf_alloc(size_t size, enum dma_data_direction direction,
|
||||
gfp_t flags);
|
||||
@ -97,7 +97,7 @@ static void rpcrdma_regbuf_free(struct rpcrdma_regbuf *rb);
|
||||
*/
|
||||
static void rpcrdma_xprt_drain(struct rpcrdma_xprt *r_xprt)
|
||||
{
|
||||
struct rdma_cm_id *id = r_xprt->rx_ep.re_id;
|
||||
struct rdma_cm_id *id = r_xprt->rx_ep->re_id;
|
||||
|
||||
/* Flush Receives, then wait for deferred Reply work
|
||||
* to complete.
|
||||
@ -139,8 +139,8 @@ void rpcrdma_flush_disconnect(struct ib_cq *cq, struct ib_wc *wc)
|
||||
struct rpc_xprt *xprt = &r_xprt->rx_xprt;
|
||||
|
||||
if (wc->status != IB_WC_SUCCESS &&
|
||||
r_xprt->rx_ep.re_connect_status == 1) {
|
||||
r_xprt->rx_ep.re_connect_status = -ECONNABORTED;
|
||||
r_xprt->rx_ep->re_connect_status == 1) {
|
||||
r_xprt->rx_ep->re_connect_status = -ECONNABORTED;
|
||||
trace_xprtrdma_flush_dct(r_xprt, wc->status);
|
||||
xprt_force_disconnect(xprt);
|
||||
}
|
||||
@ -179,7 +179,7 @@ static void rpcrdma_wc_receive(struct ib_cq *cq, struct ib_wc *wc)
|
||||
|
||||
/* WARNING: Only wr_cqe and status are reliable at this point */
|
||||
trace_xprtrdma_wc_receive(wc);
|
||||
--r_xprt->rx_ep.re_receive_count;
|
||||
--r_xprt->rx_ep->re_receive_count;
|
||||
if (wc->status != IB_WC_SUCCESS)
|
||||
goto out_flushed;
|
||||
|
||||
@ -239,13 +239,11 @@ static int
|
||||
rpcrdma_cm_event_handler(struct rdma_cm_id *id, struct rdma_cm_event *event)
|
||||
{
|
||||
struct sockaddr *sap = (struct sockaddr *)&id->route.addr.dst_addr;
|
||||
struct rpcrdma_xprt *r_xprt = id->context;
|
||||
struct rpcrdma_ep *ep = &r_xprt->rx_ep;
|
||||
struct rpc_xprt *xprt = &r_xprt->rx_xprt;
|
||||
struct rpcrdma_ep *ep = id->context;
|
||||
struct rpc_xprt *xprt = ep->re_xprt;
|
||||
|
||||
might_sleep();
|
||||
|
||||
trace_xprtrdma_cm_event(r_xprt, event);
|
||||
switch (event->event) {
|
||||
case RDMA_CM_EVENT_ADDR_RESOLVED:
|
||||
case RDMA_CM_EVENT_ROUTE_RESOLVED:
|
||||
@ -263,16 +261,13 @@ rpcrdma_cm_event_handler(struct rdma_cm_id *id, struct rdma_cm_event *event)
|
||||
case RDMA_CM_EVENT_DEVICE_REMOVAL:
|
||||
pr_info("rpcrdma: removing device %s for %pISpc\n",
|
||||
ep->re_id->device->name, sap);
|
||||
init_completion(&ep->re_remove_done);
|
||||
/* fall through */
|
||||
case RDMA_CM_EVENT_ADDR_CHANGE:
|
||||
ep->re_connect_status = -ENODEV;
|
||||
xprt_force_disconnect(xprt);
|
||||
wait_for_completion(&ep->re_remove_done);
|
||||
trace_xprtrdma_remove(ep);
|
||||
|
||||
/* Return 1 to ensure the core destroys the id. */
|
||||
return 1;
|
||||
goto disconnected;
|
||||
case RDMA_CM_EVENT_ESTABLISHED:
|
||||
++xprt->connect_cookie;
|
||||
kref_get(&ep->re_kref);
|
||||
ep->re_connect_status = 1;
|
||||
rpcrdma_update_cm_private(ep, &event->param.conn);
|
||||
trace_xprtrdma_inline_thresh(ep);
|
||||
@ -294,9 +289,7 @@ rpcrdma_cm_event_handler(struct rdma_cm_id *id, struct rdma_cm_event *event)
|
||||
case RDMA_CM_EVENT_DISCONNECTED:
|
||||
ep->re_connect_status = -ECONNABORTED;
|
||||
disconnected:
|
||||
xprt_force_disconnect(xprt);
|
||||
wake_up_all(&ep->re_connect_wait);
|
||||
break;
|
||||
return rpcrdma_ep_destroy(ep);
|
||||
default:
|
||||
break;
|
||||
}
|
||||
@ -316,7 +309,7 @@ static struct rdma_cm_id *rpcrdma_create_id(struct rpcrdma_xprt *r_xprt,
|
||||
|
||||
init_completion(&ep->re_done);
|
||||
|
||||
id = rdma_create_id(xprt->xprt_net, rpcrdma_cm_event_handler, r_xprt,
|
||||
id = rdma_create_id(xprt->xprt_net, rpcrdma_cm_event_handler, ep,
|
||||
RDMA_PS_TCP, IB_QPT_RC);
|
||||
if (IS_ERR(id))
|
||||
return id;
|
||||
@ -352,25 +345,66 @@ out:
|
||||
return ERR_PTR(rc);
|
||||
}
|
||||
|
||||
/*
|
||||
* Exported functions.
|
||||
static void rpcrdma_ep_put(struct kref *kref)
|
||||
{
|
||||
struct rpcrdma_ep *ep = container_of(kref, struct rpcrdma_ep, re_kref);
|
||||
|
||||
if (ep->re_id->qp) {
|
||||
rdma_destroy_qp(ep->re_id);
|
||||
ep->re_id->qp = NULL;
|
||||
}
|
||||
|
||||
if (ep->re_attr.recv_cq)
|
||||
ib_free_cq(ep->re_attr.recv_cq);
|
||||
ep->re_attr.recv_cq = NULL;
|
||||
if (ep->re_attr.send_cq)
|
||||
ib_free_cq(ep->re_attr.send_cq);
|
||||
ep->re_attr.send_cq = NULL;
|
||||
|
||||
if (ep->re_pd)
|
||||
ib_dealloc_pd(ep->re_pd);
|
||||
ep->re_pd = NULL;
|
||||
|
||||
kfree(ep);
|
||||
module_put(THIS_MODULE);
|
||||
}
|
||||
|
||||
/* Returns:
|
||||
* %0 if @ep still has a positive kref count, or
|
||||
* %1 if @ep was destroyed successfully.
|
||||
*/
|
||||
static int rpcrdma_ep_destroy(struct rpcrdma_ep *ep)
|
||||
{
|
||||
return kref_put(&ep->re_kref, rpcrdma_ep_put);
|
||||
}
|
||||
|
||||
static int rpcrdma_ep_create(struct rpcrdma_xprt *r_xprt)
|
||||
{
|
||||
struct rpcrdma_ep *ep = &r_xprt->rx_ep;
|
||||
struct rpcrdma_connect_private *pmsg = &ep->re_cm_private;
|
||||
struct rpcrdma_connect_private *pmsg;
|
||||
struct ib_device *device;
|
||||
struct rdma_cm_id *id;
|
||||
struct rpcrdma_ep *ep;
|
||||
int rc;
|
||||
|
||||
ep = kzalloc(sizeof(*ep), GFP_NOFS);
|
||||
if (!ep)
|
||||
return -EAGAIN;
|
||||
ep->re_xprt = &r_xprt->rx_xprt;
|
||||
kref_init(&ep->re_kref);
|
||||
|
||||
id = rpcrdma_create_id(r_xprt, ep);
|
||||
if (IS_ERR(id))
|
||||
return PTR_ERR(id);
|
||||
if (IS_ERR(id)) {
|
||||
rc = PTR_ERR(id);
|
||||
goto out_free;
|
||||
}
|
||||
__module_get(THIS_MODULE);
|
||||
device = id->device;
|
||||
ep->re_id = id;
|
||||
|
||||
ep->re_max_requests = r_xprt->rx_xprt.max_reqs;
|
||||
ep->re_inline_send = xprt_rdma_max_inline_write;
|
||||
ep->re_inline_recv = xprt_rdma_max_inline_read;
|
||||
rc = frwr_query_device(ep, id->device);
|
||||
rc = frwr_query_device(ep, device);
|
||||
if (rc)
|
||||
goto out_destroy;
|
||||
|
||||
@ -396,7 +430,7 @@ static int rpcrdma_ep_create(struct rpcrdma_xprt *r_xprt)
|
||||
ep->re_send_count = ep->re_send_batch;
|
||||
init_waitqueue_head(&ep->re_connect_wait);
|
||||
|
||||
ep->re_attr.send_cq = ib_alloc_cq_any(id->device, r_xprt,
|
||||
ep->re_attr.send_cq = ib_alloc_cq_any(device, r_xprt,
|
||||
ep->re_attr.cap.max_send_wr,
|
||||
IB_POLL_WORKQUEUE);
|
||||
if (IS_ERR(ep->re_attr.send_cq)) {
|
||||
@ -404,7 +438,7 @@ static int rpcrdma_ep_create(struct rpcrdma_xprt *r_xprt)
|
||||
goto out_destroy;
|
||||
}
|
||||
|
||||
ep->re_attr.recv_cq = ib_alloc_cq_any(id->device, r_xprt,
|
||||
ep->re_attr.recv_cq = ib_alloc_cq_any(device, r_xprt,
|
||||
ep->re_attr.cap.max_recv_wr,
|
||||
IB_POLL_WORKQUEUE);
|
||||
if (IS_ERR(ep->re_attr.recv_cq)) {
|
||||
@ -417,6 +451,7 @@ static int rpcrdma_ep_create(struct rpcrdma_xprt *r_xprt)
|
||||
memset(&ep->re_remote_cma, 0, sizeof(ep->re_remote_cma));
|
||||
|
||||
/* Prepare RDMA-CM private message */
|
||||
pmsg = &ep->re_cm_private;
|
||||
pmsg->cp_magic = rpcrdma_cmp_magic;
|
||||
pmsg->cp_version = RPCRDMA_CMP_VERSION;
|
||||
pmsg->cp_flags |= RPCRDMA_CMP_F_SND_W_INV_OK;
|
||||
@ -428,7 +463,7 @@ static int rpcrdma_ep_create(struct rpcrdma_xprt *r_xprt)
|
||||
/* Client offers RDMA Read but does not initiate */
|
||||
ep->re_remote_cma.initiator_depth = 0;
|
||||
ep->re_remote_cma.responder_resources =
|
||||
min_t(int, U8_MAX, id->device->attrs.max_qp_rd_atom);
|
||||
min_t(int, U8_MAX, device->attrs.max_qp_rd_atom);
|
||||
|
||||
/* Limit transport retries so client can detect server
|
||||
* GID changes quickly. RPC layer handles re-establishing
|
||||
@ -443,7 +478,7 @@ static int rpcrdma_ep_create(struct rpcrdma_xprt *r_xprt)
|
||||
ep->re_remote_cma.flow_control = 0;
|
||||
ep->re_remote_cma.rnr_retry_count = 0;
|
||||
|
||||
ep->re_pd = ib_alloc_pd(id->device, 0);
|
||||
ep->re_pd = ib_alloc_pd(device, 0);
|
||||
if (IS_ERR(ep->re_pd)) {
|
||||
rc = PTR_ERR(ep->re_pd);
|
||||
goto out_destroy;
|
||||
@ -453,50 +488,36 @@ static int rpcrdma_ep_create(struct rpcrdma_xprt *r_xprt)
|
||||
if (rc)
|
||||
goto out_destroy;
|
||||
|
||||
ep->re_id = id;
|
||||
r_xprt->rx_ep = ep;
|
||||
return 0;
|
||||
|
||||
out_destroy:
|
||||
rpcrdma_ep_destroy(r_xprt);
|
||||
rpcrdma_ep_destroy(ep);
|
||||
rdma_destroy_id(id);
|
||||
out_free:
|
||||
kfree(ep);
|
||||
r_xprt->rx_ep = NULL;
|
||||
return rc;
|
||||
}
|
||||
|
||||
static void rpcrdma_ep_destroy(struct rpcrdma_xprt *r_xprt)
|
||||
{
|
||||
struct rpcrdma_ep *ep = &r_xprt->rx_ep;
|
||||
|
||||
if (ep->re_id && ep->re_id->qp) {
|
||||
rdma_destroy_qp(ep->re_id);
|
||||
ep->re_id->qp = NULL;
|
||||
}
|
||||
|
||||
if (ep->re_attr.recv_cq)
|
||||
ib_free_cq(ep->re_attr.recv_cq);
|
||||
ep->re_attr.recv_cq = NULL;
|
||||
if (ep->re_attr.send_cq)
|
||||
ib_free_cq(ep->re_attr.send_cq);
|
||||
ep->re_attr.send_cq = NULL;
|
||||
|
||||
if (ep->re_pd)
|
||||
ib_dealloc_pd(ep->re_pd);
|
||||
ep->re_pd = NULL;
|
||||
}
|
||||
|
||||
/*
|
||||
* Connect unconnected endpoint.
|
||||
/**
|
||||
* rpcrdma_xprt_connect - Connect an unconnected transport
|
||||
* @r_xprt: controlling transport instance
|
||||
*
|
||||
* Returns 0 on success or a negative errno.
|
||||
*/
|
||||
int rpcrdma_xprt_connect(struct rpcrdma_xprt *r_xprt)
|
||||
{
|
||||
struct rpc_xprt *xprt = &r_xprt->rx_xprt;
|
||||
struct rpcrdma_ep *ep = &r_xprt->rx_ep;
|
||||
struct rpcrdma_ep *ep;
|
||||
int rc;
|
||||
|
||||
retry:
|
||||
rpcrdma_xprt_disconnect(r_xprt);
|
||||
rc = rpcrdma_ep_create(r_xprt);
|
||||
if (rc)
|
||||
goto out_noupdate;
|
||||
return rc;
|
||||
ep = r_xprt->rx_ep;
|
||||
|
||||
ep->re_connect_status = 0;
|
||||
xprt_clear_connected(xprt);
|
||||
@ -533,8 +554,6 @@ retry:
|
||||
out:
|
||||
if (rc)
|
||||
ep->re_connect_status = rc;
|
||||
|
||||
out_noupdate:
|
||||
trace_xprtrdma_connect(r_xprt, rc);
|
||||
return rc;
|
||||
}
|
||||
@ -545,40 +564,33 @@ out_noupdate:
|
||||
*
|
||||
* Caller serializes. Either the transport send lock is held,
|
||||
* or we're being called to destroy the transport.
|
||||
*
|
||||
* On return, @r_xprt is completely divested of all hardware
|
||||
* resources and prepared for the next ->connect operation.
|
||||
*/
|
||||
void rpcrdma_xprt_disconnect(struct rpcrdma_xprt *r_xprt)
|
||||
{
|
||||
struct rpcrdma_ep *ep = &r_xprt->rx_ep;
|
||||
struct rdma_cm_id *id = ep->re_id;
|
||||
int rc, status = ep->re_connect_status;
|
||||
struct rpcrdma_ep *ep = r_xprt->rx_ep;
|
||||
struct rdma_cm_id *id;
|
||||
int rc;
|
||||
|
||||
might_sleep();
|
||||
|
||||
if (!id)
|
||||
if (!ep)
|
||||
return;
|
||||
|
||||
id = ep->re_id;
|
||||
rc = rdma_disconnect(id);
|
||||
if (!rc)
|
||||
wait_event_interruptible(ep->re_connect_wait,
|
||||
ep->re_connect_status != 1);
|
||||
else
|
||||
ep->re_connect_status = rc;
|
||||
trace_xprtrdma_disconnect(r_xprt, rc);
|
||||
|
||||
if (id->qp)
|
||||
rpcrdma_xprt_drain(r_xprt);
|
||||
rpcrdma_xprt_drain(r_xprt);
|
||||
rpcrdma_reps_unmap(r_xprt);
|
||||
rpcrdma_reqs_reset(r_xprt);
|
||||
rpcrdma_mrs_destroy(r_xprt);
|
||||
rpcrdma_sendctxs_destroy(r_xprt);
|
||||
|
||||
rpcrdma_ep_destroy(r_xprt);
|
||||
|
||||
if (status == -ENODEV)
|
||||
complete(&ep->re_remove_done);
|
||||
else
|
||||
if (rpcrdma_ep_destroy(ep))
|
||||
rdma_destroy_id(id);
|
||||
ep->re_id = NULL;
|
||||
|
||||
r_xprt->rx_ep = NULL;
|
||||
}
|
||||
|
||||
/* Fixed-size circular FIFO queue. This implementation is wait-free and
|
||||
@ -635,14 +647,14 @@ static int rpcrdma_sendctxs_create(struct rpcrdma_xprt *r_xprt)
|
||||
* the ->send_request call to fail temporarily before too many
|
||||
* Sends are posted.
|
||||
*/
|
||||
i = r_xprt->rx_ep.re_max_requests + RPCRDMA_MAX_BC_REQUESTS;
|
||||
i = r_xprt->rx_ep->re_max_requests + RPCRDMA_MAX_BC_REQUESTS;
|
||||
buf->rb_sc_ctxs = kcalloc(i, sizeof(sc), GFP_KERNEL);
|
||||
if (!buf->rb_sc_ctxs)
|
||||
return -ENOMEM;
|
||||
|
||||
buf->rb_sc_last = i - 1;
|
||||
for (i = 0; i <= buf->rb_sc_last; i++) {
|
||||
sc = rpcrdma_sendctx_create(&r_xprt->rx_ep);
|
||||
sc = rpcrdma_sendctx_create(r_xprt->rx_ep);
|
||||
if (!sc)
|
||||
return -ENOMEM;
|
||||
|
||||
@ -746,7 +758,7 @@ static void
|
||||
rpcrdma_mrs_create(struct rpcrdma_xprt *r_xprt)
|
||||
{
|
||||
struct rpcrdma_buffer *buf = &r_xprt->rx_buf;
|
||||
struct rpcrdma_ep *ep = &r_xprt->rx_ep;
|
||||
struct rpcrdma_ep *ep = r_xprt->rx_ep;
|
||||
unsigned int count;
|
||||
|
||||
for (count = 0; count < ep->re_max_rdma_segs; count++) {
|
||||
@ -793,7 +805,7 @@ rpcrdma_mr_refresh_worker(struct work_struct *work)
|
||||
void rpcrdma_mrs_refresh(struct rpcrdma_xprt *r_xprt)
|
||||
{
|
||||
struct rpcrdma_buffer *buf = &r_xprt->rx_buf;
|
||||
struct rpcrdma_ep *ep = &r_xprt->rx_ep;
|
||||
struct rpcrdma_ep *ep = r_xprt->rx_ep;
|
||||
|
||||
/* If there is no underlying connection, it's no use
|
||||
* to wake the refresh worker.
|
||||
@ -862,7 +874,7 @@ int rpcrdma_req_setup(struct rpcrdma_xprt *r_xprt, struct rpcrdma_req *req)
|
||||
|
||||
/* Compute maximum header buffer size in bytes */
|
||||
maxhdrsize = rpcrdma_fixed_maxsz + 3 +
|
||||
r_xprt->rx_ep.re_max_rdma_segs * rpcrdma_readchunk_maxsz;
|
||||
r_xprt->rx_ep->re_max_rdma_segs * rpcrdma_readchunk_maxsz;
|
||||
maxhdrsize *= sizeof(__be32);
|
||||
rb = rpcrdma_regbuf_alloc(__roundup_pow_of_two(maxhdrsize),
|
||||
DMA_TO_DEVICE, GFP_KERNEL);
|
||||
@ -940,7 +952,7 @@ struct rpcrdma_rep *rpcrdma_rep_create(struct rpcrdma_xprt *r_xprt,
|
||||
if (rep == NULL)
|
||||
goto out;
|
||||
|
||||
rep->rr_rdmabuf = rpcrdma_regbuf_alloc(r_xprt->rx_ep.re_inline_recv,
|
||||
rep->rr_rdmabuf = rpcrdma_regbuf_alloc(r_xprt->rx_ep->re_inline_recv,
|
||||
DMA_FROM_DEVICE, GFP_KERNEL);
|
||||
if (!rep->rr_rdmabuf)
|
||||
goto out_free;
|
||||
@ -1165,7 +1177,7 @@ void rpcrdma_mr_put(struct rpcrdma_mr *mr)
|
||||
|
||||
if (mr->mr_dir != DMA_NONE) {
|
||||
trace_xprtrdma_mr_unmap(mr);
|
||||
ib_dma_unmap_sg(r_xprt->rx_ep.re_id->device,
|
||||
ib_dma_unmap_sg(r_xprt->rx_ep->re_id->device,
|
||||
mr->mr_sg, mr->mr_nents, mr->mr_dir);
|
||||
mr->mr_dir = DMA_NONE;
|
||||
}
|
||||
@ -1283,7 +1295,7 @@ bool rpcrdma_regbuf_realloc(struct rpcrdma_regbuf *rb, size_t size, gfp_t flags)
|
||||
bool __rpcrdma_regbuf_dma_map(struct rpcrdma_xprt *r_xprt,
|
||||
struct rpcrdma_regbuf *rb)
|
||||
{
|
||||
struct ib_device *device = r_xprt->rx_ep.re_id->device;
|
||||
struct ib_device *device = r_xprt->rx_ep->re_id->device;
|
||||
|
||||
if (rb->rg_direction == DMA_NONE)
|
||||
return false;
|
||||
@ -1296,7 +1308,7 @@ bool __rpcrdma_regbuf_dma_map(struct rpcrdma_xprt *r_xprt,
|
||||
}
|
||||
|
||||
rb->rg_device = device;
|
||||
rb->rg_iov.lkey = r_xprt->rx_ep.re_pd->local_dma_lkey;
|
||||
rb->rg_iov.lkey = r_xprt->rx_ep->re_pd->local_dma_lkey;
|
||||
return true;
|
||||
}
|
||||
|
||||
@ -1332,7 +1344,7 @@ static void rpcrdma_regbuf_free(struct rpcrdma_regbuf *rb)
|
||||
int rpcrdma_post_sends(struct rpcrdma_xprt *r_xprt, struct rpcrdma_req *req)
|
||||
{
|
||||
struct ib_send_wr *send_wr = &req->rl_wr;
|
||||
struct rpcrdma_ep *ep = &r_xprt->rx_ep;
|
||||
struct rpcrdma_ep *ep = r_xprt->rx_ep;
|
||||
int rc;
|
||||
|
||||
if (!ep->re_send_count || kref_read(&req->rl_kref) > 1) {
|
||||
@ -1359,7 +1371,7 @@ int rpcrdma_post_sends(struct rpcrdma_xprt *r_xprt, struct rpcrdma_req *req)
|
||||
void rpcrdma_post_recvs(struct rpcrdma_xprt *r_xprt, bool temp)
|
||||
{
|
||||
struct rpcrdma_buffer *buf = &r_xprt->rx_buf;
|
||||
struct rpcrdma_ep *ep = &r_xprt->rx_ep;
|
||||
struct rpcrdma_ep *ep = r_xprt->rx_ep;
|
||||
struct ib_recv_wr *wr, *bad_wr;
|
||||
struct rpcrdma_rep *rep;
|
||||
int needed, count, rc;
|
||||
@ -1396,7 +1408,7 @@ void rpcrdma_post_recvs(struct rpcrdma_xprt *r_xprt, bool temp)
|
||||
if (!wr)
|
||||
goto out;
|
||||
|
||||
rc = ib_post_recv(r_xprt->rx_ep.re_id->qp, wr,
|
||||
rc = ib_post_recv(ep->re_id->qp, wr,
|
||||
(const struct ib_recv_wr **)&bad_wr);
|
||||
out:
|
||||
trace_xprtrdma_post_recvs(r_xprt, count, rc);
|
||||
|
@ -68,6 +68,7 @@
|
||||
* RDMA Endpoint -- connection endpoint details
|
||||
*/
|
||||
struct rpcrdma_ep {
|
||||
struct kref re_kref;
|
||||
struct rdma_cm_id *re_id;
|
||||
struct ib_pd *re_pd;
|
||||
unsigned int re_max_rdma_segs;
|
||||
@ -75,7 +76,6 @@ struct rpcrdma_ep {
|
||||
bool re_implicit_roundup;
|
||||
enum ib_mr_type re_mrtype;
|
||||
struct completion re_done;
|
||||
struct completion re_remove_done;
|
||||
unsigned int re_send_count;
|
||||
unsigned int re_send_batch;
|
||||
unsigned int re_max_inline_send;
|
||||
@ -83,7 +83,8 @@ struct rpcrdma_ep {
|
||||
int re_async_rc;
|
||||
int re_connect_status;
|
||||
struct ib_qp_init_attr re_attr;
|
||||
wait_queue_head_t re_connect_wait;
|
||||
wait_queue_head_t re_connect_wait;
|
||||
struct rpc_xprt *re_xprt;
|
||||
struct rpcrdma_connect_private
|
||||
re_cm_private;
|
||||
struct rdma_conn_param re_remote_cma;
|
||||
@ -411,7 +412,7 @@ struct rpcrdma_stats {
|
||||
*/
|
||||
struct rpcrdma_xprt {
|
||||
struct rpc_xprt rx_xprt;
|
||||
struct rpcrdma_ep rx_ep;
|
||||
struct rpcrdma_ep *rx_ep;
|
||||
struct rpcrdma_buffer rx_buf;
|
||||
struct delayed_work rx_connect_worker;
|
||||
struct rpc_timeout rx_timeout;
|
||||
|
Loading…
Reference in New Issue
Block a user