License cleanup: add SPDX GPL-2.0 license identifier to files with no license
Many source files in the tree are missing licensing information, which
makes it harder for compliance tools to determine the correct license.
By default all files without license information are under the default
license of the kernel, which is GPL version 2.
Update the files which contain no license information with the 'GPL-2.0'
SPDX license identifier. The SPDX identifier is a legally binding
shorthand, which can be used instead of the full boiler plate text.
This patch is based on work done by Thomas Gleixner and Kate Stewart and
Philippe Ombredanne.
How this work was done:
Patches were generated and checked against linux-4.14-rc6 for a subset of
the use cases:
- file had no licensing information it it.
- file was a */uapi/* one with no licensing information in it,
- file was a */uapi/* one with existing licensing information,
Further patches will be generated in subsequent months to fix up cases
where non-standard license headers were used, and references to license
had to be inferred by heuristics based on keywords.
The analysis to determine which SPDX License Identifier to be applied to
a file was done in a spreadsheet of side by side results from of the
output of two independent scanners (ScanCode & Windriver) producing SPDX
tag:value files created by Philippe Ombredanne. Philippe prepared the
base worksheet, and did an initial spot review of a few 1000 files.
The 4.13 kernel was the starting point of the analysis with 60,537 files
assessed. Kate Stewart did a file by file comparison of the scanner
results in the spreadsheet to determine which SPDX license identifier(s)
to be applied to the file. She confirmed any determination that was not
immediately clear with lawyers working with the Linux Foundation.
Criteria used to select files for SPDX license identifier tagging was:
- Files considered eligible had to be source code files.
- Make and config files were included as candidates if they contained >5
lines of source
- File already had some variant of a license header in it (even if <5
lines).
All documentation files were explicitly excluded.
The following heuristics were used to determine which SPDX license
identifiers to apply.
- when both scanners couldn't find any license traces, file was
considered to have no license information in it, and the top level
COPYING file license applied.
For non */uapi/* files that summary was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 11139
and resulted in the first patch in this series.
If that file was a */uapi/* path one, it was "GPL-2.0 WITH
Linux-syscall-note" otherwise it was "GPL-2.0". Results of that was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 WITH Linux-syscall-note 930
and resulted in the second patch in this series.
- if a file had some form of licensing information in it, and was one
of the */uapi/* ones, it was denoted with the Linux-syscall-note if
any GPL family license was found in the file or had no licensing in
it (per prior point). Results summary:
SPDX license identifier # files
---------------------------------------------------|------
GPL-2.0 WITH Linux-syscall-note 270
GPL-2.0+ WITH Linux-syscall-note 169
((GPL-2.0 WITH Linux-syscall-note) OR BSD-2-Clause) 21
((GPL-2.0 WITH Linux-syscall-note) OR BSD-3-Clause) 17
LGPL-2.1+ WITH Linux-syscall-note 15
GPL-1.0+ WITH Linux-syscall-note 14
((GPL-2.0+ WITH Linux-syscall-note) OR BSD-3-Clause) 5
LGPL-2.0+ WITH Linux-syscall-note 4
LGPL-2.1 WITH Linux-syscall-note 3
((GPL-2.0 WITH Linux-syscall-note) OR MIT) 3
((GPL-2.0 WITH Linux-syscall-note) AND MIT) 1
and that resulted in the third patch in this series.
- when the two scanners agreed on the detected license(s), that became
the concluded license(s).
- when there was disagreement between the two scanners (one detected a
license but the other didn't, or they both detected different
licenses) a manual inspection of the file occurred.
- In most cases a manual inspection of the information in the file
resulted in a clear resolution of the license that should apply (and
which scanner probably needed to revisit its heuristics).
- When it was not immediately clear, the license identifier was
confirmed with lawyers working with the Linux Foundation.
- If there was any question as to the appropriate license identifier,
the file was flagged for further research and to be revisited later
in time.
In total, over 70 hours of logged manual review was done on the
spreadsheet to determine the SPDX license identifiers to apply to the
source files by Kate, Philippe, Thomas and, in some cases, confirmation
by lawyers working with the Linux Foundation.
Kate also obtained a third independent scan of the 4.13 code base from
FOSSology, and compared selected files where the other two scanners
disagreed against that SPDX file, to see if there was new insights. The
Windriver scanner is based on an older version of FOSSology in part, so
they are related.
Thomas did random spot checks in about 500 files from the spreadsheets
for the uapi headers and agreed with SPDX license identifier in the
files he inspected. For the non-uapi files Thomas did random spot checks
in about 15000 files.
In initial set of patches against 4.14-rc6, 3 files were found to have
copy/paste license identifier errors, and have been fixed to reflect the
correct identifier.
Additionally Philippe spent 10 hours this week doing a detailed manual
inspection and review of the 12,461 patched files from the initial patch
version early this week with:
- a full scancode scan run, collecting the matched texts, detected
license ids and scores
- reviewing anything where there was a license detected (about 500+
files) to ensure that the applied SPDX license was correct
- reviewing anything where there was no detection but the patch license
was not GPL-2.0 WITH Linux-syscall-note to ensure that the applied
SPDX license was correct
This produced a worksheet with 20 files needing minor correction. This
worksheet was then exported into 3 different .csv files for the
different types of files to be modified.
These .csv files were then reviewed by Greg. Thomas wrote a script to
parse the csv files and add the proper SPDX tag to the file, in the
format that the file expected. This script was further refined by Greg
based on the output to detect more types of files automatically and to
distinguish between header and source .c files (which need different
comment types.) Finally Greg ran the script using the .csv files to
generate the patches.
Reviewed-by: Kate Stewart <kstewart@linuxfoundation.org>
Reviewed-by: Philippe Ombredanne <pombredanne@nexb.com>
Reviewed-by: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
2017-11-01 22:07:57 +08:00
|
|
|
// SPDX-License-Identifier: GPL-2.0
|
2015-03-31 02:34:21 +08:00
|
|
|
/*
|
2017-12-15 09:57:47 +08:00
|
|
|
* Copyright (c) 2015, 2017 Oracle. All rights reserved.
|
2015-03-31 02:34:21 +08:00
|
|
|
* Copyright (c) 2003-2007 Network Appliance, Inc. All rights reserved.
|
|
|
|
*/
|
|
|
|
|
|
|
|
/* Lightweight memory registration using Fast Registration Work
|
2017-12-15 09:57:47 +08:00
|
|
|
* Requests (FRWR).
|
2015-03-31 02:34:21 +08:00
|
|
|
*
|
2019-08-20 06:37:52 +08:00
|
|
|
* FRWR features ordered asynchronous registration and invalidation
|
|
|
|
* of arbitrarily-sized memory regions. This is the fastest and safest
|
2015-03-31 02:34:21 +08:00
|
|
|
* but most complex memory registration mode.
|
|
|
|
*/
|
|
|
|
|
2015-05-26 23:52:35 +08:00
|
|
|
/* Normal operation
|
|
|
|
*
|
2019-08-20 06:37:52 +08:00
|
|
|
* A Memory Region is prepared for RDMA Read or Write using a FAST_REG
|
2018-12-19 23:59:01 +08:00
|
|
|
* Work Request (frwr_map). When the RDMA operation is finished, this
|
2015-05-26 23:52:35 +08:00
|
|
|
* Memory Region is invalidated using a LOCAL_INV Work Request
|
2019-08-20 06:37:52 +08:00
|
|
|
* (frwr_unmap_async and frwr_unmap_sync).
|
2015-05-26 23:52:35 +08:00
|
|
|
*
|
2019-08-20 06:37:52 +08:00
|
|
|
* Typically FAST_REG Work Requests are not signaled, and neither are
|
|
|
|
* RDMA Send Work Requests (with the exception of signaling occasionally
|
|
|
|
* to prevent provider work queue overflows). This greatly reduces HCA
|
2015-05-26 23:52:35 +08:00
|
|
|
* interrupt workload.
|
|
|
|
*/
|
|
|
|
|
|
|
|
/* Transport recovery
|
|
|
|
*
|
2019-08-20 06:37:52 +08:00
|
|
|
* frwr_map and frwr_unmap_* cannot run at the same time the transport
|
|
|
|
* connect worker is running. The connect worker holds the transport
|
|
|
|
* send lock, just as ->send_request does. This prevents frwr_map and
|
|
|
|
* the connect worker from running concurrently. When a connection is
|
|
|
|
* closed, the Receive completion queue is drained before the allowing
|
|
|
|
* the connect worker to get control. This prevents frwr_unmap and the
|
|
|
|
* connect worker from running concurrently.
|
|
|
|
*
|
|
|
|
* When the underlying transport disconnects, MRs that are in flight
|
2019-10-10 01:07:48 +08:00
|
|
|
* are flushed and are likely unusable. Thus all MRs are destroyed.
|
|
|
|
* New MRs are created on demand.
|
2015-05-26 23:52:35 +08:00
|
|
|
*/
|
|
|
|
|
2018-05-08 03:27:16 +08:00
|
|
|
#include <linux/sunrpc/svc_rdma.h>
|
2016-09-15 22:57:16 +08:00
|
|
|
|
2015-03-31 02:34:21 +08:00
|
|
|
#include "xprt_rdma.h"
|
2018-05-08 03:27:05 +08:00
|
|
|
#include <trace/events/rpcrdma.h>
|
2015-03-31 02:34:21 +08:00
|
|
|
|
2021-04-20 02:03:56 +08:00
|
|
|
static void frwr_cid_init(struct rpcrdma_ep *ep,
|
|
|
|
struct rpcrdma_mr *mr)
|
|
|
|
{
|
|
|
|
struct rpc_rdma_cid *cid = &mr->mr_cid;
|
|
|
|
|
|
|
|
cid->ci_queue_id = ep->re_attr.send_cq->res.id;
|
2021-04-20 02:04:21 +08:00
|
|
|
cid->ci_completion_id = mr->mr_ibmr->res.id;
|
2021-04-20 02:03:56 +08:00
|
|
|
}
|
|
|
|
|
2021-04-20 02:03:12 +08:00
|
|
|
static void frwr_mr_unmap(struct rpcrdma_xprt *r_xprt, struct rpcrdma_mr *mr)
|
|
|
|
{
|
|
|
|
if (mr->mr_device) {
|
|
|
|
trace_xprtrdma_mr_unmap(mr);
|
|
|
|
ib_dma_unmap_sg(mr->mr_device, mr->mr_sg, mr->mr_nents,
|
|
|
|
mr->mr_dir);
|
|
|
|
mr->mr_device = NULL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-12-19 23:59:01 +08:00
|
|
|
/**
|
2021-04-20 02:03:00 +08:00
|
|
|
* frwr_mr_release - Destroy one MR
|
2020-02-22 06:00:17 +08:00
|
|
|
* @mr: MR allocated by frwr_mr_init
|
2018-12-19 23:59:01 +08:00
|
|
|
*
|
|
|
|
*/
|
2021-04-20 02:03:00 +08:00
|
|
|
void frwr_mr_release(struct rpcrdma_mr *mr)
|
2018-10-02 02:25:25 +08:00
|
|
|
{
|
|
|
|
int rc;
|
|
|
|
|
2021-04-20 02:03:12 +08:00
|
|
|
frwr_mr_unmap(mr->mr_xprt, mr);
|
|
|
|
|
2021-04-20 02:04:21 +08:00
|
|
|
rc = ib_dereg_mr(mr->mr_ibmr);
|
2018-10-02 02:25:25 +08:00
|
|
|
if (rc)
|
2018-12-20 00:00:06 +08:00
|
|
|
trace_xprtrdma_frwr_dereg(mr, rc);
|
2018-10-02 02:25:25 +08:00
|
|
|
kfree(mr->mr_sg);
|
|
|
|
kfree(mr);
|
|
|
|
}
|
|
|
|
|
2020-11-10 03:40:14 +08:00
|
|
|
static void frwr_mr_put(struct rpcrdma_mr *mr)
|
|
|
|
{
|
|
|
|
frwr_mr_unmap(mr->mr_xprt, mr);
|
|
|
|
|
|
|
|
/* The MR is returned to the req's MR free list instead
|
|
|
|
* of to the xprt's MR free list. No spinlock is needed.
|
|
|
|
*/
|
|
|
|
rpcrdma_mr_push(mr, &mr->mr_req->rl_free_mrs);
|
|
|
|
}
|
|
|
|
|
2019-06-19 22:33:04 +08:00
|
|
|
/* frwr_reset - Place MRs back on the free list
|
|
|
|
* @req: request to reset
|
|
|
|
*
|
|
|
|
* Used after a failed marshal. For FRWR, this means the MRs
|
|
|
|
* don't have to be fully released and recreated.
|
|
|
|
*
|
|
|
|
* NB: This is safe only as long as none of @req's MRs are
|
|
|
|
* involved with an ongoing asynchronous FAST_REG or LOCAL_INV
|
|
|
|
* Work Request.
|
|
|
|
*/
|
|
|
|
void frwr_reset(struct rpcrdma_req *req)
|
|
|
|
{
|
2019-08-20 06:44:04 +08:00
|
|
|
struct rpcrdma_mr *mr;
|
2019-06-19 22:33:04 +08:00
|
|
|
|
2019-08-20 06:44:04 +08:00
|
|
|
while ((mr = rpcrdma_mr_pop(&req->rl_registered)))
|
2020-11-10 03:40:14 +08:00
|
|
|
frwr_mr_put(mr);
|
2019-06-19 22:33:04 +08:00
|
|
|
}
|
|
|
|
|
2018-12-19 23:59:01 +08:00
|
|
|
/**
|
2020-02-22 06:00:17 +08:00
|
|
|
* frwr_mr_init - Initialize one MR
|
|
|
|
* @r_xprt: controlling transport instance
|
2018-12-19 23:59:01 +08:00
|
|
|
* @mr: generic MR to prepare for FRWR
|
|
|
|
*
|
|
|
|
* Returns zero if successful. Otherwise a negative errno
|
|
|
|
* is returned.
|
|
|
|
*/
|
2020-02-22 06:00:17 +08:00
|
|
|
int frwr_mr_init(struct rpcrdma_xprt *r_xprt, struct rpcrdma_mr *mr)
|
2016-06-30 01:52:29 +08:00
|
|
|
{
|
2020-02-22 06:01:05 +08:00
|
|
|
struct rpcrdma_ep *ep = r_xprt->rx_ep;
|
2020-02-22 06:00:54 +08:00
|
|
|
unsigned int depth = ep->re_max_fr_depth;
|
2018-12-20 00:00:48 +08:00
|
|
|
struct scatterlist *sg;
|
|
|
|
struct ib_mr *frmr;
|
2016-06-30 01:52:29 +08:00
|
|
|
|
2022-09-23 21:06:24 +08:00
|
|
|
sg = kcalloc_node(depth, sizeof(*sg), XPRTRDMA_GFP_FLAGS,
|
|
|
|
ibdev_to_node(ep->re_id->device));
|
|
|
|
if (!sg)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
2020-02-22 06:00:54 +08:00
|
|
|
frmr = ib_alloc_mr(ep->re_pd, ep->re_mrtype, depth);
|
2018-12-20 00:00:48 +08:00
|
|
|
if (IS_ERR(frmr))
|
2016-06-30 01:52:29 +08:00
|
|
|
goto out_mr_err;
|
|
|
|
|
2020-02-22 06:00:17 +08:00
|
|
|
mr->mr_xprt = r_xprt;
|
2021-04-20 02:04:21 +08:00
|
|
|
mr->mr_ibmr = frmr;
|
2020-11-10 03:40:19 +08:00
|
|
|
mr->mr_device = NULL;
|
xprtrdma: Fix list corruption / DMAR errors during MR recovery
The ro_release_mr methods check whether mr->mr_list is empty.
Therefore, be sure to always use list_del_init when removing an MR
linked into a list using that field. Otherwise, when recovering from
transport failures or device removal, list corruption can result, or
MRs can get mapped or unmapped an odd number of times, resulting in
IOMMU-related failures.
In general this fix is appropriate back to v4.8. However, code
changes since then make it impossible to apply this patch directly
to stable kernels. The fix would have to be applied by hand or
reworked for kernels earlier than v4.16.
Backport guidance -- there are several cases:
- When creating an MR, initialize mr_list so that using list_empty
on an as-yet-unused MR is safe.
- When an MR is being handled by the remote invalidation path,
ensure that mr_list is reinitialized when it is removed from
rl_registered.
- When an MR is being handled by rpcrdma_destroy_mrs, it is removed
from mr_all, but it may still be on an rl_registered list. In
that case, the MR needs to be removed from that list before being
released.
- Other cases are covered by using list_del_init in rpcrdma_mr_pop.
Fixes: 9d6b04097882 ('xprtrdma: Place registered MWs on a ... ')
Signed-off-by: Chuck Lever <chuck.lever@oracle.com>
Signed-off-by: Anna Schumaker <Anna.Schumaker@Netapp.com>
2018-05-01 23:37:14 +08:00
|
|
|
INIT_LIST_HEAD(&mr->mr_list);
|
2021-04-20 02:04:09 +08:00
|
|
|
init_completion(&mr->mr_linv_done);
|
2021-04-20 02:03:56 +08:00
|
|
|
frwr_cid_init(ep, mr);
|
2018-12-20 00:00:48 +08:00
|
|
|
|
|
|
|
sg_init_table(sg, depth);
|
|
|
|
mr->mr_sg = sg;
|
2016-06-30 01:52:29 +08:00
|
|
|
return 0;
|
|
|
|
|
|
|
|
out_mr_err:
|
2022-09-23 21:06:24 +08:00
|
|
|
kfree(sg);
|
2022-09-28 21:00:48 +08:00
|
|
|
trace_xprtrdma_frwr_alloc(mr, PTR_ERR(frmr));
|
2022-09-23 21:06:24 +08:00
|
|
|
return PTR_ERR(frmr);
|
2016-06-30 01:52:29 +08:00
|
|
|
}
|
|
|
|
|
2018-12-19 23:59:01 +08:00
|
|
|
/**
|
2020-01-04 00:56:48 +08:00
|
|
|
* frwr_query_device - Prepare a transport for use with FRWR
|
2020-02-22 06:00:54 +08:00
|
|
|
* @ep: endpoint to fill in
|
2020-01-04 00:56:48 +08:00
|
|
|
* @device: RDMA device to query
|
2018-12-19 23:59:01 +08:00
|
|
|
*
|
|
|
|
* On success, sets:
|
2020-02-22 06:00:54 +08:00
|
|
|
* ep->re_attr
|
|
|
|
* ep->re_max_requests
|
|
|
|
* ep->re_max_rdma_segs
|
|
|
|
* ep->re_max_fr_depth
|
|
|
|
* ep->re_mrtype
|
2018-12-19 23:59:01 +08:00
|
|
|
*
|
2020-01-04 00:56:48 +08:00
|
|
|
* Return values:
|
|
|
|
* On success, returns zero.
|
|
|
|
* %-EINVAL - the device does not support FRWR memory registration
|
|
|
|
* %-ENOMEM - the device is not sufficiently capable for NFS/RDMA
|
2018-05-05 03:34:48 +08:00
|
|
|
*/
|
2020-02-22 06:00:54 +08:00
|
|
|
int frwr_query_device(struct rpcrdma_ep *ep, const struct ib_device *device)
|
2015-03-31 02:35:26 +08:00
|
|
|
{
|
2020-01-04 00:56:48 +08:00
|
|
|
const struct ib_device_attr *attrs = &device->attrs;
|
2018-05-05 03:34:48 +08:00
|
|
|
int max_qp_wr, depth, delta;
|
2020-01-04 00:56:27 +08:00
|
|
|
unsigned int max_sge;
|
|
|
|
|
2020-01-04 00:56:48 +08:00
|
|
|
if (!(attrs->device_cap_flags & IB_DEVICE_MEM_MGT_EXTENSIONS) ||
|
|
|
|
attrs->max_fast_reg_page_list_len == 0) {
|
|
|
|
pr_err("rpcrdma: 'frwr' mode is not supported by device %s\n",
|
|
|
|
device->name);
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
2020-01-04 00:56:27 +08:00
|
|
|
max_sge = min_t(unsigned int, attrs->max_send_sge,
|
|
|
|
RPCRDMA_MAX_SEND_SGES);
|
|
|
|
if (max_sge < RPCRDMA_MIN_SEND_SGES) {
|
|
|
|
pr_err("rpcrdma: HCA provides only %u send SGEs\n", max_sge);
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
2020-02-22 06:00:54 +08:00
|
|
|
ep->re_attr.cap.max_send_sge = max_sge;
|
|
|
|
ep->re_attr.cap.max_recv_sge = 1;
|
2015-03-31 02:35:26 +08:00
|
|
|
|
2020-02-22 06:00:54 +08:00
|
|
|
ep->re_mrtype = IB_MR_TYPE_MEM_REG;
|
2022-04-04 23:26:42 +08:00
|
|
|
if (attrs->kernel_cap_flags & IBK_SG_GAPS_REG)
|
2020-02-22 06:00:54 +08:00
|
|
|
ep->re_mrtype = IB_MR_TYPE_SG_GAPS;
|
xprtrdma: Support for SG_GAP devices
Some devices (such as the Mellanox CX-4) can register, under a
single R_key, a set of memory regions that are not contiguous. When
this is done, all the segments in a Reply list, say, can then be
invalidated in a single LocalInv Work Request (or via Remote
Invalidation, which can invalidate exactly one R_key when completing
a Receive).
This means a single FastReg WR is used to register, and one or zero
LocalInv WRs can invalidate, the memory involved with RDMA transfers
on behalf of an RPC.
In addition, xprtrdma constructs some Reply chunks from three or
more segments. By registering them with SG_GAP, only one segment
is needed for the Reply chunk, allowing the whole chunk to be
invalidated remotely.
Signed-off-by: Chuck Lever <chuck.lever@oracle.com>
Signed-off-by: Anna Schumaker <Anna.Schumaker@Netapp.com>
2016-11-29 23:52:24 +08:00
|
|
|
|
2018-12-19 23:58:51 +08:00
|
|
|
/* Quirk: Some devices advertise a large max_fast_reg_page_list_len
|
|
|
|
* capability, but perform optimally when the MRs are not larger
|
|
|
|
* than a page.
|
|
|
|
*/
|
2020-01-04 00:56:43 +08:00
|
|
|
if (attrs->max_sge_rd > RPCRDMA_MAX_HDR_SEGS)
|
2020-02-22 06:00:54 +08:00
|
|
|
ep->re_max_fr_depth = attrs->max_sge_rd;
|
2018-12-19 23:58:51 +08:00
|
|
|
else
|
2020-02-22 06:00:54 +08:00
|
|
|
ep->re_max_fr_depth = attrs->max_fast_reg_page_list_len;
|
|
|
|
if (ep->re_max_fr_depth > RPCRDMA_MAX_DATA_SEGS)
|
|
|
|
ep->re_max_fr_depth = RPCRDMA_MAX_DATA_SEGS;
|
2017-12-15 09:57:47 +08:00
|
|
|
|
|
|
|
/* Add room for frwr register and invalidate WRs.
|
|
|
|
* 1. FRWR reg WR for head
|
|
|
|
* 2. FRWR invalidate WR for head
|
|
|
|
* 3. N FRWR reg WRs for pagelist
|
|
|
|
* 4. N FRWR invalidate WRs for pagelist
|
|
|
|
* 5. FRWR reg WR for tail
|
|
|
|
* 6. FRWR invalidate WR for tail
|
2015-03-31 02:35:26 +08:00
|
|
|
* 7. The RDMA_SEND WR
|
|
|
|
*/
|
|
|
|
depth = 7;
|
|
|
|
|
2017-12-15 09:57:47 +08:00
|
|
|
/* Calculate N if the device max FRWR depth is smaller than
|
2015-03-31 02:35:26 +08:00
|
|
|
* RPCRDMA_MAX_DATA_SEGS.
|
|
|
|
*/
|
2020-02-22 06:00:54 +08:00
|
|
|
if (ep->re_max_fr_depth < RPCRDMA_MAX_DATA_SEGS) {
|
|
|
|
delta = RPCRDMA_MAX_DATA_SEGS - ep->re_max_fr_depth;
|
2015-03-31 02:35:26 +08:00
|
|
|
do {
|
2017-12-15 09:57:47 +08:00
|
|
|
depth += 2; /* FRWR reg + invalidate */
|
2020-02-22 06:00:54 +08:00
|
|
|
delta -= ep->re_max_fr_depth;
|
2015-03-31 02:35:26 +08:00
|
|
|
} while (delta > 0);
|
|
|
|
}
|
|
|
|
|
2020-01-04 00:56:48 +08:00
|
|
|
max_qp_wr = attrs->max_qp_wr;
|
2018-05-05 03:34:48 +08:00
|
|
|
max_qp_wr -= RPCRDMA_BACKWARD_WRS;
|
|
|
|
max_qp_wr -= 1;
|
|
|
|
if (max_qp_wr < RPCRDMA_MIN_SLOT_TABLE)
|
|
|
|
return -ENOMEM;
|
2020-02-22 06:00:54 +08:00
|
|
|
if (ep->re_max_requests > max_qp_wr)
|
|
|
|
ep->re_max_requests = max_qp_wr;
|
|
|
|
ep->re_attr.cap.max_send_wr = ep->re_max_requests * depth;
|
|
|
|
if (ep->re_attr.cap.max_send_wr > max_qp_wr) {
|
|
|
|
ep->re_max_requests = max_qp_wr / depth;
|
|
|
|
if (!ep->re_max_requests)
|
2020-01-04 00:56:48 +08:00
|
|
|
return -ENOMEM;
|
2020-02-22 06:00:54 +08:00
|
|
|
ep->re_attr.cap.max_send_wr = ep->re_max_requests * depth;
|
2015-03-31 02:35:26 +08:00
|
|
|
}
|
2020-02-22 06:00:54 +08:00
|
|
|
ep->re_attr.cap.max_send_wr += RPCRDMA_BACKWARD_WRS;
|
|
|
|
ep->re_attr.cap.max_send_wr += 1; /* for ib_drain_sq */
|
|
|
|
ep->re_attr.cap.max_recv_wr = ep->re_max_requests;
|
|
|
|
ep->re_attr.cap.max_recv_wr += RPCRDMA_BACKWARD_WRS;
|
2021-04-20 02:02:03 +08:00
|
|
|
ep->re_attr.cap.max_recv_wr += RPCRDMA_MAX_RECV_BATCH;
|
2020-02-22 06:00:54 +08:00
|
|
|
ep->re_attr.cap.max_recv_wr += 1; /* for ib_drain_rq */
|
|
|
|
|
|
|
|
ep->re_max_rdma_segs =
|
|
|
|
DIV_ROUND_UP(RPCRDMA_MAX_DATA_SEGS, ep->re_max_fr_depth);
|
2018-12-19 23:58:45 +08:00
|
|
|
/* Reply chunks require segments for head and tail buffers */
|
2020-02-22 06:00:54 +08:00
|
|
|
ep->re_max_rdma_segs += 2;
|
|
|
|
if (ep->re_max_rdma_segs > RPCRDMA_MAX_HDR_SEGS)
|
|
|
|
ep->re_max_rdma_segs = RPCRDMA_MAX_HDR_SEGS;
|
2020-01-04 00:56:43 +08:00
|
|
|
|
|
|
|
/* Ensure the underlying device is capable of conveying the
|
|
|
|
* largest r/wsize NFS will ask for. This guarantees that
|
|
|
|
* failing over from one RDMA device to another will not
|
|
|
|
* break NFS I/O.
|
|
|
|
*/
|
2020-02-22 06:00:54 +08:00
|
|
|
if ((ep->re_max_rdma_segs * ep->re_max_fr_depth) < RPCRDMA_MAX_SEGS)
|
2020-01-04 00:56:43 +08:00
|
|
|
return -ENOMEM;
|
2015-03-31 02:34:30 +08:00
|
|
|
|
2020-01-04 00:56:43 +08:00
|
|
|
return 0;
|
2015-03-31 02:34:30 +08:00
|
|
|
}
|
|
|
|
|
2018-12-19 23:59:01 +08:00
|
|
|
/**
|
|
|
|
* frwr_map - Register a memory region
|
|
|
|
* @r_xprt: controlling transport
|
|
|
|
* @seg: memory region co-ordinates
|
|
|
|
* @nsegs: number of segments remaining
|
|
|
|
* @writing: true when RDMA Write will be used
|
2018-12-19 23:59:07 +08:00
|
|
|
* @xid: XID of RPC using the registered memory
|
2019-08-20 06:45:37 +08:00
|
|
|
* @mr: MR to fill in
|
2018-12-19 23:59:01 +08:00
|
|
|
*
|
|
|
|
* Prepare a REG_MR Work Request to register a memory region
|
2015-03-31 02:34:39 +08:00
|
|
|
* for remote access via RDMA READ or RDMA WRITE.
|
2018-12-19 23:59:01 +08:00
|
|
|
*
|
|
|
|
* Returns the next segment or a negative errno pointer.
|
2019-08-20 06:45:37 +08:00
|
|
|
* On success, @mr is filled in.
|
2015-03-31 02:34:39 +08:00
|
|
|
*/
|
2018-12-19 23:59:01 +08:00
|
|
|
struct rpcrdma_mr_seg *frwr_map(struct rpcrdma_xprt *r_xprt,
|
|
|
|
struct rpcrdma_mr_seg *seg,
|
2019-02-12 00:23:44 +08:00
|
|
|
int nsegs, bool writing, __be32 xid,
|
2019-08-20 06:45:37 +08:00
|
|
|
struct rpcrdma_mr *mr)
|
2015-03-31 02:34:39 +08:00
|
|
|
{
|
2020-02-22 06:01:05 +08:00
|
|
|
struct rpcrdma_ep *ep = r_xprt->rx_ep;
|
2015-12-17 06:22:31 +08:00
|
|
|
struct ib_reg_wr *reg_wr;
|
2020-02-13 00:12:30 +08:00
|
|
|
int i, n, dma_nents;
|
2019-08-20 06:45:37 +08:00
|
|
|
struct ib_mr *ibmr;
|
2015-03-31 02:34:39 +08:00
|
|
|
u8 key;
|
|
|
|
|
2020-02-22 06:00:54 +08:00
|
|
|
if (nsegs > ep->re_max_fr_depth)
|
|
|
|
nsegs = ep->re_max_fr_depth;
|
2015-10-14 00:11:35 +08:00
|
|
|
for (i = 0; i < nsegs;) {
|
2021-02-05 00:59:13 +08:00
|
|
|
sg_set_page(&mr->mr_sg[i], seg->mr_page,
|
|
|
|
seg->mr_len, seg->mr_offset);
|
2015-10-14 00:11:35 +08:00
|
|
|
|
2015-03-31 02:34:39 +08:00
|
|
|
++seg;
|
|
|
|
++i;
|
2020-02-22 06:00:54 +08:00
|
|
|
if (ep->re_mrtype == IB_MR_TYPE_SG_GAPS)
|
xprtrdma: Support for SG_GAP devices
Some devices (such as the Mellanox CX-4) can register, under a
single R_key, a set of memory regions that are not contiguous. When
this is done, all the segments in a Reply list, say, can then be
invalidated in a single LocalInv Work Request (or via Remote
Invalidation, which can invalidate exactly one R_key when completing
a Receive).
This means a single FastReg WR is used to register, and one or zero
LocalInv WRs can invalidate, the memory involved with RDMA transfers
on behalf of an RPC.
In addition, xprtrdma constructs some Reply chunks from three or
more segments. By registering them with SG_GAP, only one segment
is needed for the Reply chunk, allowing the whole chunk to be
invalidated remotely.
Signed-off-by: Chuck Lever <chuck.lever@oracle.com>
Signed-off-by: Anna Schumaker <Anna.Schumaker@Netapp.com>
2016-11-29 23:52:24 +08:00
|
|
|
continue;
|
2021-02-05 00:59:13 +08:00
|
|
|
if ((i < nsegs && seg->mr_offset) ||
|
2015-03-31 02:34:39 +08:00
|
|
|
offset_in_page((seg-1)->mr_offset + (seg-1)->mr_len))
|
|
|
|
break;
|
|
|
|
}
|
2017-12-15 09:57:55 +08:00
|
|
|
mr->mr_dir = rpcrdma_data_dir(writing);
|
2020-02-13 00:12:30 +08:00
|
|
|
mr->mr_nents = i;
|
2015-10-14 00:11:35 +08:00
|
|
|
|
2020-02-22 06:00:54 +08:00
|
|
|
dma_nents = ib_dma_map_sg(ep->re_id->device, mr->mr_sg, mr->mr_nents,
|
2020-02-13 00:12:30 +08:00
|
|
|
mr->mr_dir);
|
|
|
|
if (!dma_nents)
|
2016-06-30 01:52:21 +08:00
|
|
|
goto out_dmamap_err;
|
2020-11-10 03:40:19 +08:00
|
|
|
mr->mr_device = ep->re_id->device;
|
2016-06-30 01:52:21 +08:00
|
|
|
|
2021-04-20 02:04:21 +08:00
|
|
|
ibmr = mr->mr_ibmr;
|
2020-02-13 00:12:30 +08:00
|
|
|
n = ib_map_mr_sg(ibmr, mr->mr_sg, dma_nents, NULL, PAGE_SIZE);
|
|
|
|
if (n != dma_nents)
|
2016-06-30 01:52:21 +08:00
|
|
|
goto out_mapmr_err;
|
2015-10-14 00:11:35 +08:00
|
|
|
|
2018-12-19 23:59:07 +08:00
|
|
|
ibmr->iova &= 0x00000000ffffffff;
|
2019-02-12 00:23:44 +08:00
|
|
|
ibmr->iova |= ((u64)be32_to_cpu(xid)) << 32;
|
2017-12-15 09:57:55 +08:00
|
|
|
key = (u8)(ibmr->rkey & 0x000000FF);
|
|
|
|
ib_update_fast_reg_key(ibmr, ++key);
|
2015-10-14 00:11:35 +08:00
|
|
|
|
2021-04-20 02:04:15 +08:00
|
|
|
reg_wr = &mr->mr_regwr;
|
2017-12-15 09:57:55 +08:00
|
|
|
reg_wr->mr = ibmr;
|
|
|
|
reg_wr->key = ibmr->rkey;
|
2015-12-17 06:22:31 +08:00
|
|
|
reg_wr->access = writing ?
|
|
|
|
IB_ACCESS_REMOTE_WRITE | IB_ACCESS_LOCAL_WRITE :
|
|
|
|
IB_ACCESS_REMOTE_READ;
|
2015-03-31 02:34:39 +08:00
|
|
|
|
2017-12-15 09:57:55 +08:00
|
|
|
mr->mr_handle = ibmr->rkey;
|
|
|
|
mr->mr_length = ibmr->length;
|
|
|
|
mr->mr_offset = ibmr->iova;
|
2018-12-19 23:59:55 +08:00
|
|
|
trace_xprtrdma_mr_map(mr);
|
2015-10-14 00:11:35 +08:00
|
|
|
|
2017-08-15 03:38:30 +08:00
|
|
|
return seg;
|
2016-06-30 01:52:21 +08:00
|
|
|
|
|
|
|
out_dmamap_err:
|
2018-12-20 00:00:06 +08:00
|
|
|
trace_xprtrdma_frwr_sgerr(mr, i);
|
2017-08-15 03:38:30 +08:00
|
|
|
return ERR_PTR(-EIO);
|
2016-06-30 01:52:21 +08:00
|
|
|
|
|
|
|
out_mapmr_err:
|
2018-12-20 00:00:06 +08:00
|
|
|
trace_xprtrdma_frwr_maperr(mr, n);
|
2017-08-15 03:38:30 +08:00
|
|
|
return ERR_PTR(-EIO);
|
2018-03-01 04:30:59 +08:00
|
|
|
}
|
2015-03-31 02:34:39 +08:00
|
|
|
|
2019-06-19 22:32:59 +08:00
|
|
|
/**
|
|
|
|
* frwr_wc_fastreg - Invoked by RDMA provider for a flushed FastReg WC
|
2020-02-22 06:00:49 +08:00
|
|
|
* @cq: completion queue
|
|
|
|
* @wc: WCE for a completed FastReg WR
|
2019-06-19 22:32:59 +08:00
|
|
|
*
|
2021-04-20 02:03:12 +08:00
|
|
|
* Each flushed MR gets destroyed after the QP has drained.
|
2019-06-19 22:32:59 +08:00
|
|
|
*/
|
|
|
|
static void frwr_wc_fastreg(struct ib_cq *cq, struct ib_wc *wc)
|
|
|
|
{
|
|
|
|
struct ib_cqe *cqe = wc->wr_cqe;
|
2021-04-20 02:04:03 +08:00
|
|
|
struct rpcrdma_mr *mr = container_of(cqe, struct rpcrdma_mr, mr_cqe);
|
2019-06-19 22:32:59 +08:00
|
|
|
|
|
|
|
/* WARNING: Only wr_cqe and status are reliable at this point */
|
2021-04-20 02:03:56 +08:00
|
|
|
trace_xprtrdma_wc_fastreg(wc, &mr->mr_cid);
|
2020-02-22 06:00:49 +08:00
|
|
|
|
2020-06-15 21:21:02 +08:00
|
|
|
rpcrdma_flush_disconnect(cq->cq_context, wc);
|
2019-06-19 22:32:59 +08:00
|
|
|
}
|
|
|
|
|
2018-12-19 23:59:01 +08:00
|
|
|
/**
|
2020-02-22 06:00:23 +08:00
|
|
|
* frwr_send - post Send WRs containing the RPC Call message
|
|
|
|
* @r_xprt: controlling transport instance
|
|
|
|
* @req: prepared RPC Call
|
2018-03-01 04:30:59 +08:00
|
|
|
*
|
2018-12-20 00:00:27 +08:00
|
|
|
* For FRWR, chain any FastReg WRs to the Send WR. Only a
|
2018-03-01 04:30:59 +08:00
|
|
|
* single ib_post_send call is needed to register memory
|
|
|
|
* and then post the Send WR.
|
2018-12-19 23:59:01 +08:00
|
|
|
*
|
2020-02-22 06:00:23 +08:00
|
|
|
* Returns the return code from ib_post_send.
|
|
|
|
*
|
|
|
|
* Caller must hold the transport send lock to ensure that the
|
|
|
|
* pointers to the transport's rdma_cm_id and QP are stable.
|
2018-03-01 04:30:59 +08:00
|
|
|
*/
|
2020-02-22 06:00:23 +08:00
|
|
|
int frwr_send(struct rpcrdma_xprt *r_xprt, struct rpcrdma_req *req)
|
2018-03-01 04:30:59 +08:00
|
|
|
{
|
2021-04-20 02:03:25 +08:00
|
|
|
struct ib_send_wr *post_wr, *send_wr = &req->rl_wr;
|
2020-11-10 03:39:31 +08:00
|
|
|
struct rpcrdma_ep *ep = r_xprt->rx_ep;
|
2018-03-01 04:30:59 +08:00
|
|
|
struct rpcrdma_mr *mr;
|
2021-04-20 02:03:25 +08:00
|
|
|
unsigned int num_wrs;
|
2021-08-03 02:44:36 +08:00
|
|
|
int ret;
|
2018-03-01 04:30:59 +08:00
|
|
|
|
2021-04-20 02:03:25 +08:00
|
|
|
num_wrs = 1;
|
|
|
|
post_wr = send_wr;
|
2018-03-01 04:30:59 +08:00
|
|
|
list_for_each_entry(mr, &req->rl_registered, mr_list) {
|
2021-04-20 02:03:31 +08:00
|
|
|
trace_xprtrdma_mr_fastreg(mr);
|
2018-03-01 04:30:59 +08:00
|
|
|
|
2021-04-20 02:04:03 +08:00
|
|
|
mr->mr_cqe.done = frwr_wc_fastreg;
|
2021-04-20 02:04:15 +08:00
|
|
|
mr->mr_regwr.wr.next = post_wr;
|
|
|
|
mr->mr_regwr.wr.wr_cqe = &mr->mr_cqe;
|
|
|
|
mr->mr_regwr.wr.num_sge = 0;
|
|
|
|
mr->mr_regwr.wr.opcode = IB_WR_REG_MR;
|
|
|
|
mr->mr_regwr.wr.send_flags = 0;
|
|
|
|
post_wr = &mr->mr_regwr.wr;
|
2021-04-20 02:03:25 +08:00
|
|
|
++num_wrs;
|
2018-03-01 04:30:59 +08:00
|
|
|
}
|
|
|
|
|
2021-04-20 02:03:25 +08:00
|
|
|
if ((kref_read(&req->rl_kref) > 1) || num_wrs > ep->re_send_count) {
|
|
|
|
send_wr->send_flags |= IB_SEND_SIGNALED;
|
|
|
|
ep->re_send_count = min_t(unsigned int, ep->re_send_batch,
|
|
|
|
num_wrs - ep->re_send_count);
|
|
|
|
} else {
|
|
|
|
send_wr->send_flags &= ~IB_SEND_SIGNALED;
|
|
|
|
ep->re_send_count -= num_wrs;
|
|
|
|
}
|
|
|
|
|
|
|
|
trace_xprtrdma_post_send(req);
|
2021-08-03 02:44:36 +08:00
|
|
|
ret = ib_post_send(ep->re_id->qp, post_wr, NULL);
|
|
|
|
if (ret)
|
|
|
|
trace_xprtrdma_post_send_err(r_xprt, req, ret);
|
|
|
|
return ret;
|
2015-03-31 02:34:39 +08:00
|
|
|
}
|
|
|
|
|
2018-12-19 23:59:01 +08:00
|
|
|
/**
|
|
|
|
* frwr_reminv - handle a remotely invalidated mr on the @mrs list
|
|
|
|
* @rep: Received reply
|
|
|
|
* @mrs: list of MRs to check
|
|
|
|
*
|
2017-12-15 09:56:26 +08:00
|
|
|
*/
|
2018-12-19 23:59:01 +08:00
|
|
|
void frwr_reminv(struct rpcrdma_rep *rep, struct list_head *mrs)
|
2017-12-15 09:56:26 +08:00
|
|
|
{
|
2017-12-15 09:57:55 +08:00
|
|
|
struct rpcrdma_mr *mr;
|
2017-12-15 09:56:26 +08:00
|
|
|
|
2017-12-15 09:57:55 +08:00
|
|
|
list_for_each_entry(mr, mrs, mr_list)
|
|
|
|
if (mr->mr_handle == rep->rr_inv_rkey) {
|
xprtrdma: Fix list corruption / DMAR errors during MR recovery
The ro_release_mr methods check whether mr->mr_list is empty.
Therefore, be sure to always use list_del_init when removing an MR
linked into a list using that field. Otherwise, when recovering from
transport failures or device removal, list corruption can result, or
MRs can get mapped or unmapped an odd number of times, resulting in
IOMMU-related failures.
In general this fix is appropriate back to v4.8. However, code
changes since then make it impossible to apply this patch directly
to stable kernels. The fix would have to be applied by hand or
reworked for kernels earlier than v4.16.
Backport guidance -- there are several cases:
- When creating an MR, initialize mr_list so that using list_empty
on an as-yet-unused MR is safe.
- When an MR is being handled by the remote invalidation path,
ensure that mr_list is reinitialized when it is removed from
rl_registered.
- When an MR is being handled by rpcrdma_destroy_mrs, it is removed
from mr_all, but it may still be on an rl_registered list. In
that case, the MR needs to be removed from that list before being
released.
- Other cases are covered by using list_del_init in rpcrdma_mr_pop.
Fixes: 9d6b04097882 ('xprtrdma: Place registered MWs on a ... ')
Signed-off-by: Chuck Lever <chuck.lever@oracle.com>
Signed-off-by: Anna Schumaker <Anna.Schumaker@Netapp.com>
2018-05-01 23:37:14 +08:00
|
|
|
list_del_init(&mr->mr_list);
|
2021-04-20 02:03:31 +08:00
|
|
|
trace_xprtrdma_mr_reminv(mr);
|
2020-11-10 03:40:14 +08:00
|
|
|
frwr_mr_put(mr);
|
2017-12-15 09:56:26 +08:00
|
|
|
break; /* only one invalidated MR per RPC */
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-11-10 03:40:14 +08:00
|
|
|
static void frwr_mr_done(struct ib_wc *wc, struct rpcrdma_mr *mr)
|
2019-06-19 22:32:59 +08:00
|
|
|
{
|
2021-04-20 02:03:12 +08:00
|
|
|
if (likely(wc->status == IB_WC_SUCCESS))
|
2020-11-10 03:40:14 +08:00
|
|
|
frwr_mr_put(mr);
|
2019-06-19 22:32:59 +08:00
|
|
|
}
|
|
|
|
|
2018-12-19 23:59:01 +08:00
|
|
|
/**
|
2019-06-19 22:32:59 +08:00
|
|
|
* frwr_wc_localinv - Invoked by RDMA provider for a LOCAL_INV WC
|
2020-02-22 06:00:49 +08:00
|
|
|
* @cq: completion queue
|
|
|
|
* @wc: WCE for a completed LocalInv WR
|
2015-12-17 06:22:47 +08:00
|
|
|
*
|
2019-06-19 22:32:59 +08:00
|
|
|
*/
|
|
|
|
static void frwr_wc_localinv(struct ib_cq *cq, struct ib_wc *wc)
|
|
|
|
{
|
|
|
|
struct ib_cqe *cqe = wc->wr_cqe;
|
2021-04-20 02:04:03 +08:00
|
|
|
struct rpcrdma_mr *mr = container_of(cqe, struct rpcrdma_mr, mr_cqe);
|
2019-06-19 22:32:59 +08:00
|
|
|
|
|
|
|
/* WARNING: Only wr_cqe and status are reliable at this point */
|
2021-04-20 02:03:56 +08:00
|
|
|
trace_xprtrdma_wc_li(wc, &mr->mr_cid);
|
2020-11-10 03:40:14 +08:00
|
|
|
frwr_mr_done(wc, mr);
|
2020-02-22 06:00:49 +08:00
|
|
|
|
2020-06-15 21:21:02 +08:00
|
|
|
rpcrdma_flush_disconnect(cq->cq_context, wc);
|
2019-06-19 22:32:59 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* frwr_wc_localinv_wake - Invoked by RDMA provider for a LOCAL_INV WC
|
2020-02-22 06:00:49 +08:00
|
|
|
* @cq: completion queue
|
|
|
|
* @wc: WCE for a completed LocalInv WR
|
2016-06-30 01:54:16 +08:00
|
|
|
*
|
2019-06-19 22:32:59 +08:00
|
|
|
* Awaken anyone waiting for an MR to finish being fenced.
|
2015-12-17 06:22:47 +08:00
|
|
|
*/
|
2019-06-19 22:32:59 +08:00
|
|
|
static void frwr_wc_localinv_wake(struct ib_cq *cq, struct ib_wc *wc)
|
|
|
|
{
|
|
|
|
struct ib_cqe *cqe = wc->wr_cqe;
|
2021-04-20 02:04:03 +08:00
|
|
|
struct rpcrdma_mr *mr = container_of(cqe, struct rpcrdma_mr, mr_cqe);
|
2019-06-19 22:32:59 +08:00
|
|
|
|
|
|
|
/* WARNING: Only wr_cqe and status are reliable at this point */
|
2021-04-20 02:03:56 +08:00
|
|
|
trace_xprtrdma_wc_li_wake(wc, &mr->mr_cid);
|
2020-11-10 03:40:14 +08:00
|
|
|
frwr_mr_done(wc, mr);
|
2021-04-20 02:04:09 +08:00
|
|
|
complete(&mr->mr_linv_done);
|
2020-02-22 06:00:49 +08:00
|
|
|
|
2020-06-15 21:21:02 +08:00
|
|
|
rpcrdma_flush_disconnect(cq->cq_context, wc);
|
2019-06-19 22:32:59 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* frwr_unmap_sync - invalidate memory regions that were registered for @req
|
|
|
|
* @r_xprt: controlling transport instance
|
|
|
|
* @req: rpcrdma_req with a non-empty list of MRs to process
|
|
|
|
*
|
|
|
|
* Sleeps until it is safe for the host CPU to access the previously mapped
|
2019-06-19 22:33:10 +08:00
|
|
|
* memory regions. This guarantees that registered MRs are properly fenced
|
|
|
|
* from the server before the RPC consumer accesses the data in them. It
|
|
|
|
* also ensures proper Send flow control: waking the next RPC waits until
|
|
|
|
* this RPC has relinquished all its Send Queue entries.
|
2019-06-19 22:32:59 +08:00
|
|
|
*/
|
|
|
|
void frwr_unmap_sync(struct rpcrdma_xprt *r_xprt, struct rpcrdma_req *req)
|
2015-12-17 06:22:47 +08:00
|
|
|
{
|
2018-07-19 00:25:32 +08:00
|
|
|
struct ib_send_wr *first, **prev, *last;
|
2020-11-10 03:39:31 +08:00
|
|
|
struct rpcrdma_ep *ep = r_xprt->rx_ep;
|
2018-07-19 00:25:32 +08:00
|
|
|
const struct ib_send_wr *bad_wr;
|
2017-12-15 09:57:55 +08:00
|
|
|
struct rpcrdma_mr *mr;
|
2019-06-19 22:32:59 +08:00
|
|
|
int rc;
|
2015-12-17 06:22:47 +08:00
|
|
|
|
2017-06-08 23:52:04 +08:00
|
|
|
/* ORDER: Invalidate all of the MRs first
|
2015-12-17 06:22:47 +08:00
|
|
|
*
|
|
|
|
* Chain the LOCAL_INV Work Requests and post them with
|
|
|
|
* a single ib_post_send() call.
|
|
|
|
*/
|
2016-11-29 23:52:57 +08:00
|
|
|
prev = &first;
|
2021-11-03 02:48:59 +08:00
|
|
|
mr = rpcrdma_mr_pop(&req->rl_registered);
|
|
|
|
do {
|
2018-10-02 02:25:30 +08:00
|
|
|
trace_xprtrdma_mr_localinv(mr);
|
2019-06-19 22:32:59 +08:00
|
|
|
r_xprt->rx_stats.local_inv_needed++;
|
2016-11-29 23:52:57 +08:00
|
|
|
|
2021-04-20 02:04:15 +08:00
|
|
|
last = &mr->mr_invwr;
|
2019-06-19 22:32:59 +08:00
|
|
|
last->next = NULL;
|
2021-04-20 02:04:03 +08:00
|
|
|
last->wr_cqe = &mr->mr_cqe;
|
2019-06-19 22:32:59 +08:00
|
|
|
last->sg_list = NULL;
|
|
|
|
last->num_sge = 0;
|
2016-11-29 23:52:57 +08:00
|
|
|
last->opcode = IB_WR_LOCAL_INV;
|
2019-06-19 22:32:59 +08:00
|
|
|
last->send_flags = IB_SEND_SIGNALED;
|
2017-12-15 09:57:55 +08:00
|
|
|
last->ex.invalidate_rkey = mr->mr_handle;
|
2015-12-17 06:22:47 +08:00
|
|
|
|
2021-04-20 02:04:03 +08:00
|
|
|
last->wr_cqe->done = frwr_wc_localinv;
|
|
|
|
|
2016-11-29 23:52:57 +08:00
|
|
|
*prev = last;
|
|
|
|
prev = &last->next;
|
2021-11-03 02:48:59 +08:00
|
|
|
} while ((mr = rpcrdma_mr_pop(&req->rl_registered)));
|
|
|
|
|
2021-05-02 03:38:02 +08:00
|
|
|
mr = container_of(last, struct rpcrdma_mr, mr_invwr);
|
2015-12-17 06:22:47 +08:00
|
|
|
|
|
|
|
/* Strong send queue ordering guarantees that when the
|
|
|
|
* last WR in the chain completes, all WRs in the chain
|
|
|
|
* are complete.
|
|
|
|
*/
|
2021-04-20 02:04:03 +08:00
|
|
|
last->wr_cqe->done = frwr_wc_localinv_wake;
|
2021-04-20 02:04:09 +08:00
|
|
|
reinit_completion(&mr->mr_linv_done);
|
2016-11-29 23:52:16 +08:00
|
|
|
|
2015-12-17 06:22:47 +08:00
|
|
|
/* Transport disconnect drains the receive CQ before it
|
|
|
|
* replaces the QP. The RPC reply handler won't call us
|
2020-02-22 06:00:54 +08:00
|
|
|
* unless re_id->qp is a valid pointer.
|
2015-12-17 06:22:47 +08:00
|
|
|
*/
|
2017-06-08 23:52:28 +08:00
|
|
|
bad_wr = NULL;
|
2020-11-10 03:39:31 +08:00
|
|
|
rc = ib_post_send(ep->re_id->qp, first, &bad_wr);
|
2015-12-17 06:22:47 +08:00
|
|
|
|
2019-06-19 22:32:59 +08:00
|
|
|
/* The final LOCAL_INV WR in the chain is supposed to
|
|
|
|
* do the wake. If it was never posted, the wake will
|
|
|
|
* not happen, so don't wait in that case.
|
2015-12-17 06:22:47 +08:00
|
|
|
*/
|
2019-06-19 22:32:59 +08:00
|
|
|
if (bad_wr != first)
|
2021-04-20 02:04:09 +08:00
|
|
|
wait_for_completion(&mr->mr_linv_done);
|
2019-06-19 22:32:59 +08:00
|
|
|
if (!rc)
|
|
|
|
return;
|
2015-03-31 02:34:48 +08:00
|
|
|
|
2021-04-20 02:03:12 +08:00
|
|
|
/* On error, the MRs get destroyed once the QP has drained. */
|
2020-11-10 03:39:37 +08:00
|
|
|
trace_xprtrdma_post_linv_err(req, rc);
|
2021-08-03 02:44:17 +08:00
|
|
|
|
|
|
|
/* Force a connection loss to ensure complete recovery.
|
|
|
|
*/
|
|
|
|
rpcrdma_force_disconnect(ep);
|
2015-12-17 06:22:47 +08:00
|
|
|
}
|
2019-06-19 22:33:10 +08:00
|
|
|
|
|
|
|
/**
|
|
|
|
* frwr_wc_localinv_done - Invoked by RDMA provider for a signaled LOCAL_INV WC
|
2020-02-22 06:00:49 +08:00
|
|
|
* @cq: completion queue
|
|
|
|
* @wc: WCE for a completed LocalInv WR
|
2019-06-19 22:33:10 +08:00
|
|
|
*
|
|
|
|
*/
|
|
|
|
static void frwr_wc_localinv_done(struct ib_cq *cq, struct ib_wc *wc)
|
|
|
|
{
|
|
|
|
struct ib_cqe *cqe = wc->wr_cqe;
|
2021-04-20 02:04:03 +08:00
|
|
|
struct rpcrdma_mr *mr = container_of(cqe, struct rpcrdma_mr, mr_cqe);
|
2021-04-20 02:03:06 +08:00
|
|
|
struct rpcrdma_rep *rep;
|
2019-06-19 22:33:10 +08:00
|
|
|
|
|
|
|
/* WARNING: Only wr_cqe and status are reliable at this point */
|
2021-04-20 02:03:56 +08:00
|
|
|
trace_xprtrdma_wc_li_done(wc, &mr->mr_cid);
|
2019-08-20 06:47:10 +08:00
|
|
|
|
2021-04-20 02:03:06 +08:00
|
|
|
/* Ensure that @rep is generated before the MR is released */
|
|
|
|
rep = mr->mr_req->rl_reply;
|
2019-08-20 06:47:10 +08:00
|
|
|
smp_rmb();
|
2021-04-20 02:03:06 +08:00
|
|
|
|
2021-04-20 02:03:19 +08:00
|
|
|
if (wc->status != IB_WC_SUCCESS) {
|
|
|
|
if (rep)
|
|
|
|
rpcrdma_unpin_rqst(rep);
|
|
|
|
rpcrdma_flush_disconnect(cq->cq_context, wc);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
frwr_mr_put(mr);
|
2019-08-20 06:47:10 +08:00
|
|
|
rpcrdma_complete_rqst(rep);
|
2019-06-19 22:33:10 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* frwr_unmap_async - invalidate memory regions that were registered for @req
|
|
|
|
* @r_xprt: controlling transport instance
|
|
|
|
* @req: rpcrdma_req with a non-empty list of MRs to process
|
|
|
|
*
|
|
|
|
* This guarantees that registered MRs are properly fenced from the
|
|
|
|
* server before the RPC consumer accesses the data in them. It also
|
|
|
|
* ensures proper Send flow control: waking the next RPC waits until
|
|
|
|
* this RPC has relinquished all its Send Queue entries.
|
|
|
|
*/
|
|
|
|
void frwr_unmap_async(struct rpcrdma_xprt *r_xprt, struct rpcrdma_req *req)
|
|
|
|
{
|
|
|
|
struct ib_send_wr *first, *last, **prev;
|
2020-11-10 03:39:31 +08:00
|
|
|
struct rpcrdma_ep *ep = r_xprt->rx_ep;
|
2019-06-19 22:33:10 +08:00
|
|
|
struct rpcrdma_mr *mr;
|
|
|
|
int rc;
|
|
|
|
|
|
|
|
/* Chain the LOCAL_INV Work Requests and post them with
|
|
|
|
* a single ib_post_send() call.
|
|
|
|
*/
|
|
|
|
prev = &first;
|
2021-11-03 02:48:59 +08:00
|
|
|
mr = rpcrdma_mr_pop(&req->rl_registered);
|
|
|
|
do {
|
2019-06-19 22:33:10 +08:00
|
|
|
trace_xprtrdma_mr_localinv(mr);
|
|
|
|
r_xprt->rx_stats.local_inv_needed++;
|
|
|
|
|
2021-04-20 02:04:15 +08:00
|
|
|
last = &mr->mr_invwr;
|
2019-06-19 22:33:10 +08:00
|
|
|
last->next = NULL;
|
2021-04-20 02:04:03 +08:00
|
|
|
last->wr_cqe = &mr->mr_cqe;
|
2019-06-19 22:33:10 +08:00
|
|
|
last->sg_list = NULL;
|
|
|
|
last->num_sge = 0;
|
|
|
|
last->opcode = IB_WR_LOCAL_INV;
|
|
|
|
last->send_flags = IB_SEND_SIGNALED;
|
|
|
|
last->ex.invalidate_rkey = mr->mr_handle;
|
|
|
|
|
2021-04-20 02:04:03 +08:00
|
|
|
last->wr_cqe->done = frwr_wc_localinv;
|
|
|
|
|
2019-06-19 22:33:10 +08:00
|
|
|
*prev = last;
|
|
|
|
prev = &last->next;
|
2021-11-03 02:48:59 +08:00
|
|
|
} while ((mr = rpcrdma_mr_pop(&req->rl_registered)));
|
2019-06-19 22:33:10 +08:00
|
|
|
|
|
|
|
/* Strong send queue ordering guarantees that when the
|
|
|
|
* last WR in the chain completes, all WRs in the chain
|
|
|
|
* are complete. The last completion will wake up the
|
|
|
|
* RPC waiter.
|
|
|
|
*/
|
2021-04-20 02:04:03 +08:00
|
|
|
last->wr_cqe->done = frwr_wc_localinv_done;
|
2019-06-19 22:33:10 +08:00
|
|
|
|
|
|
|
/* Transport disconnect drains the receive CQ before it
|
|
|
|
* replaces the QP. The RPC reply handler won't call us
|
2020-02-22 06:00:54 +08:00
|
|
|
* unless re_id->qp is a valid pointer.
|
2019-06-19 22:33:10 +08:00
|
|
|
*/
|
2021-04-20 02:03:12 +08:00
|
|
|
rc = ib_post_send(ep->re_id->qp, first, NULL);
|
2019-06-19 22:33:10 +08:00
|
|
|
if (!rc)
|
|
|
|
return;
|
|
|
|
|
2021-04-20 02:03:12 +08:00
|
|
|
/* On error, the MRs get destroyed once the QP has drained. */
|
2020-11-10 03:39:37 +08:00
|
|
|
trace_xprtrdma_post_linv_err(req, rc);
|
2019-06-19 22:33:10 +08:00
|
|
|
|
|
|
|
/* The final LOCAL_INV WR in the chain is supposed to
|
2021-04-20 02:03:19 +08:00
|
|
|
* do the wake. If it was never posted, the wake does
|
|
|
|
* not happen. Unpin the rqst in preparation for its
|
|
|
|
* retransmission.
|
2019-06-19 22:33:10 +08:00
|
|
|
*/
|
2021-04-20 02:03:19 +08:00
|
|
|
rpcrdma_unpin_rqst(req->rl_reply);
|
2021-08-03 02:44:17 +08:00
|
|
|
|
|
|
|
/* Force a connection loss to ensure complete recovery.
|
|
|
|
*/
|
|
|
|
rpcrdma_force_disconnect(ep);
|
2019-06-19 22:33:10 +08:00
|
|
|
}
|
2021-10-05 22:17:59 +08:00
|
|
|
|
|
|
|
/**
|
|
|
|
* frwr_wp_create - Create an MR for padding Write chunks
|
|
|
|
* @r_xprt: transport resources to use
|
|
|
|
*
|
|
|
|
* Return 0 on success, negative errno on failure.
|
|
|
|
*/
|
|
|
|
int frwr_wp_create(struct rpcrdma_xprt *r_xprt)
|
|
|
|
{
|
|
|
|
struct rpcrdma_ep *ep = r_xprt->rx_ep;
|
|
|
|
struct rpcrdma_mr_seg seg;
|
|
|
|
struct rpcrdma_mr *mr;
|
|
|
|
|
|
|
|
mr = rpcrdma_mr_get(r_xprt);
|
|
|
|
if (!mr)
|
|
|
|
return -EAGAIN;
|
|
|
|
mr->mr_req = NULL;
|
|
|
|
ep->re_write_pad_mr = mr;
|
|
|
|
|
|
|
|
seg.mr_len = XDR_UNIT;
|
|
|
|
seg.mr_page = virt_to_page(ep->re_write_pad);
|
|
|
|
seg.mr_offset = offset_in_page(ep->re_write_pad);
|
|
|
|
if (IS_ERR(frwr_map(r_xprt, &seg, 1, true, xdr_zero, mr)))
|
|
|
|
return -EIO;
|
|
|
|
trace_xprtrdma_mr_fastreg(mr);
|
|
|
|
|
|
|
|
mr->mr_cqe.done = frwr_wc_fastreg;
|
|
|
|
mr->mr_regwr.wr.next = NULL;
|
|
|
|
mr->mr_regwr.wr.wr_cqe = &mr->mr_cqe;
|
|
|
|
mr->mr_regwr.wr.num_sge = 0;
|
|
|
|
mr->mr_regwr.wr.opcode = IB_WR_REG_MR;
|
|
|
|
mr->mr_regwr.wr.send_flags = 0;
|
|
|
|
|
|
|
|
return ib_post_send(ep->re_id->qp, &mr->mr_regwr.wr, NULL);
|
|
|
|
}
|