2022-05-25 01:56:42 +08:00
|
|
|
// SPDX-License-Identifier: GPL-2.0
|
|
|
|
#include <linux/kernel.h>
|
|
|
|
#include <linux/errno.h>
|
|
|
|
#include <linux/fs.h>
|
|
|
|
#include <linux/file.h>
|
|
|
|
#include <linux/io_uring.h>
|
|
|
|
|
|
|
|
#include <uapi/linux/io_uring.h>
|
|
|
|
|
|
|
|
#include "io_uring.h"
|
2024-10-27 22:59:10 +08:00
|
|
|
#include "rsrc.h"
|
2022-05-25 01:56:42 +08:00
|
|
|
#include "nop.h"
|
|
|
|
|
2024-05-10 11:50:28 +08:00
|
|
|
struct io_nop {
|
|
|
|
/* NOTE: kiocb has the file as the first member, so don't do it here */
|
|
|
|
struct file *file;
|
|
|
|
int result;
|
2024-10-27 22:59:10 +08:00
|
|
|
int fd;
|
|
|
|
int buffer;
|
|
|
|
unsigned int flags;
|
2024-05-10 11:50:28 +08:00
|
|
|
};
|
|
|
|
|
2024-10-27 22:59:10 +08:00
|
|
|
#define NOP_FLAGS (IORING_NOP_INJECT_RESULT | IORING_NOP_FIXED_FILE | \
|
|
|
|
IORING_NOP_FIXED_BUFFER | IORING_NOP_FILE)
|
|
|
|
|
2022-05-25 01:56:42 +08:00
|
|
|
int io_nop_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe)
|
|
|
|
{
|
2024-05-10 11:50:28 +08:00
|
|
|
struct io_nop *nop = io_kiocb_to_cmd(req, struct io_nop);
|
|
|
|
|
2024-10-27 22:59:10 +08:00
|
|
|
nop->flags = READ_ONCE(sqe->nop_flags);
|
|
|
|
if (nop->flags & ~NOP_FLAGS)
|
2024-05-10 11:50:27 +08:00
|
|
|
return -EINVAL;
|
2024-05-10 11:50:28 +08:00
|
|
|
|
2024-10-27 22:59:10 +08:00
|
|
|
if (nop->flags & IORING_NOP_INJECT_RESULT)
|
2024-05-10 11:50:28 +08:00
|
|
|
nop->result = READ_ONCE(sqe->len);
|
|
|
|
else
|
|
|
|
nop->result = 0;
|
2024-10-27 22:59:10 +08:00
|
|
|
if (nop->flags & IORING_NOP_FIXED_FILE)
|
|
|
|
nop->fd = READ_ONCE(sqe->fd);
|
|
|
|
if (nop->flags & IORING_NOP_FIXED_BUFFER)
|
|
|
|
nop->buffer = READ_ONCE(sqe->buf_index);
|
2022-05-25 01:56:42 +08:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
int io_nop(struct io_kiocb *req, unsigned int issue_flags)
|
|
|
|
{
|
2024-05-10 11:50:28 +08:00
|
|
|
struct io_nop *nop = io_kiocb_to_cmd(req, struct io_nop);
|
2024-10-27 22:59:10 +08:00
|
|
|
int ret = nop->result;
|
|
|
|
|
|
|
|
if (nop->flags & IORING_NOP_FILE) {
|
|
|
|
if (nop->flags & IORING_NOP_FIXED_FILE) {
|
|
|
|
req->file = io_file_get_fixed(req, nop->fd, issue_flags);
|
|
|
|
req->flags |= REQ_F_FIXED_FILE;
|
|
|
|
} else {
|
|
|
|
req->file = io_file_get_normal(req, nop->fd);
|
|
|
|
}
|
|
|
|
if (!req->file) {
|
|
|
|
ret = -EBADF;
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (nop->flags & IORING_NOP_FIXED_BUFFER) {
|
|
|
|
struct io_ring_ctx *ctx = req->ctx;
|
io_uring/rsrc: get rid of per-ring io_rsrc_node list
Work in progress, but get rid of the per-ring serialization of resource
nodes, like registered buffers and files. Main issue here is that one
node can otherwise hold up a bunch of other nodes from getting freed,
which is especially a problem for file resource nodes and networked
workloads where some descriptors may not see activity in a long time.
As an example, instantiate an io_uring ring fd and create a sparse
registered file table. Even 2 will do. Then create a socket and register
it as fixed file 0, F0. The number of open files in the app is now 5,
with 0/1/2 being the usual stdin/out/err, 3 being the ring fd, and 4
being the socket. Register this socket (eg "the listener") in slot 0 of
the registered file table. Now add an operation on the socket that uses
slot 0. Finally, loop N times, where each loop creates a new socket,
registers said socket as a file, then unregisters the socket, and
finally closes the socket. This is roughly similar to what a basic
accept loop would look like.
At the end of this loop, it's not unreasonable to expect that there
would still be 5 open files. Each socket created and registered in the
loop is also unregistered and closed. But since the listener socket
registered first still has references to its resource node due to still
being active, each subsequent socket unregistration is stuck behind it
for reclaim. Hence 5 + N files are still open at that point, where N is
awaiting the final put held up by the listener socket.
Rewrite the io_rsrc_node handling to NOT rely on serialization. Struct
io_kiocb now gets explicit resource nodes assigned, with each holding a
reference to the parent node. A parent node is either of type FILE or
BUFFER, which are the two types of nodes that exist. A request can have
two nodes assigned, if it's using both registered files and buffers.
Since request issue and task_work completion is both under the ring
private lock, no atomics are needed to handle these references. It's a
simple unlocked inc/dec. As before, the registered buffer or file table
each hold a reference as well to the registered nodes. Final put of the
node will remove the node and free the underlying resource, eg unmap the
buffer or put the file.
Outside of removing the stall in resource reclaim described above, it
has the following advantages:
1) It's a lot simpler than the previous scheme, and easier to follow.
No need to specific quiesce handling anymore.
2) There are no resource node allocations in the fast path, all of that
happens at resource registration time.
3) The structs related to resource handling can all get simplified
quite a bit, like io_rsrc_node and io_rsrc_data. io_rsrc_put can
go away completely.
4) Handling of resource tags is much simpler, and doesn't require
persistent storage as it can simply get assigned up front at
registration time. Just copy them in one-by-one at registration time
and assign to the resource node.
The only real downside is that a request is now explicitly limited to
pinning 2 resources, one file and one buffer, where before just
assigning a resource node to a request would pin all of them. The upside
is that it's easier to follow now, as an individual resource is
explicitly referenced and assigned to the request.
With this in place, the above mentioned example will be using exactly 5
files at the end of the loop, not N.
Signed-off-by: Jens Axboe <axboe@kernel.dk>
2024-10-26 09:27:39 +08:00
|
|
|
struct io_rsrc_node *node;
|
2024-05-10 11:50:28 +08:00
|
|
|
|
2024-10-27 22:59:10 +08:00
|
|
|
ret = -EFAULT;
|
|
|
|
io_ring_submit_lock(ctx, issue_flags);
|
2024-10-27 23:08:31 +08:00
|
|
|
node = io_rsrc_node_lookup(&ctx->buf_table, nop->buffer);
|
|
|
|
if (node) {
|
2024-11-07 19:01:36 +08:00
|
|
|
io_req_assign_buf_node(req, node);
|
2024-10-27 22:59:10 +08:00
|
|
|
ret = 0;
|
|
|
|
}
|
|
|
|
io_ring_submit_unlock(ctx, issue_flags);
|
|
|
|
}
|
|
|
|
done:
|
|
|
|
if (ret < 0)
|
2024-05-10 11:50:28 +08:00
|
|
|
req_set_fail(req);
|
|
|
|
io_req_set_res(req, nop->result, 0);
|
2022-05-25 01:56:42 +08:00
|
|
|
return IOU_OK;
|
|
|
|
}
|