mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-14 15:54:15 +08:00
06a5464be8
Add a new io_uring opcode IORING_OP_SENDZC. The main distinction from IORING_OP_SEND is that the user should specify a notification slot index in sqe::notification_idx and the buffers are safe to reuse only when the used notification is flushed and completes. Signed-off-by: Pavel Begunkov <asml.silence@gmail.com> Link: https://lore.kernel.org/r/a80387c6a68ce9cf99b3b6ef6f71068468761fb7.1657643355.git.asml.silence@gmail.com Signed-off-by: Jens Axboe <axboe@kernel.dk>
64 lines
2.0 KiB
C
64 lines
2.0 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
|
|
#include <linux/net.h>
|
|
#include <linux/uio.h>
|
|
|
|
#include "alloc_cache.h"
|
|
|
|
#if defined(CONFIG_NET)
|
|
struct io_async_msghdr {
|
|
union {
|
|
struct iovec fast_iov[UIO_FASTIOV];
|
|
struct {
|
|
struct iovec fast_iov_one;
|
|
__kernel_size_t controllen;
|
|
int namelen;
|
|
__kernel_size_t payloadlen;
|
|
};
|
|
struct io_cache_entry cache;
|
|
};
|
|
/* points to an allocated iov, if NULL we use fast_iov instead */
|
|
struct iovec *free_iov;
|
|
struct sockaddr __user *uaddr;
|
|
struct msghdr msg;
|
|
struct sockaddr_storage addr;
|
|
};
|
|
|
|
struct io_async_connect {
|
|
struct sockaddr_storage address;
|
|
};
|
|
|
|
int io_shutdown_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
|
int io_shutdown(struct io_kiocb *req, unsigned int issue_flags);
|
|
|
|
int io_sendmsg_prep_async(struct io_kiocb *req);
|
|
void io_sendmsg_recvmsg_cleanup(struct io_kiocb *req);
|
|
int io_sendmsg_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
|
int io_sendmsg(struct io_kiocb *req, unsigned int issue_flags);
|
|
int io_send(struct io_kiocb *req, unsigned int issue_flags);
|
|
|
|
int io_recvmsg_prep_async(struct io_kiocb *req);
|
|
int io_recvmsg_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
|
int io_recvmsg(struct io_kiocb *req, unsigned int issue_flags);
|
|
int io_recv(struct io_kiocb *req, unsigned int issue_flags);
|
|
|
|
int io_accept_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
|
int io_accept(struct io_kiocb *req, unsigned int issue_flags);
|
|
|
|
int io_socket_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
|
int io_socket(struct io_kiocb *req, unsigned int issue_flags);
|
|
|
|
int io_connect_prep_async(struct io_kiocb *req);
|
|
int io_connect_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
|
int io_connect(struct io_kiocb *req, unsigned int issue_flags);
|
|
|
|
int io_sendzc(struct io_kiocb *req, unsigned int issue_flags);
|
|
int io_sendzc_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
|
|
|
void io_netmsg_cache_free(struct io_cache_entry *entry);
|
|
#else
|
|
static inline void io_netmsg_cache_free(struct io_cache_entry *entry)
|
|
{
|
|
}
|
|
#endif
|