mirror of
https://github.com/qemu/qemu.git
synced 2024-11-23 19:03:38 +08:00
aio-posix: Fix concurrent aio_poll/set_fd_handler.
It is possible for an io_poll callback to be concurrently executed along with an aio_set_fd_handlers. This can cause all sorts of problems, like a NULL callback or a bad opaque pointer. This changes set_fd_handlers so that it no longer modify existing handlers entries and instead, always insert those after having proper initialisation. Tested-by: Stefan Hajnoczi <stefanha@redhat.com> Signed-off-by: Remy Noel <remy.noel@blade-group.com> Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> Reviewed-by: Paolo Bonzini <pbonzini@redhat.com> Message-id: 20181220152030.28035-3-remy.noel@blade-group.com Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
This commit is contained in:
parent
8821b34a73
commit
fef1660132
@ -200,6 +200,31 @@ static AioHandler *find_aio_handler(AioContext *ctx, int fd)
|
||||
return NULL;
|
||||
}
|
||||
|
||||
static bool aio_remove_fd_handler(AioContext *ctx, AioHandler *node)
|
||||
{
|
||||
/* If the GSource is in the process of being destroyed then
|
||||
* g_source_remove_poll() causes an assertion failure. Skip
|
||||
* removal in that case, because glib cleans up its state during
|
||||
* destruction anyway.
|
||||
*/
|
||||
if (!g_source_is_destroyed(&ctx->source)) {
|
||||
g_source_remove_poll(&ctx->source, &node->pfd);
|
||||
}
|
||||
|
||||
/* If a read is in progress, just mark the node as deleted */
|
||||
if (qemu_lockcnt_count(&ctx->list_lock)) {
|
||||
node->deleted = 1;
|
||||
node->pfd.revents = 0;
|
||||
return false;
|
||||
}
|
||||
/* Otherwise, delete it for real. We can't just mark it as
|
||||
* deleted because deleted nodes are only cleaned up while
|
||||
* no one is walking the handlers list.
|
||||
*/
|
||||
QLIST_REMOVE(node, node);
|
||||
return true;
|
||||
}
|
||||
|
||||
void aio_set_fd_handler(AioContext *ctx,
|
||||
int fd,
|
||||
bool is_external,
|
||||
@ -209,6 +234,7 @@ void aio_set_fd_handler(AioContext *ctx,
|
||||
void *opaque)
|
||||
{
|
||||
AioHandler *node;
|
||||
AioHandler *new_node = NULL;
|
||||
bool is_new = false;
|
||||
bool deleted = false;
|
||||
int poll_disable_change;
|
||||
@ -223,28 +249,6 @@ void aio_set_fd_handler(AioContext *ctx,
|
||||
qemu_lockcnt_unlock(&ctx->list_lock);
|
||||
return;
|
||||
}
|
||||
|
||||
/* If the GSource is in the process of being destroyed then
|
||||
* g_source_remove_poll() causes an assertion failure. Skip
|
||||
* removal in that case, because glib cleans up its state during
|
||||
* destruction anyway.
|
||||
*/
|
||||
if (!g_source_is_destroyed(&ctx->source)) {
|
||||
g_source_remove_poll(&ctx->source, &node->pfd);
|
||||
}
|
||||
|
||||
/* If a read is in progress, just mark the node as deleted */
|
||||
if (qemu_lockcnt_count(&ctx->list_lock)) {
|
||||
node->deleted = 1;
|
||||
node->pfd.revents = 0;
|
||||
} else {
|
||||
/* Otherwise, delete it for real. We can't just mark it as
|
||||
* deleted because deleted nodes are only cleaned up while
|
||||
* no one is walking the handlers list.
|
||||
*/
|
||||
QLIST_REMOVE(node, node);
|
||||
deleted = true;
|
||||
}
|
||||
/* Clean events in order to unregister fd from the ctx epoll. */
|
||||
node->pfd.events = 0;
|
||||
|
||||
@ -252,24 +256,32 @@ void aio_set_fd_handler(AioContext *ctx,
|
||||
} else {
|
||||
poll_disable_change = !io_poll - (node && !node->io_poll);
|
||||
if (node == NULL) {
|
||||
/* Alloc and insert if it's not already there */
|
||||
node = g_new0(AioHandler, 1);
|
||||
node->pfd.fd = fd;
|
||||
QLIST_INSERT_HEAD_RCU(&ctx->aio_handlers, node, node);
|
||||
|
||||
g_source_add_poll(&ctx->source, &node->pfd);
|
||||
is_new = true;
|
||||
}
|
||||
/* Alloc and insert if it's not already there */
|
||||
new_node = g_new0(AioHandler, 1);
|
||||
|
||||
/* Update handler with latest information */
|
||||
node->io_read = io_read;
|
||||
node->io_write = io_write;
|
||||
node->io_poll = io_poll;
|
||||
node->opaque = opaque;
|
||||
node->is_external = is_external;
|
||||
new_node->io_read = io_read;
|
||||
new_node->io_write = io_write;
|
||||
new_node->io_poll = io_poll;
|
||||
new_node->opaque = opaque;
|
||||
new_node->is_external = is_external;
|
||||
|
||||
node->pfd.events = (io_read ? G_IO_IN | G_IO_HUP | G_IO_ERR : 0);
|
||||
node->pfd.events |= (io_write ? G_IO_OUT | G_IO_ERR : 0);
|
||||
if (is_new) {
|
||||
new_node->pfd.fd = fd;
|
||||
} else {
|
||||
new_node->pfd = node->pfd;
|
||||
}
|
||||
g_source_add_poll(&ctx->source, &new_node->pfd);
|
||||
|
||||
new_node->pfd.events = (io_read ? G_IO_IN | G_IO_HUP | G_IO_ERR : 0);
|
||||
new_node->pfd.events |= (io_write ? G_IO_OUT | G_IO_ERR : 0);
|
||||
|
||||
QLIST_INSERT_HEAD_RCU(&ctx->aio_handlers, new_node, node);
|
||||
}
|
||||
if (node) {
|
||||
deleted = aio_remove_fd_handler(ctx, node);
|
||||
}
|
||||
|
||||
/* No need to order poll_disable_cnt writes against other updates;
|
||||
@ -281,7 +293,12 @@ void aio_set_fd_handler(AioContext *ctx,
|
||||
atomic_set(&ctx->poll_disable_cnt,
|
||||
atomic_read(&ctx->poll_disable_cnt) + poll_disable_change);
|
||||
|
||||
aio_epoll_update(ctx, node, is_new);
|
||||
if (new_node) {
|
||||
aio_epoll_update(ctx, new_node, is_new);
|
||||
} else if (node) {
|
||||
/* Unregister deleted fd_handler */
|
||||
aio_epoll_update(ctx, node, false);
|
||||
}
|
||||
qemu_lockcnt_unlock(&ctx->list_lock);
|
||||
aio_notify(ctx);
|
||||
|
||||
|
@ -35,6 +35,22 @@ struct AioHandler {
|
||||
QLIST_ENTRY(AioHandler) node;
|
||||
};
|
||||
|
||||
static void aio_remove_fd_handler(AioContext *ctx, AioHandler *node)
|
||||
{
|
||||
/* If aio_poll is in progress, just mark the node as deleted */
|
||||
if (qemu_lockcnt_count(&ctx->list_lock)) {
|
||||
node->deleted = 1;
|
||||
node->pfd.revents = 0;
|
||||
} else {
|
||||
/* Otherwise, delete it for real. We can't just mark it as
|
||||
* deleted because deleted nodes are only cleaned up after
|
||||
* releasing the list_lock.
|
||||
*/
|
||||
QLIST_REMOVE(node, node);
|
||||
g_free(node);
|
||||
}
|
||||
}
|
||||
|
||||
void aio_set_fd_handler(AioContext *ctx,
|
||||
int fd,
|
||||
bool is_external,
|
||||
@ -44,41 +60,23 @@ void aio_set_fd_handler(AioContext *ctx,
|
||||
void *opaque)
|
||||
{
|
||||
/* fd is a SOCKET in our case */
|
||||
AioHandler *node;
|
||||
AioHandler *old_node;
|
||||
AioHandler *node = NULL;
|
||||
|
||||
qemu_lockcnt_lock(&ctx->list_lock);
|
||||
QLIST_FOREACH(node, &ctx->aio_handlers, node) {
|
||||
if (node->pfd.fd == fd && !node->deleted) {
|
||||
QLIST_FOREACH(old_node, &ctx->aio_handlers, node) {
|
||||
if (old_node->pfd.fd == fd && !old_node->deleted) {
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
/* Are we deleting the fd handler? */
|
||||
if (!io_read && !io_write) {
|
||||
if (node) {
|
||||
/* If aio_poll is in progress, just mark the node as deleted */
|
||||
if (qemu_lockcnt_count(&ctx->list_lock)) {
|
||||
node->deleted = 1;
|
||||
node->pfd.revents = 0;
|
||||
} else {
|
||||
/* Otherwise, delete it for real. We can't just mark it as
|
||||
* deleted because deleted nodes are only cleaned up after
|
||||
* releasing the list_lock.
|
||||
*/
|
||||
QLIST_REMOVE(node, node);
|
||||
g_free(node);
|
||||
}
|
||||
}
|
||||
} else {
|
||||
if (io_read || io_write) {
|
||||
HANDLE event;
|
||||
long bitmask = 0;
|
||||
|
||||
if (node == NULL) {
|
||||
/* Alloc and insert if it's not already there */
|
||||
node = g_new0(AioHandler, 1);
|
||||
node->pfd.fd = fd;
|
||||
QLIST_INSERT_HEAD_RCU(&ctx->aio_handlers, node, node);
|
||||
}
|
||||
/* Alloc and insert if it's not already there */
|
||||
node = g_new0(AioHandler, 1);
|
||||
node->pfd.fd = fd;
|
||||
|
||||
node->pfd.events = 0;
|
||||
if (node->io_read) {
|
||||
@ -104,9 +102,13 @@ void aio_set_fd_handler(AioContext *ctx,
|
||||
bitmask |= FD_WRITE | FD_CONNECT;
|
||||
}
|
||||
|
||||
QLIST_INSERT_HEAD_RCU(&ctx->aio_handlers, node, node);
|
||||
event = event_notifier_get_handle(&ctx->notifier);
|
||||
WSAEventSelect(node->pfd.fd, event, bitmask);
|
||||
}
|
||||
if (old_node) {
|
||||
aio_remove_fd_handler(ctx, old_node);
|
||||
}
|
||||
|
||||
qemu_lockcnt_unlock(&ctx->list_lock);
|
||||
aio_notify(ctx);
|
||||
@ -139,18 +141,7 @@ void aio_set_event_notifier(AioContext *ctx,
|
||||
if (node) {
|
||||
g_source_remove_poll(&ctx->source, &node->pfd);
|
||||
|
||||
/* aio_poll is in progress, just mark the node as deleted */
|
||||
if (qemu_lockcnt_count(&ctx->list_lock)) {
|
||||
node->deleted = 1;
|
||||
node->pfd.revents = 0;
|
||||
} else {
|
||||
/* Otherwise, delete it for real. We can't just mark it as
|
||||
* deleted because deleted nodes are only cleaned up after
|
||||
* releasing the list_lock.
|
||||
*/
|
||||
QLIST_REMOVE(node, node);
|
||||
g_free(node);
|
||||
}
|
||||
aio_remove_fd_handler(ctx, node);
|
||||
}
|
||||
} else {
|
||||
if (node == NULL) {
|
||||
|
Loading…
Reference in New Issue
Block a user