mirror of https://github.com/xqemu/xqemu.git
aio: make ctx->list_lock a QemuLockCnt, subsuming ctx->walking_bh
This will make it possible to walk the list of bottom halves without holding the AioContext lock---and in turn to call bottom half handlers without holding the lock. Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> Reviewed-by: Fam Zheng <famz@redhat.com> Message-id: 20170112180800.21085-4-pbonzini@redhat.com Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
This commit is contained in:
parent
51dee5e465
commit
d7c99a1282
35
async.c
35
async.c
|
@ -53,14 +53,14 @@ void aio_bh_schedule_oneshot(AioContext *ctx, QEMUBHFunc *cb, void *opaque)
|
||||||
.cb = cb,
|
.cb = cb,
|
||||||
.opaque = opaque,
|
.opaque = opaque,
|
||||||
};
|
};
|
||||||
qemu_mutex_lock(&ctx->list_lock);
|
qemu_lockcnt_lock(&ctx->list_lock);
|
||||||
bh->next = ctx->first_bh;
|
bh->next = ctx->first_bh;
|
||||||
bh->scheduled = 1;
|
bh->scheduled = 1;
|
||||||
bh->deleted = 1;
|
bh->deleted = 1;
|
||||||
/* Make sure that the members are ready before putting bh into list */
|
/* Make sure that the members are ready before putting bh into list */
|
||||||
smp_wmb();
|
smp_wmb();
|
||||||
ctx->first_bh = bh;
|
ctx->first_bh = bh;
|
||||||
qemu_mutex_unlock(&ctx->list_lock);
|
qemu_lockcnt_unlock(&ctx->list_lock);
|
||||||
aio_notify(ctx);
|
aio_notify(ctx);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -73,12 +73,12 @@ QEMUBH *aio_bh_new(AioContext *ctx, QEMUBHFunc *cb, void *opaque)
|
||||||
.cb = cb,
|
.cb = cb,
|
||||||
.opaque = opaque,
|
.opaque = opaque,
|
||||||
};
|
};
|
||||||
qemu_mutex_lock(&ctx->list_lock);
|
qemu_lockcnt_lock(&ctx->list_lock);
|
||||||
bh->next = ctx->first_bh;
|
bh->next = ctx->first_bh;
|
||||||
/* Make sure that the members are ready before putting bh into list */
|
/* Make sure that the members are ready before putting bh into list */
|
||||||
smp_wmb();
|
smp_wmb();
|
||||||
ctx->first_bh = bh;
|
ctx->first_bh = bh;
|
||||||
qemu_mutex_unlock(&ctx->list_lock);
|
qemu_lockcnt_unlock(&ctx->list_lock);
|
||||||
return bh;
|
return bh;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -93,13 +93,11 @@ int aio_bh_poll(AioContext *ctx)
|
||||||
QEMUBH *bh, **bhp, *next;
|
QEMUBH *bh, **bhp, *next;
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
ctx->walking_bh++;
|
qemu_lockcnt_inc(&ctx->list_lock);
|
||||||
|
|
||||||
ret = 0;
|
ret = 0;
|
||||||
for (bh = ctx->first_bh; bh; bh = next) {
|
for (bh = atomic_rcu_read(&ctx->first_bh); bh; bh = next) {
|
||||||
/* Make sure that fetching bh happens before accessing its members */
|
next = atomic_rcu_read(&bh->next);
|
||||||
smp_read_barrier_depends();
|
|
||||||
next = bh->next;
|
|
||||||
/* The atomic_xchg is paired with the one in qemu_bh_schedule. The
|
/* The atomic_xchg is paired with the one in qemu_bh_schedule. The
|
||||||
* implicit memory barrier ensures that the callback sees all writes
|
* implicit memory barrier ensures that the callback sees all writes
|
||||||
* done by the scheduling thread. It also ensures that the scheduling
|
* done by the scheduling thread. It also ensures that the scheduling
|
||||||
|
@ -116,11 +114,8 @@ int aio_bh_poll(AioContext *ctx)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
ctx->walking_bh--;
|
|
||||||
|
|
||||||
/* remove deleted bhs */
|
/* remove deleted bhs */
|
||||||
if (!ctx->walking_bh) {
|
if (qemu_lockcnt_dec_and_lock(&ctx->list_lock)) {
|
||||||
qemu_mutex_lock(&ctx->list_lock);
|
|
||||||
bhp = &ctx->first_bh;
|
bhp = &ctx->first_bh;
|
||||||
while (*bhp) {
|
while (*bhp) {
|
||||||
bh = *bhp;
|
bh = *bhp;
|
||||||
|
@ -131,7 +126,7 @@ int aio_bh_poll(AioContext *ctx)
|
||||||
bhp = &bh->next;
|
bhp = &bh->next;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
qemu_mutex_unlock(&ctx->list_lock);
|
qemu_lockcnt_unlock(&ctx->list_lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
return ret;
|
return ret;
|
||||||
|
@ -187,7 +182,8 @@ aio_compute_timeout(AioContext *ctx)
|
||||||
int timeout = -1;
|
int timeout = -1;
|
||||||
QEMUBH *bh;
|
QEMUBH *bh;
|
||||||
|
|
||||||
for (bh = ctx->first_bh; bh; bh = bh->next) {
|
for (bh = atomic_rcu_read(&ctx->first_bh); bh;
|
||||||
|
bh = atomic_rcu_read(&bh->next)) {
|
||||||
if (bh->scheduled) {
|
if (bh->scheduled) {
|
||||||
if (bh->idle) {
|
if (bh->idle) {
|
||||||
/* idle bottom halves will be polled at least
|
/* idle bottom halves will be polled at least
|
||||||
|
@ -270,7 +266,8 @@ aio_ctx_finalize(GSource *source)
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
qemu_mutex_lock(&ctx->list_lock);
|
qemu_lockcnt_lock(&ctx->list_lock);
|
||||||
|
assert(!qemu_lockcnt_count(&ctx->list_lock));
|
||||||
while (ctx->first_bh) {
|
while (ctx->first_bh) {
|
||||||
QEMUBH *next = ctx->first_bh->next;
|
QEMUBH *next = ctx->first_bh->next;
|
||||||
|
|
||||||
|
@ -280,12 +277,12 @@ aio_ctx_finalize(GSource *source)
|
||||||
g_free(ctx->first_bh);
|
g_free(ctx->first_bh);
|
||||||
ctx->first_bh = next;
|
ctx->first_bh = next;
|
||||||
}
|
}
|
||||||
qemu_mutex_unlock(&ctx->list_lock);
|
qemu_lockcnt_unlock(&ctx->list_lock);
|
||||||
|
|
||||||
aio_set_event_notifier(ctx, &ctx->notifier, false, NULL, NULL);
|
aio_set_event_notifier(ctx, &ctx->notifier, false, NULL, NULL);
|
||||||
event_notifier_cleanup(&ctx->notifier);
|
event_notifier_cleanup(&ctx->notifier);
|
||||||
qemu_rec_mutex_destroy(&ctx->lock);
|
qemu_rec_mutex_destroy(&ctx->lock);
|
||||||
qemu_mutex_destroy(&ctx->list_lock);
|
qemu_lockcnt_destroy(&ctx->list_lock);
|
||||||
timerlistgroup_deinit(&ctx->tlg);
|
timerlistgroup_deinit(&ctx->tlg);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -372,6 +369,7 @@ AioContext *aio_context_new(Error **errp)
|
||||||
goto fail;
|
goto fail;
|
||||||
}
|
}
|
||||||
g_source_set_can_recurse(&ctx->source, true);
|
g_source_set_can_recurse(&ctx->source, true);
|
||||||
|
qemu_lockcnt_init(&ctx->list_lock);
|
||||||
aio_set_event_notifier(ctx, &ctx->notifier,
|
aio_set_event_notifier(ctx, &ctx->notifier,
|
||||||
false,
|
false,
|
||||||
(EventNotifierHandler *)
|
(EventNotifierHandler *)
|
||||||
|
@ -381,7 +379,6 @@ AioContext *aio_context_new(Error **errp)
|
||||||
ctx->linux_aio = NULL;
|
ctx->linux_aio = NULL;
|
||||||
#endif
|
#endif
|
||||||
ctx->thread_pool = NULL;
|
ctx->thread_pool = NULL;
|
||||||
qemu_mutex_init(&ctx->list_lock);
|
|
||||||
qemu_rec_mutex_init(&ctx->lock);
|
qemu_rec_mutex_init(&ctx->lock);
|
||||||
timerlistgroup_init(&ctx->tlg, aio_timerlist_notify, ctx);
|
timerlistgroup_init(&ctx->tlg, aio_timerlist_notify, ctx);
|
||||||
|
|
||||||
|
|
|
@ -90,17 +90,15 @@ struct AioContext {
|
||||||
*/
|
*/
|
||||||
uint32_t notify_me;
|
uint32_t notify_me;
|
||||||
|
|
||||||
/* lock to protect between bh's adders and deleter */
|
/* A lock to protect between bh's adders and deleter, and to ensure
|
||||||
QemuMutex list_lock;
|
* that no callbacks are removed while we're walking and dispatching
|
||||||
|
* them.
|
||||||
|
*/
|
||||||
|
QemuLockCnt list_lock;
|
||||||
|
|
||||||
/* Anchor of the list of Bottom Halves belonging to the context */
|
/* Anchor of the list of Bottom Halves belonging to the context */
|
||||||
struct QEMUBH *first_bh;
|
struct QEMUBH *first_bh;
|
||||||
|
|
||||||
/* A simple lock used to protect the first_bh list, and ensure that
|
|
||||||
* no callbacks are removed while we're walking and dispatching callbacks.
|
|
||||||
*/
|
|
||||||
int walking_bh;
|
|
||||||
|
|
||||||
/* Used by aio_notify.
|
/* Used by aio_notify.
|
||||||
*
|
*
|
||||||
* "notified" is used to avoid expensive event_notifier_test_and_clear
|
* "notified" is used to avoid expensive event_notifier_test_and_clear
|
||||||
|
|
Loading…
Reference in New Issue