mirror of https://github.com/xemu-project/xemu.git
atomics: eliminate mb_read/mb_set
qatomic_mb_read and qatomic_mb_set were the very first atomic primitives introduced for QEMU; their semantics are unclear and they provide a false sense of safety. The last use of qatomic_mb_read() has been removed, so delete it. qatomic_mb_set() instead can survive as an optimized qatomic_set()+smp_mb(), similar to Linux's smp_store_mb(), but rename it to qatomic_set_mb() to match the order of the two operations. Reviewed-by: Richard Henderson <richard.henderson@linaro.org> Signed-off-by: Paolo Bonzini <pbonzini@redhat.com>
This commit is contained in:
parent
09a49afeae
commit
06831001ac
|
@ -774,7 +774,7 @@ static inline bool cpu_handle_interrupt(CPUState *cpu,
|
||||||
* Ensure zeroing happens before reading cpu->exit_request or
|
* Ensure zeroing happens before reading cpu->exit_request or
|
||||||
* cpu->interrupt_request (see also smp_wmb in cpu_exit())
|
* cpu->interrupt_request (see also smp_wmb in cpu_exit())
|
||||||
*/
|
*/
|
||||||
qatomic_mb_set(&cpu_neg(cpu)->icount_decr.u16.high, 0);
|
qatomic_set_mb(&cpu_neg(cpu)->icount_decr.u16.high, 0);
|
||||||
|
|
||||||
if (unlikely(qatomic_read(&cpu->interrupt_request))) {
|
if (unlikely(qatomic_read(&cpu->interrupt_request))) {
|
||||||
int interrupt_request;
|
int interrupt_request;
|
||||||
|
|
|
@ -119,7 +119,7 @@ static void *mttcg_cpu_thread_fn(void *arg)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
qatomic_mb_set(&cpu->exit_request, 0);
|
qatomic_set_mb(&cpu->exit_request, 0);
|
||||||
qemu_wait_io_event(cpu);
|
qemu_wait_io_event(cpu);
|
||||||
} while (!cpu->unplug || cpu_can_run(cpu));
|
} while (!cpu->unplug || cpu_can_run(cpu));
|
||||||
|
|
||||||
|
|
|
@ -244,7 +244,7 @@ static void *rr_cpu_thread_fn(void *arg)
|
||||||
|
|
||||||
while (cpu && cpu_work_list_empty(cpu) && !cpu->exit_request) {
|
while (cpu && cpu_work_list_empty(cpu) && !cpu->exit_request) {
|
||||||
/* Store rr_current_cpu before evaluating cpu_can_run(). */
|
/* Store rr_current_cpu before evaluating cpu_can_run(). */
|
||||||
qatomic_mb_set(&rr_current_cpu, cpu);
|
qatomic_set_mb(&rr_current_cpu, cpu);
|
||||||
|
|
||||||
current_cpu = cpu;
|
current_cpu = cpu;
|
||||||
|
|
||||||
|
@ -287,7 +287,7 @@ static void *rr_cpu_thread_fn(void *arg)
|
||||||
qatomic_set(&rr_current_cpu, NULL);
|
qatomic_set(&rr_current_cpu, NULL);
|
||||||
|
|
||||||
if (cpu && cpu->exit_request) {
|
if (cpu && cpu->exit_request) {
|
||||||
qatomic_mb_set(&cpu->exit_request, 0);
|
qatomic_set_mb(&cpu->exit_request, 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (icount_enabled() && all_cpu_threads_idle()) {
|
if (icount_enabled() && all_cpu_threads_idle()) {
|
||||||
|
|
|
@ -102,28 +102,10 @@ Similar operations return the new value of ``*ptr``::
|
||||||
typeof(*ptr) qatomic_or_fetch(ptr, val)
|
typeof(*ptr) qatomic_or_fetch(ptr, val)
|
||||||
typeof(*ptr) qatomic_xor_fetch(ptr, val)
|
typeof(*ptr) qatomic_xor_fetch(ptr, val)
|
||||||
|
|
||||||
``qemu/atomic.h`` also provides loads and stores that cannot be reordered
|
``qemu/atomic.h`` also provides an optimized shortcut for
|
||||||
with each other::
|
``qatomic_set`` followed by ``smp_mb``::
|
||||||
|
|
||||||
typeof(*ptr) qatomic_mb_read(ptr)
|
void qatomic_set_mb(ptr, val)
|
||||||
void qatomic_mb_set(ptr, val)
|
|
||||||
|
|
||||||
However these do not provide sequential consistency and, in particular,
|
|
||||||
they do not participate in the total ordering enforced by
|
|
||||||
sequentially-consistent operations. For this reason they are deprecated.
|
|
||||||
They should instead be replaced with any of the following (ordered from
|
|
||||||
easiest to hardest):
|
|
||||||
|
|
||||||
- accesses inside a mutex or spinlock
|
|
||||||
|
|
||||||
- lightweight synchronization primitives such as ``QemuEvent``
|
|
||||||
|
|
||||||
- RCU operations (``qatomic_rcu_read``, ``qatomic_rcu_set``) when publishing
|
|
||||||
or accessing a new version of a data structure
|
|
||||||
|
|
||||||
- other atomic accesses: ``qatomic_read`` and ``qatomic_load_acquire`` for
|
|
||||||
loads, ``qatomic_set`` and ``qatomic_store_release`` for stores, ``smp_mb``
|
|
||||||
to forbid reordering subsequent loads before a store.
|
|
||||||
|
|
||||||
|
|
||||||
Weak atomic access and manual memory barriers
|
Weak atomic access and manual memory barriers
|
||||||
|
@ -523,8 +505,7 @@ and memory barriers, and the equivalents in QEMU:
|
||||||
| :: |
|
| :: |
|
||||||
| |
|
| |
|
||||||
| a = qatomic_read(&x); |
|
| a = qatomic_read(&x); |
|
||||||
| qatomic_set(&x, a + 2); |
|
| qatomic_set_mb(&x, a + 2); |
|
||||||
| smp_mb(); |
|
|
||||||
| b = qatomic_read(&y); |
|
| b = qatomic_read(&y); |
|
||||||
+--------------------------------+
|
+--------------------------------+
|
||||||
|
|
||||||
|
|
|
@ -259,24 +259,17 @@
|
||||||
# define smp_mb__after_rmw() smp_mb()
|
# define smp_mb__after_rmw() smp_mb()
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
/* qatomic_mb_read/set semantics map Java volatile variables. They are
|
/*
|
||||||
* less expensive on some platforms (notably POWER) than fully
|
* On some architectures, qatomic_set_mb is more efficient than a store
|
||||||
* sequentially consistent operations.
|
* plus a fence.
|
||||||
*
|
|
||||||
* As long as they are used as paired operations they are safe to
|
|
||||||
* use. See docs/devel/atomics.rst for more discussion.
|
|
||||||
*/
|
*/
|
||||||
|
|
||||||
#define qatomic_mb_read(ptr) \
|
|
||||||
qatomic_load_acquire(ptr)
|
|
||||||
|
|
||||||
#if !defined(QEMU_SANITIZE_THREAD) && \
|
#if !defined(QEMU_SANITIZE_THREAD) && \
|
||||||
(defined(__i386__) || defined(__x86_64__) || defined(__s390x__))
|
(defined(__i386__) || defined(__x86_64__) || defined(__s390x__))
|
||||||
/* This is more efficient than a store plus a fence. */
|
# define qatomic_set_mb(ptr, i) \
|
||||||
# define qatomic_mb_set(ptr, i) \
|
|
||||||
({ (void)qatomic_xchg(ptr, i); smp_mb__after_rmw(); })
|
({ (void)qatomic_xchg(ptr, i); smp_mb__after_rmw(); })
|
||||||
#else
|
#else
|
||||||
# define qatomic_mb_set(ptr, i) \
|
# define qatomic_set_mb(ptr, i) \
|
||||||
({ qatomic_store_release(ptr, i); smp_mb(); })
|
({ qatomic_store_release(ptr, i); smp_mb(); })
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
|
|
@ -246,7 +246,7 @@ static QMPRequest *monitor_qmp_dispatcher_pop_any(void)
|
||||||
*
|
*
|
||||||
* Clear qmp_dispatcher_co_busy before reading request.
|
* Clear qmp_dispatcher_co_busy before reading request.
|
||||||
*/
|
*/
|
||||||
qatomic_mb_set(&qmp_dispatcher_co_busy, false);
|
qatomic_set_mb(&qmp_dispatcher_co_busy, false);
|
||||||
|
|
||||||
WITH_QEMU_LOCK_GUARD(&monitor_lock) {
|
WITH_QEMU_LOCK_GUARD(&monitor_lock) {
|
||||||
QMPRequest *req_obj;
|
QMPRequest *req_obj;
|
||||||
|
|
|
@ -405,7 +405,7 @@ static void qemu_cpu_stop(CPUState *cpu, bool exit)
|
||||||
|
|
||||||
void qemu_wait_io_event_common(CPUState *cpu)
|
void qemu_wait_io_event_common(CPUState *cpu)
|
||||||
{
|
{
|
||||||
qatomic_mb_set(&cpu->thread_kicked, false);
|
qatomic_set_mb(&cpu->thread_kicked, false);
|
||||||
if (cpu->stop) {
|
if (cpu->stop) {
|
||||||
qemu_cpu_stop(cpu, false);
|
qemu_cpu_stop(cpu, false);
|
||||||
}
|
}
|
||||||
|
|
|
@ -3132,7 +3132,7 @@ void address_space_unmap(AddressSpace *as, void *buffer, hwaddr len,
|
||||||
bounce.buffer = NULL;
|
bounce.buffer = NULL;
|
||||||
memory_region_unref(bounce.mr);
|
memory_region_unref(bounce.mr);
|
||||||
/* Clear in_use before reading map_client_list. */
|
/* Clear in_use before reading map_client_list. */
|
||||||
qatomic_mb_set(&bounce.in_use, false);
|
qatomic_set_mb(&bounce.in_use, false);
|
||||||
cpu_notify_map_clients();
|
cpu_notify_map_clients();
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1229,7 +1229,7 @@ static void hvf_wait_for_ipi(CPUState *cpu, struct timespec *ts)
|
||||||
* Use pselect to sleep so that other threads can IPI us while we're
|
* Use pselect to sleep so that other threads can IPI us while we're
|
||||||
* sleeping.
|
* sleeping.
|
||||||
*/
|
*/
|
||||||
qatomic_mb_set(&cpu->thread_kicked, false);
|
qatomic_set_mb(&cpu->thread_kicked, false);
|
||||||
qemu_mutex_unlock_iothread();
|
qemu_mutex_unlock_iothread();
|
||||||
pselect(0, 0, 0, 0, ts, &cpu->hvf->unblock_ipi_mask);
|
pselect(0, 0, 0, 0, ts, &cpu->hvf->unblock_ipi_mask);
|
||||||
qemu_mutex_lock_iothread();
|
qemu_mutex_lock_iothread();
|
||||||
|
|
|
@ -154,7 +154,7 @@ static coroutine_fn void test_multi_co_schedule_entry(void *opaque)
|
||||||
n = g_test_rand_int_range(0, NUM_CONTEXTS);
|
n = g_test_rand_int_range(0, NUM_CONTEXTS);
|
||||||
schedule_next(n);
|
schedule_next(n);
|
||||||
|
|
||||||
qatomic_mb_set(&to_schedule[id], qemu_coroutine_self());
|
qatomic_set_mb(&to_schedule[id], qemu_coroutine_self());
|
||||||
/* finish_cb can run here. */
|
/* finish_cb can run here. */
|
||||||
qemu_coroutine_yield();
|
qemu_coroutine_yield();
|
||||||
g_assert(to_schedule[id] == NULL);
|
g_assert(to_schedule[id] == NULL);
|
||||||
|
|
|
@ -202,7 +202,7 @@ static void coroutine_fn qemu_co_mutex_lock_slowpath(AioContext *ctx,
|
||||||
push_waiter(mutex, &w);
|
push_waiter(mutex, &w);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Add waiter before reading mutex->handoff. Pairs with qatomic_mb_set
|
* Add waiter before reading mutex->handoff. Pairs with qatomic_set_mb
|
||||||
* in qemu_co_mutex_unlock.
|
* in qemu_co_mutex_unlock.
|
||||||
*/
|
*/
|
||||||
smp_mb__after_rmw();
|
smp_mb__after_rmw();
|
||||||
|
@ -310,7 +310,7 @@ void coroutine_fn qemu_co_mutex_unlock(CoMutex *mutex)
|
||||||
|
|
||||||
our_handoff = mutex->sequence;
|
our_handoff = mutex->sequence;
|
||||||
/* Set handoff before checking for waiters. */
|
/* Set handoff before checking for waiters. */
|
||||||
qatomic_mb_set(&mutex->handoff, our_handoff);
|
qatomic_set_mb(&mutex->handoff, our_handoff);
|
||||||
if (!has_waiters(mutex)) {
|
if (!has_waiters(mutex)) {
|
||||||
/* The concurrent lock has not added itself yet, so it
|
/* The concurrent lock has not added itself yet, so it
|
||||||
* will be able to pick our handoff.
|
* will be able to pick our handoff.
|
||||||
|
|
Loading…
Reference in New Issue