mirror of https://github.com/xemu-project/xemu.git
cputlb: Move cpu->pending_tlb_flush to env->tlb_c.pending_flush
Protect it with the tlb_lock instead of using atomics. The move puts it in or near the same cacheline as the lock; using the lock means we don't need a second atomic operation in order to perform the update. Which makes it cheap to also update pending_flush in tlb_flush_by_mmuidx_async_work. Tested-by: Emilio G. Cota <cota@braap.org> Reviewed-by: Emilio G. Cota <cota@braap.org> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
This commit is contained in:
parent
8ab102667e
commit
60a2ad7d86
|
@ -133,6 +133,7 @@ static void tlb_flush_nocheck(CPUState *cpu)
|
||||||
* that do not hold the lock are performed by the same owner thread.
|
* that do not hold the lock are performed by the same owner thread.
|
||||||
*/
|
*/
|
||||||
qemu_spin_lock(&env->tlb_c.lock);
|
qemu_spin_lock(&env->tlb_c.lock);
|
||||||
|
env->tlb_c.pending_flush = 0;
|
||||||
memset(env->tlb_table, -1, sizeof(env->tlb_table));
|
memset(env->tlb_table, -1, sizeof(env->tlb_table));
|
||||||
memset(env->tlb_v_table, -1, sizeof(env->tlb_v_table));
|
memset(env->tlb_v_table, -1, sizeof(env->tlb_v_table));
|
||||||
qemu_spin_unlock(&env->tlb_c.lock);
|
qemu_spin_unlock(&env->tlb_c.lock);
|
||||||
|
@ -142,8 +143,6 @@ static void tlb_flush_nocheck(CPUState *cpu)
|
||||||
env->vtlb_index = 0;
|
env->vtlb_index = 0;
|
||||||
env->tlb_flush_addr = -1;
|
env->tlb_flush_addr = -1;
|
||||||
env->tlb_flush_mask = 0;
|
env->tlb_flush_mask = 0;
|
||||||
|
|
||||||
atomic_mb_set(&cpu->pending_tlb_flush, 0);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static void tlb_flush_global_async_work(CPUState *cpu, run_on_cpu_data data)
|
static void tlb_flush_global_async_work(CPUState *cpu, run_on_cpu_data data)
|
||||||
|
@ -154,8 +153,15 @@ static void tlb_flush_global_async_work(CPUState *cpu, run_on_cpu_data data)
|
||||||
void tlb_flush(CPUState *cpu)
|
void tlb_flush(CPUState *cpu)
|
||||||
{
|
{
|
||||||
if (cpu->created && !qemu_cpu_is_self(cpu)) {
|
if (cpu->created && !qemu_cpu_is_self(cpu)) {
|
||||||
if (atomic_mb_read(&cpu->pending_tlb_flush) != ALL_MMUIDX_BITS) {
|
CPUArchState *env = cpu->env_ptr;
|
||||||
atomic_mb_set(&cpu->pending_tlb_flush, ALL_MMUIDX_BITS);
|
uint16_t pending;
|
||||||
|
|
||||||
|
qemu_spin_lock(&env->tlb_c.lock);
|
||||||
|
pending = env->tlb_c.pending_flush;
|
||||||
|
env->tlb_c.pending_flush = ALL_MMUIDX_BITS;
|
||||||
|
qemu_spin_unlock(&env->tlb_c.lock);
|
||||||
|
|
||||||
|
if (pending != ALL_MMUIDX_BITS) {
|
||||||
async_run_on_cpu(cpu, tlb_flush_global_async_work,
|
async_run_on_cpu(cpu, tlb_flush_global_async_work,
|
||||||
RUN_ON_CPU_NULL);
|
RUN_ON_CPU_NULL);
|
||||||
}
|
}
|
||||||
|
@ -189,6 +195,8 @@ static void tlb_flush_by_mmuidx_async_work(CPUState *cpu, run_on_cpu_data data)
|
||||||
tlb_debug("start: mmu_idx:0x%04lx\n", mmu_idx_bitmask);
|
tlb_debug("start: mmu_idx:0x%04lx\n", mmu_idx_bitmask);
|
||||||
|
|
||||||
qemu_spin_lock(&env->tlb_c.lock);
|
qemu_spin_lock(&env->tlb_c.lock);
|
||||||
|
env->tlb_c.pending_flush &= ~mmu_idx_bitmask;
|
||||||
|
|
||||||
for (mmu_idx = 0; mmu_idx < NB_MMU_MODES; mmu_idx++) {
|
for (mmu_idx = 0; mmu_idx < NB_MMU_MODES; mmu_idx++) {
|
||||||
|
|
||||||
if (test_bit(mmu_idx, &mmu_idx_bitmask)) {
|
if (test_bit(mmu_idx, &mmu_idx_bitmask)) {
|
||||||
|
@ -210,19 +218,22 @@ void tlb_flush_by_mmuidx(CPUState *cpu, uint16_t idxmap)
|
||||||
tlb_debug("mmu_idx: 0x%" PRIx16 "\n", idxmap);
|
tlb_debug("mmu_idx: 0x%" PRIx16 "\n", idxmap);
|
||||||
|
|
||||||
if (!qemu_cpu_is_self(cpu)) {
|
if (!qemu_cpu_is_self(cpu)) {
|
||||||
uint16_t pending_flushes = idxmap;
|
CPUArchState *env = cpu->env_ptr;
|
||||||
pending_flushes &= ~atomic_mb_read(&cpu->pending_tlb_flush);
|
uint16_t pending, to_clean;
|
||||||
|
|
||||||
if (pending_flushes) {
|
qemu_spin_lock(&env->tlb_c.lock);
|
||||||
tlb_debug("reduced mmu_idx: 0x%" PRIx16 "\n", pending_flushes);
|
pending = env->tlb_c.pending_flush;
|
||||||
|
to_clean = idxmap & ~pending;
|
||||||
|
env->tlb_c.pending_flush = pending | idxmap;
|
||||||
|
qemu_spin_unlock(&env->tlb_c.lock);
|
||||||
|
|
||||||
atomic_or(&cpu->pending_tlb_flush, pending_flushes);
|
if (to_clean) {
|
||||||
|
tlb_debug("reduced mmu_idx: 0x%" PRIx16 "\n", to_clean);
|
||||||
async_run_on_cpu(cpu, tlb_flush_by_mmuidx_async_work,
|
async_run_on_cpu(cpu, tlb_flush_by_mmuidx_async_work,
|
||||||
RUN_ON_CPU_HOST_INT(pending_flushes));
|
RUN_ON_CPU_HOST_INT(to_clean));
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
tlb_flush_by_mmuidx_async_work(cpu,
|
tlb_flush_by_mmuidx_async_work(cpu, RUN_ON_CPU_HOST_INT(idxmap));
|
||||||
RUN_ON_CPU_HOST_INT(idxmap));
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -145,8 +145,14 @@ typedef struct CPUIOTLBEntry {
|
||||||
* Data elements that are shared between all MMU modes.
|
* Data elements that are shared between all MMU modes.
|
||||||
*/
|
*/
|
||||||
typedef struct CPUTLBCommon {
|
typedef struct CPUTLBCommon {
|
||||||
/* lock serializes updates to tlb_table and tlb_v_table */
|
/* Serialize updates to tlb_table and tlb_v_table, and others as noted. */
|
||||||
QemuSpin lock;
|
QemuSpin lock;
|
||||||
|
/*
|
||||||
|
* Within pending_flush, for each bit N, there exists an outstanding
|
||||||
|
* cross-cpu flush for mmu_idx N. Further cross-cpu flushes to that
|
||||||
|
* mmu_idx may be discarded. Protected by tlb_c.lock.
|
||||||
|
*/
|
||||||
|
uint16_t pending_flush;
|
||||||
} CPUTLBCommon;
|
} CPUTLBCommon;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
|
|
@ -429,12 +429,6 @@ struct CPUState {
|
||||||
|
|
||||||
struct hax_vcpu_state *hax_vcpu;
|
struct hax_vcpu_state *hax_vcpu;
|
||||||
|
|
||||||
/* The pending_tlb_flush flag is set and cleared atomically to
|
|
||||||
* avoid potential races. The aim of the flag is to avoid
|
|
||||||
* unnecessary flushes.
|
|
||||||
*/
|
|
||||||
uint16_t pending_tlb_flush;
|
|
||||||
|
|
||||||
int hvf_fd;
|
int hvf_fd;
|
||||||
|
|
||||||
/* track IOMMUs whose translations we've cached in the TCG TLB */
|
/* track IOMMUs whose translations we've cached in the TCG TLB */
|
||||||
|
|
Loading…
Reference in New Issue