mirror of https://github.com/xqemu/xqemu.git
tcg: move tb_ctx.tb_phys_invalidate_count to tcg_ctx
Thereby making it per-TCGContext. Once we remove tb_lock, this will avoid an atomic increment every time a TB is invalidated. Reviewed-by: Richard Henderson <richard.henderson@linaro.org> Reviewed-by: Alex Bennée <alex.bennee@linaro.org> Signed-off-by: Emilio G. Cota <cota@braap.org> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
This commit is contained in:
parent
be2cdc5e35
commit
128ed2278c
|
@ -1069,7 +1069,8 @@ void tb_phys_invalidate(TranslationBlock *tb, tb_page_addr_t page_addr)
|
||||||
/* suppress any remaining jumps to this TB */
|
/* suppress any remaining jumps to this TB */
|
||||||
tb_jmp_unlink(tb);
|
tb_jmp_unlink(tb);
|
||||||
|
|
||||||
tb_ctx.tb_phys_invalidate_count++;
|
atomic_set(&tcg_ctx->tb_phys_invalidate_count,
|
||||||
|
tcg_ctx->tb_phys_invalidate_count + 1);
|
||||||
}
|
}
|
||||||
|
|
||||||
#ifdef CONFIG_SOFTMMU
|
#ifdef CONFIG_SOFTMMU
|
||||||
|
@ -1855,7 +1856,7 @@ void dump_exec_info(FILE *f, fprintf_function cpu_fprintf)
|
||||||
cpu_fprintf(f, "\nStatistics:\n");
|
cpu_fprintf(f, "\nStatistics:\n");
|
||||||
cpu_fprintf(f, "TB flush count %u\n",
|
cpu_fprintf(f, "TB flush count %u\n",
|
||||||
atomic_read(&tb_ctx.tb_flush_count));
|
atomic_read(&tb_ctx.tb_flush_count));
|
||||||
cpu_fprintf(f, "TB invalidate count %d\n", tb_ctx.tb_phys_invalidate_count);
|
cpu_fprintf(f, "TB invalidate count %zu\n", tcg_tb_phys_invalidate_count());
|
||||||
cpu_fprintf(f, "TLB flush count %zu\n", tlb_flush_count());
|
cpu_fprintf(f, "TLB flush count %zu\n", tlb_flush_count());
|
||||||
tcg_dump_info(f, cpu_fprintf);
|
tcg_dump_info(f, cpu_fprintf);
|
||||||
}
|
}
|
||||||
|
|
|
@ -37,7 +37,6 @@ struct TBContext {
|
||||||
|
|
||||||
/* statistics */
|
/* statistics */
|
||||||
unsigned tb_flush_count;
|
unsigned tb_flush_count;
|
||||||
int tb_phys_invalidate_count;
|
|
||||||
};
|
};
|
||||||
|
|
||||||
extern TBContext tb_ctx;
|
extern TBContext tb_ctx;
|
||||||
|
|
14
tcg/tcg.c
14
tcg/tcg.c
|
@ -791,6 +791,20 @@ size_t tcg_code_capacity(void)
|
||||||
return capacity;
|
return capacity;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
size_t tcg_tb_phys_invalidate_count(void)
|
||||||
|
{
|
||||||
|
unsigned int n_ctxs = atomic_read(&n_tcg_ctxs);
|
||||||
|
unsigned int i;
|
||||||
|
size_t total = 0;
|
||||||
|
|
||||||
|
for (i = 0; i < n_ctxs; i++) {
|
||||||
|
const TCGContext *s = atomic_read(&tcg_ctxs[i]);
|
||||||
|
|
||||||
|
total += atomic_read(&s->tb_phys_invalidate_count);
|
||||||
|
}
|
||||||
|
return total;
|
||||||
|
}
|
||||||
|
|
||||||
/* pool based memory allocation */
|
/* pool based memory allocation */
|
||||||
void *tcg_malloc_internal(TCGContext *s, int size)
|
void *tcg_malloc_internal(TCGContext *s, int size)
|
||||||
{
|
{
|
||||||
|
|
|
@ -695,6 +695,8 @@ struct TCGContext {
|
||||||
/* Threshold to flush the translated code buffer. */
|
/* Threshold to flush the translated code buffer. */
|
||||||
void *code_gen_highwater;
|
void *code_gen_highwater;
|
||||||
|
|
||||||
|
size_t tb_phys_invalidate_count;
|
||||||
|
|
||||||
/* Track which vCPU triggers events */
|
/* Track which vCPU triggers events */
|
||||||
CPUState *cpu; /* *_trans */
|
CPUState *cpu; /* *_trans */
|
||||||
|
|
||||||
|
@ -868,6 +870,7 @@ size_t tcg_code_capacity(void);
|
||||||
|
|
||||||
void tcg_tb_insert(TranslationBlock *tb);
|
void tcg_tb_insert(TranslationBlock *tb);
|
||||||
void tcg_tb_remove(TranslationBlock *tb);
|
void tcg_tb_remove(TranslationBlock *tb);
|
||||||
|
size_t tcg_tb_phys_invalidate_count(void);
|
||||||
TranslationBlock *tcg_tb_lookup(uintptr_t tc_ptr);
|
TranslationBlock *tcg_tb_lookup(uintptr_t tc_ptr);
|
||||||
void tcg_tb_foreach(GTraverseFunc func, gpointer user_data);
|
void tcg_tb_foreach(GTraverseFunc func, gpointer user_data);
|
||||||
size_t tcg_nb_tbs(void);
|
size_t tcg_nb_tbs(void);
|
||||||
|
|
Loading…
Reference in New Issue