mirror of https://github.com/xemu-project/xemu.git
memory: move bitmap synchronization to its own function
We want to have all the functions that handle directly the dirty bitmap near. We will change it later. Signed-off-by: Juan Quintela <quintela@redhat.com> Reviewed-by: Orit Wasserman <owasserm@redhat.com>
This commit is contained in:
parent
c9dd46fc0d
commit
5ff7fb77b3
|
@ -79,6 +79,37 @@ static inline void cpu_physical_memory_set_dirty_range(ram_addr_t start,
|
||||||
xen_modified_memory(start, length);
|
xen_modified_memory(start, length);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static inline void cpu_physical_memory_set_dirty_lebitmap(unsigned long *bitmap,
|
||||||
|
ram_addr_t start,
|
||||||
|
ram_addr_t pages)
|
||||||
|
{
|
||||||
|
unsigned int i, j;
|
||||||
|
unsigned long page_number, c;
|
||||||
|
hwaddr addr;
|
||||||
|
ram_addr_t ram_addr;
|
||||||
|
unsigned int len = (pages + HOST_LONG_BITS - 1) / HOST_LONG_BITS;
|
||||||
|
unsigned long hpratio = getpagesize() / TARGET_PAGE_SIZE;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* bitmap-traveling is faster than memory-traveling (for addr...)
|
||||||
|
* especially when most of the memory is not dirty.
|
||||||
|
*/
|
||||||
|
for (i = 0; i < len; i++) {
|
||||||
|
if (bitmap[i] != 0) {
|
||||||
|
c = leul_to_cpu(bitmap[i]);
|
||||||
|
do {
|
||||||
|
j = ffsl(c) - 1;
|
||||||
|
c &= ~(1ul << j);
|
||||||
|
page_number = (i * HOST_LONG_BITS + j) * hpratio;
|
||||||
|
addr = page_number * TARGET_PAGE_SIZE;
|
||||||
|
ram_addr = start + addr;
|
||||||
|
cpu_physical_memory_set_dirty_range(ram_addr,
|
||||||
|
TARGET_PAGE_SIZE * hpratio);
|
||||||
|
} while (c != 0);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
static inline void cpu_physical_memory_clear_dirty_range(ram_addr_t start,
|
static inline void cpu_physical_memory_clear_dirty_range(ram_addr_t start,
|
||||||
ram_addr_t length,
|
ram_addr_t length,
|
||||||
unsigned client)
|
unsigned client)
|
||||||
|
|
27
kvm-all.c
27
kvm-all.c
|
@ -380,33 +380,10 @@ static int kvm_set_migration_log(int enable)
|
||||||
static int kvm_get_dirty_pages_log_range(MemoryRegionSection *section,
|
static int kvm_get_dirty_pages_log_range(MemoryRegionSection *section,
|
||||||
unsigned long *bitmap)
|
unsigned long *bitmap)
|
||||||
{
|
{
|
||||||
unsigned int i, j;
|
|
||||||
unsigned long page_number, c;
|
|
||||||
hwaddr addr;
|
|
||||||
ram_addr_t start = section->offset_within_region + section->mr->ram_addr;
|
ram_addr_t start = section->offset_within_region + section->mr->ram_addr;
|
||||||
ram_addr_t ram_addr;
|
ram_addr_t pages = int128_get64(section->size) / getpagesize();
|
||||||
unsigned int pages = int128_get64(section->size) / getpagesize();
|
|
||||||
unsigned int len = (pages + HOST_LONG_BITS - 1) / HOST_LONG_BITS;
|
|
||||||
unsigned long hpratio = getpagesize() / TARGET_PAGE_SIZE;
|
|
||||||
|
|
||||||
/*
|
cpu_physical_memory_set_dirty_lebitmap(bitmap, start, pages);
|
||||||
* bitmap-traveling is faster than memory-traveling (for addr...)
|
|
||||||
* especially when most of the memory is not dirty.
|
|
||||||
*/
|
|
||||||
for (i = 0; i < len; i++) {
|
|
||||||
if (bitmap[i] != 0) {
|
|
||||||
c = leul_to_cpu(bitmap[i]);
|
|
||||||
do {
|
|
||||||
j = ffsl(c) - 1;
|
|
||||||
c &= ~(1ul << j);
|
|
||||||
page_number = (i * HOST_LONG_BITS + j) * hpratio;
|
|
||||||
addr = page_number * TARGET_PAGE_SIZE;
|
|
||||||
ram_addr = start + addr;
|
|
||||||
cpu_physical_memory_set_dirty_range(ram_addr,
|
|
||||||
TARGET_PAGE_SIZE * hpratio);
|
|
||||||
} while (c != 0);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue