mirror of https://github.com/xqemu/xqemu.git
ram: Make RAMState dynamic
We create the variable while we are at migration and we remove it after migration. Signed-off-by: Juan Quintela <quintela@redhat.com> Reviewed-by: Dr. David Alan Gilbert <dgilbert@redhat.com>
This commit is contained in:
parent
9360447d34
commit
53518d9448
|
@ -204,11 +204,11 @@ struct RAMState {
|
||||||
};
|
};
|
||||||
typedef struct RAMState RAMState;
|
typedef struct RAMState RAMState;
|
||||||
|
|
||||||
static RAMState ram_state;
|
static RAMState *ram_state;
|
||||||
|
|
||||||
uint64_t ram_bytes_remaining(void)
|
uint64_t ram_bytes_remaining(void)
|
||||||
{
|
{
|
||||||
return ram_state.migration_dirty_pages * TARGET_PAGE_SIZE;
|
return ram_state->migration_dirty_pages * TARGET_PAGE_SIZE;
|
||||||
}
|
}
|
||||||
|
|
||||||
MigrationStats ram_counters;
|
MigrationStats ram_counters;
|
||||||
|
@ -793,7 +793,7 @@ static int ram_save_page(RAMState *rs, PageSearchStatus *pss, bool last_stage)
|
||||||
static int do_compress_ram_page(QEMUFile *f, RAMBlock *block,
|
static int do_compress_ram_page(QEMUFile *f, RAMBlock *block,
|
||||||
ram_addr_t offset)
|
ram_addr_t offset)
|
||||||
{
|
{
|
||||||
RAMState *rs = &ram_state;
|
RAMState *rs = ram_state;
|
||||||
int bytes_sent, blen;
|
int bytes_sent, blen;
|
||||||
uint8_t *p = block->host + (offset & TARGET_PAGE_MASK);
|
uint8_t *p = block->host + (offset & TARGET_PAGE_MASK);
|
||||||
|
|
||||||
|
@ -1140,7 +1140,7 @@ static void migration_page_queue_free(RAMState *rs)
|
||||||
int ram_save_queue_pages(const char *rbname, ram_addr_t start, ram_addr_t len)
|
int ram_save_queue_pages(const char *rbname, ram_addr_t start, ram_addr_t len)
|
||||||
{
|
{
|
||||||
RAMBlock *ramblock;
|
RAMBlock *ramblock;
|
||||||
RAMState *rs = &ram_state;
|
RAMState *rs = ram_state;
|
||||||
|
|
||||||
ram_counters.postcopy_requests++;
|
ram_counters.postcopy_requests++;
|
||||||
rcu_read_lock();
|
rcu_read_lock();
|
||||||
|
@ -1361,7 +1361,7 @@ void free_xbzrle_decoded_buf(void)
|
||||||
|
|
||||||
static void ram_migration_cleanup(void *opaque)
|
static void ram_migration_cleanup(void *opaque)
|
||||||
{
|
{
|
||||||
RAMState *rs = opaque;
|
RAMState **rsp = opaque;
|
||||||
RAMBlock *block;
|
RAMBlock *block;
|
||||||
|
|
||||||
/* caller have hold iothread lock or is in a bh, so there is
|
/* caller have hold iothread lock or is in a bh, so there is
|
||||||
|
@ -1388,7 +1388,9 @@ static void ram_migration_cleanup(void *opaque)
|
||||||
XBZRLE.zero_target_page = NULL;
|
XBZRLE.zero_target_page = NULL;
|
||||||
}
|
}
|
||||||
XBZRLE_cache_unlock();
|
XBZRLE_cache_unlock();
|
||||||
migration_page_queue_free(rs);
|
migration_page_queue_free(*rsp);
|
||||||
|
g_free(*rsp);
|
||||||
|
*rsp = NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void ram_state_reset(RAMState *rs)
|
static void ram_state_reset(RAMState *rs)
|
||||||
|
@ -1559,7 +1561,7 @@ static void postcopy_chunk_hostpages_pass(MigrationState *ms, bool unsent_pass,
|
||||||
RAMBlock *block,
|
RAMBlock *block,
|
||||||
PostcopyDiscardState *pds)
|
PostcopyDiscardState *pds)
|
||||||
{
|
{
|
||||||
RAMState *rs = &ram_state;
|
RAMState *rs = ram_state;
|
||||||
unsigned long *bitmap = block->bmap;
|
unsigned long *bitmap = block->bmap;
|
||||||
unsigned long *unsentmap = block->unsentmap;
|
unsigned long *unsentmap = block->unsentmap;
|
||||||
unsigned int host_ratio = block->page_size / TARGET_PAGE_SIZE;
|
unsigned int host_ratio = block->page_size / TARGET_PAGE_SIZE;
|
||||||
|
@ -1714,7 +1716,7 @@ static int postcopy_chunk_hostpages(MigrationState *ms, RAMBlock *block)
|
||||||
*/
|
*/
|
||||||
int ram_postcopy_send_discard_bitmap(MigrationState *ms)
|
int ram_postcopy_send_discard_bitmap(MigrationState *ms)
|
||||||
{
|
{
|
||||||
RAMState *rs = &ram_state;
|
RAMState *rs = ram_state;
|
||||||
RAMBlock *block;
|
RAMBlock *block;
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
|
@ -1797,12 +1799,13 @@ err:
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int ram_state_init(RAMState *rs)
|
static int ram_state_init(RAMState **rsp)
|
||||||
{
|
{
|
||||||
memset(rs, 0, sizeof(*rs));
|
*rsp = g_new0(RAMState, 1);
|
||||||
qemu_mutex_init(&rs->bitmap_mutex);
|
|
||||||
qemu_mutex_init(&rs->src_page_req_mutex);
|
qemu_mutex_init(&(*rsp)->bitmap_mutex);
|
||||||
QSIMPLEQ_INIT(&rs->src_page_requests);
|
qemu_mutex_init(&(*rsp)->src_page_req_mutex);
|
||||||
|
QSIMPLEQ_INIT(&(*rsp)->src_page_requests);
|
||||||
|
|
||||||
if (migrate_use_xbzrle()) {
|
if (migrate_use_xbzrle()) {
|
||||||
XBZRLE_cache_lock();
|
XBZRLE_cache_lock();
|
||||||
|
@ -1813,6 +1816,8 @@ static int ram_state_init(RAMState *rs)
|
||||||
if (!XBZRLE.cache) {
|
if (!XBZRLE.cache) {
|
||||||
XBZRLE_cache_unlock();
|
XBZRLE_cache_unlock();
|
||||||
error_report("Error creating cache");
|
error_report("Error creating cache");
|
||||||
|
g_free(*rsp);
|
||||||
|
*rsp = NULL;
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
XBZRLE_cache_unlock();
|
XBZRLE_cache_unlock();
|
||||||
|
@ -1821,6 +1826,8 @@ static int ram_state_init(RAMState *rs)
|
||||||
XBZRLE.encoded_buf = g_try_malloc0(TARGET_PAGE_SIZE);
|
XBZRLE.encoded_buf = g_try_malloc0(TARGET_PAGE_SIZE);
|
||||||
if (!XBZRLE.encoded_buf) {
|
if (!XBZRLE.encoded_buf) {
|
||||||
error_report("Error allocating encoded_buf");
|
error_report("Error allocating encoded_buf");
|
||||||
|
g_free(*rsp);
|
||||||
|
*rsp = NULL;
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1829,6 +1836,8 @@ static int ram_state_init(RAMState *rs)
|
||||||
error_report("Error allocating current_buf");
|
error_report("Error allocating current_buf");
|
||||||
g_free(XBZRLE.encoded_buf);
|
g_free(XBZRLE.encoded_buf);
|
||||||
XBZRLE.encoded_buf = NULL;
|
XBZRLE.encoded_buf = NULL;
|
||||||
|
g_free(*rsp);
|
||||||
|
*rsp = NULL;
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1838,7 +1847,7 @@ static int ram_state_init(RAMState *rs)
|
||||||
|
|
||||||
qemu_mutex_lock_ramlist();
|
qemu_mutex_lock_ramlist();
|
||||||
rcu_read_lock();
|
rcu_read_lock();
|
||||||
ram_state_reset(rs);
|
ram_state_reset(*rsp);
|
||||||
|
|
||||||
/* Skip setting bitmap if there is no RAM */
|
/* Skip setting bitmap if there is no RAM */
|
||||||
if (ram_bytes_total()) {
|
if (ram_bytes_total()) {
|
||||||
|
@ -1860,10 +1869,10 @@ static int ram_state_init(RAMState *rs)
|
||||||
* Count the total number of pages used by ram blocks not including any
|
* Count the total number of pages used by ram blocks not including any
|
||||||
* gaps due to alignment or unplugs.
|
* gaps due to alignment or unplugs.
|
||||||
*/
|
*/
|
||||||
rs->migration_dirty_pages = ram_bytes_total() >> TARGET_PAGE_BITS;
|
(*rsp)->migration_dirty_pages = ram_bytes_total() >> TARGET_PAGE_BITS;
|
||||||
|
|
||||||
memory_global_dirty_log_start();
|
memory_global_dirty_log_start();
|
||||||
migration_bitmap_sync(rs);
|
migration_bitmap_sync(*rsp);
|
||||||
qemu_mutex_unlock_ramlist();
|
qemu_mutex_unlock_ramlist();
|
||||||
qemu_mutex_unlock_iothread();
|
qemu_mutex_unlock_iothread();
|
||||||
rcu_read_unlock();
|
rcu_read_unlock();
|
||||||
|
@ -1888,16 +1897,16 @@ static int ram_state_init(RAMState *rs)
|
||||||
*/
|
*/
|
||||||
static int ram_save_setup(QEMUFile *f, void *opaque)
|
static int ram_save_setup(QEMUFile *f, void *opaque)
|
||||||
{
|
{
|
||||||
RAMState *rs = opaque;
|
RAMState **rsp = opaque;
|
||||||
RAMBlock *block;
|
RAMBlock *block;
|
||||||
|
|
||||||
/* migration has already setup the bitmap, reuse it. */
|
/* migration has already setup the bitmap, reuse it. */
|
||||||
if (!migration_in_colo_state()) {
|
if (!migration_in_colo_state()) {
|
||||||
if (ram_state_init(rs) < 0) {
|
if (ram_state_init(rsp) != 0) {
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
rs->f = f;
|
(*rsp)->f = f;
|
||||||
|
|
||||||
rcu_read_lock();
|
rcu_read_lock();
|
||||||
|
|
||||||
|
@ -1932,7 +1941,8 @@ static int ram_save_setup(QEMUFile *f, void *opaque)
|
||||||
*/
|
*/
|
||||||
static int ram_save_iterate(QEMUFile *f, void *opaque)
|
static int ram_save_iterate(QEMUFile *f, void *opaque)
|
||||||
{
|
{
|
||||||
RAMState *rs = opaque;
|
RAMState **temp = opaque;
|
||||||
|
RAMState *rs = *temp;
|
||||||
int ret;
|
int ret;
|
||||||
int i;
|
int i;
|
||||||
int64_t t0;
|
int64_t t0;
|
||||||
|
@ -2007,7 +2017,8 @@ static int ram_save_iterate(QEMUFile *f, void *opaque)
|
||||||
*/
|
*/
|
||||||
static int ram_save_complete(QEMUFile *f, void *opaque)
|
static int ram_save_complete(QEMUFile *f, void *opaque)
|
||||||
{
|
{
|
||||||
RAMState *rs = opaque;
|
RAMState **temp = opaque;
|
||||||
|
RAMState *rs = *temp;
|
||||||
|
|
||||||
rcu_read_lock();
|
rcu_read_lock();
|
||||||
|
|
||||||
|
@ -2044,7 +2055,8 @@ static void ram_save_pending(QEMUFile *f, void *opaque, uint64_t max_size,
|
||||||
uint64_t *non_postcopiable_pending,
|
uint64_t *non_postcopiable_pending,
|
||||||
uint64_t *postcopiable_pending)
|
uint64_t *postcopiable_pending)
|
||||||
{
|
{
|
||||||
RAMState *rs = opaque;
|
RAMState **temp = opaque;
|
||||||
|
RAMState *rs = *temp;
|
||||||
uint64_t remaining_size;
|
uint64_t remaining_size;
|
||||||
|
|
||||||
remaining_size = rs->migration_dirty_pages * TARGET_PAGE_SIZE;
|
remaining_size = rs->migration_dirty_pages * TARGET_PAGE_SIZE;
|
||||||
|
|
Loading…
Reference in New Issue