migration: Merge ram_counters and ram_atomic_counters

Using MgrationStats as type for ram_counters mean that we didn't have
to re-declare each value in another struct. The need of atomic
counters have make us to create MigrationAtomicStats for this atomic
counters.

Create RAMStats type which is a merge of MigrationStats and
MigrationAtomicStats removing unused members.

Signed-off-by: Juan Quintela <quintela@redhat.com>
Reviewed-by: Peter Xu <peterx@redhat.com>

---

Fix typos found by David Edmondson
This commit is contained in:
Juan Quintela 2023-02-22 15:56:45 +01:00
parent 8ebb6ecc37
commit abce5fa16d
4 changed files with 37 additions and 42 deletions

View File

@ -1140,12 +1140,12 @@ static void populate_ram_info(MigrationInfo *info, MigrationState *s)
size_t page_size = qemu_target_page_size(); size_t page_size = qemu_target_page_size();
info->ram = g_malloc0(sizeof(*info->ram)); info->ram = g_malloc0(sizeof(*info->ram));
info->ram->transferred = stat64_get(&ram_atomic_counters.transferred); info->ram->transferred = stat64_get(&ram_counters.transferred);
info->ram->total = ram_bytes_total(); info->ram->total = ram_bytes_total();
info->ram->duplicate = stat64_get(&ram_atomic_counters.duplicate); info->ram->duplicate = stat64_get(&ram_counters.duplicate);
/* legacy value. It is not used anymore */ /* legacy value. It is not used anymore */
info->ram->skipped = 0; info->ram->skipped = 0;
info->ram->normal = stat64_get(&ram_atomic_counters.normal); info->ram->normal = stat64_get(&ram_counters.normal);
info->ram->normal_bytes = info->ram->normal * page_size; info->ram->normal_bytes = info->ram->normal * page_size;
info->ram->mbps = s->mbps; info->ram->mbps = s->mbps;
info->ram->dirty_sync_count = ram_counters.dirty_sync_count; info->ram->dirty_sync_count = ram_counters.dirty_sync_count;
@ -1157,7 +1157,7 @@ static void populate_ram_info(MigrationInfo *info, MigrationState *s)
info->ram->pages_per_second = s->pages_per_second; info->ram->pages_per_second = s->pages_per_second;
info->ram->precopy_bytes = ram_counters.precopy_bytes; info->ram->precopy_bytes = ram_counters.precopy_bytes;
info->ram->downtime_bytes = ram_counters.downtime_bytes; info->ram->downtime_bytes = ram_counters.downtime_bytes;
info->ram->postcopy_bytes = stat64_get(&ram_atomic_counters.postcopy_bytes); info->ram->postcopy_bytes = stat64_get(&ram_counters.postcopy_bytes);
if (migrate_use_xbzrle()) { if (migrate_use_xbzrle()) {
info->xbzrle_cache = g_malloc0(sizeof(*info->xbzrle_cache)); info->xbzrle_cache = g_malloc0(sizeof(*info->xbzrle_cache));

View File

@ -433,7 +433,7 @@ static int multifd_send_pages(QEMUFile *f)
transferred = ((uint64_t) pages->num) * p->page_size + p->packet_len; transferred = ((uint64_t) pages->num) * p->page_size + p->packet_len;
qemu_file_acct_rate_limit(f, transferred); qemu_file_acct_rate_limit(f, transferred);
ram_counters.multifd_bytes += transferred; ram_counters.multifd_bytes += transferred;
stat64_add(&ram_atomic_counters.transferred, transferred); stat64_add(&ram_counters.transferred, transferred);
qemu_mutex_unlock(&p->mutex); qemu_mutex_unlock(&p->mutex);
qemu_sem_post(&p->sem); qemu_sem_post(&p->sem);
@ -628,7 +628,7 @@ int multifd_send_sync_main(QEMUFile *f)
p->pending_job++; p->pending_job++;
qemu_file_acct_rate_limit(f, p->packet_len); qemu_file_acct_rate_limit(f, p->packet_len);
ram_counters.multifd_bytes += p->packet_len; ram_counters.multifd_bytes += p->packet_len;
stat64_add(&ram_atomic_counters.transferred, p->packet_len); stat64_add(&ram_counters.transferred, p->packet_len);
qemu_mutex_unlock(&p->mutex); qemu_mutex_unlock(&p->mutex);
qemu_sem_post(&p->sem); qemu_sem_post(&p->sem);
} }

View File

@ -458,25 +458,18 @@ uint64_t ram_bytes_remaining(void)
0; 0;
} }
/* RAMStats ram_counters;
* NOTE: not all stats in ram_counters are used in reality. See comments
* for struct MigrationAtomicStats. The ultimate result of ram migration
* counters will be a merged version with both ram_counters and the atomic
* fields in ram_atomic_counters.
*/
MigrationStats ram_counters;
MigrationAtomicStats ram_atomic_counters;
void ram_transferred_add(uint64_t bytes) void ram_transferred_add(uint64_t bytes)
{ {
if (runstate_is_running()) { if (runstate_is_running()) {
ram_counters.precopy_bytes += bytes; ram_counters.precopy_bytes += bytes;
} else if (migration_in_postcopy()) { } else if (migration_in_postcopy()) {
stat64_add(&ram_atomic_counters.postcopy_bytes, bytes); stat64_add(&ram_counters.postcopy_bytes, bytes);
} else { } else {
ram_counters.downtime_bytes += bytes; ram_counters.downtime_bytes += bytes;
} }
stat64_add(&ram_atomic_counters.transferred, bytes); stat64_add(&ram_counters.transferred, bytes);
} }
void dirty_sync_missed_zero_copy(void) void dirty_sync_missed_zero_copy(void)
@ -756,7 +749,7 @@ void mig_throttle_counter_reset(void)
rs->time_last_bitmap_sync = qemu_clock_get_ms(QEMU_CLOCK_REALTIME); rs->time_last_bitmap_sync = qemu_clock_get_ms(QEMU_CLOCK_REALTIME);
rs->num_dirty_pages_period = 0; rs->num_dirty_pages_period = 0;
rs->bytes_xfer_prev = stat64_get(&ram_atomic_counters.transferred); rs->bytes_xfer_prev = stat64_get(&ram_counters.transferred);
} }
/** /**
@ -1130,8 +1123,8 @@ uint64_t ram_pagesize_summary(void)
uint64_t ram_get_total_transferred_pages(void) uint64_t ram_get_total_transferred_pages(void)
{ {
return stat64_get(&ram_atomic_counters.normal) + return stat64_get(&ram_counters.normal) +
stat64_get(&ram_atomic_counters.duplicate) + stat64_get(&ram_counters.duplicate) +
compression_counters.pages + xbzrle_counters.pages; compression_counters.pages + xbzrle_counters.pages;
} }
@ -1192,7 +1185,7 @@ static void migration_trigger_throttle(RAMState *rs)
MigrationState *s = migrate_get_current(); MigrationState *s = migrate_get_current();
uint64_t threshold = s->parameters.throttle_trigger_threshold; uint64_t threshold = s->parameters.throttle_trigger_threshold;
uint64_t bytes_xfer_period = uint64_t bytes_xfer_period =
stat64_get(&ram_atomic_counters.transferred) - rs->bytes_xfer_prev; stat64_get(&ram_counters.transferred) - rs->bytes_xfer_prev;
uint64_t bytes_dirty_period = rs->num_dirty_pages_period * TARGET_PAGE_SIZE; uint64_t bytes_dirty_period = rs->num_dirty_pages_period * TARGET_PAGE_SIZE;
uint64_t bytes_dirty_threshold = bytes_xfer_period * threshold / 100; uint64_t bytes_dirty_threshold = bytes_xfer_period * threshold / 100;
@ -1255,7 +1248,7 @@ static void migration_bitmap_sync(RAMState *rs)
/* reset period counters */ /* reset period counters */
rs->time_last_bitmap_sync = end_time; rs->time_last_bitmap_sync = end_time;
rs->num_dirty_pages_period = 0; rs->num_dirty_pages_period = 0;
rs->bytes_xfer_prev = stat64_get(&ram_atomic_counters.transferred); rs->bytes_xfer_prev = stat64_get(&ram_counters.transferred);
} }
if (migrate_use_events()) { if (migrate_use_events()) {
qapi_event_send_migration_pass(ram_counters.dirty_sync_count); qapi_event_send_migration_pass(ram_counters.dirty_sync_count);
@ -1331,7 +1324,7 @@ static int save_zero_page(PageSearchStatus *pss, QEMUFile *f, RAMBlock *block,
int len = save_zero_page_to_file(pss, f, block, offset); int len = save_zero_page_to_file(pss, f, block, offset);
if (len) { if (len) {
stat64_add(&ram_atomic_counters.duplicate, 1); stat64_add(&ram_counters.duplicate, 1);
ram_transferred_add(len); ram_transferred_add(len);
return 1; return 1;
} }
@ -1368,9 +1361,9 @@ static bool control_save_page(PageSearchStatus *pss, RAMBlock *block,
} }
if (bytes_xmit > 0) { if (bytes_xmit > 0) {
stat64_add(&ram_atomic_counters.normal, 1); stat64_add(&ram_counters.normal, 1);
} else if (bytes_xmit == 0) { } else if (bytes_xmit == 0) {
stat64_add(&ram_atomic_counters.duplicate, 1); stat64_add(&ram_counters.duplicate, 1);
} }
return true; return true;
@ -1402,7 +1395,7 @@ static int save_normal_page(PageSearchStatus *pss, RAMBlock *block,
qemu_put_buffer(file, buf, TARGET_PAGE_SIZE); qemu_put_buffer(file, buf, TARGET_PAGE_SIZE);
} }
ram_transferred_add(TARGET_PAGE_SIZE); ram_transferred_add(TARGET_PAGE_SIZE);
stat64_add(&ram_atomic_counters.normal, 1); stat64_add(&ram_counters.normal, 1);
return 1; return 1;
} }
@ -1458,7 +1451,7 @@ static int ram_save_multifd_page(QEMUFile *file, RAMBlock *block,
if (multifd_queue_page(file, block, offset) < 0) { if (multifd_queue_page(file, block, offset) < 0) {
return -1; return -1;
} }
stat64_add(&ram_atomic_counters.normal, 1); stat64_add(&ram_counters.normal, 1);
return 1; return 1;
} }
@ -1497,7 +1490,7 @@ update_compress_thread_counts(const CompressParam *param, int bytes_xmit)
ram_transferred_add(bytes_xmit); ram_transferred_add(bytes_xmit);
if (param->zero_page) { if (param->zero_page) {
stat64_add(&ram_atomic_counters.duplicate, 1); stat64_add(&ram_counters.duplicate, 1);
return; return;
} }
@ -2632,9 +2625,9 @@ void acct_update_position(QEMUFile *f, size_t size, bool zero)
uint64_t pages = size / TARGET_PAGE_SIZE; uint64_t pages = size / TARGET_PAGE_SIZE;
if (zero) { if (zero) {
stat64_add(&ram_atomic_counters.duplicate, pages); stat64_add(&ram_counters.duplicate, pages);
} else { } else {
stat64_add(&ram_atomic_counters.normal, pages); stat64_add(&ram_counters.normal, pages);
ram_transferred_add(size); ram_transferred_add(size);
qemu_file_credit_transfer(f, size); qemu_file_credit_transfer(f, size);
} }

View File

@ -35,25 +35,27 @@
#include "qemu/stats64.h" #include "qemu/stats64.h"
/* /*
* These are the migration statistic counters that need to be updated using * These are the ram migration statistic counters. It is loosely
* atomic ops (can be accessed by more than one thread). Here since we * based on MigrationStats. We change to Stat64 any counter that
* cannot modify MigrationStats directly to use Stat64 as it was defined in * needs to be updated using atomic ops (can be accessed by more than
* the QAPI scheme, we define an internal structure to hold them, and we * one thread).
* propagate the real values when QMP queries happen.
*
* IOW, the corresponding fields within ram_counters on these specific
* fields will be always zero and not being used at all; they're just
* placeholders to make it QAPI-compatible.
*/ */
typedef struct { typedef struct {
Stat64 transferred; int64_t dirty_pages_rate;
int64_t dirty_sync_count;
uint64_t dirty_sync_missed_zero_copy;
uint64_t downtime_bytes;
Stat64 duplicate; Stat64 duplicate;
uint64_t multifd_bytes;
Stat64 normal; Stat64 normal;
Stat64 postcopy_bytes; Stat64 postcopy_bytes;
} MigrationAtomicStats; int64_t postcopy_requests;
uint64_t precopy_bytes;
int64_t remaining;
Stat64 transferred;
} RAMStats;
extern MigrationAtomicStats ram_atomic_counters; extern RAMStats ram_counters;
extern MigrationStats ram_counters;
extern XBZRLECacheStats xbzrle_counters; extern XBZRLECacheStats xbzrle_counters;
extern CompressionStats compression_counters; extern CompressionStats compression_counters;