[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[Qemu-devel] [PULL 23/27] migration: cleanup stats update into function
From: |
Juan Quintela |
Subject: |
[Qemu-devel] [PULL 23/27] migration: cleanup stats update into function |
Date: |
Mon, 15 Jan 2018 12:53:05 +0100 |
From: Peter Xu <address@hidden>
We have quite a few lines in migration_thread() that calculates some
statistics for the migration interations. Isolate it into a single
function to improve readability.
Signed-off-by: Peter Xu <address@hidden>
Reviewed-by: Juan Quintela <address@hidden>
Signed-off-by: Juan Quintela <address@hidden>
---
migration/migration.c | 84 ++++++++++++++++++++++++++++++---------------------
migration/migration.h | 11 +++++++
2 files changed, 60 insertions(+), 35 deletions(-)
diff --git a/migration/migration.c b/migration/migration.c
index 6023218ffa..d7f85aa460 100644
--- a/migration/migration.c
+++ b/migration/migration.c
@@ -1303,6 +1303,8 @@ MigrationState *migrate_init(void)
s->start_time = qemu_clock_get_ms(QEMU_CLOCK_REALTIME);
s->total_time = 0;
s->vm_was_running = false;
+ s->iteration_initial_bytes = 0;
+ s->threshold_size = 0;
return s;
}
@@ -2209,6 +2211,43 @@ static void migration_calculate_complete(MigrationState
*s)
}
}
+static void migration_update_counters(MigrationState *s,
+ int64_t current_time)
+{
+ uint64_t transferred, time_spent;
+ int64_t threshold_size;
+ double bandwidth;
+
+ if (current_time < s->iteration_start_time + BUFFER_DELAY) {
+ return;
+ }
+
+ transferred = qemu_ftell(s->to_dst_file) - s->iteration_initial_bytes;
+ time_spent = current_time - s->iteration_start_time;
+ bandwidth = (double)transferred / time_spent;
+ threshold_size = bandwidth * s->parameters.downtime_limit;
+
+ s->mbps = (((double) transferred * 8.0) /
+ ((double) time_spent / 1000.0)) / 1000.0 / 1000.0;
+
+ /*
+ * if we haven't sent anything, we don't want to
+ * recalculate. 10000 is a small enough number for our purposes
+ */
+ if (ram_counters.dirty_pages_rate && transferred > 10000) {
+ s->expected_downtime = ram_counters.dirty_pages_rate *
+ qemu_target_page_size() / bandwidth;
+ }
+
+ qemu_file_reset_rate_limit(s->to_dst_file);
+
+ s->iteration_start_time = current_time;
+ s->iteration_initial_bytes = qemu_ftell(s->to_dst_file);
+
+ trace_migrate_transferred(transferred, time_spent,
+ bandwidth, threshold_size);
+}
+
/*
* Master migration thread on the source VM.
* It drives the migration and pumps the data down the outgoing channel.
@@ -2216,22 +2255,15 @@ static void migration_calculate_complete(MigrationState
*s)
static void *migration_thread(void *opaque)
{
MigrationState *s = opaque;
- /* Used by the bandwidth calcs, updated later */
- int64_t initial_time = qemu_clock_get_ms(QEMU_CLOCK_REALTIME);
int64_t setup_start = qemu_clock_get_ms(QEMU_CLOCK_HOST);
- int64_t initial_bytes = 0;
- /*
- * The final stage happens when the remaining data is smaller than
- * this threshold; it's calculated from the requested downtime and
- * measured bandwidth
- */
- int64_t threshold_size = 0;
bool entered_postcopy = false;
/* The active state we expect to be in; ACTIVE or POSTCOPY_ACTIVE */
enum MigrationStatus current_active_state = MIGRATION_STATUS_ACTIVE;
rcu_register_thread();
+ s->iteration_start_time = qemu_clock_get_ms(QEMU_CLOCK_REALTIME);
+
qemu_savevm_state_header(s->to_dst_file);
/*
@@ -2271,17 +2303,17 @@ static void *migration_thread(void *opaque)
if (!qemu_file_rate_limit(s->to_dst_file)) {
uint64_t pend_post, pend_nonpost;
- qemu_savevm_state_pending(s->to_dst_file, threshold_size,
+ qemu_savevm_state_pending(s->to_dst_file, s->threshold_size,
&pend_nonpost, &pend_post);
pending_size = pend_nonpost + pend_post;
- trace_migrate_pending(pending_size, threshold_size,
+ trace_migrate_pending(pending_size, s->threshold_size,
pend_post, pend_nonpost);
- if (pending_size && pending_size >= threshold_size) {
+ if (pending_size && pending_size >= s->threshold_size) {
/* Still a significant amount to transfer */
if (migrate_postcopy() &&
s->state != MIGRATION_STATUS_POSTCOPY_ACTIVE &&
- pend_nonpost <= threshold_size &&
+ pend_nonpost <= s->threshold_size &&
atomic_read(&s->start_postcopy)) {
if (!postcopy_start(s)) {
@@ -2306,33 +2338,15 @@ static void *migration_thread(void *opaque)
trace_migration_thread_file_err();
break;
}
+
current_time = qemu_clock_get_ms(QEMU_CLOCK_REALTIME);
- if (current_time >= initial_time + BUFFER_DELAY) {
- uint64_t transferred_bytes = qemu_ftell(s->to_dst_file) -
- initial_bytes;
- uint64_t time_spent = current_time - initial_time;
- double bandwidth = (double)transferred_bytes / time_spent;
- threshold_size = bandwidth * s->parameters.downtime_limit;
- s->mbps = (((double) transferred_bytes * 8.0) /
- ((double) time_spent / 1000.0)) / 1000.0 / 1000.0;
+ migration_update_counters(s, current_time);
- trace_migrate_transferred(transferred_bytes, time_spent,
- bandwidth, threshold_size);
- /* if we haven't sent anything, we don't want to recalculate
- 10000 is a small enough number for our purposes */
- if (ram_counters.dirty_pages_rate && transferred_bytes > 10000) {
- s->expected_downtime = ram_counters.dirty_pages_rate *
- qemu_target_page_size() / bandwidth;
- }
-
- qemu_file_reset_rate_limit(s->to_dst_file);
- initial_time = current_time;
- initial_bytes = qemu_ftell(s->to_dst_file);
- }
if (qemu_file_rate_limit(s->to_dst_file)) {
/* usleep expects microseconds */
- g_usleep((initial_time + BUFFER_DELAY - current_time)*1000);
+ g_usleep((s->iteration_start_time + BUFFER_DELAY -
+ current_time) * 1000);
}
}
diff --git a/migration/migration.h b/migration/migration.h
index 0aad12f9a1..f2bc1aaf85 100644
--- a/migration/migration.h
+++ b/migration/migration.h
@@ -102,6 +102,17 @@ struct MigrationState
QEMUBH *cleanup_bh;
QEMUFile *to_dst_file;
+ /* bytes already send at the beggining of current interation */
+ uint64_t iteration_initial_bytes;
+ /* time at the start of current iteration */
+ int64_t iteration_start_time;
+ /*
+ * The final stage happens when the remaining data is smaller than
+ * this threshold; it's calculated from the requested downtime and
+ * measured bandwidth
+ */
+ int64_t threshold_size;
+
/* params from 'migrate-set-parameters' */
MigrationParameters parameters;
--
2.14.3
- [Qemu-devel] [PULL 12/27] migration: add postcopy total blocktime into query-migrate, (continued)
- [Qemu-devel] [PULL 12/27] migration: add postcopy total blocktime into query-migrate, Juan Quintela, 2018/01/15
- [Qemu-devel] [PULL 13/27] migration: Guard ram_bytes_remaining against early call, Juan Quintela, 2018/01/15
- [Qemu-devel] [PULL 14/27] migration: finalize current_migration object, Juan Quintela, 2018/01/15
- [Qemu-devel] [PULL 16/27] migration: qemu_savevm_state_cleanup() in cleanup, Juan Quintela, 2018/01/15
- [Qemu-devel] [PULL 15/27] migration: assert colo instead of check, Juan Quintela, 2018/01/15
- [Qemu-devel] [PULL 17/27] migration: remove "enable_colo" var, Juan Quintela, 2018/01/15
- [Qemu-devel] [PULL 18/27] migration: split use of MigrationState.total_time, Juan Quintela, 2018/01/15
- [Qemu-devel] [PULL 19/27] migration: move vm_old_running into global state, Juan Quintela, 2018/01/15
- [Qemu-devel] [PULL 20/27] migration: introduce downtime_start, Juan Quintela, 2018/01/15
- [Qemu-devel] [PULL 21/27] migration: introduce migrate_calculate_complete, Juan Quintela, 2018/01/15
- [Qemu-devel] [PULL 23/27] migration: cleanup stats update into function,
Juan Quintela <=
- [Qemu-devel] [PULL 22/27] migration: use switch at the end of migration, Juan Quintela, 2018/01/15
- [Qemu-devel] [PULL 24/27] migration: major cleanup for migrate iterations, Juan Quintela, 2018/01/15
- [Qemu-devel] [PULL 26/27] migration: remove some block_cleanup_parameters(), Juan Quintela, 2018/01/15
- [Qemu-devel] [PULL 25/27] migration: put the finish part into a new function, Juan Quintela, 2018/01/15
- [Qemu-devel] [PULL 27/27] migration: remove notify in fd_error, Juan Quintela, 2018/01/15
- Re: [Qemu-devel] [PULL 00/27] Migration pull, Peter Maydell, 2018/01/15