When VM is configured with huge memory, the current throttle logic
doesn't look like to scale, because migration_trigger_throttle()
is only called for each iteration, so it won't be invoked for a long
time if one iteration can take a long time.
The background sync and throttle aim to fix the above issue by
synchronizing the remote dirty bitmap and triggering the throttle
once detect that iteration lasts a long time.
This is a trade-off between synchronization overhead and CPU throttle
impact.
Signed-off-by: Hyman Huang <yong.huang@smartx.com>
---
migration/migration.c | 12 +++++++++++
tests/qtest/migration-test.c | 39 ++++++++++++++++++++++++++++++++++++
2 files changed, 51 insertions(+)
diff --git a/migration/migration.c b/migration/migration.c
index 055d527ff6..af8b22fa15 100644
--- a/migration/migration.c
+++ b/migration/migration.c
@@ -1416,6 +1416,7 @@ static void migrate_fd_cleanup(MigrationState *s)
trace_migrate_fd_cleanup();
bql_unlock();
+ migration_background_sync_cleanup();
if (s->migration_thread_running) {
qemu_thread_join(&s->thread);
s->migration_thread_running = false;
@@ -3263,6 +3264,7 @@ static MigIterateState migration_iteration_run(MigrationState *s)
if ((!pending_size || pending_size < s->threshold_size) && can_switchover) {
trace_migration_thread_low_pending(pending_size);
+ migration_background_sync_cleanup();
migration_completion(s);
return MIG_ITERATE_BREAK;
}
@@ -3508,6 +3510,16 @@ static void *migration_thread(void *opaque)
ret = qemu_savevm_state_setup(s->to_dst_file, &local_err);
bql_unlock();
+ if (!migrate_dirty_limit()) {
+ /*
+ * Initiate the background sync watcher in order to guarantee
+ * that the CPU throttling acts appropriately. Dirty Limit
+ * doesn't use CPU throttle to make guest down, so ignore that
+ * case.
+ */
+ migration_background_sync_setup();
+ }
+
qemu_savevm_wait_unplug(s, MIGRATION_STATUS_SETUP,
MIGRATION_STATUS_ACTIVE);
diff --git a/tests/qtest/migration-test.c b/tests/qtest/migration-test.c
index b796a90cad..e0e94d26be 100644
--- a/tests/qtest/migration-test.c
+++ b/tests/qtest/migration-test.c
@@ -281,6 +281,11 @@ static uint64_t get_migration_pass(QTestState *who)
return read_ram_property_int(who, "iteration-count");
}
+static uint64_t get_dirty_sync_count(QTestState *who)
+{
+ return read_ram_property_int(who, "dirty-sync-count");
+}
+
static void read_blocktime(QTestState *who)
{
QDict *rsp_return;
@@ -468,6 +473,12 @@ static void migrate_ensure_converge(QTestState *who)
migrate_set_parameter_int(who, "downtime-limit", 30 * 1000);
}
+static void migrate_ensure_iteration_last_long(QTestState *who)
+{
+ /* Set 10Byte/s bandwidth limit to make the iteration last long enough */
+ migrate_set_parameter_int(who, "max-bandwidth", 10);
+}
+
/*
* Our goal is to ensure that we run a single full migration
* iteration, and also dirty memory, ensuring that at least
@@ -2791,6 +2802,10 @@ static void test_migrate_auto_converge(void)
* so we need to decrease a bandwidth.
*/
const int64_t init_pct = 5, inc_pct = 25, max_pct = 95;
+ uint64_t prev_iter_cnt = 0, iter_cnt;
+ uint64_t iter_cnt_changes = 0;
+ uint64_t prev_dirty_sync_cnt = 0, dirty_sync_cnt;
+ uint64_t dirty_sync_cnt_changes = 0;
if (test_migrate_start(&from, &to, uri, &args)) {
return;
@@ -2827,6 +2842,30 @@ static void test_migrate_auto_converge(void)
} while (true);
/* The first percentage of throttling should be at least init_pct */
g_assert_cmpint(percentage, >=, init_pct);
+
+ /* Make sure the iteration take a long time enough */
+ migrate_ensure_iteration_last_long(from);
+
+ /*
+ * End the loop when the dirty sync count or iteration count changes.
+ */
+ while (iter_cnt_changes < 2 && dirty_sync_cnt_changes < 2) {
+ usleep(1000 * 1000);
+ iter_cnt = get_migration_pass(from);
+ iter_cnt_changes += (iter_cnt != prev_iter_cnt);
+ prev_iter_cnt = iter_cnt;
+
+ dirty_sync_cnt = get_dirty_sync_count(from);
+ dirty_sync_cnt_changes += (dirty_sync_cnt != prev_dirty_sync_cnt);
+ prev_dirty_sync_cnt = dirty_sync_cnt;
+ }
+
+ /*
+ * The dirty sync count must have changed because we are in the same
+ * iteration.
+ */
+ g_assert_cmpint(iter_cnt_changes , < , dirty_sync_cnt_changes);
+
/* Now, when we tested that throttling works, let it converge */
migrate_ensure_converge(from);
--
2.39.1
Hyman Huang <yong.huang@smartx.com> writes: > When VM is configured with huge memory, the current throttle logic > doesn't look like to scale, because migration_trigger_throttle() > is only called for each iteration, so it won't be invoked for a long > time if one iteration can take a long time. > > The background sync and throttle aim to fix the above issue by > synchronizing the remote dirty bitmap and triggering the throttle > once detect that iteration lasts a long time. > > This is a trade-off between synchronization overhead and CPU throttle > impact. > > Signed-off-by: Hyman Huang <yong.huang@smartx.com> > --- > migration/migration.c | 12 +++++++++++ > tests/qtest/migration-test.c | 39 ++++++++++++++++++++++++++++++++++++ > 2 files changed, 51 insertions(+) > > diff --git a/migration/migration.c b/migration/migration.c > index 055d527ff6..af8b22fa15 100644 > --- a/migration/migration.c > +++ b/migration/migration.c > @@ -1416,6 +1416,7 @@ static void migrate_fd_cleanup(MigrationState *s) > > trace_migrate_fd_cleanup(); > bql_unlock(); > + migration_background_sync_cleanup(); > if (s->migration_thread_running) { > qemu_thread_join(&s->thread); > s->migration_thread_running = false; > @@ -3263,6 +3264,7 @@ static MigIterateState migration_iteration_run(MigrationState *s) > > if ((!pending_size || pending_size < s->threshold_size) && can_switchover) { > trace_migration_thread_low_pending(pending_size); > + migration_background_sync_cleanup(); This one is redundant with the migrate_fd_cleanup() call at the end of migration_iteration_finish(). > migration_completion(s); > return MIG_ITERATE_BREAK; > } > @@ -3508,6 +3510,16 @@ static void *migration_thread(void *opaque) > ret = qemu_savevm_state_setup(s->to_dst_file, &local_err); > bql_unlock(); > > + if (!migrate_dirty_limit()) { > + /* > + * Initiate the background sync watcher in order to guarantee > + * that the CPU throttling acts appropriately. Dirty Limit > + * doesn't use CPU throttle to make guest down, so ignore that > + * case. > + */ > + migration_background_sync_setup(); > + } > + > qemu_savevm_wait_unplug(s, MIGRATION_STATUS_SETUP, > MIGRATION_STATUS_ACTIVE); > > diff --git a/tests/qtest/migration-test.c b/tests/qtest/migration-test.c > index b796a90cad..e0e94d26be 100644 > --- a/tests/qtest/migration-test.c > +++ b/tests/qtest/migration-test.c > @@ -281,6 +281,11 @@ static uint64_t get_migration_pass(QTestState *who) > return read_ram_property_int(who, "iteration-count"); > } > > +static uint64_t get_dirty_sync_count(QTestState *who) > +{ > + return read_ram_property_int(who, "dirty-sync-count"); > +} > + > static void read_blocktime(QTestState *who) > { > QDict *rsp_return; > @@ -468,6 +473,12 @@ static void migrate_ensure_converge(QTestState *who) > migrate_set_parameter_int(who, "downtime-limit", 30 * 1000); > } > > +static void migrate_ensure_iteration_last_long(QTestState *who) > +{ > + /* Set 10Byte/s bandwidth limit to make the iteration last long enough */ > + migrate_set_parameter_int(who, "max-bandwidth", 10); > +} > + > /* > * Our goal is to ensure that we run a single full migration > * iteration, and also dirty memory, ensuring that at least > @@ -2791,6 +2802,10 @@ static void test_migrate_auto_converge(void) > * so we need to decrease a bandwidth. > */ > const int64_t init_pct = 5, inc_pct = 25, max_pct = 95; > + uint64_t prev_iter_cnt = 0, iter_cnt; > + uint64_t iter_cnt_changes = 0; > + uint64_t prev_dirty_sync_cnt = 0, dirty_sync_cnt; > + uint64_t dirty_sync_cnt_changes = 0; > > if (test_migrate_start(&from, &to, uri, &args)) { > return; > @@ -2827,6 +2842,30 @@ static void test_migrate_auto_converge(void) > } while (true); > /* The first percentage of throttling should be at least init_pct */ > g_assert_cmpint(percentage, >=, init_pct); > + > + /* Make sure the iteration take a long time enough */ > + migrate_ensure_iteration_last_long(from); > + > + /* > + * End the loop when the dirty sync count or iteration count changes. > + */ > + while (iter_cnt_changes < 2 && dirty_sync_cnt_changes < 2) { > + usleep(1000 * 1000); > + iter_cnt = get_migration_pass(from); > + iter_cnt_changes += (iter_cnt != prev_iter_cnt); > + prev_iter_cnt = iter_cnt; > + > + dirty_sync_cnt = get_dirty_sync_count(from); > + dirty_sync_cnt_changes += (dirty_sync_cnt != prev_dirty_sync_cnt); > + prev_dirty_sync_cnt = dirty_sync_cnt; > + } > + > + /* > + * The dirty sync count must have changed because we are in the same > + * iteration. > + */ > + g_assert_cmpint(iter_cnt_changes , < , dirty_sync_cnt_changes); > + > /* Now, when we tested that throttling works, let it converge */ > migrate_ensure_converge(from);
© 2016 - 2024 Red Hat, Inc.