When VM is configured with huge memory, the current throttle logic
doesn't look like to scale, because migration_trigger_throttle()
is only called for each iteration, so it won't be invoked for a long
time if one iteration can take a long time.
The background sync and throttle aim to fix the above issue by
synchronizing the remote dirty bitmap and triggering the throttle
once detect that iteration lasts a long time.
This is a trade-off between synchronization overhead and CPU throttle
impact.
Signed-off-by: Hyman Huang <yong.huang@smartx.com>
---
migration/migration.c | 12 +++++++++++
tests/qtest/migration-test.c | 39 ++++++++++++++++++++++++++++++++++++
2 files changed, 51 insertions(+)
diff --git a/migration/migration.c b/migration/migration.c
index 055d527ff6..af8b22fa15 100644
--- a/migration/migration.c
+++ b/migration/migration.c
@@ -1416,6 +1416,7 @@ static void migrate_fd_cleanup(MigrationState *s)
trace_migrate_fd_cleanup();
bql_unlock();
+ migration_background_sync_cleanup();
if (s->migration_thread_running) {
qemu_thread_join(&s->thread);
s->migration_thread_running = false;
@@ -3263,6 +3264,7 @@ static MigIterateState migration_iteration_run(MigrationState *s)
if ((!pending_size || pending_size < s->threshold_size) && can_switchover) {
trace_migration_thread_low_pending(pending_size);
+ migration_background_sync_cleanup();
migration_completion(s);
return MIG_ITERATE_BREAK;
}
@@ -3508,6 +3510,16 @@ static void *migration_thread(void *opaque)
ret = qemu_savevm_state_setup(s->to_dst_file, &local_err);
bql_unlock();
+ if (!migrate_dirty_limit()) {
+ /*
+ * Initiate the background sync watcher in order to guarantee
+ * that the CPU throttling acts appropriately. Dirty Limit
+ * doesn't use CPU throttle to make guest down, so ignore that
+ * case.
+ */
+ migration_background_sync_setup();
+ }
+
qemu_savevm_wait_unplug(s, MIGRATION_STATUS_SETUP,
MIGRATION_STATUS_ACTIVE);
diff --git a/tests/qtest/migration-test.c b/tests/qtest/migration-test.c
index b796a90cad..e0e94d26be 100644
--- a/tests/qtest/migration-test.c
+++ b/tests/qtest/migration-test.c
@@ -281,6 +281,11 @@ static uint64_t get_migration_pass(QTestState *who)
return read_ram_property_int(who, "iteration-count");
}
+static uint64_t get_dirty_sync_count(QTestState *who)
+{
+ return read_ram_property_int(who, "dirty-sync-count");
+}
+
static void read_blocktime(QTestState *who)
{
QDict *rsp_return;
@@ -468,6 +473,12 @@ static void migrate_ensure_converge(QTestState *who)
migrate_set_parameter_int(who, "downtime-limit", 30 * 1000);
}
+static void migrate_ensure_iteration_last_long(QTestState *who)
+{
+ /* Set 10Byte/s bandwidth limit to make the iteration last long enough */
+ migrate_set_parameter_int(who, "max-bandwidth", 10);
+}
+
/*
* Our goal is to ensure that we run a single full migration
* iteration, and also dirty memory, ensuring that at least
@@ -2791,6 +2802,10 @@ static void test_migrate_auto_converge(void)
* so we need to decrease a bandwidth.
*/
const int64_t init_pct = 5, inc_pct = 25, max_pct = 95;
+ uint64_t prev_iter_cnt = 0, iter_cnt;
+ uint64_t iter_cnt_changes = 0;
+ uint64_t prev_dirty_sync_cnt = 0, dirty_sync_cnt;
+ uint64_t dirty_sync_cnt_changes = 0;
if (test_migrate_start(&from, &to, uri, &args)) {
return;
@@ -2827,6 +2842,30 @@ static void test_migrate_auto_converge(void)
} while (true);
/* The first percentage of throttling should be at least init_pct */
g_assert_cmpint(percentage, >=, init_pct);
+
+ /* Make sure the iteration take a long time enough */
+ migrate_ensure_iteration_last_long(from);
+
+ /*
+ * End the loop when the dirty sync count or iteration count changes.
+ */
+ while (iter_cnt_changes < 2 && dirty_sync_cnt_changes < 2) {
+ usleep(1000 * 1000);
+ iter_cnt = get_migration_pass(from);
+ iter_cnt_changes += (iter_cnt != prev_iter_cnt);
+ prev_iter_cnt = iter_cnt;
+
+ dirty_sync_cnt = get_dirty_sync_count(from);
+ dirty_sync_cnt_changes += (dirty_sync_cnt != prev_dirty_sync_cnt);
+ prev_dirty_sync_cnt = dirty_sync_cnt;
+ }
+
+ /*
+ * The dirty sync count must have changed because we are in the same
+ * iteration.
+ */
+ g_assert_cmpint(iter_cnt_changes , < , dirty_sync_cnt_changes);
+
/* Now, when we tested that throttling works, let it converge */
migrate_ensure_converge(from);
--
2.39.1
Hyman Huang <yong.huang@smartx.com> writes:
> When VM is configured with huge memory, the current throttle logic
> doesn't look like to scale, because migration_trigger_throttle()
> is only called for each iteration, so it won't be invoked for a long
> time if one iteration can take a long time.
>
> The background sync and throttle aim to fix the above issue by
> synchronizing the remote dirty bitmap and triggering the throttle
> once detect that iteration lasts a long time.
>
> This is a trade-off between synchronization overhead and CPU throttle
> impact.
>
> Signed-off-by: Hyman Huang <yong.huang@smartx.com>
> ---
> migration/migration.c | 12 +++++++++++
> tests/qtest/migration-test.c | 39 ++++++++++++++++++++++++++++++++++++
> 2 files changed, 51 insertions(+)
>
> diff --git a/migration/migration.c b/migration/migration.c
> index 055d527ff6..af8b22fa15 100644
> --- a/migration/migration.c
> +++ b/migration/migration.c
> @@ -1416,6 +1416,7 @@ static void migrate_fd_cleanup(MigrationState *s)
>
> trace_migrate_fd_cleanup();
> bql_unlock();
> + migration_background_sync_cleanup();
> if (s->migration_thread_running) {
> qemu_thread_join(&s->thread);
> s->migration_thread_running = false;
> @@ -3263,6 +3264,7 @@ static MigIterateState migration_iteration_run(MigrationState *s)
>
> if ((!pending_size || pending_size < s->threshold_size) && can_switchover) {
> trace_migration_thread_low_pending(pending_size);
> + migration_background_sync_cleanup();
This one is redundant with the migrate_fd_cleanup() call at the end of
migration_iteration_finish().
> migration_completion(s);
> return MIG_ITERATE_BREAK;
> }
> @@ -3508,6 +3510,16 @@ static void *migration_thread(void *opaque)
> ret = qemu_savevm_state_setup(s->to_dst_file, &local_err);
> bql_unlock();
>
> + if (!migrate_dirty_limit()) {
> + /*
> + * Initiate the background sync watcher in order to guarantee
> + * that the CPU throttling acts appropriately. Dirty Limit
> + * doesn't use CPU throttle to make guest down, so ignore that
> + * case.
> + */
> + migration_background_sync_setup();
> + }
> +
> qemu_savevm_wait_unplug(s, MIGRATION_STATUS_SETUP,
> MIGRATION_STATUS_ACTIVE);
>
> diff --git a/tests/qtest/migration-test.c b/tests/qtest/migration-test.c
> index b796a90cad..e0e94d26be 100644
> --- a/tests/qtest/migration-test.c
> +++ b/tests/qtest/migration-test.c
> @@ -281,6 +281,11 @@ static uint64_t get_migration_pass(QTestState *who)
> return read_ram_property_int(who, "iteration-count");
> }
>
> +static uint64_t get_dirty_sync_count(QTestState *who)
> +{
> + return read_ram_property_int(who, "dirty-sync-count");
> +}
> +
> static void read_blocktime(QTestState *who)
> {
> QDict *rsp_return;
> @@ -468,6 +473,12 @@ static void migrate_ensure_converge(QTestState *who)
> migrate_set_parameter_int(who, "downtime-limit", 30 * 1000);
> }
>
> +static void migrate_ensure_iteration_last_long(QTestState *who)
> +{
> + /* Set 10Byte/s bandwidth limit to make the iteration last long enough */
> + migrate_set_parameter_int(who, "max-bandwidth", 10);
> +}
> +
> /*
> * Our goal is to ensure that we run a single full migration
> * iteration, and also dirty memory, ensuring that at least
> @@ -2791,6 +2802,10 @@ static void test_migrate_auto_converge(void)
> * so we need to decrease a bandwidth.
> */
> const int64_t init_pct = 5, inc_pct = 25, max_pct = 95;
> + uint64_t prev_iter_cnt = 0, iter_cnt;
> + uint64_t iter_cnt_changes = 0;
> + uint64_t prev_dirty_sync_cnt = 0, dirty_sync_cnt;
> + uint64_t dirty_sync_cnt_changes = 0;
>
> if (test_migrate_start(&from, &to, uri, &args)) {
> return;
> @@ -2827,6 +2842,30 @@ static void test_migrate_auto_converge(void)
> } while (true);
> /* The first percentage of throttling should be at least init_pct */
> g_assert_cmpint(percentage, >=, init_pct);
> +
> + /* Make sure the iteration take a long time enough */
> + migrate_ensure_iteration_last_long(from);
> +
> + /*
> + * End the loop when the dirty sync count or iteration count changes.
> + */
> + while (iter_cnt_changes < 2 && dirty_sync_cnt_changes < 2) {
> + usleep(1000 * 1000);
> + iter_cnt = get_migration_pass(from);
> + iter_cnt_changes += (iter_cnt != prev_iter_cnt);
> + prev_iter_cnt = iter_cnt;
> +
> + dirty_sync_cnt = get_dirty_sync_count(from);
> + dirty_sync_cnt_changes += (dirty_sync_cnt != prev_dirty_sync_cnt);
> + prev_dirty_sync_cnt = dirty_sync_cnt;
> + }
> +
> + /*
> + * The dirty sync count must have changed because we are in the same
> + * iteration.
> + */
> + g_assert_cmpint(iter_cnt_changes , < , dirty_sync_cnt_changes);
> +
> /* Now, when we tested that throttling works, let it converge */
> migrate_ensure_converge(from);
© 2016 - 2026 Red Hat, Inc.