From nobody Mon Feb 9 19:09:33 2026 Delivered-To: importer@patchew.org Authentication-Results: mx.zohomail.com; dkim=pass; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=pass(p=none dis=none) header.from=linaro.org ARC-Seal: i=1; a=rsa-sha256; t=1768535227; cv=none; d=zohomail.com; s=zohoarc; b=TUL/duAcyj3dryMr1R6AxQl1E/TAlI7AFBq8/R5s7Xl0IokrmNlqR/xSltaDN/PXO6VmEmJm9GbcTn9oyabjF1NNZ5q8TsZ8q/imuzr6NgJch9x0uIYK9Mh2VHlkasBggvKIVK48TlVB+R1at95fHbKKLVujDUxnZgdedvhpi6U= ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=zohomail.com; s=zohoarc; t=1768535227; h=Content-Transfer-Encoding:Cc:Cc:Date:Date:From:From:In-Reply-To:List-Subscribe:List-Post:List-Id:List-Archive:List-Help:List-Unsubscribe:MIME-Version:Message-ID:References:Sender:Subject:Subject:To:To:Message-Id:Reply-To; bh=Ub91UhZKFBIAg2VzAb1rWE9bViNfsMn6m+Mycdb+wOk=; b=fYolX5nJcSZ0dzOqqBWSfDsKXazuo1EwBvOxDlk3Q71+Q+lT66hiZ8Z9M4c8H0Ykal+Yu7XsYJW2SaqTiRGotHqnjLWuE4UiSfysICbn2F5+diQXsXSRQkCNfollcjudUWpddg9j5lXGhpF1fV/tAH7oia3vnZtcw5V57SVhS6w= ARC-Authentication-Results: i=1; mx.zohomail.com; dkim=pass; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=pass header.from= (p=none dis=none) Return-Path: Received: from lists.gnu.org (lists.gnu.org [209.51.188.17]) by mx.zohomail.com with SMTPS id 1768535227092206.57201510480513; Thu, 15 Jan 2026 19:47:07 -0800 (PST) Received: from localhost ([::1] helo=lists1p.gnu.org) by lists.gnu.org with esmtp (Exim 4.90_1) (envelope-from ) id 1vgagf-0005Oo-97; Thu, 15 Jan 2026 22:39:49 -0500 Received: from eggs.gnu.org ([2001:470:142:3::10]) by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1vgagd-0005NW-Gi for qemu-devel@nongnu.org; Thu, 15 Jan 2026 22:39:47 -0500 Received: from mail-pl1-x62d.google.com ([2607:f8b0:4864:20::62d]) by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_128_GCM_SHA256:128) (Exim 4.90_1) (envelope-from ) id 1vgaga-00005x-M3 for qemu-devel@nongnu.org; Thu, 15 Jan 2026 22:39:47 -0500 Received: by mail-pl1-x62d.google.com with SMTP id d9443c01a7336-2a0ac29fca1so10955765ad.2 for ; Thu, 15 Jan 2026 19:39:40 -0800 (PST) Received: from stoup.. ([180.233.125.201]) by smtp.gmail.com with ESMTPSA id d9443c01a7336-2a7190ca9d5sm6527975ad.25.2026.01.15.19.39.35 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Thu, 15 Jan 2026 19:39:38 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; t=1768534779; x=1769139579; darn=nongnu.org; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:from:to:cc:subject:date :message-id:reply-to; bh=Ub91UhZKFBIAg2VzAb1rWE9bViNfsMn6m+Mycdb+wOk=; b=m6VVg++XKXjmsPCExg3xwIKyniAXJf7hiyRoXB62WyaPNb1+P4s9/jbRdxuJVdyWEj 6hwM+bAHnvIAJhplTuEGaVkqRf5SLb87hewjJFcOzwiGoPmVgm+1xf0LJaDUw43l6tsa 1nvI3g7G2Phy4TZZn/seDWkDI04Lwzwsa4RoWUq2+ee+hpgEGdpkpA3Q+R6afdjP1gt9 unj2yh6kkjIiyqV1iRIFId0ckJDJloVYwdPMxbJhSKGoYkIIb5qkfEDBB2mjqnzUcDDQ XrBpD5EaWY/PZDPZ6JRylzKvAKvDgFeNvCT3RUejfdislLOzwQMa5VK/ybapAABC0aRN UPCA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1768534779; x=1769139579; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:x-gm-gg:x-gm-message-state:from :to:cc:subject:date:message-id:reply-to; bh=Ub91UhZKFBIAg2VzAb1rWE9bViNfsMn6m+Mycdb+wOk=; b=r587IKHKsNi3zspNnMbBwKt7rN0RDlPWNw0Gqfc4xsKHiKaEveBL4et/U5L2LPBOpC i11XWIzGmIuPm63HiheCw+RdZlt4GR+jNwfhGpwIfi9QNrHOZLZFiVEQWpA5c2aEINBA FLUDtsKdidlbAHpVFLI++34oSmrwTNgoQW7YYa01cIl5I01N6HHTB6jDgpP9zZhLBPFX 3EZfCKOXgr5XLEho91Q8HJBczo8oj5zIqKQTWDxeCzCNmtx1ng/ILt8tcHDIWTqNO6Hb aH/smE70fvjdyjHr27gVgu6ORy5sCOm/EIHcxFdbKqQRFsJJwOI0XoRB9U2rbTxuPGVe d64w== X-Gm-Message-State: AOJu0Yyk07P7BatqtB3trCSQ7mkm5pGCl1EzeOAvLFXcrJY4nbEPHe1h JJklyh0mV44MF1NftuLehPrYx/rMlxOrH43gs0MeFPP1jvWrRR1q/kN1nltNfg0Uodb1cz89uTp mN4NFEyKhwg== X-Gm-Gg: AY/fxX6D2BLoMPgKhB2HRa8hL9JQoT9mThd8+utg6oK0INlEOqgOfM/J4Q9/i0dl3f+ lw5lmttCmaeyuezVS8V57lw7IQf0gkbnOh55gIDJ9KP6/VvyzXv1mHyvM5oF5HR3IU1Xd7Uc3sI LpSW608crpukinMFsJO8Xp+tNEUljBwlbr9jteK+7lTLn63O2zn+FiQkQmiJZdUW39XLhly3sla QbtF+J4URSyOjIb46SER3mXeT5kKXu5cKIy9Kwgrc5yQj7ZbAxeX7jpfZUWmnRJja19BuycooFK E3l2xCizHxoqCmXCitNmaqSaCtSC+GkjaNKyd4ZGFNNAbl//zk+xWmZIZrXobnntU2vmBmrUCxY IfmniisaotWdgKrOEfKSKXqshXqLkK+DehPHDaQ3GolrtVGxfJbzg+xba6CL0kNaFqzwwbn9UbL H/WAlS14ICWV/6fAYQfw== X-Received: by 2002:a17:902:d492:b0:2a0:8859:3722 with SMTP id d9443c01a7336-2a717532f61mr14928695ad.25.1768534779293; Thu, 15 Jan 2026 19:39:39 -0800 (PST) From: Richard Henderson To: qemu-devel@nongnu.org Cc: ktokunaga.mail@gmail.com, berrange@redhat.com, pierrick.bouvier@linaro.org, thuth@redhat.com, pbonzini@redhat.com, philmd@linaro.org Subject: [PATCH v2 53/58] migration: Drop use of Stat64 Date: Fri, 16 Jan 2026 14:32:59 +1100 Message-ID: <20260116033305.51162-54-richard.henderson@linaro.org> X-Mailer: git-send-email 2.43.0 In-Reply-To: <20260116033305.51162-1-richard.henderson@linaro.org> References: <20260116033305.51162-1-richard.henderson@linaro.org> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable Received-SPF: pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) client-ip=209.51.188.17; envelope-from=qemu-devel-bounces+importer=patchew.org@nongnu.org; helo=lists.gnu.org; Received-SPF: pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) client-ip=209.51.188.17; envelope-from=qemu-devel-bounces+importer=patchew.org@nongnu.org; helo=lists.gnu.org; Received-SPF: pass client-ip=2607:f8b0:4864:20::62d; envelope-from=richard.henderson@linaro.org; helo=mail-pl1-x62d.google.com X-Spam_score_int: -20 X-Spam_score: -2.1 X-Spam_bar: -- X-Spam_report: (-2.1 / 5.0 requ) BAYES_00=-1.9, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, DKIM_VALID_EF=-0.1, RCVD_IN_DNSWL_NONE=-0.0001, SPF_HELO_NONE=0.001, SPF_PASS=-0.001 autolearn=ham autolearn_force=no X-Spam_action: no action X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: qemu development List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: qemu-devel-bounces+importer=patchew.org@nongnu.org Sender: qemu-devel-bounces+importer=patchew.org@nongnu.org X-ZohoMail-DKIM: pass (identity @linaro.org) X-ZM-MESSAGEID: 1768535242049158500 Content-Type: text/plain; charset="utf-8" The Stat64 structure is an aid for 32-bit hosts, and is no longer required. Use plain 64-bit types. Reviewed-by: Pierrick Bouvier Signed-off-by: Richard Henderson Reviewed-by: Philippe Mathieu-Daud=C3=A9 --- migration/migration-stats.h | 36 ++++++++++++++++------------------- migration/cpu-throttle.c | 4 ++-- migration/migration-stats.c | 16 ++++++++-------- migration/migration.c | 24 +++++++++++------------ migration/multifd-nocomp.c | 2 +- migration/multifd-zero-page.c | 4 ++-- migration/multifd.c | 12 ++++-------- migration/qemu-file.c | 6 +++--- migration/ram.c | 30 ++++++++++++++--------------- migration/rdma.c | 8 ++++---- 10 files changed, 67 insertions(+), 75 deletions(-) diff --git a/migration/migration-stats.h b/migration/migration-stats.h index 05290ade76..c0f50144c9 100644 --- a/migration/migration-stats.h +++ b/migration/migration-stats.h @@ -13,8 +13,6 @@ #ifndef QEMU_MIGRATION_STATS_H #define QEMU_MIGRATION_STATS_H =20 -#include "qemu/stats64.h" - /* * Amount of time to allocate to each "chunk" of bandwidth-throttled * data. @@ -29,9 +27,7 @@ =20 /* * These are the ram migration statistic counters. It is loosely - * based on MigrationStats. We change to Stat64 any counter that - * needs to be updated using atomic ops (can be accessed by more than - * one thread). + * based on MigrationStats. */ typedef struct { /* @@ -41,66 +37,66 @@ typedef struct { * since last iteration, not counting what the guest has dirtied * since we synchronized bitmaps. */ - Stat64 dirty_bytes_last_sync; + uint64_t dirty_bytes_last_sync; /* * Number of pages dirtied per second. */ - Stat64 dirty_pages_rate; + uint64_t dirty_pages_rate; /* * Number of times we have synchronized guest bitmaps. */ - Stat64 dirty_sync_count; + uint64_t dirty_sync_count; /* * Number of times zero copy failed to send any page using zero * copy. */ - Stat64 dirty_sync_missed_zero_copy; + uint64_t dirty_sync_missed_zero_copy; /* * Number of bytes sent at migration completion stage while the * guest is stopped. */ - Stat64 downtime_bytes; + uint64_t downtime_bytes; /* * Number of bytes sent through multifd channels. */ - Stat64 multifd_bytes; + uint64_t multifd_bytes; /* * Number of pages transferred that were not full of zeros. */ - Stat64 normal_pages; + uint64_t normal_pages; /* * Number of bytes sent during postcopy. */ - Stat64 postcopy_bytes; + uint64_t postcopy_bytes; /* * Number of postcopy page faults that we have handled during * postcopy stage. */ - Stat64 postcopy_requests; + uint64_t postcopy_requests; /* * Number of bytes sent during precopy stage. */ - Stat64 precopy_bytes; + uint64_t precopy_bytes; /* * Number of bytes transferred with QEMUFile. */ - Stat64 qemu_file_transferred; + uint64_t qemu_file_transferred; /* * Amount of transferred data at the start of current cycle. */ - Stat64 rate_limit_start; + uint64_t rate_limit_start; /* * Maximum amount of data we can send in a cycle. */ - Stat64 rate_limit_max; + uint64_t rate_limit_max; /* * Number of bytes sent through RDMA. */ - Stat64 rdma_bytes; + uint64_t rdma_bytes; /* * Number of pages transferred that were full of zeros. */ - Stat64 zero_pages; + uint64_t zero_pages; } MigrationAtomicStats; =20 extern MigrationAtomicStats mig_stats; diff --git a/migration/cpu-throttle.c b/migration/cpu-throttle.c index 0642e6bdea..3b4d4aea52 100644 --- a/migration/cpu-throttle.c +++ b/migration/cpu-throttle.c @@ -134,7 +134,7 @@ int cpu_throttle_get_percentage(void) =20 void cpu_throttle_dirty_sync_timer_tick(void *opaque) { - uint64_t sync_cnt =3D stat64_get(&mig_stats.dirty_sync_count); + uint64_t sync_cnt =3D qatomic_read(&mig_stats.dirty_sync_count); =20 /* * The first iteration copies all memory anyhow and has no @@ -153,7 +153,7 @@ void cpu_throttle_dirty_sync_timer_tick(void *opaque) } =20 end: - throttle_dirty_sync_count_prev =3D stat64_get(&mig_stats.dirty_sync_co= unt); + throttle_dirty_sync_count_prev =3D qatomic_read(&mig_stats.dirty_sync_= count); =20 timer_mod(throttle_dirty_sync_timer, qemu_clock_get_ms(QEMU_CLOCK_VIRTUAL_RT) + diff --git a/migration/migration-stats.c b/migration/migration-stats.c index f690b98a03..3f17b6ac5c 100644 --- a/migration/migration-stats.c +++ b/migration/migration-stats.c @@ -11,7 +11,7 @@ */ =20 #include "qemu/osdep.h" -#include "qemu/stats64.h" +#include "qemu/atomic.h" #include "qemu-file.h" #include "trace.h" #include "migration-stats.h" @@ -29,7 +29,7 @@ bool migration_rate_exceeded(QEMUFile *f) return false; } =20 - uint64_t rate_limit_start =3D stat64_get(&mig_stats.rate_limit_start); + uint64_t rate_limit_start =3D qatomic_read(&mig_stats.rate_limit_start= ); uint64_t rate_limit_current =3D migration_transferred_bytes(); uint64_t rate_limit_used =3D rate_limit_current - rate_limit_start; =20 @@ -41,7 +41,7 @@ bool migration_rate_exceeded(QEMUFile *f) =20 uint64_t migration_rate_get(void) { - return stat64_get(&mig_stats.rate_limit_max); + return qatomic_read(&mig_stats.rate_limit_max); } =20 #define XFER_LIMIT_RATIO (1000 / BUFFER_DELAY) @@ -51,19 +51,19 @@ void migration_rate_set(uint64_t limit) /* * 'limit' is per second. But we check it each BUFFER_DELAY milliseco= nds. */ - stat64_set(&mig_stats.rate_limit_max, limit / XFER_LIMIT_RATIO); + qatomic_set(&mig_stats.rate_limit_max, limit / XFER_LIMIT_RATIO); } =20 void migration_rate_reset(void) { - stat64_set(&mig_stats.rate_limit_start, migration_transferred_bytes()); + qatomic_set(&mig_stats.rate_limit_start, migration_transferred_bytes()= ); } =20 uint64_t migration_transferred_bytes(void) { - uint64_t multifd =3D stat64_get(&mig_stats.multifd_bytes); - uint64_t rdma =3D stat64_get(&mig_stats.rdma_bytes); - uint64_t qemu_file =3D stat64_get(&mig_stats.qemu_file_transferred); + uint64_t multifd =3D qatomic_read(&mig_stats.multifd_bytes); + uint64_t rdma =3D qatomic_read(&mig_stats.rdma_bytes); + uint64_t qemu_file =3D qatomic_read(&mig_stats.qemu_file_transferred); =20 trace_migration_transferred_bytes(qemu_file, multifd, rdma); return qemu_file + multifd + rdma; diff --git a/migration/migration.c b/migration/migration.c index 1c34d8d432..1bcde301f7 100644 --- a/migration/migration.c +++ b/migration/migration.c @@ -1266,22 +1266,22 @@ static void populate_ram_info(MigrationInfo *info, = MigrationState *s) info->ram =3D g_malloc0(sizeof(*info->ram)); info->ram->transferred =3D migration_transferred_bytes(); info->ram->total =3D ram_bytes_total(); - info->ram->duplicate =3D stat64_get(&mig_stats.zero_pages); - info->ram->normal =3D stat64_get(&mig_stats.normal_pages); + info->ram->duplicate =3D qatomic_read(&mig_stats.zero_pages); + info->ram->normal =3D qatomic_read(&mig_stats.normal_pages); info->ram->normal_bytes =3D info->ram->normal * page_size; info->ram->mbps =3D s->mbps; info->ram->dirty_sync_count =3D - stat64_get(&mig_stats.dirty_sync_count); + qatomic_read(&mig_stats.dirty_sync_count); info->ram->dirty_sync_missed_zero_copy =3D - stat64_get(&mig_stats.dirty_sync_missed_zero_copy); + qatomic_read(&mig_stats.dirty_sync_missed_zero_copy); info->ram->postcopy_requests =3D - stat64_get(&mig_stats.postcopy_requests); + qatomic_read(&mig_stats.postcopy_requests); info->ram->page_size =3D page_size; - info->ram->multifd_bytes =3D stat64_get(&mig_stats.multifd_bytes); + info->ram->multifd_bytes =3D qatomic_read(&mig_stats.multifd_bytes); info->ram->pages_per_second =3D s->pages_per_second; - info->ram->precopy_bytes =3D stat64_get(&mig_stats.precopy_bytes); - info->ram->downtime_bytes =3D stat64_get(&mig_stats.downtime_bytes); - info->ram->postcopy_bytes =3D stat64_get(&mig_stats.postcopy_bytes); + info->ram->precopy_bytes =3D qatomic_read(&mig_stats.precopy_bytes); + info->ram->downtime_bytes =3D qatomic_read(&mig_stats.downtime_bytes); + info->ram->postcopy_bytes =3D qatomic_read(&mig_stats.postcopy_bytes); =20 if (migrate_xbzrle()) { info->xbzrle_cache =3D g_malloc0(sizeof(*info->xbzrle_cache)); @@ -1302,7 +1302,7 @@ static void populate_ram_info(MigrationInfo *info, Mi= grationState *s) if (s->state !=3D MIGRATION_STATUS_COMPLETED) { info->ram->remaining =3D ram_bytes_remaining(); info->ram->dirty_pages_rate =3D - stat64_get(&mig_stats.dirty_pages_rate); + qatomic_read(&mig_stats.dirty_pages_rate); } =20 if (migrate_dirty_limit() && dirtylimit_in_service()) { @@ -3420,10 +3420,10 @@ static void migration_update_counters(MigrationStat= e *s, * if we haven't sent anything, we don't want to * recalculate. 10000 is a small enough number for our purposes */ - if (stat64_get(&mig_stats.dirty_pages_rate) && + if (qatomic_read(&mig_stats.dirty_pages_rate) && transferred > 10000) { s->expected_downtime =3D - stat64_get(&mig_stats.dirty_bytes_last_sync) / expected_bw_per= _ms; + qatomic_read(&mig_stats.dirty_bytes_last_sync) / expected_bw_p= er_ms; } =20 migration_rate_reset(); diff --git a/migration/multifd-nocomp.c b/migration/multifd-nocomp.c index b48eae3d86..9be79b3b8e 100644 --- a/migration/multifd-nocomp.c +++ b/migration/multifd-nocomp.c @@ -141,7 +141,7 @@ static int multifd_nocomp_send_prepare(MultiFDSendParam= s *p, Error **errp) return -1; } =20 - stat64_add(&mig_stats.multifd_bytes, p->packet_len); + qatomic_add(&mig_stats.multifd_bytes, p->packet_len); } =20 return 0; diff --git a/migration/multifd-zero-page.c b/migration/multifd-zero-page.c index 4cde868159..00c330416a 100644 --- a/migration/multifd-zero-page.c +++ b/migration/multifd-zero-page.c @@ -77,8 +77,8 @@ void multifd_send_zero_page_detect(MultiFDSendParams *p) pages->normal_num =3D i; =20 out: - stat64_add(&mig_stats.normal_pages, pages->normal_num); - stat64_add(&mig_stats.zero_pages, pages->num - pages->normal_num); + qatomic_add(&mig_stats.normal_pages, pages->normal_num); + qatomic_add(&mig_stats.zero_pages, pages->num - pages->normal_num); } =20 void multifd_recv_zero_page_process(MultiFDRecvParams *p) diff --git a/migration/multifd.c b/migration/multifd.c index bf6da85af8..c9d4a67a46 100644 --- a/migration/multifd.c +++ b/migration/multifd.c @@ -58,10 +58,6 @@ struct { * operations on both 32bit / 64 bits hosts. It means on 32bit systems * multifd will overflow the packet_num easier, but that should be * fine. - * - * Another option is to use QEMU's Stat64 then it'll be 64 bits on all - * hosts, however so far it does not support atomic fetch_add() yet. - * Make it easy for now. */ uintptr_t packet_num; /* @@ -174,7 +170,7 @@ static int multifd_send_initial_packet(MultiFDSendParam= s *p, Error **errp) if (ret !=3D 0) { return -1; } - stat64_add(&mig_stats.multifd_bytes, size); + qatomic_add(&mig_stats.multifd_bytes, size); return 0; } =20 @@ -607,7 +603,7 @@ static int multifd_zero_copy_flush(QIOChannel *c) return -1; } if (ret =3D=3D 1) { - stat64_add(&mig_stats.dirty_sync_missed_zero_copy, 1); + qatomic_add(&mig_stats.dirty_sync_missed_zero_copy, 1); } =20 return ret; @@ -735,7 +731,7 @@ static void *multifd_send_thread(void *opaque) break; } =20 - stat64_add(&mig_stats.multifd_bytes, total_size); + qatomic_add(&mig_stats.multifd_bytes, total_size); =20 p->next_packet_size =3D 0; multifd_send_data_clear(p->data); @@ -766,7 +762,7 @@ static void *multifd_send_thread(void *opaque) break; } /* p->next_packet_size will always be zero for a SYNC pack= et */ - stat64_add(&mig_stats.multifd_bytes, p->packet_len); + qatomic_add(&mig_stats.multifd_bytes, p->packet_len); } =20 qatomic_set(&p->pending_sync, MULTIFD_SYNC_NONE); diff --git a/migration/qemu-file.c b/migration/qemu-file.c index 4b5a409a80..8d82d94416 100644 --- a/migration/qemu-file.c +++ b/migration/qemu-file.c @@ -295,7 +295,7 @@ int qemu_fflush(QEMUFile *f) qemu_file_set_error_obj(f, -EIO, local_error); } else { uint64_t size =3D iov_size(f->iov, f->iovcnt); - stat64_add(&mig_stats.qemu_file_transferred, size); + qatomic_add(&mig_stats.qemu_file_transferred, size); } =20 qemu_iovec_release_ram(f); @@ -552,7 +552,7 @@ void qemu_put_buffer_at(QEMUFile *f, const uint8_t *buf= , size_t buflen, return; } =20 - stat64_add(&mig_stats.qemu_file_transferred, buflen); + qatomic_add(&mig_stats.qemu_file_transferred, buflen); } =20 =20 @@ -785,7 +785,7 @@ int coroutine_mixed_fn qemu_get_byte(QEMUFile *f) =20 uint64_t qemu_file_transferred(QEMUFile *f) { - uint64_t ret =3D stat64_get(&mig_stats.qemu_file_transferred); + uint64_t ret =3D qatomic_read(&mig_stats.qemu_file_transferred); int i; =20 g_assert(qemu_file_is_writable(f)); diff --git a/migration/ram.c b/migration/ram.c index 04958c5603..fc7ece2c1a 100644 --- a/migration/ram.c +++ b/migration/ram.c @@ -479,11 +479,11 @@ uint64_t ram_bytes_remaining(void) void ram_transferred_add(uint64_t bytes) { if (runstate_is_running()) { - stat64_add(&mig_stats.precopy_bytes, bytes); + qatomic_add(&mig_stats.precopy_bytes, bytes); } else if (migration_in_postcopy()) { - stat64_add(&mig_stats.postcopy_bytes, bytes); + qatomic_add(&mig_stats.postcopy_bytes, bytes); } else { - stat64_add(&mig_stats.downtime_bytes, bytes); + qatomic_add(&mig_stats.downtime_bytes, bytes); } } =20 @@ -605,7 +605,7 @@ static void xbzrle_cache_zero_page(ram_addr_t current_a= ddr) /* We don't care if this fails to allocate a new cache page * as long as it updated an old one */ cache_insert(XBZRLE.cache, current_addr, XBZRLE.zero_target_page, - stat64_get(&mig_stats.dirty_sync_count)); + qatomic_read(&mig_stats.dirty_sync_count)); } =20 #define ENCODING_FLAG_XBZRLE 0x1 @@ -631,7 +631,7 @@ static int save_xbzrle_page(RAMState *rs, PageSearchSta= tus *pss, int encoded_len =3D 0, bytes_xbzrle; uint8_t *prev_cached_page; QEMUFile *file =3D pss->pss_channel; - uint64_t generation =3D stat64_get(&mig_stats.dirty_sync_count); + uint64_t generation =3D qatomic_read(&mig_stats.dirty_sync_count); =20 if (!cache_is_cached(XBZRLE.cache, current_addr, generation)) { xbzrle_counters.cache_miss++; @@ -1035,9 +1035,9 @@ uint64_t ram_pagesize_summary(void) =20 uint64_t ram_get_total_transferred_pages(void) { - return stat64_get(&mig_stats.normal_pages) + - stat64_get(&mig_stats.zero_pages) + - xbzrle_counters.pages; + return (qatomic_read(&mig_stats.normal_pages) + + qatomic_read(&mig_stats.zero_pages) + + xbzrle_counters.pages); } =20 static void migration_update_rates(RAMState *rs, int64_t end_time) @@ -1045,7 +1045,7 @@ static void migration_update_rates(RAMState *rs, int6= 4_t end_time) uint64_t page_count =3D rs->target_page_count - rs->target_page_count_= prev; =20 /* calculate period counters */ - stat64_set(&mig_stats.dirty_pages_rate, + qatomic_set(&mig_stats.dirty_pages_rate, rs->num_dirty_pages_period * 1000 / (end_time - rs->time_last_bitmap_sync)); =20 @@ -1136,7 +1136,7 @@ static void migration_bitmap_sync(RAMState *rs, bool = last_stage) RAMBlock *block; int64_t end_time; =20 - stat64_add(&mig_stats.dirty_sync_count, 1); + qatomic_add(&mig_stats.dirty_sync_count, 1); =20 if (!rs->time_last_bitmap_sync) { rs->time_last_bitmap_sync =3D qemu_clock_get_ms(QEMU_CLOCK_REALTIM= E); @@ -1150,7 +1150,7 @@ static void migration_bitmap_sync(RAMState *rs, bool = last_stage) RAMBLOCK_FOREACH_NOT_IGNORED(block) { ramblock_sync_dirty_bitmap(rs, block); } - stat64_set(&mig_stats.dirty_bytes_last_sync, ram_bytes_remaini= ng()); + qatomic_set(&mig_stats.dirty_bytes_last_sync, ram_bytes_remain= ing()); } } =20 @@ -1173,7 +1173,7 @@ static void migration_bitmap_sync(RAMState *rs, bool = last_stage) rs->bytes_xfer_prev =3D migration_transferred_bytes(); } if (migrate_events()) { - uint64_t generation =3D stat64_get(&mig_stats.dirty_sync_count); + uint64_t generation =3D qatomic_read(&mig_stats.dirty_sync_count); qapi_event_send_migration_pass(generation); } } @@ -1232,7 +1232,7 @@ static int save_zero_page(RAMState *rs, PageSearchSta= tus *pss, return 0; } =20 - stat64_add(&mig_stats.zero_pages, 1); + qatomic_add(&mig_stats.zero_pages, 1); =20 if (migrate_mapped_ram()) { /* zero pages are not transferred with mapped-ram */ @@ -1291,7 +1291,7 @@ static int save_normal_page(PageSearchStatus *pss, RA= MBlock *block, } } ram_transferred_add(TARGET_PAGE_SIZE); - stat64_add(&mig_stats.normal_pages, 1); + qatomic_add(&mig_stats.normal_pages, 1); return 1; } =20 @@ -1943,7 +1943,7 @@ int ram_save_queue_pages(const char *rbname, ram_addr= _t start, ram_addr_t len, RAMBlock *ramblock; RAMState *rs =3D ram_state; =20 - stat64_add(&mig_stats.postcopy_requests, 1); + qatomic_add(&mig_stats.postcopy_requests, 1); RCU_READ_LOCK_GUARD(); =20 if (!rbname) { diff --git a/migration/rdma.c b/migration/rdma.c index 9e301cf917..cced173379 100644 --- a/migration/rdma.c +++ b/migration/rdma.c @@ -1936,8 +1936,8 @@ retry: * would think that head.len would be the more similar * thing to a correct value. */ - stat64_add(&mig_stats.zero_pages, - sge.length / qemu_target_page_size()); + qatomic_add(&mig_stats.zero_pages, + sge.length / qemu_target_page_size()); return 1; } =20 @@ -2045,7 +2045,7 @@ retry: } =20 set_bit(chunk, block->transit_bitmap); - stat64_add(&mig_stats.normal_pages, sge.length / qemu_target_page_size= ()); + qatomic_add(&mig_stats.normal_pages, sge.length / qemu_target_page_siz= e()); /* * We are adding to transferred the amount of data written, but no * overhead at all. I will assume that RDMA is magicaly and don't @@ -2055,7 +2055,7 @@ retry: * sizeof(send_wr) + sge.length * but this being RDMA, who knows. */ - stat64_add(&mig_stats.rdma_bytes, sge.length); + qatomic_add(&mig_stats.rdma_bytes, sge.length); ram_transferred_add(sge.length); rdma->total_writes++; =20 --=20 2.43.0