From nobody Wed Apr 24 14:58:26 2024 Delivered-To: importer@patchew.org Received-SPF: pass (zohomail.com: domain of lists.xenproject.org designates 192.237.175.120 as permitted sender) client-ip=192.237.175.120; envelope-from=xen-devel-bounces@lists.xenproject.org; helo=lists.xenproject.org; Authentication-Results: mx.zohomail.com; dkim=pass; spf=pass (zohomail.com: domain of lists.xenproject.org designates 192.237.175.120 as permitted sender) smtp.mailfrom=xen-devel-bounces@lists.xenproject.org; dmarc=pass(p=none dis=none) header.from=redhat.com ARC-Seal: i=1; a=rsa-sha256; t=1684361468; cv=none; d=zohomail.com; s=zohoarc; b=Qj106ZcIq93FJjjwmOzJl0vJdW9SHPzaUtz3ZgVFgQGJMUtxcieJFK4iZLJ3cgtc2b7/OIq+lbPGWbq4xP8BNlq+VWkPfPhm/cfPbn8AET7tZcTi1bZ03s0AbChSRC0P4zpYt8bz4ACWQZPf7iSKoL1oH7OqBjvmCB8UqYEC9xk= ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=zohomail.com; s=zohoarc; t=1684361468; h=Content-Transfer-Encoding:Cc:Date:From:In-Reply-To:List-Subscribe:List-Post:List-Id:List-Help:List-Unsubscribe:MIME-Version:Message-ID:References:Sender:Subject:To; bh=/rde1IRAOm4sc1Tal7ezRnPv14rgbAWjAI/I5FB1Dhc=; b=RXhTBxKlITistDh5m+LFHPKBX/X6MBFQQ1cYgVZN4S+w7SjuVfThgMbBCpwKH1GXvUFAk0E1FqTDB6pmORZ2K+LNMwrEyrh7Pjz9i4hPQHcgnVE58j7aKpUDtu9X9tjae/NJq9RBxAonMaydp6ft3DKGwltTTv8PKRlgRulbJH0= ARC-Authentication-Results: i=1; mx.zohomail.com; dkim=pass; spf=pass (zohomail.com: domain of lists.xenproject.org designates 192.237.175.120 as permitted sender) smtp.mailfrom=xen-devel-bounces@lists.xenproject.org; dmarc=pass header.from= (p=none dis=none) Return-Path: Received: from lists.xenproject.org (lists.xenproject.org [192.237.175.120]) by mx.zohomail.com with SMTPS id 1684361468412630.0345631471579; Wed, 17 May 2023 15:11:08 -0700 (PDT) Received: from list by lists.xenproject.org with outflank-mailman.536080.834200 (Exim 4.92) (envelope-from ) id 1pzPLz-00058r-8e; Wed, 17 May 2023 22:10:39 +0000 Received: by outflank-mailman (output) from mailman id 536080.834200; Wed, 17 May 2023 22:10:39 +0000 Received: from localhost ([127.0.0.1] helo=lists.xenproject.org) by lists.xenproject.org with esmtp (Exim 4.92) (envelope-from ) id 1pzPLz-00058k-5o; Wed, 17 May 2023 22:10:39 +0000 Received: by outflank-mailman (input) for mailman id 536080; Wed, 17 May 2023 22:10:37 +0000 Received: from se1-gles-flk1-in.inumbo.com ([94.247.172.50] helo=se1-gles-flk1.inumbo.com) by lists.xenproject.org with esmtp (Exim 4.92) (envelope-from ) id 1pzPLx-0004sV-Fc for xen-devel@lists.xenproject.org; Wed, 17 May 2023 22:10:37 +0000 Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.133.124]) by se1-gles-flk1.inumbo.com (Halon) with ESMTPS id a49c1865-f4ff-11ed-8611-37d641c3527e; Thu, 18 May 2023 00:10:34 +0200 (CEST) Received: from mimecast-mx02.redhat.com (mimecast-mx02.redhat.com [66.187.233.88]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id us-mta-48-VfyR7csfO32fVtGOTuGAng-1; Wed, 17 May 2023 18:10:29 -0400 Received: from smtp.corp.redhat.com (int-mx06.intmail.prod.int.rdu2.redhat.com [10.11.54.6]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx02.redhat.com (Postfix) with ESMTPS id 8DD1A805F5A; Wed, 17 May 2023 22:10:28 +0000 (UTC) Received: from localhost (unknown [10.39.192.14]) by smtp.corp.redhat.com (Postfix) with ESMTP id AEC6C2166B31; Wed, 17 May 2023 22:10:27 +0000 (UTC) X-Outflank-Mailman: Message body and most headers restored to incoming version X-BeenThere: xen-devel@lists.xenproject.org List-Id: Xen developer discussion List-Unsubscribe: , List-Post: List-Help: List-Subscribe: , Errors-To: xen-devel-bounces@lists.xenproject.org Precedence: list Sender: "Xen-devel" X-Inumbo-ID: a49c1865-f4ff-11ed-8611-37d641c3527e DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1684361433; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=/rde1IRAOm4sc1Tal7ezRnPv14rgbAWjAI/I5FB1Dhc=; b=FWyCOElPIxRnQkix7GgSn8cUOgF64boWhooWqybUHODsgKsfjSwXJJpq7PiYYTm6w6oBMf lzvjwuRQIXiK+0LcJ3pwIlMoUd1pFuIOQg5AfgbyJnsufP2fLGXk6OvgoNDJGQWtZtnGuQ e3+vK3+ZDOI8rO50IlXxcx4UMFrBPsI= X-MC-Unique: VfyR7csfO32fVtGOTuGAng-1 From: Stefan Hajnoczi To: qemu-devel@nongnu.org Cc: xen-devel@lists.xenproject.org, Julia Suvorova , "Michael S. Tsirkin" , Stefano Stabellini , Stefano Garzarella , Kevin Wolf , Fam Zheng , =?UTF-8?q?Philippe=20Mathieu-Daud=C3=A9?= , Paolo Bonzini , qemu-block@nongnu.org, Paul Durrant , Anthony Perard , Aarushi Mehta , Stefan Hajnoczi , Hanna Reitz Subject: [PATCH 1/6] block: add blk_io_plug_call() API Date: Wed, 17 May 2023 18:10:17 -0400 Message-Id: <20230517221022.325091-2-stefanha@redhat.com> In-Reply-To: <20230517221022.325091-1-stefanha@redhat.com> References: <20230517221022.325091-1-stefanha@redhat.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable X-Scanned-By: MIMEDefang 3.1 on 10.11.54.6 X-ZohoMail-DKIM: pass (identity @redhat.com) X-ZM-MESSAGEID: 1684361470605100001 Content-Type: text/plain; charset="utf-8" Introduce a new API for thread-local blk_io_plug() that does not traverse the block graph. The goal is to make blk_io_plug() multi-queue friendly. Instead of having block drivers track whether or not we're in a plugged section, provide an API that allows them to defer a function call until we're unplugged: blk_io_plug_call(fn, opaque). If blk_io_plug_call() is called multiple times with the same fn/opaque pair, then fn() is only called once at the end of the function - resulting in batching. This patch introduces the API and changes blk_io_plug()/blk_io_unplug(). blk_io_plug()/blk_io_unplug() no longer require a BlockBackend argument because the plug state is now thread-local. Later patches convert block drivers to blk_io_plug_call() and then we can finally remove .bdrv_co_io_plug() once all block drivers have been converted. Signed-off-by: Stefan Hajnoczi Reviewed-by: Eric Blake --- MAINTAINERS | 1 + include/sysemu/block-backend-io.h | 13 +-- block/block-backend.c | 22 ----- block/plug.c | 159 ++++++++++++++++++++++++++++++ hw/block/dataplane/xen-block.c | 8 +- hw/block/virtio-blk.c | 4 +- hw/scsi/virtio-scsi.c | 6 +- block/meson.build | 1 + 8 files changed, 173 insertions(+), 41 deletions(-) create mode 100644 block/plug.c diff --git a/MAINTAINERS b/MAINTAINERS index 50585117a0..574202295c 100644 --- a/MAINTAINERS +++ b/MAINTAINERS @@ -2644,6 +2644,7 @@ F: util/aio-*.c F: util/aio-*.h F: util/fdmon-*.c F: block/io.c +F: block/plug.c F: migration/block* F: include/block/aio.h F: include/block/aio-wait.h diff --git a/include/sysemu/block-backend-io.h b/include/sysemu/block-backe= nd-io.h index d62a7ee773..be4dcef59d 100644 --- a/include/sysemu/block-backend-io.h +++ b/include/sysemu/block-backend-io.h @@ -100,16 +100,9 @@ void blk_iostatus_set_err(BlockBackend *blk, int error= ); int blk_get_max_iov(BlockBackend *blk); int blk_get_max_hw_iov(BlockBackend *blk); =20 -/* - * blk_io_plug/unplug are thread-local operations. This means that multiple - * IOThreads can simultaneously call plug/unplug, but the caller must ensu= re - * that each unplug() is called in the same IOThread of the matching plug(= ). - */ -void coroutine_fn blk_co_io_plug(BlockBackend *blk); -void co_wrapper blk_io_plug(BlockBackend *blk); - -void coroutine_fn blk_co_io_unplug(BlockBackend *blk); -void co_wrapper blk_io_unplug(BlockBackend *blk); +void blk_io_plug(void); +void blk_io_unplug(void); +void blk_io_plug_call(void (*fn)(void *), void *opaque); =20 AioContext *blk_get_aio_context(BlockBackend *blk); BlockAcctStats *blk_get_stats(BlockBackend *blk); diff --git a/block/block-backend.c b/block/block-backend.c index ca537cd0ad..1f1d226ba6 100644 --- a/block/block-backend.c +++ b/block/block-backend.c @@ -2568,28 +2568,6 @@ void blk_add_insert_bs_notifier(BlockBackend *blk, N= otifier *notify) notifier_list_add(&blk->insert_bs_notifiers, notify); } =20 -void coroutine_fn blk_co_io_plug(BlockBackend *blk) -{ - BlockDriverState *bs =3D blk_bs(blk); - IO_CODE(); - GRAPH_RDLOCK_GUARD(); - - if (bs) { - bdrv_co_io_plug(bs); - } -} - -void coroutine_fn blk_co_io_unplug(BlockBackend *blk) -{ - BlockDriverState *bs =3D blk_bs(blk); - IO_CODE(); - GRAPH_RDLOCK_GUARD(); - - if (bs) { - bdrv_co_io_unplug(bs); - } -} - BlockAcctStats *blk_get_stats(BlockBackend *blk) { IO_CODE(); diff --git a/block/plug.c b/block/plug.c new file mode 100644 index 0000000000..6738a568ba --- /dev/null +++ b/block/plug.c @@ -0,0 +1,159 @@ +/* SPDX-License-Identifier: GPL-2.0-or-later */ +/* + * Block I/O plugging + * + * Copyright Red Hat. + * + * This API defers a function call within a blk_io_plug()/blk_io_unplug() + * section, allowing multiple calls to batch up. This is a performance + * optimization that is used in the block layer to submit several I/O requ= ests + * at once instead of individually: + * + * blk_io_plug(); <-- start of plugged region + * ... + * blk_io_plug_call(my_func, my_obj); <-- deferred my_func(my_obj) call + * blk_io_plug_call(my_func, my_obj); <-- another + * blk_io_plug_call(my_func, my_obj); <-- another + * ... + * blk_io_unplug(); <-- end of plugged region, my_func(my_obj) is called= once + * + * This code is actually generic and not tied to the block layer. If anoth= er + * subsystem needs this functionality, it could be renamed. + */ + +#include "qemu/osdep.h" +#include "qemu/coroutine-tls.h" +#include "qemu/notify.h" +#include "qemu/thread.h" +#include "sysemu/block-backend.h" + +/* A function call that has been deferred until unplug() */ +typedef struct { + void (*fn)(void *); + void *opaque; +} UnplugFn; + +/* Per-thread state */ +typedef struct { + unsigned count; /* how many times has plug() been called? */ + GArray *unplug_fns; /* functions to call at unplug time */ +} Plug; + +/* Use get_ptr_plug() to fetch this thread-local value */ +QEMU_DEFINE_STATIC_CO_TLS(Plug, plug); + +/* Called at thread cleanup time */ +static void blk_io_plug_atexit(Notifier *n, void *value) +{ + Plug *plug =3D get_ptr_plug(); + g_array_free(plug->unplug_fns, TRUE); +} + +/* This won't involve coroutines, so use __thread */ +static __thread Notifier blk_io_plug_atexit_notifier; + +/** + * blk_io_plug_call: + * @fn: a function pointer to be invoked + * @opaque: a user-defined argument to @fn() + * + * Call @fn(@opaque) immediately if not within a blk_io_plug()/blk_io_unpl= ug() + * section. + * + * Otherwise defer the call until the end of the outermost + * blk_io_plug()/blk_io_unplug() section in this thread. If the same + * @fn/@opaque pair has already been deferred, it will only be called once= upon + * blk_io_unplug() so that accumulated calls are batched into a single cal= l. + * + * The caller must ensure that @opaque is not be freed before @fn() is inv= oked. + */ +void blk_io_plug_call(void (*fn)(void *), void *opaque) +{ + Plug *plug =3D get_ptr_plug(); + + /* Call immediately if we're not plugged */ + if (plug->count =3D=3D 0) { + fn(opaque); + return; + } + + GArray *array =3D plug->unplug_fns; + if (!array) { + array =3D g_array_new(FALSE, FALSE, sizeof(UnplugFn)); + plug->unplug_fns =3D array; + blk_io_plug_atexit_notifier.notify =3D blk_io_plug_atexit; + qemu_thread_atexit_add(&blk_io_plug_atexit_notifier); + } + + UnplugFn *fns =3D (UnplugFn *)array->data; + UnplugFn new_fn =3D { + .fn =3D fn, + .opaque =3D opaque, + }; + + /* + * There won't be many, so do a linear search. If this becomes a bottl= eneck + * then a binary search (glib 2.62+) or different data structure could= be + * used. + */ + for (guint i =3D 0; i < array->len; i++) { + if (memcmp(&fns[i], &new_fn, sizeof(new_fn)) =3D=3D 0) { + return; /* already exists */ + } + } + + g_array_append_val(array, new_fn); +} + +/** + * blk_io_plug: Defer blk_io_plug_call() functions until blk_io_unplug() + * + * blk_io_plug/unplug are thread-local operations. This means that multiple + * threads can simultaneously call plug/unplug, but the caller must ensure= that + * each unplug() is called in the same thread of the matching plug(). + * + * Nesting is supported. blk_io_plug_call() functions are only called at t= he + * outermost blk_io_unplug(). + */ +void blk_io_plug(void) +{ + Plug *plug =3D get_ptr_plug(); + + assert(plug->count < UINT32_MAX); + + plug->count++; +} + +/** + * blk_io_unplug: Run any pending blk_io_plug_call() functions + * + * There must have been a matching blk_io_plug() call in the same thread p= rior + * to this blk_io_unplug() call. + */ +void blk_io_unplug(void) +{ + Plug *plug =3D get_ptr_plug(); + + assert(plug->count > 0); + + if (--plug->count > 0) { + return; + } + + GArray *array =3D plug->unplug_fns; + if (!array) { + return; + } + + UnplugFn *fns =3D (UnplugFn *)array->data; + + for (guint i =3D 0; i < array->len; i++) { + fns[i].fn(fns[i].opaque); + } + + /* + * This resets the array without freeing memory so that appending is c= heap + * in the future. + */ + g_array_set_size(array, 0); +} diff --git a/hw/block/dataplane/xen-block.c b/hw/block/dataplane/xen-block.c index d8bc39d359..e49c24f63d 100644 --- a/hw/block/dataplane/xen-block.c +++ b/hw/block/dataplane/xen-block.c @@ -537,7 +537,7 @@ static bool xen_block_handle_requests(XenBlockDataPlane= *dataplane) * is below us. */ if (inflight_atstart > IO_PLUG_THRESHOLD) { - blk_io_plug(dataplane->blk); + blk_io_plug(); } while (rc !=3D rp) { /* pull request from ring */ @@ -577,12 +577,12 @@ static bool xen_block_handle_requests(XenBlockDataPla= ne *dataplane) =20 if (inflight_atstart > IO_PLUG_THRESHOLD && batched >=3D inflight_atstart) { - blk_io_unplug(dataplane->blk); + blk_io_unplug(); } xen_block_do_aio(request); if (inflight_atstart > IO_PLUG_THRESHOLD) { if (batched >=3D inflight_atstart) { - blk_io_plug(dataplane->blk); + blk_io_plug(); batched =3D 0; } else { batched++; @@ -590,7 +590,7 @@ static bool xen_block_handle_requests(XenBlockDataPlane= *dataplane) } } if (inflight_atstart > IO_PLUG_THRESHOLD) { - blk_io_unplug(dataplane->blk); + blk_io_unplug(); } =20 return done_something; diff --git a/hw/block/virtio-blk.c b/hw/block/virtio-blk.c index 8f65ea4659..b4286424c1 100644 --- a/hw/block/virtio-blk.c +++ b/hw/block/virtio-blk.c @@ -1134,7 +1134,7 @@ void virtio_blk_handle_vq(VirtIOBlock *s, VirtQueue *= vq) bool suppress_notifications =3D virtio_queue_get_notification(vq); =20 aio_context_acquire(blk_get_aio_context(s->blk)); - blk_io_plug(s->blk); + blk_io_plug(); =20 do { if (suppress_notifications) { @@ -1158,7 +1158,7 @@ void virtio_blk_handle_vq(VirtIOBlock *s, VirtQueue *= vq) virtio_blk_submit_multireq(s, &mrb); } =20 - blk_io_unplug(s->blk); + blk_io_unplug(); aio_context_release(blk_get_aio_context(s->blk)); } =20 diff --git a/hw/scsi/virtio-scsi.c b/hw/scsi/virtio-scsi.c index 612c525d9d..534a44ee07 100644 --- a/hw/scsi/virtio-scsi.c +++ b/hw/scsi/virtio-scsi.c @@ -799,7 +799,7 @@ static int virtio_scsi_handle_cmd_req_prepare(VirtIOSCS= I *s, VirtIOSCSIReq *req) return -ENOBUFS; } scsi_req_ref(req->sreq); - blk_io_plug(d->conf.blk); + blk_io_plug(); object_unref(OBJECT(d)); return 0; } @@ -810,7 +810,7 @@ static void virtio_scsi_handle_cmd_req_submit(VirtIOSCS= I *s, VirtIOSCSIReq *req) if (scsi_req_enqueue(sreq)) { scsi_req_continue(sreq); } - blk_io_unplug(sreq->dev->conf.blk); + blk_io_unplug(); scsi_req_unref(sreq); } =20 @@ -836,7 +836,7 @@ static void virtio_scsi_handle_cmd_vq(VirtIOSCSI *s, Vi= rtQueue *vq) while (!QTAILQ_EMPTY(&reqs)) { req =3D QTAILQ_FIRST(&reqs); QTAILQ_REMOVE(&reqs, req, next); - blk_io_unplug(req->sreq->dev->conf.blk); + blk_io_unplug(); scsi_req_unref(req->sreq); virtqueue_detach_element(req->vq, &req->elem, 0); virtio_scsi_free_req(req); diff --git a/block/meson.build b/block/meson.build index 486dda8b85..fb4332bd66 100644 --- a/block/meson.build +++ b/block/meson.build @@ -23,6 +23,7 @@ block_ss.add(files( 'mirror.c', 'nbd.c', 'null.c', + 'plug.c', 'qapi.c', 'qcow2-bitmap.c', 'qcow2-cache.c', --=20 2.40.1 From nobody Wed Apr 24 14:58:26 2024 Delivered-To: importer@patchew.org Received-SPF: pass (zohomail.com: domain of lists.xenproject.org designates 192.237.175.120 as permitted sender) client-ip=192.237.175.120; envelope-from=xen-devel-bounces@lists.xenproject.org; helo=lists.xenproject.org; Authentication-Results: mx.zohomail.com; dkim=pass; spf=pass (zohomail.com: domain of lists.xenproject.org designates 192.237.175.120 as permitted sender) smtp.mailfrom=xen-devel-bounces@lists.xenproject.org; dmarc=pass(p=none dis=none) header.from=redhat.com ARC-Seal: i=1; a=rsa-sha256; t=1684361482; cv=none; d=zohomail.com; s=zohoarc; b=iLCynnMLVkxIIEV90i7VajsPS/OVejeE5dS/bo/6k6gruMYTfp35VK9EEhcyVX+gO86CwN/RQ58KB1ScBqN2i1lh6xDRdroh810f/V0lseI4bzFM3SmTuspZyrdWQW+5oRNEWvUov+NIAIGq5dlygottXzy/snByPR9dS0Yr8VA= ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=zohomail.com; s=zohoarc; t=1684361482; h=Content-Transfer-Encoding:Cc:Date:From:In-Reply-To:List-Subscribe:List-Post:List-Id:List-Help:List-Unsubscribe:MIME-Version:Message-ID:References:Sender:Subject:To; bh=gFn0D1QeS7K9r5hk8We93ICMQFKJ73KryVejH0WL6Ew=; b=n7r+qcvW6XHWaHhToghacOiUEfmxsHD0cxNx93O6KYLFW1KO0f2NKnnbqHgDX1+c8uTUjr+cpcKn2INFblcc05K0rekxoZ8ijWgKcVraGL+ENXj799bnpwXXTtzCuMI4jQwNjX48SjlY1TJRlJkLMOMEHvr4NxDbHrqXGOK42NI= ARC-Authentication-Results: i=1; mx.zohomail.com; dkim=pass; spf=pass (zohomail.com: domain of lists.xenproject.org designates 192.237.175.120 as permitted sender) smtp.mailfrom=xen-devel-bounces@lists.xenproject.org; dmarc=pass header.from= (p=none dis=none) Return-Path: Received: from lists.xenproject.org (lists.xenproject.org [192.237.175.120]) by mx.zohomail.com with SMTPS id 1684361482269555.832509196461; Wed, 17 May 2023 15:11:22 -0700 (PDT) Received: from list by lists.xenproject.org with outflank-mailman.536081.834205 (Exim 4.92) (envelope-from ) id 1pzPLz-0005CP-JP; Wed, 17 May 2023 22:10:39 +0000 Received: by outflank-mailman (output) from mailman id 536081.834205; Wed, 17 May 2023 22:10:39 +0000 Received: from localhost ([127.0.0.1] helo=lists.xenproject.org) by lists.xenproject.org with esmtp (Exim 4.92) (envelope-from ) id 1pzPLz-0005BQ-EJ; Wed, 17 May 2023 22:10:39 +0000 Received: by outflank-mailman (input) for mailman id 536081; Wed, 17 May 2023 22:10:37 +0000 Received: from se1-gles-sth1-in.inumbo.com ([159.253.27.254] helo=se1-gles-sth1.inumbo.com) by lists.xenproject.org with esmtp (Exim 4.92) (envelope-from ) id 1pzPLx-0004sP-Ny for xen-devel@lists.xenproject.org; Wed, 17 May 2023 22:10:37 +0000 Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.133.124]) by se1-gles-sth1.inumbo.com (Halon) with ESMTPS id a639c45e-f4ff-11ed-b22a-6b7b168915f2; Thu, 18 May 2023 00:10:37 +0200 (CEST) Received: from mimecast-mx02.redhat.com (mx3-rdu2.redhat.com [66.187.233.73]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id us-mta-192-C0RqvpuqMluZCT-d_XeyiQ-1; Wed, 17 May 2023 18:10:32 -0400 Received: from smtp.corp.redhat.com (int-mx07.intmail.prod.int.rdu2.redhat.com [10.11.54.7]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx02.redhat.com (Postfix) with ESMTPS id 5F6833C0BE57; Wed, 17 May 2023 22:10:31 +0000 (UTC) Received: from localhost (unknown [10.39.192.14]) by smtp.corp.redhat.com (Postfix) with ESMTP id B3F2614171C0; Wed, 17 May 2023 22:10:30 +0000 (UTC) X-Outflank-Mailman: Message body and most headers restored to incoming version X-BeenThere: xen-devel@lists.xenproject.org List-Id: Xen developer discussion List-Unsubscribe: , List-Post: List-Help: List-Subscribe: , Errors-To: xen-devel-bounces@lists.xenproject.org Precedence: list Sender: "Xen-devel" X-Inumbo-ID: a639c45e-f4ff-11ed-b22a-6b7b168915f2 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1684361435; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=gFn0D1QeS7K9r5hk8We93ICMQFKJ73KryVejH0WL6Ew=; b=ZZqmvM2fOa51dy/lEyym2F0cKXavI3OdGqa87KJGo54j/9zAiOMmfVu4r29yppxmq23am8 D0Rpt1N7hoXfyQ4aRhQPNLqOfHDNP2S1YkIpYMTHps2q3udYjbJ/UQsB7Ns3DXSOBHw9vj 3AzaBs42j1us/eOQaGLu5nS5imBhrsI= X-MC-Unique: C0RqvpuqMluZCT-d_XeyiQ-1 From: Stefan Hajnoczi To: qemu-devel@nongnu.org Cc: xen-devel@lists.xenproject.org, Julia Suvorova , "Michael S. Tsirkin" , Stefano Stabellini , Stefano Garzarella , Kevin Wolf , Fam Zheng , =?UTF-8?q?Philippe=20Mathieu-Daud=C3=A9?= , Paolo Bonzini , qemu-block@nongnu.org, Paul Durrant , Anthony Perard , Aarushi Mehta , Stefan Hajnoczi , Hanna Reitz Subject: [PATCH 2/6] block/nvme: convert to blk_io_plug_call() API Date: Wed, 17 May 2023 18:10:18 -0400 Message-Id: <20230517221022.325091-3-stefanha@redhat.com> In-Reply-To: <20230517221022.325091-1-stefanha@redhat.com> References: <20230517221022.325091-1-stefanha@redhat.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable X-Scanned-By: MIMEDefang 3.1 on 10.11.54.7 X-ZohoMail-DKIM: pass (identity @redhat.com) X-ZM-MESSAGEID: 1684361483513100009 Content-Type: text/plain; charset="utf-8" Stop using the .bdrv_co_io_plug() API because it is not multi-queue block layer friendly. Use the new blk_io_plug_call() API to batch I/O submission instead. Signed-off-by: Stefan Hajnoczi Reviewed-by: Eric Blake --- block/nvme.c | 44 ++++++++++++-------------------------------- 1 file changed, 12 insertions(+), 32 deletions(-) diff --git a/block/nvme.c b/block/nvme.c index 5b744c2bda..100b38b592 100644 --- a/block/nvme.c +++ b/block/nvme.c @@ -25,6 +25,7 @@ #include "qemu/vfio-helpers.h" #include "block/block-io.h" #include "block/block_int.h" +#include "sysemu/block-backend.h" #include "sysemu/replay.h" #include "trace.h" =20 @@ -119,7 +120,6 @@ struct BDRVNVMeState { int blkshift; =20 uint64_t max_transfer; - bool plugged; =20 bool supports_write_zeroes; bool supports_discard; @@ -282,7 +282,7 @@ static void nvme_kick(NVMeQueuePair *q) { BDRVNVMeState *s =3D q->s; =20 - if (s->plugged || !q->need_kick) { + if (!q->need_kick) { return; } trace_nvme_kick(s, q->index); @@ -387,10 +387,6 @@ static bool nvme_process_completion(NVMeQueuePair *q) NvmeCqe *c; =20 trace_nvme_process_completion(s, q->index, q->inflight); - if (s->plugged) { - trace_nvme_process_completion_queue_plugged(s, q->index); - return false; - } =20 /* * Support re-entrancy when a request cb() function invokes aio_poll(). @@ -480,6 +476,15 @@ static void nvme_trace_command(const NvmeCmd *cmd) } } =20 +static void nvme_unplug_fn(void *opaque) +{ + NVMeQueuePair *q =3D opaque; + + QEMU_LOCK_GUARD(&q->lock); + nvme_kick(q); + nvme_process_completion(q); +} + static void nvme_submit_command(NVMeQueuePair *q, NVMeRequest *req, NvmeCmd *cmd, BlockCompletionFunc cb, void *opaque) @@ -496,8 +501,7 @@ static void nvme_submit_command(NVMeQueuePair *q, NVMeR= equest *req, q->sq.tail * NVME_SQ_ENTRY_BYTES, cmd, sizeof(*cmd)); q->sq.tail =3D (q->sq.tail + 1) % NVME_QUEUE_SIZE; q->need_kick++; - nvme_kick(q); - nvme_process_completion(q); + blk_io_plug_call(nvme_unplug_fn, q); qemu_mutex_unlock(&q->lock); } =20 @@ -1567,27 +1571,6 @@ static void nvme_attach_aio_context(BlockDriverState= *bs, } } =20 -static void coroutine_fn nvme_co_io_plug(BlockDriverState *bs) -{ - BDRVNVMeState *s =3D bs->opaque; - assert(!s->plugged); - s->plugged =3D true; -} - -static void coroutine_fn nvme_co_io_unplug(BlockDriverState *bs) -{ - BDRVNVMeState *s =3D bs->opaque; - assert(s->plugged); - s->plugged =3D false; - for (unsigned i =3D INDEX_IO(0); i < s->queue_count; i++) { - NVMeQueuePair *q =3D s->queues[i]; - qemu_mutex_lock(&q->lock); - nvme_kick(q); - nvme_process_completion(q); - qemu_mutex_unlock(&q->lock); - } -} - static bool nvme_register_buf(BlockDriverState *bs, void *host, size_t siz= e, Error **errp) { @@ -1664,9 +1647,6 @@ static BlockDriver bdrv_nvme =3D { .bdrv_detach_aio_context =3D nvme_detach_aio_context, .bdrv_attach_aio_context =3D nvme_attach_aio_context, =20 - .bdrv_co_io_plug =3D nvme_co_io_plug, - .bdrv_co_io_unplug =3D nvme_co_io_unplug, - .bdrv_register_buf =3D nvme_register_buf, .bdrv_unregister_buf =3D nvme_unregister_buf, }; --=20 2.40.1 From nobody Wed Apr 24 14:58:26 2024 Delivered-To: importer@patchew.org Received-SPF: pass (zohomail.com: domain of lists.xenproject.org designates 192.237.175.120 as permitted sender) client-ip=192.237.175.120; envelope-from=xen-devel-bounces@lists.xenproject.org; helo=lists.xenproject.org; Authentication-Results: mx.zohomail.com; dkim=pass; spf=pass (zohomail.com: domain of lists.xenproject.org designates 192.237.175.120 as permitted sender) smtp.mailfrom=xen-devel-bounces@lists.xenproject.org; dmarc=pass(p=none dis=none) header.from=redhat.com ARC-Seal: i=1; a=rsa-sha256; t=1684361479; cv=none; d=zohomail.com; s=zohoarc; b=TYoYutjgKsiuvhInhM7MWobzacwArjMiiWxAsI38F4LG6Mwg93gbT52gJTd9+L6GcMba/6EW+aZG6HSjENp8g01pXPNRGuiJEaaGTvm5xXLbYhHTmeDnZ6yG1msqaxXP66zKVftNhXKm9gH2hmZNP7in2sH7t9+SHp1/5zhJ5wg= ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=zohomail.com; s=zohoarc; t=1684361479; h=Content-Transfer-Encoding:Cc:Date:From:In-Reply-To:List-Subscribe:List-Post:List-Id:List-Help:List-Unsubscribe:MIME-Version:Message-ID:References:Sender:Subject:To; bh=RXMA7sqy2vuXoTTqUl+VUv8v7FFhFPNT89EZySBaASY=; b=ZLkseFf+ttLnqN/4fPSysXr7qZML8AELXhHhWpGhm9y4DOiwgXfqp1VxoMSrYeYrcyG/7JTmfjTgaAcgHujIydkbUQfjC3iereCVRAD6+C/6Vtp45zZbIjB0UI2MHHY2hO2IEKPnAeuymC/QedZcuJC+FIo7qBsJTiCZHIFVYb8= ARC-Authentication-Results: i=1; mx.zohomail.com; dkim=pass; spf=pass (zohomail.com: domain of lists.xenproject.org designates 192.237.175.120 as permitted sender) smtp.mailfrom=xen-devel-bounces@lists.xenproject.org; dmarc=pass header.from= (p=none dis=none) Return-Path: Received: from lists.xenproject.org (lists.xenproject.org [192.237.175.120]) by mx.zohomail.com with SMTPS id 1684361479238472.3861629539191; Wed, 17 May 2023 15:11:19 -0700 (PDT) Received: from list by lists.xenproject.org with outflank-mailman.536082.834220 (Exim 4.92) (envelope-from ) id 1pzPM0-0005bT-QJ; Wed, 17 May 2023 22:10:40 +0000 Received: by outflank-mailman (output) from mailman id 536082.834220; Wed, 17 May 2023 22:10:40 +0000 Received: from localhost ([127.0.0.1] helo=lists.xenproject.org) by lists.xenproject.org with esmtp (Exim 4.92) (envelope-from ) id 1pzPM0-0005bJ-NK; Wed, 17 May 2023 22:10:40 +0000 Received: by outflank-mailman (input) for mailman id 536082; Wed, 17 May 2023 22:10:39 +0000 Received: from se1-gles-sth1-in.inumbo.com ([159.253.27.254] helo=se1-gles-sth1.inumbo.com) by lists.xenproject.org with esmtp (Exim 4.92) (envelope-from ) id 1pzPLz-0004sP-Ok for xen-devel@lists.xenproject.org; Wed, 17 May 2023 22:10:39 +0000 Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.133.124]) by se1-gles-sth1.inumbo.com (Halon) with ESMTPS id a78f9fea-f4ff-11ed-b22a-6b7b168915f2; Thu, 18 May 2023 00:10:39 +0200 (CEST) Received: from mimecast-mx02.redhat.com (mimecast-mx02.redhat.com [66.187.233.88]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id us-mta-274-PzJp_SQIOtapVDQqpnz-qw-1; Wed, 17 May 2023 18:10:34 -0400 Received: from smtp.corp.redhat.com (int-mx06.intmail.prod.int.rdu2.redhat.com [10.11.54.6]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx02.redhat.com (Postfix) with ESMTPS id BAB96857FB2; Wed, 17 May 2023 22:10:33 +0000 (UTC) Received: from localhost (unknown [10.39.192.14]) by smtp.corp.redhat.com (Postfix) with ESMTP id 1EF702166B31; Wed, 17 May 2023 22:10:32 +0000 (UTC) X-Outflank-Mailman: Message body and most headers restored to incoming version X-BeenThere: xen-devel@lists.xenproject.org List-Id: Xen developer discussion List-Unsubscribe: , List-Post: List-Help: List-Subscribe: , Errors-To: xen-devel-bounces@lists.xenproject.org Precedence: list Sender: "Xen-devel" X-Inumbo-ID: a78f9fea-f4ff-11ed-b22a-6b7b168915f2 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1684361438; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=RXMA7sqy2vuXoTTqUl+VUv8v7FFhFPNT89EZySBaASY=; b=e8S3k/NQQUr+/Q52PVXxi9RUiPOH8ulNON7h0Nhd4Ewvpbi70kPdGP4T+NZ/MOm1FUdkaz kLiU7y2ExV9qAyFf5c2bi0gOKnmFUm/AreJ0nyFS/CSWjccRBESx9gIPZ2p8AxPudAyBnn aZ9I+zRlNO4HQ6uXYcHZNzfXtM7RHmI= X-MC-Unique: PzJp_SQIOtapVDQqpnz-qw-1 From: Stefan Hajnoczi To: qemu-devel@nongnu.org Cc: xen-devel@lists.xenproject.org, Julia Suvorova , "Michael S. Tsirkin" , Stefano Stabellini , Stefano Garzarella , Kevin Wolf , Fam Zheng , =?UTF-8?q?Philippe=20Mathieu-Daud=C3=A9?= , Paolo Bonzini , qemu-block@nongnu.org, Paul Durrant , Anthony Perard , Aarushi Mehta , Stefan Hajnoczi , Hanna Reitz Subject: [PATCH 3/6] block/blkio: convert to blk_io_plug_call() API Date: Wed, 17 May 2023 18:10:19 -0400 Message-Id: <20230517221022.325091-4-stefanha@redhat.com> In-Reply-To: <20230517221022.325091-1-stefanha@redhat.com> References: <20230517221022.325091-1-stefanha@redhat.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable X-Scanned-By: MIMEDefang 3.1 on 10.11.54.6 X-ZohoMail-DKIM: pass (identity @redhat.com) X-ZM-MESSAGEID: 1684361480358100005 Content-Type: text/plain; charset="utf-8" Stop using the .bdrv_co_io_plug() API because it is not multi-queue block layer friendly. Use the new blk_io_plug_call() API to batch I/O submission instead. Signed-off-by: Stefan Hajnoczi Reviewed-by: Eric Blake --- block/blkio.c | 40 +++++++++++++++++++++------------------- 1 file changed, 21 insertions(+), 19 deletions(-) diff --git a/block/blkio.c b/block/blkio.c index 0cdc99a729..f2a1dc1fb2 100644 --- a/block/blkio.c +++ b/block/blkio.c @@ -325,16 +325,28 @@ static void blkio_detach_aio_context(BlockDriverState= *bs) false, NULL, NULL, NULL, NULL, NULL); } =20 -/* Call with s->blkio_lock held to submit I/O after enqueuing a new reques= t */ -static void blkio_submit_io(BlockDriverState *bs) +/* + * Called by blk_io_unplug() or immediately if not plugged. Called without + * blkio_lock. + */ +static void blkio_unplug_fn(BlockDriverState *bs) { - if (qatomic_read(&bs->io_plugged) =3D=3D 0) { - BDRVBlkioState *s =3D bs->opaque; + BDRVBlkioState *s =3D bs->opaque; =20 + WITH_QEMU_LOCK_GUARD(&s->blkio_lock) { blkioq_do_io(s->blkioq, NULL, 0, 0, NULL); } } =20 +/* + * Schedule I/O submission after enqueuing a new request. Called without + * blkio_lock. + */ +static void blkio_submit_io(BlockDriverState *bs) +{ + blk_io_plug_call(blkio_unplug_fn, bs); +} + static int coroutine_fn blkio_co_pdiscard(BlockDriverState *bs, int64_t offset, int64_t bytes) { @@ -345,9 +357,9 @@ blkio_co_pdiscard(BlockDriverState *bs, int64_t offset,= int64_t bytes) =20 WITH_QEMU_LOCK_GUARD(&s->blkio_lock) { blkioq_discard(s->blkioq, offset, bytes, &cod, 0); - blkio_submit_io(bs); } =20 + blkio_submit_io(bs); qemu_coroutine_yield(); return cod.ret; } @@ -378,9 +390,9 @@ blkio_co_preadv(BlockDriverState *bs, int64_t offset, i= nt64_t bytes, =20 WITH_QEMU_LOCK_GUARD(&s->blkio_lock) { blkioq_readv(s->blkioq, offset, iov, iovcnt, &cod, 0); - blkio_submit_io(bs); } =20 + blkio_submit_io(bs); qemu_coroutine_yield(); =20 if (use_bounce_buffer) { @@ -423,9 +435,9 @@ static int coroutine_fn blkio_co_pwritev(BlockDriverSta= te *bs, int64_t offset, =20 WITH_QEMU_LOCK_GUARD(&s->blkio_lock) { blkioq_writev(s->blkioq, offset, iov, iovcnt, &cod, blkio_flags); - blkio_submit_io(bs); } =20 + blkio_submit_io(bs); qemu_coroutine_yield(); =20 if (use_bounce_buffer) { @@ -444,9 +456,9 @@ static int coroutine_fn blkio_co_flush(BlockDriverState= *bs) =20 WITH_QEMU_LOCK_GUARD(&s->blkio_lock) { blkioq_flush(s->blkioq, &cod, 0); - blkio_submit_io(bs); } =20 + blkio_submit_io(bs); qemu_coroutine_yield(); return cod.ret; } @@ -472,22 +484,13 @@ static int coroutine_fn blkio_co_pwrite_zeroes(BlockD= riverState *bs, =20 WITH_QEMU_LOCK_GUARD(&s->blkio_lock) { blkioq_write_zeroes(s->blkioq, offset, bytes, &cod, blkio_flags); - blkio_submit_io(bs); } =20 + blkio_submit_io(bs); qemu_coroutine_yield(); return cod.ret; } =20 -static void coroutine_fn blkio_co_io_unplug(BlockDriverState *bs) -{ - BDRVBlkioState *s =3D bs->opaque; - - WITH_QEMU_LOCK_GUARD(&s->blkio_lock) { - blkio_submit_io(bs); - } -} - typedef enum { BMRR_OK, BMRR_SKIP, @@ -1009,7 +1012,6 @@ static void blkio_refresh_limits(BlockDriverState *bs= , Error **errp) .bdrv_co_pwritev =3D blkio_co_pwritev, \ .bdrv_co_flush_to_disk =3D blkio_co_flush, \ .bdrv_co_pwrite_zeroes =3D blkio_co_pwrite_zeroes, \ - .bdrv_co_io_unplug =3D blkio_co_io_unplug, \ .bdrv_refresh_limits =3D blkio_refresh_limits, \ .bdrv_register_buf =3D blkio_register_buf, \ .bdrv_unregister_buf =3D blkio_unregister_buf, \ --=20 2.40.1 From nobody Wed Apr 24 14:58:26 2024 Delivered-To: importer@patchew.org Received-SPF: pass (zohomail.com: domain of lists.xenproject.org designates 192.237.175.120 as permitted sender) client-ip=192.237.175.120; envelope-from=xen-devel-bounces@lists.xenproject.org; helo=lists.xenproject.org; Authentication-Results: mx.zohomail.com; dkim=pass; spf=pass (zohomail.com: domain of lists.xenproject.org designates 192.237.175.120 as permitted sender) smtp.mailfrom=xen-devel-bounces@lists.xenproject.org; dmarc=pass(p=none dis=none) header.from=redhat.com ARC-Seal: i=1; a=rsa-sha256; t=1684361478; cv=none; d=zohomail.com; s=zohoarc; b=Kddk/FnXe8M/cSKqctUiRGUF27NRg3N5hN2k6BckUPcsxzVHRE4XdjANgYPG4Z9r0mZIQfEhLeQK9FMSZTvytJ6XA3KLAkIK+IuaO1h9+Y7txnVlE/uUMsMC8PU6A3G4IGGzIui4Ea6w2nkv4tLXMeg9w00wUzE2NiNJbRk3wP8= ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=zohomail.com; s=zohoarc; t=1684361478; h=Content-Transfer-Encoding:Cc:Date:From:In-Reply-To:List-Subscribe:List-Post:List-Id:List-Help:List-Unsubscribe:MIME-Version:Message-ID:References:Sender:Subject:To; bh=Blp/dScqtxXcjtaEje+/rVMH7PoeKSgRMwItRYa7aR4=; b=SC0dQ/AMQmEl7idCdQDSsT4DolTNSFVLWaznQPnRgWV3AaP1dilCysatbYMQPcV2w9RmL3zjb9y9nmCvEYYehEh89mroeN8o+3Xivzz4GGqpD9O2BFUjw0P6RuzwNK9MnKQuVhfqP0LZ+9OEZs2KFzYp5+u8+s4mzI8exl7YSro= ARC-Authentication-Results: i=1; mx.zohomail.com; dkim=pass; spf=pass (zohomail.com: domain of lists.xenproject.org designates 192.237.175.120 as permitted sender) smtp.mailfrom=xen-devel-bounces@lists.xenproject.org; dmarc=pass header.from= (p=none dis=none) Return-Path: Received: from lists.xenproject.org (lists.xenproject.org [192.237.175.120]) by mx.zohomail.com with SMTPS id 1684361478220775.4571956930777; Wed, 17 May 2023 15:11:18 -0700 (PDT) Received: from list by lists.xenproject.org with outflank-mailman.536083.834231 (Exim 4.92) (envelope-from ) id 1pzPM4-0005wt-Bj; Wed, 17 May 2023 22:10:44 +0000 Received: by outflank-mailman (output) from mailman id 536083.834231; Wed, 17 May 2023 22:10:44 +0000 Received: from localhost ([127.0.0.1] helo=lists.xenproject.org) by lists.xenproject.org with esmtp (Exim 4.92) (envelope-from ) id 1pzPM4-0005wd-69; Wed, 17 May 2023 22:10:44 +0000 Received: by outflank-mailman (input) for mailman id 536083; Wed, 17 May 2023 22:10:43 +0000 Received: from se1-gles-flk1-in.inumbo.com ([94.247.172.50] helo=se1-gles-flk1.inumbo.com) by lists.xenproject.org with esmtp (Exim 4.92) (envelope-from ) id 1pzPM3-0004sV-0H for xen-devel@lists.xenproject.org; Wed, 17 May 2023 22:10:43 +0000 Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.133.124]) by se1-gles-flk1.inumbo.com (Halon) with ESMTPS id a8b30104-f4ff-11ed-8611-37d641c3527e; Thu, 18 May 2023 00:10:41 +0200 (CEST) Received: from mimecast-mx02.redhat.com (mimecast-mx02.redhat.com [66.187.233.88]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id us-mta-500-iuCQhDZoOHmdirDNUMxiUg-1; Wed, 17 May 2023 18:10:36 -0400 Received: from smtp.corp.redhat.com (int-mx02.intmail.prod.int.rdu2.redhat.com [10.11.54.2]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx02.redhat.com (Postfix) with ESMTPS id 34265101A54F; Wed, 17 May 2023 22:10:36 +0000 (UTC) Received: from localhost (unknown [10.39.192.14]) by smtp.corp.redhat.com (Postfix) with ESMTP id 8660D40C6EC4; Wed, 17 May 2023 22:10:35 +0000 (UTC) X-Outflank-Mailman: Message body and most headers restored to incoming version X-BeenThere: xen-devel@lists.xenproject.org List-Id: Xen developer discussion List-Unsubscribe: , List-Post: List-Help: List-Subscribe: , Errors-To: xen-devel-bounces@lists.xenproject.org Precedence: list Sender: "Xen-devel" X-Inumbo-ID: a8b30104-f4ff-11ed-8611-37d641c3527e DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1684361440; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=Blp/dScqtxXcjtaEje+/rVMH7PoeKSgRMwItRYa7aR4=; b=RyxjFtLx6ZGCwF0gH2t7Mcrg5hJdfkf//cJheEgAcDX5wfwZ5k8wcl6R44dnECe10LDYO1 lda7AcE07j3Y2qTzlh7G8kfYOOIHbAy1j1fDPH/qgfe60NZ/HGxaJHtznQ5/eo5Lc+dAzi mWyKICdj0xKQZ7mIizyyjxcVDtpa9/U= X-MC-Unique: iuCQhDZoOHmdirDNUMxiUg-1 From: Stefan Hajnoczi To: qemu-devel@nongnu.org Cc: xen-devel@lists.xenproject.org, Julia Suvorova , "Michael S. Tsirkin" , Stefano Stabellini , Stefano Garzarella , Kevin Wolf , Fam Zheng , =?UTF-8?q?Philippe=20Mathieu-Daud=C3=A9?= , Paolo Bonzini , qemu-block@nongnu.org, Paul Durrant , Anthony Perard , Aarushi Mehta , Stefan Hajnoczi , Hanna Reitz Subject: [PATCH 4/6] block/io_uring: convert to blk_io_plug_call() API Date: Wed, 17 May 2023 18:10:20 -0400 Message-Id: <20230517221022.325091-5-stefanha@redhat.com> In-Reply-To: <20230517221022.325091-1-stefanha@redhat.com> References: <20230517221022.325091-1-stefanha@redhat.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable X-Scanned-By: MIMEDefang 3.1 on 10.11.54.2 X-ZohoMail-DKIM: pass (identity @redhat.com) X-ZM-MESSAGEID: 1684361480255100001 Content-Type: text/plain; charset="utf-8" Stop using the .bdrv_co_io_plug() API because it is not multi-queue block layer friendly. Use the new blk_io_plug_call() API to batch I/O submission instead. Signed-off-by: Stefan Hajnoczi Reviewed-by: Eric Blake --- include/block/raw-aio.h | 7 ------- block/file-posix.c | 10 --------- block/io_uring.c | 45 ++++++++++++++++------------------------- block/trace-events | 5 ++--- 4 files changed, 19 insertions(+), 48 deletions(-) diff --git a/include/block/raw-aio.h b/include/block/raw-aio.h index 0fe85ade77..da60ca13ef 100644 --- a/include/block/raw-aio.h +++ b/include/block/raw-aio.h @@ -81,13 +81,6 @@ int coroutine_fn luring_co_submit(BlockDriverState *bs, = int fd, uint64_t offset, QEMUIOVector *qiov, int type); void luring_detach_aio_context(LuringState *s, AioContext *old_context); void luring_attach_aio_context(LuringState *s, AioContext *new_context); - -/* - * luring_io_plug/unplug work in the thread's current AioContext, therefor= e the - * caller must ensure that they are paired in the same IOThread. - */ -void luring_io_plug(void); -void luring_io_unplug(void); #endif =20 #ifdef _WIN32 diff --git a/block/file-posix.c b/block/file-posix.c index 0ab158efba..7baa8491dd 100644 --- a/block/file-posix.c +++ b/block/file-posix.c @@ -2558,11 +2558,6 @@ static void coroutine_fn raw_co_io_plug(BlockDriverS= tate *bs) laio_io_plug(); } #endif -#ifdef CONFIG_LINUX_IO_URING - if (s->use_linux_io_uring) { - luring_io_plug(); - } -#endif } =20 static void coroutine_fn raw_co_io_unplug(BlockDriverState *bs) @@ -2573,11 +2568,6 @@ static void coroutine_fn raw_co_io_unplug(BlockDrive= rState *bs) laio_io_unplug(s->aio_max_batch); } #endif -#ifdef CONFIG_LINUX_IO_URING - if (s->use_linux_io_uring) { - luring_io_unplug(); - } -#endif } =20 static int coroutine_fn raw_co_flush_to_disk(BlockDriverState *bs) diff --git a/block/io_uring.c b/block/io_uring.c index 82cab6a5bd..9a45e5fb8b 100644 --- a/block/io_uring.c +++ b/block/io_uring.c @@ -16,6 +16,7 @@ #include "block/raw-aio.h" #include "qemu/coroutine.h" #include "qapi/error.h" +#include "sysemu/block-backend.h" #include "trace.h" =20 /* Only used for assertions. */ @@ -41,7 +42,6 @@ typedef struct LuringAIOCB { } LuringAIOCB; =20 typedef struct LuringQueue { - int plugged; unsigned int in_queue; unsigned int in_flight; bool blocked; @@ -267,7 +267,7 @@ static void luring_process_completions_and_submit(Lurin= gState *s) { luring_process_completions(s); =20 - if (!s->io_q.plugged && s->io_q.in_queue > 0) { + if (s->io_q.in_queue > 0) { ioq_submit(s); } } @@ -301,29 +301,17 @@ static void qemu_luring_poll_ready(void *opaque) static void ioq_init(LuringQueue *io_q) { QSIMPLEQ_INIT(&io_q->submit_queue); - io_q->plugged =3D 0; io_q->in_queue =3D 0; io_q->in_flight =3D 0; io_q->blocked =3D false; } =20 -void luring_io_plug(void) +static void luring_unplug_fn(void *opaque) { - AioContext *ctx =3D qemu_get_current_aio_context(); - LuringState *s =3D aio_get_linux_io_uring(ctx); - trace_luring_io_plug(s); - s->io_q.plugged++; -} - -void luring_io_unplug(void) -{ - AioContext *ctx =3D qemu_get_current_aio_context(); - LuringState *s =3D aio_get_linux_io_uring(ctx); - assert(s->io_q.plugged); - trace_luring_io_unplug(s, s->io_q.blocked, s->io_q.plugged, - s->io_q.in_queue, s->io_q.in_flight); - if (--s->io_q.plugged =3D=3D 0 && - !s->io_q.blocked && s->io_q.in_queue > 0) { + LuringState *s =3D opaque; + trace_luring_unplug_fn(s, s->io_q.blocked, s->io_q.in_queue, + s->io_q.in_flight); + if (!s->io_q.blocked && s->io_q.in_queue > 0) { ioq_submit(s); } } @@ -337,7 +325,6 @@ void luring_io_unplug(void) * @type: type of request * * Fetches sqes from ring, adds to pending queue and preps them - * */ static int luring_do_submit(int fd, LuringAIOCB *luringcb, LuringState *s, uint64_t offset, int type) @@ -370,14 +357,16 @@ static int luring_do_submit(int fd, LuringAIOCB *luri= ngcb, LuringState *s, =20 QSIMPLEQ_INSERT_TAIL(&s->io_q.submit_queue, luringcb, next); s->io_q.in_queue++; - trace_luring_do_submit(s, s->io_q.blocked, s->io_q.plugged, - s->io_q.in_queue, s->io_q.in_flight); - if (!s->io_q.blocked && - (!s->io_q.plugged || - s->io_q.in_flight + s->io_q.in_queue >=3D MAX_ENTRIES)) { - ret =3D ioq_submit(s); - trace_luring_do_submit_done(s, ret); - return ret; + trace_luring_do_submit(s, s->io_q.blocked, s->io_q.in_queue, + s->io_q.in_flight); + if (!s->io_q.blocked) { + if (s->io_q.in_flight + s->io_q.in_queue >=3D MAX_ENTRIES) { + ret =3D ioq_submit(s); + trace_luring_do_submit_done(s, ret); + return ret; + } + + blk_io_plug_call(luring_unplug_fn, s); } return 0; } diff --git a/block/trace-events b/block/trace-events index 32665158d6..c22fb1ed43 100644 --- a/block/trace-events +++ b/block/trace-events @@ -64,9 +64,8 @@ file_paio_submit(void *acb, void *opaque, int64_t offset,= int count, int type) " # io_uring.c luring_init_state(void *s, size_t size) "s %p size %zu" luring_cleanup_state(void *s) "%p freed" -luring_io_plug(void *s) "LuringState %p plug" -luring_io_unplug(void *s, int blocked, int plugged, int queued, int inflig= ht) "LuringState %p blocked %d plugged %d queued %d inflight %d" -luring_do_submit(void *s, int blocked, int plugged, int queued, int inflig= ht) "LuringState %p blocked %d plugged %d queued %d inflight %d" +luring_unplug_fn(void *s, int blocked, int queued, int inflight) "LuringSt= ate %p blocked %d queued %d inflight %d" +luring_do_submit(void *s, int blocked, int queued, int inflight) "LuringSt= ate %p blocked %d queued %d inflight %d" luring_do_submit_done(void *s, int ret) "LuringState %p submitted to kerne= l %d" luring_co_submit(void *bs, void *s, void *luringcb, int fd, uint64_t offse= t, size_t nbytes, int type) "bs %p s %p luringcb %p fd %d offset %" PRId64 = " nbytes %zd type %d" luring_process_completion(void *s, void *aiocb, int ret) "LuringState %p l= uringcb %p ret %d" --=20 2.40.1 From nobody Wed Apr 24 14:58:26 2024 Delivered-To: importer@patchew.org Received-SPF: pass (zohomail.com: domain of lists.xenproject.org designates 192.237.175.120 as permitted sender) client-ip=192.237.175.120; envelope-from=xen-devel-bounces@lists.xenproject.org; helo=lists.xenproject.org; Authentication-Results: mx.zohomail.com; dkim=pass; spf=pass (zohomail.com: domain of lists.xenproject.org designates 192.237.175.120 as permitted sender) smtp.mailfrom=xen-devel-bounces@lists.xenproject.org; dmarc=pass(p=none dis=none) header.from=redhat.com ARC-Seal: i=1; a=rsa-sha256; t=1684361480; cv=none; d=zohomail.com; s=zohoarc; b=hFzzat495gbRIkyT5T8oqrO37DJUJ404JB5MrgAx5kH9nKQE7X5xq/+P7903EKlnWJRKQaV53a1rmGYHxkslJfLCuMUfLAWEXNiGNtAgk5qW8rRzgSGGfxhTYTEXs8tC8Thl23L27gl/LngZB5JfFWdvMrL8Zg4M5/Jg/k2AYMI= ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=zohomail.com; s=zohoarc; t=1684361480; h=Content-Transfer-Encoding:Cc:Date:From:In-Reply-To:List-Subscribe:List-Post:List-Id:List-Help:List-Unsubscribe:MIME-Version:Message-ID:References:Sender:Subject:To; bh=wAL0IqCXxu/U5jmrcTSsPXlVFdfkdt5JIfR3AjEbRH4=; b=k6lTaPH/DpCdhp11r31OG8XdpQl7HS7+S8OGEZVk0DXY5N7TNDrKigKk6GLX2RuWo6Xd/6LW/CLaE8r0MscZS949g4Q4QvukmZBH/2y725R4AroKV5xiUxn7KBMj+eXmFV0QNnTYOj4WAaBqMJVrtYNbr0e+WLqDj5uDJ2xQzsw= ARC-Authentication-Results: i=1; mx.zohomail.com; dkim=pass; spf=pass (zohomail.com: domain of lists.xenproject.org designates 192.237.175.120 as permitted sender) smtp.mailfrom=xen-devel-bounces@lists.xenproject.org; dmarc=pass header.from= (p=none dis=none) Return-Path: Received: from lists.xenproject.org (lists.xenproject.org [192.237.175.120]) by mx.zohomail.com with SMTPS id 1684361480418830.2821456030238; Wed, 17 May 2023 15:11:20 -0700 (PDT) Received: from list by lists.xenproject.org with outflank-mailman.536084.834240 (Exim 4.92) (envelope-from ) id 1pzPM6-0006En-KS; Wed, 17 May 2023 22:10:46 +0000 Received: by outflank-mailman (output) from mailman id 536084.834240; Wed, 17 May 2023 22:10:46 +0000 Received: from localhost ([127.0.0.1] helo=lists.xenproject.org) by lists.xenproject.org with esmtp (Exim 4.92) (envelope-from ) id 1pzPM6-0006EY-Fb; Wed, 17 May 2023 22:10:46 +0000 Received: by outflank-mailman (input) for mailman id 536084; Wed, 17 May 2023 22:10:44 +0000 Received: from se1-gles-sth1-in.inumbo.com ([159.253.27.254] helo=se1-gles-sth1.inumbo.com) by lists.xenproject.org with esmtp (Exim 4.92) (envelope-from ) id 1pzPM4-0004sP-Nd for xen-devel@lists.xenproject.org; Wed, 17 May 2023 22:10:44 +0000 Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.133.124]) by se1-gles-sth1.inumbo.com (Halon) with ESMTPS id aa518528-f4ff-11ed-b22a-6b7b168915f2; Thu, 18 May 2023 00:10:44 +0200 (CEST) Received: from mimecast-mx02.redhat.com (mimecast-mx02.redhat.com [66.187.233.88]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id us-mta-84-iMg22q66OE2OQnfjQYyXYg-1; Wed, 17 May 2023 18:10:39 -0400 Received: from smtp.corp.redhat.com (int-mx06.intmail.prod.int.rdu2.redhat.com [10.11.54.6]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx02.redhat.com (Postfix) with ESMTPS id E198A870820; Wed, 17 May 2023 22:10:38 +0000 (UTC) Received: from localhost (unknown [10.39.192.14]) by smtp.corp.redhat.com (Postfix) with ESMTP id 592832166B31; Wed, 17 May 2023 22:10:38 +0000 (UTC) X-Outflank-Mailman: Message body and most headers restored to incoming version X-BeenThere: xen-devel@lists.xenproject.org List-Id: Xen developer discussion List-Unsubscribe: , List-Post: List-Help: List-Subscribe: , Errors-To: xen-devel-bounces@lists.xenproject.org Precedence: list Sender: "Xen-devel" X-Inumbo-ID: aa518528-f4ff-11ed-b22a-6b7b168915f2 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1684361442; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=wAL0IqCXxu/U5jmrcTSsPXlVFdfkdt5JIfR3AjEbRH4=; b=eUaOjwnN+BLneCHx6+v2uLzbz2heKkbp8hVGIKkrjqKPMsGVfE04zVgrWPy5DkyJYLZqOK C5GAzhcqm80uv6tTBi5WPC9gdNMNCVUhW/MEhjcutCaOhpjSrPKGzcT8J7AXps8taJN6Rd AhHmmlFYUOiYYLTxD6mw4Rl5/XrbvxE= X-MC-Unique: iMg22q66OE2OQnfjQYyXYg-1 From: Stefan Hajnoczi To: qemu-devel@nongnu.org Cc: xen-devel@lists.xenproject.org, Julia Suvorova , "Michael S. Tsirkin" , Stefano Stabellini , Stefano Garzarella , Kevin Wolf , Fam Zheng , =?UTF-8?q?Philippe=20Mathieu-Daud=C3=A9?= , Paolo Bonzini , qemu-block@nongnu.org, Paul Durrant , Anthony Perard , Aarushi Mehta , Stefan Hajnoczi , Hanna Reitz Subject: [PATCH 5/6] block/linux-aio: convert to blk_io_plug_call() API Date: Wed, 17 May 2023 18:10:21 -0400 Message-Id: <20230517221022.325091-6-stefanha@redhat.com> In-Reply-To: <20230517221022.325091-1-stefanha@redhat.com> References: <20230517221022.325091-1-stefanha@redhat.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable X-Scanned-By: MIMEDefang 3.1 on 10.11.54.6 X-ZohoMail-DKIM: pass (identity @redhat.com) X-ZM-MESSAGEID: 1684361481737100007 Content-Type: text/plain; charset="utf-8" Stop using the .bdrv_co_io_plug() API because it is not multi-queue block layer friendly. Use the new blk_io_plug_call() API to batch I/O submission instead. Signed-off-by: Stefan Hajnoczi Reviewed-by: Eric Blake --- include/block/raw-aio.h | 7 ------- block/file-posix.c | 28 ---------------------------- block/linux-aio.c | 41 +++++++++++------------------------------ 3 files changed, 11 insertions(+), 65 deletions(-) diff --git a/include/block/raw-aio.h b/include/block/raw-aio.h index da60ca13ef..0f63c2800c 100644 --- a/include/block/raw-aio.h +++ b/include/block/raw-aio.h @@ -62,13 +62,6 @@ int coroutine_fn laio_co_submit(int fd, uint64_t offset,= QEMUIOVector *qiov, =20 void laio_detach_aio_context(LinuxAioState *s, AioContext *old_context); void laio_attach_aio_context(LinuxAioState *s, AioContext *new_context); - -/* - * laio_io_plug/unplug work in the thread's current AioContext, therefore = the - * caller must ensure that they are paired in the same IOThread. - */ -void laio_io_plug(void); -void laio_io_unplug(uint64_t dev_max_batch); #endif /* io_uring.c - Linux io_uring implementation */ #ifdef CONFIG_LINUX_IO_URING diff --git a/block/file-posix.c b/block/file-posix.c index 7baa8491dd..ac1ed54811 100644 --- a/block/file-posix.c +++ b/block/file-posix.c @@ -2550,26 +2550,6 @@ static int coroutine_fn raw_co_pwritev(BlockDriverSt= ate *bs, int64_t offset, return raw_co_prw(bs, offset, bytes, qiov, QEMU_AIO_WRITE); } =20 -static void coroutine_fn raw_co_io_plug(BlockDriverState *bs) -{ - BDRVRawState __attribute__((unused)) *s =3D bs->opaque; -#ifdef CONFIG_LINUX_AIO - if (s->use_linux_aio) { - laio_io_plug(); - } -#endif -} - -static void coroutine_fn raw_co_io_unplug(BlockDriverState *bs) -{ - BDRVRawState __attribute__((unused)) *s =3D bs->opaque; -#ifdef CONFIG_LINUX_AIO - if (s->use_linux_aio) { - laio_io_unplug(s->aio_max_batch); - } -#endif -} - static int coroutine_fn raw_co_flush_to_disk(BlockDriverState *bs) { BDRVRawState *s =3D bs->opaque; @@ -3914,8 +3894,6 @@ BlockDriver bdrv_file =3D { .bdrv_co_copy_range_from =3D raw_co_copy_range_from, .bdrv_co_copy_range_to =3D raw_co_copy_range_to, .bdrv_refresh_limits =3D raw_refresh_limits, - .bdrv_co_io_plug =3D raw_co_io_plug, - .bdrv_co_io_unplug =3D raw_co_io_unplug, .bdrv_attach_aio_context =3D raw_aio_attach_aio_context, =20 .bdrv_co_truncate =3D raw_co_truncate, @@ -4286,8 +4264,6 @@ static BlockDriver bdrv_host_device =3D { .bdrv_co_copy_range_from =3D raw_co_copy_range_from, .bdrv_co_copy_range_to =3D raw_co_copy_range_to, .bdrv_refresh_limits =3D raw_refresh_limits, - .bdrv_co_io_plug =3D raw_co_io_plug, - .bdrv_co_io_unplug =3D raw_co_io_unplug, .bdrv_attach_aio_context =3D raw_aio_attach_aio_context, =20 .bdrv_co_truncate =3D raw_co_truncate, @@ -4424,8 +4400,6 @@ static BlockDriver bdrv_host_cdrom =3D { .bdrv_co_pwritev =3D raw_co_pwritev, .bdrv_co_flush_to_disk =3D raw_co_flush_to_disk, .bdrv_refresh_limits =3D cdrom_refresh_limits, - .bdrv_co_io_plug =3D raw_co_io_plug, - .bdrv_co_io_unplug =3D raw_co_io_unplug, .bdrv_attach_aio_context =3D raw_aio_attach_aio_context, =20 .bdrv_co_truncate =3D raw_co_truncate, @@ -4552,8 +4526,6 @@ static BlockDriver bdrv_host_cdrom =3D { .bdrv_co_pwritev =3D raw_co_pwritev, .bdrv_co_flush_to_disk =3D raw_co_flush_to_disk, .bdrv_refresh_limits =3D cdrom_refresh_limits, - .bdrv_co_io_plug =3D raw_co_io_plug, - .bdrv_co_io_unplug =3D raw_co_io_unplug, .bdrv_attach_aio_context =3D raw_aio_attach_aio_context, =20 .bdrv_co_truncate =3D raw_co_truncate, diff --git a/block/linux-aio.c b/block/linux-aio.c index 442c86209b..5021aed68f 100644 --- a/block/linux-aio.c +++ b/block/linux-aio.c @@ -15,6 +15,7 @@ #include "qemu/event_notifier.h" #include "qemu/coroutine.h" #include "qapi/error.h" +#include "sysemu/block-backend.h" =20 /* Only used for assertions. */ #include "qemu/coroutine_int.h" @@ -46,7 +47,6 @@ struct qemu_laiocb { }; =20 typedef struct { - int plugged; unsigned int in_queue; unsigned int in_flight; bool blocked; @@ -236,7 +236,7 @@ static void qemu_laio_process_completions_and_submit(Li= nuxAioState *s) { qemu_laio_process_completions(s); =20 - if (!s->io_q.plugged && !QSIMPLEQ_EMPTY(&s->io_q.pending)) { + if (!QSIMPLEQ_EMPTY(&s->io_q.pending)) { ioq_submit(s); } } @@ -277,7 +277,6 @@ static void qemu_laio_poll_ready(EventNotifier *opaque) static void ioq_init(LaioQueue *io_q) { QSIMPLEQ_INIT(&io_q->pending); - io_q->plugged =3D 0; io_q->in_queue =3D 0; io_q->in_flight =3D 0; io_q->blocked =3D false; @@ -354,31 +353,11 @@ static uint64_t laio_max_batch(LinuxAioState *s, uint= 64_t dev_max_batch) return max_batch; } =20 -void laio_io_plug(void) +static void laio_unplug_fn(void *opaque) { - AioContext *ctx =3D qemu_get_current_aio_context(); - LinuxAioState *s =3D aio_get_linux_aio(ctx); + LinuxAioState *s =3D opaque; =20 - s->io_q.plugged++; -} - -void laio_io_unplug(uint64_t dev_max_batch) -{ - AioContext *ctx =3D qemu_get_current_aio_context(); - LinuxAioState *s =3D aio_get_linux_aio(ctx); - - assert(s->io_q.plugged); - s->io_q.plugged--; - - /* - * Why max batch checking is performed here: - * Another BDS may have queued requests with a higher dev_max_batch and - * therefore in_queue could now exceed our dev_max_batch. Re-check the= max - * batch so we can honor our device's dev_max_batch. - */ - if (s->io_q.in_queue >=3D laio_max_batch(s, dev_max_batch) || - (!s->io_q.plugged && - !s->io_q.blocked && !QSIMPLEQ_EMPTY(&s->io_q.pending))) { + if (!s->io_q.blocked && !QSIMPLEQ_EMPTY(&s->io_q.pending)) { ioq_submit(s); } } @@ -410,10 +389,12 @@ static int laio_do_submit(int fd, struct qemu_laiocb = *laiocb, off_t offset, =20 QSIMPLEQ_INSERT_TAIL(&s->io_q.pending, laiocb, next); s->io_q.in_queue++; - if (!s->io_q.blocked && - (!s->io_q.plugged || - s->io_q.in_queue >=3D laio_max_batch(s, dev_max_batch))) { - ioq_submit(s); + if (!s->io_q.blocked) { + if (s->io_q.in_queue >=3D laio_max_batch(s, dev_max_batch)) { + ioq_submit(s); + } else { + blk_io_plug_call(laio_unplug_fn, s); + } } =20 return 0; --=20 2.40.1 From nobody Wed Apr 24 14:58:26 2024 Delivered-To: importer@patchew.org Authentication-Results: mx.zohomail.com; dkim=pass; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=pass(p=none dis=none) header.from=redhat.com ARC-Seal: i=1; a=rsa-sha256; t=1684361499; cv=none; d=zohomail.com; s=zohoarc; b=ff7Iib3+4UgIQHqqLrRzTbmwJQGEqgbTrW1pgEBI2q4dXMLsRpo7WVnPlqUDq0l5srcAdDyN6H5rqQFzp5cSfu60lxau5qoanJdwstyIPR4v8YSvw5rGxdD2WXJPn9nJxbelBHOXDZNMhGkc5D6hPuRN0xuOqOhEaCggp+B0F90= ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=zohomail.com; s=zohoarc; t=1684361499; h=Content-Transfer-Encoding:Cc:Date:From:In-Reply-To:List-Subscribe:List-Post:List-Id:List-Archive:List-Help:List-Unsubscribe:MIME-Version:Message-ID:References:Sender:Subject:To; bh=oAMrkGRvsGP9WognGT4SAJgv9CJG1xPxVEQ07vb4i/g=; b=K0Ud72g0+eDA22yoE0pEf5mx6VJR+JaAYRH8Cp383rli5ngdOuVCEkPV6DWf1zijyrsZpKtOlPsiz5bnapZ/Gg5Lbj/hmVWVVQTSiHecdDbvJ17LEJ6i+2bHvSH1IpCIRipDxUjoZYfscBFcXKADZ9KidvB6d2KaEQYBjUtvhMo= ARC-Authentication-Results: i=1; mx.zohomail.com; dkim=pass; spf=pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) smtp.mailfrom=qemu-devel-bounces+importer=patchew.org@nongnu.org; dmarc=pass header.from= (p=none dis=none) Return-Path: Received: from lists.gnu.org (lists.gnu.org [209.51.188.17]) by mx.zohomail.com with SMTPS id 1684361499210911.9921938427731; Wed, 17 May 2023 15:11:39 -0700 (PDT) Received: from localhost ([::1] helo=lists1p.gnu.org) by lists.gnu.org with esmtp (Exim 4.90_1) (envelope-from ) id 1pzPMC-0003ep-Gv; Wed, 17 May 2023 18:10:52 -0400 Received: from eggs.gnu.org ([2001:470:142:3::10]) by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1pzPMA-0003du-IY for qemu-devel@nongnu.org; Wed, 17 May 2023 18:10:50 -0400 Received: from us-smtp-delivery-124.mimecast.com ([170.10.129.124]) by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1pzPM8-0002Ud-3b for qemu-devel@nongnu.org; Wed, 17 May 2023 18:10:50 -0400 Received: from mimecast-mx02.redhat.com (mx3-rdu2.redhat.com [66.187.233.73]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id us-mta-283-ZuGjwf9_O02pHtA8FBXCrg-1; Wed, 17 May 2023 18:10:42 -0400 Received: from smtp.corp.redhat.com (int-mx04.intmail.prod.int.rdu2.redhat.com [10.11.54.4]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx02.redhat.com (Postfix) with ESMTPS id 6832838184E6; Wed, 17 May 2023 22:10:41 +0000 (UTC) Received: from localhost (unknown [10.39.192.14]) by smtp.corp.redhat.com (Postfix) with ESMTP id C65272026D16; Wed, 17 May 2023 22:10:40 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1684361447; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=oAMrkGRvsGP9WognGT4SAJgv9CJG1xPxVEQ07vb4i/g=; b=amJNMPzWL/RO4jJFq+LhTlN4+Wi1zf5Mp3q3lTdv24Qt80yy9RV9x2GIFhp0uIH23j4SFz QWiOyyp+V5sjDi0ulKMK53n8GaDhvp6nzojBYTHdebZZu/258vteTzinRuzDv8oKVzTPPN AJQ+KW/Fu2WgPy++Q5s3cqepqUhjyfs= X-MC-Unique: ZuGjwf9_O02pHtA8FBXCrg-1 From: Stefan Hajnoczi To: qemu-devel@nongnu.org Cc: xen-devel@lists.xenproject.org, Julia Suvorova , "Michael S. Tsirkin" , Stefano Stabellini , Stefano Garzarella , Kevin Wolf , Fam Zheng , =?UTF-8?q?Philippe=20Mathieu-Daud=C3=A9?= , Paolo Bonzini , qemu-block@nongnu.org, Paul Durrant , Anthony Perard , Aarushi Mehta , Stefan Hajnoczi , Hanna Reitz Subject: [PATCH 6/6] block: remove bdrv_co_io_plug() API Date: Wed, 17 May 2023 18:10:22 -0400 Message-Id: <20230517221022.325091-7-stefanha@redhat.com> In-Reply-To: <20230517221022.325091-1-stefanha@redhat.com> References: <20230517221022.325091-1-stefanha@redhat.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable X-Scanned-By: MIMEDefang 3.1 on 10.11.54.4 Received-SPF: pass (zohomail.com: domain of gnu.org designates 209.51.188.17 as permitted sender) client-ip=209.51.188.17; envelope-from=qemu-devel-bounces+importer=patchew.org@nongnu.org; helo=lists.gnu.org; Received-SPF: pass client-ip=170.10.129.124; envelope-from=stefanha@redhat.com; helo=us-smtp-delivery-124.mimecast.com X-Spam_score_int: -20 X-Spam_score: -2.1 X-Spam_bar: -- X-Spam_report: (-2.1 / 5.0 requ) BAYES_00=-1.9, DKIMWL_WL_HIGH=-0.001, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, DKIM_VALID_EF=-0.1, RCVD_IN_DNSWL_NONE=-0.0001, SPF_HELO_NONE=0.001, SPF_PASS=-0.001, T_SCC_BODY_TEXT_LINE=-0.01 autolearn=unavailable autolearn_force=no X-Spam_action: no action X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: qemu-devel-bounces+importer=patchew.org@nongnu.org Sender: qemu-devel-bounces+importer=patchew.org@nongnu.org X-ZohoMail-DKIM: pass (identity @redhat.com) X-ZM-MESSAGEID: 1684361499721100001 Content-Type: text/plain; charset="utf-8" No block driver implements .bdrv_co_io_plug() anymore. Get rid of the function pointers. Signed-off-by: Stefan Hajnoczi Reviewed-by: Eric Blake --- include/block/block-io.h | 3 --- include/block/block_int-common.h | 11 ---------- block/io.c | 37 -------------------------------- 3 files changed, 51 deletions(-) diff --git a/include/block/block-io.h b/include/block/block-io.h index a27e471a87..43af816d75 100644 --- a/include/block/block-io.h +++ b/include/block/block-io.h @@ -259,9 +259,6 @@ void coroutine_fn bdrv_co_leave(BlockDriverState *bs, A= ioContext *old_ctx); =20 AioContext *child_of_bds_get_parent_aio_context(BdrvChild *c); =20 -void coroutine_fn GRAPH_RDLOCK bdrv_co_io_plug(BlockDriverState *bs); -void coroutine_fn GRAPH_RDLOCK bdrv_co_io_unplug(BlockDriverState *bs); - bool coroutine_fn GRAPH_RDLOCK bdrv_co_can_store_new_dirty_bitmap(BlockDriverState *bs, const char *name, uint32_t granularity, Error **errp); diff --git a/include/block/block_int-common.h b/include/block/block_int-com= mon.h index dbec0e3bb4..fa369d83dd 100644 --- a/include/block/block_int-common.h +++ b/include/block/block_int-common.h @@ -753,11 +753,6 @@ struct BlockDriver { void coroutine_fn GRAPH_RDLOCK_PTR (*bdrv_co_debug_event)( BlockDriverState *bs, BlkdebugEvent event); =20 - /* io queue for linux-aio */ - void coroutine_fn GRAPH_RDLOCK_PTR (*bdrv_co_io_plug)(BlockDriverState= *bs); - void coroutine_fn GRAPH_RDLOCK_PTR (*bdrv_co_io_unplug)( - BlockDriverState *bs); - /** * bdrv_drain_begin is called if implemented in the beginning of a * drain operation to drain and stop any internal sources of requests = in @@ -1227,12 +1222,6 @@ struct BlockDriverState { unsigned int in_flight; unsigned int serialising_in_flight; =20 - /* - * counter for nested bdrv_io_plug. - * Accessed with atomic ops. - */ - unsigned io_plugged; - /* do we need to tell the quest if we have a volatile write cache? */ int enable_write_cache; =20 diff --git a/block/io.c b/block/io.c index 4d54fda593..56b0c1ce6c 100644 --- a/block/io.c +++ b/block/io.c @@ -3219,43 +3219,6 @@ void *qemu_try_blockalign0(BlockDriverState *bs, siz= e_t size) return mem; } =20 -void coroutine_fn bdrv_co_io_plug(BlockDriverState *bs) -{ - BdrvChild *child; - IO_CODE(); - assert_bdrv_graph_readable(); - - QLIST_FOREACH(child, &bs->children, next) { - bdrv_co_io_plug(child->bs); - } - - if (qatomic_fetch_inc(&bs->io_plugged) =3D=3D 0) { - BlockDriver *drv =3D bs->drv; - if (drv && drv->bdrv_co_io_plug) { - drv->bdrv_co_io_plug(bs); - } - } -} - -void coroutine_fn bdrv_co_io_unplug(BlockDriverState *bs) -{ - BdrvChild *child; - IO_CODE(); - assert_bdrv_graph_readable(); - - assert(bs->io_plugged); - if (qatomic_fetch_dec(&bs->io_plugged) =3D=3D 1) { - BlockDriver *drv =3D bs->drv; - if (drv && drv->bdrv_co_io_unplug) { - drv->bdrv_co_io_unplug(bs); - } - } - - QLIST_FOREACH(child, &bs->children, next) { - bdrv_co_io_unplug(child->bs); - } -} - /* Helper that undoes bdrv_register_buf() when it fails partway through */ static void GRAPH_RDLOCK bdrv_register_buf_rollback(BlockDriverState *bs, void *host, size_t size, --=20 2.40.1