1
The following changes since commit 9d662a6b22a0838a85c5432385f35db2488a33a5:
1
The following changes since commit c6a5fc2ac76c5ab709896ee1b0edd33685a67ed1:
2
2
3
Merge remote-tracking branch 'remotes/legoater/tags/pull-ppc-20220305' into staging (2022-03-05 18:03:15 +0000)
3
decodetree: Add --output-null for meson testing (2023-05-31 19:56:42 -0700)
4
4
5
are available in the Git repository at:
5
are available in the Git repository at:
6
6
7
https://gitlab.com/hreitz/qemu.git tags/pull-block-2022-03-07
7
https://gitlab.com/stefanha/qemu.git tags/block-pull-request
8
8
9
for you to fetch changes up to 743da0b401cdc3ee94bc519975e339a3cdbe0ad1:
9
for you to fetch changes up to 98b126f5e3228a346c774e569e26689943b401dd:
10
10
11
iotests/image-fleecing: test push backup with fleecing (2022-03-07 09:33:31 +0100)
11
qapi: add '@fdset' feature for BlockdevOptionsVirtioBlkVhostVdpa (2023-06-01 11:08:21 -0400)
12
12
13
----------------------------------------------------------------
13
----------------------------------------------------------------
14
Block patches for 7.0-rc0:
14
Pull request
15
- New fleecing backup scheme
15
16
- iotest fixes
16
- Stefano Garzarella's blkio block driver 'fd' parameter
17
- Fixes for the curl block driver
17
- My thread-local blk_io_plug() series
18
- Fix for the preallocate block driver
19
- IDE fix for zero-length TRIM requests
20
18
21
----------------------------------------------------------------
19
----------------------------------------------------------------
22
Hanna Reitz (2):
23
ide: Increment BB in-flight counter for TRIM BH
24
iotests: Write test output to TEST_DIR
25
20
26
Peter Maydell (2):
21
Stefan Hajnoczi (6):
27
block/curl.c: Set error message string if curl_init_state() fails
22
block: add blk_io_plug_call() API
28
block/curl.c: Check error return from curl_easy_setopt()
23
block/nvme: convert to blk_io_plug_call() API
24
block/blkio: convert to blk_io_plug_call() API
25
block/io_uring: convert to blk_io_plug_call() API
26
block/linux-aio: convert to blk_io_plug_call() API
27
block: remove bdrv_co_io_plug() API
29
28
30
Thomas Huth (2):
29
Stefano Garzarella (2):
31
tests/qemu-iotests/040: Skip TestCommitWithFilters without 'throttle'
30
block/blkio: use qemu_open() to support fd passing for virtio-blk
32
tests/qemu-iotests/testrunner: Quote "case not run" lines in TAP mode
31
qapi: add '@fdset' feature for BlockdevOptionsVirtioBlkVhostVdpa
33
32
34
Vladimir Sementsov-Ogievskiy (17):
33
MAINTAINERS | 1 +
35
block: fix preallocate filter: don't do unaligned preallocate requests
34
qapi/block-core.json | 6 ++
36
block/block-copy: move copy_bitmap initialization to
35
meson.build | 4 +
37
block_copy_state_new()
36
include/block/block-io.h | 3 -
38
block/dirty-bitmap: bdrv_merge_dirty_bitmap(): add return value
37
include/block/block_int-common.h | 11 ---
39
block/block-copy: block_copy_state_new(): add bitmap parameter
38
include/block/raw-aio.h | 14 ---
40
block/copy-before-write: add bitmap open parameter
39
include/sysemu/block-backend-io.h | 13 +--
41
block/block-copy: add block_copy_reset()
40
block/blkio.c | 96 ++++++++++++------
42
block: intoduce reqlist
41
block/block-backend.c | 22 -----
43
block/reqlist: reqlist_find_conflict(): use ranges_overlap()
42
block/file-posix.c | 38 -------
44
block/dirty-bitmap: introduce bdrv_dirty_bitmap_status()
43
block/io.c | 37 -------
45
block/reqlist: add reqlist_wait_all()
44
block/io_uring.c | 44 ++++-----
46
block/io: introduce block driver snapshot-access API
45
block/linux-aio.c | 41 +++-----
47
block: introduce snapshot-access block driver
46
block/nvme.c | 44 +++------
48
block: copy-before-write: realize snapshot-access API
47
block/plug.c | 159 ++++++++++++++++++++++++++++++
49
iotests/image-fleecing: add test-case for fleecing format node
48
hw/block/dataplane/xen-block.c | 8 +-
50
iotests.py: add qemu_io_pipe_and_status()
49
hw/block/virtio-blk.c | 4 +-
51
iotests/image-fleecing: add test case with bitmap
50
hw/scsi/virtio-scsi.c | 6 +-
52
iotests/image-fleecing: test push backup with fleecing
51
block/meson.build | 1 +
53
52
block/trace-events | 6 +-
54
qapi/block-core.json | 14 +-
53
20 files changed, 293 insertions(+), 265 deletions(-)
55
include/block/block-common.h | 3 +-
54
create mode 100644 block/plug.c
56
include/block/block-copy.h | 2 +
57
include/block/block_int-common.h | 24 ++
58
include/block/block_int-io.h | 9 +
59
include/block/dirty-bitmap.h | 4 +-
60
include/block/reqlist.h | 75 ++++++
61
include/qemu/hbitmap.h | 12 +
62
block/block-copy.c | 150 +++++------
63
block/copy-before-write.c | 265 +++++++++++++++++++-
64
block/curl.c | 92 ++++---
65
block/dirty-bitmap.c | 15 +-
66
block/io.c | 76 ++++++
67
block/monitor/bitmap-qmp-cmds.c | 5 +-
68
block/preallocate.c | 15 +-
69
block/reqlist.c | 85 +++++++
70
block/snapshot-access.c | 132 ++++++++++
71
hw/ide/core.c | 7 +
72
util/hbitmap.c | 33 +++
73
MAINTAINERS | 5 +-
74
block/meson.build | 2 +
75
tests/qemu-iotests/040 | 1 +
76
tests/qemu-iotests/257.out | 224 +++++++++++++++++
77
tests/qemu-iotests/common.rc | 6 +-
78
tests/qemu-iotests/iotests.py | 8 +-
79
tests/qemu-iotests/testenv.py | 5 +-
80
tests/qemu-iotests/testrunner.py | 19 +-
81
tests/qemu-iotests/tests/image-fleecing | 185 +++++++++++---
82
tests/qemu-iotests/tests/image-fleecing.out | 221 +++++++++++++++-
83
29 files changed, 1499 insertions(+), 195 deletions(-)
84
create mode 100644 include/block/reqlist.h
85
create mode 100644 block/reqlist.c
86
create mode 100644 block/snapshot-access.c
87
55
88
--
56
--
89
2.34.1
57
2.40.1
diff view generated by jsdifflib
Deleted patch
1
When we still have an AIOCB registered for DMA operations, we try to
2
settle the respective operation by draining the BlockBackend associated
3
with the IDE device.
4
1
5
However, this assumes that every DMA operation is associated with an
6
increment of the BlockBackend’s in-flight counter (e.g. through some
7
ongoing I/O operation), so that draining the BB until its in-flight
8
counter reaches 0 will settle all DMA operations. That is not the case:
9
For TRIM, the guest can issue a zero-length operation that will not
10
result in any I/O operation forwarded to the BlockBackend, and also not
11
increment the in-flight counter in any other way. In such a case,
12
blk_drain() will be a no-op if no other operations are in flight.
13
14
It is clear that if blk_drain() is a no-op, the value of
15
s->bus->dma->aiocb will not change between checking it in the `if`
16
condition and asserting that it is NULL after blk_drain().
17
18
The particular problem is that ide_issue_trim() creates a BH
19
(ide_trim_bh_cb()) to settle the TRIM request: iocb->common.cb() is
20
ide_dma_cb(), which will either create a new request, or find the
21
transfer to be done and call ide_set_inactive(), which clears
22
s->bus->dma->aiocb. Therefore, the blk_drain() must wait for
23
ide_trim_bh_cb() to run, which currently it will not always do.
24
25
To fix this issue, we increment the BlockBackend's in-flight counter
26
when the TRIM operation begins (in ide_issue_trim(), when the
27
ide_trim_bh_cb() BH is created) and decrement it when ide_trim_bh_cb()
28
is done.
29
30
Buglink: https://bugzilla.redhat.com/show_bug.cgi?id=2029980
31
Suggested-by: Paolo Bonzini <pbonzini@redhat.com>
32
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
33
Message-Id: <20220120142259.120189-1-hreitz@redhat.com>
34
Reviewed-by: Paolo Bonzini <pbonzini@redhat.com>
35
Reviewed-by: John Snow <jsnow@redhat.com>
36
Tested-by: John Snow <jsnow@redhat.com>
37
---
38
hw/ide/core.c | 7 +++++++
39
1 file changed, 7 insertions(+)
40
41
diff --git a/hw/ide/core.c b/hw/ide/core.c
42
index XXXXXXX..XXXXXXX 100644
43
--- a/hw/ide/core.c
44
+++ b/hw/ide/core.c
45
@@ -XXX,XX +XXX,XX @@ static const AIOCBInfo trim_aiocb_info = {
46
static void ide_trim_bh_cb(void *opaque)
47
{
48
TrimAIOCB *iocb = opaque;
49
+ BlockBackend *blk = iocb->s->blk;
50
51
iocb->common.cb(iocb->common.opaque, iocb->ret);
52
53
qemu_bh_delete(iocb->bh);
54
iocb->bh = NULL;
55
qemu_aio_unref(iocb);
56
+
57
+ /* Paired with an increment in ide_issue_trim() */
58
+ blk_dec_in_flight(blk);
59
}
60
61
static void ide_issue_trim_cb(void *opaque, int ret)
62
@@ -XXX,XX +XXX,XX @@ BlockAIOCB *ide_issue_trim(
63
IDEState *s = opaque;
64
TrimAIOCB *iocb;
65
66
+ /* Paired with a decrement in ide_trim_bh_cb() */
67
+ blk_inc_in_flight(s->blk);
68
+
69
iocb = blk_aio_get(&trim_aiocb_info, s->blk, cb, cb_opaque);
70
iocb->s = s;
71
iocb->bh = qemu_bh_new(ide_trim_bh_cb, iocb);
72
--
73
2.34.1
74
75
diff view generated by jsdifflib
Deleted patch
1
From: Peter Maydell <peter.maydell@linaro.org>
2
1
3
In curl_open(), the 'out' label assumes that the state->errmsg string
4
has been set (either by curl_easy_perform() or by manually copying a
5
string into it); however if curl_init_state() fails we will jump to
6
that label without setting the string. Add the missing error string
7
setup.
8
9
(We can't be specific about the cause of failure: the documentation
10
of curl_easy_init() just says "If this function returns NULL,
11
something went wrong".)
12
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
14
Message-Id: <20220222152341.850419-2-peter.maydell@linaro.org>
15
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
16
Reviewed-by: Hanna Reitz <hreitz@redhat.com>
17
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
18
---
19
block/curl.c | 2 ++
20
1 file changed, 2 insertions(+)
21
22
diff --git a/block/curl.c b/block/curl.c
23
index XXXXXXX..XXXXXXX 100644
24
--- a/block/curl.c
25
+++ b/block/curl.c
26
@@ -XXX,XX +XXX,XX @@ static int curl_open(BlockDriverState *bs, QDict *options, int flags,
27
// Get file size
28
29
if (curl_init_state(s, state) < 0) {
30
+ pstrcpy(state->errmsg, CURL_ERROR_SIZE,
31
+ "curl library initialization failed.");
32
goto out;
33
}
34
35
--
36
2.34.1
37
38
diff view generated by jsdifflib
1
From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
1
Introduce a new API for thread-local blk_io_plug() that does not
2
2
traverse the block graph. The goal is to make blk_io_plug() multi-queue
3
The new block driver simply utilizes snapshot-access API of underlying
3
friendly.
4
block node.
4
5
5
Instead of having block drivers track whether or not we're in a plugged
6
In further patches we want to use it like this:
6
section, provide an API that allows them to defer a function call until
7
7
we're unplugged: blk_io_plug_call(fn, opaque). If blk_io_plug_call() is
8
[guest] [NBD export]
8
called multiple times with the same fn/opaque pair, then fn() is only
9
| |
9
called once at the end of the function - resulting in batching.
10
| root | root
10
11
v file v
11
This patch introduces the API and changes blk_io_plug()/blk_io_unplug().
12
[copy-before-write]<------[snapshot-access]
12
blk_io_plug()/blk_io_unplug() no longer require a BlockBackend argument
13
| |
13
because the plug state is now thread-local.
14
| file | target
14
15
v v
15
Later patches convert block drivers to blk_io_plug_call() and then we
16
[active-disk] [temp.img]
16
can finally remove .bdrv_co_io_plug() once all block drivers have been
17
17
converted.
18
This way, NBD client will be able to read snapshotted state of active
18
19
disk, when active disk is continued to be written by guest. This is
19
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
20
known as "fleecing", and currently uses another scheme based on qcow2
20
Reviewed-by: Eric Blake <eblake@redhat.com>
21
temporary image which backing file is active-disk. New scheme comes
21
Reviewed-by: Stefano Garzarella <sgarzare@redhat.com>
22
with benefits - see next commit.
22
Acked-by: Kevin Wolf <kwolf@redhat.com>
23
23
Message-id: 20230530180959.1108766-2-stefanha@redhat.com
24
The other possible application is exporting internal snapshots of
24
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
25
qcow2, like this:
26
27
[guest] [NBD export]
28
| |
29
| root | root
30
v file v
31
[qcow2]<---------[snapshot-access]
32
33
For this, we'll need to implement snapshot-access API handlers in
34
qcow2 driver, and improve snapshot-access block driver (and API) to
35
make it possible to select snapshot by name. Another thing to improve
36
is size of snapshot. Now for simplicity we just use size of bs->file,
37
which is OK for backup, but for qcow2 snapshots export we'll need to
38
imporve snapshot-access API to get size of snapshot.
39
40
Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
41
Message-Id: <20220303194349.2304213-12-vsementsov@virtuozzo.com>
42
[hreitz: Rebased on block GS/IO split]
43
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
44
---
25
---
45
qapi/block-core.json | 4 +-
26
MAINTAINERS | 1 +
46
include/block/block_int-common.h | 6 ++
27
include/sysemu/block-backend-io.h | 13 +--
47
block/snapshot-access.c | 132 +++++++++++++++++++++++++++++++
28
block/block-backend.c | 22 -----
48
MAINTAINERS | 1 +
29
block/plug.c | 159 ++++++++++++++++++++++++++++++
49
block/meson.build | 1 +
30
hw/block/dataplane/xen-block.c | 8 +-
50
5 files changed, 143 insertions(+), 1 deletion(-)
31
hw/block/virtio-blk.c | 4 +-
51
create mode 100644 block/snapshot-access.c
32
hw/scsi/virtio-scsi.c | 6 +-
52
33
block/meson.build | 1 +
53
diff --git a/qapi/block-core.json b/qapi/block-core.json
34
8 files changed, 173 insertions(+), 41 deletions(-)
54
index XXXXXXX..XXXXXXX 100644
35
create mode 100644 block/plug.c
55
--- a/qapi/block-core.json
36
56
+++ b/qapi/block-core.json
37
diff --git a/MAINTAINERS b/MAINTAINERS
57
@@ -XXX,XX +XXX,XX @@
38
index XXXXXXX..XXXXXXX 100644
58
# @blkreplay: Since 4.2
39
--- a/MAINTAINERS
59
# @compress: Since 5.0
40
+++ b/MAINTAINERS
60
# @copy-before-write: Since 6.2
41
@@ -XXX,XX +XXX,XX @@ F: util/aio-*.c
61
+# @snapshot-access: Since 7.0
42
F: util/aio-*.h
62
#
43
F: util/fdmon-*.c
63
# Since: 2.9
44
F: block/io.c
64
##
45
+F: block/plug.c
65
{ 'enum': 'BlockdevDriver',
46
F: migration/block*
66
'data': [ 'blkdebug', 'blklogwrites', 'blkreplay', 'blkverify', 'bochs',
47
F: include/block/aio.h
67
'cloop', 'compress', 'copy-before-write', 'copy-on-read', 'dmg',
48
F: include/block/aio-wait.h
68
- 'file', 'ftp', 'ftps', 'gluster',
49
diff --git a/include/sysemu/block-backend-io.h b/include/sysemu/block-backend-io.h
69
+ 'file', 'snapshot-access', 'ftp', 'ftps', 'gluster',
50
index XXXXXXX..XXXXXXX 100644
70
{'name': 'host_cdrom', 'if': 'HAVE_HOST_BLOCK_DEVICE' },
51
--- a/include/sysemu/block-backend-io.h
71
{'name': 'host_device', 'if': 'HAVE_HOST_BLOCK_DEVICE' },
52
+++ b/include/sysemu/block-backend-io.h
72
'http', 'https', 'iscsi',
53
@@ -XXX,XX +XXX,XX @@ void blk_iostatus_set_err(BlockBackend *blk, int error);
73
@@ -XXX,XX +XXX,XX @@
54
int blk_get_max_iov(BlockBackend *blk);
74
'rbd': 'BlockdevOptionsRbd',
55
int blk_get_max_hw_iov(BlockBackend *blk);
75
'replication': { 'type': 'BlockdevOptionsReplication',
56
76
'if': 'CONFIG_REPLICATION' },
57
-/*
77
+ 'snapshot-access': 'BlockdevOptionsGenericFormat',
58
- * blk_io_plug/unplug are thread-local operations. This means that multiple
78
'ssh': 'BlockdevOptionsSsh',
59
- * IOThreads can simultaneously call plug/unplug, but the caller must ensure
79
'throttle': 'BlockdevOptionsThrottle',
60
- * that each unplug() is called in the same IOThread of the matching plug().
80
'vdi': 'BlockdevOptionsGenericFormat',
61
- */
81
diff --git a/include/block/block_int-common.h b/include/block/block_int-common.h
62
-void coroutine_fn blk_co_io_plug(BlockBackend *blk);
82
index XXXXXXX..XXXXXXX 100644
63
-void co_wrapper blk_io_plug(BlockBackend *blk);
83
--- a/include/block/block_int-common.h
64
-
84
+++ b/include/block/block_int-common.h
65
-void coroutine_fn blk_co_io_unplug(BlockBackend *blk);
85
@@ -XXX,XX +XXX,XX @@ struct BlockDriver {
66
-void co_wrapper blk_io_unplug(BlockBackend *blk);
86
* in generic block-layer: no serializing, no alignment, no tracked
67
+void blk_io_plug(void);
87
* requests. So, block-driver that realizes these APIs is fully responsible
68
+void blk_io_unplug(void);
88
* for synchronization between snapshot-access API and normal IO requests.
69
+void blk_io_plug_call(void (*fn)(void *), void *opaque);
89
+ *
70
90
+ * TODO: To be able to support qcow2's internal snapshots, this API will
71
AioContext *blk_get_aio_context(BlockBackend *blk);
91
+ * need to be extended to:
72
BlockAcctStats *blk_get_stats(BlockBackend *blk);
92
+ * - be able to select a specific snapshot
73
diff --git a/block/block-backend.c b/block/block-backend.c
93
+ * - receive the snapshot's actual length (which may differ from bs's
74
index XXXXXXX..XXXXXXX 100644
94
+ * length)
75
--- a/block/block-backend.c
95
*/
76
+++ b/block/block-backend.c
96
int coroutine_fn (*bdrv_co_preadv_snapshot)(BlockDriverState *bs,
77
@@ -XXX,XX +XXX,XX @@ void blk_add_insert_bs_notifier(BlockBackend *blk, Notifier *notify)
97
int64_t offset, int64_t bytes, QEMUIOVector *qiov, size_t qiov_offset);
78
notifier_list_add(&blk->insert_bs_notifiers, notify);
98
diff --git a/block/snapshot-access.c b/block/snapshot-access.c
79
}
80
81
-void coroutine_fn blk_co_io_plug(BlockBackend *blk)
82
-{
83
- BlockDriverState *bs = blk_bs(blk);
84
- IO_CODE();
85
- GRAPH_RDLOCK_GUARD();
86
-
87
- if (bs) {
88
- bdrv_co_io_plug(bs);
89
- }
90
-}
91
-
92
-void coroutine_fn blk_co_io_unplug(BlockBackend *blk)
93
-{
94
- BlockDriverState *bs = blk_bs(blk);
95
- IO_CODE();
96
- GRAPH_RDLOCK_GUARD();
97
-
98
- if (bs) {
99
- bdrv_co_io_unplug(bs);
100
- }
101
-}
102
-
103
BlockAcctStats *blk_get_stats(BlockBackend *blk)
104
{
105
IO_CODE();
106
diff --git a/block/plug.c b/block/plug.c
99
new file mode 100644
107
new file mode 100644
100
index XXXXXXX..XXXXXXX
108
index XXXXXXX..XXXXXXX
101
--- /dev/null
109
--- /dev/null
102
+++ b/block/snapshot-access.c
110
+++ b/block/plug.c
103
@@ -XXX,XX +XXX,XX @@
111
@@ -XXX,XX +XXX,XX @@
112
+/* SPDX-License-Identifier: GPL-2.0-or-later */
104
+/*
113
+/*
105
+ * snapshot_access block driver
114
+ * Block I/O plugging
106
+ *
115
+ *
107
+ * Copyright (c) 2022 Virtuozzo International GmbH.
116
+ * Copyright Red Hat.
108
+ *
117
+ *
109
+ * Author:
118
+ * This API defers a function call within a blk_io_plug()/blk_io_unplug()
110
+ * Sementsov-Ogievskiy Vladimir <vsementsov@virtuozzo.com>
119
+ * section, allowing multiple calls to batch up. This is a performance
111
+ *
120
+ * optimization that is used in the block layer to submit several I/O requests
112
+ * This program is free software; you can redistribute it and/or modify
121
+ * at once instead of individually:
113
+ * it under the terms of the GNU General Public License as published by
122
+ *
114
+ * the Free Software Foundation; either version 2 of the License, or
123
+ * blk_io_plug(); <-- start of plugged region
115
+ * (at your option) any later version.
124
+ * ...
116
+ *
125
+ * blk_io_plug_call(my_func, my_obj); <-- deferred my_func(my_obj) call
117
+ * This program is distributed in the hope that it will be useful,
126
+ * blk_io_plug_call(my_func, my_obj); <-- another
118
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
127
+ * blk_io_plug_call(my_func, my_obj); <-- another
119
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
128
+ * ...
120
+ * GNU General Public License for more details.
129
+ * blk_io_unplug(); <-- end of plugged region, my_func(my_obj) is called once
121
+ *
130
+ *
122
+ * You should have received a copy of the GNU General Public License
131
+ * This code is actually generic and not tied to the block layer. If another
123
+ * along with this program. If not, see <http://www.gnu.org/licenses/>.
132
+ * subsystem needs this functionality, it could be renamed.
124
+ */
133
+ */
125
+
134
+
126
+#include "qemu/osdep.h"
135
+#include "qemu/osdep.h"
127
+
136
+#include "qemu/coroutine-tls.h"
137
+#include "qemu/notify.h"
138
+#include "qemu/thread.h"
128
+#include "sysemu/block-backend.h"
139
+#include "sysemu/block-backend.h"
129
+#include "qemu/cutils.h"
140
+
130
+#include "block/block_int.h"
141
+/* A function call that has been deferred until unplug() */
131
+
142
+typedef struct {
132
+static coroutine_fn int
143
+ void (*fn)(void *);
133
+snapshot_access_co_preadv_part(BlockDriverState *bs,
144
+ void *opaque;
134
+ int64_t offset, int64_t bytes,
145
+} UnplugFn;
135
+ QEMUIOVector *qiov, size_t qiov_offset,
146
+
136
+ BdrvRequestFlags flags)
147
+/* Per-thread state */
148
+typedef struct {
149
+ unsigned count; /* how many times has plug() been called? */
150
+ GArray *unplug_fns; /* functions to call at unplug time */
151
+} Plug;
152
+
153
+/* Use get_ptr_plug() to fetch this thread-local value */
154
+QEMU_DEFINE_STATIC_CO_TLS(Plug, plug);
155
+
156
+/* Called at thread cleanup time */
157
+static void blk_io_plug_atexit(Notifier *n, void *value)
137
+{
158
+{
138
+ if (flags) {
159
+ Plug *plug = get_ptr_plug();
139
+ return -ENOTSUP;
160
+ g_array_free(plug->unplug_fns, TRUE);
140
+ }
141
+
142
+ return bdrv_co_preadv_snapshot(bs->file, offset, bytes, qiov, qiov_offset);
143
+}
161
+}
144
+
162
+
145
+static int coroutine_fn
163
+/* This won't involve coroutines, so use __thread */
146
+snapshot_access_co_block_status(BlockDriverState *bs,
164
+static __thread Notifier blk_io_plug_atexit_notifier;
147
+ bool want_zero, int64_t offset,
165
+
148
+ int64_t bytes, int64_t *pnum,
166
+/**
149
+ int64_t *map, BlockDriverState **file)
167
+ * blk_io_plug_call:
168
+ * @fn: a function pointer to be invoked
169
+ * @opaque: a user-defined argument to @fn()
170
+ *
171
+ * Call @fn(@opaque) immediately if not within a blk_io_plug()/blk_io_unplug()
172
+ * section.
173
+ *
174
+ * Otherwise defer the call until the end of the outermost
175
+ * blk_io_plug()/blk_io_unplug() section in this thread. If the same
176
+ * @fn/@opaque pair has already been deferred, it will only be called once upon
177
+ * blk_io_unplug() so that accumulated calls are batched into a single call.
178
+ *
179
+ * The caller must ensure that @opaque is not freed before @fn() is invoked.
180
+ */
181
+void blk_io_plug_call(void (*fn)(void *), void *opaque)
150
+{
182
+{
151
+ return bdrv_co_snapshot_block_status(bs->file->bs, want_zero, offset,
183
+ Plug *plug = get_ptr_plug();
152
+ bytes, pnum, map, file);
184
+
185
+ /* Call immediately if we're not plugged */
186
+ if (plug->count == 0) {
187
+ fn(opaque);
188
+ return;
189
+ }
190
+
191
+ GArray *array = plug->unplug_fns;
192
+ if (!array) {
193
+ array = g_array_new(FALSE, FALSE, sizeof(UnplugFn));
194
+ plug->unplug_fns = array;
195
+ blk_io_plug_atexit_notifier.notify = blk_io_plug_atexit;
196
+ qemu_thread_atexit_add(&blk_io_plug_atexit_notifier);
197
+ }
198
+
199
+ UnplugFn *fns = (UnplugFn *)array->data;
200
+ UnplugFn new_fn = {
201
+ .fn = fn,
202
+ .opaque = opaque,
203
+ };
204
+
205
+ /*
206
+ * There won't be many, so do a linear search. If this becomes a bottleneck
207
+ * then a binary search (glib 2.62+) or different data structure could be
208
+ * used.
209
+ */
210
+ for (guint i = 0; i < array->len; i++) {
211
+ if (memcmp(&fns[i], &new_fn, sizeof(new_fn)) == 0) {
212
+ return; /* already exists */
213
+ }
214
+ }
215
+
216
+ g_array_append_val(array, new_fn);
153
+}
217
+}
154
+
218
+
155
+static int coroutine_fn snapshot_access_co_pdiscard(BlockDriverState *bs,
219
+/**
156
+ int64_t offset, int64_t bytes)
220
+ * blk_io_plug: Defer blk_io_plug_call() functions until blk_io_unplug()
221
+ *
222
+ * blk_io_plug/unplug are thread-local operations. This means that multiple
223
+ * threads can simultaneously call plug/unplug, but the caller must ensure that
224
+ * each unplug() is called in the same thread of the matching plug().
225
+ *
226
+ * Nesting is supported. blk_io_plug_call() functions are only called at the
227
+ * outermost blk_io_unplug().
228
+ */
229
+void blk_io_plug(void)
157
+{
230
+{
158
+ return bdrv_co_pdiscard_snapshot(bs->file->bs, offset, bytes);
231
+ Plug *plug = get_ptr_plug();
232
+
233
+ assert(plug->count < UINT32_MAX);
234
+
235
+ plug->count++;
159
+}
236
+}
160
+
237
+
161
+static int coroutine_fn
238
+/**
162
+snapshot_access_co_pwrite_zeroes(BlockDriverState *bs,
239
+ * blk_io_unplug: Run any pending blk_io_plug_call() functions
163
+ int64_t offset, int64_t bytes,
240
+ *
164
+ BdrvRequestFlags flags)
241
+ * There must have been a matching blk_io_plug() call in the same thread prior
242
+ * to this blk_io_unplug() call.
243
+ */
244
+void blk_io_unplug(void)
165
+{
245
+{
166
+ return -ENOTSUP;
246
+ Plug *plug = get_ptr_plug();
247
+
248
+ assert(plug->count > 0);
249
+
250
+ if (--plug->count > 0) {
251
+ return;
252
+ }
253
+
254
+ GArray *array = plug->unplug_fns;
255
+ if (!array) {
256
+ return;
257
+ }
258
+
259
+ UnplugFn *fns = (UnplugFn *)array->data;
260
+
261
+ for (guint i = 0; i < array->len; i++) {
262
+ fns[i].fn(fns[i].opaque);
263
+ }
264
+
265
+ /*
266
+ * This resets the array without freeing memory so that appending is cheap
267
+ * in the future.
268
+ */
269
+ g_array_set_size(array, 0);
167
+}
270
+}
168
+
271
diff --git a/hw/block/dataplane/xen-block.c b/hw/block/dataplane/xen-block.c
169
+static coroutine_fn int
272
index XXXXXXX..XXXXXXX 100644
170
+snapshot_access_co_pwritev_part(BlockDriverState *bs,
273
--- a/hw/block/dataplane/xen-block.c
171
+ int64_t offset, int64_t bytes,
274
+++ b/hw/block/dataplane/xen-block.c
172
+ QEMUIOVector *qiov, size_t qiov_offset,
275
@@ -XXX,XX +XXX,XX @@ static bool xen_block_handle_requests(XenBlockDataPlane *dataplane)
173
+ BdrvRequestFlags flags)
276
* is below us.
174
+{
277
*/
175
+ return -ENOTSUP;
278
if (inflight_atstart > IO_PLUG_THRESHOLD) {
176
+}
279
- blk_io_plug(dataplane->blk);
177
+
280
+ blk_io_plug();
178
+
281
}
179
+static void snapshot_access_refresh_filename(BlockDriverState *bs)
282
while (rc != rp) {
180
+{
283
/* pull request from ring */
181
+ pstrcpy(bs->exact_filename, sizeof(bs->exact_filename),
284
@@ -XXX,XX +XXX,XX @@ static bool xen_block_handle_requests(XenBlockDataPlane *dataplane)
182
+ bs->file->bs->filename);
285
183
+}
286
if (inflight_atstart > IO_PLUG_THRESHOLD &&
184
+
287
batched >= inflight_atstart) {
185
+static int snapshot_access_open(BlockDriverState *bs, QDict *options, int flags,
288
- blk_io_unplug(dataplane->blk);
186
+ Error **errp)
289
+ blk_io_unplug();
187
+{
290
}
188
+ bs->file = bdrv_open_child(NULL, options, "file", bs, &child_of_bds,
291
xen_block_do_aio(request);
189
+ BDRV_CHILD_DATA | BDRV_CHILD_PRIMARY,
292
if (inflight_atstart > IO_PLUG_THRESHOLD) {
190
+ false, errp);
293
if (batched >= inflight_atstart) {
191
+ if (!bs->file) {
294
- blk_io_plug(dataplane->blk);
192
+ return -EINVAL;
295
+ blk_io_plug();
193
+ }
296
batched = 0;
194
+
297
} else {
195
+ bs->total_sectors = bs->file->bs->total_sectors;
298
batched++;
196
+
299
@@ -XXX,XX +XXX,XX @@ static bool xen_block_handle_requests(XenBlockDataPlane *dataplane)
197
+ return 0;
300
}
198
+}
301
}
199
+
302
if (inflight_atstart > IO_PLUG_THRESHOLD) {
200
+static void snapshot_access_child_perm(BlockDriverState *bs, BdrvChild *c,
303
- blk_io_unplug(dataplane->blk);
201
+ BdrvChildRole role,
304
+ blk_io_unplug();
202
+ BlockReopenQueue *reopen_queue,
305
}
203
+ uint64_t perm, uint64_t shared,
306
204
+ uint64_t *nperm, uint64_t *nshared)
307
return done_something;
205
+{
308
diff --git a/hw/block/virtio-blk.c b/hw/block/virtio-blk.c
206
+ /*
309
index XXXXXXX..XXXXXXX 100644
207
+ * Currently, we don't need any permissions. If bs->file provides
310
--- a/hw/block/virtio-blk.c
208
+ * snapshot-access API, we can use it.
311
+++ b/hw/block/virtio-blk.c
209
+ */
312
@@ -XXX,XX +XXX,XX @@ void virtio_blk_handle_vq(VirtIOBlock *s, VirtQueue *vq)
210
+ *nperm = 0;
313
bool suppress_notifications = virtio_queue_get_notification(vq);
211
+ *nshared = BLK_PERM_ALL;
314
212
+}
315
aio_context_acquire(blk_get_aio_context(s->blk));
213
+
316
- blk_io_plug(s->blk);
214
+BlockDriver bdrv_snapshot_access_drv = {
317
+ blk_io_plug();
215
+ .format_name = "snapshot-access",
318
216
+
319
do {
217
+ .bdrv_open = snapshot_access_open,
320
if (suppress_notifications) {
218
+
321
@@ -XXX,XX +XXX,XX @@ void virtio_blk_handle_vq(VirtIOBlock *s, VirtQueue *vq)
219
+ .bdrv_co_preadv_part = snapshot_access_co_preadv_part,
322
virtio_blk_submit_multireq(s, &mrb);
220
+ .bdrv_co_pwritev_part = snapshot_access_co_pwritev_part,
323
}
221
+ .bdrv_co_pwrite_zeroes = snapshot_access_co_pwrite_zeroes,
324
222
+ .bdrv_co_pdiscard = snapshot_access_co_pdiscard,
325
- blk_io_unplug(s->blk);
223
+ .bdrv_co_block_status = snapshot_access_co_block_status,
326
+ blk_io_unplug();
224
+
327
aio_context_release(blk_get_aio_context(s->blk));
225
+ .bdrv_refresh_filename = snapshot_access_refresh_filename,
328
}
226
+
329
227
+ .bdrv_child_perm = snapshot_access_child_perm,
330
diff --git a/hw/scsi/virtio-scsi.c b/hw/scsi/virtio-scsi.c
228
+};
331
index XXXXXXX..XXXXXXX 100644
229
+
332
--- a/hw/scsi/virtio-scsi.c
230
+static void snapshot_access_init(void)
333
+++ b/hw/scsi/virtio-scsi.c
231
+{
334
@@ -XXX,XX +XXX,XX @@ static int virtio_scsi_handle_cmd_req_prepare(VirtIOSCSI *s, VirtIOSCSIReq *req)
232
+ bdrv_register(&bdrv_snapshot_access_drv);
335
return -ENOBUFS;
233
+}
336
}
234
+
337
scsi_req_ref(req->sreq);
235
+block_init(snapshot_access_init);
338
- blk_io_plug(d->conf.blk);
236
diff --git a/MAINTAINERS b/MAINTAINERS
339
+ blk_io_plug();
237
index XXXXXXX..XXXXXXX 100644
340
object_unref(OBJECT(d));
238
--- a/MAINTAINERS
341
return 0;
239
+++ b/MAINTAINERS
342
}
240
@@ -XXX,XX +XXX,XX @@ F: block/reqlist.c
343
@@ -XXX,XX +XXX,XX @@ static void virtio_scsi_handle_cmd_req_submit(VirtIOSCSI *s, VirtIOSCSIReq *req)
241
F: include/block/reqlist.h
344
if (scsi_req_enqueue(sreq)) {
242
F: block/copy-before-write.h
345
scsi_req_continue(sreq);
243
F: block/copy-before-write.c
346
}
244
+F: block/snapshot-access.c
347
- blk_io_unplug(sreq->dev->conf.blk);
245
F: include/block/aio_task.h
348
+ blk_io_unplug();
246
F: block/aio_task.c
349
scsi_req_unref(sreq);
247
F: util/qemu-co-shared-resource.c
350
}
351
352
@@ -XXX,XX +XXX,XX @@ static void virtio_scsi_handle_cmd_vq(VirtIOSCSI *s, VirtQueue *vq)
353
while (!QTAILQ_EMPTY(&reqs)) {
354
req = QTAILQ_FIRST(&reqs);
355
QTAILQ_REMOVE(&reqs, req, next);
356
- blk_io_unplug(req->sreq->dev->conf.blk);
357
+ blk_io_unplug();
358
scsi_req_unref(req->sreq);
359
virtqueue_detach_element(req->vq, &req->elem, 0);
360
virtio_scsi_free_req(req);
248
diff --git a/block/meson.build b/block/meson.build
361
diff --git a/block/meson.build b/block/meson.build
249
index XXXXXXX..XXXXXXX 100644
362
index XXXXXXX..XXXXXXX 100644
250
--- a/block/meson.build
363
--- a/block/meson.build
251
+++ b/block/meson.build
364
+++ b/block/meson.build
252
@@ -XXX,XX +XXX,XX @@ block_ss.add(files(
365
@@ -XXX,XX +XXX,XX @@ block_ss.add(files(
253
'raw-format.c',
366
'mirror.c',
254
'reqlist.c',
367
'nbd.c',
255
'snapshot.c',
368
'null.c',
256
+ 'snapshot-access.c',
369
+ 'plug.c',
257
'throttle-groups.c',
370
'qapi.c',
258
'throttle.c',
371
'qcow2-bitmap.c',
259
'vhdx-endian.c',
372
'qcow2-cache.c',
260
--
373
--
261
2.34.1
374
2.40.1
diff view generated by jsdifflib
1
From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
1
Stop using the .bdrv_co_io_plug() API because it is not multi-queue
2
block layer friendly. Use the new blk_io_plug_call() API to batch I/O
3
submission instead.
2
4
3
Split intersecting-requests functionality out of block-copy to be
5
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
4
reused in copy-before-write filter.
6
Reviewed-by: Eric Blake <eblake@redhat.com>
7
Reviewed-by: Stefano Garzarella <sgarzare@redhat.com>
8
Acked-by: Kevin Wolf <kwolf@redhat.com>
9
Message-id: 20230530180959.1108766-3-stefanha@redhat.com
10
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
11
---
12
block/nvme.c | 44 ++++++++++++--------------------------------
13
block/trace-events | 1 -
14
2 files changed, 12 insertions(+), 33 deletions(-)
5
15
6
Note: while being here, fix tiny typo in MAINTAINERS.
16
diff --git a/block/nvme.c b/block/nvme.c
7
17
index XXXXXXX..XXXXXXX 100644
8
Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
18
--- a/block/nvme.c
9
Reviewed-by: Hanna Reitz <hreitz@redhat.com>
19
+++ b/block/nvme.c
10
Message-Id: <20220303194349.2304213-7-vsementsov@virtuozzo.com>
11
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
12
---
13
include/block/reqlist.h | 67 +++++++++++++++++++++++
14
block/block-copy.c | 116 +++++++++++++---------------------------
15
block/reqlist.c | 76 ++++++++++++++++++++++++++
16
MAINTAINERS | 4 +-
17
block/meson.build | 1 +
18
5 files changed, 184 insertions(+), 80 deletions(-)
19
create mode 100644 include/block/reqlist.h
20
create mode 100644 block/reqlist.c
21
22
diff --git a/include/block/reqlist.h b/include/block/reqlist.h
23
new file mode 100644
24
index XXXXXXX..XXXXXXX
25
--- /dev/null
26
+++ b/include/block/reqlist.h
27
@@ -XXX,XX +XXX,XX @@
20
@@ -XXX,XX +XXX,XX @@
28
+/*
21
#include "qemu/vfio-helpers.h"
29
+ * reqlist API
22
#include "block/block-io.h"
30
+ *
23
#include "block/block_int.h"
31
+ * Copyright (C) 2013 Proxmox Server Solutions
24
+#include "sysemu/block-backend.h"
32
+ * Copyright (c) 2021 Virtuozzo International GmbH.
25
#include "sysemu/replay.h"
33
+ *
26
#include "trace.h"
34
+ * Authors:
27
35
+ * Dietmar Maurer (dietmar@proxmox.com)
28
@@ -XXX,XX +XXX,XX @@ struct BDRVNVMeState {
36
+ * Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
29
int blkshift;
37
+ *
30
38
+ * This work is licensed under the terms of the GNU GPL, version 2 or later.
31
uint64_t max_transfer;
39
+ * See the COPYING file in the top-level directory.
32
- bool plugged;
40
+ */
33
34
bool supports_write_zeroes;
35
bool supports_discard;
36
@@ -XXX,XX +XXX,XX @@ static void nvme_kick(NVMeQueuePair *q)
37
{
38
BDRVNVMeState *s = q->s;
39
40
- if (s->plugged || !q->need_kick) {
41
+ if (!q->need_kick) {
42
return;
43
}
44
trace_nvme_kick(s, q->index);
45
@@ -XXX,XX +XXX,XX @@ static bool nvme_process_completion(NVMeQueuePair *q)
46
NvmeCqe *c;
47
48
trace_nvme_process_completion(s, q->index, q->inflight);
49
- if (s->plugged) {
50
- trace_nvme_process_completion_queue_plugged(s, q->index);
51
- return false;
52
- }
53
54
/*
55
* Support re-entrancy when a request cb() function invokes aio_poll().
56
@@ -XXX,XX +XXX,XX @@ static void nvme_trace_command(const NvmeCmd *cmd)
57
}
58
}
59
60
+static void nvme_unplug_fn(void *opaque)
61
+{
62
+ NVMeQueuePair *q = opaque;
41
+
63
+
42
+#ifndef REQLIST_H
64
+ QEMU_LOCK_GUARD(&q->lock);
43
+#define REQLIST_H
65
+ nvme_kick(q);
66
+ nvme_process_completion(q);
67
+}
44
+
68
+
45
+#include "qemu/coroutine.h"
69
static void nvme_submit_command(NVMeQueuePair *q, NVMeRequest *req,
46
+
70
NvmeCmd *cmd, BlockCompletionFunc cb,
47
+/*
71
void *opaque)
48
+ * The API is not thread-safe and shouldn't be. The struct is public to be part
72
@@ -XXX,XX +XXX,XX @@ static void nvme_submit_command(NVMeQueuePair *q, NVMeRequest *req,
49
+ * of other structures and protected by third-party locks, see
73
q->sq.tail * NVME_SQ_ENTRY_BYTES, cmd, sizeof(*cmd));
50
+ * block/block-copy.c for example.
74
q->sq.tail = (q->sq.tail + 1) % NVME_QUEUE_SIZE;
51
+ */
75
q->need_kick++;
52
+
76
- nvme_kick(q);
53
+typedef struct BlockReq {
77
- nvme_process_completion(q);
54
+ int64_t offset;
78
+ blk_io_plug_call(nvme_unplug_fn, q);
55
+ int64_t bytes;
79
qemu_mutex_unlock(&q->lock);
56
+
57
+ CoQueue wait_queue; /* coroutines blocked on this req */
58
+ QLIST_ENTRY(BlockReq) list;
59
+} BlockReq;
60
+
61
+typedef QLIST_HEAD(, BlockReq) BlockReqList;
62
+
63
+/*
64
+ * Initialize new request and add it to the list. Caller must be sure that
65
+ * there are no conflicting requests in the list.
66
+ */
67
+void reqlist_init_req(BlockReqList *reqs, BlockReq *req, int64_t offset,
68
+ int64_t bytes);
69
+/* Search for request in the list intersecting with @offset/@bytes area. */
70
+BlockReq *reqlist_find_conflict(BlockReqList *reqs, int64_t offset,
71
+ int64_t bytes);
72
+
73
+/*
74
+ * If there are no intersecting requests return false. Otherwise, wait for the
75
+ * first found intersecting request to finish and return true.
76
+ *
77
+ * @lock is passed to qemu_co_queue_wait()
78
+ * False return value proves that lock was released at no point.
79
+ */
80
+bool coroutine_fn reqlist_wait_one(BlockReqList *reqs, int64_t offset,
81
+ int64_t bytes, CoMutex *lock);
82
+
83
+/*
84
+ * Shrink request and wake all waiting coroutines (maybe some of them are not
85
+ * intersecting with shrunk request).
86
+ */
87
+void coroutine_fn reqlist_shrink_req(BlockReq *req, int64_t new_bytes);
88
+
89
+/*
90
+ * Remove request and wake all waiting coroutines. Do not release any memory.
91
+ */
92
+void coroutine_fn reqlist_remove_req(BlockReq *req);
93
+
94
+#endif /* REQLIST_H */
95
diff --git a/block/block-copy.c b/block/block-copy.c
96
index XXXXXXX..XXXXXXX 100644
97
--- a/block/block-copy.c
98
+++ b/block/block-copy.c
99
@@ -XXX,XX +XXX,XX @@
100
#include "trace.h"
101
#include "qapi/error.h"
102
#include "block/block-copy.h"
103
+#include "block/reqlist.h"
104
#include "sysemu/block-backend.h"
105
#include "qemu/units.h"
106
#include "qemu/coroutine.h"
107
@@ -XXX,XX +XXX,XX @@ typedef struct BlockCopyTask {
108
*/
109
BlockCopyState *s;
110
BlockCopyCallState *call_state;
111
- int64_t offset;
112
/*
113
* @method can also be set again in the while loop of
114
* block_copy_dirty_clusters(), but it is never accessed concurrently
115
@@ -XXX,XX +XXX,XX @@ typedef struct BlockCopyTask {
116
BlockCopyMethod method;
117
118
/*
119
- * Fields whose state changes throughout the execution
120
- * Protected by lock in BlockCopyState.
121
+ * Generally, req is protected by lock in BlockCopyState, Still req.offset
122
+ * is only set on task creation, so may be read concurrently after creation.
123
+ * req.bytes is changed at most once, and need only protecting the case of
124
+ * parallel read while updating @bytes value in block_copy_task_shrink().
125
*/
126
- CoQueue wait_queue; /* coroutines blocked on this task */
127
- /*
128
- * Only protect the case of parallel read while updating @bytes
129
- * value in block_copy_task_shrink().
130
- */
131
- int64_t bytes;
132
- QLIST_ENTRY(BlockCopyTask) list;
133
+ BlockReq req;
134
} BlockCopyTask;
135
136
static int64_t task_end(BlockCopyTask *task)
137
{
138
- return task->offset + task->bytes;
139
+ return task->req.offset + task->req.bytes;
140
}
80
}
141
81
142
typedef struct BlockCopyState {
82
@@ -XXX,XX +XXX,XX @@ static void nvme_attach_aio_context(BlockDriverState *bs,
143
@@ -XXX,XX +XXX,XX @@ typedef struct BlockCopyState {
83
}
144
CoMutex lock;
84
}
145
int64_t in_flight_bytes;
85
146
BlockCopyMethod method;
86
-static void coroutine_fn nvme_co_io_plug(BlockDriverState *bs)
147
- QLIST_HEAD(, BlockCopyTask) tasks; /* All tasks from all block-copy calls */
148
+ BlockReqList reqs;
149
QLIST_HEAD(, BlockCopyCallState) calls;
150
/*
151
* skip_unallocated:
152
@@ -XXX,XX +XXX,XX @@ typedef struct BlockCopyState {
153
RateLimit rate_limit;
154
} BlockCopyState;
155
156
-/* Called with lock held */
157
-static BlockCopyTask *find_conflicting_task(BlockCopyState *s,
158
- int64_t offset, int64_t bytes)
159
-{
87
-{
160
- BlockCopyTask *t;
88
- BDRVNVMeState *s = bs->opaque;
161
-
89
- assert(!s->plugged);
162
- QLIST_FOREACH(t, &s->tasks, list) {
90
- s->plugged = true;
163
- if (offset + bytes > t->offset && offset < t->offset + t->bytes) {
164
- return t;
165
- }
166
- }
167
-
168
- return NULL;
169
-}
91
-}
170
-
92
-
171
-/*
93
-static void coroutine_fn nvme_co_io_unplug(BlockDriverState *bs)
172
- * If there are no intersecting tasks return false. Otherwise, wait for the
173
- * first found intersecting tasks to finish and return true.
174
- *
175
- * Called with lock held. May temporary release the lock.
176
- * Return value of 0 proves that lock was NOT released.
177
- */
178
-static bool coroutine_fn block_copy_wait_one(BlockCopyState *s, int64_t offset,
179
- int64_t bytes)
180
-{
94
-{
181
- BlockCopyTask *task = find_conflicting_task(s, offset, bytes);
95
- BDRVNVMeState *s = bs->opaque;
182
-
96
- assert(s->plugged);
183
- if (!task) {
97
- s->plugged = false;
184
- return false;
98
- for (unsigned i = INDEX_IO(0); i < s->queue_count; i++) {
99
- NVMeQueuePair *q = s->queues[i];
100
- qemu_mutex_lock(&q->lock);
101
- nvme_kick(q);
102
- nvme_process_completion(q);
103
- qemu_mutex_unlock(&q->lock);
185
- }
104
- }
186
-
187
- qemu_co_queue_wait(&task->wait_queue, &s->lock);
188
-
189
- return true;
190
-}
105
-}
191
-
106
-
192
/* Called with lock held */
107
static bool nvme_register_buf(BlockDriverState *bs, void *host, size_t size,
193
static int64_t block_copy_chunk_size(BlockCopyState *s)
108
Error **errp)
194
{
109
{
195
@@ -XXX,XX +XXX,XX @@ block_copy_task_create(BlockCopyState *s, BlockCopyCallState *call_state,
110
@@ -XXX,XX +XXX,XX @@ static BlockDriver bdrv_nvme = {
196
bytes = QEMU_ALIGN_UP(bytes, s->cluster_size);
111
.bdrv_detach_aio_context = nvme_detach_aio_context,
197
112
.bdrv_attach_aio_context = nvme_attach_aio_context,
198
/* region is dirty, so no existent tasks possible in it */
113
199
- assert(!find_conflicting_task(s, offset, bytes));
114
- .bdrv_co_io_plug = nvme_co_io_plug,
200
+ assert(!reqlist_find_conflict(&s->reqs, offset, bytes));
115
- .bdrv_co_io_unplug = nvme_co_io_unplug,
201
116
-
202
bdrv_reset_dirty_bitmap(s->copy_bitmap, offset, bytes);
117
.bdrv_register_buf = nvme_register_buf,
203
s->in_flight_bytes += bytes;
118
.bdrv_unregister_buf = nvme_unregister_buf,
204
@@ -XXX,XX +XXX,XX @@ block_copy_task_create(BlockCopyState *s, BlockCopyCallState *call_state,
119
};
205
.task.func = block_copy_task_entry,
120
diff --git a/block/trace-events b/block/trace-events
206
.s = s,
207
.call_state = call_state,
208
- .offset = offset,
209
- .bytes = bytes,
210
.method = s->method,
211
};
212
- qemu_co_queue_init(&task->wait_queue);
213
- QLIST_INSERT_HEAD(&s->tasks, task, list);
214
+ reqlist_init_req(&s->reqs, &task->req, offset, bytes);
215
216
return task;
217
}
218
@@ -XXX,XX +XXX,XX @@ static void coroutine_fn block_copy_task_shrink(BlockCopyTask *task,
219
int64_t new_bytes)
220
{
221
QEMU_LOCK_GUARD(&task->s->lock);
222
- if (new_bytes == task->bytes) {
223
+ if (new_bytes == task->req.bytes) {
224
return;
225
}
226
227
- assert(new_bytes > 0 && new_bytes < task->bytes);
228
+ assert(new_bytes > 0 && new_bytes < task->req.bytes);
229
230
- task->s->in_flight_bytes -= task->bytes - new_bytes;
231
+ task->s->in_flight_bytes -= task->req.bytes - new_bytes;
232
bdrv_set_dirty_bitmap(task->s->copy_bitmap,
233
- task->offset + new_bytes, task->bytes - new_bytes);
234
+ task->req.offset + new_bytes,
235
+ task->req.bytes - new_bytes);
236
237
- task->bytes = new_bytes;
238
- qemu_co_queue_restart_all(&task->wait_queue);
239
+ reqlist_shrink_req(&task->req, new_bytes);
240
}
241
242
static void coroutine_fn block_copy_task_end(BlockCopyTask *task, int ret)
243
{
244
QEMU_LOCK_GUARD(&task->s->lock);
245
- task->s->in_flight_bytes -= task->bytes;
246
+ task->s->in_flight_bytes -= task->req.bytes;
247
if (ret < 0) {
248
- bdrv_set_dirty_bitmap(task->s->copy_bitmap, task->offset, task->bytes);
249
+ bdrv_set_dirty_bitmap(task->s->copy_bitmap, task->req.offset,
250
+ task->req.bytes);
251
}
252
- QLIST_REMOVE(task, list);
253
if (task->s->progress) {
254
progress_set_remaining(task->s->progress,
255
bdrv_get_dirty_count(task->s->copy_bitmap) +
256
task->s->in_flight_bytes);
257
}
258
- qemu_co_queue_restart_all(&task->wait_queue);
259
+ reqlist_remove_req(&task->req);
260
}
261
262
void block_copy_state_free(BlockCopyState *s)
263
@@ -XXX,XX +XXX,XX @@ BlockCopyState *block_copy_state_new(BdrvChild *source, BdrvChild *target,
264
265
ratelimit_init(&s->rate_limit);
266
qemu_co_mutex_init(&s->lock);
267
- QLIST_INIT(&s->tasks);
268
+ QLIST_INIT(&s->reqs);
269
QLIST_INIT(&s->calls);
270
271
return s;
272
@@ -XXX,XX +XXX,XX @@ static coroutine_fn int block_copy_task_run(AioTaskPool *pool,
273
274
aio_task_pool_wait_slot(pool);
275
if (aio_task_pool_status(pool) < 0) {
276
- co_put_to_shres(task->s->mem, task->bytes);
277
+ co_put_to_shres(task->s->mem, task->req.bytes);
278
block_copy_task_end(task, -ECANCELED);
279
g_free(task);
280
return -ECANCELED;
281
@@ -XXX,XX +XXX,XX @@ static coroutine_fn int block_copy_task_entry(AioTask *task)
282
BlockCopyMethod method = t->method;
283
int ret;
284
285
- ret = block_copy_do_copy(s, t->offset, t->bytes, &method, &error_is_read);
286
+ ret = block_copy_do_copy(s, t->req.offset, t->req.bytes, &method,
287
+ &error_is_read);
288
289
WITH_QEMU_LOCK_GUARD(&s->lock) {
290
if (s->method == t->method) {
291
@@ -XXX,XX +XXX,XX @@ static coroutine_fn int block_copy_task_entry(AioTask *task)
292
t->call_state->error_is_read = error_is_read;
293
}
294
} else if (s->progress) {
295
- progress_work_done(s->progress, t->bytes);
296
+ progress_work_done(s->progress, t->req.bytes);
297
}
298
}
299
- co_put_to_shres(s->mem, t->bytes);
300
+ co_put_to_shres(s->mem, t->req.bytes);
301
block_copy_task_end(t, ret);
302
303
return ret;
304
@@ -XXX,XX +XXX,XX @@ block_copy_dirty_clusters(BlockCopyCallState *call_state)
305
trace_block_copy_skip_range(s, offset, bytes);
306
break;
307
}
308
- if (task->offset > offset) {
309
- trace_block_copy_skip_range(s, offset, task->offset - offset);
310
+ if (task->req.offset > offset) {
311
+ trace_block_copy_skip_range(s, offset, task->req.offset - offset);
312
}
313
314
found_dirty = true;
315
316
- ret = block_copy_block_status(s, task->offset, task->bytes,
317
+ ret = block_copy_block_status(s, task->req.offset, task->req.bytes,
318
&status_bytes);
319
assert(ret >= 0); /* never fail */
320
- if (status_bytes < task->bytes) {
321
+ if (status_bytes < task->req.bytes) {
322
block_copy_task_shrink(task, status_bytes);
323
}
324
if (qatomic_read(&s->skip_unallocated) &&
325
!(ret & BDRV_BLOCK_ALLOCATED)) {
326
block_copy_task_end(task, 0);
327
- trace_block_copy_skip_range(s, task->offset, task->bytes);
328
+ trace_block_copy_skip_range(s, task->req.offset, task->req.bytes);
329
offset = task_end(task);
330
bytes = end - offset;
331
g_free(task);
332
@@ -XXX,XX +XXX,XX @@ block_copy_dirty_clusters(BlockCopyCallState *call_state)
333
}
334
}
335
336
- ratelimit_calculate_delay(&s->rate_limit, task->bytes);
337
+ ratelimit_calculate_delay(&s->rate_limit, task->req.bytes);
338
339
- trace_block_copy_process(s, task->offset);
340
+ trace_block_copy_process(s, task->req.offset);
341
342
- co_get_from_shres(s->mem, task->bytes);
343
+ co_get_from_shres(s->mem, task->req.bytes);
344
345
offset = task_end(task);
346
bytes = end - offset;
347
@@ -XXX,XX +XXX,XX @@ static int coroutine_fn block_copy_common(BlockCopyCallState *call_state)
348
* Check that there is no task we still need to
349
* wait to complete
350
*/
351
- ret = block_copy_wait_one(s, call_state->offset,
352
- call_state->bytes);
353
+ ret = reqlist_wait_one(&s->reqs, call_state->offset,
354
+ call_state->bytes, &s->lock);
355
if (ret == 0) {
356
/*
357
* No pending tasks, but check again the bitmap in this
358
@@ -XXX,XX +XXX,XX @@ static int coroutine_fn block_copy_common(BlockCopyCallState *call_state)
359
* between this and the critical section in
360
* block_copy_dirty_clusters().
361
*
362
- * block_copy_wait_one return value 0 also means that it
363
+ * reqlist_wait_one return value 0 also means that it
364
* didn't release the lock. So, we are still in the same
365
* critical section, not interrupted by any concurrent
366
* access to state.
367
diff --git a/block/reqlist.c b/block/reqlist.c
368
new file mode 100644
369
index XXXXXXX..XXXXXXX
370
--- /dev/null
371
+++ b/block/reqlist.c
372
@@ -XXX,XX +XXX,XX @@
373
+/*
374
+ * reqlist API
375
+ *
376
+ * Copyright (C) 2013 Proxmox Server Solutions
377
+ * Copyright (c) 2021 Virtuozzo International GmbH.
378
+ *
379
+ * Authors:
380
+ * Dietmar Maurer (dietmar@proxmox.com)
381
+ * Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
382
+ *
383
+ * This work is licensed under the terms of the GNU GPL, version 2 or later.
384
+ * See the COPYING file in the top-level directory.
385
+ */
386
+
387
+#include "qemu/osdep.h"
388
+
389
+#include "block/reqlist.h"
390
+
391
+void reqlist_init_req(BlockReqList *reqs, BlockReq *req, int64_t offset,
392
+ int64_t bytes)
393
+{
394
+ assert(!reqlist_find_conflict(reqs, offset, bytes));
395
+
396
+ *req = (BlockReq) {
397
+ .offset = offset,
398
+ .bytes = bytes,
399
+ };
400
+ qemu_co_queue_init(&req->wait_queue);
401
+ QLIST_INSERT_HEAD(reqs, req, list);
402
+}
403
+
404
+BlockReq *reqlist_find_conflict(BlockReqList *reqs, int64_t offset,
405
+ int64_t bytes)
406
+{
407
+ BlockReq *r;
408
+
409
+ QLIST_FOREACH(r, reqs, list) {
410
+ if (offset + bytes > r->offset && offset < r->offset + r->bytes) {
411
+ return r;
412
+ }
413
+ }
414
+
415
+ return NULL;
416
+}
417
+
418
+bool coroutine_fn reqlist_wait_one(BlockReqList *reqs, int64_t offset,
419
+ int64_t bytes, CoMutex *lock)
420
+{
421
+ BlockReq *r = reqlist_find_conflict(reqs, offset, bytes);
422
+
423
+ if (!r) {
424
+ return false;
425
+ }
426
+
427
+ qemu_co_queue_wait(&r->wait_queue, lock);
428
+
429
+ return true;
430
+}
431
+
432
+void coroutine_fn reqlist_shrink_req(BlockReq *req, int64_t new_bytes)
433
+{
434
+ if (new_bytes == req->bytes) {
435
+ return;
436
+ }
437
+
438
+ assert(new_bytes > 0 && new_bytes < req->bytes);
439
+
440
+ req->bytes = new_bytes;
441
+ qemu_co_queue_restart_all(&req->wait_queue);
442
+}
443
+
444
+void coroutine_fn reqlist_remove_req(BlockReq *req)
445
+{
446
+ QLIST_REMOVE(req, list);
447
+ qemu_co_queue_restart_all(&req->wait_queue);
448
+}
449
diff --git a/MAINTAINERS b/MAINTAINERS
450
index XXXXXXX..XXXXXXX 100644
121
index XXXXXXX..XXXXXXX 100644
451
--- a/MAINTAINERS
122
--- a/block/trace-events
452
+++ b/MAINTAINERS
123
+++ b/block/trace-events
453
@@ -XXX,XX +XXX,XX @@ F: block/stream.c
124
@@ -XXX,XX +XXX,XX @@ nvme_kick(void *s, unsigned q_index) "s %p q #%u"
454
F: block/mirror.c
125
nvme_dma_flush_queue_wait(void *s) "s %p"
455
F: qapi/job.json
126
nvme_error(int cmd_specific, int sq_head, int sqid, int cid, int status) "cmd_specific %d sq_head %d sqid %d cid %d status 0x%x"
456
F: block/block-copy.c
127
nvme_process_completion(void *s, unsigned q_index, int inflight) "s %p q #%u inflight %d"
457
-F: include/block/block-copy.c
128
-nvme_process_completion_queue_plugged(void *s, unsigned q_index) "s %p q #%u"
458
+F: include/block/block-copy.h
129
nvme_complete_command(void *s, unsigned q_index, int cid) "s %p q #%u cid %d"
459
+F: block/reqlist.c
130
nvme_submit_command(void *s, unsigned q_index, int cid) "s %p q #%u cid %d"
460
+F: include/block/reqlist.h
131
nvme_submit_command_raw(int c0, int c1, int c2, int c3, int c4, int c5, int c6, int c7) "%02x %02x %02x %02x %02x %02x %02x %02x"
461
F: block/copy-before-write.h
462
F: block/copy-before-write.c
463
F: include/block/aio_task.h
464
diff --git a/block/meson.build b/block/meson.build
465
index XXXXXXX..XXXXXXX 100644
466
--- a/block/meson.build
467
+++ b/block/meson.build
468
@@ -XXX,XX +XXX,XX @@ block_ss.add(files(
469
'qcow2.c',
470
'quorum.c',
471
'raw-format.c',
472
+ 'reqlist.c',
473
'snapshot.c',
474
'throttle-groups.c',
475
'throttle.c',
476
--
132
--
477
2.34.1
133
2.40.1
diff view generated by jsdifflib
1
From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
1
Stop using the .bdrv_co_io_plug() API because it is not multi-queue
2
block layer friendly. Use the new blk_io_plug_call() API to batch I/O
3
submission instead.
2
4
3
Split block_copy_reset() out of block_copy_reset_unallocated() to be
5
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
4
used separately later.
6
Reviewed-by: Eric Blake <eblake@redhat.com>
7
Reviewed-by: Stefano Garzarella <sgarzare@redhat.com>
8
Acked-by: Kevin Wolf <kwolf@redhat.com>
9
Message-id: 20230530180959.1108766-4-stefanha@redhat.com
10
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
11
---
12
block/blkio.c | 43 ++++++++++++++++++++++++-------------------
13
1 file changed, 24 insertions(+), 19 deletions(-)
5
14
6
Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
15
diff --git a/block/blkio.c b/block/blkio.c
7
Reviewed-by: Hanna Reitz <hreitz@redhat.com>
8
Message-Id: <20220303194349.2304213-6-vsementsov@virtuozzo.com>
9
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
10
---
11
include/block/block-copy.h | 1 +
12
block/block-copy.c | 21 +++++++++++++--------
13
2 files changed, 14 insertions(+), 8 deletions(-)
14
15
diff --git a/include/block/block-copy.h b/include/block/block-copy.h
16
index XXXXXXX..XXXXXXX 100644
16
index XXXXXXX..XXXXXXX 100644
17
--- a/include/block/block-copy.h
17
--- a/block/blkio.c
18
+++ b/include/block/block-copy.h
18
+++ b/block/blkio.c
19
@@ -XXX,XX +XXX,XX @@ void block_copy_set_progress_meter(BlockCopyState *s, ProgressMeter *pm);
19
@@ -XXX,XX +XXX,XX @@
20
20
#include "qemu/error-report.h"
21
void block_copy_state_free(BlockCopyState *s);
21
#include "qapi/qmp/qdict.h"
22
22
#include "qemu/module.h"
23
+void block_copy_reset(BlockCopyState *s, int64_t offset, int64_t bytes);
23
+#include "sysemu/block-backend.h"
24
int64_t block_copy_reset_unallocated(BlockCopyState *s,
24
#include "exec/memory.h" /* for ram_block_discard_disable() */
25
int64_t offset, int64_t *count);
25
26
26
#include "block/block-io.h"
27
diff --git a/block/block-copy.c b/block/block-copy.c
27
@@ -XXX,XX +XXX,XX @@ static void blkio_detach_aio_context(BlockDriverState *bs)
28
index XXXXXXX..XXXXXXX 100644
28
NULL, NULL, NULL);
29
--- a/block/block-copy.c
29
}
30
+++ b/block/block-copy.c
30
31
@@ -XXX,XX +XXX,XX @@ static int block_copy_is_cluster_allocated(BlockCopyState *s, int64_t offset,
31
-/* Call with s->blkio_lock held to submit I/O after enqueuing a new request */
32
-static void blkio_submit_io(BlockDriverState *bs)
33
+/*
34
+ * Called by blk_io_unplug() or immediately if not plugged. Called without
35
+ * blkio_lock.
36
+ */
37
+static void blkio_unplug_fn(void *opaque)
38
{
39
- if (qatomic_read(&bs->io_plugged) == 0) {
40
- BDRVBlkioState *s = bs->opaque;
41
+ BDRVBlkioState *s = opaque;
42
43
+ WITH_QEMU_LOCK_GUARD(&s->blkio_lock) {
44
blkioq_do_io(s->blkioq, NULL, 0, 0, NULL);
32
}
45
}
33
}
46
}
34
47
35
+void block_copy_reset(BlockCopyState *s, int64_t offset, int64_t bytes)
48
+/*
49
+ * Schedule I/O submission after enqueuing a new request. Called without
50
+ * blkio_lock.
51
+ */
52
+static void blkio_submit_io(BlockDriverState *bs)
36
+{
53
+{
37
+ QEMU_LOCK_GUARD(&s->lock);
54
+ BDRVBlkioState *s = bs->opaque;
38
+
55
+
39
+ bdrv_reset_dirty_bitmap(s->copy_bitmap, offset, bytes);
56
+ blk_io_plug_call(blkio_unplug_fn, s);
40
+ if (s->progress) {
41
+ progress_set_remaining(s->progress,
42
+ bdrv_get_dirty_count(s->copy_bitmap) +
43
+ s->in_flight_bytes);
44
+ }
45
+}
57
+}
46
+
58
+
47
/*
59
static int coroutine_fn
48
* Reset bits in copy_bitmap starting at offset if they represent unallocated
60
blkio_co_pdiscard(BlockDriverState *bs, int64_t offset, int64_t bytes)
49
* data in the image. May reset subsequent contiguous bits.
61
{
50
@@ -XXX,XX +XXX,XX @@ int64_t block_copy_reset_unallocated(BlockCopyState *s,
62
@@ -XXX,XX +XXX,XX @@ blkio_co_pdiscard(BlockDriverState *bs, int64_t offset, int64_t bytes)
51
bytes = clusters * s->cluster_size;
63
52
64
WITH_QEMU_LOCK_GUARD(&s->blkio_lock) {
53
if (!ret) {
65
blkioq_discard(s->blkioq, offset, bytes, &cod, 0);
54
- qemu_co_mutex_lock(&s->lock);
66
- blkio_submit_io(bs);
55
- bdrv_reset_dirty_bitmap(s->copy_bitmap, offset, bytes);
56
- if (s->progress) {
57
- progress_set_remaining(s->progress,
58
- bdrv_get_dirty_count(s->copy_bitmap) +
59
- s->in_flight_bytes);
60
- }
61
- qemu_co_mutex_unlock(&s->lock);
62
+ block_copy_reset(s, offset, bytes);
63
}
67
}
64
68
65
*count = bytes;
69
+ blkio_submit_io(bs);
70
qemu_coroutine_yield();
71
return cod.ret;
72
}
73
@@ -XXX,XX +XXX,XX @@ blkio_co_preadv(BlockDriverState *bs, int64_t offset, int64_t bytes,
74
75
WITH_QEMU_LOCK_GUARD(&s->blkio_lock) {
76
blkioq_readv(s->blkioq, offset, iov, iovcnt, &cod, 0);
77
- blkio_submit_io(bs);
78
}
79
80
+ blkio_submit_io(bs);
81
qemu_coroutine_yield();
82
83
if (use_bounce_buffer) {
84
@@ -XXX,XX +XXX,XX @@ static int coroutine_fn blkio_co_pwritev(BlockDriverState *bs, int64_t offset,
85
86
WITH_QEMU_LOCK_GUARD(&s->blkio_lock) {
87
blkioq_writev(s->blkioq, offset, iov, iovcnt, &cod, blkio_flags);
88
- blkio_submit_io(bs);
89
}
90
91
+ blkio_submit_io(bs);
92
qemu_coroutine_yield();
93
94
if (use_bounce_buffer) {
95
@@ -XXX,XX +XXX,XX @@ static int coroutine_fn blkio_co_flush(BlockDriverState *bs)
96
97
WITH_QEMU_LOCK_GUARD(&s->blkio_lock) {
98
blkioq_flush(s->blkioq, &cod, 0);
99
- blkio_submit_io(bs);
100
}
101
102
+ blkio_submit_io(bs);
103
qemu_coroutine_yield();
104
return cod.ret;
105
}
106
@@ -XXX,XX +XXX,XX @@ static int coroutine_fn blkio_co_pwrite_zeroes(BlockDriverState *bs,
107
108
WITH_QEMU_LOCK_GUARD(&s->blkio_lock) {
109
blkioq_write_zeroes(s->blkioq, offset, bytes, &cod, blkio_flags);
110
- blkio_submit_io(bs);
111
}
112
113
+ blkio_submit_io(bs);
114
qemu_coroutine_yield();
115
return cod.ret;
116
}
117
118
-static void coroutine_fn blkio_co_io_unplug(BlockDriverState *bs)
119
-{
120
- BDRVBlkioState *s = bs->opaque;
121
-
122
- WITH_QEMU_LOCK_GUARD(&s->blkio_lock) {
123
- blkio_submit_io(bs);
124
- }
125
-}
126
-
127
typedef enum {
128
BMRR_OK,
129
BMRR_SKIP,
130
@@ -XXX,XX +XXX,XX @@ static void blkio_refresh_limits(BlockDriverState *bs, Error **errp)
131
.bdrv_co_pwritev = blkio_co_pwritev, \
132
.bdrv_co_flush_to_disk = blkio_co_flush, \
133
.bdrv_co_pwrite_zeroes = blkio_co_pwrite_zeroes, \
134
- .bdrv_co_io_unplug = blkio_co_io_unplug, \
135
.bdrv_refresh_limits = blkio_refresh_limits, \
136
.bdrv_register_buf = blkio_register_buf, \
137
.bdrv_unregister_buf = blkio_unregister_buf, \
66
--
138
--
67
2.34.1
139
2.40.1
diff view generated by jsdifflib
1
From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
1
Stop using the .bdrv_co_io_plug() API because it is not multi-queue
2
block layer friendly. Use the new blk_io_plug_call() API to batch I/O
3
submission instead.
2
4
3
Current scheme of image fleecing looks like this:
5
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
6
Reviewed-by: Eric Blake <eblake@redhat.com>
7
Reviewed-by: Stefano Garzarella <sgarzare@redhat.com>
8
Acked-by: Kevin Wolf <kwolf@redhat.com>
9
Message-id: 20230530180959.1108766-5-stefanha@redhat.com
10
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
11
---
12
include/block/raw-aio.h | 7 -------
13
block/file-posix.c | 10 ----------
14
block/io_uring.c | 44 ++++++++++++++++-------------------------
15
block/trace-events | 5 ++---
16
4 files changed, 19 insertions(+), 47 deletions(-)
4
17
5
[guest] [NBD export]
18
diff --git a/include/block/raw-aio.h b/include/block/raw-aio.h
6
| |
7
|root | root
8
v v
9
[copy-before-write] -----> [temp.qcow2]
10
| target |
11
|file |backing
12
v |
13
[active disk] <-------------+
14
15
- On guest writes copy-before-write filter copies old data from active
16
disk to temp.qcow2. So fleecing client (NBD export) when reads
17
changed regions from temp.qcow2 image and unchanged from active disk
18
through backing link.
19
20
This patch makes possible new image fleecing scheme:
21
22
[guest] [NBD export]
23
| |
24
| root | root
25
v file v
26
[copy-before-write]<------[snapshot-access]
27
| |
28
| file | target
29
v v
30
[active-disk] [temp.img]
31
32
- copy-before-write does CBW operations and also provides
33
snapshot-access API. The API may be accessed through
34
snapshot-access driver.
35
36
Benefits of new scheme:
37
38
1. Access control: if remote client try to read data that not covered
39
by original dirty bitmap used on copy-before-write open, client gets
40
-EACCES.
41
42
2. Discard support: if remote client do DISCARD, this additionally to
43
discarding data in temp.img informs block-copy process to not copy
44
these clusters. Next read from discarded area will return -EACCES.
45
This is significant thing: when fleecing user reads data that was
46
not yet copied to temp.img, we can avoid copying it on further guest
47
write.
48
49
3. Synchronisation between client reads and block-copy write is more
50
efficient. In old scheme we just rely on BDRV_REQ_SERIALISING flag
51
used for writes to temp.qcow2. New scheme is less blocking:
52
- fleecing reads are never blocked: if data region is untouched or
53
in-flight, we just read from active-disk, otherwise we read from
54
temp.img
55
- writes to temp.img are not blocked by fleecing reads
56
- still, guest writes of-course are blocked by in-flight fleecing
57
reads, that currently read from active-disk - it's the minimum
58
necessary blocking
59
60
4. Temporary image may be of any format, as we don't rely on backing
61
feature.
62
63
5. Permission relation are simplified. With old scheme we have to share
64
write permission on target child of copy-before-write, otherwise
65
backing link conflicts with copy-before-write file child write
66
permissions. With new scheme we don't have backing link, and
67
copy-before-write node may have unshared access to temporary node.
68
(Not realized in this commit, will be in future).
69
70
6. Having control on fleecing reads we'll be able to implement
71
alternative behavior on failed copy-before-write operations.
72
Currently we just break guest request (that's a historical behavior
73
of backup). But in some scenarios it's a bad behavior: better
74
is to drop the backup as failed but don't break guest request.
75
With new scheme we can simply unset some bits in a bitmap on CBW
76
failure and further fleecing reads will -EACCES, or something like
77
this. (Not implemented in this commit, will be in future)
78
Additional application for this is implementing timeout for CBW
79
operations.
80
81
Iotest 257 output is updated, as two more bitmaps now live in
82
copy-before-write filter.
83
84
Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
85
Message-Id: <20220303194349.2304213-13-vsementsov@virtuozzo.com>
86
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
87
---
88
block/copy-before-write.c | 212 ++++++++++++++++++++++++++++++++++-
89
tests/qemu-iotests/257.out | 224 +++++++++++++++++++++++++++++++++++++
90
2 files changed, 435 insertions(+), 1 deletion(-)
91
92
diff --git a/block/copy-before-write.c b/block/copy-before-write.c
93
index XXXXXXX..XXXXXXX 100644
19
index XXXXXXX..XXXXXXX 100644
94
--- a/block/copy-before-write.c
20
--- a/include/block/raw-aio.h
95
+++ b/block/copy-before-write.c
21
+++ b/include/block/raw-aio.h
22
@@ -XXX,XX +XXX,XX @@ int coroutine_fn luring_co_submit(BlockDriverState *bs, int fd, uint64_t offset,
23
QEMUIOVector *qiov, int type);
24
void luring_detach_aio_context(LuringState *s, AioContext *old_context);
25
void luring_attach_aio_context(LuringState *s, AioContext *new_context);
26
-
27
-/*
28
- * luring_io_plug/unplug work in the thread's current AioContext, therefore the
29
- * caller must ensure that they are paired in the same IOThread.
30
- */
31
-void luring_io_plug(void);
32
-void luring_io_unplug(void);
33
#endif
34
35
#ifdef _WIN32
36
diff --git a/block/file-posix.c b/block/file-posix.c
37
index XXXXXXX..XXXXXXX 100644
38
--- a/block/file-posix.c
39
+++ b/block/file-posix.c
40
@@ -XXX,XX +XXX,XX @@ static void coroutine_fn raw_co_io_plug(BlockDriverState *bs)
41
laio_io_plug();
42
}
43
#endif
44
-#ifdef CONFIG_LINUX_IO_URING
45
- if (s->use_linux_io_uring) {
46
- luring_io_plug();
47
- }
48
-#endif
49
}
50
51
static void coroutine_fn raw_co_io_unplug(BlockDriverState *bs)
52
@@ -XXX,XX +XXX,XX @@ static void coroutine_fn raw_co_io_unplug(BlockDriverState *bs)
53
laio_io_unplug(s->aio_max_batch);
54
}
55
#endif
56
-#ifdef CONFIG_LINUX_IO_URING
57
- if (s->use_linux_io_uring) {
58
- luring_io_unplug();
59
- }
60
-#endif
61
}
62
63
static int coroutine_fn raw_co_flush_to_disk(BlockDriverState *bs)
64
diff --git a/block/io_uring.c b/block/io_uring.c
65
index XXXXXXX..XXXXXXX 100644
66
--- a/block/io_uring.c
67
+++ b/block/io_uring.c
96
@@ -XXX,XX +XXX,XX @@
68
@@ -XXX,XX +XXX,XX @@
97
#include "block/block-copy.h"
69
#include "block/raw-aio.h"
98
70
#include "qemu/coroutine.h"
99
#include "block/copy-before-write.h"
71
#include "qapi/error.h"
100
+#include "block/reqlist.h"
72
+#include "sysemu/block-backend.h"
101
73
#include "trace.h"
102
#include "qapi/qapi-visit-block-core.h"
74
103
75
/* Only used for assertions. */
104
typedef struct BDRVCopyBeforeWriteState {
76
@@ -XXX,XX +XXX,XX @@ typedef struct LuringAIOCB {
105
BlockCopyState *bcs;
77
} LuringAIOCB;
106
BdrvChild *target;
78
107
+
79
typedef struct LuringQueue {
108
+ /*
80
- int plugged;
109
+ * @lock: protects access to @access_bitmap, @done_bitmap and
81
unsigned int in_queue;
110
+ * @frozen_read_reqs
82
unsigned int in_flight;
111
+ */
83
bool blocked;
112
+ CoMutex lock;
84
@@ -XXX,XX +XXX,XX @@ static void luring_process_completions_and_submit(LuringState *s)
113
+
85
{
114
+ /*
86
luring_process_completions(s);
115
+ * @access_bitmap: represents areas allowed for reading by fleecing user.
87
116
+ * Reading from non-dirty areas leads to -EACCES.
88
- if (!s->io_q.plugged && s->io_q.in_queue > 0) {
117
+ */
89
+ if (s->io_q.in_queue > 0) {
118
+ BdrvDirtyBitmap *access_bitmap;
90
ioq_submit(s);
119
+
91
}
120
+ /*
121
+ * @done_bitmap: represents areas that was successfully copied to @target by
122
+ * copy-before-write operations.
123
+ */
124
+ BdrvDirtyBitmap *done_bitmap;
125
+
126
+ /*
127
+ * @frozen_read_reqs: current read requests for fleecing user in bs->file
128
+ * node. These areas must not be rewritten by guest.
129
+ */
130
+ BlockReqList frozen_read_reqs;
131
} BDRVCopyBeforeWriteState;
132
133
static coroutine_fn int cbw_co_preadv(
134
@@ -XXX,XX +XXX,XX @@ static coroutine_fn int cbw_co_preadv(
135
return bdrv_co_preadv(bs->file, offset, bytes, qiov, flags);
136
}
92
}
137
93
@@ -XXX,XX +XXX,XX @@ static void qemu_luring_poll_ready(void *opaque)
138
+/*
94
static void ioq_init(LuringQueue *io_q)
139
+ * Do copy-before-write operation.
140
+ *
141
+ * On failure guest request must be failed too.
142
+ *
143
+ * On success, we also wait for all in-flight fleecing read requests in source
144
+ * node, and it's guaranteed that after cbw_do_copy_before_write() successful
145
+ * return there are no such requests and they will never appear.
146
+ */
147
static coroutine_fn int cbw_do_copy_before_write(BlockDriverState *bs,
148
uint64_t offset, uint64_t bytes, BdrvRequestFlags flags)
149
{
95
{
150
BDRVCopyBeforeWriteState *s = bs->opaque;
96
QSIMPLEQ_INIT(&io_q->submit_queue);
151
+ int ret;
97
- io_q->plugged = 0;
152
uint64_t off, end;
98
io_q->in_queue = 0;
153
int64_t cluster_size = block_copy_cluster_size(s->bcs);
99
io_q->in_flight = 0;
154
100
io_q->blocked = false;
155
@@ -XXX,XX +XXX,XX @@ static coroutine_fn int cbw_do_copy_before_write(BlockDriverState *bs,
156
off = QEMU_ALIGN_DOWN(offset, cluster_size);
157
end = QEMU_ALIGN_UP(offset + bytes, cluster_size);
158
159
- return block_copy(s->bcs, off, end - off, true);
160
+ ret = block_copy(s->bcs, off, end - off, true);
161
+ if (ret < 0) {
162
+ return ret;
163
+ }
164
+
165
+ WITH_QEMU_LOCK_GUARD(&s->lock) {
166
+ bdrv_set_dirty_bitmap(s->done_bitmap, off, end - off);
167
+ reqlist_wait_all(&s->frozen_read_reqs, off, end - off, &s->lock);
168
+ }
169
+
170
+ return 0;
171
}
101
}
172
102
173
static int coroutine_fn cbw_co_pdiscard(BlockDriverState *bs,
103
-void luring_io_plug(void)
174
@@ -XXX,XX +XXX,XX @@ static int coroutine_fn cbw_co_flush(BlockDriverState *bs)
104
+static void luring_unplug_fn(void *opaque)
175
return bdrv_co_flush(bs->file->bs);
105
{
106
- AioContext *ctx = qemu_get_current_aio_context();
107
- LuringState *s = aio_get_linux_io_uring(ctx);
108
- trace_luring_io_plug(s);
109
- s->io_q.plugged++;
110
-}
111
-
112
-void luring_io_unplug(void)
113
-{
114
- AioContext *ctx = qemu_get_current_aio_context();
115
- LuringState *s = aio_get_linux_io_uring(ctx);
116
- assert(s->io_q.plugged);
117
- trace_luring_io_unplug(s, s->io_q.blocked, s->io_q.plugged,
118
- s->io_q.in_queue, s->io_q.in_flight);
119
- if (--s->io_q.plugged == 0 &&
120
- !s->io_q.blocked && s->io_q.in_queue > 0) {
121
+ LuringState *s = opaque;
122
+ trace_luring_unplug_fn(s, s->io_q.blocked, s->io_q.in_queue,
123
+ s->io_q.in_flight);
124
+ if (!s->io_q.blocked && s->io_q.in_queue > 0) {
125
ioq_submit(s);
126
}
176
}
127
}
177
128
@@ -XXX,XX +XXX,XX @@ static int luring_do_submit(int fd, LuringAIOCB *luringcb, LuringState *s,
178
+/*
129
179
+ * If @offset not accessible - return NULL.
130
QSIMPLEQ_INSERT_TAIL(&s->io_q.submit_queue, luringcb, next);
180
+ *
131
s->io_q.in_queue++;
181
+ * Otherwise, set @pnum to some bytes that accessible from @file (@file is set
132
- trace_luring_do_submit(s, s->io_q.blocked, s->io_q.plugged,
182
+ * to bs->file or to s->target). Return newly allocated BlockReq object that
133
- s->io_q.in_queue, s->io_q.in_flight);
183
+ * should be than passed to cbw_snapshot_read_unlock().
134
- if (!s->io_q.blocked &&
184
+ *
135
- (!s->io_q.plugged ||
185
+ * It's guaranteed that guest writes will not interact in the region until
136
- s->io_q.in_flight + s->io_q.in_queue >= MAX_ENTRIES)) {
186
+ * cbw_snapshot_read_unlock() called.
137
- ret = ioq_submit(s);
187
+ */
138
- trace_luring_do_submit_done(s, ret);
188
+static BlockReq *cbw_snapshot_read_lock(BlockDriverState *bs,
139
- return ret;
189
+ int64_t offset, int64_t bytes,
140
+ trace_luring_do_submit(s, s->io_q.blocked, s->io_q.in_queue,
190
+ int64_t *pnum, BdrvChild **file)
141
+ s->io_q.in_flight);
191
+{
142
+ if (!s->io_q.blocked) {
192
+ BDRVCopyBeforeWriteState *s = bs->opaque;
143
+ if (s->io_q.in_flight + s->io_q.in_queue >= MAX_ENTRIES) {
193
+ BlockReq *req = g_new(BlockReq, 1);
144
+ ret = ioq_submit(s);
194
+ bool done;
145
+ trace_luring_do_submit_done(s, ret);
195
+
196
+ QEMU_LOCK_GUARD(&s->lock);
197
+
198
+ if (bdrv_dirty_bitmap_next_zero(s->access_bitmap, offset, bytes) != -1) {
199
+ g_free(req);
200
+ return NULL;
201
+ }
202
+
203
+ done = bdrv_dirty_bitmap_status(s->done_bitmap, offset, bytes, pnum);
204
+ if (done) {
205
+ /*
206
+ * Special invalid BlockReq, that is handled in
207
+ * cbw_snapshot_read_unlock(). We don't need to lock something to read
208
+ * from s->target.
209
+ */
210
+ *req = (BlockReq) {.offset = -1, .bytes = -1};
211
+ *file = s->target;
212
+ } else {
213
+ reqlist_init_req(&s->frozen_read_reqs, req, offset, bytes);
214
+ *file = bs->file;
215
+ }
216
+
217
+ return req;
218
+}
219
+
220
+static void cbw_snapshot_read_unlock(BlockDriverState *bs, BlockReq *req)
221
+{
222
+ BDRVCopyBeforeWriteState *s = bs->opaque;
223
+
224
+ if (req->offset == -1 && req->bytes == -1) {
225
+ g_free(req);
226
+ return;
227
+ }
228
+
229
+ QEMU_LOCK_GUARD(&s->lock);
230
+
231
+ reqlist_remove_req(req);
232
+ g_free(req);
233
+}
234
+
235
+static coroutine_fn int
236
+cbw_co_preadv_snapshot(BlockDriverState *bs, int64_t offset, int64_t bytes,
237
+ QEMUIOVector *qiov, size_t qiov_offset)
238
+{
239
+ BlockReq *req;
240
+ BdrvChild *file;
241
+ int ret;
242
+
243
+ /* TODO: upgrade to async loop using AioTask */
244
+ while (bytes) {
245
+ int64_t cur_bytes;
246
+
247
+ req = cbw_snapshot_read_lock(bs, offset, bytes, &cur_bytes, &file);
248
+ if (!req) {
249
+ return -EACCES;
250
+ }
251
+
252
+ ret = bdrv_co_preadv_part(file, offset, cur_bytes,
253
+ qiov, qiov_offset, 0);
254
+ cbw_snapshot_read_unlock(bs, req);
255
+ if (ret < 0) {
256
+ return ret;
146
+ return ret;
257
+ }
147
+ }
258
+
148
+
259
+ bytes -= cur_bytes;
149
+ blk_io_plug_call(luring_unplug_fn, s);
260
+ offset += cur_bytes;
261
+ qiov_offset += cur_bytes;
262
+ }
263
+
264
+ return 0;
265
+}
266
+
267
+static int coroutine_fn
268
+cbw_co_snapshot_block_status(BlockDriverState *bs,
269
+ bool want_zero, int64_t offset, int64_t bytes,
270
+ int64_t *pnum, int64_t *map,
271
+ BlockDriverState **file)
272
+{
273
+ BDRVCopyBeforeWriteState *s = bs->opaque;
274
+ BlockReq *req;
275
+ int ret;
276
+ int64_t cur_bytes;
277
+ BdrvChild *child;
278
+
279
+ req = cbw_snapshot_read_lock(bs, offset, bytes, &cur_bytes, &child);
280
+ if (!req) {
281
+ return -EACCES;
282
+ }
283
+
284
+ ret = bdrv_block_status(child->bs, offset, cur_bytes, pnum, map, file);
285
+ if (child == s->target) {
286
+ /*
287
+ * We refer to s->target only for areas that we've written to it.
288
+ * And we can not report unallocated blocks in s->target: this will
289
+ * break generic block-status-above logic, that will go to
290
+ * copy-before-write filtered child in this case.
291
+ */
292
+ assert(ret & BDRV_BLOCK_ALLOCATED);
293
+ }
294
+
295
+ cbw_snapshot_read_unlock(bs, req);
296
+
297
+ return ret;
298
+}
299
+
300
+static int coroutine_fn cbw_co_pdiscard_snapshot(BlockDriverState *bs,
301
+ int64_t offset, int64_t bytes)
302
+{
303
+ BDRVCopyBeforeWriteState *s = bs->opaque;
304
+
305
+ WITH_QEMU_LOCK_GUARD(&s->lock) {
306
+ bdrv_reset_dirty_bitmap(s->access_bitmap, offset, bytes);
307
+ }
308
+
309
+ block_copy_reset(s->bcs, offset, bytes);
310
+
311
+ return bdrv_co_pdiscard(s->target, offset, bytes);
312
+}
313
+
314
static void cbw_refresh_filename(BlockDriverState *bs)
315
{
316
pstrcpy(bs->exact_filename, sizeof(bs->exact_filename),
317
@@ -XXX,XX +XXX,XX @@ static int cbw_open(BlockDriverState *bs, QDict *options, int flags,
318
{
319
BDRVCopyBeforeWriteState *s = bs->opaque;
320
BdrvDirtyBitmap *bitmap = NULL;
321
+ int64_t cluster_size;
322
323
bs->file = bdrv_open_child(NULL, options, "file", bs, &child_of_bds,
324
BDRV_CHILD_FILTERED | BDRV_CHILD_PRIMARY,
325
@@ -XXX,XX +XXX,XX @@ static int cbw_open(BlockDriverState *bs, QDict *options, int flags,
326
return -EINVAL;
327
}
150
}
328
329
+ cluster_size = block_copy_cluster_size(s->bcs);
330
+
331
+ s->done_bitmap = bdrv_create_dirty_bitmap(bs, cluster_size, NULL, errp);
332
+ if (!s->done_bitmap) {
333
+ return -EINVAL;
334
+ }
335
+ bdrv_disable_dirty_bitmap(s->done_bitmap);
336
+
337
+ /* s->access_bitmap starts equal to bcs bitmap */
338
+ s->access_bitmap = bdrv_create_dirty_bitmap(bs, cluster_size, NULL, errp);
339
+ if (!s->access_bitmap) {
340
+ return -EINVAL;
341
+ }
342
+ bdrv_disable_dirty_bitmap(s->access_bitmap);
343
+ bdrv_dirty_bitmap_merge_internal(s->access_bitmap,
344
+ block_copy_dirty_bitmap(s->bcs), NULL,
345
+ true);
346
+
347
+ qemu_co_mutex_init(&s->lock);
348
+ QLIST_INIT(&s->frozen_read_reqs);
349
+
350
return 0;
151
return 0;
351
}
152
}
352
153
diff --git a/block/trace-events b/block/trace-events
353
@@ -XXX,XX +XXX,XX @@ static void cbw_close(BlockDriverState *bs)
354
{
355
BDRVCopyBeforeWriteState *s = bs->opaque;
356
357
+ bdrv_release_dirty_bitmap(s->access_bitmap);
358
+ bdrv_release_dirty_bitmap(s->done_bitmap);
359
+
360
block_copy_state_free(s->bcs);
361
s->bcs = NULL;
362
}
363
@@ -XXX,XX +XXX,XX @@ BlockDriver bdrv_cbw_filter = {
364
.bdrv_co_pdiscard = cbw_co_pdiscard,
365
.bdrv_co_flush = cbw_co_flush,
366
367
+ .bdrv_co_preadv_snapshot = cbw_co_preadv_snapshot,
368
+ .bdrv_co_pdiscard_snapshot = cbw_co_pdiscard_snapshot,
369
+ .bdrv_co_snapshot_block_status = cbw_co_snapshot_block_status,
370
+
371
.bdrv_refresh_filename = cbw_refresh_filename,
372
373
.bdrv_child_perm = cbw_child_perm,
374
diff --git a/tests/qemu-iotests/257.out b/tests/qemu-iotests/257.out
375
index XXXXXXX..XXXXXXX 100644
154
index XXXXXXX..XXXXXXX 100644
376
--- a/tests/qemu-iotests/257.out
155
--- a/block/trace-events
377
+++ b/tests/qemu-iotests/257.out
156
+++ b/block/trace-events
378
@@ -XXX,XX +XXX,XX @@ write -P0x67 0x3fe0000 0x20000
157
@@ -XXX,XX +XXX,XX @@ file_paio_submit(void *acb, void *opaque, int64_t offset, int count, int type) "
379
{"return": ""}
158
# io_uring.c
380
{
159
luring_init_state(void *s, size_t size) "s %p size %zu"
381
"bitmaps": {
160
luring_cleanup_state(void *s) "%p freed"
382
+ "backup-top": [
161
-luring_io_plug(void *s) "LuringState %p plug"
383
+ {
162
-luring_io_unplug(void *s, int blocked, int plugged, int queued, int inflight) "LuringState %p blocked %d plugged %d queued %d inflight %d"
384
+ "busy": false,
163
-luring_do_submit(void *s, int blocked, int plugged, int queued, int inflight) "LuringState %p blocked %d plugged %d queued %d inflight %d"
385
+ "count": 67108864,
164
+luring_unplug_fn(void *s, int blocked, int queued, int inflight) "LuringState %p blocked %d queued %d inflight %d"
386
+ "granularity": 65536,
165
+luring_do_submit(void *s, int blocked, int queued, int inflight) "LuringState %p blocked %d queued %d inflight %d"
387
+ "persistent": false,
166
luring_do_submit_done(void *s, int ret) "LuringState %p submitted to kernel %d"
388
+ "recording": false
167
luring_co_submit(void *bs, void *s, void *luringcb, int fd, uint64_t offset, size_t nbytes, int type) "bs %p s %p luringcb %p fd %d offset %" PRId64 " nbytes %zd type %d"
389
+ },
168
luring_process_completion(void *s, void *aiocb, int ret) "LuringState %p luringcb %p ret %d"
390
+ {
391
+ "busy": false,
392
+ "count": 458752,
393
+ "granularity": 65536,
394
+ "persistent": false,
395
+ "recording": false
396
+ }
397
+ ],
398
"drive0": [
399
{
400
"busy": false,
401
@@ -XXX,XX +XXX,XX @@ write -P0x67 0x3fe0000 0x20000
402
{"return": ""}
403
{
404
"bitmaps": {
405
+ "backup-top": [
406
+ {
407
+ "busy": false,
408
+ "count": 67108864,
409
+ "granularity": 65536,
410
+ "persistent": false,
411
+ "recording": false
412
+ },
413
+ {
414
+ "busy": false,
415
+ "count": 458752,
416
+ "granularity": 65536,
417
+ "persistent": false,
418
+ "recording": false
419
+ }
420
+ ],
421
"drive0": [
422
{
423
"busy": false,
424
@@ -XXX,XX +XXX,XX @@ write -P0x67 0x3fe0000 0x20000
425
{"return": ""}
426
{
427
"bitmaps": {
428
+ "backup-top": [
429
+ {
430
+ "busy": false,
431
+ "count": 67108864,
432
+ "granularity": 65536,
433
+ "persistent": false,
434
+ "recording": false
435
+ },
436
+ {
437
+ "busy": false,
438
+ "count": 458752,
439
+ "granularity": 65536,
440
+ "persistent": false,
441
+ "recording": false
442
+ }
443
+ ],
444
"drive0": [
445
{
446
"busy": false,
447
@@ -XXX,XX +XXX,XX @@ write -P0x67 0x3fe0000 0x20000
448
{"return": ""}
449
{
450
"bitmaps": {
451
+ "backup-top": [
452
+ {
453
+ "busy": false,
454
+ "count": 67108864,
455
+ "granularity": 65536,
456
+ "persistent": false,
457
+ "recording": false
458
+ },
459
+ {
460
+ "busy": false,
461
+ "count": 458752,
462
+ "granularity": 65536,
463
+ "persistent": false,
464
+ "recording": false
465
+ }
466
+ ],
467
"drive0": [
468
{
469
"busy": false,
470
@@ -XXX,XX +XXX,XX @@ write -P0x67 0x3fe0000 0x20000
471
{"return": ""}
472
{
473
"bitmaps": {
474
+ "backup-top": [
475
+ {
476
+ "busy": false,
477
+ "count": 67108864,
478
+ "granularity": 65536,
479
+ "persistent": false,
480
+ "recording": false
481
+ },
482
+ {
483
+ "busy": false,
484
+ "count": 458752,
485
+ "granularity": 65536,
486
+ "persistent": false,
487
+ "recording": false
488
+ }
489
+ ],
490
"drive0": [
491
{
492
"busy": false,
493
@@ -XXX,XX +XXX,XX @@ write -P0x67 0x3fe0000 0x20000
494
{"return": ""}
495
{
496
"bitmaps": {
497
+ "backup-top": [
498
+ {
499
+ "busy": false,
500
+ "count": 67108864,
501
+ "granularity": 65536,
502
+ "persistent": false,
503
+ "recording": false
504
+ },
505
+ {
506
+ "busy": false,
507
+ "count": 458752,
508
+ "granularity": 65536,
509
+ "persistent": false,
510
+ "recording": false
511
+ }
512
+ ],
513
"drive0": [
514
{
515
"busy": false,
516
@@ -XXX,XX +XXX,XX @@ write -P0x67 0x3fe0000 0x20000
517
{"return": ""}
518
{
519
"bitmaps": {
520
+ "backup-top": [
521
+ {
522
+ "busy": false,
523
+ "count": 67108864,
524
+ "granularity": 65536,
525
+ "persistent": false,
526
+ "recording": false
527
+ },
528
+ {
529
+ "busy": false,
530
+ "count": 458752,
531
+ "granularity": 65536,
532
+ "persistent": false,
533
+ "recording": false
534
+ }
535
+ ],
536
"drive0": [
537
{
538
"busy": false,
539
@@ -XXX,XX +XXX,XX @@ write -P0x67 0x3fe0000 0x20000
540
{"return": ""}
541
{
542
"bitmaps": {
543
+ "backup-top": [
544
+ {
545
+ "busy": false,
546
+ "count": 67108864,
547
+ "granularity": 65536,
548
+ "persistent": false,
549
+ "recording": false
550
+ },
551
+ {
552
+ "busy": false,
553
+ "count": 458752,
554
+ "granularity": 65536,
555
+ "persistent": false,
556
+ "recording": false
557
+ }
558
+ ],
559
"drive0": [
560
{
561
"busy": false,
562
@@ -XXX,XX +XXX,XX @@ write -P0x67 0x3fe0000 0x20000
563
{"return": ""}
564
{
565
"bitmaps": {
566
+ "backup-top": [
567
+ {
568
+ "busy": false,
569
+ "count": 67108864,
570
+ "granularity": 65536,
571
+ "persistent": false,
572
+ "recording": false
573
+ },
574
+ {
575
+ "busy": false,
576
+ "count": 458752,
577
+ "granularity": 65536,
578
+ "persistent": false,
579
+ "recording": false
580
+ }
581
+ ],
582
"drive0": [
583
{
584
"busy": false,
585
@@ -XXX,XX +XXX,XX @@ write -P0x67 0x3fe0000 0x20000
586
{"return": ""}
587
{
588
"bitmaps": {
589
+ "backup-top": [
590
+ {
591
+ "busy": false,
592
+ "count": 67108864,
593
+ "granularity": 65536,
594
+ "persistent": false,
595
+ "recording": false
596
+ },
597
+ {
598
+ "busy": false,
599
+ "count": 458752,
600
+ "granularity": 65536,
601
+ "persistent": false,
602
+ "recording": false
603
+ }
604
+ ],
605
"drive0": [
606
{
607
"busy": false,
608
@@ -XXX,XX +XXX,XX @@ write -P0x67 0x3fe0000 0x20000
609
{"return": ""}
610
{
611
"bitmaps": {
612
+ "backup-top": [
613
+ {
614
+ "busy": false,
615
+ "count": 67108864,
616
+ "granularity": 65536,
617
+ "persistent": false,
618
+ "recording": false
619
+ },
620
+ {
621
+ "busy": false,
622
+ "count": 458752,
623
+ "granularity": 65536,
624
+ "persistent": false,
625
+ "recording": false
626
+ }
627
+ ],
628
"drive0": [
629
{
630
"busy": false,
631
@@ -XXX,XX +XXX,XX @@ write -P0x67 0x3fe0000 0x20000
632
{"return": ""}
633
{
634
"bitmaps": {
635
+ "backup-top": [
636
+ {
637
+ "busy": false,
638
+ "count": 67108864,
639
+ "granularity": 65536,
640
+ "persistent": false,
641
+ "recording": false
642
+ },
643
+ {
644
+ "busy": false,
645
+ "count": 458752,
646
+ "granularity": 65536,
647
+ "persistent": false,
648
+ "recording": false
649
+ }
650
+ ],
651
"drive0": [
652
{
653
"busy": false,
654
@@ -XXX,XX +XXX,XX @@ write -P0x67 0x3fe0000 0x20000
655
{"return": ""}
656
{
657
"bitmaps": {
658
+ "backup-top": [
659
+ {
660
+ "busy": false,
661
+ "count": 67108864,
662
+ "granularity": 65536,
663
+ "persistent": false,
664
+ "recording": false
665
+ },
666
+ {
667
+ "busy": false,
668
+ "count": 458752,
669
+ "granularity": 65536,
670
+ "persistent": false,
671
+ "recording": false
672
+ }
673
+ ],
674
"drive0": [
675
{
676
"busy": false,
677
@@ -XXX,XX +XXX,XX @@ write -P0x67 0x3fe0000 0x20000
678
{"return": ""}
679
{
680
"bitmaps": {
681
+ "backup-top": [
682
+ {
683
+ "busy": false,
684
+ "count": 67108864,
685
+ "granularity": 65536,
686
+ "persistent": false,
687
+ "recording": false
688
+ },
689
+ {
690
+ "busy": false,
691
+ "count": 458752,
692
+ "granularity": 65536,
693
+ "persistent": false,
694
+ "recording": false
695
+ }
696
+ ],
697
"drive0": [
698
{
699
"busy": false,
700
--
169
--
701
2.34.1
170
2.40.1
diff view generated by jsdifflib
1
From: Peter Maydell <peter.maydell@linaro.org>
1
Stop using the .bdrv_co_io_plug() API because it is not multi-queue
2
2
block layer friendly. Use the new blk_io_plug_call() API to batch I/O
3
Coverity points out that we aren't checking the return value
3
submission instead.
4
from curl_easy_setopt() for any of the calls to it we make
4
5
in block/curl.c.
5
Note that a dev_max_batch check is dropped in laio_io_unplug() because
6
6
the semantics of unplug_fn() are different from .bdrv_co_unplug():
7
Some of these options are documented as always succeeding (e.g.
7
1. unplug_fn() is only called when the last blk_io_unplug() call occurs,
8
CURLOPT_VERBOSE) but others have documented failure cases (e.g.
8
not every time blk_io_unplug() is called.
9
CURLOPT_URL). For consistency we check every call, even the ones
9
2. unplug_fn() is per-thread, not per-BlockDriverState, so there is no
10
that theoretically cannot fail.
10
way to get per-BlockDriverState fields like dev_max_batch.
11
11
12
Fixes: Coverity CID 1459336, 1459482, 1460331
12
Therefore this condition cannot be moved to laio_unplug_fn(). It is not
13
Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
13
obvious that this condition affects performance in practice, so I am
14
Message-Id: <20220222152341.850419-3-peter.maydell@linaro.org>
14
removing it instead of trying to come up with a more complex mechanism
15
Reviewed-by: Hanna Reitz <hreitz@redhat.com>
15
to preserve the condition.
16
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
16
17
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
18
Reviewed-by: Eric Blake <eblake@redhat.com>
19
Acked-by: Kevin Wolf <kwolf@redhat.com>
20
Reviewed-by: Stefano Garzarella <sgarzare@redhat.com>
21
Message-id: 20230530180959.1108766-6-stefanha@redhat.com
22
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
17
---
23
---
18
block/curl.c | 90 +++++++++++++++++++++++++++++++++-------------------
24
include/block/raw-aio.h | 7 -------
19
1 file changed, 57 insertions(+), 33 deletions(-)
25
block/file-posix.c | 28 ----------------------------
20
26
block/linux-aio.c | 41 +++++++++++------------------------------
21
diff --git a/block/curl.c b/block/curl.c
27
3 files changed, 11 insertions(+), 65 deletions(-)
28
29
diff --git a/include/block/raw-aio.h b/include/block/raw-aio.h
22
index XXXXXXX..XXXXXXX 100644
30
index XXXXXXX..XXXXXXX 100644
23
--- a/block/curl.c
31
--- a/include/block/raw-aio.h
24
+++ b/block/curl.c
32
+++ b/include/block/raw-aio.h
25
@@ -XXX,XX +XXX,XX @@ static int curl_init_state(BDRVCURLState *s, CURLState *state)
33
@@ -XXX,XX +XXX,XX @@ int coroutine_fn laio_co_submit(int fd, uint64_t offset, QEMUIOVector *qiov,
26
if (!state->curl) {
34
27
return -EIO;
35
void laio_detach_aio_context(LinuxAioState *s, AioContext *old_context);
28
}
36
void laio_attach_aio_context(LinuxAioState *s, AioContext *new_context);
29
- curl_easy_setopt(state->curl, CURLOPT_URL, s->url);
37
-
30
- curl_easy_setopt(state->curl, CURLOPT_SSL_VERIFYPEER,
38
-/*
31
- (long) s->sslverify);
39
- * laio_io_plug/unplug work in the thread's current AioContext, therefore the
32
- curl_easy_setopt(state->curl, CURLOPT_SSL_VERIFYHOST,
40
- * caller must ensure that they are paired in the same IOThread.
33
- s->sslverify ? 2L : 0L);
41
- */
34
+ if (curl_easy_setopt(state->curl, CURLOPT_URL, s->url) ||
42
-void laio_io_plug(void);
35
+ curl_easy_setopt(state->curl, CURLOPT_SSL_VERIFYPEER,
43
-void laio_io_unplug(uint64_t dev_max_batch);
36
+ (long) s->sslverify) ||
44
#endif
37
+ curl_easy_setopt(state->curl, CURLOPT_SSL_VERIFYHOST,
45
/* io_uring.c - Linux io_uring implementation */
38
+ s->sslverify ? 2L : 0L)) {
46
#ifdef CONFIG_LINUX_IO_URING
39
+ goto err;
47
diff --git a/block/file-posix.c b/block/file-posix.c
48
index XXXXXXX..XXXXXXX 100644
49
--- a/block/file-posix.c
50
+++ b/block/file-posix.c
51
@@ -XXX,XX +XXX,XX @@ static int coroutine_fn raw_co_pwritev(BlockDriverState *bs, int64_t offset,
52
return raw_co_prw(bs, offset, bytes, qiov, QEMU_AIO_WRITE);
53
}
54
55
-static void coroutine_fn raw_co_io_plug(BlockDriverState *bs)
56
-{
57
- BDRVRawState __attribute__((unused)) *s = bs->opaque;
58
-#ifdef CONFIG_LINUX_AIO
59
- if (s->use_linux_aio) {
60
- laio_io_plug();
61
- }
62
-#endif
63
-}
64
-
65
-static void coroutine_fn raw_co_io_unplug(BlockDriverState *bs)
66
-{
67
- BDRVRawState __attribute__((unused)) *s = bs->opaque;
68
-#ifdef CONFIG_LINUX_AIO
69
- if (s->use_linux_aio) {
70
- laio_io_unplug(s->aio_max_batch);
71
- }
72
-#endif
73
-}
74
-
75
static int coroutine_fn raw_co_flush_to_disk(BlockDriverState *bs)
76
{
77
BDRVRawState *s = bs->opaque;
78
@@ -XXX,XX +XXX,XX @@ BlockDriver bdrv_file = {
79
.bdrv_co_copy_range_from = raw_co_copy_range_from,
80
.bdrv_co_copy_range_to = raw_co_copy_range_to,
81
.bdrv_refresh_limits = raw_refresh_limits,
82
- .bdrv_co_io_plug = raw_co_io_plug,
83
- .bdrv_co_io_unplug = raw_co_io_unplug,
84
.bdrv_attach_aio_context = raw_aio_attach_aio_context,
85
86
.bdrv_co_truncate = raw_co_truncate,
87
@@ -XXX,XX +XXX,XX @@ static BlockDriver bdrv_host_device = {
88
.bdrv_co_copy_range_from = raw_co_copy_range_from,
89
.bdrv_co_copy_range_to = raw_co_copy_range_to,
90
.bdrv_refresh_limits = raw_refresh_limits,
91
- .bdrv_co_io_plug = raw_co_io_plug,
92
- .bdrv_co_io_unplug = raw_co_io_unplug,
93
.bdrv_attach_aio_context = raw_aio_attach_aio_context,
94
95
.bdrv_co_truncate = raw_co_truncate,
96
@@ -XXX,XX +XXX,XX @@ static BlockDriver bdrv_host_cdrom = {
97
.bdrv_co_pwritev = raw_co_pwritev,
98
.bdrv_co_flush_to_disk = raw_co_flush_to_disk,
99
.bdrv_refresh_limits = cdrom_refresh_limits,
100
- .bdrv_co_io_plug = raw_co_io_plug,
101
- .bdrv_co_io_unplug = raw_co_io_unplug,
102
.bdrv_attach_aio_context = raw_aio_attach_aio_context,
103
104
.bdrv_co_truncate = raw_co_truncate,
105
@@ -XXX,XX +XXX,XX @@ static BlockDriver bdrv_host_cdrom = {
106
.bdrv_co_pwritev = raw_co_pwritev,
107
.bdrv_co_flush_to_disk = raw_co_flush_to_disk,
108
.bdrv_refresh_limits = cdrom_refresh_limits,
109
- .bdrv_co_io_plug = raw_co_io_plug,
110
- .bdrv_co_io_unplug = raw_co_io_unplug,
111
.bdrv_attach_aio_context = raw_aio_attach_aio_context,
112
113
.bdrv_co_truncate = raw_co_truncate,
114
diff --git a/block/linux-aio.c b/block/linux-aio.c
115
index XXXXXXX..XXXXXXX 100644
116
--- a/block/linux-aio.c
117
+++ b/block/linux-aio.c
118
@@ -XXX,XX +XXX,XX @@
119
#include "qemu/event_notifier.h"
120
#include "qemu/coroutine.h"
121
#include "qapi/error.h"
122
+#include "sysemu/block-backend.h"
123
124
/* Only used for assertions. */
125
#include "qemu/coroutine_int.h"
126
@@ -XXX,XX +XXX,XX @@ struct qemu_laiocb {
127
};
128
129
typedef struct {
130
- int plugged;
131
unsigned int in_queue;
132
unsigned int in_flight;
133
bool blocked;
134
@@ -XXX,XX +XXX,XX @@ static void qemu_laio_process_completions_and_submit(LinuxAioState *s)
135
{
136
qemu_laio_process_completions(s);
137
138
- if (!s->io_q.plugged && !QSIMPLEQ_EMPTY(&s->io_q.pending)) {
139
+ if (!QSIMPLEQ_EMPTY(&s->io_q.pending)) {
140
ioq_submit(s);
141
}
142
}
143
@@ -XXX,XX +XXX,XX @@ static void qemu_laio_poll_ready(EventNotifier *opaque)
144
static void ioq_init(LaioQueue *io_q)
145
{
146
QSIMPLEQ_INIT(&io_q->pending);
147
- io_q->plugged = 0;
148
io_q->in_queue = 0;
149
io_q->in_flight = 0;
150
io_q->blocked = false;
151
@@ -XXX,XX +XXX,XX @@ static uint64_t laio_max_batch(LinuxAioState *s, uint64_t dev_max_batch)
152
return max_batch;
153
}
154
155
-void laio_io_plug(void)
156
+static void laio_unplug_fn(void *opaque)
157
{
158
- AioContext *ctx = qemu_get_current_aio_context();
159
- LinuxAioState *s = aio_get_linux_aio(ctx);
160
+ LinuxAioState *s = opaque;
161
162
- s->io_q.plugged++;
163
-}
164
-
165
-void laio_io_unplug(uint64_t dev_max_batch)
166
-{
167
- AioContext *ctx = qemu_get_current_aio_context();
168
- LinuxAioState *s = aio_get_linux_aio(ctx);
169
-
170
- assert(s->io_q.plugged);
171
- s->io_q.plugged--;
172
-
173
- /*
174
- * Why max batch checking is performed here:
175
- * Another BDS may have queued requests with a higher dev_max_batch and
176
- * therefore in_queue could now exceed our dev_max_batch. Re-check the max
177
- * batch so we can honor our device's dev_max_batch.
178
- */
179
- if (s->io_q.in_queue >= laio_max_batch(s, dev_max_batch) ||
180
- (!s->io_q.plugged &&
181
- !s->io_q.blocked && !QSIMPLEQ_EMPTY(&s->io_q.pending))) {
182
+ if (!s->io_q.blocked && !QSIMPLEQ_EMPTY(&s->io_q.pending)) {
183
ioq_submit(s);
184
}
185
}
186
@@ -XXX,XX +XXX,XX @@ static int laio_do_submit(int fd, struct qemu_laiocb *laiocb, off_t offset,
187
188
QSIMPLEQ_INSERT_TAIL(&s->io_q.pending, laiocb, next);
189
s->io_q.in_queue++;
190
- if (!s->io_q.blocked &&
191
- (!s->io_q.plugged ||
192
- s->io_q.in_queue >= laio_max_batch(s, dev_max_batch))) {
193
- ioq_submit(s);
194
+ if (!s->io_q.blocked) {
195
+ if (s->io_q.in_queue >= laio_max_batch(s, dev_max_batch)) {
196
+ ioq_submit(s);
197
+ } else {
198
+ blk_io_plug_call(laio_unplug_fn, s);
40
+ }
199
+ }
41
if (s->cookie) {
42
- curl_easy_setopt(state->curl, CURLOPT_COOKIE, s->cookie);
43
+ if (curl_easy_setopt(state->curl, CURLOPT_COOKIE, s->cookie)) {
44
+ goto err;
45
+ }
46
+ }
47
+ if (curl_easy_setopt(state->curl, CURLOPT_TIMEOUT, (long)s->timeout) ||
48
+ curl_easy_setopt(state->curl, CURLOPT_WRITEFUNCTION,
49
+ (void *)curl_read_cb) ||
50
+ curl_easy_setopt(state->curl, CURLOPT_WRITEDATA, (void *)state) ||
51
+ curl_easy_setopt(state->curl, CURLOPT_PRIVATE, (void *)state) ||
52
+ curl_easy_setopt(state->curl, CURLOPT_AUTOREFERER, 1) ||
53
+ curl_easy_setopt(state->curl, CURLOPT_FOLLOWLOCATION, 1) ||
54
+ curl_easy_setopt(state->curl, CURLOPT_NOSIGNAL, 1) ||
55
+ curl_easy_setopt(state->curl, CURLOPT_ERRORBUFFER, state->errmsg) ||
56
+ curl_easy_setopt(state->curl, CURLOPT_FAILONERROR, 1)) {
57
+ goto err;
58
}
59
- curl_easy_setopt(state->curl, CURLOPT_TIMEOUT, (long)s->timeout);
60
- curl_easy_setopt(state->curl, CURLOPT_WRITEFUNCTION,
61
- (void *)curl_read_cb);
62
- curl_easy_setopt(state->curl, CURLOPT_WRITEDATA, (void *)state);
63
- curl_easy_setopt(state->curl, CURLOPT_PRIVATE, (void *)state);
64
- curl_easy_setopt(state->curl, CURLOPT_AUTOREFERER, 1);
65
- curl_easy_setopt(state->curl, CURLOPT_FOLLOWLOCATION, 1);
66
- curl_easy_setopt(state->curl, CURLOPT_NOSIGNAL, 1);
67
- curl_easy_setopt(state->curl, CURLOPT_ERRORBUFFER, state->errmsg);
68
- curl_easy_setopt(state->curl, CURLOPT_FAILONERROR, 1);
69
-
70
if (s->username) {
71
- curl_easy_setopt(state->curl, CURLOPT_USERNAME, s->username);
72
+ if (curl_easy_setopt(state->curl, CURLOPT_USERNAME, s->username)) {
73
+ goto err;
74
+ }
75
}
76
if (s->password) {
77
- curl_easy_setopt(state->curl, CURLOPT_PASSWORD, s->password);
78
+ if (curl_easy_setopt(state->curl, CURLOPT_PASSWORD, s->password)) {
79
+ goto err;
80
+ }
81
}
82
if (s->proxyusername) {
83
- curl_easy_setopt(state->curl,
84
- CURLOPT_PROXYUSERNAME, s->proxyusername);
85
+ if (curl_easy_setopt(state->curl,
86
+ CURLOPT_PROXYUSERNAME, s->proxyusername)) {
87
+ goto err;
88
+ }
89
}
90
if (s->proxypassword) {
91
- curl_easy_setopt(state->curl,
92
- CURLOPT_PROXYPASSWORD, s->proxypassword);
93
+ if (curl_easy_setopt(state->curl,
94
+ CURLOPT_PROXYPASSWORD, s->proxypassword)) {
95
+ goto err;
96
+ }
97
}
98
99
/* Restrict supported protocols to avoid security issues in the more
100
@@ -XXX,XX +XXX,XX @@ static int curl_init_state(BDRVCURLState *s, CURLState *state)
101
* Restricting protocols is only supported from 7.19.4 upwards.
102
*/
103
#if LIBCURL_VERSION_NUM >= 0x071304
104
- curl_easy_setopt(state->curl, CURLOPT_PROTOCOLS, PROTOCOLS);
105
- curl_easy_setopt(state->curl, CURLOPT_REDIR_PROTOCOLS, PROTOCOLS);
106
+ if (curl_easy_setopt(state->curl, CURLOPT_PROTOCOLS, PROTOCOLS) ||
107
+ curl_easy_setopt(state->curl, CURLOPT_REDIR_PROTOCOLS, PROTOCOLS)) {
108
+ goto err;
109
+ }
110
#endif
111
112
#ifdef DEBUG_VERBOSE
113
- curl_easy_setopt(state->curl, CURLOPT_VERBOSE, 1);
114
+ if (curl_easy_setopt(state->curl, CURLOPT_VERBOSE, 1)) {
115
+ goto err;
116
+ }
117
#endif
118
}
200
}
119
201
120
state->s = s;
121
122
return 0;
202
return 0;
123
+
124
+err:
125
+ curl_easy_cleanup(state->curl);
126
+ state->curl = NULL;
127
+ return -EIO;
128
}
129
130
/* Called with s->mutex held. */
131
@@ -XXX,XX +XXX,XX @@ static int curl_open(BlockDriverState *bs, QDict *options, int flags,
132
}
133
134
s->accept_range = false;
135
- curl_easy_setopt(state->curl, CURLOPT_NOBODY, 1);
136
- curl_easy_setopt(state->curl, CURLOPT_HEADERFUNCTION,
137
- curl_header_cb);
138
- curl_easy_setopt(state->curl, CURLOPT_HEADERDATA, s);
139
+ if (curl_easy_setopt(state->curl, CURLOPT_NOBODY, 1) ||
140
+ curl_easy_setopt(state->curl, CURLOPT_HEADERFUNCTION, curl_header_cb) ||
141
+ curl_easy_setopt(state->curl, CURLOPT_HEADERDATA, s)) {
142
+ pstrcpy(state->errmsg, CURL_ERROR_SIZE,
143
+ "curl library initialization failed.");
144
+ goto out;
145
+ }
146
if (curl_easy_perform(state->curl))
147
goto out;
148
if (curl_easy_getinfo(state->curl, CURLINFO_CONTENT_LENGTH_DOWNLOAD, &d)) {
149
@@ -XXX,XX +XXX,XX @@ static void curl_setup_preadv(BlockDriverState *bs, CURLAIOCB *acb)
150
151
snprintf(state->range, 127, "%" PRIu64 "-%" PRIu64, start, end);
152
trace_curl_setup_preadv(acb->bytes, start, state->range);
153
- curl_easy_setopt(state->curl, CURLOPT_RANGE, state->range);
154
-
155
- if (curl_multi_add_handle(s->multi, state->curl) != CURLM_OK) {
156
+ if (curl_easy_setopt(state->curl, CURLOPT_RANGE, state->range) ||
157
+ curl_multi_add_handle(s->multi, state->curl) != CURLM_OK) {
158
state->acb[0] = NULL;
159
acb->ret = -EIO;
160
161
--
203
--
162
2.34.1
204
2.40.1
diff view generated by jsdifflib
Deleted patch
1
From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
2
1
3
There is a bug in handling BDRV_REQ_NO_WAIT flag: we still may wait in
4
wait_serialising_requests() if request is unaligned. And this is
5
possible for the only user of this flag (preallocate filter) if
6
underlying file is unaligned to its request_alignment on start.
7
8
So, we have to fix preallocate filter to do only aligned preallocate
9
requests.
10
11
Next, we should fix generic block/io.c somehow. Keeping in mind that
12
preallocate is the only user of BDRV_REQ_NO_WAIT and that we have to
13
fix its behavior now, it seems more safe to just assert that we never
14
use BDRV_REQ_NO_WAIT with unaligned requests and add corresponding
15
comment. Let's do so.
16
17
Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
18
Reviewed-by: Denis V. Lunev <den@openvz.org>
19
Message-Id: <20220215121609.38570-1-vsementsov@virtuozzo.com>
20
[hreitz: Rebased on block GS/IO split]
21
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
22
---
23
include/block/block-common.h | 3 ++-
24
block/io.c | 4 ++++
25
block/preallocate.c | 15 ++++++++++++---
26
3 files changed, 18 insertions(+), 4 deletions(-)
27
28
diff --git a/include/block/block-common.h b/include/block/block-common.h
29
index XXXXXXX..XXXXXXX 100644
30
--- a/include/block/block-common.h
31
+++ b/include/block/block-common.h
32
@@ -XXX,XX +XXX,XX @@ typedef enum {
33
34
/*
35
* If we need to wait for other requests, just fail immediately. Used
36
- * only together with BDRV_REQ_SERIALISING.
37
+ * only together with BDRV_REQ_SERIALISING. Used only with requests aligned
38
+ * to request_alignment (corresponding assertions are in block/io.c).
39
*/
40
BDRV_REQ_NO_WAIT = 0x400,
41
42
diff --git a/block/io.c b/block/io.c
43
index XXXXXXX..XXXXXXX 100644
44
--- a/block/io.c
45
+++ b/block/io.c
46
@@ -XXX,XX +XXX,XX @@ static int coroutine_fn bdrv_co_do_zero_pwritev(BdrvChild *child,
47
48
padding = bdrv_init_padding(bs, offset, bytes, &pad);
49
if (padding) {
50
+ assert(!(flags & BDRV_REQ_NO_WAIT));
51
bdrv_make_request_serialising(req, align);
52
53
bdrv_padding_rmw_read(child, req, &pad, true);
54
@@ -XXX,XX +XXX,XX @@ int coroutine_fn bdrv_co_pwritev_part(BdrvChild *child,
55
* serialize the request to prevent interactions of the
56
* widened region with other transactions.
57
*/
58
+ assert(!(flags & BDRV_REQ_NO_WAIT));
59
bdrv_make_request_serialising(&req, align);
60
bdrv_padding_rmw_read(child, &req, &pad, false);
61
}
62
@@ -XXX,XX +XXX,XX @@ static int coroutine_fn bdrv_co_copy_range_internal(
63
/* TODO We can support BDRV_REQ_NO_FALLBACK here */
64
assert(!(read_flags & BDRV_REQ_NO_FALLBACK));
65
assert(!(write_flags & BDRV_REQ_NO_FALLBACK));
66
+ assert(!(read_flags & BDRV_REQ_NO_WAIT));
67
+ assert(!(write_flags & BDRV_REQ_NO_WAIT));
68
69
if (!dst || !dst->bs || !bdrv_is_inserted(dst->bs)) {
70
return -ENOMEDIUM;
71
diff --git a/block/preallocate.c b/block/preallocate.c
72
index XXXXXXX..XXXXXXX 100644
73
--- a/block/preallocate.c
74
+++ b/block/preallocate.c
75
@@ -XXX,XX +XXX,XX @@ static bool coroutine_fn handle_write(BlockDriverState *bs, int64_t offset,
76
int64_t end = offset + bytes;
77
int64_t prealloc_start, prealloc_end;
78
int ret;
79
+ uint32_t file_align = bs->file->bs->bl.request_alignment;
80
+ uint32_t prealloc_align = MAX(s->opts.prealloc_align, file_align);
81
+
82
+ assert(QEMU_IS_ALIGNED(prealloc_align, file_align));
83
84
if (!has_prealloc_perms(bs)) {
85
/* We don't have state neither should try to recover it */
86
@@ -XXX,XX +XXX,XX @@ static bool coroutine_fn handle_write(BlockDriverState *bs, int64_t offset,
87
88
/* Now we want new preallocation, as request writes beyond s->file_end. */
89
90
- prealloc_start = want_merge_zero ? MIN(offset, s->file_end) : s->file_end;
91
- prealloc_end = QEMU_ALIGN_UP(end + s->opts.prealloc_size,
92
- s->opts.prealloc_align);
93
+ prealloc_start = QEMU_ALIGN_UP(
94
+ want_merge_zero ? MIN(offset, s->file_end) : s->file_end,
95
+ file_align);
96
+ prealloc_end = QEMU_ALIGN_UP(
97
+ MAX(prealloc_start, end) + s->opts.prealloc_size,
98
+ prealloc_align);
99
+
100
+ want_merge_zero = want_merge_zero && (prealloc_start <= offset);
101
102
ret = bdrv_co_pwrite_zeroes(
103
bs->file, prealloc_start, prealloc_end - prealloc_start,
104
--
105
2.34.1
diff view generated by jsdifflib
Deleted patch
1
From: Thomas Huth <thuth@redhat.com>
2
1
3
iotest 040 already has some checks for the availability of the 'throttle'
4
driver, but some new code has been added in the course of time that
5
depends on 'throttle' but does not check for its availability. Add
6
a check to the TestCommitWithFilters class so that this iotest now
7
also passes again if 'throttle' has not been enabled in the QEMU
8
binaries.
9
10
Signed-off-by: Thomas Huth <thuth@redhat.com>
11
Message-Id: <20220223123127.3206042-1-thuth@redhat.com>
12
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
13
---
14
tests/qemu-iotests/040 | 1 +
15
1 file changed, 1 insertion(+)
16
17
diff --git a/tests/qemu-iotests/040 b/tests/qemu-iotests/040
18
index XXXXXXX..XXXXXXX 100755
19
--- a/tests/qemu-iotests/040
20
+++ b/tests/qemu-iotests/040
21
@@ -XXX,XX +XXX,XX @@ class TestCommitWithFilters(iotests.QMPTestCase):
22
pattern_file)
23
self.assertFalse('Pattern verification failed' in result)
24
25
+ @iotests.skip_if_unsupported(['throttle'])
26
def setUp(self):
27
qemu_img('create', '-f', iotests.imgfmt, self.img0, '64M')
28
qemu_img('create', '-f', iotests.imgfmt, self.img1, '64M')
29
--
30
2.34.1
diff view generated by jsdifflib
Deleted patch
1
From: Thomas Huth <thuth@redhat.com>
2
1
3
In TAP mode, the stdout is reserved for the TAP protocol, so we
4
have to make sure to mark other lines with a comment '#' character
5
at the beginning to avoid that the TAP parser at the other end
6
gets confused.
7
8
To test this condition, run "configure" for example with:
9
10
--block-drv-rw-whitelist=copy-before-write,qcow2,raw,file,host_device,blkdebug,null-co,copy-on-read
11
12
so that iotest 041 will report that some tests are not run due to
13
the missing "quorum" driver. Without this change, "make check-block"
14
fails since the meson tap parser gets confused by these messages.
15
16
Signed-off-by: Thomas Huth <thuth@redhat.com>
17
Message-Id: <20220223124353.3273898-1-thuth@redhat.com>
18
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
19
---
20
tests/qemu-iotests/testrunner.py | 5 ++++-
21
1 file changed, 4 insertions(+), 1 deletion(-)
22
23
diff --git a/tests/qemu-iotests/testrunner.py b/tests/qemu-iotests/testrunner.py
24
index XXXXXXX..XXXXXXX 100644
25
--- a/tests/qemu-iotests/testrunner.py
26
+++ b/tests/qemu-iotests/testrunner.py
27
@@ -XXX,XX +XXX,XX @@ def run_test(self, test: str,
28
description=res.description)
29
30
if res.casenotrun:
31
- print(res.casenotrun)
32
+ if self.tap:
33
+ print('#' + res.casenotrun.replace('\n', '\n#'))
34
+ else:
35
+ print(res.casenotrun)
36
37
return res
38
39
--
40
2.34.1
diff view generated by jsdifflib
Deleted patch
1
Drop the use of OUTPUT_DIR (test/qemu-iotests under the build
2
directory), and instead write test output files (.out.bad, .notrun, and
3
.casenotrun) to TEST_DIR.
4
1
5
With this, the same test can be run concurrently without the separate
6
instances interfering, because they will need separate TEST_DIRs anyway.
7
Running the same test separately is useful when running the iotests with
8
various format/protocol combinations in parallel, or when you just want
9
to aggressively exercise a single test (e.g. when it fails only
10
sporadically).
11
12
Putting this output into TEST_DIR means that it will stick around for
13
inspection after the test run is done (though running the same test in
14
the same TEST_DIR will overwrite it, just as it used to be); but given
15
that TEST_DIR is a scratch directory, it should be clear that users can
16
delete all of its content at any point. (And if TEST_DIR is on tmpfs,
17
it will just disappear on shutdown.) Contrarily, alternative approaches
18
that would put these output files into OUTPUT_DIR with some prefix to
19
differentiate between separate test runs might easily lead to cluttering
20
OUTPUT_DIR.
21
22
(This change means OUTPUT_DIR is no longer written to by the iotests, so
23
we can drop its usage altogether.)
24
25
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
26
Message-Id: <20220221172909.762858-1-hreitz@redhat.com>
27
[hreitz: Simplified `Path(os.path.join(x, y))` to `Path(x, y)`, as
28
suggested by Vladimir; and rebased on 9086c7639822b6
29
("tests/qemu-iotests: Rework the checks and spots using GNU
30
sed")]
31
Reviewed-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
32
---
33
tests/qemu-iotests/common.rc | 6 +++---
34
tests/qemu-iotests/iotests.py | 5 ++---
35
tests/qemu-iotests/testenv.py | 5 +----
36
tests/qemu-iotests/testrunner.py | 14 ++++++++------
37
4 files changed, 14 insertions(+), 16 deletions(-)
38
39
diff --git a/tests/qemu-iotests/common.rc b/tests/qemu-iotests/common.rc
40
index XXXXXXX..XXXXXXX 100644
41
--- a/tests/qemu-iotests/common.rc
42
+++ b/tests/qemu-iotests/common.rc
43
@@ -XXX,XX +XXX,XX @@
44
# bail out, setting up .notrun file
45
_notrun()
46
{
47
- echo "$*" >"$OUTPUT_DIR/$seq.notrun"
48
+ echo "$*" >"$TEST_DIR/$seq.notrun"
49
echo "$seq not run: $*"
50
status=0
51
exit
52
@@ -XXX,XX +XXX,XX @@ _img_info()
53
#
54
_casenotrun()
55
{
56
- echo " [case not run] $*" >>"$OUTPUT_DIR/$seq.casenotrun"
57
+ echo " [case not run] $*" >>"$TEST_DIR/$seq.casenotrun"
58
}
59
60
# just plain bail out
61
#
62
_fail()
63
{
64
- echo "$*" | tee -a "$OUTPUT_DIR/$seq.full"
65
+ echo "$*" | tee -a "$TEST_DIR/$seq.full"
66
echo "(see $seq.full for details)"
67
status=1
68
exit 1
69
diff --git a/tests/qemu-iotests/iotests.py b/tests/qemu-iotests/iotests.py
70
index XXXXXXX..XXXXXXX 100644
71
--- a/tests/qemu-iotests/iotests.py
72
+++ b/tests/qemu-iotests/iotests.py
73
@@ -XXX,XX +XXX,XX @@
74
75
imgfmt = os.environ.get('IMGFMT', 'raw')
76
imgproto = os.environ.get('IMGPROTO', 'file')
77
-output_dir = os.environ.get('OUTPUT_DIR', '.')
78
79
try:
80
test_dir = os.environ['TEST_DIR']
81
@@ -XXX,XX +XXX,XX @@ def notrun(reason):
82
# Each test in qemu-iotests has a number ("seq")
83
seq = os.path.basename(sys.argv[0])
84
85
- with open('%s/%s.notrun' % (output_dir, seq), 'w', encoding='utf-8') \
86
+ with open('%s/%s.notrun' % (test_dir, seq), 'w', encoding='utf-8') \
87
as outfile:
88
outfile.write(reason + '\n')
89
logger.warning("%s not run: %s", seq, reason)
90
@@ -XXX,XX +XXX,XX @@ def case_notrun(reason):
91
# Each test in qemu-iotests has a number ("seq")
92
seq = os.path.basename(sys.argv[0])
93
94
- with open('%s/%s.casenotrun' % (output_dir, seq), 'a', encoding='utf-8') \
95
+ with open('%s/%s.casenotrun' % (test_dir, seq), 'a', encoding='utf-8') \
96
as outfile:
97
outfile.write(' [case not run] ' + reason + '\n')
98
99
diff --git a/tests/qemu-iotests/testenv.py b/tests/qemu-iotests/testenv.py
100
index XXXXXXX..XXXXXXX 100644
101
--- a/tests/qemu-iotests/testenv.py
102
+++ b/tests/qemu-iotests/testenv.py
103
@@ -XXX,XX +XXX,XX @@ class TestEnv(ContextManager['TestEnv']):
104
# pylint: disable=too-many-instance-attributes
105
106
env_variables = ['PYTHONPATH', 'TEST_DIR', 'SOCK_DIR', 'SAMPLE_IMG_DIR',
107
- 'OUTPUT_DIR', 'PYTHON', 'QEMU_PROG', 'QEMU_IMG_PROG',
108
+ 'PYTHON', 'QEMU_PROG', 'QEMU_IMG_PROG',
109
'QEMU_IO_PROG', 'QEMU_NBD_PROG', 'QSD_PROG',
110
'QEMU_OPTIONS', 'QEMU_IMG_OPTIONS',
111
'QEMU_IO_OPTIONS', 'QEMU_IO_OPTIONS_NO_FMT',
112
@@ -XXX,XX +XXX,XX @@ def init_directories(self) -> None:
113
TEST_DIR
114
SOCK_DIR
115
SAMPLE_IMG_DIR
116
- OUTPUT_DIR
117
"""
118
119
# Path where qemu goodies live in this source tree.
120
@@ -XXX,XX +XXX,XX @@ def init_directories(self) -> None:
121
os.path.join(self.source_iotests,
122
'sample_images'))
123
124
- self.output_dir = os.getcwd() # OUTPUT_DIR
125
-
126
def init_binaries(self) -> None:
127
"""Init binary path variables:
128
PYTHON (for bash tests)
129
diff --git a/tests/qemu-iotests/testrunner.py b/tests/qemu-iotests/testrunner.py
130
index XXXXXXX..XXXXXXX 100644
131
--- a/tests/qemu-iotests/testrunner.py
132
+++ b/tests/qemu-iotests/testrunner.py
133
@@ -XXX,XX +XXX,XX @@ def do_run_test(self, test: str, mp: bool) -> TestResult:
134
"""
135
136
f_test = Path(test)
137
- f_bad = Path(f_test.name + '.out.bad')
138
- f_notrun = Path(f_test.name + '.notrun')
139
- f_casenotrun = Path(f_test.name + '.casenotrun')
140
f_reference = Path(self.find_reference(test))
141
142
if not f_test.exists():
143
@@ -XXX,XX +XXX,XX @@ def do_run_test(self, test: str, mp: bool) -> TestResult:
144
description='No qualified output '
145
f'(expected {f_reference})')
146
147
- for p in (f_bad, f_notrun, f_casenotrun):
148
- silent_unlink(p)
149
-
150
args = [str(f_test.resolve())]
151
env = self.env.prepare_subprocess(args)
152
if mp:
153
@@ -XXX,XX +XXX,XX @@ def do_run_test(self, test: str, mp: bool) -> TestResult:
154
env[d] = os.path.join(env[d], f_test.name)
155
Path(env[d]).mkdir(parents=True, exist_ok=True)
156
157
+ test_dir = env['TEST_DIR']
158
+ f_bad = Path(test_dir, f_test.name + '.out.bad')
159
+ f_notrun = Path(test_dir, f_test.name + '.notrun')
160
+ f_casenotrun = Path(test_dir, f_test.name + '.casenotrun')
161
+
162
+ for p in (f_notrun, f_casenotrun):
163
+ silent_unlink(p)
164
+
165
t0 = time.time()
166
with f_bad.open('w', encoding="utf-8") as f:
167
with subprocess.Popen(args, cwd=str(f_test.parent), env=env,
168
--
169
2.34.1
diff view generated by jsdifflib
1
From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
1
No block driver implements .bdrv_co_io_plug() anymore. Get rid of the
2
function pointers.
2
3
3
Add new block driver handlers and corresponding generic wrappers.
4
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
4
It will be used to allow copy-before-write filter to provide
5
Reviewed-by: Eric Blake <eblake@redhat.com>
5
reach fleecing interface in further commit.
6
Reviewed-by: Stefano Garzarella <sgarzare@redhat.com>
7
Acked-by: Kevin Wolf <kwolf@redhat.com>
8
Message-id: 20230530180959.1108766-7-stefanha@redhat.com
9
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
10
---
11
include/block/block-io.h | 3 ---
12
include/block/block_int-common.h | 11 ----------
13
block/io.c | 37 --------------------------------
14
3 files changed, 51 deletions(-)
6
15
7
In future this approach may be used to allow reading qcow2 internal
16
diff --git a/include/block/block-io.h b/include/block/block-io.h
8
snapshots, for example to export them through NBD.
17
index XXXXXXX..XXXXXXX 100644
9
18
--- a/include/block/block-io.h
10
Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
19
+++ b/include/block/block-io.h
11
Reviewed-by: Hanna Reitz <hreitz@redhat.com>
20
@@ -XXX,XX +XXX,XX @@ void coroutine_fn bdrv_co_leave(BlockDriverState *bs, AioContext *old_ctx);
12
Message-Id: <20220303194349.2304213-11-vsementsov@virtuozzo.com>
21
13
[hreitz: Rebased on block GS/IO split]
22
AioContext *child_of_bds_get_parent_aio_context(BdrvChild *c);
14
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
23
15
---
24
-void coroutine_fn GRAPH_RDLOCK bdrv_co_io_plug(BlockDriverState *bs);
16
include/block/block_int-common.h | 18 ++++++++
25
-void coroutine_fn GRAPH_RDLOCK bdrv_co_io_unplug(BlockDriverState *bs);
17
include/block/block_int-io.h | 9 ++++
26
-
18
block/io.c | 72 ++++++++++++++++++++++++++++++++
27
bool coroutine_fn GRAPH_RDLOCK
19
3 files changed, 99 insertions(+)
28
bdrv_co_can_store_new_dirty_bitmap(BlockDriverState *bs, const char *name,
20
29
uint32_t granularity, Error **errp);
21
diff --git a/include/block/block_int-common.h b/include/block/block_int-common.h
30
diff --git a/include/block/block_int-common.h b/include/block/block_int-common.h
22
index XXXXXXX..XXXXXXX 100644
31
index XXXXXXX..XXXXXXX 100644
23
--- a/include/block/block_int-common.h
32
--- a/include/block/block_int-common.h
24
+++ b/include/block/block_int-common.h
33
+++ b/include/block/block_int-common.h
25
@@ -XXX,XX +XXX,XX @@ struct BlockDriver {
34
@@ -XXX,XX +XXX,XX @@ struct BlockDriver {
26
bool want_zero, int64_t offset, int64_t bytes, int64_t *pnum,
35
void coroutine_fn GRAPH_RDLOCK_PTR (*bdrv_co_debug_event)(
27
int64_t *map, BlockDriverState **file);
36
BlockDriverState *bs, BlkdebugEvent event);
28
37
29
+ /*
38
- /* io queue for linux-aio */
30
+ * Snapshot-access API.
39
- void coroutine_fn GRAPH_RDLOCK_PTR (*bdrv_co_io_plug)(BlockDriverState *bs);
31
+ *
40
- void coroutine_fn GRAPH_RDLOCK_PTR (*bdrv_co_io_unplug)(
32
+ * Block-driver may provide snapshot-access API: special functions to access
41
- BlockDriverState *bs);
33
+ * some internal "snapshot". The functions are similar with normal
42
-
34
+ * read/block_status/discard handler, but don't have any specific handling
43
bool (*bdrv_supports_persistent_dirty_bitmap)(BlockDriverState *bs);
35
+ * in generic block-layer: no serializing, no alignment, no tracked
44
36
+ * requests. So, block-driver that realizes these APIs is fully responsible
45
bool coroutine_fn GRAPH_RDLOCK_PTR (*bdrv_co_can_store_new_dirty_bitmap)(
37
+ * for synchronization between snapshot-access API and normal IO requests.
46
@@ -XXX,XX +XXX,XX @@ struct BlockDriverState {
38
+ */
47
unsigned int in_flight;
39
+ int coroutine_fn (*bdrv_co_preadv_snapshot)(BlockDriverState *bs,
48
unsigned int serialising_in_flight;
40
+ int64_t offset, int64_t bytes, QEMUIOVector *qiov, size_t qiov_offset);
49
41
+ int coroutine_fn (*bdrv_co_snapshot_block_status)(BlockDriverState *bs,
50
- /*
42
+ bool want_zero, int64_t offset, int64_t bytes, int64_t *pnum,
51
- * counter for nested bdrv_io_plug.
43
+ int64_t *map, BlockDriverState **file);
52
- * Accessed with atomic ops.
44
+ int coroutine_fn (*bdrv_co_pdiscard_snapshot)(BlockDriverState *bs,
53
- */
45
+ int64_t offset, int64_t bytes);
54
- unsigned io_plugged;
46
+
55
-
47
/*
56
/* do we need to tell the quest if we have a volatile write cache? */
48
* Invalidate any cached meta-data.
57
int enable_write_cache;
49
*/
58
50
diff --git a/include/block/block_int-io.h b/include/block/block_int-io.h
51
index XXXXXXX..XXXXXXX 100644
52
--- a/include/block/block_int-io.h
53
+++ b/include/block/block_int-io.h
54
@@ -XXX,XX +XXX,XX @@
55
* the I/O API.
56
*/
57
58
+int coroutine_fn bdrv_co_preadv_snapshot(BdrvChild *child,
59
+ int64_t offset, int64_t bytes, QEMUIOVector *qiov, size_t qiov_offset);
60
+int coroutine_fn bdrv_co_snapshot_block_status(BlockDriverState *bs,
61
+ bool want_zero, int64_t offset, int64_t bytes, int64_t *pnum,
62
+ int64_t *map, BlockDriverState **file);
63
+int coroutine_fn bdrv_co_pdiscard_snapshot(BlockDriverState *bs,
64
+ int64_t offset, int64_t bytes);
65
+
66
+
67
int coroutine_fn bdrv_co_preadv(BdrvChild *child,
68
int64_t offset, int64_t bytes, QEMUIOVector *qiov,
69
BdrvRequestFlags flags);
70
diff --git a/block/io.c b/block/io.c
59
diff --git a/block/io.c b/block/io.c
71
index XXXXXXX..XXXXXXX 100644
60
index XXXXXXX..XXXXXXX 100644
72
--- a/block/io.c
61
--- a/block/io.c
73
+++ b/block/io.c
62
+++ b/block/io.c
74
@@ -XXX,XX +XXX,XX @@ void bdrv_cancel_in_flight(BlockDriverState *bs)
63
@@ -XXX,XX +XXX,XX @@ void *qemu_try_blockalign0(BlockDriverState *bs, size_t size)
75
bs->drv->bdrv_cancel_in_flight(bs);
64
return mem;
76
}
77
}
65
}
78
+
66
79
+int coroutine_fn
67
-void coroutine_fn bdrv_co_io_plug(BlockDriverState *bs)
80
+bdrv_co_preadv_snapshot(BdrvChild *child, int64_t offset, int64_t bytes,
68
-{
81
+ QEMUIOVector *qiov, size_t qiov_offset)
69
- BdrvChild *child;
82
+{
70
- IO_CODE();
83
+ BlockDriverState *bs = child->bs;
71
- assert_bdrv_graph_readable();
84
+ BlockDriver *drv = bs->drv;
72
-
85
+ int ret;
73
- QLIST_FOREACH(child, &bs->children, next) {
86
+ IO_CODE();
74
- bdrv_co_io_plug(child->bs);
87
+
75
- }
88
+ if (!drv) {
76
-
89
+ return -ENOMEDIUM;
77
- if (qatomic_fetch_inc(&bs->io_plugged) == 0) {
90
+ }
78
- BlockDriver *drv = bs->drv;
91
+
79
- if (drv && drv->bdrv_co_io_plug) {
92
+ if (!drv->bdrv_co_preadv_snapshot) {
80
- drv->bdrv_co_io_plug(bs);
93
+ return -ENOTSUP;
81
- }
94
+ }
82
- }
95
+
83
-}
96
+ bdrv_inc_in_flight(bs);
84
-
97
+ ret = drv->bdrv_co_preadv_snapshot(bs, offset, bytes, qiov, qiov_offset);
85
-void coroutine_fn bdrv_co_io_unplug(BlockDriverState *bs)
98
+ bdrv_dec_in_flight(bs);
86
-{
99
+
87
- BdrvChild *child;
100
+ return ret;
88
- IO_CODE();
101
+}
89
- assert_bdrv_graph_readable();
102
+
90
-
103
+int coroutine_fn
91
- assert(bs->io_plugged);
104
+bdrv_co_snapshot_block_status(BlockDriverState *bs,
92
- if (qatomic_fetch_dec(&bs->io_plugged) == 1) {
105
+ bool want_zero, int64_t offset, int64_t bytes,
93
- BlockDriver *drv = bs->drv;
106
+ int64_t *pnum, int64_t *map,
94
- if (drv && drv->bdrv_co_io_unplug) {
107
+ BlockDriverState **file)
95
- drv->bdrv_co_io_unplug(bs);
108
+{
96
- }
109
+ BlockDriver *drv = bs->drv;
97
- }
110
+ int ret;
98
-
111
+ IO_CODE();
99
- QLIST_FOREACH(child, &bs->children, next) {
112
+
100
- bdrv_co_io_unplug(child->bs);
113
+ if (!drv) {
101
- }
114
+ return -ENOMEDIUM;
102
-}
115
+ }
103
-
116
+
104
/* Helper that undoes bdrv_register_buf() when it fails partway through */
117
+ if (!drv->bdrv_co_snapshot_block_status) {
105
static void GRAPH_RDLOCK
118
+ return -ENOTSUP;
106
bdrv_register_buf_rollback(BlockDriverState *bs, void *host, size_t size,
119
+ }
120
+
121
+ bdrv_inc_in_flight(bs);
122
+ ret = drv->bdrv_co_snapshot_block_status(bs, want_zero, offset, bytes,
123
+ pnum, map, file);
124
+ bdrv_dec_in_flight(bs);
125
+
126
+ return ret;
127
+}
128
+
129
+int coroutine_fn
130
+bdrv_co_pdiscard_snapshot(BlockDriverState *bs, int64_t offset, int64_t bytes)
131
+{
132
+ BlockDriver *drv = bs->drv;
133
+ int ret;
134
+ IO_CODE();
135
+
136
+ if (!drv) {
137
+ return -ENOMEDIUM;
138
+ }
139
+
140
+ if (!drv->bdrv_co_pdiscard_snapshot) {
141
+ return -ENOTSUP;
142
+ }
143
+
144
+ bdrv_inc_in_flight(bs);
145
+ ret = drv->bdrv_co_pdiscard_snapshot(bs, offset, bytes);
146
+ bdrv_dec_in_flight(bs);
147
+
148
+ return ret;
149
+}
150
--
107
--
151
2.34.1
108
2.40.1
diff view generated by jsdifflib
1
From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
1
From: Stefano Garzarella <sgarzare@redhat.com>
2
2
3
We are going to complicate bitmap initialization in the further
3
Some virtio-blk drivers (e.g. virtio-blk-vhost-vdpa) supports the fd
4
commit. And in future, backup job will be able to work without filter
4
passing. Let's expose this to the user, so the management layer
5
(when source is immutable), so we'll need same bitmap initialization in
5
can pass the file descriptor of an already opened path.
6
copy-before-write filter and in backup job. So, it's reasonable to do
7
it in block-copy.
8
6
9
Note that for now cbw_open() is the only caller of
7
If the libblkio virtio-blk driver supports fd passing, let's always
10
block_copy_state_new().
8
use qemu_open() to open the `path`, so we can handle fd passing
9
from the management layer through the "/dev/fdset/N" special path.
11
10
12
Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
11
Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com>
13
Reviewed-by: Hanna Reitz <hreitz@redhat.com>
12
Signed-off-by: Stefano Garzarella <sgarzare@redhat.com>
14
Message-Id: <20220303194349.2304213-2-vsementsov@virtuozzo.com>
13
Message-id: 20230530071941.8954-2-sgarzare@redhat.com
15
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
14
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
16
---
15
---
17
block/block-copy.c | 1 +
16
block/blkio.c | 53 ++++++++++++++++++++++++++++++++++++++++++---------
18
block/copy-before-write.c | 4 ----
17
1 file changed, 44 insertions(+), 9 deletions(-)
19
2 files changed, 1 insertion(+), 4 deletions(-)
20
18
21
diff --git a/block/block-copy.c b/block/block-copy.c
19
diff --git a/block/blkio.c b/block/blkio.c
22
index XXXXXXX..XXXXXXX 100644
20
index XXXXXXX..XXXXXXX 100644
23
--- a/block/block-copy.c
21
--- a/block/blkio.c
24
+++ b/block/block-copy.c
22
+++ b/block/blkio.c
25
@@ -XXX,XX +XXX,XX @@ BlockCopyState *block_copy_state_new(BdrvChild *source, BdrvChild *target,
23
@@ -XXX,XX +XXX,XX @@ static int blkio_virtio_blk_common_open(BlockDriverState *bs,
26
return NULL;
27
}
28
bdrv_disable_dirty_bitmap(copy_bitmap);
29
+ bdrv_set_dirty_bitmap(copy_bitmap, 0, bdrv_dirty_bitmap_size(copy_bitmap));
30
31
/*
32
* If source is in backing chain of target assume that target is going to be
33
diff --git a/block/copy-before-write.c b/block/copy-before-write.c
34
index XXXXXXX..XXXXXXX 100644
35
--- a/block/copy-before-write.c
36
+++ b/block/copy-before-write.c
37
@@ -XXX,XX +XXX,XX @@ static int cbw_open(BlockDriverState *bs, QDict *options, int flags,
38
Error **errp)
39
{
24
{
40
BDRVCopyBeforeWriteState *s = bs->opaque;
25
const char *path = qdict_get_try_str(options, "path");
41
- BdrvDirtyBitmap *copy_bitmap;
26
BDRVBlkioState *s = bs->opaque;
42
27
- int ret;
43
bs->file = bdrv_open_child(NULL, options, "file", bs, &child_of_bds,
28
+ bool fd_supported = false;
44
BDRV_CHILD_FILTERED | BDRV_CHILD_PRIMARY,
29
+ int fd, ret;
45
@@ -XXX,XX +XXX,XX @@ static int cbw_open(BlockDriverState *bs, QDict *options, int flags,
30
31
if (!path) {
32
error_setg(errp, "missing 'path' option");
46
return -EINVAL;
33
return -EINVAL;
47
}
34
}
48
35
49
- copy_bitmap = block_copy_dirty_bitmap(s->bcs);
36
- ret = blkio_set_str(s->blkio, "path", path);
50
- bdrv_set_dirty_bitmap(copy_bitmap, 0, bdrv_dirty_bitmap_size(copy_bitmap));
37
- qdict_del(options, "path");
38
- if (ret < 0) {
39
- error_setg_errno(errp, -ret, "failed to set path: %s",
40
- blkio_get_error_msg());
41
- return ret;
42
- }
51
-
43
-
44
if (!(flags & BDRV_O_NOCACHE)) {
45
error_setg(errp, "cache.direct=off is not supported");
46
return -EINVAL;
47
}
48
+
49
+ if (blkio_get_int(s->blkio, "fd", &fd) == 0) {
50
+ fd_supported = true;
51
+ }
52
+
53
+ /*
54
+ * If the libblkio driver supports fd passing, let's always use qemu_open()
55
+ * to open the `path`, so we can handle fd passing from the management
56
+ * layer through the "/dev/fdset/N" special path.
57
+ */
58
+ if (fd_supported) {
59
+ int open_flags;
60
+
61
+ if (flags & BDRV_O_RDWR) {
62
+ open_flags = O_RDWR;
63
+ } else {
64
+ open_flags = O_RDONLY;
65
+ }
66
+
67
+ fd = qemu_open(path, open_flags, errp);
68
+ if (fd < 0) {
69
+ return -EINVAL;
70
+ }
71
+
72
+ ret = blkio_set_int(s->blkio, "fd", fd);
73
+ if (ret < 0) {
74
+ error_setg_errno(errp, -ret, "failed to set fd: %s",
75
+ blkio_get_error_msg());
76
+ qemu_close(fd);
77
+ return ret;
78
+ }
79
+ } else {
80
+ ret = blkio_set_str(s->blkio, "path", path);
81
+ if (ret < 0) {
82
+ error_setg_errno(errp, -ret, "failed to set path: %s",
83
+ blkio_get_error_msg());
84
+ return ret;
85
+ }
86
+ }
87
+
88
+ qdict_del(options, "path");
89
+
52
return 0;
90
return 0;
53
}
91
}
54
92
55
--
93
--
56
2.34.1
94
2.40.1
diff view generated by jsdifflib
Deleted patch
1
From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
2
1
3
That simplifies handling failure in existing code and in further new
4
usage of bdrv_merge_dirty_bitmap().
5
6
Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
7
Reviewed-by: Hanna Reitz <hreitz@redhat.com>
8
Message-Id: <20220303194349.2304213-3-vsementsov@virtuozzo.com>
9
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
10
---
11
include/block/dirty-bitmap.h | 2 +-
12
block/dirty-bitmap.c | 9 +++++++--
13
block/monitor/bitmap-qmp-cmds.c | 5 +----
14
3 files changed, 9 insertions(+), 7 deletions(-)
15
16
diff --git a/include/block/dirty-bitmap.h b/include/block/dirty-bitmap.h
17
index XXXXXXX..XXXXXXX 100644
18
--- a/include/block/dirty-bitmap.h
19
+++ b/include/block/dirty-bitmap.h
20
@@ -XXX,XX +XXX,XX @@ void bdrv_dirty_bitmap_set_persistence(BdrvDirtyBitmap *bitmap,
21
bool persistent);
22
void bdrv_dirty_bitmap_set_inconsistent(BdrvDirtyBitmap *bitmap);
23
void bdrv_dirty_bitmap_set_busy(BdrvDirtyBitmap *bitmap, bool busy);
24
-void bdrv_merge_dirty_bitmap(BdrvDirtyBitmap *dest, const BdrvDirtyBitmap *src,
25
+bool bdrv_merge_dirty_bitmap(BdrvDirtyBitmap *dest, const BdrvDirtyBitmap *src,
26
HBitmap **backup, Error **errp);
27
void bdrv_dirty_bitmap_skip_store(BdrvDirtyBitmap *bitmap, bool skip);
28
bool bdrv_dirty_bitmap_get(BdrvDirtyBitmap *bitmap, int64_t offset);
29
diff --git a/block/dirty-bitmap.c b/block/dirty-bitmap.c
30
index XXXXXXX..XXXXXXX 100644
31
--- a/block/dirty-bitmap.c
32
+++ b/block/dirty-bitmap.c
33
@@ -XXX,XX +XXX,XX @@ bool bdrv_dirty_bitmap_next_dirty_area(BdrvDirtyBitmap *bitmap,
34
* Ensures permissions on bitmaps are reasonable; use for public API.
35
*
36
* @backup: If provided, make a copy of dest here prior to merge.
37
+ *
38
+ * Returns true on success, false on failure. In case of failure bitmaps are
39
+ * untouched.
40
*/
41
-void bdrv_merge_dirty_bitmap(BdrvDirtyBitmap *dest, const BdrvDirtyBitmap *src,
42
+bool bdrv_merge_dirty_bitmap(BdrvDirtyBitmap *dest, const BdrvDirtyBitmap *src,
43
HBitmap **backup, Error **errp)
44
{
45
- bool ret;
46
+ bool ret = false;
47
48
bdrv_dirty_bitmaps_lock(dest->bs);
49
if (src->bs != dest->bs) {
50
@@ -XXX,XX +XXX,XX @@ out:
51
if (src->bs != dest->bs) {
52
bdrv_dirty_bitmaps_unlock(src->bs);
53
}
54
+
55
+ return ret;
56
}
57
58
/**
59
diff --git a/block/monitor/bitmap-qmp-cmds.c b/block/monitor/bitmap-qmp-cmds.c
60
index XXXXXXX..XXXXXXX 100644
61
--- a/block/monitor/bitmap-qmp-cmds.c
62
+++ b/block/monitor/bitmap-qmp-cmds.c
63
@@ -XXX,XX +XXX,XX @@ BdrvDirtyBitmap *block_dirty_bitmap_merge(const char *node, const char *target,
64
BlockDriverState *bs;
65
BdrvDirtyBitmap *dst, *src, *anon;
66
BlockDirtyBitmapMergeSourceList *lst;
67
- Error *local_err = NULL;
68
69
GLOBAL_STATE_CODE();
70
71
@@ -XXX,XX +XXX,XX @@ BdrvDirtyBitmap *block_dirty_bitmap_merge(const char *node, const char *target,
72
abort();
73
}
74
75
- bdrv_merge_dirty_bitmap(anon, src, NULL, &local_err);
76
- if (local_err) {
77
- error_propagate(errp, local_err);
78
+ if (!bdrv_merge_dirty_bitmap(anon, src, NULL, errp)) {
79
dst = NULL;
80
goto out;
81
}
82
--
83
2.34.1
diff view generated by jsdifflib
Deleted patch
1
From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
2
1
3
This will be used in the following commit to bring "incremental" mode
4
to copy-before-write filter.
5
6
Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
7
Reviewed-by: Hanna Reitz <hreitz@redhat.com>
8
Message-Id: <20220303194349.2304213-4-vsementsov@virtuozzo.com>
9
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
10
---
11
include/block/block-copy.h | 1 +
12
block/block-copy.c | 14 +++++++++++++-
13
block/copy-before-write.c | 2 +-
14
3 files changed, 15 insertions(+), 2 deletions(-)
15
16
diff --git a/include/block/block-copy.h b/include/block/block-copy.h
17
index XXXXXXX..XXXXXXX 100644
18
--- a/include/block/block-copy.h
19
+++ b/include/block/block-copy.h
20
@@ -XXX,XX +XXX,XX @@ typedef struct BlockCopyState BlockCopyState;
21
typedef struct BlockCopyCallState BlockCopyCallState;
22
23
BlockCopyState *block_copy_state_new(BdrvChild *source, BdrvChild *target,
24
+ const BdrvDirtyBitmap *bitmap,
25
Error **errp);
26
27
/* Function should be called prior any actual copy request */
28
diff --git a/block/block-copy.c b/block/block-copy.c
29
index XXXXXXX..XXXXXXX 100644
30
--- a/block/block-copy.c
31
+++ b/block/block-copy.c
32
@@ -XXX,XX +XXX,XX @@ static int64_t block_copy_calculate_cluster_size(BlockDriverState *target,
33
}
34
35
BlockCopyState *block_copy_state_new(BdrvChild *source, BdrvChild *target,
36
+ const BdrvDirtyBitmap *bitmap,
37
Error **errp)
38
{
39
+ ERRP_GUARD();
40
BlockCopyState *s;
41
int64_t cluster_size;
42
BdrvDirtyBitmap *copy_bitmap;
43
@@ -XXX,XX +XXX,XX @@ BlockCopyState *block_copy_state_new(BdrvChild *source, BdrvChild *target,
44
return NULL;
45
}
46
bdrv_disable_dirty_bitmap(copy_bitmap);
47
- bdrv_set_dirty_bitmap(copy_bitmap, 0, bdrv_dirty_bitmap_size(copy_bitmap));
48
+ if (bitmap) {
49
+ if (!bdrv_merge_dirty_bitmap(copy_bitmap, bitmap, NULL, errp)) {
50
+ error_prepend(errp, "Failed to merge bitmap '%s' to internal "
51
+ "copy-bitmap: ", bdrv_dirty_bitmap_name(bitmap));
52
+ bdrv_release_dirty_bitmap(copy_bitmap);
53
+ return NULL;
54
+ }
55
+ } else {
56
+ bdrv_set_dirty_bitmap(copy_bitmap, 0,
57
+ bdrv_dirty_bitmap_size(copy_bitmap));
58
+ }
59
60
/*
61
* If source is in backing chain of target assume that target is going to be
62
diff --git a/block/copy-before-write.c b/block/copy-before-write.c
63
index XXXXXXX..XXXXXXX 100644
64
--- a/block/copy-before-write.c
65
+++ b/block/copy-before-write.c
66
@@ -XXX,XX +XXX,XX @@ static int cbw_open(BlockDriverState *bs, QDict *options, int flags,
67
((BDRV_REQ_FUA | BDRV_REQ_MAY_UNMAP | BDRV_REQ_NO_FALLBACK) &
68
bs->file->bs->supported_zero_flags);
69
70
- s->bcs = block_copy_state_new(bs->file, s->target, errp);
71
+ s->bcs = block_copy_state_new(bs->file, s->target, NULL, errp);
72
if (!s->bcs) {
73
error_prepend(errp, "Cannot create block-copy-state: ");
74
return -EINVAL;
75
--
76
2.34.1
diff view generated by jsdifflib
1
From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
1
From: Stefano Garzarella <sgarzare@redhat.com>
2
2
3
This brings "incremental" mode to copy-before-write filter: user can
3
The virtio-blk-vhost-vdpa driver in libblkio 1.3.0 supports the fd
4
specify bitmap so that filter will copy only "dirty" areas.
4
passing through the new 'fd' property.
5
5
6
Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
6
Since now we are using qemu_open() on '@path' if the virtio-blk driver
7
Message-Id: <20220303194349.2304213-5-vsementsov@virtuozzo.com>
7
supports the fd passing, let's announce it.
8
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
8
In this way, the management layer can pass the file descriptor of an
9
already opened vhost-vdpa character device. This is useful especially
10
when the device can only be accessed with certain privileges.
11
12
Add the '@fdset' feature only when the virtio-blk-vhost-vdpa driver
13
in libblkio supports it.
14
15
Suggested-by: Markus Armbruster <armbru@redhat.com>
16
Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com>
17
Signed-off-by: Stefano Garzarella <sgarzare@redhat.com>
18
Message-id: 20230530071941.8954-3-sgarzare@redhat.com
19
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
9
---
20
---
10
qapi/block-core.json | 10 +++++++-
21
qapi/block-core.json | 6 ++++++
11
block/copy-before-write.c | 51 ++++++++++++++++++++++++++++++++++++++-
22
meson.build | 4 ++++
12
2 files changed, 59 insertions(+), 2 deletions(-)
23
2 files changed, 10 insertions(+)
13
24
14
diff --git a/qapi/block-core.json b/qapi/block-core.json
25
diff --git a/qapi/block-core.json b/qapi/block-core.json
15
index XXXXXXX..XXXXXXX 100644
26
index XXXXXXX..XXXXXXX 100644
16
--- a/qapi/block-core.json
27
--- a/qapi/block-core.json
17
+++ b/qapi/block-core.json
28
+++ b/qapi/block-core.json
18
@@ -XXX,XX +XXX,XX @@
29
@@ -XXX,XX +XXX,XX @@
19
#
30
#
20
# @target: The target for copy-before-write operations.
31
# @path: path to the vhost-vdpa character device.
21
#
32
#
22
+# @bitmap: If specified, copy-before-write filter will do
33
+# Features:
23
+# copy-before-write operations only for dirty regions of the
34
+# @fdset: Member @path supports the special "/dev/fdset/N" path
24
+# bitmap. Bitmap size must be equal to length of file and
35
+# (since 8.1)
25
+# target child of the filter. Note also, that bitmap is used
26
+# only to initialize internal bitmap of the process, so further
27
+# modifications (or removing) of specified bitmap doesn't
28
+# influence the filter. (Since 7.0)
29
+#
36
+#
30
# Since: 6.2
37
# Since: 7.2
31
##
38
##
32
{ 'struct': 'BlockdevOptionsCbw',
39
{ 'struct': 'BlockdevOptionsVirtioBlkVhostVdpa',
33
'base': 'BlockdevOptionsGenericFormat',
40
'data': { 'path': 'str' },
34
- 'data': { 'target': 'BlockdevRef' } }
41
+ 'features': [ { 'name' :'fdset',
35
+ 'data': { 'target': 'BlockdevRef', '*bitmap': 'BlockDirtyBitmap' } }
42
+ 'if': 'CONFIG_BLKIO_VHOST_VDPA_FD' } ],
43
'if': 'CONFIG_BLKIO' }
36
44
37
##
45
##
38
# @BlockdevOptions:
46
diff --git a/meson.build b/meson.build
39
diff --git a/block/copy-before-write.c b/block/copy-before-write.c
40
index XXXXXXX..XXXXXXX 100644
47
index XXXXXXX..XXXXXXX 100644
41
--- a/block/copy-before-write.c
48
--- a/meson.build
42
+++ b/block/copy-before-write.c
49
+++ b/meson.build
43
@@ -XXX,XX +XXX,XX @@
50
@@ -XXX,XX +XXX,XX @@ config_host_data.set('CONFIG_LZO', lzo.found())
44
51
config_host_data.set('CONFIG_MPATH', mpathpersist.found())
45
#include "block/copy-before-write.h"
52
config_host_data.set('CONFIG_MPATH_NEW_API', mpathpersist_new_api)
46
53
config_host_data.set('CONFIG_BLKIO', blkio.found())
47
+#include "qapi/qapi-visit-block-core.h"
54
+if blkio.found()
48
+
55
+ config_host_data.set('CONFIG_BLKIO_VHOST_VDPA_FD',
49
typedef struct BDRVCopyBeforeWriteState {
56
+ blkio.version().version_compare('>=1.3.0'))
50
BlockCopyState *bcs;
57
+endif
51
BdrvChild *target;
58
config_host_data.set('CONFIG_CURL', curl.found())
52
@@ -XXX,XX +XXX,XX @@ static void cbw_child_perm(BlockDriverState *bs, BdrvChild *c,
59
config_host_data.set('CONFIG_CURSES', curses.found())
53
}
60
config_host_data.set('CONFIG_GBM', gbm.found())
54
}
55
56
+static bool cbw_parse_bitmap_option(QDict *options, BdrvDirtyBitmap **bitmap,
57
+ Error **errp)
58
+{
59
+ QDict *bitmap_qdict = NULL;
60
+ BlockDirtyBitmap *bmp_param = NULL;
61
+ Visitor *v = NULL;
62
+ bool ret = false;
63
+
64
+ *bitmap = NULL;
65
+
66
+ qdict_extract_subqdict(options, &bitmap_qdict, "bitmap.");
67
+ if (!qdict_size(bitmap_qdict)) {
68
+ ret = true;
69
+ goto out;
70
+ }
71
+
72
+ v = qobject_input_visitor_new_flat_confused(bitmap_qdict, errp);
73
+ if (!v) {
74
+ goto out;
75
+ }
76
+
77
+ visit_type_BlockDirtyBitmap(v, NULL, &bmp_param, errp);
78
+ if (!bmp_param) {
79
+ goto out;
80
+ }
81
+
82
+ *bitmap = block_dirty_bitmap_lookup(bmp_param->node, bmp_param->name, NULL,
83
+ errp);
84
+ if (!*bitmap) {
85
+ goto out;
86
+ }
87
+
88
+ ret = true;
89
+
90
+out:
91
+ qapi_free_BlockDirtyBitmap(bmp_param);
92
+ visit_free(v);
93
+ qobject_unref(bitmap_qdict);
94
+
95
+ return ret;
96
+}
97
+
98
static int cbw_open(BlockDriverState *bs, QDict *options, int flags,
99
Error **errp)
100
{
101
BDRVCopyBeforeWriteState *s = bs->opaque;
102
+ BdrvDirtyBitmap *bitmap = NULL;
103
104
bs->file = bdrv_open_child(NULL, options, "file", bs, &child_of_bds,
105
BDRV_CHILD_FILTERED | BDRV_CHILD_PRIMARY,
106
@@ -XXX,XX +XXX,XX @@ static int cbw_open(BlockDriverState *bs, QDict *options, int flags,
107
return -EINVAL;
108
}
109
110
+ if (!cbw_parse_bitmap_option(options, &bitmap, errp)) {
111
+ return -EINVAL;
112
+ }
113
+
114
bs->total_sectors = bs->file->bs->total_sectors;
115
bs->supported_write_flags = BDRV_REQ_WRITE_UNCHANGED |
116
(BDRV_REQ_FUA & bs->file->bs->supported_write_flags);
117
@@ -XXX,XX +XXX,XX @@ static int cbw_open(BlockDriverState *bs, QDict *options, int flags,
118
((BDRV_REQ_FUA | BDRV_REQ_MAY_UNMAP | BDRV_REQ_NO_FALLBACK) &
119
bs->file->bs->supported_zero_flags);
120
121
- s->bcs = block_copy_state_new(bs->file, s->target, NULL, errp);
122
+ s->bcs = block_copy_state_new(bs->file, s->target, bitmap, errp);
123
if (!s->bcs) {
124
error_prepend(errp, "Cannot create block-copy-state: ");
125
return -EINVAL;
126
--
61
--
127
2.34.1
62
2.40.1
diff view generated by jsdifflib
Deleted patch
1
From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
2
1
3
Let's reuse convenient helper.
4
5
Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
6
Reviewed-by: Hanna Reitz <hreitz@redhat.com>
7
Message-Id: <20220303194349.2304213-8-vsementsov@virtuozzo.com>
8
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
9
---
10
block/reqlist.c | 3 ++-
11
1 file changed, 2 insertions(+), 1 deletion(-)
12
13
diff --git a/block/reqlist.c b/block/reqlist.c
14
index XXXXXXX..XXXXXXX 100644
15
--- a/block/reqlist.c
16
+++ b/block/reqlist.c
17
@@ -XXX,XX +XXX,XX @@
18
*/
19
20
#include "qemu/osdep.h"
21
+#include "qemu/range.h"
22
23
#include "block/reqlist.h"
24
25
@@ -XXX,XX +XXX,XX @@ BlockReq *reqlist_find_conflict(BlockReqList *reqs, int64_t offset,
26
BlockReq *r;
27
28
QLIST_FOREACH(r, reqs, list) {
29
- if (offset + bytes > r->offset && offset < r->offset + r->bytes) {
30
+ if (ranges_overlap(offset, bytes, r->offset, r->bytes)) {
31
return r;
32
}
33
}
34
--
35
2.34.1
diff view generated by jsdifflib
Deleted patch
1
From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
2
1
3
Add a convenient function similar with bdrv_block_status() to get
4
status of dirty bitmap.
5
6
Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
7
Reviewed-by: Hanna Reitz <hreitz@redhat.com>
8
Message-Id: <20220303194349.2304213-9-vsementsov@virtuozzo.com>
9
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
10
---
11
include/block/dirty-bitmap.h | 2 ++
12
include/qemu/hbitmap.h | 12 ++++++++++++
13
block/dirty-bitmap.c | 6 ++++++
14
util/hbitmap.c | 33 +++++++++++++++++++++++++++++++++
15
4 files changed, 53 insertions(+)
16
17
diff --git a/include/block/dirty-bitmap.h b/include/block/dirty-bitmap.h
18
index XXXXXXX..XXXXXXX 100644
19
--- a/include/block/dirty-bitmap.h
20
+++ b/include/block/dirty-bitmap.h
21
@@ -XXX,XX +XXX,XX @@ int64_t bdrv_dirty_bitmap_next_zero(BdrvDirtyBitmap *bitmap, int64_t offset,
22
bool bdrv_dirty_bitmap_next_dirty_area(BdrvDirtyBitmap *bitmap,
23
int64_t start, int64_t end, int64_t max_dirty_count,
24
int64_t *dirty_start, int64_t *dirty_count);
25
+bool bdrv_dirty_bitmap_status(BdrvDirtyBitmap *bitmap, int64_t offset,
26
+ int64_t bytes, int64_t *count);
27
BdrvDirtyBitmap *bdrv_reclaim_dirty_bitmap_locked(BdrvDirtyBitmap *bitmap,
28
Error **errp);
29
30
diff --git a/include/qemu/hbitmap.h b/include/qemu/hbitmap.h
31
index XXXXXXX..XXXXXXX 100644
32
--- a/include/qemu/hbitmap.h
33
+++ b/include/qemu/hbitmap.h
34
@@ -XXX,XX +XXX,XX @@ bool hbitmap_next_dirty_area(const HBitmap *hb, int64_t start, int64_t end,
35
int64_t max_dirty_count,
36
int64_t *dirty_start, int64_t *dirty_count);
37
38
+/*
39
+ * bdrv_dirty_bitmap_status:
40
+ * @hb: The HBitmap to operate on
41
+ * @start: The bit to start from
42
+ * @count: Number of bits to proceed
43
+ * @pnum: Out-parameter. How many bits has same value starting from @start
44
+ *
45
+ * Returns true if bitmap is dirty at @start, false otherwise.
46
+ */
47
+bool hbitmap_status(const HBitmap *hb, int64_t start, int64_t count,
48
+ int64_t *pnum);
49
+
50
/**
51
* hbitmap_iter_next:
52
* @hbi: HBitmapIter to operate on.
53
diff --git a/block/dirty-bitmap.c b/block/dirty-bitmap.c
54
index XXXXXXX..XXXXXXX 100644
55
--- a/block/dirty-bitmap.c
56
+++ b/block/dirty-bitmap.c
57
@@ -XXX,XX +XXX,XX @@ bool bdrv_dirty_bitmap_next_dirty_area(BdrvDirtyBitmap *bitmap,
58
dirty_start, dirty_count);
59
}
60
61
+bool bdrv_dirty_bitmap_status(BdrvDirtyBitmap *bitmap, int64_t offset,
62
+ int64_t bytes, int64_t *count)
63
+{
64
+ return hbitmap_status(bitmap->bitmap, offset, bytes, count);
65
+}
66
+
67
/**
68
* bdrv_merge_dirty_bitmap: merge src into dest.
69
* Ensures permissions on bitmaps are reasonable; use for public API.
70
diff --git a/util/hbitmap.c b/util/hbitmap.c
71
index XXXXXXX..XXXXXXX 100644
72
--- a/util/hbitmap.c
73
+++ b/util/hbitmap.c
74
@@ -XXX,XX +XXX,XX @@ bool hbitmap_next_dirty_area(const HBitmap *hb, int64_t start, int64_t end,
75
return true;
76
}
77
78
+bool hbitmap_status(const HBitmap *hb, int64_t start, int64_t count,
79
+ int64_t *pnum)
80
+{
81
+ int64_t next_dirty, next_zero;
82
+
83
+ assert(start >= 0);
84
+ assert(count > 0);
85
+ assert(start + count <= hb->orig_size);
86
+
87
+ next_dirty = hbitmap_next_dirty(hb, start, count);
88
+ if (next_dirty == -1) {
89
+ *pnum = count;
90
+ return false;
91
+ }
92
+
93
+ if (next_dirty > start) {
94
+ *pnum = next_dirty - start;
95
+ return false;
96
+ }
97
+
98
+ assert(next_dirty == start);
99
+
100
+ next_zero = hbitmap_next_zero(hb, start, count);
101
+ if (next_zero == -1) {
102
+ *pnum = count;
103
+ return true;
104
+ }
105
+
106
+ assert(next_zero > start);
107
+ *pnum = next_zero - start;
108
+ return false;
109
+}
110
+
111
bool hbitmap_empty(const HBitmap *hb)
112
{
113
return hb->count == 0;
114
--
115
2.34.1
diff view generated by jsdifflib
Deleted patch
1
From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
2
1
3
Add function to wait for all intersecting requests.
4
To be used in the further commit.
5
6
Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
7
Reviewed-by: Nikita Lapshin <nikita.lapshin@virtuozzo.com>
8
Reviewed-by: Hanna Reitz <hreitz@redhat.com>
9
Message-Id: <20220303194349.2304213-10-vsementsov@virtuozzo.com>
10
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
11
---
12
include/block/reqlist.h | 8 ++++++++
13
block/reqlist.c | 8 ++++++++
14
2 files changed, 16 insertions(+)
15
16
diff --git a/include/block/reqlist.h b/include/block/reqlist.h
17
index XXXXXXX..XXXXXXX 100644
18
--- a/include/block/reqlist.h
19
+++ b/include/block/reqlist.h
20
@@ -XXX,XX +XXX,XX @@ BlockReq *reqlist_find_conflict(BlockReqList *reqs, int64_t offset,
21
bool coroutine_fn reqlist_wait_one(BlockReqList *reqs, int64_t offset,
22
int64_t bytes, CoMutex *lock);
23
24
+/*
25
+ * Wait for all intersecting requests. It just calls reqlist_wait_one() in a
26
+ * loop, caller is responsible to stop producing new requests in this region
27
+ * in parallel, otherwise reqlist_wait_all() may never return.
28
+ */
29
+void coroutine_fn reqlist_wait_all(BlockReqList *reqs, int64_t offset,
30
+ int64_t bytes, CoMutex *lock);
31
+
32
/*
33
* Shrink request and wake all waiting coroutines (maybe some of them are not
34
* intersecting with shrunk request).
35
diff --git a/block/reqlist.c b/block/reqlist.c
36
index XXXXXXX..XXXXXXX 100644
37
--- a/block/reqlist.c
38
+++ b/block/reqlist.c
39
@@ -XXX,XX +XXX,XX @@ bool coroutine_fn reqlist_wait_one(BlockReqList *reqs, int64_t offset,
40
return true;
41
}
42
43
+void coroutine_fn reqlist_wait_all(BlockReqList *reqs, int64_t offset,
44
+ int64_t bytes, CoMutex *lock)
45
+{
46
+ while (reqlist_wait_one(reqs, offset, bytes, lock)) {
47
+ /* continue */
48
+ }
49
+}
50
+
51
void coroutine_fn reqlist_shrink_req(BlockReq *req, int64_t new_bytes)
52
{
53
if (new_bytes == req->bytes) {
54
--
55
2.34.1
diff view generated by jsdifflib
Deleted patch
1
From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
2
1
3
Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
4
Reviewed-by: Hanna Reitz <hreitz@redhat.com>
5
Message-Id: <20220303194349.2304213-14-vsementsov@virtuozzo.com>
6
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
7
---
8
tests/qemu-iotests/tests/image-fleecing | 64 +++++++++++++-----
9
tests/qemu-iotests/tests/image-fleecing.out | 74 ++++++++++++++++++++-
10
2 files changed, 119 insertions(+), 19 deletions(-)
11
12
diff --git a/tests/qemu-iotests/tests/image-fleecing b/tests/qemu-iotests/tests/image-fleecing
13
index XXXXXXX..XXXXXXX 100755
14
--- a/tests/qemu-iotests/tests/image-fleecing
15
+++ b/tests/qemu-iotests/tests/image-fleecing
16
@@ -XXX,XX +XXX,XX @@ remainder = [('0xd5', '0x108000', '32k'), # Right-end of partial-left [1]
17
('0xdc', '32M', '32k'), # Left-end of partial-right [2]
18
('0xcd', '0x3ff0000', '64k')] # patterns[3]
19
20
-def do_test(use_cbw, base_img_path, fleece_img_path, nbd_sock_path, vm):
21
+def do_test(use_cbw, use_snapshot_access_filter, base_img_path,
22
+ fleece_img_path, nbd_sock_path, vm):
23
log('--- Setting up images ---')
24
log('')
25
26
assert qemu_img('create', '-f', iotests.imgfmt, base_img_path, '64M') == 0
27
- assert qemu_img('create', '-f', 'qcow2', fleece_img_path, '64M') == 0
28
+ if use_snapshot_access_filter:
29
+ assert use_cbw
30
+ assert qemu_img('create', '-f', 'raw', fleece_img_path, '64M') == 0
31
+ else:
32
+ assert qemu_img('create', '-f', 'qcow2', fleece_img_path, '64M') == 0
33
34
for p in patterns:
35
qemu_io('-f', iotests.imgfmt,
36
@@ -XXX,XX +XXX,XX @@ def do_test(use_cbw, base_img_path, fleece_img_path, nbd_sock_path, vm):
37
log('')
38
39
40
- # create tmp_node backed by src_node
41
- log(vm.qmp('blockdev-add', {
42
- 'driver': 'qcow2',
43
- 'node-name': tmp_node,
44
- 'file': {
45
+ if use_snapshot_access_filter:
46
+ log(vm.qmp('blockdev-add', {
47
+ 'node-name': tmp_node,
48
'driver': 'file',
49
'filename': fleece_img_path,
50
- },
51
- 'backing': src_node,
52
- }))
53
+ }))
54
+ else:
55
+ # create tmp_node backed by src_node
56
+ log(vm.qmp('blockdev-add', {
57
+ 'driver': 'qcow2',
58
+ 'node-name': tmp_node,
59
+ 'file': {
60
+ 'driver': 'file',
61
+ 'filename': fleece_img_path,
62
+ },
63
+ 'backing': src_node,
64
+ }))
65
66
# Establish CBW from source to fleecing node
67
if use_cbw:
68
@@ -XXX,XX +XXX,XX @@ def do_test(use_cbw, base_img_path, fleece_img_path, nbd_sock_path, vm):
69
}))
70
71
log(vm.qmp('qom-set', path=qom_path, property='drive', value='fl-cbw'))
72
+
73
+ if use_snapshot_access_filter:
74
+ log(vm.qmp('blockdev-add', {
75
+ 'driver': 'snapshot-access',
76
+ 'node-name': 'fl-access',
77
+ 'file': 'fl-cbw',
78
+ }))
79
else:
80
log(vm.qmp('blockdev-backup',
81
job_id='fleecing',
82
@@ -XXX,XX +XXX,XX @@ def do_test(use_cbw, base_img_path, fleece_img_path, nbd_sock_path, vm):
83
target=tmp_node,
84
sync='none'))
85
86
+ export_node = 'fl-access' if use_snapshot_access_filter else tmp_node
87
+
88
log('')
89
log('--- Setting up NBD Export ---')
90
log('')
91
92
- nbd_uri = 'nbd+unix:///%s?socket=%s' % (tmp_node, nbd_sock_path)
93
+ nbd_uri = 'nbd+unix:///%s?socket=%s' % (export_node, nbd_sock_path)
94
log(vm.qmp('nbd-server-start',
95
{'addr': {'type': 'unix',
96
'data': {'path': nbd_sock_path}}}))
97
98
- log(vm.qmp('nbd-server-add', device=tmp_node))
99
+ log(vm.qmp('nbd-server-add', device=export_node))
100
101
log('')
102
log('--- Sanity Check ---')
103
@@ -XXX,XX +XXX,XX @@ def do_test(use_cbw, base_img_path, fleece_img_path, nbd_sock_path, vm):
104
log('--- Cleanup ---')
105
log('')
106
107
+ log(vm.qmp('nbd-server-stop'))
108
+
109
if use_cbw:
110
+ if use_snapshot_access_filter:
111
+ log(vm.qmp('blockdev-del', node_name='fl-access'))
112
log(vm.qmp('qom-set', path=qom_path, property='drive', value=src_node))
113
log(vm.qmp('blockdev-del', node_name='fl-cbw'))
114
else:
115
@@ -XXX,XX +XXX,XX @@ def do_test(use_cbw, base_img_path, fleece_img_path, nbd_sock_path, vm):
116
assert e is not None
117
log(e, filters=[iotests.filter_qmp_event])
118
119
- log(vm.qmp('nbd-server-stop'))
120
log(vm.qmp('blockdev-del', node_name=tmp_node))
121
vm.shutdown()
122
123
@@ -XXX,XX +XXX,XX @@ def do_test(use_cbw, base_img_path, fleece_img_path, nbd_sock_path, vm):
124
log('Done')
125
126
127
-def test(use_cbw):
128
+def test(use_cbw, use_snapshot_access_filter):
129
with iotests.FilePath('base.img') as base_img_path, \
130
iotests.FilePath('fleece.img') as fleece_img_path, \
131
iotests.FilePath('nbd.sock',
132
base_dir=iotests.sock_dir) as nbd_sock_path, \
133
iotests.VM() as vm:
134
- do_test(use_cbw, base_img_path, fleece_img_path, nbd_sock_path, vm)
135
+ do_test(use_cbw, use_snapshot_access_filter, base_img_path,
136
+ fleece_img_path, nbd_sock_path, vm)
137
138
139
log('=== Test backup(sync=none) based fleecing ===\n')
140
-test(False)
141
+test(False, False)
142
+
143
+log('=== Test cbw-filter based fleecing ===\n')
144
+test(True, False)
145
146
-log('=== Test filter based fleecing ===\n')
147
-test(True)
148
+log('=== Test fleecing-format based fleecing ===\n')
149
+test(True, True)
150
diff --git a/tests/qemu-iotests/tests/image-fleecing.out b/tests/qemu-iotests/tests/image-fleecing.out
151
index XXXXXXX..XXXXXXX 100644
152
--- a/tests/qemu-iotests/tests/image-fleecing.out
153
+++ b/tests/qemu-iotests/tests/image-fleecing.out
154
@@ -XXX,XX +XXX,XX @@ read -P0 0x3fe0000 64k
155
156
--- Cleanup ---
157
158
+{"return": {}}
159
{"return": {}}
160
{"data": {"device": "fleecing", "len": 67108864, "offset": 393216, "speed": 0, "type": "backup"}, "event": "BLOCK_JOB_CANCELLED", "timestamp": {"microseconds": "USECS", "seconds": "SECS"}}
161
{"return": {}}
162
+
163
+--- Confirming writes ---
164
+
165
+read -P0xab 0 64k
166
+read -P0xad 0x00f8000 64k
167
+read -P0x1d 0x2008000 64k
168
+read -P0xea 0x3fe0000 64k
169
+read -P0xd5 0x108000 32k
170
+read -P0xdc 32M 32k
171
+read -P0xcd 0x3ff0000 64k
172
+
173
+Done
174
+=== Test cbw-filter based fleecing ===
175
+
176
+--- Setting up images ---
177
+
178
+Done
179
+
180
+--- Launching VM ---
181
+
182
+Done
183
+
184
+--- Setting up Fleecing Graph ---
185
+
186
+{"return": {}}
187
+{"return": {}}
188
+{"return": {}}
189
+
190
+--- Setting up NBD Export ---
191
+
192
+{"return": {}}
193
+{"return": {}}
194
+
195
+--- Sanity Check ---
196
+
197
+read -P0x5d 0 64k
198
+read -P0xd5 1M 64k
199
+read -P0xdc 32M 64k
200
+read -P0xcd 0x3ff0000 64k
201
+read -P0 0x00f8000 32k
202
+read -P0 0x2010000 32k
203
+read -P0 0x3fe0000 64k
204
+
205
+--- Testing COW ---
206
+
207
+write -P0xab 0 64k
208
+{"return": ""}
209
+write -P0xad 0x00f8000 64k
210
+{"return": ""}
211
+write -P0x1d 0x2008000 64k
212
+{"return": ""}
213
+write -P0xea 0x3fe0000 64k
214
+{"return": ""}
215
+
216
+--- Verifying Data ---
217
+
218
+read -P0x5d 0 64k
219
+read -P0xd5 1M 64k
220
+read -P0xdc 32M 64k
221
+read -P0xcd 0x3ff0000 64k
222
+read -P0 0x00f8000 32k
223
+read -P0 0x2010000 32k
224
+read -P0 0x3fe0000 64k
225
+
226
+--- Cleanup ---
227
+
228
+{"return": {}}
229
+{"return": {}}
230
+{"return": {}}
231
{"return": {}}
232
233
--- Confirming writes ---
234
@@ -XXX,XX +XXX,XX @@ read -P0xdc 32M 32k
235
read -P0xcd 0x3ff0000 64k
236
237
Done
238
-=== Test filter based fleecing ===
239
+=== Test fleecing-format based fleecing ===
240
241
--- Setting up images ---
242
243
@@ -XXX,XX +XXX,XX @@ Done
244
{"return": {}}
245
{"return": {}}
246
{"return": {}}
247
+{"return": {}}
248
249
--- Setting up NBD Export ---
250
251
@@ -XXX,XX +XXX,XX @@ read -P0 0x3fe0000 64k
252
{"return": {}}
253
{"return": {}}
254
{"return": {}}
255
+{"return": {}}
256
257
--- Confirming writes ---
258
259
--
260
2.34.1
diff view generated by jsdifflib
Deleted patch
1
From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
2
1
3
Add helper that returns both status and output, to be used in the
4
following commit
5
6
Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
7
Message-Id: <20220303194349.2304213-15-vsementsov@virtuozzo.com>
8
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
9
---
10
tests/qemu-iotests/iotests.py | 3 +++
11
1 file changed, 3 insertions(+)
12
13
diff --git a/tests/qemu-iotests/iotests.py b/tests/qemu-iotests/iotests.py
14
index XXXXXXX..XXXXXXX 100644
15
--- a/tests/qemu-iotests/iotests.py
16
+++ b/tests/qemu-iotests/iotests.py
17
@@ -XXX,XX +XXX,XX @@ def qemu_io(*args):
18
'''Run qemu-io and return the stdout data'''
19
return qemu_tool_pipe_and_status('qemu-io', qemu_io_wrap_args(args))[0]
20
21
+def qemu_io_pipe_and_status(*args):
22
+ return qemu_tool_pipe_and_status('qemu-io', qemu_io_wrap_args(args))
23
+
24
def qemu_io_log(*args):
25
result = qemu_io(*args)
26
log(result, filters=[filter_testfiles, filter_qemu_io])
27
--
28
2.34.1
diff view generated by jsdifflib
Deleted patch
1
From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
2
1
3
Note that reads zero areas (not dirty in the bitmap) fails, that's
4
correct.
5
6
Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
7
Message-Id: <20220303194349.2304213-16-vsementsov@virtuozzo.com>
8
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
9
---
10
tests/qemu-iotests/tests/image-fleecing | 38 +++++++---
11
tests/qemu-iotests/tests/image-fleecing.out | 84 +++++++++++++++++++++
12
2 files changed, 113 insertions(+), 9 deletions(-)
13
14
diff --git a/tests/qemu-iotests/tests/image-fleecing b/tests/qemu-iotests/tests/image-fleecing
15
index XXXXXXX..XXXXXXX 100755
16
--- a/tests/qemu-iotests/tests/image-fleecing
17
+++ b/tests/qemu-iotests/tests/image-fleecing
18
@@ -XXX,XX +XXX,XX @@
19
# Creator/Owner: John Snow <jsnow@redhat.com>
20
21
import iotests
22
-from iotests import log, qemu_img, qemu_io, qemu_io_silent
23
+from iotests import log, qemu_img, qemu_io, qemu_io_silent, \
24
+ qemu_io_pipe_and_status
25
26
iotests.script_initialize(
27
- supported_fmts=['qcow2', 'qcow', 'qed', 'vmdk', 'vhdx', 'raw'],
28
+ supported_fmts=['qcow2'],
29
supported_platforms=['linux'],
30
required_fmts=['copy-before-write'],
31
+ unsupported_imgopts=['compat']
32
)
33
34
patterns = [('0x5d', '0', '64k'),
35
@@ -XXX,XX +XXX,XX @@ remainder = [('0xd5', '0x108000', '32k'), # Right-end of partial-left [1]
36
('0xcd', '0x3ff0000', '64k')] # patterns[3]
37
38
def do_test(use_cbw, use_snapshot_access_filter, base_img_path,
39
- fleece_img_path, nbd_sock_path, vm):
40
+ fleece_img_path, nbd_sock_path, vm,
41
+ bitmap=False):
42
log('--- Setting up images ---')
43
log('')
44
45
assert qemu_img('create', '-f', iotests.imgfmt, base_img_path, '64M') == 0
46
+ if bitmap:
47
+ assert qemu_img('bitmap', '--add', base_img_path, 'bitmap0') == 0
48
+
49
if use_snapshot_access_filter:
50
assert use_cbw
51
assert qemu_img('create', '-f', 'raw', fleece_img_path, '64M') == 0
52
@@ -XXX,XX +XXX,XX @@ def do_test(use_cbw, use_snapshot_access_filter, base_img_path,
53
54
# Establish CBW from source to fleecing node
55
if use_cbw:
56
- log(vm.qmp('blockdev-add', {
57
+ fl_cbw = {
58
'driver': 'copy-before-write',
59
'node-name': 'fl-cbw',
60
'file': src_node,
61
'target': tmp_node
62
- }))
63
+ }
64
+
65
+ if bitmap:
66
+ fl_cbw['bitmap'] = {'node': src_node, 'name': 'bitmap0'}
67
+
68
+ log(vm.qmp('blockdev-add', fl_cbw))
69
70
log(vm.qmp('qom-set', path=qom_path, property='drive', value='fl-cbw'))
71
72
@@ -XXX,XX +XXX,XX @@ def do_test(use_cbw, use_snapshot_access_filter, base_img_path,
73
for p in patterns + zeroes:
74
cmd = 'read -P%s %s %s' % p
75
log(cmd)
76
- assert qemu_io_silent('-r', '-f', 'raw', '-c', cmd, nbd_uri) == 0
77
+ out, ret = qemu_io_pipe_and_status('-r', '-f', 'raw', '-c', cmd,
78
+ nbd_uri)
79
+ if ret != 0:
80
+ print(out)
81
82
log('')
83
log('--- Testing COW ---')
84
@@ -XXX,XX +XXX,XX @@ def do_test(use_cbw, use_snapshot_access_filter, base_img_path,
85
for p in patterns + zeroes:
86
cmd = 'read -P%s %s %s' % p
87
log(cmd)
88
- assert qemu_io_silent('-r', '-f', 'raw', '-c', cmd, nbd_uri) == 0
89
+ out, ret = qemu_io_pipe_and_status('-r', '-f', 'raw', '-c', cmd,
90
+ nbd_uri)
91
+ if ret != 0:
92
+ print(out)
93
94
log('')
95
log('--- Cleanup ---')
96
@@ -XXX,XX +XXX,XX @@ def do_test(use_cbw, use_snapshot_access_filter, base_img_path,
97
log('Done')
98
99
100
-def test(use_cbw, use_snapshot_access_filter):
101
+def test(use_cbw, use_snapshot_access_filter, bitmap=False):
102
with iotests.FilePath('base.img') as base_img_path, \
103
iotests.FilePath('fleece.img') as fleece_img_path, \
104
iotests.FilePath('nbd.sock',
105
base_dir=iotests.sock_dir) as nbd_sock_path, \
106
iotests.VM() as vm:
107
do_test(use_cbw, use_snapshot_access_filter, base_img_path,
108
- fleece_img_path, nbd_sock_path, vm)
109
+ fleece_img_path, nbd_sock_path, vm, bitmap=bitmap)
110
111
112
log('=== Test backup(sync=none) based fleecing ===\n')
113
@@ -XXX,XX +XXX,XX @@ test(True, False)
114
115
log('=== Test fleecing-format based fleecing ===\n')
116
test(True, True)
117
+
118
+log('=== Test fleecing-format based fleecing with bitmap ===\n')
119
+test(True, True, bitmap=True)
120
diff --git a/tests/qemu-iotests/tests/image-fleecing.out b/tests/qemu-iotests/tests/image-fleecing.out
121
index XXXXXXX..XXXXXXX 100644
122
--- a/tests/qemu-iotests/tests/image-fleecing.out
123
+++ b/tests/qemu-iotests/tests/image-fleecing.out
124
@@ -XXX,XX +XXX,XX @@ read -P0 0x00f8000 32k
125
read -P0 0x2010000 32k
126
read -P0 0x3fe0000 64k
127
128
+--- Cleanup ---
129
+
130
+{"return": {}}
131
+{"return": {}}
132
+{"return": {}}
133
+{"return": {}}
134
+{"return": {}}
135
+
136
+--- Confirming writes ---
137
+
138
+read -P0xab 0 64k
139
+read -P0xad 0x00f8000 64k
140
+read -P0x1d 0x2008000 64k
141
+read -P0xea 0x3fe0000 64k
142
+read -P0xd5 0x108000 32k
143
+read -P0xdc 32M 32k
144
+read -P0xcd 0x3ff0000 64k
145
+
146
+Done
147
+=== Test fleecing-format based fleecing with bitmap ===
148
+
149
+--- Setting up images ---
150
+
151
+Done
152
+
153
+--- Launching VM ---
154
+
155
+Done
156
+
157
+--- Setting up Fleecing Graph ---
158
+
159
+{"return": {}}
160
+{"return": {}}
161
+{"return": {}}
162
+{"return": {}}
163
+
164
+--- Setting up NBD Export ---
165
+
166
+{"return": {}}
167
+{"return": {}}
168
+
169
+--- Sanity Check ---
170
+
171
+read -P0x5d 0 64k
172
+read -P0xd5 1M 64k
173
+read -P0xdc 32M 64k
174
+read -P0xcd 0x3ff0000 64k
175
+read -P0 0x00f8000 32k
176
+read failed: Invalid argument
177
+
178
+read -P0 0x2010000 32k
179
+read failed: Invalid argument
180
+
181
+read -P0 0x3fe0000 64k
182
+read failed: Invalid argument
183
+
184
+
185
+--- Testing COW ---
186
+
187
+write -P0xab 0 64k
188
+{"return": ""}
189
+write -P0xad 0x00f8000 64k
190
+{"return": ""}
191
+write -P0x1d 0x2008000 64k
192
+{"return": ""}
193
+write -P0xea 0x3fe0000 64k
194
+{"return": ""}
195
+
196
+--- Verifying Data ---
197
+
198
+read -P0x5d 0 64k
199
+read -P0xd5 1M 64k
200
+read -P0xdc 32M 64k
201
+read -P0xcd 0x3ff0000 64k
202
+read -P0 0x00f8000 32k
203
+read failed: Invalid argument
204
+
205
+read -P0 0x2010000 32k
206
+read failed: Invalid argument
207
+
208
+read -P0 0x3fe0000 64k
209
+read failed: Invalid argument
210
+
211
+
212
--- Cleanup ---
213
214
{"return": {}}
215
--
216
2.34.1
diff view generated by jsdifflib
Deleted patch
1
From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
2
1
3
Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com>
4
Message-Id: <20220303194349.2304213-17-vsementsov@virtuozzo.com>
5
Signed-off-by: Hanna Reitz <hreitz@redhat.com>
6
---
7
tests/qemu-iotests/tests/image-fleecing | 125 +++++++++++++++-----
8
tests/qemu-iotests/tests/image-fleecing.out | 63 ++++++++++
9
2 files changed, 156 insertions(+), 32 deletions(-)
10
11
diff --git a/tests/qemu-iotests/tests/image-fleecing b/tests/qemu-iotests/tests/image-fleecing
12
index XXXXXXX..XXXXXXX 100755
13
--- a/tests/qemu-iotests/tests/image-fleecing
14
+++ b/tests/qemu-iotests/tests/image-fleecing
15
@@ -XXX,XX +XXX,XX @@ remainder = [('0xd5', '0x108000', '32k'), # Right-end of partial-left [1]
16
('0xdc', '32M', '32k'), # Left-end of partial-right [2]
17
('0xcd', '0x3ff0000', '64k')] # patterns[3]
18
19
-def do_test(use_cbw, use_snapshot_access_filter, base_img_path,
20
- fleece_img_path, nbd_sock_path, vm,
21
+def do_test(vm, use_cbw, use_snapshot_access_filter, base_img_path,
22
+ fleece_img_path, nbd_sock_path=None,
23
+ target_img_path=None,
24
bitmap=False):
25
+ push_backup = target_img_path is not None
26
+ assert (nbd_sock_path is not None) != push_backup
27
+ if push_backup:
28
+ assert use_cbw
29
+
30
log('--- Setting up images ---')
31
log('')
32
33
@@ -XXX,XX +XXX,XX @@ def do_test(use_cbw, use_snapshot_access_filter, base_img_path,
34
else:
35
assert qemu_img('create', '-f', 'qcow2', fleece_img_path, '64M') == 0
36
37
+ if push_backup:
38
+ assert qemu_img('create', '-f', 'qcow2', target_img_path, '64M') == 0
39
+
40
for p in patterns:
41
qemu_io('-f', iotests.imgfmt,
42
'-c', 'write -P%s %s %s' % p, base_img_path)
43
@@ -XXX,XX +XXX,XX @@ def do_test(use_cbw, use_snapshot_access_filter, base_img_path,
44
45
export_node = 'fl-access' if use_snapshot_access_filter else tmp_node
46
47
- log('')
48
- log('--- Setting up NBD Export ---')
49
- log('')
50
+ if push_backup:
51
+ log('')
52
+ log('--- Starting actual backup ---')
53
+ log('')
54
55
- nbd_uri = 'nbd+unix:///%s?socket=%s' % (export_node, nbd_sock_path)
56
- log(vm.qmp('nbd-server-start',
57
- {'addr': {'type': 'unix',
58
- 'data': {'path': nbd_sock_path}}}))
59
+ log(vm.qmp('blockdev-add', **{
60
+ 'driver': iotests.imgfmt,
61
+ 'node-name': 'target',
62
+ 'file': {
63
+ 'driver': 'file',
64
+ 'filename': target_img_path
65
+ }
66
+ }))
67
+ log(vm.qmp('blockdev-backup', device=export_node,
68
+ sync='full', target='target',
69
+ job_id='push-backup', speed=1))
70
+ else:
71
+ log('')
72
+ log('--- Setting up NBD Export ---')
73
+ log('')
74
75
- log(vm.qmp('nbd-server-add', device=export_node))
76
+ nbd_uri = 'nbd+unix:///%s?socket=%s' % (export_node, nbd_sock_path)
77
+ log(vm.qmp('nbd-server-start',
78
+ {'addr': { 'type': 'unix',
79
+ 'data': { 'path': nbd_sock_path } } }))
80
81
- log('')
82
- log('--- Sanity Check ---')
83
- log('')
84
+ log(vm.qmp('nbd-server-add', device=export_node))
85
86
- for p in patterns + zeroes:
87
- cmd = 'read -P%s %s %s' % p
88
- log(cmd)
89
- out, ret = qemu_io_pipe_and_status('-r', '-f', 'raw', '-c', cmd,
90
- nbd_uri)
91
- if ret != 0:
92
- print(out)
93
+ log('')
94
+ log('--- Sanity Check ---')
95
+ log('')
96
+
97
+ for p in patterns + zeroes:
98
+ cmd = 'read -P%s %s %s' % p
99
+ log(cmd)
100
+ out, ret = qemu_io_pipe_and_status('-r', '-f', 'raw', '-c', cmd,
101
+ nbd_uri)
102
+ if ret != 0:
103
+ print(out)
104
105
log('')
106
log('--- Testing COW ---')
107
@@ -XXX,XX +XXX,XX @@ def do_test(use_cbw, use_snapshot_access_filter, base_img_path,
108
log(cmd)
109
log(vm.hmp_qemu_io(qom_path, cmd, qdev=True))
110
111
+ if push_backup:
112
+ # Check that previous operations were done during backup, not after
113
+ # If backup is already finished, it's possible that it was finished
114
+ # even before hmp qemu_io write, and we didn't actually test
115
+ # copy-before-write operation. This should not happen, as we use
116
+ # speed=1. But worth checking.
117
+ result = vm.qmp('query-block-jobs')
118
+ assert len(result['return']) == 1
119
+
120
+ result = vm.qmp('block-job-set-speed', device='push-backup', speed=0)
121
+ assert result == {'return': {}}
122
+
123
+ log(vm.event_wait(name='BLOCK_JOB_COMPLETED',
124
+ match={'data': {'device': 'push-backup'}}),
125
+ filters=[iotests.filter_qmp_event])
126
+ log(vm.qmp('blockdev-del', node_name='target'))
127
+
128
log('')
129
log('--- Verifying Data ---')
130
log('')
131
@@ -XXX,XX +XXX,XX @@ def do_test(use_cbw, use_snapshot_access_filter, base_img_path,
132
for p in patterns + zeroes:
133
cmd = 'read -P%s %s %s' % p
134
log(cmd)
135
- out, ret = qemu_io_pipe_and_status('-r', '-f', 'raw', '-c', cmd,
136
- nbd_uri)
137
+ args = ['-r', '-c', cmd]
138
+ if push_backup:
139
+ args += [target_img_path]
140
+ else:
141
+ args += ['-f', 'raw', nbd_uri]
142
+ out, ret = qemu_io_pipe_and_status(*args)
143
if ret != 0:
144
print(out)
145
146
@@ -XXX,XX +XXX,XX @@ def do_test(use_cbw, use_snapshot_access_filter, base_img_path,
147
log('--- Cleanup ---')
148
log('')
149
150
- log(vm.qmp('nbd-server-stop'))
151
+ if not push_backup:
152
+ log(vm.qmp('nbd-server-stop'))
153
154
if use_cbw:
155
if use_snapshot_access_filter:
156
@@ -XXX,XX +XXX,XX @@ def do_test(use_cbw, use_snapshot_access_filter, base_img_path,
157
log('Done')
158
159
160
-def test(use_cbw, use_snapshot_access_filter, bitmap=False):
161
+def test(use_cbw, use_snapshot_access_filter,
162
+ nbd_sock_path=None, target_img_path=None, bitmap=False):
163
with iotests.FilePath('base.img') as base_img_path, \
164
iotests.FilePath('fleece.img') as fleece_img_path, \
165
- iotests.FilePath('nbd.sock',
166
- base_dir=iotests.sock_dir) as nbd_sock_path, \
167
iotests.VM() as vm:
168
- do_test(use_cbw, use_snapshot_access_filter, base_img_path,
169
- fleece_img_path, nbd_sock_path, vm, bitmap=bitmap)
170
+ do_test(vm, use_cbw, use_snapshot_access_filter, base_img_path,
171
+ fleece_img_path, nbd_sock_path, target_img_path,
172
+ bitmap=bitmap)
173
+
174
+def test_pull(use_cbw, use_snapshot_access_filter, bitmap=False):
175
+ with iotests.FilePath('nbd.sock',
176
+ base_dir=iotests.sock_dir) as nbd_sock_path:
177
+ test(use_cbw, use_snapshot_access_filter, nbd_sock_path, None,
178
+ bitmap=bitmap)
179
+
180
+def test_push():
181
+ with iotests.FilePath('target.img') as target_img_path:
182
+ test(True, True, None, target_img_path)
183
184
185
log('=== Test backup(sync=none) based fleecing ===\n')
186
-test(False, False)
187
+test_pull(False, False)
188
189
log('=== Test cbw-filter based fleecing ===\n')
190
-test(True, False)
191
+test_pull(True, False)
192
193
log('=== Test fleecing-format based fleecing ===\n')
194
-test(True, True)
195
+test_pull(True, True)
196
197
log('=== Test fleecing-format based fleecing with bitmap ===\n')
198
-test(True, True, bitmap=True)
199
+test_pull(True, True, bitmap=True)
200
+
201
+log('=== Test push backup with fleecing ===\n')
202
+test_push()
203
diff --git a/tests/qemu-iotests/tests/image-fleecing.out b/tests/qemu-iotests/tests/image-fleecing.out
204
index XXXXXXX..XXXXXXX 100644
205
--- a/tests/qemu-iotests/tests/image-fleecing.out
206
+++ b/tests/qemu-iotests/tests/image-fleecing.out
207
@@ -XXX,XX +XXX,XX @@ read -P0xdc 32M 32k
208
read -P0xcd 0x3ff0000 64k
209
210
Done
211
+=== Test push backup with fleecing ===
212
+
213
+--- Setting up images ---
214
+
215
+Done
216
+
217
+--- Launching VM ---
218
+
219
+Done
220
+
221
+--- Setting up Fleecing Graph ---
222
+
223
+{"return": {}}
224
+{"return": {}}
225
+{"return": {}}
226
+{"return": {}}
227
+
228
+--- Starting actual backup ---
229
+
230
+{"return": {}}
231
+{"return": {}}
232
+
233
+--- Testing COW ---
234
+
235
+write -P0xab 0 64k
236
+{"return": ""}
237
+write -P0xad 0x00f8000 64k
238
+{"return": ""}
239
+write -P0x1d 0x2008000 64k
240
+{"return": ""}
241
+write -P0xea 0x3fe0000 64k
242
+{"return": ""}
243
+{"data": {"device": "push-backup", "len": 67108864, "offset": 67108864, "speed": 0, "type": "backup"}, "event": "BLOCK_JOB_COMPLETED", "timestamp": {"microseconds": "USECS", "seconds": "SECS"}}
244
+{"return": {}}
245
+
246
+--- Verifying Data ---
247
+
248
+read -P0x5d 0 64k
249
+read -P0xd5 1M 64k
250
+read -P0xdc 32M 64k
251
+read -P0xcd 0x3ff0000 64k
252
+read -P0 0x00f8000 32k
253
+read -P0 0x2010000 32k
254
+read -P0 0x3fe0000 64k
255
+
256
+--- Cleanup ---
257
+
258
+{"return": {}}
259
+{"return": {}}
260
+{"return": {}}
261
+{"return": {}}
262
+
263
+--- Confirming writes ---
264
+
265
+read -P0xab 0 64k
266
+read -P0xad 0x00f8000 64k
267
+read -P0x1d 0x2008000 64k
268
+read -P0xea 0x3fe0000 64k
269
+read -P0xd5 0x108000 32k
270
+read -P0xdc 32M 32k
271
+read -P0xcd 0x3ff0000 64k
272
+
273
+Done
274
--
275
2.34.1
diff view generated by jsdifflib