1 | The following changes since commit 848a6caa88b9f082c89c9b41afa975761262981d: | 1 | The following changes since commit ac793156f650ae2d77834932d72224175ee69086: |
---|---|---|---|
2 | 2 | ||
3 | Merge tag 'migration-20230602-pull-request' of https://gitlab.com/juan.quintela/qemu into staging (2023-06-02 17:33:29 -0700) | 3 | Merge remote-tracking branch 'remotes/pmaydell/tags/pull-target-arm-20201020-1' into staging (2020-10-20 21:11:35 +0100) |
4 | 4 | ||
5 | are available in the Git repository at: | 5 | are available in the Git repository at: |
6 | 6 | ||
7 | https://gitlab.com/hreitz/qemu.git tags/pull-block-2023-06-05 | 7 | https://gitlab.com/stefanha/qemu.git tags/block-pull-request |
8 | 8 | ||
9 | for you to fetch changes up to 42a2890a76f4783cd1c212f27856edcf2b5e8a75: | 9 | for you to fetch changes up to 32a3fd65e7e3551337fd26bfc0e2f899d70c028c: |
10 | 10 | ||
11 | qcow2: add discard-no-unref option (2023-06-05 13:15:42 +0200) | 11 | iotests: add commit top->base cases to 274 (2020-10-22 09:55:39 +0100) |
12 | 12 | ||
13 | ---------------------------------------------------------------- | 13 | ---------------------------------------------------------------- |
14 | Block patches | 14 | Pull request |
15 | 15 | ||
16 | - Fix padding of unaligned vectored requests to match the host alignment | 16 | v2: |
17 | for vectors with 1023 or 1024 buffers | 17 | * Fix format string issues on 32-bit hosts [Peter] |
18 | - Refactor and fix bugs in parallels's image check functionality | 18 | * Fix qemu-nbd.c CONFIG_POSIX ifdef issue [Eric] |
19 | - Add an option to the qcow2 driver to retain (qcow2-level) allocations | 19 | * Fix missing eventfd.h header on macOS [Peter] |
20 | on discard requests from the guest (while still forwarding the discard | 20 | * Drop unreliable vhost-user-blk test (will send a new patch when ready) [Peter] |
21 | to the lower level and marking the range as zero) | 21 | |
22 | This pull request contains the vhost-user-blk server by Coiby Xu along with my | ||
23 | additions, block/nvme.c alignment and hardware error statistics by Philippe | ||
24 | Mathieu-Daudé, and bdrv_co_block_status_above() fixes by Vladimir | ||
25 | Sementsov-Ogievskiy. | ||
22 | 26 | ||
23 | ---------------------------------------------------------------- | 27 | ---------------------------------------------------------------- |
24 | Alexander Ivanov (12): | ||
25 | parallels: Out of image offset in BAT leads to image inflation | ||
26 | parallels: Fix high_off calculation in parallels_co_check() | ||
27 | parallels: Fix image_end_offset and data_end after out-of-image check | ||
28 | parallels: create parallels_set_bat_entry_helper() to assign BAT value | ||
29 | parallels: Use generic infrastructure for BAT writing in | ||
30 | parallels_co_check() | ||
31 | parallels: Move check of unclean image to a separate function | ||
32 | parallels: Move check of cluster outside image to a separate function | ||
33 | parallels: Fix statistics calculation | ||
34 | parallels: Move check of leaks to a separate function | ||
35 | parallels: Move statistic collection to a separate function | ||
36 | parallels: Replace qemu_co_mutex_lock by WITH_QEMU_LOCK_GUARD | ||
37 | parallels: Incorrect condition in out-of-image check | ||
38 | 28 | ||
39 | Hanna Czenczek (4): | 29 | Coiby Xu (6): |
40 | util/iov: Make qiov_slice() public | 30 | libvhost-user: Allow vu_message_read to be replaced |
41 | block: Collapse padded I/O vecs exceeding IOV_MAX | 31 | libvhost-user: remove watch for kick_fd when de-initialize vu-dev |
42 | util/iov: Remove qemu_iovec_init_extended() | 32 | util/vhost-user-server: generic vhost user server |
43 | iotests/iov-padding: New test | 33 | block: move logical block size check function to a common utility |
34 | function | ||
35 | block/export: vhost-user block device backend server | ||
36 | MAINTAINERS: Add vhost-user block device backend server maintainer | ||
44 | 37 | ||
45 | Jean-Louis Dupond (1): | 38 | Philippe Mathieu-Daudé (1): |
46 | qcow2: add discard-no-unref option | 39 | block/nvme: Add driver statistics for access alignment and hw errors |
47 | 40 | ||
48 | qapi/block-core.json | 12 ++ | 41 | Stefan Hajnoczi (16): |
49 | block/qcow2.h | 3 + | 42 | util/vhost-user-server: s/fileds/fields/ typo fix |
50 | include/qemu/iov.h | 8 +- | 43 | util/vhost-user-server: drop unnecessary QOM cast |
51 | block/io.c | 166 ++++++++++++++++++-- | 44 | util/vhost-user-server: drop unnecessary watch deletion |
52 | block/parallels.c | 190 ++++++++++++++++------- | 45 | block/export: consolidate request structs into VuBlockReq |
53 | block/qcow2-cluster.c | 32 +++- | 46 | util/vhost-user-server: drop unused DevicePanicNotifier |
54 | block/qcow2.c | 18 +++ | 47 | util/vhost-user-server: fix memory leak in vu_message_read() |
55 | util/iov.c | 89 ++--------- | 48 | util/vhost-user-server: check EOF when reading payload |
56 | qemu-options.hx | 12 ++ | 49 | util/vhost-user-server: rework vu_client_trip() coroutine lifecycle |
57 | tests/qemu-iotests/tests/iov-padding | 85 ++++++++++ | 50 | block/export: report flush errors |
58 | tests/qemu-iotests/tests/iov-padding.out | 59 +++++++ | 51 | block/export: convert vhost-user-blk server to block export API |
59 | 11 files changed, 523 insertions(+), 151 deletions(-) | 52 | util/vhost-user-server: move header to include/ |
60 | create mode 100755 tests/qemu-iotests/tests/iov-padding | 53 | util/vhost-user-server: use static library in meson.build |
61 | create mode 100644 tests/qemu-iotests/tests/iov-padding.out | 54 | qemu-storage-daemon: avoid compiling blockdev_ss twice |
55 | block: move block exports to libblockdev | ||
56 | block/export: add iothread and fixed-iothread options | ||
57 | block/export: add vhost-user-blk multi-queue support | ||
58 | |||
59 | Vladimir Sementsov-Ogievskiy (5): | ||
60 | block/io: fix bdrv_co_block_status_above | ||
61 | block/io: bdrv_common_block_status_above: support include_base | ||
62 | block/io: bdrv_common_block_status_above: support bs == base | ||
63 | block/io: fix bdrv_is_allocated_above | ||
64 | iotests: add commit top->base cases to 274 | ||
65 | |||
66 | MAINTAINERS | 9 + | ||
67 | qapi/block-core.json | 24 +- | ||
68 | qapi/block-export.json | 36 +- | ||
69 | block/coroutines.h | 2 + | ||
70 | block/export/vhost-user-blk-server.h | 19 + | ||
71 | contrib/libvhost-user/libvhost-user.h | 21 + | ||
72 | include/qemu/vhost-user-server.h | 65 +++ | ||
73 | util/block-helpers.h | 19 + | ||
74 | block/export/export.c | 37 +- | ||
75 | block/export/vhost-user-blk-server.c | 431 ++++++++++++++++++++ | ||
76 | block/io.c | 132 +++--- | ||
77 | block/nvme.c | 27 ++ | ||
78 | block/qcow2.c | 16 +- | ||
79 | contrib/libvhost-user/libvhost-user-glib.c | 2 +- | ||
80 | contrib/libvhost-user/libvhost-user.c | 15 +- | ||
81 | hw/core/qdev-properties-system.c | 31 +- | ||
82 | nbd/server.c | 2 - | ||
83 | qemu-nbd.c | 21 +- | ||
84 | softmmu/vl.c | 4 + | ||
85 | stubs/blk-exp-close-all.c | 7 + | ||
86 | tests/vhost-user-bridge.c | 2 + | ||
87 | tools/virtiofsd/fuse_virtio.c | 4 +- | ||
88 | util/block-helpers.c | 46 +++ | ||
89 | util/vhost-user-server.c | 446 +++++++++++++++++++++ | ||
90 | block/export/meson.build | 3 +- | ||
91 | contrib/libvhost-user/meson.build | 1 + | ||
92 | meson.build | 22 +- | ||
93 | nbd/meson.build | 2 + | ||
94 | storage-daemon/meson.build | 3 +- | ||
95 | stubs/meson.build | 1 + | ||
96 | tests/qemu-iotests/274 | 20 + | ||
97 | tests/qemu-iotests/274.out | 68 ++++ | ||
98 | util/meson.build | 4 + | ||
99 | 33 files changed, 1420 insertions(+), 122 deletions(-) | ||
100 | create mode 100644 block/export/vhost-user-blk-server.h | ||
101 | create mode 100644 include/qemu/vhost-user-server.h | ||
102 | create mode 100644 util/block-helpers.h | ||
103 | create mode 100644 block/export/vhost-user-blk-server.c | ||
104 | create mode 100644 stubs/blk-exp-close-all.c | ||
105 | create mode 100644 util/block-helpers.c | ||
106 | create mode 100644 util/vhost-user-server.c | ||
62 | 107 | ||
63 | -- | 108 | -- |
64 | 2.40.1 | 109 | 2.26.2 |
110 | diff view generated by jsdifflib |
1 | From: Jean-Louis Dupond <jean-louis@dupond.be> | 1 | From: Philippe Mathieu-Daudé <philmd@redhat.com> |
---|---|---|---|
2 | 2 | ||
3 | When we for example have a sparse qcow2 image and discard: unmap is enabled, | 3 | Keep statistics of some hardware errors, and number of |
4 | there can be a lot of fragmentation in the image after some time. Especially on VM's | 4 | aligned/unaligned I/O accesses. |
5 | that do a lot of writes/deletes. | ||
6 | This causes the qcow2 image to grow even over 110% of its virtual size, | ||
7 | because the free gaps in the image get too small to allocate new | ||
8 | continuous clusters. So it allocates new space at the end of the image. | ||
9 | 5 | ||
10 | Disabling discard is not an option, as discard is needed to keep the | 6 | QMP example booting a full RHEL 8.3 aarch64 guest: |
11 | incremental backup size as low as possible. Without discard, the | ||
12 | incremental backups would become large, as qemu thinks it's just dirty | ||
13 | blocks but it doesn't know the blocks are unneeded. | ||
14 | So we need to avoid fragmentation but also 'empty' the unneeded blocks in | ||
15 | the image to have a small incremental backup. | ||
16 | 7 | ||
17 | In addition, we also want to send the discards further down the stack, so | 8 | { "execute": "query-blockstats" } |
18 | the underlying blocks are still discarded. | 9 | { |
10 | "return": [ | ||
11 | { | ||
12 | "device": "", | ||
13 | "node-name": "drive0", | ||
14 | "stats": { | ||
15 | "flush_total_time_ns": 6026948, | ||
16 | "wr_highest_offset": 3383991230464, | ||
17 | "wr_total_time_ns": 807450995, | ||
18 | "failed_wr_operations": 0, | ||
19 | "failed_rd_operations": 0, | ||
20 | "wr_merged": 3, | ||
21 | "wr_bytes": 50133504, | ||
22 | "failed_unmap_operations": 0, | ||
23 | "failed_flush_operations": 0, | ||
24 | "account_invalid": false, | ||
25 | "rd_total_time_ns": 1846979900, | ||
26 | "flush_operations": 130, | ||
27 | "wr_operations": 659, | ||
28 | "rd_merged": 1192, | ||
29 | "rd_bytes": 218244096, | ||
30 | "account_failed": false, | ||
31 | "idle_time_ns": 2678641497, | ||
32 | "rd_operations": 7406, | ||
33 | }, | ||
34 | "driver-specific": { | ||
35 | "driver": "nvme", | ||
36 | "completion-errors": 0, | ||
37 | "unaligned-accesses": 2959, | ||
38 | "aligned-accesses": 4477 | ||
39 | }, | ||
40 | "qdev": "/machine/peripheral-anon/device[0]/virtio-backend" | ||
41 | } | ||
42 | ] | ||
43 | } | ||
19 | 44 | ||
20 | Therefor we introduce a new qcow2 option "discard-no-unref". | 45 | Suggested-by: Stefan Hajnoczi <stefanha@gmail.com> |
21 | When setting this option to true, discards will no longer have the qcow2 | 46 | Signed-off-by: Philippe Mathieu-Daudé <philmd@redhat.com> |
22 | driver relinquish cluster allocations. Other than that, the request is | 47 | Acked-by: Markus Armbruster <armbru@redhat.com> |
23 | handled as normal: All clusters in range are marked as zero, and, if | 48 | Message-id: 20201001162939.1567915-1-philmd@redhat.com |
24 | pass-discard-request is true, it is passed further down the stack. | 49 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
25 | The only difference is that the now-zero clusters are preallocated | ||
26 | instead of being unallocated. | ||
27 | This will avoid fragmentation on the qcow2 image. | ||
28 | |||
29 | Fixes: https://gitlab.com/qemu-project/qemu/-/issues/1621 | ||
30 | Signed-off-by: Jean-Louis Dupond <jean-louis@dupond.be> | ||
31 | Message-Id: <20230605084523.34134-2-jean-louis@dupond.be> | ||
32 | Reviewed-by: Hanna Czenczek <hreitz@redhat.com> | ||
33 | Signed-off-by: Hanna Czenczek <hreitz@redhat.com> | ||
34 | --- | 50 | --- |
35 | qapi/block-core.json | 12 ++++++++++++ | 51 | qapi/block-core.json | 24 +++++++++++++++++++++++- |
36 | block/qcow2.h | 3 +++ | 52 | block/nvme.c | 27 +++++++++++++++++++++++++++ |
37 | block/qcow2-cluster.c | 32 ++++++++++++++++++++++++++++---- | 53 | 2 files changed, 50 insertions(+), 1 deletion(-) |
38 | block/qcow2.c | 18 ++++++++++++++++++ | ||
39 | qemu-options.hx | 12 ++++++++++++ | ||
40 | 5 files changed, 73 insertions(+), 4 deletions(-) | ||
41 | 54 | ||
42 | diff --git a/qapi/block-core.json b/qapi/block-core.json | 55 | diff --git a/qapi/block-core.json b/qapi/block-core.json |
43 | index XXXXXXX..XXXXXXX 100644 | 56 | index XXXXXXX..XXXXXXX 100644 |
44 | --- a/qapi/block-core.json | 57 | --- a/qapi/block-core.json |
45 | +++ b/qapi/block-core.json | 58 | +++ b/qapi/block-core.json |
46 | @@ -XXX,XX +XXX,XX @@ | 59 | @@ -XXX,XX +XXX,XX @@ |
47 | # @pass-discard-other: whether discard requests for the data source | 60 | 'discard-nb-failed': 'uint64', |
48 | # should be issued on other occasions where a cluster gets freed | 61 | 'discard-bytes-ok': 'uint64' } } |
49 | # | 62 | |
50 | +# @discard-no-unref: when enabled, discards from the guest will not cause | 63 | +## |
51 | +# cluster allocations to be relinquished. This prevents qcow2 fragmentation | 64 | +# @BlockStatsSpecificNvme: |
52 | +# that would be caused by such discards. Besides potential | ||
53 | +# performance degradation, such fragmentation can lead to increased | ||
54 | +# allocation of clusters past the end of the image file, | ||
55 | +# resulting in image files whose file length can grow much larger | ||
56 | +# than their guest disk size would suggest. | ||
57 | +# If image file length is of concern (e.g. when storing qcow2 | ||
58 | +# images directly on block devices), you should consider enabling | ||
59 | +# this option. (since 8.1) | ||
60 | +# | 65 | +# |
61 | # @overlap-check: which overlap checks to perform for writes to the | 66 | +# NVMe driver statistics |
62 | # image, defaults to 'cached' (since 2.2) | 67 | +# |
68 | +# @completion-errors: The number of completion errors. | ||
69 | +# | ||
70 | +# @aligned-accesses: The number of aligned accesses performed by | ||
71 | +# the driver. | ||
72 | +# | ||
73 | +# @unaligned-accesses: The number of unaligned accesses performed by | ||
74 | +# the driver. | ||
75 | +# | ||
76 | +# Since: 5.2 | ||
77 | +## | ||
78 | +{ 'struct': 'BlockStatsSpecificNvme', | ||
79 | + 'data': { | ||
80 | + 'completion-errors': 'uint64', | ||
81 | + 'aligned-accesses': 'uint64', | ||
82 | + 'unaligned-accesses': 'uint64' } } | ||
83 | + | ||
84 | ## | ||
85 | # @BlockStatsSpecific: | ||
63 | # | 86 | # |
64 | @@ -XXX,XX +XXX,XX @@ | 87 | @@ -XXX,XX +XXX,XX @@ |
65 | '*pass-discard-request': 'bool', | 88 | 'discriminator': 'driver', |
66 | '*pass-discard-snapshot': 'bool', | 89 | 'data': { |
67 | '*pass-discard-other': 'bool', | 90 | 'file': 'BlockStatsSpecificFile', |
68 | + '*discard-no-unref': 'bool', | 91 | - 'host_device': 'BlockStatsSpecificFile' } } |
69 | '*overlap-check': 'Qcow2OverlapChecks', | 92 | + 'host_device': 'BlockStatsSpecificFile', |
70 | '*cache-size': 'int', | 93 | + 'nvme': 'BlockStatsSpecificNvme' } } |
71 | '*l2-cache-size': 'int', | 94 | |
72 | diff --git a/block/qcow2.h b/block/qcow2.h | 95 | ## |
96 | # @BlockStats: | ||
97 | diff --git a/block/nvme.c b/block/nvme.c | ||
73 | index XXXXXXX..XXXXXXX 100644 | 98 | index XXXXXXX..XXXXXXX 100644 |
74 | --- a/block/qcow2.h | 99 | --- a/block/nvme.c |
75 | +++ b/block/qcow2.h | 100 | +++ b/block/nvme.c |
76 | @@ -XXX,XX +XXX,XX @@ | 101 | @@ -XXX,XX +XXX,XX @@ struct BDRVNVMeState { |
77 | #define QCOW2_OPT_DISCARD_REQUEST "pass-discard-request" | 102 | |
78 | #define QCOW2_OPT_DISCARD_SNAPSHOT "pass-discard-snapshot" | 103 | /* PCI address (required for nvme_refresh_filename()) */ |
79 | #define QCOW2_OPT_DISCARD_OTHER "pass-discard-other" | 104 | char *device; |
80 | +#define QCOW2_OPT_DISCARD_NO_UNREF "discard-no-unref" | ||
81 | #define QCOW2_OPT_OVERLAP "overlap-check" | ||
82 | #define QCOW2_OPT_OVERLAP_TEMPLATE "overlap-check.template" | ||
83 | #define QCOW2_OPT_OVERLAP_MAIN_HEADER "overlap-check.main-header" | ||
84 | @@ -XXX,XX +XXX,XX @@ typedef struct BDRVQcow2State { | ||
85 | |||
86 | bool discard_passthrough[QCOW2_DISCARD_MAX]; | ||
87 | |||
88 | + bool discard_no_unref; | ||
89 | + | 105 | + |
90 | int overlap_check; /* bitmask of Qcow2MetadataOverlap values */ | 106 | + struct { |
91 | bool signaled_corruption; | 107 | + uint64_t completion_errors; |
92 | 108 | + uint64_t aligned_accesses; | |
93 | diff --git a/block/qcow2-cluster.c b/block/qcow2-cluster.c | 109 | + uint64_t unaligned_accesses; |
94 | index XXXXXXX..XXXXXXX 100644 | 110 | + } stats; |
95 | --- a/block/qcow2-cluster.c | 111 | }; |
96 | +++ b/block/qcow2-cluster.c | 112 | |
97 | @@ -XXX,XX +XXX,XX @@ static int discard_in_l2_slice(BlockDriverState *bs, uint64_t offset, | 113 | #define NVME_BLOCK_OPT_DEVICE "device" |
98 | uint64_t new_l2_bitmap = old_l2_bitmap; | 114 | @@ -XXX,XX +XXX,XX @@ static bool nvme_process_completion(NVMeQueuePair *q) |
99 | QCow2ClusterType cluster_type = | 115 | break; |
100 | qcow2_get_cluster_type(bs, old_l2_entry); | ||
101 | + bool keep_reference = (cluster_type != QCOW2_CLUSTER_COMPRESSED) && | ||
102 | + !full_discard && | ||
103 | + (s->discard_no_unref && | ||
104 | + type == QCOW2_DISCARD_REQUEST); | ||
105 | |||
106 | /* | ||
107 | * If full_discard is true, the cluster should not read back as zeroes, | ||
108 | @@ -XXX,XX +XXX,XX @@ static int discard_in_l2_slice(BlockDriverState *bs, uint64_t offset, | ||
109 | new_l2_entry = new_l2_bitmap = 0; | ||
110 | } else if (bs->backing || qcow2_cluster_is_allocated(cluster_type)) { | ||
111 | if (has_subclusters(s)) { | ||
112 | - new_l2_entry = 0; | ||
113 | + if (keep_reference) { | ||
114 | + new_l2_entry = old_l2_entry; | ||
115 | + } else { | ||
116 | + new_l2_entry = 0; | ||
117 | + } | ||
118 | new_l2_bitmap = QCOW_L2_BITMAP_ALL_ZEROES; | ||
119 | } else { | ||
120 | - new_l2_entry = s->qcow_version >= 3 ? QCOW_OFLAG_ZERO : 0; | ||
121 | + if (s->qcow_version >= 3) { | ||
122 | + if (keep_reference) { | ||
123 | + new_l2_entry |= QCOW_OFLAG_ZERO; | ||
124 | + } else { | ||
125 | + new_l2_entry = QCOW_OFLAG_ZERO; | ||
126 | + } | ||
127 | + } else { | ||
128 | + new_l2_entry = 0; | ||
129 | + } | ||
130 | } | ||
131 | } | 116 | } |
132 | 117 | ret = nvme_translate_error(c); | |
133 | @@ -XXX,XX +XXX,XX @@ static int discard_in_l2_slice(BlockDriverState *bs, uint64_t offset, | 118 | + if (ret) { |
134 | if (has_subclusters(s)) { | 119 | + s->stats.completion_errors++; |
135 | set_l2_bitmap(s, l2_slice, l2_index + i, new_l2_bitmap); | 120 | + } |
136 | } | 121 | q->cq.head = (q->cq.head + 1) % NVME_QUEUE_SIZE; |
137 | - /* Then decrease the refcount */ | 122 | if (!q->cq.head) { |
138 | - qcow2_free_any_cluster(bs, old_l2_entry, type); | 123 | q->cq_phase = !q->cq_phase; |
139 | + if (!keep_reference) { | 124 | @@ -XXX,XX +XXX,XX @@ static int nvme_co_prw(BlockDriverState *bs, uint64_t offset, uint64_t bytes, |
140 | + /* Then decrease the refcount */ | 125 | assert(QEMU_IS_ALIGNED(bytes, s->page_size)); |
141 | + qcow2_free_any_cluster(bs, old_l2_entry, type); | 126 | assert(bytes <= s->max_transfer); |
142 | + } else if (s->discard_passthrough[type] && | 127 | if (nvme_qiov_aligned(bs, qiov)) { |
143 | + (cluster_type == QCOW2_CLUSTER_NORMAL || | 128 | + s->stats.aligned_accesses++; |
144 | + cluster_type == QCOW2_CLUSTER_ZERO_ALLOC)) { | 129 | return nvme_co_prw_aligned(bs, offset, bytes, qiov, is_write, flags); |
145 | + /* If we keep the reference, pass on the discard still */ | ||
146 | + bdrv_pdiscard(s->data_file, old_l2_entry & L2E_OFFSET_MASK, | ||
147 | + s->cluster_size); | ||
148 | + } | ||
149 | } | 130 | } |
150 | 131 | + s->stats.unaligned_accesses++; | |
151 | qcow2_cache_put(s->l2_table_cache, (void **) &l2_slice); | 132 | trace_nvme_prw_buffered(s, offset, bytes, qiov->niov, is_write); |
152 | diff --git a/block/qcow2.c b/block/qcow2.c | 133 | buf = qemu_try_memalign(s->page_size, bytes); |
153 | index XXXXXXX..XXXXXXX 100644 | 134 | |
154 | --- a/block/qcow2.c | 135 | @@ -XXX,XX +XXX,XX @@ static void nvme_unregister_buf(BlockDriverState *bs, void *host) |
155 | +++ b/block/qcow2.c | 136 | qemu_vfio_dma_unmap(s->vfio, host); |
156 | @@ -XXX,XX +XXX,XX @@ static const char *const mutable_opts[] = { | 137 | } |
157 | QCOW2_OPT_DISCARD_REQUEST, | 138 | |
158 | QCOW2_OPT_DISCARD_SNAPSHOT, | 139 | +static BlockStatsSpecific *nvme_get_specific_stats(BlockDriverState *bs) |
159 | QCOW2_OPT_DISCARD_OTHER, | 140 | +{ |
160 | + QCOW2_OPT_DISCARD_NO_UNREF, | 141 | + BlockStatsSpecific *stats = g_new(BlockStatsSpecific, 1); |
161 | QCOW2_OPT_OVERLAP, | 142 | + BDRVNVMeState *s = bs->opaque; |
162 | QCOW2_OPT_OVERLAP_TEMPLATE, | ||
163 | QCOW2_OPT_OVERLAP_MAIN_HEADER, | ||
164 | @@ -XXX,XX +XXX,XX @@ static QemuOptsList qcow2_runtime_opts = { | ||
165 | .type = QEMU_OPT_BOOL, | ||
166 | .help = "Generate discard requests when other clusters are freed", | ||
167 | }, | ||
168 | + { | ||
169 | + .name = QCOW2_OPT_DISCARD_NO_UNREF, | ||
170 | + .type = QEMU_OPT_BOOL, | ||
171 | + .help = "Do not unreference discarded clusters", | ||
172 | + }, | ||
173 | { | ||
174 | .name = QCOW2_OPT_OVERLAP, | ||
175 | .type = QEMU_OPT_STRING, | ||
176 | @@ -XXX,XX +XXX,XX @@ typedef struct Qcow2ReopenState { | ||
177 | bool use_lazy_refcounts; | ||
178 | int overlap_check; | ||
179 | bool discard_passthrough[QCOW2_DISCARD_MAX]; | ||
180 | + bool discard_no_unref; | ||
181 | uint64_t cache_clean_interval; | ||
182 | QCryptoBlockOpenOptions *crypto_opts; /* Disk encryption runtime options */ | ||
183 | } Qcow2ReopenState; | ||
184 | @@ -XXX,XX +XXX,XX @@ static int qcow2_update_options_prepare(BlockDriverState *bs, | ||
185 | r->discard_passthrough[QCOW2_DISCARD_OTHER] = | ||
186 | qemu_opt_get_bool(opts, QCOW2_OPT_DISCARD_OTHER, false); | ||
187 | |||
188 | + r->discard_no_unref = qemu_opt_get_bool(opts, QCOW2_OPT_DISCARD_NO_UNREF, | ||
189 | + false); | ||
190 | + if (r->discard_no_unref && s->qcow_version < 3) { | ||
191 | + error_setg(errp, | ||
192 | + "discard-no-unref is only supported since qcow2 version 3"); | ||
193 | + ret = -EINVAL; | ||
194 | + goto fail; | ||
195 | + } | ||
196 | + | 143 | + |
197 | switch (s->crypt_method_header) { | 144 | + stats->driver = BLOCKDEV_DRIVER_NVME; |
198 | case QCOW_CRYPT_NONE: | 145 | + stats->u.nvme = (BlockStatsSpecificNvme) { |
199 | if (encryptfmt) { | 146 | + .completion_errors = s->stats.completion_errors, |
200 | @@ -XXX,XX +XXX,XX @@ static void qcow2_update_options_commit(BlockDriverState *bs, | 147 | + .aligned_accesses = s->stats.aligned_accesses, |
201 | s->discard_passthrough[i] = r->discard_passthrough[i]; | 148 | + .unaligned_accesses = s->stats.unaligned_accesses, |
202 | } | 149 | + }; |
203 | |||
204 | + s->discard_no_unref = r->discard_no_unref; | ||
205 | + | 150 | + |
206 | if (s->cache_clean_interval != r->cache_clean_interval) { | 151 | + return stats; |
207 | cache_clean_timer_del(bs); | 152 | +} |
208 | s->cache_clean_interval = r->cache_clean_interval; | ||
209 | diff --git a/qemu-options.hx b/qemu-options.hx | ||
210 | index XXXXXXX..XXXXXXX 100644 | ||
211 | --- a/qemu-options.hx | ||
212 | +++ b/qemu-options.hx | ||
213 | @@ -XXX,XX +XXX,XX @@ SRST | ||
214 | issued on other occasions where a cluster gets freed | ||
215 | (on/off; default: off) | ||
216 | |||
217 | + ``discard-no-unref`` | ||
218 | + When enabled, discards from the guest will not cause cluster | ||
219 | + allocations to be relinquished. This prevents qcow2 fragmentation | ||
220 | + that would be caused by such discards. Besides potential | ||
221 | + performance degradation, such fragmentation can lead to increased | ||
222 | + allocation of clusters past the end of the image file, | ||
223 | + resulting in image files whose file length can grow much larger | ||
224 | + than their guest disk size would suggest. | ||
225 | + If image file length is of concern (e.g. when storing qcow2 | ||
226 | + images directly on block devices), you should consider enabling | ||
227 | + this option. | ||
228 | + | 153 | + |
229 | ``overlap-check`` | 154 | static const char *const nvme_strong_runtime_opts[] = { |
230 | Which overlap checks to perform for writes to the image | 155 | NVME_BLOCK_OPT_DEVICE, |
231 | (none/constant/cached/all; default: cached). For details or | 156 | NVME_BLOCK_OPT_NAMESPACE, |
157 | @@ -XXX,XX +XXX,XX @@ static BlockDriver bdrv_nvme = { | ||
158 | .bdrv_refresh_filename = nvme_refresh_filename, | ||
159 | .bdrv_refresh_limits = nvme_refresh_limits, | ||
160 | .strong_runtime_opts = nvme_strong_runtime_opts, | ||
161 | + .bdrv_get_specific_stats = nvme_get_specific_stats, | ||
162 | |||
163 | .bdrv_detach_aio_context = nvme_detach_aio_context, | ||
164 | .bdrv_attach_aio_context = nvme_attach_aio_context, | ||
232 | -- | 165 | -- |
233 | 2.40.1 | 166 | 2.26.2 |
167 | diff view generated by jsdifflib |
1 | From: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | 1 | From: Coiby Xu <coiby.xu@gmail.com> |
---|---|---|---|
2 | 2 | ||
3 | This helper will be reused in next patches during parallels_co_check | 3 | Allow vu_message_read to be replaced by one which will make use of the |
4 | rework to simplify its code. | 4 | QIOChannel functions. Thus reading vhost-user message won't stall the |
5 | guest. For slave channel, we still use the default vu_message_read. | ||
5 | 6 | ||
6 | Signed-off-by: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | 7 | Reviewed-by: Marc-André Lureau <marcandre.lureau@redhat.com> |
7 | Reviewed-by: Denis V. Lunev <den@openvz.org> | 8 | Signed-off-by: Coiby Xu <coiby.xu@gmail.com> |
8 | Reviewed-by: Vladimir Sementsov-Ogievskiy <vsementsov@yandex-team.ru> | 9 | Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> |
9 | Message-Id: <20230424093147.197643-5-alexander.ivanov@virtuozzo.com> | 10 | Message-id: 20200918080912.321299-2-coiby.xu@gmail.com |
10 | Reviewed-by: Hanna Czenczek <hreitz@redhat.com> | 11 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
11 | Signed-off-by: Hanna Czenczek <hreitz@redhat.com> | ||
12 | --- | 12 | --- |
13 | block/parallels.c | 11 ++++++++--- | 13 | contrib/libvhost-user/libvhost-user.h | 21 +++++++++++++++++++++ |
14 | 1 file changed, 8 insertions(+), 3 deletions(-) | 14 | contrib/libvhost-user/libvhost-user-glib.c | 2 +- |
15 | contrib/libvhost-user/libvhost-user.c | 14 +++++++------- | ||
16 | tests/vhost-user-bridge.c | 2 ++ | ||
17 | tools/virtiofsd/fuse_virtio.c | 4 ++-- | ||
18 | 5 files changed, 33 insertions(+), 10 deletions(-) | ||
15 | 19 | ||
16 | diff --git a/block/parallels.c b/block/parallels.c | 20 | diff --git a/contrib/libvhost-user/libvhost-user.h b/contrib/libvhost-user/libvhost-user.h |
17 | index XXXXXXX..XXXXXXX 100644 | 21 | index XXXXXXX..XXXXXXX 100644 |
18 | --- a/block/parallels.c | 22 | --- a/contrib/libvhost-user/libvhost-user.h |
19 | +++ b/block/parallels.c | 23 | +++ b/contrib/libvhost-user/libvhost-user.h |
20 | @@ -XXX,XX +XXX,XX @@ static int64_t block_status(BDRVParallelsState *s, int64_t sector_num, | 24 | @@ -XXX,XX +XXX,XX @@ |
21 | return start_off; | 25 | */ |
26 | #define VHOST_USER_MAX_RAM_SLOTS 32 | ||
27 | |||
28 | +#define VHOST_USER_HDR_SIZE offsetof(VhostUserMsg, payload.u64) | ||
29 | + | ||
30 | typedef enum VhostSetConfigType { | ||
31 | VHOST_SET_CONFIG_TYPE_MASTER = 0, | ||
32 | VHOST_SET_CONFIG_TYPE_MIGRATION = 1, | ||
33 | @@ -XXX,XX +XXX,XX @@ typedef uint64_t (*vu_get_features_cb) (VuDev *dev); | ||
34 | typedef void (*vu_set_features_cb) (VuDev *dev, uint64_t features); | ||
35 | typedef int (*vu_process_msg_cb) (VuDev *dev, VhostUserMsg *vmsg, | ||
36 | int *do_reply); | ||
37 | +typedef bool (*vu_read_msg_cb) (VuDev *dev, int sock, VhostUserMsg *vmsg); | ||
38 | typedef void (*vu_queue_set_started_cb) (VuDev *dev, int qidx, bool started); | ||
39 | typedef bool (*vu_queue_is_processed_in_order_cb) (VuDev *dev, int qidx); | ||
40 | typedef int (*vu_get_config_cb) (VuDev *dev, uint8_t *config, uint32_t len); | ||
41 | @@ -XXX,XX +XXX,XX @@ struct VuDev { | ||
42 | bool broken; | ||
43 | uint16_t max_queues; | ||
44 | |||
45 | + /* @read_msg: custom method to read vhost-user message | ||
46 | + * | ||
47 | + * Read data from vhost_user socket fd and fill up | ||
48 | + * the passed VhostUserMsg *vmsg struct. | ||
49 | + * | ||
50 | + * If reading fails, it should close the received set of file | ||
51 | + * descriptors as socket message's auxiliary data. | ||
52 | + * | ||
53 | + * For the details, please refer to vu_message_read in libvhost-user.c | ||
54 | + * which will be used by default if not custom method is provided when | ||
55 | + * calling vu_init | ||
56 | + * | ||
57 | + * Returns: true if vhost-user message successfully received, | ||
58 | + * otherwise return false. | ||
59 | + * | ||
60 | + */ | ||
61 | + vu_read_msg_cb read_msg; | ||
62 | /* @set_watch: add or update the given fd to the watch set, | ||
63 | * call cb when condition is met */ | ||
64 | vu_set_watch_cb set_watch; | ||
65 | @@ -XXX,XX +XXX,XX @@ bool vu_init(VuDev *dev, | ||
66 | uint16_t max_queues, | ||
67 | int socket, | ||
68 | vu_panic_cb panic, | ||
69 | + vu_read_msg_cb read_msg, | ||
70 | vu_set_watch_cb set_watch, | ||
71 | vu_remove_watch_cb remove_watch, | ||
72 | const VuDevIface *iface); | ||
73 | diff --git a/contrib/libvhost-user/libvhost-user-glib.c b/contrib/libvhost-user/libvhost-user-glib.c | ||
74 | index XXXXXXX..XXXXXXX 100644 | ||
75 | --- a/contrib/libvhost-user/libvhost-user-glib.c | ||
76 | +++ b/contrib/libvhost-user/libvhost-user-glib.c | ||
77 | @@ -XXX,XX +XXX,XX @@ vug_init(VugDev *dev, uint16_t max_queues, int socket, | ||
78 | g_assert(dev); | ||
79 | g_assert(iface); | ||
80 | |||
81 | - if (!vu_init(&dev->parent, max_queues, socket, panic, set_watch, | ||
82 | + if (!vu_init(&dev->parent, max_queues, socket, panic, NULL, set_watch, | ||
83 | remove_watch, iface)) { | ||
84 | return false; | ||
85 | } | ||
86 | diff --git a/contrib/libvhost-user/libvhost-user.c b/contrib/libvhost-user/libvhost-user.c | ||
87 | index XXXXXXX..XXXXXXX 100644 | ||
88 | --- a/contrib/libvhost-user/libvhost-user.c | ||
89 | +++ b/contrib/libvhost-user/libvhost-user.c | ||
90 | @@ -XXX,XX +XXX,XX @@ | ||
91 | /* The version of inflight buffer */ | ||
92 | #define INFLIGHT_VERSION 1 | ||
93 | |||
94 | -#define VHOST_USER_HDR_SIZE offsetof(VhostUserMsg, payload.u64) | ||
95 | - | ||
96 | /* The version of the protocol we support */ | ||
97 | #define VHOST_USER_VERSION 1 | ||
98 | #define LIBVHOST_USER_DEBUG 0 | ||
99 | @@ -XXX,XX +XXX,XX @@ have_userfault(void) | ||
22 | } | 100 | } |
23 | 101 | ||
24 | +static void parallels_set_bat_entry(BDRVParallelsState *s, | 102 | static bool |
25 | + uint32_t index, uint32_t offset) | 103 | -vu_message_read(VuDev *dev, int conn_fd, VhostUserMsg *vmsg) |
26 | +{ | 104 | +vu_message_read_default(VuDev *dev, int conn_fd, VhostUserMsg *vmsg) |
27 | + s->bat_bitmap[index] = cpu_to_le32(offset); | 105 | { |
28 | + bitmap_set(s->bat_dirty_bmap, bat_entry_off(index) / s->bat_dirty_block, 1); | 106 | char control[CMSG_SPACE(VHOST_MEMORY_BASELINE_NREGIONS * sizeof(int))] = {}; |
29 | +} | 107 | struct iovec iov = { |
30 | + | 108 | @@ -XXX,XX +XXX,XX @@ vu_process_message_reply(VuDev *dev, const VhostUserMsg *vmsg) |
31 | static int64_t coroutine_fn GRAPH_RDLOCK | 109 | goto out; |
32 | allocate_clusters(BlockDriverState *bs, int64_t sector_num, | ||
33 | int nb_sectors, int *pnum) | ||
34 | @@ -XXX,XX +XXX,XX @@ allocate_clusters(BlockDriverState *bs, int64_t sector_num, | ||
35 | } | 110 | } |
36 | 111 | ||
37 | for (i = 0; i < to_allocate; i++) { | 112 | - if (!vu_message_read(dev, dev->slave_fd, &msg_reply)) { |
38 | - s->bat_bitmap[idx + i] = cpu_to_le32(s->data_end / s->off_multiplier); | 113 | + if (!vu_message_read_default(dev, dev->slave_fd, &msg_reply)) { |
39 | + parallels_set_bat_entry(s, idx + i, s->data_end / s->off_multiplier); | 114 | goto out; |
40 | s->data_end += s->tracks; | ||
41 | - bitmap_set(s->bat_dirty_bmap, | ||
42 | - bat_entry_off(idx + i) / s->bat_dirty_block, 1); | ||
43 | } | 115 | } |
44 | 116 | ||
45 | return bat2sect(s, idx) + sector_num % s->tracks; | 117 | @@ -XXX,XX +XXX,XX @@ vu_set_mem_table_exec_postcopy(VuDev *dev, VhostUserMsg *vmsg) |
118 | /* Wait for QEMU to confirm that it's registered the handler for the | ||
119 | * faults. | ||
120 | */ | ||
121 | - if (!vu_message_read(dev, dev->sock, vmsg) || | ||
122 | + if (!dev->read_msg(dev, dev->sock, vmsg) || | ||
123 | vmsg->size != sizeof(vmsg->payload.u64) || | ||
124 | vmsg->payload.u64 != 0) { | ||
125 | vu_panic(dev, "failed to receive valid ack for postcopy set-mem-table"); | ||
126 | @@ -XXX,XX +XXX,XX @@ vu_dispatch(VuDev *dev) | ||
127 | int reply_requested; | ||
128 | bool need_reply, success = false; | ||
129 | |||
130 | - if (!vu_message_read(dev, dev->sock, &vmsg)) { | ||
131 | + if (!dev->read_msg(dev, dev->sock, &vmsg)) { | ||
132 | goto end; | ||
133 | } | ||
134 | |||
135 | @@ -XXX,XX +XXX,XX @@ vu_init(VuDev *dev, | ||
136 | uint16_t max_queues, | ||
137 | int socket, | ||
138 | vu_panic_cb panic, | ||
139 | + vu_read_msg_cb read_msg, | ||
140 | vu_set_watch_cb set_watch, | ||
141 | vu_remove_watch_cb remove_watch, | ||
142 | const VuDevIface *iface) | ||
143 | @@ -XXX,XX +XXX,XX @@ vu_init(VuDev *dev, | ||
144 | |||
145 | dev->sock = socket; | ||
146 | dev->panic = panic; | ||
147 | + dev->read_msg = read_msg ? read_msg : vu_message_read_default; | ||
148 | dev->set_watch = set_watch; | ||
149 | dev->remove_watch = remove_watch; | ||
150 | dev->iface = iface; | ||
151 | @@ -XXX,XX +XXX,XX @@ static void _vu_queue_notify(VuDev *dev, VuVirtq *vq, bool sync) | ||
152 | |||
153 | vu_message_write(dev, dev->slave_fd, &vmsg); | ||
154 | if (ack) { | ||
155 | - vu_message_read(dev, dev->slave_fd, &vmsg); | ||
156 | + vu_message_read_default(dev, dev->slave_fd, &vmsg); | ||
157 | } | ||
158 | return; | ||
159 | } | ||
160 | diff --git a/tests/vhost-user-bridge.c b/tests/vhost-user-bridge.c | ||
161 | index XXXXXXX..XXXXXXX 100644 | ||
162 | --- a/tests/vhost-user-bridge.c | ||
163 | +++ b/tests/vhost-user-bridge.c | ||
164 | @@ -XXX,XX +XXX,XX @@ vubr_accept_cb(int sock, void *ctx) | ||
165 | VHOST_USER_BRIDGE_MAX_QUEUES, | ||
166 | conn_fd, | ||
167 | vubr_panic, | ||
168 | + NULL, | ||
169 | vubr_set_watch, | ||
170 | vubr_remove_watch, | ||
171 | &vuiface)) { | ||
172 | @@ -XXX,XX +XXX,XX @@ vubr_new(const char *path, bool client) | ||
173 | VHOST_USER_BRIDGE_MAX_QUEUES, | ||
174 | dev->sock, | ||
175 | vubr_panic, | ||
176 | + NULL, | ||
177 | vubr_set_watch, | ||
178 | vubr_remove_watch, | ||
179 | &vuiface)) { | ||
180 | diff --git a/tools/virtiofsd/fuse_virtio.c b/tools/virtiofsd/fuse_virtio.c | ||
181 | index XXXXXXX..XXXXXXX 100644 | ||
182 | --- a/tools/virtiofsd/fuse_virtio.c | ||
183 | +++ b/tools/virtiofsd/fuse_virtio.c | ||
184 | @@ -XXX,XX +XXX,XX @@ int virtio_session_mount(struct fuse_session *se) | ||
185 | se->vu_socketfd = data_sock; | ||
186 | se->virtio_dev->se = se; | ||
187 | pthread_rwlock_init(&se->virtio_dev->vu_dispatch_rwlock, NULL); | ||
188 | - vu_init(&se->virtio_dev->dev, 2, se->vu_socketfd, fv_panic, fv_set_watch, | ||
189 | - fv_remove_watch, &fv_iface); | ||
190 | + vu_init(&se->virtio_dev->dev, 2, se->vu_socketfd, fv_panic, NULL, | ||
191 | + fv_set_watch, fv_remove_watch, &fv_iface); | ||
192 | |||
193 | return 0; | ||
194 | } | ||
46 | -- | 195 | -- |
47 | 2.40.1 | 196 | 2.26.2 |
197 | diff view generated by jsdifflib |
1 | From: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | 1 | From: Coiby Xu <coiby.xu@gmail.com> |
---|---|---|---|
2 | 2 | ||
3 | Don't let high_off be more than the file size even if we don't fix the | 3 | When the client is running in gdb and quit command is run in gdb, |
4 | image. | 4 | QEMU will still dispatch the event which will cause segment fault in |
5 | the callback function. | ||
5 | 6 | ||
6 | Signed-off-by: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | 7 | Signed-off-by: Coiby Xu <coiby.xu@gmail.com> |
7 | Reviewed-by: Denis V. Lunev <den@openvz.org> | 8 | Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> |
8 | Reviewed-by: Vladimir Sementsov-Ogievskiy <vsementsov@yandex-team.ru> | 9 | Reviewed-by: Marc-André Lureau <marcandre.lureau@redhat.com> |
9 | Message-Id: <20230424093147.197643-3-alexander.ivanov@virtuozzo.com> | 10 | Message-id: 20200918080912.321299-3-coiby.xu@gmail.com |
10 | Reviewed-by: Hanna Czenczek <hreitz@redhat.com> | 11 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
11 | Signed-off-by: Hanna Czenczek <hreitz@redhat.com> | ||
12 | --- | 12 | --- |
13 | block/parallels.c | 4 ++-- | 13 | contrib/libvhost-user/libvhost-user.c | 1 + |
14 | 1 file changed, 2 insertions(+), 2 deletions(-) | 14 | 1 file changed, 1 insertion(+) |
15 | 15 | ||
16 | diff --git a/block/parallels.c b/block/parallels.c | 16 | diff --git a/contrib/libvhost-user/libvhost-user.c b/contrib/libvhost-user/libvhost-user.c |
17 | index XXXXXXX..XXXXXXX 100644 | 17 | index XXXXXXX..XXXXXXX 100644 |
18 | --- a/block/parallels.c | 18 | --- a/contrib/libvhost-user/libvhost-user.c |
19 | +++ b/block/parallels.c | 19 | +++ b/contrib/libvhost-user/libvhost-user.c |
20 | @@ -XXX,XX +XXX,XX @@ parallels_co_check(BlockDriverState *bs, BdrvCheckResult *res, | 20 | @@ -XXX,XX +XXX,XX @@ vu_deinit(VuDev *dev) |
21 | fix & BDRV_FIX_ERRORS ? "Repairing" : "ERROR", i); | ||
22 | res->corruptions++; | ||
23 | if (fix & BDRV_FIX_ERRORS) { | ||
24 | - prev_off = 0; | ||
25 | s->bat_bitmap[i] = 0; | ||
26 | res->corruptions_fixed++; | ||
27 | flush_bat = true; | ||
28 | - continue; | ||
29 | } | ||
30 | + prev_off = 0; | ||
31 | + continue; | ||
32 | } | 21 | } |
33 | 22 | ||
34 | res->bfi.allocated_clusters++; | 23 | if (vq->kick_fd != -1) { |
24 | + dev->remove_watch(dev, vq->kick_fd); | ||
25 | close(vq->kick_fd); | ||
26 | vq->kick_fd = -1; | ||
27 | } | ||
35 | -- | 28 | -- |
36 | 2.40.1 | 29 | 2.26.2 |
30 | diff view generated by jsdifflib |
New patch | |||
---|---|---|---|
1 | From: Coiby Xu <coiby.xu@gmail.com> | ||
1 | 2 | ||
3 | Sharing QEMU devices via vhost-user protocol. | ||
4 | |||
5 | Only one vhost-user client can connect to the server one time. | ||
6 | |||
7 | Suggested-by: Kevin Wolf <kwolf@redhat.com> | ||
8 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
9 | Signed-off-by: Coiby Xu <coiby.xu@gmail.com> | ||
10 | Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
11 | Reviewed-by: Marc-André Lureau <marcandre.lureau@redhat.com> | ||
12 | Message-id: 20200918080912.321299-4-coiby.xu@gmail.com | ||
13 | [Fixed size_t %lu -> %zu format string compiler error. | ||
14 | --Stefan] | ||
15 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
16 | --- | ||
17 | util/vhost-user-server.h | 65 ++++++ | ||
18 | util/vhost-user-server.c | 428 +++++++++++++++++++++++++++++++++++++++ | ||
19 | util/meson.build | 1 + | ||
20 | 3 files changed, 494 insertions(+) | ||
21 | create mode 100644 util/vhost-user-server.h | ||
22 | create mode 100644 util/vhost-user-server.c | ||
23 | |||
24 | diff --git a/util/vhost-user-server.h b/util/vhost-user-server.h | ||
25 | new file mode 100644 | ||
26 | index XXXXXXX..XXXXXXX | ||
27 | --- /dev/null | ||
28 | +++ b/util/vhost-user-server.h | ||
29 | @@ -XXX,XX +XXX,XX @@ | ||
30 | +/* | ||
31 | + * Sharing QEMU devices via vhost-user protocol | ||
32 | + * | ||
33 | + * Copyright (c) Coiby Xu <coiby.xu@gmail.com>. | ||
34 | + * Copyright (c) 2020 Red Hat, Inc. | ||
35 | + * | ||
36 | + * This work is licensed under the terms of the GNU GPL, version 2 or | ||
37 | + * later. See the COPYING file in the top-level directory. | ||
38 | + */ | ||
39 | + | ||
40 | +#ifndef VHOST_USER_SERVER_H | ||
41 | +#define VHOST_USER_SERVER_H | ||
42 | + | ||
43 | +#include "contrib/libvhost-user/libvhost-user.h" | ||
44 | +#include "io/channel-socket.h" | ||
45 | +#include "io/channel-file.h" | ||
46 | +#include "io/net-listener.h" | ||
47 | +#include "qemu/error-report.h" | ||
48 | +#include "qapi/error.h" | ||
49 | +#include "standard-headers/linux/virtio_blk.h" | ||
50 | + | ||
51 | +typedef struct VuFdWatch { | ||
52 | + VuDev *vu_dev; | ||
53 | + int fd; /*kick fd*/ | ||
54 | + void *pvt; | ||
55 | + vu_watch_cb cb; | ||
56 | + bool processing; | ||
57 | + QTAILQ_ENTRY(VuFdWatch) next; | ||
58 | +} VuFdWatch; | ||
59 | + | ||
60 | +typedef struct VuServer VuServer; | ||
61 | +typedef void DevicePanicNotifierFn(VuServer *server); | ||
62 | + | ||
63 | +struct VuServer { | ||
64 | + QIONetListener *listener; | ||
65 | + AioContext *ctx; | ||
66 | + DevicePanicNotifierFn *device_panic_notifier; | ||
67 | + int max_queues; | ||
68 | + const VuDevIface *vu_iface; | ||
69 | + VuDev vu_dev; | ||
70 | + QIOChannel *ioc; /* The I/O channel with the client */ | ||
71 | + QIOChannelSocket *sioc; /* The underlying data channel with the client */ | ||
72 | + /* IOChannel for fd provided via VHOST_USER_SET_SLAVE_REQ_FD */ | ||
73 | + QIOChannel *ioc_slave; | ||
74 | + QIOChannelSocket *sioc_slave; | ||
75 | + Coroutine *co_trip; /* coroutine for processing VhostUserMsg */ | ||
76 | + QTAILQ_HEAD(, VuFdWatch) vu_fd_watches; | ||
77 | + /* restart coroutine co_trip if AIOContext is changed */ | ||
78 | + bool aio_context_changed; | ||
79 | + bool processing_msg; | ||
80 | +}; | ||
81 | + | ||
82 | +bool vhost_user_server_start(VuServer *server, | ||
83 | + SocketAddress *unix_socket, | ||
84 | + AioContext *ctx, | ||
85 | + uint16_t max_queues, | ||
86 | + DevicePanicNotifierFn *device_panic_notifier, | ||
87 | + const VuDevIface *vu_iface, | ||
88 | + Error **errp); | ||
89 | + | ||
90 | +void vhost_user_server_stop(VuServer *server); | ||
91 | + | ||
92 | +void vhost_user_server_set_aio_context(VuServer *server, AioContext *ctx); | ||
93 | + | ||
94 | +#endif /* VHOST_USER_SERVER_H */ | ||
95 | diff --git a/util/vhost-user-server.c b/util/vhost-user-server.c | ||
96 | new file mode 100644 | ||
97 | index XXXXXXX..XXXXXXX | ||
98 | --- /dev/null | ||
99 | +++ b/util/vhost-user-server.c | ||
100 | @@ -XXX,XX +XXX,XX @@ | ||
101 | +/* | ||
102 | + * Sharing QEMU devices via vhost-user protocol | ||
103 | + * | ||
104 | + * Copyright (c) Coiby Xu <coiby.xu@gmail.com>. | ||
105 | + * Copyright (c) 2020 Red Hat, Inc. | ||
106 | + * | ||
107 | + * This work is licensed under the terms of the GNU GPL, version 2 or | ||
108 | + * later. See the COPYING file in the top-level directory. | ||
109 | + */ | ||
110 | +#include "qemu/osdep.h" | ||
111 | +#include "qemu/main-loop.h" | ||
112 | +#include "vhost-user-server.h" | ||
113 | + | ||
114 | +static void vmsg_close_fds(VhostUserMsg *vmsg) | ||
115 | +{ | ||
116 | + int i; | ||
117 | + for (i = 0; i < vmsg->fd_num; i++) { | ||
118 | + close(vmsg->fds[i]); | ||
119 | + } | ||
120 | +} | ||
121 | + | ||
122 | +static void vmsg_unblock_fds(VhostUserMsg *vmsg) | ||
123 | +{ | ||
124 | + int i; | ||
125 | + for (i = 0; i < vmsg->fd_num; i++) { | ||
126 | + qemu_set_nonblock(vmsg->fds[i]); | ||
127 | + } | ||
128 | +} | ||
129 | + | ||
130 | +static void vu_accept(QIONetListener *listener, QIOChannelSocket *sioc, | ||
131 | + gpointer opaque); | ||
132 | + | ||
133 | +static void close_client(VuServer *server) | ||
134 | +{ | ||
135 | + /* | ||
136 | + * Before closing the client | ||
137 | + * | ||
138 | + * 1. Let vu_client_trip stop processing new vhost-user msg | ||
139 | + * | ||
140 | + * 2. remove kick_handler | ||
141 | + * | ||
142 | + * 3. wait for the kick handler to be finished | ||
143 | + * | ||
144 | + * 4. wait for the current vhost-user msg to be finished processing | ||
145 | + */ | ||
146 | + | ||
147 | + QIOChannelSocket *sioc = server->sioc; | ||
148 | + /* When this is set vu_client_trip will stop new processing vhost-user message */ | ||
149 | + server->sioc = NULL; | ||
150 | + | ||
151 | + VuFdWatch *vu_fd_watch, *next; | ||
152 | + QTAILQ_FOREACH_SAFE(vu_fd_watch, &server->vu_fd_watches, next, next) { | ||
153 | + aio_set_fd_handler(server->ioc->ctx, vu_fd_watch->fd, true, NULL, | ||
154 | + NULL, NULL, NULL); | ||
155 | + } | ||
156 | + | ||
157 | + while (!QTAILQ_EMPTY(&server->vu_fd_watches)) { | ||
158 | + QTAILQ_FOREACH_SAFE(vu_fd_watch, &server->vu_fd_watches, next, next) { | ||
159 | + if (!vu_fd_watch->processing) { | ||
160 | + QTAILQ_REMOVE(&server->vu_fd_watches, vu_fd_watch, next); | ||
161 | + g_free(vu_fd_watch); | ||
162 | + } | ||
163 | + } | ||
164 | + } | ||
165 | + | ||
166 | + while (server->processing_msg) { | ||
167 | + if (server->ioc->read_coroutine) { | ||
168 | + server->ioc->read_coroutine = NULL; | ||
169 | + qio_channel_set_aio_fd_handler(server->ioc, server->ioc->ctx, NULL, | ||
170 | + NULL, server->ioc); | ||
171 | + server->processing_msg = false; | ||
172 | + } | ||
173 | + } | ||
174 | + | ||
175 | + vu_deinit(&server->vu_dev); | ||
176 | + object_unref(OBJECT(sioc)); | ||
177 | + object_unref(OBJECT(server->ioc)); | ||
178 | +} | ||
179 | + | ||
180 | +static void panic_cb(VuDev *vu_dev, const char *buf) | ||
181 | +{ | ||
182 | + VuServer *server = container_of(vu_dev, VuServer, vu_dev); | ||
183 | + | ||
184 | + /* avoid while loop in close_client */ | ||
185 | + server->processing_msg = false; | ||
186 | + | ||
187 | + if (buf) { | ||
188 | + error_report("vu_panic: %s", buf); | ||
189 | + } | ||
190 | + | ||
191 | + if (server->sioc) { | ||
192 | + close_client(server); | ||
193 | + } | ||
194 | + | ||
195 | + if (server->device_panic_notifier) { | ||
196 | + server->device_panic_notifier(server); | ||
197 | + } | ||
198 | + | ||
199 | + /* | ||
200 | + * Set the callback function for network listener so another | ||
201 | + * vhost-user client can connect to this server | ||
202 | + */ | ||
203 | + qio_net_listener_set_client_func(server->listener, | ||
204 | + vu_accept, | ||
205 | + server, | ||
206 | + NULL); | ||
207 | +} | ||
208 | + | ||
209 | +static bool coroutine_fn | ||
210 | +vu_message_read(VuDev *vu_dev, int conn_fd, VhostUserMsg *vmsg) | ||
211 | +{ | ||
212 | + struct iovec iov = { | ||
213 | + .iov_base = (char *)vmsg, | ||
214 | + .iov_len = VHOST_USER_HDR_SIZE, | ||
215 | + }; | ||
216 | + int rc, read_bytes = 0; | ||
217 | + Error *local_err = NULL; | ||
218 | + /* | ||
219 | + * Store fds/nfds returned from qio_channel_readv_full into | ||
220 | + * temporary variables. | ||
221 | + * | ||
222 | + * VhostUserMsg is a packed structure, gcc will complain about passing | ||
223 | + * pointer to a packed structure member if we pass &VhostUserMsg.fd_num | ||
224 | + * and &VhostUserMsg.fds directly when calling qio_channel_readv_full, | ||
225 | + * thus two temporary variables nfds and fds are used here. | ||
226 | + */ | ||
227 | + size_t nfds = 0, nfds_t = 0; | ||
228 | + const size_t max_fds = G_N_ELEMENTS(vmsg->fds); | ||
229 | + int *fds_t = NULL; | ||
230 | + VuServer *server = container_of(vu_dev, VuServer, vu_dev); | ||
231 | + QIOChannel *ioc = server->ioc; | ||
232 | + | ||
233 | + if (!ioc) { | ||
234 | + error_report_err(local_err); | ||
235 | + goto fail; | ||
236 | + } | ||
237 | + | ||
238 | + assert(qemu_in_coroutine()); | ||
239 | + do { | ||
240 | + /* | ||
241 | + * qio_channel_readv_full may have short reads, keeping calling it | ||
242 | + * until getting VHOST_USER_HDR_SIZE or 0 bytes in total | ||
243 | + */ | ||
244 | + rc = qio_channel_readv_full(ioc, &iov, 1, &fds_t, &nfds_t, &local_err); | ||
245 | + if (rc < 0) { | ||
246 | + if (rc == QIO_CHANNEL_ERR_BLOCK) { | ||
247 | + qio_channel_yield(ioc, G_IO_IN); | ||
248 | + continue; | ||
249 | + } else { | ||
250 | + error_report_err(local_err); | ||
251 | + return false; | ||
252 | + } | ||
253 | + } | ||
254 | + read_bytes += rc; | ||
255 | + if (nfds_t > 0) { | ||
256 | + if (nfds + nfds_t > max_fds) { | ||
257 | + error_report("A maximum of %zu fds are allowed, " | ||
258 | + "however got %zu fds now", | ||
259 | + max_fds, nfds + nfds_t); | ||
260 | + goto fail; | ||
261 | + } | ||
262 | + memcpy(vmsg->fds + nfds, fds_t, | ||
263 | + nfds_t *sizeof(vmsg->fds[0])); | ||
264 | + nfds += nfds_t; | ||
265 | + g_free(fds_t); | ||
266 | + } | ||
267 | + if (read_bytes == VHOST_USER_HDR_SIZE || rc == 0) { | ||
268 | + break; | ||
269 | + } | ||
270 | + iov.iov_base = (char *)vmsg + read_bytes; | ||
271 | + iov.iov_len = VHOST_USER_HDR_SIZE - read_bytes; | ||
272 | + } while (true); | ||
273 | + | ||
274 | + vmsg->fd_num = nfds; | ||
275 | + /* qio_channel_readv_full will make socket fds blocking, unblock them */ | ||
276 | + vmsg_unblock_fds(vmsg); | ||
277 | + if (vmsg->size > sizeof(vmsg->payload)) { | ||
278 | + error_report("Error: too big message request: %d, " | ||
279 | + "size: vmsg->size: %u, " | ||
280 | + "while sizeof(vmsg->payload) = %zu", | ||
281 | + vmsg->request, vmsg->size, sizeof(vmsg->payload)); | ||
282 | + goto fail; | ||
283 | + } | ||
284 | + | ||
285 | + struct iovec iov_payload = { | ||
286 | + .iov_base = (char *)&vmsg->payload, | ||
287 | + .iov_len = vmsg->size, | ||
288 | + }; | ||
289 | + if (vmsg->size) { | ||
290 | + rc = qio_channel_readv_all_eof(ioc, &iov_payload, 1, &local_err); | ||
291 | + if (rc == -1) { | ||
292 | + error_report_err(local_err); | ||
293 | + goto fail; | ||
294 | + } | ||
295 | + } | ||
296 | + | ||
297 | + return true; | ||
298 | + | ||
299 | +fail: | ||
300 | + vmsg_close_fds(vmsg); | ||
301 | + | ||
302 | + return false; | ||
303 | +} | ||
304 | + | ||
305 | + | ||
306 | +static void vu_client_start(VuServer *server); | ||
307 | +static coroutine_fn void vu_client_trip(void *opaque) | ||
308 | +{ | ||
309 | + VuServer *server = opaque; | ||
310 | + | ||
311 | + while (!server->aio_context_changed && server->sioc) { | ||
312 | + server->processing_msg = true; | ||
313 | + vu_dispatch(&server->vu_dev); | ||
314 | + server->processing_msg = false; | ||
315 | + } | ||
316 | + | ||
317 | + if (server->aio_context_changed && server->sioc) { | ||
318 | + server->aio_context_changed = false; | ||
319 | + vu_client_start(server); | ||
320 | + } | ||
321 | +} | ||
322 | + | ||
323 | +static void vu_client_start(VuServer *server) | ||
324 | +{ | ||
325 | + server->co_trip = qemu_coroutine_create(vu_client_trip, server); | ||
326 | + aio_co_enter(server->ctx, server->co_trip); | ||
327 | +} | ||
328 | + | ||
329 | +/* | ||
330 | + * a wrapper for vu_kick_cb | ||
331 | + * | ||
332 | + * since aio_dispatch can only pass one user data pointer to the | ||
333 | + * callback function, pack VuDev and pvt into a struct. Then unpack it | ||
334 | + * and pass them to vu_kick_cb | ||
335 | + */ | ||
336 | +static void kick_handler(void *opaque) | ||
337 | +{ | ||
338 | + VuFdWatch *vu_fd_watch = opaque; | ||
339 | + vu_fd_watch->processing = true; | ||
340 | + vu_fd_watch->cb(vu_fd_watch->vu_dev, 0, vu_fd_watch->pvt); | ||
341 | + vu_fd_watch->processing = false; | ||
342 | +} | ||
343 | + | ||
344 | + | ||
345 | +static VuFdWatch *find_vu_fd_watch(VuServer *server, int fd) | ||
346 | +{ | ||
347 | + | ||
348 | + VuFdWatch *vu_fd_watch, *next; | ||
349 | + QTAILQ_FOREACH_SAFE(vu_fd_watch, &server->vu_fd_watches, next, next) { | ||
350 | + if (vu_fd_watch->fd == fd) { | ||
351 | + return vu_fd_watch; | ||
352 | + } | ||
353 | + } | ||
354 | + return NULL; | ||
355 | +} | ||
356 | + | ||
357 | +static void | ||
358 | +set_watch(VuDev *vu_dev, int fd, int vu_evt, | ||
359 | + vu_watch_cb cb, void *pvt) | ||
360 | +{ | ||
361 | + | ||
362 | + VuServer *server = container_of(vu_dev, VuServer, vu_dev); | ||
363 | + g_assert(vu_dev); | ||
364 | + g_assert(fd >= 0); | ||
365 | + g_assert(cb); | ||
366 | + | ||
367 | + VuFdWatch *vu_fd_watch = find_vu_fd_watch(server, fd); | ||
368 | + | ||
369 | + if (!vu_fd_watch) { | ||
370 | + VuFdWatch *vu_fd_watch = g_new0(VuFdWatch, 1); | ||
371 | + | ||
372 | + QTAILQ_INSERT_TAIL(&server->vu_fd_watches, vu_fd_watch, next); | ||
373 | + | ||
374 | + vu_fd_watch->fd = fd; | ||
375 | + vu_fd_watch->cb = cb; | ||
376 | + qemu_set_nonblock(fd); | ||
377 | + aio_set_fd_handler(server->ioc->ctx, fd, true, kick_handler, | ||
378 | + NULL, NULL, vu_fd_watch); | ||
379 | + vu_fd_watch->vu_dev = vu_dev; | ||
380 | + vu_fd_watch->pvt = pvt; | ||
381 | + } | ||
382 | +} | ||
383 | + | ||
384 | + | ||
385 | +static void remove_watch(VuDev *vu_dev, int fd) | ||
386 | +{ | ||
387 | + VuServer *server; | ||
388 | + g_assert(vu_dev); | ||
389 | + g_assert(fd >= 0); | ||
390 | + | ||
391 | + server = container_of(vu_dev, VuServer, vu_dev); | ||
392 | + | ||
393 | + VuFdWatch *vu_fd_watch = find_vu_fd_watch(server, fd); | ||
394 | + | ||
395 | + if (!vu_fd_watch) { | ||
396 | + return; | ||
397 | + } | ||
398 | + aio_set_fd_handler(server->ioc->ctx, fd, true, NULL, NULL, NULL, NULL); | ||
399 | + | ||
400 | + QTAILQ_REMOVE(&server->vu_fd_watches, vu_fd_watch, next); | ||
401 | + g_free(vu_fd_watch); | ||
402 | +} | ||
403 | + | ||
404 | + | ||
405 | +static void vu_accept(QIONetListener *listener, QIOChannelSocket *sioc, | ||
406 | + gpointer opaque) | ||
407 | +{ | ||
408 | + VuServer *server = opaque; | ||
409 | + | ||
410 | + if (server->sioc) { | ||
411 | + warn_report("Only one vhost-user client is allowed to " | ||
412 | + "connect the server one time"); | ||
413 | + return; | ||
414 | + } | ||
415 | + | ||
416 | + if (!vu_init(&server->vu_dev, server->max_queues, sioc->fd, panic_cb, | ||
417 | + vu_message_read, set_watch, remove_watch, server->vu_iface)) { | ||
418 | + error_report("Failed to initialize libvhost-user"); | ||
419 | + return; | ||
420 | + } | ||
421 | + | ||
422 | + /* | ||
423 | + * Unset the callback function for network listener to make another | ||
424 | + * vhost-user client keeping waiting until this client disconnects | ||
425 | + */ | ||
426 | + qio_net_listener_set_client_func(server->listener, | ||
427 | + NULL, | ||
428 | + NULL, | ||
429 | + NULL); | ||
430 | + server->sioc = sioc; | ||
431 | + /* | ||
432 | + * Increase the object reference, so sioc will not freed by | ||
433 | + * qio_net_listener_channel_func which will call object_unref(OBJECT(sioc)) | ||
434 | + */ | ||
435 | + object_ref(OBJECT(server->sioc)); | ||
436 | + qio_channel_set_name(QIO_CHANNEL(sioc), "vhost-user client"); | ||
437 | + server->ioc = QIO_CHANNEL(sioc); | ||
438 | + object_ref(OBJECT(server->ioc)); | ||
439 | + qio_channel_attach_aio_context(server->ioc, server->ctx); | ||
440 | + qio_channel_set_blocking(QIO_CHANNEL(server->sioc), false, NULL); | ||
441 | + vu_client_start(server); | ||
442 | +} | ||
443 | + | ||
444 | + | ||
445 | +void vhost_user_server_stop(VuServer *server) | ||
446 | +{ | ||
447 | + if (server->sioc) { | ||
448 | + close_client(server); | ||
449 | + } | ||
450 | + | ||
451 | + if (server->listener) { | ||
452 | + qio_net_listener_disconnect(server->listener); | ||
453 | + object_unref(OBJECT(server->listener)); | ||
454 | + } | ||
455 | + | ||
456 | +} | ||
457 | + | ||
458 | +void vhost_user_server_set_aio_context(VuServer *server, AioContext *ctx) | ||
459 | +{ | ||
460 | + VuFdWatch *vu_fd_watch, *next; | ||
461 | + void *opaque = NULL; | ||
462 | + IOHandler *io_read = NULL; | ||
463 | + bool attach; | ||
464 | + | ||
465 | + server->ctx = ctx ? ctx : qemu_get_aio_context(); | ||
466 | + | ||
467 | + if (!server->sioc) { | ||
468 | + /* not yet serving any client*/ | ||
469 | + return; | ||
470 | + } | ||
471 | + | ||
472 | + if (ctx) { | ||
473 | + qio_channel_attach_aio_context(server->ioc, ctx); | ||
474 | + server->aio_context_changed = true; | ||
475 | + io_read = kick_handler; | ||
476 | + attach = true; | ||
477 | + } else { | ||
478 | + qio_channel_detach_aio_context(server->ioc); | ||
479 | + /* server->ioc->ctx keeps the old AioConext */ | ||
480 | + ctx = server->ioc->ctx; | ||
481 | + attach = false; | ||
482 | + } | ||
483 | + | ||
484 | + QTAILQ_FOREACH_SAFE(vu_fd_watch, &server->vu_fd_watches, next, next) { | ||
485 | + if (vu_fd_watch->cb) { | ||
486 | + opaque = attach ? vu_fd_watch : NULL; | ||
487 | + aio_set_fd_handler(ctx, vu_fd_watch->fd, true, | ||
488 | + io_read, NULL, NULL, | ||
489 | + opaque); | ||
490 | + } | ||
491 | + } | ||
492 | +} | ||
493 | + | ||
494 | + | ||
495 | +bool vhost_user_server_start(VuServer *server, | ||
496 | + SocketAddress *socket_addr, | ||
497 | + AioContext *ctx, | ||
498 | + uint16_t max_queues, | ||
499 | + DevicePanicNotifierFn *device_panic_notifier, | ||
500 | + const VuDevIface *vu_iface, | ||
501 | + Error **errp) | ||
502 | +{ | ||
503 | + QIONetListener *listener = qio_net_listener_new(); | ||
504 | + if (qio_net_listener_open_sync(listener, socket_addr, 1, | ||
505 | + errp) < 0) { | ||
506 | + object_unref(OBJECT(listener)); | ||
507 | + return false; | ||
508 | + } | ||
509 | + | ||
510 | + /* zero out unspecified fileds */ | ||
511 | + *server = (VuServer) { | ||
512 | + .listener = listener, | ||
513 | + .vu_iface = vu_iface, | ||
514 | + .max_queues = max_queues, | ||
515 | + .ctx = ctx, | ||
516 | + .device_panic_notifier = device_panic_notifier, | ||
517 | + }; | ||
518 | + | ||
519 | + qio_net_listener_set_name(server->listener, "vhost-user-backend-listener"); | ||
520 | + | ||
521 | + qio_net_listener_set_client_func(server->listener, | ||
522 | + vu_accept, | ||
523 | + server, | ||
524 | + NULL); | ||
525 | + | ||
526 | + QTAILQ_INIT(&server->vu_fd_watches); | ||
527 | + return true; | ||
528 | +} | ||
529 | diff --git a/util/meson.build b/util/meson.build | ||
530 | index XXXXXXX..XXXXXXX 100644 | ||
531 | --- a/util/meson.build | ||
532 | +++ b/util/meson.build | ||
533 | @@ -XXX,XX +XXX,XX @@ if have_block | ||
534 | util_ss.add(files('main-loop.c')) | ||
535 | util_ss.add(files('nvdimm-utils.c')) | ||
536 | util_ss.add(files('qemu-coroutine.c', 'qemu-coroutine-lock.c', 'qemu-coroutine-io.c')) | ||
537 | + util_ss.add(when: 'CONFIG_LINUX', if_true: files('vhost-user-server.c')) | ||
538 | util_ss.add(files('qemu-coroutine-sleep.c')) | ||
539 | util_ss.add(files('qemu-co-shared-resource.c')) | ||
540 | util_ss.add(files('thread-pool.c', 'qemu-timer.c')) | ||
541 | -- | ||
542 | 2.26.2 | ||
543 | diff view generated by jsdifflib |
1 | From: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | 1 | From: Coiby Xu <coiby.xu@gmail.com> |
---|---|---|---|
2 | 2 | ||
3 | We will add more and more checks so we need a better code structure | 3 | Move the constants from hw/core/qdev-properties.c to |
4 | in parallels_co_check. Let each check performs in a separate loop | 4 | util/block-helpers.h so that knowledge of the min/max values is |
5 | in a separate helper. | ||
6 | 5 | ||
7 | Signed-off-by: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | 6 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
8 | Reviewed-by: Denis V. Lunev <den@openvz.org> | 7 | Signed-off-by: Coiby Xu <coiby.xu@gmail.com> |
9 | Reviewed-by: Vladimir Sementsov-Ogievskiy <vsementsov@yandex-team.ru> | 8 | Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> |
10 | Message-Id: <20230424093147.197643-7-alexander.ivanov@virtuozzo.com> | 9 | Reviewed-by: Marc-André Lureau <marcandre.lureau@redhat.com> |
11 | Reviewed-by: Hanna Czenczek <hreitz@redhat.com> | 10 | Acked-by: Eduardo Habkost <ehabkost@redhat.com> |
12 | Signed-off-by: Hanna Czenczek <hreitz@redhat.com> | 11 | Message-id: 20200918080912.321299-5-coiby.xu@gmail.com |
12 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
13 | --- | 13 | --- |
14 | block/parallels.c | 31 +++++++++++++++++++++---------- | 14 | util/block-helpers.h | 19 +++++++++++++ |
15 | 1 file changed, 21 insertions(+), 10 deletions(-) | 15 | hw/core/qdev-properties-system.c | 31 ++++----------------- |
16 | util/block-helpers.c | 46 ++++++++++++++++++++++++++++++++ | ||
17 | util/meson.build | 1 + | ||
18 | 4 files changed, 71 insertions(+), 26 deletions(-) | ||
19 | create mode 100644 util/block-helpers.h | ||
20 | create mode 100644 util/block-helpers.c | ||
16 | 21 | ||
17 | diff --git a/block/parallels.c b/block/parallels.c | 22 | diff --git a/util/block-helpers.h b/util/block-helpers.h |
23 | new file mode 100644 | ||
24 | index XXXXXXX..XXXXXXX | ||
25 | --- /dev/null | ||
26 | +++ b/util/block-helpers.h | ||
27 | @@ -XXX,XX +XXX,XX @@ | ||
28 | +#ifndef BLOCK_HELPERS_H | ||
29 | +#define BLOCK_HELPERS_H | ||
30 | + | ||
31 | +#include "qemu/units.h" | ||
32 | + | ||
33 | +/* lower limit is sector size */ | ||
34 | +#define MIN_BLOCK_SIZE INT64_C(512) | ||
35 | +#define MIN_BLOCK_SIZE_STR "512 B" | ||
36 | +/* | ||
37 | + * upper limit is arbitrary, 2 MiB looks sufficient for all sensible uses, and | ||
38 | + * matches qcow2 cluster size limit | ||
39 | + */ | ||
40 | +#define MAX_BLOCK_SIZE (2 * MiB) | ||
41 | +#define MAX_BLOCK_SIZE_STR "2 MiB" | ||
42 | + | ||
43 | +void check_block_size(const char *id, const char *name, int64_t value, | ||
44 | + Error **errp); | ||
45 | + | ||
46 | +#endif /* BLOCK_HELPERS_H */ | ||
47 | diff --git a/hw/core/qdev-properties-system.c b/hw/core/qdev-properties-system.c | ||
18 | index XXXXXXX..XXXXXXX 100644 | 48 | index XXXXXXX..XXXXXXX 100644 |
19 | --- a/block/parallels.c | 49 | --- a/hw/core/qdev-properties-system.c |
20 | +++ b/block/parallels.c | 50 | +++ b/hw/core/qdev-properties-system.c |
21 | @@ -XXX,XX +XXX,XX @@ parallels_co_readv(BlockDriverState *bs, int64_t sector_num, int nb_sectors, | 51 | @@ -XXX,XX +XXX,XX @@ |
22 | return ret; | 52 | #include "sysemu/blockdev.h" |
53 | #include "net/net.h" | ||
54 | #include "hw/pci/pci.h" | ||
55 | +#include "util/block-helpers.h" | ||
56 | |||
57 | static bool check_prop_still_unset(DeviceState *dev, const char *name, | ||
58 | const void *old_val, const char *new_val, | ||
59 | @@ -XXX,XX +XXX,XX @@ const PropertyInfo qdev_prop_losttickpolicy = { | ||
60 | |||
61 | /* --- blocksize --- */ | ||
62 | |||
63 | -/* lower limit is sector size */ | ||
64 | -#define MIN_BLOCK_SIZE 512 | ||
65 | -#define MIN_BLOCK_SIZE_STR "512 B" | ||
66 | -/* | ||
67 | - * upper limit is arbitrary, 2 MiB looks sufficient for all sensible uses, and | ||
68 | - * matches qcow2 cluster size limit | ||
69 | - */ | ||
70 | -#define MAX_BLOCK_SIZE (2 * MiB) | ||
71 | -#define MAX_BLOCK_SIZE_STR "2 MiB" | ||
72 | - | ||
73 | static void set_blocksize(Object *obj, Visitor *v, const char *name, | ||
74 | void *opaque, Error **errp) | ||
75 | { | ||
76 | @@ -XXX,XX +XXX,XX @@ static void set_blocksize(Object *obj, Visitor *v, const char *name, | ||
77 | Property *prop = opaque; | ||
78 | uint32_t *ptr = qdev_get_prop_ptr(dev, prop); | ||
79 | uint64_t value; | ||
80 | + Error *local_err = NULL; | ||
81 | |||
82 | if (dev->realized) { | ||
83 | qdev_prop_set_after_realize(dev, name, errp); | ||
84 | @@ -XXX,XX +XXX,XX @@ static void set_blocksize(Object *obj, Visitor *v, const char *name, | ||
85 | if (!visit_type_size(v, name, &value, errp)) { | ||
86 | return; | ||
87 | } | ||
88 | - /* value of 0 means "unset" */ | ||
89 | - if (value && (value < MIN_BLOCK_SIZE || value > MAX_BLOCK_SIZE)) { | ||
90 | - error_setg(errp, | ||
91 | - "Property %s.%s doesn't take value %" PRIu64 | ||
92 | - " (minimum: " MIN_BLOCK_SIZE_STR | ||
93 | - ", maximum: " MAX_BLOCK_SIZE_STR ")", | ||
94 | - dev->id ? : "", name, value); | ||
95 | + check_block_size(dev->id ? : "", name, value, &local_err); | ||
96 | + if (local_err) { | ||
97 | + error_propagate(errp, local_err); | ||
98 | return; | ||
99 | } | ||
100 | - | ||
101 | - /* We rely on power-of-2 blocksizes for bitmasks */ | ||
102 | - if ((value & (value - 1)) != 0) { | ||
103 | - error_setg(errp, | ||
104 | - "Property %s.%s doesn't take value '%" PRId64 "', " | ||
105 | - "it's not a power of 2", dev->id ?: "", name, (int64_t)value); | ||
106 | - return; | ||
107 | - } | ||
108 | - | ||
109 | *ptr = value; | ||
23 | } | 110 | } |
24 | 111 | ||
25 | +static void parallels_check_unclean(BlockDriverState *bs, | 112 | diff --git a/util/block-helpers.c b/util/block-helpers.c |
26 | + BdrvCheckResult *res, | 113 | new file mode 100644 |
27 | + BdrvCheckMode fix) | 114 | index XXXXXXX..XXXXXXX |
115 | --- /dev/null | ||
116 | +++ b/util/block-helpers.c | ||
117 | @@ -XXX,XX +XXX,XX @@ | ||
118 | +/* | ||
119 | + * Block utility functions | ||
120 | + * | ||
121 | + * Copyright IBM, Corp. 2011 | ||
122 | + * Copyright (c) 2020 Coiby Xu <coiby.xu@gmail.com> | ||
123 | + * | ||
124 | + * This work is licensed under the terms of the GNU GPL, version 2 or later. | ||
125 | + * See the COPYING file in the top-level directory. | ||
126 | + */ | ||
127 | + | ||
128 | +#include "qemu/osdep.h" | ||
129 | +#include "qapi/error.h" | ||
130 | +#include "qapi/qmp/qerror.h" | ||
131 | +#include "block-helpers.h" | ||
132 | + | ||
133 | +/** | ||
134 | + * check_block_size: | ||
135 | + * @id: The unique ID of the object | ||
136 | + * @name: The name of the property being validated | ||
137 | + * @value: The block size in bytes | ||
138 | + * @errp: A pointer to an area to store an error | ||
139 | + * | ||
140 | + * This function checks that the block size meets the following conditions: | ||
141 | + * 1. At least MIN_BLOCK_SIZE | ||
142 | + * 2. No larger than MAX_BLOCK_SIZE | ||
143 | + * 3. A power of 2 | ||
144 | + */ | ||
145 | +void check_block_size(const char *id, const char *name, int64_t value, | ||
146 | + Error **errp) | ||
28 | +{ | 147 | +{ |
29 | + BDRVParallelsState *s = bs->opaque; | 148 | + /* value of 0 means "unset" */ |
30 | + | 149 | + if (value && (value < MIN_BLOCK_SIZE || value > MAX_BLOCK_SIZE)) { |
31 | + if (!s->header_unclean) { | 150 | + error_setg(errp, QERR_PROPERTY_VALUE_OUT_OF_RANGE, |
151 | + id, name, value, MIN_BLOCK_SIZE, MAX_BLOCK_SIZE); | ||
32 | + return; | 152 | + return; |
33 | + } | 153 | + } |
34 | + | 154 | + |
35 | + fprintf(stderr, "%s image was not closed correctly\n", | 155 | + /* We rely on power-of-2 blocksizes for bitmasks */ |
36 | + fix & BDRV_FIX_ERRORS ? "Repairing" : "ERROR"); | 156 | + if ((value & (value - 1)) != 0) { |
37 | + res->corruptions++; | 157 | + error_setg(errp, |
38 | + if (fix & BDRV_FIX_ERRORS) { | 158 | + "Property %s.%s doesn't take value '%" PRId64 |
39 | + /* parallels_close will do the job right */ | 159 | + "', it's not a power of 2", |
40 | + res->corruptions_fixed++; | 160 | + id, name, value); |
41 | + s->header_unclean = false; | 161 | + return; |
42 | + } | 162 | + } |
43 | +} | 163 | +} |
44 | 164 | diff --git a/util/meson.build b/util/meson.build | |
45 | static int coroutine_fn GRAPH_RDLOCK | 165 | index XXXXXXX..XXXXXXX 100644 |
46 | parallels_co_check(BlockDriverState *bs, BdrvCheckResult *res, | 166 | --- a/util/meson.build |
47 | @@ -XXX,XX +XXX,XX @@ parallels_co_check(BlockDriverState *bs, BdrvCheckResult *res, | 167 | +++ b/util/meson.build |
48 | } | 168 | @@ -XXX,XX +XXX,XX @@ if have_block |
49 | 169 | util_ss.add(files('nvdimm-utils.c')) | |
50 | qemu_co_mutex_lock(&s->lock); | 170 | util_ss.add(files('qemu-coroutine.c', 'qemu-coroutine-lock.c', 'qemu-coroutine-io.c')) |
51 | - if (s->header_unclean) { | 171 | util_ss.add(when: 'CONFIG_LINUX', if_true: files('vhost-user-server.c')) |
52 | - fprintf(stderr, "%s image was not closed correctly\n", | 172 | + util_ss.add(files('block-helpers.c')) |
53 | - fix & BDRV_FIX_ERRORS ? "Repairing" : "ERROR"); | 173 | util_ss.add(files('qemu-coroutine-sleep.c')) |
54 | - res->corruptions++; | 174 | util_ss.add(files('qemu-co-shared-resource.c')) |
55 | - if (fix & BDRV_FIX_ERRORS) { | 175 | util_ss.add(files('thread-pool.c', 'qemu-timer.c')) |
56 | - /* parallels_close will do the job right */ | ||
57 | - res->corruptions_fixed++; | ||
58 | - s->header_unclean = false; | ||
59 | - } | ||
60 | - } | ||
61 | + | ||
62 | + parallels_check_unclean(bs, res, fix); | ||
63 | |||
64 | res->bfi.total_clusters = s->bat_size; | ||
65 | res->bfi.compressed_clusters = 0; /* compression is not supported */ | ||
66 | -- | 176 | -- |
67 | 2.40.1 | 177 | 2.26.2 |
178 | diff view generated by jsdifflib |
New patch | |||
---|---|---|---|
1 | From: Coiby Xu <coiby.xu@gmail.com> | ||
1 | 2 | ||
3 | By making use of libvhost-user, block device drive can be shared to | ||
4 | the connected vhost-user client. Only one client can connect to the | ||
5 | server one time. | ||
6 | |||
7 | Since vhost-user-server needs a block drive to be created first, delay | ||
8 | the creation of this object. | ||
9 | |||
10 | Suggested-by: Kevin Wolf <kwolf@redhat.com> | ||
11 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
12 | Signed-off-by: Coiby Xu <coiby.xu@gmail.com> | ||
13 | Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
14 | Reviewed-by: Marc-André Lureau <marcandre.lureau@redhat.com> | ||
15 | Message-id: 20200918080912.321299-6-coiby.xu@gmail.com | ||
16 | [Shorten "vhost_user_blk_server" string to "vhost_user_blk" to avoid the | ||
17 | following compiler warning: | ||
18 | ../block/export/vhost-user-blk-server.c:178:50: error: ‘%s’ directive output truncated writing 21 bytes into a region of size 20 [-Werror=format-truncation=] | ||
19 | and fix "Invalid size %ld ..." ssize_t format string arguments for | ||
20 | 32-bit hosts. | ||
21 | --Stefan] | ||
22 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
23 | --- | ||
24 | block/export/vhost-user-blk-server.h | 36 ++ | ||
25 | block/export/vhost-user-blk-server.c | 661 +++++++++++++++++++++++++++ | ||
26 | softmmu/vl.c | 4 + | ||
27 | block/meson.build | 1 + | ||
28 | 4 files changed, 702 insertions(+) | ||
29 | create mode 100644 block/export/vhost-user-blk-server.h | ||
30 | create mode 100644 block/export/vhost-user-blk-server.c | ||
31 | |||
32 | diff --git a/block/export/vhost-user-blk-server.h b/block/export/vhost-user-blk-server.h | ||
33 | new file mode 100644 | ||
34 | index XXXXXXX..XXXXXXX | ||
35 | --- /dev/null | ||
36 | +++ b/block/export/vhost-user-blk-server.h | ||
37 | @@ -XXX,XX +XXX,XX @@ | ||
38 | +/* | ||
39 | + * Sharing QEMU block devices via vhost-user protocal | ||
40 | + * | ||
41 | + * Copyright (c) Coiby Xu <coiby.xu@gmail.com>. | ||
42 | + * Copyright (c) 2020 Red Hat, Inc. | ||
43 | + * | ||
44 | + * This work is licensed under the terms of the GNU GPL, version 2 or | ||
45 | + * later. See the COPYING file in the top-level directory. | ||
46 | + */ | ||
47 | + | ||
48 | +#ifndef VHOST_USER_BLK_SERVER_H | ||
49 | +#define VHOST_USER_BLK_SERVER_H | ||
50 | +#include "util/vhost-user-server.h" | ||
51 | + | ||
52 | +typedef struct VuBlockDev VuBlockDev; | ||
53 | +#define TYPE_VHOST_USER_BLK_SERVER "vhost-user-blk-server" | ||
54 | +#define VHOST_USER_BLK_SERVER(obj) \ | ||
55 | + OBJECT_CHECK(VuBlockDev, obj, TYPE_VHOST_USER_BLK_SERVER) | ||
56 | + | ||
57 | +/* vhost user block device */ | ||
58 | +struct VuBlockDev { | ||
59 | + Object parent_obj; | ||
60 | + char *node_name; | ||
61 | + SocketAddress *addr; | ||
62 | + AioContext *ctx; | ||
63 | + VuServer vu_server; | ||
64 | + bool running; | ||
65 | + uint32_t blk_size; | ||
66 | + BlockBackend *backend; | ||
67 | + QIOChannelSocket *sioc; | ||
68 | + QTAILQ_ENTRY(VuBlockDev) next; | ||
69 | + struct virtio_blk_config blkcfg; | ||
70 | + bool writable; | ||
71 | +}; | ||
72 | + | ||
73 | +#endif /* VHOST_USER_BLK_SERVER_H */ | ||
74 | diff --git a/block/export/vhost-user-blk-server.c b/block/export/vhost-user-blk-server.c | ||
75 | new file mode 100644 | ||
76 | index XXXXXXX..XXXXXXX | ||
77 | --- /dev/null | ||
78 | +++ b/block/export/vhost-user-blk-server.c | ||
79 | @@ -XXX,XX +XXX,XX @@ | ||
80 | +/* | ||
81 | + * Sharing QEMU block devices via vhost-user protocal | ||
82 | + * | ||
83 | + * Parts of the code based on nbd/server.c. | ||
84 | + * | ||
85 | + * Copyright (c) Coiby Xu <coiby.xu@gmail.com>. | ||
86 | + * Copyright (c) 2020 Red Hat, Inc. | ||
87 | + * | ||
88 | + * This work is licensed under the terms of the GNU GPL, version 2 or | ||
89 | + * later. See the COPYING file in the top-level directory. | ||
90 | + */ | ||
91 | +#include "qemu/osdep.h" | ||
92 | +#include "block/block.h" | ||
93 | +#include "vhost-user-blk-server.h" | ||
94 | +#include "qapi/error.h" | ||
95 | +#include "qom/object_interfaces.h" | ||
96 | +#include "sysemu/block-backend.h" | ||
97 | +#include "util/block-helpers.h" | ||
98 | + | ||
99 | +enum { | ||
100 | + VHOST_USER_BLK_MAX_QUEUES = 1, | ||
101 | +}; | ||
102 | +struct virtio_blk_inhdr { | ||
103 | + unsigned char status; | ||
104 | +}; | ||
105 | + | ||
106 | +typedef struct VuBlockReq { | ||
107 | + VuVirtqElement *elem; | ||
108 | + int64_t sector_num; | ||
109 | + size_t size; | ||
110 | + struct virtio_blk_inhdr *in; | ||
111 | + struct virtio_blk_outhdr out; | ||
112 | + VuServer *server; | ||
113 | + struct VuVirtq *vq; | ||
114 | +} VuBlockReq; | ||
115 | + | ||
116 | +static void vu_block_req_complete(VuBlockReq *req) | ||
117 | +{ | ||
118 | + VuDev *vu_dev = &req->server->vu_dev; | ||
119 | + | ||
120 | + /* IO size with 1 extra status byte */ | ||
121 | + vu_queue_push(vu_dev, req->vq, req->elem, req->size + 1); | ||
122 | + vu_queue_notify(vu_dev, req->vq); | ||
123 | + | ||
124 | + if (req->elem) { | ||
125 | + free(req->elem); | ||
126 | + } | ||
127 | + | ||
128 | + g_free(req); | ||
129 | +} | ||
130 | + | ||
131 | +static VuBlockDev *get_vu_block_device_by_server(VuServer *server) | ||
132 | +{ | ||
133 | + return container_of(server, VuBlockDev, vu_server); | ||
134 | +} | ||
135 | + | ||
136 | +static int coroutine_fn | ||
137 | +vu_block_discard_write_zeroes(VuBlockReq *req, struct iovec *iov, | ||
138 | + uint32_t iovcnt, uint32_t type) | ||
139 | +{ | ||
140 | + struct virtio_blk_discard_write_zeroes desc; | ||
141 | + ssize_t size = iov_to_buf(iov, iovcnt, 0, &desc, sizeof(desc)); | ||
142 | + if (unlikely(size != sizeof(desc))) { | ||
143 | + error_report("Invalid size %zd, expect %zu", size, sizeof(desc)); | ||
144 | + return -EINVAL; | ||
145 | + } | ||
146 | + | ||
147 | + VuBlockDev *vdev_blk = get_vu_block_device_by_server(req->server); | ||
148 | + uint64_t range[2] = { le64_to_cpu(desc.sector) << 9, | ||
149 | + le32_to_cpu(desc.num_sectors) << 9 }; | ||
150 | + if (type == VIRTIO_BLK_T_DISCARD) { | ||
151 | + if (blk_co_pdiscard(vdev_blk->backend, range[0], range[1]) == 0) { | ||
152 | + return 0; | ||
153 | + } | ||
154 | + } else if (type == VIRTIO_BLK_T_WRITE_ZEROES) { | ||
155 | + if (blk_co_pwrite_zeroes(vdev_blk->backend, | ||
156 | + range[0], range[1], 0) == 0) { | ||
157 | + return 0; | ||
158 | + } | ||
159 | + } | ||
160 | + | ||
161 | + return -EINVAL; | ||
162 | +} | ||
163 | + | ||
164 | +static void coroutine_fn vu_block_flush(VuBlockReq *req) | ||
165 | +{ | ||
166 | + VuBlockDev *vdev_blk = get_vu_block_device_by_server(req->server); | ||
167 | + BlockBackend *backend = vdev_blk->backend; | ||
168 | + blk_co_flush(backend); | ||
169 | +} | ||
170 | + | ||
171 | +struct req_data { | ||
172 | + VuServer *server; | ||
173 | + VuVirtq *vq; | ||
174 | + VuVirtqElement *elem; | ||
175 | +}; | ||
176 | + | ||
177 | +static void coroutine_fn vu_block_virtio_process_req(void *opaque) | ||
178 | +{ | ||
179 | + struct req_data *data = opaque; | ||
180 | + VuServer *server = data->server; | ||
181 | + VuVirtq *vq = data->vq; | ||
182 | + VuVirtqElement *elem = data->elem; | ||
183 | + uint32_t type; | ||
184 | + VuBlockReq *req; | ||
185 | + | ||
186 | + VuBlockDev *vdev_blk = get_vu_block_device_by_server(server); | ||
187 | + BlockBackend *backend = vdev_blk->backend; | ||
188 | + | ||
189 | + struct iovec *in_iov = elem->in_sg; | ||
190 | + struct iovec *out_iov = elem->out_sg; | ||
191 | + unsigned in_num = elem->in_num; | ||
192 | + unsigned out_num = elem->out_num; | ||
193 | + /* refer to hw/block/virtio_blk.c */ | ||
194 | + if (elem->out_num < 1 || elem->in_num < 1) { | ||
195 | + error_report("virtio-blk request missing headers"); | ||
196 | + free(elem); | ||
197 | + return; | ||
198 | + } | ||
199 | + | ||
200 | + req = g_new0(VuBlockReq, 1); | ||
201 | + req->server = server; | ||
202 | + req->vq = vq; | ||
203 | + req->elem = elem; | ||
204 | + | ||
205 | + if (unlikely(iov_to_buf(out_iov, out_num, 0, &req->out, | ||
206 | + sizeof(req->out)) != sizeof(req->out))) { | ||
207 | + error_report("virtio-blk request outhdr too short"); | ||
208 | + goto err; | ||
209 | + } | ||
210 | + | ||
211 | + iov_discard_front(&out_iov, &out_num, sizeof(req->out)); | ||
212 | + | ||
213 | + if (in_iov[in_num - 1].iov_len < sizeof(struct virtio_blk_inhdr)) { | ||
214 | + error_report("virtio-blk request inhdr too short"); | ||
215 | + goto err; | ||
216 | + } | ||
217 | + | ||
218 | + /* We always touch the last byte, so just see how big in_iov is. */ | ||
219 | + req->in = (void *)in_iov[in_num - 1].iov_base | ||
220 | + + in_iov[in_num - 1].iov_len | ||
221 | + - sizeof(struct virtio_blk_inhdr); | ||
222 | + iov_discard_back(in_iov, &in_num, sizeof(struct virtio_blk_inhdr)); | ||
223 | + | ||
224 | + type = le32_to_cpu(req->out.type); | ||
225 | + switch (type & ~VIRTIO_BLK_T_BARRIER) { | ||
226 | + case VIRTIO_BLK_T_IN: | ||
227 | + case VIRTIO_BLK_T_OUT: { | ||
228 | + ssize_t ret = 0; | ||
229 | + bool is_write = type & VIRTIO_BLK_T_OUT; | ||
230 | + req->sector_num = le64_to_cpu(req->out.sector); | ||
231 | + | ||
232 | + int64_t offset = req->sector_num * vdev_blk->blk_size; | ||
233 | + QEMUIOVector qiov; | ||
234 | + if (is_write) { | ||
235 | + qemu_iovec_init_external(&qiov, out_iov, out_num); | ||
236 | + ret = blk_co_pwritev(backend, offset, qiov.size, | ||
237 | + &qiov, 0); | ||
238 | + } else { | ||
239 | + qemu_iovec_init_external(&qiov, in_iov, in_num); | ||
240 | + ret = blk_co_preadv(backend, offset, qiov.size, | ||
241 | + &qiov, 0); | ||
242 | + } | ||
243 | + if (ret >= 0) { | ||
244 | + req->in->status = VIRTIO_BLK_S_OK; | ||
245 | + } else { | ||
246 | + req->in->status = VIRTIO_BLK_S_IOERR; | ||
247 | + } | ||
248 | + break; | ||
249 | + } | ||
250 | + case VIRTIO_BLK_T_FLUSH: | ||
251 | + vu_block_flush(req); | ||
252 | + req->in->status = VIRTIO_BLK_S_OK; | ||
253 | + break; | ||
254 | + case VIRTIO_BLK_T_GET_ID: { | ||
255 | + size_t size = MIN(iov_size(&elem->in_sg[0], in_num), | ||
256 | + VIRTIO_BLK_ID_BYTES); | ||
257 | + snprintf(elem->in_sg[0].iov_base, size, "%s", "vhost_user_blk"); | ||
258 | + req->in->status = VIRTIO_BLK_S_OK; | ||
259 | + req->size = elem->in_sg[0].iov_len; | ||
260 | + break; | ||
261 | + } | ||
262 | + case VIRTIO_BLK_T_DISCARD: | ||
263 | + case VIRTIO_BLK_T_WRITE_ZEROES: { | ||
264 | + int rc; | ||
265 | + rc = vu_block_discard_write_zeroes(req, &elem->out_sg[1], | ||
266 | + out_num, type); | ||
267 | + if (rc == 0) { | ||
268 | + req->in->status = VIRTIO_BLK_S_OK; | ||
269 | + } else { | ||
270 | + req->in->status = VIRTIO_BLK_S_IOERR; | ||
271 | + } | ||
272 | + break; | ||
273 | + } | ||
274 | + default: | ||
275 | + req->in->status = VIRTIO_BLK_S_UNSUPP; | ||
276 | + break; | ||
277 | + } | ||
278 | + | ||
279 | + vu_block_req_complete(req); | ||
280 | + return; | ||
281 | + | ||
282 | +err: | ||
283 | + free(elem); | ||
284 | + g_free(req); | ||
285 | + return; | ||
286 | +} | ||
287 | + | ||
288 | +static void vu_block_process_vq(VuDev *vu_dev, int idx) | ||
289 | +{ | ||
290 | + VuServer *server; | ||
291 | + VuVirtq *vq; | ||
292 | + struct req_data *req_data; | ||
293 | + | ||
294 | + server = container_of(vu_dev, VuServer, vu_dev); | ||
295 | + assert(server); | ||
296 | + | ||
297 | + vq = vu_get_queue(vu_dev, idx); | ||
298 | + assert(vq); | ||
299 | + VuVirtqElement *elem; | ||
300 | + while (1) { | ||
301 | + elem = vu_queue_pop(vu_dev, vq, sizeof(VuVirtqElement) + | ||
302 | + sizeof(VuBlockReq)); | ||
303 | + if (elem) { | ||
304 | + req_data = g_new0(struct req_data, 1); | ||
305 | + req_data->server = server; | ||
306 | + req_data->vq = vq; | ||
307 | + req_data->elem = elem; | ||
308 | + Coroutine *co = qemu_coroutine_create(vu_block_virtio_process_req, | ||
309 | + req_data); | ||
310 | + aio_co_enter(server->ioc->ctx, co); | ||
311 | + } else { | ||
312 | + break; | ||
313 | + } | ||
314 | + } | ||
315 | +} | ||
316 | + | ||
317 | +static void vu_block_queue_set_started(VuDev *vu_dev, int idx, bool started) | ||
318 | +{ | ||
319 | + VuVirtq *vq; | ||
320 | + | ||
321 | + assert(vu_dev); | ||
322 | + | ||
323 | + vq = vu_get_queue(vu_dev, idx); | ||
324 | + vu_set_queue_handler(vu_dev, vq, started ? vu_block_process_vq : NULL); | ||
325 | +} | ||
326 | + | ||
327 | +static uint64_t vu_block_get_features(VuDev *dev) | ||
328 | +{ | ||
329 | + uint64_t features; | ||
330 | + VuServer *server = container_of(dev, VuServer, vu_dev); | ||
331 | + VuBlockDev *vdev_blk = get_vu_block_device_by_server(server); | ||
332 | + features = 1ull << VIRTIO_BLK_F_SIZE_MAX | | ||
333 | + 1ull << VIRTIO_BLK_F_SEG_MAX | | ||
334 | + 1ull << VIRTIO_BLK_F_TOPOLOGY | | ||
335 | + 1ull << VIRTIO_BLK_F_BLK_SIZE | | ||
336 | + 1ull << VIRTIO_BLK_F_FLUSH | | ||
337 | + 1ull << VIRTIO_BLK_F_DISCARD | | ||
338 | + 1ull << VIRTIO_BLK_F_WRITE_ZEROES | | ||
339 | + 1ull << VIRTIO_BLK_F_CONFIG_WCE | | ||
340 | + 1ull << VIRTIO_F_VERSION_1 | | ||
341 | + 1ull << VIRTIO_RING_F_INDIRECT_DESC | | ||
342 | + 1ull << VIRTIO_RING_F_EVENT_IDX | | ||
343 | + 1ull << VHOST_USER_F_PROTOCOL_FEATURES; | ||
344 | + | ||
345 | + if (!vdev_blk->writable) { | ||
346 | + features |= 1ull << VIRTIO_BLK_F_RO; | ||
347 | + } | ||
348 | + | ||
349 | + return features; | ||
350 | +} | ||
351 | + | ||
352 | +static uint64_t vu_block_get_protocol_features(VuDev *dev) | ||
353 | +{ | ||
354 | + return 1ull << VHOST_USER_PROTOCOL_F_CONFIG | | ||
355 | + 1ull << VHOST_USER_PROTOCOL_F_INFLIGHT_SHMFD; | ||
356 | +} | ||
357 | + | ||
358 | +static int | ||
359 | +vu_block_get_config(VuDev *vu_dev, uint8_t *config, uint32_t len) | ||
360 | +{ | ||
361 | + VuServer *server = container_of(vu_dev, VuServer, vu_dev); | ||
362 | + VuBlockDev *vdev_blk = get_vu_block_device_by_server(server); | ||
363 | + memcpy(config, &vdev_blk->blkcfg, len); | ||
364 | + | ||
365 | + return 0; | ||
366 | +} | ||
367 | + | ||
368 | +static int | ||
369 | +vu_block_set_config(VuDev *vu_dev, const uint8_t *data, | ||
370 | + uint32_t offset, uint32_t size, uint32_t flags) | ||
371 | +{ | ||
372 | + VuServer *server = container_of(vu_dev, VuServer, vu_dev); | ||
373 | + VuBlockDev *vdev_blk = get_vu_block_device_by_server(server); | ||
374 | + uint8_t wce; | ||
375 | + | ||
376 | + /* don't support live migration */ | ||
377 | + if (flags != VHOST_SET_CONFIG_TYPE_MASTER) { | ||
378 | + return -EINVAL; | ||
379 | + } | ||
380 | + | ||
381 | + if (offset != offsetof(struct virtio_blk_config, wce) || | ||
382 | + size != 1) { | ||
383 | + return -EINVAL; | ||
384 | + } | ||
385 | + | ||
386 | + wce = *data; | ||
387 | + vdev_blk->blkcfg.wce = wce; | ||
388 | + blk_set_enable_write_cache(vdev_blk->backend, wce); | ||
389 | + return 0; | ||
390 | +} | ||
391 | + | ||
392 | +/* | ||
393 | + * When the client disconnects, it sends a VHOST_USER_NONE request | ||
394 | + * and vu_process_message will simple call exit which cause the VM | ||
395 | + * to exit abruptly. | ||
396 | + * To avoid this issue, process VHOST_USER_NONE request ahead | ||
397 | + * of vu_process_message. | ||
398 | + * | ||
399 | + */ | ||
400 | +static int vu_block_process_msg(VuDev *dev, VhostUserMsg *vmsg, int *do_reply) | ||
401 | +{ | ||
402 | + if (vmsg->request == VHOST_USER_NONE) { | ||
403 | + dev->panic(dev, "disconnect"); | ||
404 | + return true; | ||
405 | + } | ||
406 | + return false; | ||
407 | +} | ||
408 | + | ||
409 | +static const VuDevIface vu_block_iface = { | ||
410 | + .get_features = vu_block_get_features, | ||
411 | + .queue_set_started = vu_block_queue_set_started, | ||
412 | + .get_protocol_features = vu_block_get_protocol_features, | ||
413 | + .get_config = vu_block_get_config, | ||
414 | + .set_config = vu_block_set_config, | ||
415 | + .process_msg = vu_block_process_msg, | ||
416 | +}; | ||
417 | + | ||
418 | +static void blk_aio_attached(AioContext *ctx, void *opaque) | ||
419 | +{ | ||
420 | + VuBlockDev *vub_dev = opaque; | ||
421 | + aio_context_acquire(ctx); | ||
422 | + vhost_user_server_set_aio_context(&vub_dev->vu_server, ctx); | ||
423 | + aio_context_release(ctx); | ||
424 | +} | ||
425 | + | ||
426 | +static void blk_aio_detach(void *opaque) | ||
427 | +{ | ||
428 | + VuBlockDev *vub_dev = opaque; | ||
429 | + AioContext *ctx = vub_dev->vu_server.ctx; | ||
430 | + aio_context_acquire(ctx); | ||
431 | + vhost_user_server_set_aio_context(&vub_dev->vu_server, NULL); | ||
432 | + aio_context_release(ctx); | ||
433 | +} | ||
434 | + | ||
435 | +static void | ||
436 | +vu_block_initialize_config(BlockDriverState *bs, | ||
437 | + struct virtio_blk_config *config, uint32_t blk_size) | ||
438 | +{ | ||
439 | + config->capacity = bdrv_getlength(bs) >> BDRV_SECTOR_BITS; | ||
440 | + config->blk_size = blk_size; | ||
441 | + config->size_max = 0; | ||
442 | + config->seg_max = 128 - 2; | ||
443 | + config->min_io_size = 1; | ||
444 | + config->opt_io_size = 1; | ||
445 | + config->num_queues = VHOST_USER_BLK_MAX_QUEUES; | ||
446 | + config->max_discard_sectors = 32768; | ||
447 | + config->max_discard_seg = 1; | ||
448 | + config->discard_sector_alignment = config->blk_size >> 9; | ||
449 | + config->max_write_zeroes_sectors = 32768; | ||
450 | + config->max_write_zeroes_seg = 1; | ||
451 | +} | ||
452 | + | ||
453 | +static VuBlockDev *vu_block_init(VuBlockDev *vu_block_device, Error **errp) | ||
454 | +{ | ||
455 | + | ||
456 | + BlockBackend *blk; | ||
457 | + Error *local_error = NULL; | ||
458 | + const char *node_name = vu_block_device->node_name; | ||
459 | + bool writable = vu_block_device->writable; | ||
460 | + uint64_t perm = BLK_PERM_CONSISTENT_READ; | ||
461 | + int ret; | ||
462 | + | ||
463 | + AioContext *ctx; | ||
464 | + | ||
465 | + BlockDriverState *bs = bdrv_lookup_bs(node_name, node_name, &local_error); | ||
466 | + | ||
467 | + if (!bs) { | ||
468 | + error_propagate(errp, local_error); | ||
469 | + return NULL; | ||
470 | + } | ||
471 | + | ||
472 | + if (bdrv_is_read_only(bs)) { | ||
473 | + writable = false; | ||
474 | + } | ||
475 | + | ||
476 | + if (writable) { | ||
477 | + perm |= BLK_PERM_WRITE; | ||
478 | + } | ||
479 | + | ||
480 | + ctx = bdrv_get_aio_context(bs); | ||
481 | + aio_context_acquire(ctx); | ||
482 | + bdrv_invalidate_cache(bs, NULL); | ||
483 | + aio_context_release(ctx); | ||
484 | + | ||
485 | + /* | ||
486 | + * Don't allow resize while the vhost user server is running, | ||
487 | + * otherwise we don't care what happens with the node. | ||
488 | + */ | ||
489 | + blk = blk_new(bdrv_get_aio_context(bs), perm, | ||
490 | + BLK_PERM_CONSISTENT_READ | BLK_PERM_WRITE_UNCHANGED | | ||
491 | + BLK_PERM_WRITE | BLK_PERM_GRAPH_MOD); | ||
492 | + ret = blk_insert_bs(blk, bs, errp); | ||
493 | + | ||
494 | + if (ret < 0) { | ||
495 | + goto fail; | ||
496 | + } | ||
497 | + | ||
498 | + blk_set_enable_write_cache(blk, false); | ||
499 | + | ||
500 | + blk_set_allow_aio_context_change(blk, true); | ||
501 | + | ||
502 | + vu_block_device->blkcfg.wce = 0; | ||
503 | + vu_block_device->backend = blk; | ||
504 | + if (!vu_block_device->blk_size) { | ||
505 | + vu_block_device->blk_size = BDRV_SECTOR_SIZE; | ||
506 | + } | ||
507 | + vu_block_device->blkcfg.blk_size = vu_block_device->blk_size; | ||
508 | + blk_set_guest_block_size(blk, vu_block_device->blk_size); | ||
509 | + vu_block_initialize_config(bs, &vu_block_device->blkcfg, | ||
510 | + vu_block_device->blk_size); | ||
511 | + return vu_block_device; | ||
512 | + | ||
513 | +fail: | ||
514 | + blk_unref(blk); | ||
515 | + return NULL; | ||
516 | +} | ||
517 | + | ||
518 | +static void vu_block_deinit(VuBlockDev *vu_block_device) | ||
519 | +{ | ||
520 | + if (vu_block_device->backend) { | ||
521 | + blk_remove_aio_context_notifier(vu_block_device->backend, blk_aio_attached, | ||
522 | + blk_aio_detach, vu_block_device); | ||
523 | + } | ||
524 | + | ||
525 | + blk_unref(vu_block_device->backend); | ||
526 | +} | ||
527 | + | ||
528 | +static void vhost_user_blk_server_stop(VuBlockDev *vu_block_device) | ||
529 | +{ | ||
530 | + vhost_user_server_stop(&vu_block_device->vu_server); | ||
531 | + vu_block_deinit(vu_block_device); | ||
532 | +} | ||
533 | + | ||
534 | +static void vhost_user_blk_server_start(VuBlockDev *vu_block_device, | ||
535 | + Error **errp) | ||
536 | +{ | ||
537 | + AioContext *ctx; | ||
538 | + SocketAddress *addr = vu_block_device->addr; | ||
539 | + | ||
540 | + if (!vu_block_init(vu_block_device, errp)) { | ||
541 | + return; | ||
542 | + } | ||
543 | + | ||
544 | + ctx = bdrv_get_aio_context(blk_bs(vu_block_device->backend)); | ||
545 | + | ||
546 | + if (!vhost_user_server_start(&vu_block_device->vu_server, addr, ctx, | ||
547 | + VHOST_USER_BLK_MAX_QUEUES, | ||
548 | + NULL, &vu_block_iface, | ||
549 | + errp)) { | ||
550 | + goto error; | ||
551 | + } | ||
552 | + | ||
553 | + blk_add_aio_context_notifier(vu_block_device->backend, blk_aio_attached, | ||
554 | + blk_aio_detach, vu_block_device); | ||
555 | + vu_block_device->running = true; | ||
556 | + return; | ||
557 | + | ||
558 | + error: | ||
559 | + vu_block_deinit(vu_block_device); | ||
560 | +} | ||
561 | + | ||
562 | +static bool vu_prop_modifiable(VuBlockDev *vus, Error **errp) | ||
563 | +{ | ||
564 | + if (vus->running) { | ||
565 | + error_setg(errp, "The property can't be modified " | ||
566 | + "while the server is running"); | ||
567 | + return false; | ||
568 | + } | ||
569 | + return true; | ||
570 | +} | ||
571 | + | ||
572 | +static void vu_set_node_name(Object *obj, const char *value, Error **errp) | ||
573 | +{ | ||
574 | + VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj); | ||
575 | + | ||
576 | + if (!vu_prop_modifiable(vus, errp)) { | ||
577 | + return; | ||
578 | + } | ||
579 | + | ||
580 | + if (vus->node_name) { | ||
581 | + g_free(vus->node_name); | ||
582 | + } | ||
583 | + | ||
584 | + vus->node_name = g_strdup(value); | ||
585 | +} | ||
586 | + | ||
587 | +static char *vu_get_node_name(Object *obj, Error **errp) | ||
588 | +{ | ||
589 | + VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj); | ||
590 | + return g_strdup(vus->node_name); | ||
591 | +} | ||
592 | + | ||
593 | +static void free_socket_addr(SocketAddress *addr) | ||
594 | +{ | ||
595 | + g_free(addr->u.q_unix.path); | ||
596 | + g_free(addr); | ||
597 | +} | ||
598 | + | ||
599 | +static void vu_set_unix_socket(Object *obj, const char *value, | ||
600 | + Error **errp) | ||
601 | +{ | ||
602 | + VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj); | ||
603 | + | ||
604 | + if (!vu_prop_modifiable(vus, errp)) { | ||
605 | + return; | ||
606 | + } | ||
607 | + | ||
608 | + if (vus->addr) { | ||
609 | + free_socket_addr(vus->addr); | ||
610 | + } | ||
611 | + | ||
612 | + SocketAddress *addr = g_new0(SocketAddress, 1); | ||
613 | + addr->type = SOCKET_ADDRESS_TYPE_UNIX; | ||
614 | + addr->u.q_unix.path = g_strdup(value); | ||
615 | + vus->addr = addr; | ||
616 | +} | ||
617 | + | ||
618 | +static char *vu_get_unix_socket(Object *obj, Error **errp) | ||
619 | +{ | ||
620 | + VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj); | ||
621 | + return g_strdup(vus->addr->u.q_unix.path); | ||
622 | +} | ||
623 | + | ||
624 | +static bool vu_get_block_writable(Object *obj, Error **errp) | ||
625 | +{ | ||
626 | + VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj); | ||
627 | + return vus->writable; | ||
628 | +} | ||
629 | + | ||
630 | +static void vu_set_block_writable(Object *obj, bool value, Error **errp) | ||
631 | +{ | ||
632 | + VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj); | ||
633 | + | ||
634 | + if (!vu_prop_modifiable(vus, errp)) { | ||
635 | + return; | ||
636 | + } | ||
637 | + | ||
638 | + vus->writable = value; | ||
639 | +} | ||
640 | + | ||
641 | +static void vu_get_blk_size(Object *obj, Visitor *v, const char *name, | ||
642 | + void *opaque, Error **errp) | ||
643 | +{ | ||
644 | + VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj); | ||
645 | + uint32_t value = vus->blk_size; | ||
646 | + | ||
647 | + visit_type_uint32(v, name, &value, errp); | ||
648 | +} | ||
649 | + | ||
650 | +static void vu_set_blk_size(Object *obj, Visitor *v, const char *name, | ||
651 | + void *opaque, Error **errp) | ||
652 | +{ | ||
653 | + VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj); | ||
654 | + | ||
655 | + Error *local_err = NULL; | ||
656 | + uint32_t value; | ||
657 | + | ||
658 | + if (!vu_prop_modifiable(vus, errp)) { | ||
659 | + return; | ||
660 | + } | ||
661 | + | ||
662 | + visit_type_uint32(v, name, &value, &local_err); | ||
663 | + if (local_err) { | ||
664 | + goto out; | ||
665 | + } | ||
666 | + | ||
667 | + check_block_size(object_get_typename(obj), name, value, &local_err); | ||
668 | + if (local_err) { | ||
669 | + goto out; | ||
670 | + } | ||
671 | + | ||
672 | + vus->blk_size = value; | ||
673 | + | ||
674 | +out: | ||
675 | + error_propagate(errp, local_err); | ||
676 | +} | ||
677 | + | ||
678 | +static void vhost_user_blk_server_instance_finalize(Object *obj) | ||
679 | +{ | ||
680 | + VuBlockDev *vub = VHOST_USER_BLK_SERVER(obj); | ||
681 | + | ||
682 | + vhost_user_blk_server_stop(vub); | ||
683 | + | ||
684 | + /* | ||
685 | + * Unlike object_property_add_str, object_class_property_add_str | ||
686 | + * doesn't have a release method. Thus manual memory freeing is | ||
687 | + * needed. | ||
688 | + */ | ||
689 | + free_socket_addr(vub->addr); | ||
690 | + g_free(vub->node_name); | ||
691 | +} | ||
692 | + | ||
693 | +static void vhost_user_blk_server_complete(UserCreatable *obj, Error **errp) | ||
694 | +{ | ||
695 | + VuBlockDev *vub = VHOST_USER_BLK_SERVER(obj); | ||
696 | + | ||
697 | + vhost_user_blk_server_start(vub, errp); | ||
698 | +} | ||
699 | + | ||
700 | +static void vhost_user_blk_server_class_init(ObjectClass *klass, | ||
701 | + void *class_data) | ||
702 | +{ | ||
703 | + UserCreatableClass *ucc = USER_CREATABLE_CLASS(klass); | ||
704 | + ucc->complete = vhost_user_blk_server_complete; | ||
705 | + | ||
706 | + object_class_property_add_bool(klass, "writable", | ||
707 | + vu_get_block_writable, | ||
708 | + vu_set_block_writable); | ||
709 | + | ||
710 | + object_class_property_add_str(klass, "node-name", | ||
711 | + vu_get_node_name, | ||
712 | + vu_set_node_name); | ||
713 | + | ||
714 | + object_class_property_add_str(klass, "unix-socket", | ||
715 | + vu_get_unix_socket, | ||
716 | + vu_set_unix_socket); | ||
717 | + | ||
718 | + object_class_property_add(klass, "logical-block-size", "uint32", | ||
719 | + vu_get_blk_size, vu_set_blk_size, | ||
720 | + NULL, NULL); | ||
721 | +} | ||
722 | + | ||
723 | +static const TypeInfo vhost_user_blk_server_info = { | ||
724 | + .name = TYPE_VHOST_USER_BLK_SERVER, | ||
725 | + .parent = TYPE_OBJECT, | ||
726 | + .instance_size = sizeof(VuBlockDev), | ||
727 | + .instance_finalize = vhost_user_blk_server_instance_finalize, | ||
728 | + .class_init = vhost_user_blk_server_class_init, | ||
729 | + .interfaces = (InterfaceInfo[]) { | ||
730 | + {TYPE_USER_CREATABLE}, | ||
731 | + {} | ||
732 | + }, | ||
733 | +}; | ||
734 | + | ||
735 | +static void vhost_user_blk_server_register_types(void) | ||
736 | +{ | ||
737 | + type_register_static(&vhost_user_blk_server_info); | ||
738 | +} | ||
739 | + | ||
740 | +type_init(vhost_user_blk_server_register_types) | ||
741 | diff --git a/softmmu/vl.c b/softmmu/vl.c | ||
742 | index XXXXXXX..XXXXXXX 100644 | ||
743 | --- a/softmmu/vl.c | ||
744 | +++ b/softmmu/vl.c | ||
745 | @@ -XXX,XX +XXX,XX @@ static bool object_create_initial(const char *type, QemuOpts *opts) | ||
746 | } | ||
747 | #endif | ||
748 | |||
749 | + /* Reason: vhost-user-blk-server property "node-name" */ | ||
750 | + if (g_str_equal(type, "vhost-user-blk-server")) { | ||
751 | + return false; | ||
752 | + } | ||
753 | /* | ||
754 | * Reason: filter-* property "netdev" etc. | ||
755 | */ | ||
756 | diff --git a/block/meson.build b/block/meson.build | ||
757 | index XXXXXXX..XXXXXXX 100644 | ||
758 | --- a/block/meson.build | ||
759 | +++ b/block/meson.build | ||
760 | @@ -XXX,XX +XXX,XX @@ block_ss.add(when: 'CONFIG_WIN32', if_true: files('file-win32.c', 'win32-aio.c') | ||
761 | block_ss.add(when: 'CONFIG_POSIX', if_true: [files('file-posix.c'), coref, iokit]) | ||
762 | block_ss.add(when: 'CONFIG_LIBISCSI', if_true: files('iscsi-opts.c')) | ||
763 | block_ss.add(when: 'CONFIG_LINUX', if_true: files('nvme.c')) | ||
764 | +block_ss.add(when: 'CONFIG_LINUX', if_true: files('export/vhost-user-blk-server.c', '../contrib/libvhost-user/libvhost-user.c')) | ||
765 | block_ss.add(when: 'CONFIG_REPLICATION', if_true: files('replication.c')) | ||
766 | block_ss.add(when: 'CONFIG_SHEEPDOG', if_true: files('sheepdog.c')) | ||
767 | block_ss.add(when: ['CONFIG_LINUX_AIO', libaio], if_true: files('linux-aio.c')) | ||
768 | -- | ||
769 | 2.26.2 | ||
770 | diff view generated by jsdifflib |
New patch | |||
---|---|---|---|
1 | From: Coiby Xu <coiby.xu@gmail.com> | ||
1 | 2 | ||
3 | Suggested-by: Stefano Garzarella <sgarzare@redhat.com> | ||
4 | Signed-off-by: Coiby Xu <coiby.xu@gmail.com> | ||
5 | Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
6 | Reviewed-by: Marc-André Lureau <marcandre.lureau@redhat.com> | ||
7 | Message-id: 20200918080912.321299-8-coiby.xu@gmail.com | ||
8 | [Removed reference to vhost-user-blk-test.c, it will be sent in a | ||
9 | separate pull request. | ||
10 | --Stefan] | ||
11 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
12 | --- | ||
13 | MAINTAINERS | 7 +++++++ | ||
14 | 1 file changed, 7 insertions(+) | ||
15 | |||
16 | diff --git a/MAINTAINERS b/MAINTAINERS | ||
17 | index XXXXXXX..XXXXXXX 100644 | ||
18 | --- a/MAINTAINERS | ||
19 | +++ b/MAINTAINERS | ||
20 | @@ -XXX,XX +XXX,XX @@ L: qemu-block@nongnu.org | ||
21 | S: Supported | ||
22 | F: tests/image-fuzzer/ | ||
23 | |||
24 | +Vhost-user block device backend server | ||
25 | +M: Coiby Xu <Coiby.Xu@gmail.com> | ||
26 | +S: Maintained | ||
27 | +F: block/export/vhost-user-blk-server.c | ||
28 | +F: util/vhost-user-server.c | ||
29 | +F: tests/qtest/libqos/vhost-user-blk.c | ||
30 | + | ||
31 | Replication | ||
32 | M: Wen Congyang <wencongyang2@huawei.com> | ||
33 | M: Xie Changlong <xiechanglong.d@gmail.com> | ||
34 | -- | ||
35 | 2.26.2 | ||
36 | diff view generated by jsdifflib |
New patch | |||
---|---|---|---|
1 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
2 | Message-id: 20200924151549.913737-3-stefanha@redhat.com | ||
3 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
4 | --- | ||
5 | util/vhost-user-server.c | 2 +- | ||
6 | 1 file changed, 1 insertion(+), 1 deletion(-) | ||
1 | 7 | ||
8 | diff --git a/util/vhost-user-server.c b/util/vhost-user-server.c | ||
9 | index XXXXXXX..XXXXXXX 100644 | ||
10 | --- a/util/vhost-user-server.c | ||
11 | +++ b/util/vhost-user-server.c | ||
12 | @@ -XXX,XX +XXX,XX @@ bool vhost_user_server_start(VuServer *server, | ||
13 | return false; | ||
14 | } | ||
15 | |||
16 | - /* zero out unspecified fileds */ | ||
17 | + /* zero out unspecified fields */ | ||
18 | *server = (VuServer) { | ||
19 | .listener = listener, | ||
20 | .vu_iface = vu_iface, | ||
21 | -- | ||
22 | 2.26.2 | ||
23 | diff view generated by jsdifflib |
1 | From: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | 1 | We already have access to the value with the correct type (ioc and sioc |
---|---|---|---|
2 | are the same QIOChannel). | ||
2 | 3 | ||
3 | All the offsets in the BAT must be lower than the file size. | 4 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
4 | Fix the check condition for correct check. | 5 | Message-id: 20200924151549.913737-4-stefanha@redhat.com |
5 | 6 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | |
6 | Signed-off-by: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | ||
7 | Reviewed-by: Denis V. Lunev <den@openvz.org> | ||
8 | Message-Id: <20230424093147.197643-13-alexander.ivanov@virtuozzo.com> | ||
9 | Reviewed-by: Hanna Czenczek <hreitz@redhat.com> | ||
10 | Signed-off-by: Hanna Czenczek <hreitz@redhat.com> | ||
11 | --- | 7 | --- |
12 | block/parallels.c | 2 +- | 8 | util/vhost-user-server.c | 2 +- |
13 | 1 file changed, 1 insertion(+), 1 deletion(-) | 9 | 1 file changed, 1 insertion(+), 1 deletion(-) |
14 | 10 | ||
15 | diff --git a/block/parallels.c b/block/parallels.c | 11 | diff --git a/util/vhost-user-server.c b/util/vhost-user-server.c |
16 | index XXXXXXX..XXXXXXX 100644 | 12 | index XXXXXXX..XXXXXXX 100644 |
17 | --- a/block/parallels.c | 13 | --- a/util/vhost-user-server.c |
18 | +++ b/block/parallels.c | 14 | +++ b/util/vhost-user-server.c |
19 | @@ -XXX,XX +XXX,XX @@ parallels_check_outside_image(BlockDriverState *bs, BdrvCheckResult *res, | 15 | @@ -XXX,XX +XXX,XX @@ static void vu_accept(QIONetListener *listener, QIOChannelSocket *sioc, |
20 | high_off = 0; | 16 | server->ioc = QIO_CHANNEL(sioc); |
21 | for (i = 0; i < s->bat_size; i++) { | 17 | object_ref(OBJECT(server->ioc)); |
22 | off = bat2sect(s, i) << BDRV_SECTOR_BITS; | 18 | qio_channel_attach_aio_context(server->ioc, server->ctx); |
23 | - if (off > size) { | 19 | - qio_channel_set_blocking(QIO_CHANNEL(server->sioc), false, NULL); |
24 | + if (off + s->cluster_size > size) { | 20 | + qio_channel_set_blocking(server->ioc, false, NULL); |
25 | fprintf(stderr, "%s cluster %u is outside image\n", | 21 | vu_client_start(server); |
26 | fix & BDRV_FIX_ERRORS ? "Repairing" : "ERROR", i); | 22 | } |
27 | res->corruptions++; | 23 | |
28 | -- | 24 | -- |
29 | 2.40.1 | 25 | 2.26.2 |
26 | diff view generated by jsdifflib |
1 | bdrv_pad_request() was the main user of qemu_iovec_init_extended(). | 1 | Explicitly deleting watches is not necessary since libvhost-user calls |
---|---|---|---|
2 | HEAD^ has removed that use, so we can remove qemu_iovec_init_extended() | 2 | remove_watch() during vu_deinit(). Add an assertion to check this |
3 | now. | 3 | though. |
4 | 4 | ||
5 | The only remaining user is qemu_iovec_init_slice(), which can easily | 5 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
6 | inline the small part it really needs. | 6 | Message-id: 20200924151549.913737-5-stefanha@redhat.com |
7 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
8 | --- | ||
9 | util/vhost-user-server.c | 19 ++++--------------- | ||
10 | 1 file changed, 4 insertions(+), 15 deletions(-) | ||
7 | 11 | ||
8 | Note that qemu_iovec_init_extended() offered a memcpy() optimization to | 12 | diff --git a/util/vhost-user-server.c b/util/vhost-user-server.c |
9 | initialize the new I/O vector. qemu_iovec_concat_iov(), which is used | ||
10 | to replace its functionality, does not, but calls qemu_iovec_add() for | ||
11 | every single element. If we decide this optimization was important, we | ||
12 | will need to re-implement it in qemu_iovec_concat_iov(), which might | ||
13 | also benefit its pre-existing users. | ||
14 | |||
15 | Reviewed-by: Eric Blake <eblake@redhat.com> | ||
16 | Reviewed-by: Vladimir Sementsov-Ogievskiy <vsementsov@yandex-team.ru> | ||
17 | Signed-off-by: Hanna Czenczek <hreitz@redhat.com> | ||
18 | Message-Id: <20230411173418.19549-4-hreitz@redhat.com> | ||
19 | --- | ||
20 | include/qemu/iov.h | 5 --- | ||
21 | util/iov.c | 79 +++++++--------------------------------------- | ||
22 | 2 files changed, 11 insertions(+), 73 deletions(-) | ||
23 | |||
24 | diff --git a/include/qemu/iov.h b/include/qemu/iov.h | ||
25 | index XXXXXXX..XXXXXXX 100644 | 13 | index XXXXXXX..XXXXXXX 100644 |
26 | --- a/include/qemu/iov.h | 14 | --- a/util/vhost-user-server.c |
27 | +++ b/include/qemu/iov.h | 15 | +++ b/util/vhost-user-server.c |
28 | @@ -XXX,XX +XXX,XX @@ static inline void *qemu_iovec_buf(QEMUIOVector *qiov) | 16 | @@ -XXX,XX +XXX,XX @@ static void close_client(VuServer *server) |
29 | 17 | /* When this is set vu_client_trip will stop new processing vhost-user message */ | |
30 | void qemu_iovec_init(QEMUIOVector *qiov, int alloc_hint); | 18 | server->sioc = NULL; |
31 | void qemu_iovec_init_external(QEMUIOVector *qiov, struct iovec *iov, int niov); | 19 | |
32 | -int qemu_iovec_init_extended( | 20 | - VuFdWatch *vu_fd_watch, *next; |
33 | - QEMUIOVector *qiov, | 21 | - QTAILQ_FOREACH_SAFE(vu_fd_watch, &server->vu_fd_watches, next, next) { |
34 | - void *head_buf, size_t head_len, | 22 | - aio_set_fd_handler(server->ioc->ctx, vu_fd_watch->fd, true, NULL, |
35 | - QEMUIOVector *mid_qiov, size_t mid_offset, size_t mid_len, | 23 | - NULL, NULL, NULL); |
36 | - void *tail_buf, size_t tail_len); | ||
37 | void qemu_iovec_init_slice(QEMUIOVector *qiov, QEMUIOVector *source, | ||
38 | size_t offset, size_t len); | ||
39 | struct iovec *qemu_iovec_slice(QEMUIOVector *qiov, | ||
40 | diff --git a/util/iov.c b/util/iov.c | ||
41 | index XXXXXXX..XXXXXXX 100644 | ||
42 | --- a/util/iov.c | ||
43 | +++ b/util/iov.c | ||
44 | @@ -XXX,XX +XXX,XX @@ int qemu_iovec_subvec_niov(QEMUIOVector *qiov, size_t offset, size_t len) | ||
45 | return niov; | ||
46 | } | ||
47 | |||
48 | -/* | ||
49 | - * Compile new iovec, combining @head_buf buffer, sub-qiov of @mid_qiov, | ||
50 | - * and @tail_buf buffer into new qiov. | ||
51 | - */ | ||
52 | -int qemu_iovec_init_extended( | ||
53 | - QEMUIOVector *qiov, | ||
54 | - void *head_buf, size_t head_len, | ||
55 | - QEMUIOVector *mid_qiov, size_t mid_offset, size_t mid_len, | ||
56 | - void *tail_buf, size_t tail_len) | ||
57 | -{ | ||
58 | - size_t mid_head, mid_tail; | ||
59 | - int total_niov, mid_niov = 0; | ||
60 | - struct iovec *p, *mid_iov = NULL; | ||
61 | - | ||
62 | - assert(mid_qiov->niov <= IOV_MAX); | ||
63 | - | ||
64 | - if (SIZE_MAX - head_len < mid_len || | ||
65 | - SIZE_MAX - head_len - mid_len < tail_len) | ||
66 | - { | ||
67 | - return -EINVAL; | ||
68 | - } | 24 | - } |
69 | - | 25 | - |
70 | - if (mid_len) { | 26 | - while (!QTAILQ_EMPTY(&server->vu_fd_watches)) { |
71 | - mid_iov = qemu_iovec_slice(mid_qiov, mid_offset, mid_len, | 27 | - QTAILQ_FOREACH_SAFE(vu_fd_watch, &server->vu_fd_watches, next, next) { |
72 | - &mid_head, &mid_tail, &mid_niov); | 28 | - if (!vu_fd_watch->processing) { |
29 | - QTAILQ_REMOVE(&server->vu_fd_watches, vu_fd_watch, next); | ||
30 | - g_free(vu_fd_watch); | ||
31 | - } | ||
32 | - } | ||
73 | - } | 33 | - } |
74 | - | 34 | - |
75 | - total_niov = !!head_len + mid_niov + !!tail_len; | 35 | while (server->processing_msg) { |
76 | - if (total_niov > IOV_MAX) { | 36 | if (server->ioc->read_coroutine) { |
77 | - return -EINVAL; | 37 | server->ioc->read_coroutine = NULL; |
78 | - } | 38 | @@ -XXX,XX +XXX,XX @@ static void close_client(VuServer *server) |
79 | - | 39 | } |
80 | - if (total_niov == 1) { | 40 | |
81 | - qemu_iovec_init_buf(qiov, NULL, 0); | 41 | vu_deinit(&server->vu_dev); |
82 | - p = &qiov->local_iov; | 42 | + |
83 | - } else { | 43 | + /* vu_deinit() should have called remove_watch() */ |
84 | - qiov->niov = qiov->nalloc = total_niov; | 44 | + assert(QTAILQ_EMPTY(&server->vu_fd_watches)); |
85 | - qiov->size = head_len + mid_len + tail_len; | 45 | + |
86 | - p = qiov->iov = g_new(struct iovec, qiov->niov); | 46 | object_unref(OBJECT(sioc)); |
87 | - } | 47 | object_unref(OBJECT(server->ioc)); |
88 | - | ||
89 | - if (head_len) { | ||
90 | - p->iov_base = head_buf; | ||
91 | - p->iov_len = head_len; | ||
92 | - p++; | ||
93 | - } | ||
94 | - | ||
95 | - assert(!mid_niov == !mid_len); | ||
96 | - if (mid_niov) { | ||
97 | - memcpy(p, mid_iov, mid_niov * sizeof(*p)); | ||
98 | - p[0].iov_base = (uint8_t *)p[0].iov_base + mid_head; | ||
99 | - p[0].iov_len -= mid_head; | ||
100 | - p[mid_niov - 1].iov_len -= mid_tail; | ||
101 | - p += mid_niov; | ||
102 | - } | ||
103 | - | ||
104 | - if (tail_len) { | ||
105 | - p->iov_base = tail_buf; | ||
106 | - p->iov_len = tail_len; | ||
107 | - } | ||
108 | - | ||
109 | - return 0; | ||
110 | -} | ||
111 | - | ||
112 | /* | ||
113 | * Check if the contents of subrange of qiov data is all zeroes. | ||
114 | */ | ||
115 | @@ -XXX,XX +XXX,XX @@ bool qemu_iovec_is_zero(QEMUIOVector *qiov, size_t offset, size_t bytes) | ||
116 | void qemu_iovec_init_slice(QEMUIOVector *qiov, QEMUIOVector *source, | ||
117 | size_t offset, size_t len) | ||
118 | { | ||
119 | - int ret; | ||
120 | + struct iovec *slice_iov; | ||
121 | + int slice_niov; | ||
122 | + size_t slice_head, slice_tail; | ||
123 | |||
124 | assert(source->size >= len); | ||
125 | assert(source->size - len >= offset); | ||
126 | |||
127 | - /* We shrink the request, so we can't overflow neither size_t nor MAX_IOV */ | ||
128 | - ret = qemu_iovec_init_extended(qiov, NULL, 0, source, offset, len, NULL, 0); | ||
129 | - assert(ret == 0); | ||
130 | + slice_iov = qemu_iovec_slice(source, offset, len, | ||
131 | + &slice_head, &slice_tail, &slice_niov); | ||
132 | + if (slice_niov == 1) { | ||
133 | + qemu_iovec_init_buf(qiov, slice_iov[0].iov_base + slice_head, len); | ||
134 | + } else { | ||
135 | + qemu_iovec_init(qiov, slice_niov); | ||
136 | + qemu_iovec_concat_iov(qiov, slice_iov, slice_niov, slice_head, len); | ||
137 | + } | ||
138 | } | 48 | } |
139 | |||
140 | void qemu_iovec_destroy(QEMUIOVector *qiov) | ||
141 | -- | 49 | -- |
142 | 2.40.1 | 50 | 2.26.2 |
51 | diff view generated by jsdifflib |
1 | From: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | 1 | Only one struct is needed per request. Drop req_data and the separate |
---|---|---|---|
2 | VuBlockReq instance. Instead let vu_queue_pop() allocate everything at | ||
3 | once. | ||
2 | 4 | ||
3 | BAT is written in the context of conventional operations over the image | 5 | This fixes the req_data memory leak in vu_block_virtio_process_req(). |
4 | inside bdrv_co_flush() when it calls parallels_co_flush_to_os() callback. | ||
5 | Thus we should not modify BAT array directly, but call | ||
6 | parallels_set_bat_entry() helper and bdrv_co_flush() further on. After | ||
7 | that there is no need to manually write BAT and track its modification. | ||
8 | 6 | ||
9 | This makes code more generic and allows to split parallels_set_bat_entry() | 7 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
10 | for independent pieces. | 8 | Message-id: 20200924151549.913737-6-stefanha@redhat.com |
9 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
10 | --- | ||
11 | block/export/vhost-user-blk-server.c | 68 +++++++++------------------- | ||
12 | 1 file changed, 21 insertions(+), 47 deletions(-) | ||
11 | 13 | ||
12 | Signed-off-by: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | 14 | diff --git a/block/export/vhost-user-blk-server.c b/block/export/vhost-user-blk-server.c |
13 | Reviewed-by: Denis V. Lunev <den@openvz.org> | ||
14 | Message-Id: <20230424093147.197643-6-alexander.ivanov@virtuozzo.com> | ||
15 | Reviewed-by: Hanna Czenczek <hreitz@redhat.com> | ||
16 | Signed-off-by: Hanna Czenczek <hreitz@redhat.com> | ||
17 | --- | ||
18 | block/parallels.c | 23 ++++++++++------------- | ||
19 | 1 file changed, 10 insertions(+), 13 deletions(-) | ||
20 | |||
21 | diff --git a/block/parallels.c b/block/parallels.c | ||
22 | index XXXXXXX..XXXXXXX 100644 | 15 | index XXXXXXX..XXXXXXX 100644 |
23 | --- a/block/parallels.c | 16 | --- a/block/export/vhost-user-blk-server.c |
24 | +++ b/block/parallels.c | 17 | +++ b/block/export/vhost-user-blk-server.c |
25 | @@ -XXX,XX +XXX,XX @@ parallels_co_check(BlockDriverState *bs, BdrvCheckResult *res, | 18 | @@ -XXX,XX +XXX,XX @@ struct virtio_blk_inhdr { |
26 | { | 19 | }; |
27 | BDRVParallelsState *s = bs->opaque; | 20 | |
28 | int64_t size, prev_off, high_off; | 21 | typedef struct VuBlockReq { |
29 | - int ret; | 22 | - VuVirtqElement *elem; |
30 | + int ret = 0; | 23 | + VuVirtqElement elem; |
31 | uint32_t i; | 24 | int64_t sector_num; |
32 | - bool flush_bat = false; | 25 | size_t size; |
33 | 26 | struct virtio_blk_inhdr *in; | |
34 | size = bdrv_getlength(bs->file->bs); | 27 | @@ -XXX,XX +XXX,XX @@ static void vu_block_req_complete(VuBlockReq *req) |
35 | if (size < 0) { | 28 | VuDev *vu_dev = &req->server->vu_dev; |
36 | @@ -XXX,XX +XXX,XX @@ parallels_co_check(BlockDriverState *bs, BdrvCheckResult *res, | 29 | |
37 | fix & BDRV_FIX_ERRORS ? "Repairing" : "ERROR", i); | 30 | /* IO size with 1 extra status byte */ |
38 | res->corruptions++; | 31 | - vu_queue_push(vu_dev, req->vq, req->elem, req->size + 1); |
39 | if (fix & BDRV_FIX_ERRORS) { | 32 | + vu_queue_push(vu_dev, req->vq, &req->elem, req->size + 1); |
40 | - s->bat_bitmap[i] = 0; | 33 | vu_queue_notify(vu_dev, req->vq); |
41 | + parallels_set_bat_entry(s, i, 0); | 34 | |
42 | res->corruptions_fixed++; | 35 | - if (req->elem) { |
43 | - flush_bat = true; | 36 | - free(req->elem); |
44 | } | ||
45 | prev_off = 0; | ||
46 | continue; | ||
47 | @@ -XXX,XX +XXX,XX @@ parallels_co_check(BlockDriverState *bs, BdrvCheckResult *res, | ||
48 | prev_off = off; | ||
49 | } | ||
50 | |||
51 | - ret = 0; | ||
52 | - if (flush_bat) { | ||
53 | - ret = bdrv_co_pwrite_sync(bs->file, 0, s->header_size, s->header, 0); | ||
54 | - if (ret < 0) { | ||
55 | - res->check_errors++; | ||
56 | - goto out; | ||
57 | - } | ||
58 | - } | 37 | - } |
59 | - | 38 | - |
60 | if (high_off == 0) { | 39 | - g_free(req); |
61 | res->image_end_offset = s->data_end << BDRV_SECTOR_BITS; | 40 | + free(req); |
62 | } else { | 41 | } |
63 | @@ -XXX,XX +XXX,XX @@ parallels_co_check(BlockDriverState *bs, BdrvCheckResult *res, | 42 | |
64 | 43 | static VuBlockDev *get_vu_block_device_by_server(VuServer *server) | |
65 | out: | 44 | @@ -XXX,XX +XXX,XX @@ static void coroutine_fn vu_block_flush(VuBlockReq *req) |
66 | qemu_co_mutex_unlock(&s->lock); | 45 | blk_co_flush(backend); |
46 | } | ||
47 | |||
48 | -struct req_data { | ||
49 | - VuServer *server; | ||
50 | - VuVirtq *vq; | ||
51 | - VuVirtqElement *elem; | ||
52 | -}; | ||
53 | - | ||
54 | static void coroutine_fn vu_block_virtio_process_req(void *opaque) | ||
55 | { | ||
56 | - struct req_data *data = opaque; | ||
57 | - VuServer *server = data->server; | ||
58 | - VuVirtq *vq = data->vq; | ||
59 | - VuVirtqElement *elem = data->elem; | ||
60 | + VuBlockReq *req = opaque; | ||
61 | + VuServer *server = req->server; | ||
62 | + VuVirtqElement *elem = &req->elem; | ||
63 | uint32_t type; | ||
64 | - VuBlockReq *req; | ||
65 | |||
66 | VuBlockDev *vdev_blk = get_vu_block_device_by_server(server); | ||
67 | BlockBackend *backend = vdev_blk->backend; | ||
68 | @@ -XXX,XX +XXX,XX @@ static void coroutine_fn vu_block_virtio_process_req(void *opaque) | ||
69 | struct iovec *out_iov = elem->out_sg; | ||
70 | unsigned in_num = elem->in_num; | ||
71 | unsigned out_num = elem->out_num; | ||
67 | + | 72 | + |
68 | + if (ret == 0) { | 73 | /* refer to hw/block/virtio_blk.c */ |
69 | + ret = bdrv_co_flush(bs); | 74 | if (elem->out_num < 1 || elem->in_num < 1) { |
70 | + if (ret < 0) { | 75 | error_report("virtio-blk request missing headers"); |
71 | + res->check_errors++; | 76 | - free(elem); |
72 | + } | 77 | - return; |
73 | + } | 78 | + goto err; |
79 | } | ||
80 | |||
81 | - req = g_new0(VuBlockReq, 1); | ||
82 | - req->server = server; | ||
83 | - req->vq = vq; | ||
84 | - req->elem = elem; | ||
85 | - | ||
86 | if (unlikely(iov_to_buf(out_iov, out_num, 0, &req->out, | ||
87 | sizeof(req->out)) != sizeof(req->out))) { | ||
88 | error_report("virtio-blk request outhdr too short"); | ||
89 | @@ -XXX,XX +XXX,XX @@ static void coroutine_fn vu_block_virtio_process_req(void *opaque) | ||
90 | |||
91 | err: | ||
92 | free(elem); | ||
93 | - g_free(req); | ||
94 | - return; | ||
95 | } | ||
96 | |||
97 | static void vu_block_process_vq(VuDev *vu_dev, int idx) | ||
98 | { | ||
99 | - VuServer *server; | ||
100 | - VuVirtq *vq; | ||
101 | - struct req_data *req_data; | ||
102 | + VuServer *server = container_of(vu_dev, VuServer, vu_dev); | ||
103 | + VuVirtq *vq = vu_get_queue(vu_dev, idx); | ||
104 | |||
105 | - server = container_of(vu_dev, VuServer, vu_dev); | ||
106 | - assert(server); | ||
107 | - | ||
108 | - vq = vu_get_queue(vu_dev, idx); | ||
109 | - assert(vq); | ||
110 | - VuVirtqElement *elem; | ||
111 | while (1) { | ||
112 | - elem = vu_queue_pop(vu_dev, vq, sizeof(VuVirtqElement) + | ||
113 | - sizeof(VuBlockReq)); | ||
114 | - if (elem) { | ||
115 | - req_data = g_new0(struct req_data, 1); | ||
116 | - req_data->server = server; | ||
117 | - req_data->vq = vq; | ||
118 | - req_data->elem = elem; | ||
119 | - Coroutine *co = qemu_coroutine_create(vu_block_virtio_process_req, | ||
120 | - req_data); | ||
121 | - aio_co_enter(server->ioc->ctx, co); | ||
122 | - } else { | ||
123 | + VuBlockReq *req; | ||
74 | + | 124 | + |
75 | return ret; | 125 | + req = vu_queue_pop(vu_dev, vq, sizeof(VuBlockReq)); |
126 | + if (!req) { | ||
127 | break; | ||
128 | } | ||
129 | + | ||
130 | + req->server = server; | ||
131 | + req->vq = vq; | ||
132 | + | ||
133 | + Coroutine *co = | ||
134 | + qemu_coroutine_create(vu_block_virtio_process_req, req); | ||
135 | + qemu_coroutine_enter(co); | ||
136 | } | ||
76 | } | 137 | } |
77 | 138 | ||
78 | -- | 139 | -- |
79 | 2.40.1 | 140 | 2.26.2 |
141 | diff view generated by jsdifflib |
New patch | |||
---|---|---|---|
1 | The device panic notifier callback is not used. Drop it. | ||
1 | 2 | ||
3 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
4 | Message-id: 20200924151549.913737-7-stefanha@redhat.com | ||
5 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
6 | --- | ||
7 | util/vhost-user-server.h | 3 --- | ||
8 | block/export/vhost-user-blk-server.c | 3 +-- | ||
9 | util/vhost-user-server.c | 6 ------ | ||
10 | 3 files changed, 1 insertion(+), 11 deletions(-) | ||
11 | |||
12 | diff --git a/util/vhost-user-server.h b/util/vhost-user-server.h | ||
13 | index XXXXXXX..XXXXXXX 100644 | ||
14 | --- a/util/vhost-user-server.h | ||
15 | +++ b/util/vhost-user-server.h | ||
16 | @@ -XXX,XX +XXX,XX @@ typedef struct VuFdWatch { | ||
17 | } VuFdWatch; | ||
18 | |||
19 | typedef struct VuServer VuServer; | ||
20 | -typedef void DevicePanicNotifierFn(VuServer *server); | ||
21 | |||
22 | struct VuServer { | ||
23 | QIONetListener *listener; | ||
24 | AioContext *ctx; | ||
25 | - DevicePanicNotifierFn *device_panic_notifier; | ||
26 | int max_queues; | ||
27 | const VuDevIface *vu_iface; | ||
28 | VuDev vu_dev; | ||
29 | @@ -XXX,XX +XXX,XX @@ bool vhost_user_server_start(VuServer *server, | ||
30 | SocketAddress *unix_socket, | ||
31 | AioContext *ctx, | ||
32 | uint16_t max_queues, | ||
33 | - DevicePanicNotifierFn *device_panic_notifier, | ||
34 | const VuDevIface *vu_iface, | ||
35 | Error **errp); | ||
36 | |||
37 | diff --git a/block/export/vhost-user-blk-server.c b/block/export/vhost-user-blk-server.c | ||
38 | index XXXXXXX..XXXXXXX 100644 | ||
39 | --- a/block/export/vhost-user-blk-server.c | ||
40 | +++ b/block/export/vhost-user-blk-server.c | ||
41 | @@ -XXX,XX +XXX,XX @@ static void vhost_user_blk_server_start(VuBlockDev *vu_block_device, | ||
42 | ctx = bdrv_get_aio_context(blk_bs(vu_block_device->backend)); | ||
43 | |||
44 | if (!vhost_user_server_start(&vu_block_device->vu_server, addr, ctx, | ||
45 | - VHOST_USER_BLK_MAX_QUEUES, | ||
46 | - NULL, &vu_block_iface, | ||
47 | + VHOST_USER_BLK_MAX_QUEUES, &vu_block_iface, | ||
48 | errp)) { | ||
49 | goto error; | ||
50 | } | ||
51 | diff --git a/util/vhost-user-server.c b/util/vhost-user-server.c | ||
52 | index XXXXXXX..XXXXXXX 100644 | ||
53 | --- a/util/vhost-user-server.c | ||
54 | +++ b/util/vhost-user-server.c | ||
55 | @@ -XXX,XX +XXX,XX @@ static void panic_cb(VuDev *vu_dev, const char *buf) | ||
56 | close_client(server); | ||
57 | } | ||
58 | |||
59 | - if (server->device_panic_notifier) { | ||
60 | - server->device_panic_notifier(server); | ||
61 | - } | ||
62 | - | ||
63 | /* | ||
64 | * Set the callback function for network listener so another | ||
65 | * vhost-user client can connect to this server | ||
66 | @@ -XXX,XX +XXX,XX @@ bool vhost_user_server_start(VuServer *server, | ||
67 | SocketAddress *socket_addr, | ||
68 | AioContext *ctx, | ||
69 | uint16_t max_queues, | ||
70 | - DevicePanicNotifierFn *device_panic_notifier, | ||
71 | const VuDevIface *vu_iface, | ||
72 | Error **errp) | ||
73 | { | ||
74 | @@ -XXX,XX +XXX,XX @@ bool vhost_user_server_start(VuServer *server, | ||
75 | .vu_iface = vu_iface, | ||
76 | .max_queues = max_queues, | ||
77 | .ctx = ctx, | ||
78 | - .device_panic_notifier = device_panic_notifier, | ||
79 | }; | ||
80 | |||
81 | qio_net_listener_set_name(server->listener, "vhost-user-backend-listener"); | ||
82 | -- | ||
83 | 2.26.2 | ||
84 | diff view generated by jsdifflib |
New patch | |||
---|---|---|---|
1 | fds[] is leaked when qio_channel_readv_full() fails. | ||
1 | 2 | ||
3 | Use vmsg->fds[] instead of keeping a local fds[] array. Then we can | ||
4 | reuse goto fail to clean up fds. vmsg->fd_num must be zeroed before the | ||
5 | loop to make this safe. | ||
6 | |||
7 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
8 | Message-id: 20200924151549.913737-8-stefanha@redhat.com | ||
9 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
10 | --- | ||
11 | util/vhost-user-server.c | 50 ++++++++++++++++++---------------------- | ||
12 | 1 file changed, 23 insertions(+), 27 deletions(-) | ||
13 | |||
14 | diff --git a/util/vhost-user-server.c b/util/vhost-user-server.c | ||
15 | index XXXXXXX..XXXXXXX 100644 | ||
16 | --- a/util/vhost-user-server.c | ||
17 | +++ b/util/vhost-user-server.c | ||
18 | @@ -XXX,XX +XXX,XX @@ vu_message_read(VuDev *vu_dev, int conn_fd, VhostUserMsg *vmsg) | ||
19 | }; | ||
20 | int rc, read_bytes = 0; | ||
21 | Error *local_err = NULL; | ||
22 | - /* | ||
23 | - * Store fds/nfds returned from qio_channel_readv_full into | ||
24 | - * temporary variables. | ||
25 | - * | ||
26 | - * VhostUserMsg is a packed structure, gcc will complain about passing | ||
27 | - * pointer to a packed structure member if we pass &VhostUserMsg.fd_num | ||
28 | - * and &VhostUserMsg.fds directly when calling qio_channel_readv_full, | ||
29 | - * thus two temporary variables nfds and fds are used here. | ||
30 | - */ | ||
31 | - size_t nfds = 0, nfds_t = 0; | ||
32 | const size_t max_fds = G_N_ELEMENTS(vmsg->fds); | ||
33 | - int *fds_t = NULL; | ||
34 | VuServer *server = container_of(vu_dev, VuServer, vu_dev); | ||
35 | QIOChannel *ioc = server->ioc; | ||
36 | |||
37 | + vmsg->fd_num = 0; | ||
38 | if (!ioc) { | ||
39 | error_report_err(local_err); | ||
40 | goto fail; | ||
41 | @@ -XXX,XX +XXX,XX @@ vu_message_read(VuDev *vu_dev, int conn_fd, VhostUserMsg *vmsg) | ||
42 | |||
43 | assert(qemu_in_coroutine()); | ||
44 | do { | ||
45 | + size_t nfds = 0; | ||
46 | + int *fds = NULL; | ||
47 | + | ||
48 | /* | ||
49 | * qio_channel_readv_full may have short reads, keeping calling it | ||
50 | * until getting VHOST_USER_HDR_SIZE or 0 bytes in total | ||
51 | */ | ||
52 | - rc = qio_channel_readv_full(ioc, &iov, 1, &fds_t, &nfds_t, &local_err); | ||
53 | + rc = qio_channel_readv_full(ioc, &iov, 1, &fds, &nfds, &local_err); | ||
54 | if (rc < 0) { | ||
55 | if (rc == QIO_CHANNEL_ERR_BLOCK) { | ||
56 | + assert(local_err == NULL); | ||
57 | qio_channel_yield(ioc, G_IO_IN); | ||
58 | continue; | ||
59 | } else { | ||
60 | error_report_err(local_err); | ||
61 | - return false; | ||
62 | + goto fail; | ||
63 | } | ||
64 | } | ||
65 | - read_bytes += rc; | ||
66 | - if (nfds_t > 0) { | ||
67 | - if (nfds + nfds_t > max_fds) { | ||
68 | + | ||
69 | + if (nfds > 0) { | ||
70 | + if (vmsg->fd_num + nfds > max_fds) { | ||
71 | error_report("A maximum of %zu fds are allowed, " | ||
72 | "however got %zu fds now", | ||
73 | - max_fds, nfds + nfds_t); | ||
74 | + max_fds, vmsg->fd_num + nfds); | ||
75 | + g_free(fds); | ||
76 | goto fail; | ||
77 | } | ||
78 | - memcpy(vmsg->fds + nfds, fds_t, | ||
79 | - nfds_t *sizeof(vmsg->fds[0])); | ||
80 | - nfds += nfds_t; | ||
81 | - g_free(fds_t); | ||
82 | + memcpy(vmsg->fds + vmsg->fd_num, fds, nfds * sizeof(vmsg->fds[0])); | ||
83 | + vmsg->fd_num += nfds; | ||
84 | + g_free(fds); | ||
85 | } | ||
86 | - if (read_bytes == VHOST_USER_HDR_SIZE || rc == 0) { | ||
87 | - break; | ||
88 | + | ||
89 | + if (rc == 0) { /* socket closed */ | ||
90 | + goto fail; | ||
91 | } | ||
92 | - iov.iov_base = (char *)vmsg + read_bytes; | ||
93 | - iov.iov_len = VHOST_USER_HDR_SIZE - read_bytes; | ||
94 | - } while (true); | ||
95 | |||
96 | - vmsg->fd_num = nfds; | ||
97 | + iov.iov_base += rc; | ||
98 | + iov.iov_len -= rc; | ||
99 | + read_bytes += rc; | ||
100 | + } while (read_bytes != VHOST_USER_HDR_SIZE); | ||
101 | + | ||
102 | /* qio_channel_readv_full will make socket fds blocking, unblock them */ | ||
103 | vmsg_unblock_fds(vmsg); | ||
104 | if (vmsg->size > sizeof(vmsg->payload)) { | ||
105 | -- | ||
106 | 2.26.2 | ||
107 | diff view generated by jsdifflib |
1 | From: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | 1 | Unexpected EOF is an error that must be reported. |
---|---|---|---|
2 | 2 | ||
3 | Set data_end to the end of the last cluster inside the image. In such a | 3 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
4 | way we can be sure that corrupted offsets in the BAT can't affect on the | 4 | Message-id: 20200924151549.913737-9-stefanha@redhat.com |
5 | image size. If there are no allocated clusters set image_end_offset by | 5 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
6 | data_end. | 6 | --- |
7 | util/vhost-user-server.c | 6 ++++-- | ||
8 | 1 file changed, 4 insertions(+), 2 deletions(-) | ||
7 | 9 | ||
8 | Signed-off-by: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | 10 | diff --git a/util/vhost-user-server.c b/util/vhost-user-server.c |
9 | Reviewed-by: Denis V. Lunev <den@openvz.org> | ||
10 | Message-Id: <20230424093147.197643-4-alexander.ivanov@virtuozzo.com> | ||
11 | Reviewed-by: Hanna Czenczek <hreitz@redhat.com> | ||
12 | Signed-off-by: Hanna Czenczek <hreitz@redhat.com> | ||
13 | --- | ||
14 | block/parallels.c | 8 +++++++- | ||
15 | 1 file changed, 7 insertions(+), 1 deletion(-) | ||
16 | |||
17 | diff --git a/block/parallels.c b/block/parallels.c | ||
18 | index XXXXXXX..XXXXXXX 100644 | 11 | index XXXXXXX..XXXXXXX 100644 |
19 | --- a/block/parallels.c | 12 | --- a/util/vhost-user-server.c |
20 | +++ b/block/parallels.c | 13 | +++ b/util/vhost-user-server.c |
21 | @@ -XXX,XX +XXX,XX @@ parallels_co_check(BlockDriverState *bs, BdrvCheckResult *res, | 14 | @@ -XXX,XX +XXX,XX @@ vu_message_read(VuDev *vu_dev, int conn_fd, VhostUserMsg *vmsg) |
15 | }; | ||
16 | if (vmsg->size) { | ||
17 | rc = qio_channel_readv_all_eof(ioc, &iov_payload, 1, &local_err); | ||
18 | - if (rc == -1) { | ||
19 | - error_report_err(local_err); | ||
20 | + if (rc != 1) { | ||
21 | + if (local_err) { | ||
22 | + error_report_err(local_err); | ||
23 | + } | ||
24 | goto fail; | ||
22 | } | 25 | } |
23 | } | 26 | } |
24 | |||
25 | - res->image_end_offset = high_off + s->cluster_size; | ||
26 | + if (high_off == 0) { | ||
27 | + res->image_end_offset = s->data_end << BDRV_SECTOR_BITS; | ||
28 | + } else { | ||
29 | + res->image_end_offset = high_off + s->cluster_size; | ||
30 | + s->data_end = res->image_end_offset >> BDRV_SECTOR_BITS; | ||
31 | + } | ||
32 | + | ||
33 | if (size > res->image_end_offset) { | ||
34 | int64_t count; | ||
35 | count = DIV_ROUND_UP(size - res->image_end_offset, s->cluster_size); | ||
36 | -- | 27 | -- |
37 | 2.40.1 | 28 | 2.26.2 |
29 | diff view generated by jsdifflib |
1 | From: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | 1 | The vu_client_trip() coroutine is leaked during AioContext switching. It |
---|---|---|---|
2 | is also unsafe to destroy the vu_dev in panic_cb() since its callers | ||
3 | still access it in some cases. | ||
2 | 4 | ||
3 | We will add more and more checks so we need a better code structure | 5 | Rework the lifecycle to solve these safety issues. |
4 | in parallels_co_check. Let each check performs in a separate loop | ||
5 | in a separate helper. | ||
6 | 6 | ||
7 | Signed-off-by: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | 7 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
8 | Reviewed-by: Denis V. Lunev <den@openvz.org> | 8 | Message-id: 20200924151549.913737-10-stefanha@redhat.com |
9 | Reviewed-by: Vladimir Sementsov-Ogievskiy <vsementsov@yandex-team.ru> | 9 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
10 | Message-Id: <20230424093147.197643-11-alexander.ivanov@virtuozzo.com> | ||
11 | Reviewed-by: Hanna Czenczek <hreitz@redhat.com> | ||
12 | Signed-off-by: Hanna Czenczek <hreitz@redhat.com> | ||
13 | --- | 10 | --- |
14 | block/parallels.c | 52 +++++++++++++++++++++++++++-------------------- | 11 | util/vhost-user-server.h | 29 ++-- |
15 | 1 file changed, 30 insertions(+), 22 deletions(-) | 12 | block/export/vhost-user-blk-server.c | 9 +- |
13 | util/vhost-user-server.c | 245 +++++++++++++++------------ | ||
14 | 3 files changed, 155 insertions(+), 128 deletions(-) | ||
16 | 15 | ||
17 | diff --git a/block/parallels.c b/block/parallels.c | 16 | diff --git a/util/vhost-user-server.h b/util/vhost-user-server.h |
18 | index XXXXXXX..XXXXXXX 100644 | 17 | index XXXXXXX..XXXXXXX 100644 |
19 | --- a/block/parallels.c | 18 | --- a/util/vhost-user-server.h |
20 | +++ b/block/parallels.c | 19 | +++ b/util/vhost-user-server.h |
21 | @@ -XXX,XX +XXX,XX @@ parallels_check_leak(BlockDriverState *bs, BdrvCheckResult *res, | 20 | @@ -XXX,XX +XXX,XX @@ |
22 | return 0; | 21 | #include "qapi/error.h" |
23 | } | 22 | #include "standard-headers/linux/virtio_blk.h" |
24 | 23 | ||
25 | -static int coroutine_fn GRAPH_RDLOCK | 24 | +/* A kick fd that we monitor on behalf of libvhost-user */ |
26 | -parallels_co_check(BlockDriverState *bs, BdrvCheckResult *res, | 25 | typedef struct VuFdWatch { |
27 | - BdrvCheckMode fix) | 26 | VuDev *vu_dev; |
28 | +static void parallels_collect_statistics(BlockDriverState *bs, | 27 | int fd; /*kick fd*/ |
29 | + BdrvCheckResult *res, | 28 | void *pvt; |
30 | + BdrvCheckMode fix) | 29 | vu_watch_cb cb; |
31 | { | 30 | - bool processing; |
32 | BDRVParallelsState *s = bs->opaque; | 31 | QTAILQ_ENTRY(VuFdWatch) next; |
33 | - int64_t prev_off; | 32 | } VuFdWatch; |
34 | - int ret; | 33 | |
35 | + int64_t off, prev_off; | 34 | -typedef struct VuServer VuServer; |
36 | uint32_t i; | 35 | - |
37 | 36 | -struct VuServer { | |
38 | - qemu_co_mutex_lock(&s->lock); | 37 | +/** |
39 | - | 38 | + * VuServer: |
40 | - parallels_check_unclean(bs, res, fix); | 39 | + * A vhost-user server instance with user-defined VuDevIface callbacks. |
41 | - | 40 | + * Vhost-user device backends can be implemented using VuServer. VuDevIface |
42 | - ret = parallels_check_outside_image(bs, res, fix); | 41 | + * callbacks and virtqueue kicks run in the given AioContext. |
43 | - if (ret < 0) { | 42 | + */ |
44 | - goto out; | 43 | +typedef struct { |
44 | QIONetListener *listener; | ||
45 | + QEMUBH *restart_listener_bh; | ||
46 | AioContext *ctx; | ||
47 | int max_queues; | ||
48 | const VuDevIface *vu_iface; | ||
49 | + | ||
50 | + /* Protected by ctx lock */ | ||
51 | VuDev vu_dev; | ||
52 | QIOChannel *ioc; /* The I/O channel with the client */ | ||
53 | QIOChannelSocket *sioc; /* The underlying data channel with the client */ | ||
54 | - /* IOChannel for fd provided via VHOST_USER_SET_SLAVE_REQ_FD */ | ||
55 | - QIOChannel *ioc_slave; | ||
56 | - QIOChannelSocket *sioc_slave; | ||
57 | - Coroutine *co_trip; /* coroutine for processing VhostUserMsg */ | ||
58 | QTAILQ_HEAD(, VuFdWatch) vu_fd_watches; | ||
59 | - /* restart coroutine co_trip if AIOContext is changed */ | ||
60 | - bool aio_context_changed; | ||
61 | - bool processing_msg; | ||
62 | -}; | ||
63 | + | ||
64 | + Coroutine *co_trip; /* coroutine for processing VhostUserMsg */ | ||
65 | +} VuServer; | ||
66 | |||
67 | bool vhost_user_server_start(VuServer *server, | ||
68 | SocketAddress *unix_socket, | ||
69 | @@ -XXX,XX +XXX,XX @@ bool vhost_user_server_start(VuServer *server, | ||
70 | |||
71 | void vhost_user_server_stop(VuServer *server); | ||
72 | |||
73 | -void vhost_user_server_set_aio_context(VuServer *server, AioContext *ctx); | ||
74 | +void vhost_user_server_attach_aio_context(VuServer *server, AioContext *ctx); | ||
75 | +void vhost_user_server_detach_aio_context(VuServer *server); | ||
76 | |||
77 | #endif /* VHOST_USER_SERVER_H */ | ||
78 | diff --git a/block/export/vhost-user-blk-server.c b/block/export/vhost-user-blk-server.c | ||
79 | index XXXXXXX..XXXXXXX 100644 | ||
80 | --- a/block/export/vhost-user-blk-server.c | ||
81 | +++ b/block/export/vhost-user-blk-server.c | ||
82 | @@ -XXX,XX +XXX,XX @@ static const VuDevIface vu_block_iface = { | ||
83 | static void blk_aio_attached(AioContext *ctx, void *opaque) | ||
84 | { | ||
85 | VuBlockDev *vub_dev = opaque; | ||
86 | - aio_context_acquire(ctx); | ||
87 | - vhost_user_server_set_aio_context(&vub_dev->vu_server, ctx); | ||
88 | - aio_context_release(ctx); | ||
89 | + vhost_user_server_attach_aio_context(&vub_dev->vu_server, ctx); | ||
90 | } | ||
91 | |||
92 | static void blk_aio_detach(void *opaque) | ||
93 | { | ||
94 | VuBlockDev *vub_dev = opaque; | ||
95 | - AioContext *ctx = vub_dev->vu_server.ctx; | ||
96 | - aio_context_acquire(ctx); | ||
97 | - vhost_user_server_set_aio_context(&vub_dev->vu_server, NULL); | ||
98 | - aio_context_release(ctx); | ||
99 | + vhost_user_server_detach_aio_context(&vub_dev->vu_server); | ||
100 | } | ||
101 | |||
102 | static void | ||
103 | diff --git a/util/vhost-user-server.c b/util/vhost-user-server.c | ||
104 | index XXXXXXX..XXXXXXX 100644 | ||
105 | --- a/util/vhost-user-server.c | ||
106 | +++ b/util/vhost-user-server.c | ||
107 | @@ -XXX,XX +XXX,XX @@ | ||
108 | */ | ||
109 | #include "qemu/osdep.h" | ||
110 | #include "qemu/main-loop.h" | ||
111 | +#include "block/aio-wait.h" | ||
112 | #include "vhost-user-server.h" | ||
113 | |||
114 | +/* | ||
115 | + * Theory of operation: | ||
116 | + * | ||
117 | + * VuServer is started and stopped by vhost_user_server_start() and | ||
118 | + * vhost_user_server_stop() from the main loop thread. Starting the server | ||
119 | + * opens a vhost-user UNIX domain socket and listens for incoming connections. | ||
120 | + * Only one connection is allowed at a time. | ||
121 | + * | ||
122 | + * The connection is handled by the vu_client_trip() coroutine in the | ||
123 | + * VuServer->ctx AioContext. The coroutine consists of a vu_dispatch() loop | ||
124 | + * where libvhost-user calls vu_message_read() to receive the next vhost-user | ||
125 | + * protocol messages over the UNIX domain socket. | ||
126 | + * | ||
127 | + * When virtqueues are set up libvhost-user calls set_watch() to monitor kick | ||
128 | + * fds. These fds are also handled in the VuServer->ctx AioContext. | ||
129 | + * | ||
130 | + * Both vu_client_trip() and kick fd monitoring can be stopped by shutting down | ||
131 | + * the socket connection. Shutting down the socket connection causes | ||
132 | + * vu_message_read() to fail since no more data can be received from the socket. | ||
133 | + * After vu_dispatch() fails, vu_client_trip() calls vu_deinit() to stop | ||
134 | + * libvhost-user before terminating the coroutine. vu_deinit() calls | ||
135 | + * remove_watch() to stop monitoring kick fds and this stops virtqueue | ||
136 | + * processing. | ||
137 | + * | ||
138 | + * When vu_client_trip() has finished cleaning up it schedules a BH in the main | ||
139 | + * loop thread to accept the next client connection. | ||
140 | + * | ||
141 | + * When libvhost-user detects an error it calls panic_cb() and sets the | ||
142 | + * dev->broken flag. Both vu_client_trip() and kick fd processing stop when | ||
143 | + * the dev->broken flag is set. | ||
144 | + * | ||
145 | + * It is possible to switch AioContexts using | ||
146 | + * vhost_user_server_detach_aio_context() and | ||
147 | + * vhost_user_server_attach_aio_context(). They stop monitoring fds in the old | ||
148 | + * AioContext and resume monitoring in the new AioContext. The vu_client_trip() | ||
149 | + * coroutine remains in a yielded state during the switch. This is made | ||
150 | + * possible by QIOChannel's support for spurious coroutine re-entry in | ||
151 | + * qio_channel_yield(). The coroutine will restart I/O when re-entered from the | ||
152 | + * new AioContext. | ||
153 | + */ | ||
154 | + | ||
155 | static void vmsg_close_fds(VhostUserMsg *vmsg) | ||
156 | { | ||
157 | int i; | ||
158 | @@ -XXX,XX +XXX,XX @@ static void vmsg_unblock_fds(VhostUserMsg *vmsg) | ||
159 | } | ||
160 | } | ||
161 | |||
162 | -static void vu_accept(QIONetListener *listener, QIOChannelSocket *sioc, | ||
163 | - gpointer opaque); | ||
164 | - | ||
165 | -static void close_client(VuServer *server) | ||
166 | -{ | ||
167 | - /* | ||
168 | - * Before closing the client | ||
169 | - * | ||
170 | - * 1. Let vu_client_trip stop processing new vhost-user msg | ||
171 | - * | ||
172 | - * 2. remove kick_handler | ||
173 | - * | ||
174 | - * 3. wait for the kick handler to be finished | ||
175 | - * | ||
176 | - * 4. wait for the current vhost-user msg to be finished processing | ||
177 | - */ | ||
178 | - | ||
179 | - QIOChannelSocket *sioc = server->sioc; | ||
180 | - /* When this is set vu_client_trip will stop new processing vhost-user message */ | ||
181 | - server->sioc = NULL; | ||
182 | - | ||
183 | - while (server->processing_msg) { | ||
184 | - if (server->ioc->read_coroutine) { | ||
185 | - server->ioc->read_coroutine = NULL; | ||
186 | - qio_channel_set_aio_fd_handler(server->ioc, server->ioc->ctx, NULL, | ||
187 | - NULL, server->ioc); | ||
188 | - server->processing_msg = false; | ||
189 | - } | ||
45 | - } | 190 | - } |
46 | - | 191 | - |
47 | - ret = parallels_check_leak(bs, res, fix); | 192 | - vu_deinit(&server->vu_dev); |
48 | - if (ret < 0) { | 193 | - |
49 | - goto out; | 194 | - /* vu_deinit() should have called remove_watch() */ |
195 | - assert(QTAILQ_EMPTY(&server->vu_fd_watches)); | ||
196 | - | ||
197 | - object_unref(OBJECT(sioc)); | ||
198 | - object_unref(OBJECT(server->ioc)); | ||
199 | -} | ||
200 | - | ||
201 | static void panic_cb(VuDev *vu_dev, const char *buf) | ||
202 | { | ||
203 | - VuServer *server = container_of(vu_dev, VuServer, vu_dev); | ||
204 | - | ||
205 | - /* avoid while loop in close_client */ | ||
206 | - server->processing_msg = false; | ||
207 | - | ||
208 | - if (buf) { | ||
209 | - error_report("vu_panic: %s", buf); | ||
50 | - } | 210 | - } |
51 | - | 211 | - |
52 | res->bfi.total_clusters = s->bat_size; | 212 | - if (server->sioc) { |
53 | res->bfi.compressed_clusters = 0; /* compression is not supported */ | 213 | - close_client(server); |
54 | 214 | - } | |
55 | prev_off = 0; | 215 | - |
56 | for (i = 0; i < s->bat_size; i++) { | 216 | - /* |
57 | - int64_t off = bat2sect(s, i) << BDRV_SECTOR_BITS; | 217 | - * Set the callback function for network listener so another |
58 | + off = bat2sect(s, i) << BDRV_SECTOR_BITS; | 218 | - * vhost-user client can connect to this server |
59 | /* | 219 | - */ |
60 | * If BDRV_FIX_ERRORS is not set, out-of-image BAT entries were not | 220 | - qio_net_listener_set_client_func(server->listener, |
61 | * fixed. Skip not allocated and out-of-image BAT entries. | 221 | - vu_accept, |
62 | @@ -XXX,XX +XXX,XX @@ parallels_co_check(BlockDriverState *bs, BdrvCheckResult *res, | 222 | - server, |
63 | continue; | 223 | - NULL); |
64 | } | 224 | + error_report("vu_panic: %s", buf); |
65 | 225 | } | |
66 | - res->bfi.allocated_clusters++; | 226 | |
67 | - | 227 | static bool coroutine_fn |
68 | if (prev_off != 0 && (prev_off + s->cluster_size) != off) { | 228 | @@ -XXX,XX +XXX,XX @@ fail: |
69 | res->bfi.fragmented_clusters++; | 229 | return false; |
70 | } | 230 | } |
71 | prev_off = off; | 231 | |
72 | + res->bfi.allocated_clusters++; | 232 | - |
233 | -static void vu_client_start(VuServer *server); | ||
234 | static coroutine_fn void vu_client_trip(void *opaque) | ||
235 | { | ||
236 | VuServer *server = opaque; | ||
237 | + VuDev *vu_dev = &server->vu_dev; | ||
238 | |||
239 | - while (!server->aio_context_changed && server->sioc) { | ||
240 | - server->processing_msg = true; | ||
241 | - vu_dispatch(&server->vu_dev); | ||
242 | - server->processing_msg = false; | ||
243 | + while (!vu_dev->broken && vu_dispatch(vu_dev)) { | ||
244 | + /* Keep running */ | ||
245 | } | ||
246 | |||
247 | - if (server->aio_context_changed && server->sioc) { | ||
248 | - server->aio_context_changed = false; | ||
249 | - vu_client_start(server); | ||
250 | - } | ||
251 | -} | ||
252 | + vu_deinit(vu_dev); | ||
253 | + | ||
254 | + /* vu_deinit() should have called remove_watch() */ | ||
255 | + assert(QTAILQ_EMPTY(&server->vu_fd_watches)); | ||
256 | + | ||
257 | + object_unref(OBJECT(server->sioc)); | ||
258 | + server->sioc = NULL; | ||
259 | |||
260 | -static void vu_client_start(VuServer *server) | ||
261 | -{ | ||
262 | - server->co_trip = qemu_coroutine_create(vu_client_trip, server); | ||
263 | - aio_co_enter(server->ctx, server->co_trip); | ||
264 | + object_unref(OBJECT(server->ioc)); | ||
265 | + server->ioc = NULL; | ||
266 | + | ||
267 | + server->co_trip = NULL; | ||
268 | + if (server->restart_listener_bh) { | ||
269 | + qemu_bh_schedule(server->restart_listener_bh); | ||
270 | + } | ||
271 | + aio_wait_kick(); | ||
272 | } | ||
273 | |||
274 | /* | ||
275 | @@ -XXX,XX +XXX,XX @@ static void vu_client_start(VuServer *server) | ||
276 | static void kick_handler(void *opaque) | ||
277 | { | ||
278 | VuFdWatch *vu_fd_watch = opaque; | ||
279 | - vu_fd_watch->processing = true; | ||
280 | - vu_fd_watch->cb(vu_fd_watch->vu_dev, 0, vu_fd_watch->pvt); | ||
281 | - vu_fd_watch->processing = false; | ||
282 | + VuDev *vu_dev = vu_fd_watch->vu_dev; | ||
283 | + | ||
284 | + vu_fd_watch->cb(vu_dev, 0, vu_fd_watch->pvt); | ||
285 | + | ||
286 | + /* Stop vu_client_trip() if an error occurred in vu_fd_watch->cb() */ | ||
287 | + if (vu_dev->broken) { | ||
288 | + VuServer *server = container_of(vu_dev, VuServer, vu_dev); | ||
289 | + | ||
290 | + qio_channel_shutdown(server->ioc, QIO_CHANNEL_SHUTDOWN_BOTH, NULL); | ||
291 | + } | ||
292 | } | ||
293 | |||
294 | - | ||
295 | static VuFdWatch *find_vu_fd_watch(VuServer *server, int fd) | ||
296 | { | ||
297 | |||
298 | @@ -XXX,XX +XXX,XX @@ static void vu_accept(QIONetListener *listener, QIOChannelSocket *sioc, | ||
299 | qio_channel_set_name(QIO_CHANNEL(sioc), "vhost-user client"); | ||
300 | server->ioc = QIO_CHANNEL(sioc); | ||
301 | object_ref(OBJECT(server->ioc)); | ||
302 | - qio_channel_attach_aio_context(server->ioc, server->ctx); | ||
303 | + | ||
304 | + /* TODO vu_message_write() spins if non-blocking! */ | ||
305 | qio_channel_set_blocking(server->ioc, false, NULL); | ||
306 | - vu_client_start(server); | ||
307 | + | ||
308 | + server->co_trip = qemu_coroutine_create(vu_client_trip, server); | ||
309 | + | ||
310 | + aio_context_acquire(server->ctx); | ||
311 | + vhost_user_server_attach_aio_context(server, server->ctx); | ||
312 | + aio_context_release(server->ctx); | ||
313 | } | ||
314 | |||
315 | - | ||
316 | void vhost_user_server_stop(VuServer *server) | ||
317 | { | ||
318 | + aio_context_acquire(server->ctx); | ||
319 | + | ||
320 | + qemu_bh_delete(server->restart_listener_bh); | ||
321 | + server->restart_listener_bh = NULL; | ||
322 | + | ||
323 | if (server->sioc) { | ||
324 | - close_client(server); | ||
325 | + VuFdWatch *vu_fd_watch; | ||
326 | + | ||
327 | + QTAILQ_FOREACH(vu_fd_watch, &server->vu_fd_watches, next) { | ||
328 | + aio_set_fd_handler(server->ctx, vu_fd_watch->fd, true, | ||
329 | + NULL, NULL, NULL, vu_fd_watch); | ||
330 | + } | ||
331 | + | ||
332 | + qio_channel_shutdown(server->ioc, QIO_CHANNEL_SHUTDOWN_BOTH, NULL); | ||
333 | + | ||
334 | + AIO_WAIT_WHILE(server->ctx, server->co_trip); | ||
335 | } | ||
336 | |||
337 | + aio_context_release(server->ctx); | ||
338 | + | ||
339 | if (server->listener) { | ||
340 | qio_net_listener_disconnect(server->listener); | ||
341 | object_unref(OBJECT(server->listener)); | ||
73 | } | 342 | } |
74 | +} | 343 | +} |
75 | + | 344 | + |
76 | +static int coroutine_fn GRAPH_RDLOCK | 345 | +/* |
77 | +parallels_co_check(BlockDriverState *bs, BdrvCheckResult *res, | 346 | + * Allow the next client to connect to the server. Called from a BH in the main |
78 | + BdrvCheckMode fix) | 347 | + * loop. |
348 | + */ | ||
349 | +static void restart_listener_bh(void *opaque) | ||
79 | +{ | 350 | +{ |
80 | + BDRVParallelsState *s = bs->opaque; | 351 | + VuServer *server = opaque; |
81 | + int ret; | 352 | |
82 | + | 353 | + qio_net_listener_set_client_func(server->listener, vu_accept, server, |
83 | + qemu_co_mutex_lock(&s->lock); | 354 | + NULL); |
84 | + | 355 | } |
85 | + parallels_check_unclean(bs, res, fix); | 356 | |
86 | + | 357 | -void vhost_user_server_set_aio_context(VuServer *server, AioContext *ctx) |
87 | + ret = parallels_check_outside_image(bs, res, fix); | 358 | +/* Called with ctx acquired */ |
88 | + if (ret < 0) { | 359 | +void vhost_user_server_attach_aio_context(VuServer *server, AioContext *ctx) |
89 | + goto out; | 360 | { |
361 | - VuFdWatch *vu_fd_watch, *next; | ||
362 | - void *opaque = NULL; | ||
363 | - IOHandler *io_read = NULL; | ||
364 | - bool attach; | ||
365 | + VuFdWatch *vu_fd_watch; | ||
366 | |||
367 | - server->ctx = ctx ? ctx : qemu_get_aio_context(); | ||
368 | + server->ctx = ctx; | ||
369 | |||
370 | if (!server->sioc) { | ||
371 | - /* not yet serving any client*/ | ||
372 | return; | ||
373 | } | ||
374 | |||
375 | - if (ctx) { | ||
376 | - qio_channel_attach_aio_context(server->ioc, ctx); | ||
377 | - server->aio_context_changed = true; | ||
378 | - io_read = kick_handler; | ||
379 | - attach = true; | ||
380 | - } else { | ||
381 | + qio_channel_attach_aio_context(server->ioc, ctx); | ||
382 | + | ||
383 | + QTAILQ_FOREACH(vu_fd_watch, &server->vu_fd_watches, next) { | ||
384 | + aio_set_fd_handler(ctx, vu_fd_watch->fd, true, kick_handler, NULL, | ||
385 | + NULL, vu_fd_watch); | ||
90 | + } | 386 | + } |
91 | + | 387 | + |
92 | + ret = parallels_check_leak(bs, res, fix); | 388 | + aio_co_schedule(ctx, server->co_trip); |
93 | + if (ret < 0) { | 389 | +} |
94 | + goto out; | 390 | + |
95 | + } | 391 | +/* Called with server->ctx acquired */ |
96 | + | 392 | +void vhost_user_server_detach_aio_context(VuServer *server) |
97 | + parallels_collect_statistics(bs, res, fix); | 393 | +{ |
98 | 394 | + if (server->sioc) { | |
99 | out: | 395 | + VuFdWatch *vu_fd_watch; |
100 | qemu_co_mutex_unlock(&s->lock); | 396 | + |
397 | + QTAILQ_FOREACH(vu_fd_watch, &server->vu_fd_watches, next) { | ||
398 | + aio_set_fd_handler(server->ctx, vu_fd_watch->fd, true, | ||
399 | + NULL, NULL, NULL, vu_fd_watch); | ||
400 | + } | ||
401 | + | ||
402 | qio_channel_detach_aio_context(server->ioc); | ||
403 | - /* server->ioc->ctx keeps the old AioConext */ | ||
404 | - ctx = server->ioc->ctx; | ||
405 | - attach = false; | ||
406 | } | ||
407 | |||
408 | - QTAILQ_FOREACH_SAFE(vu_fd_watch, &server->vu_fd_watches, next, next) { | ||
409 | - if (vu_fd_watch->cb) { | ||
410 | - opaque = attach ? vu_fd_watch : NULL; | ||
411 | - aio_set_fd_handler(ctx, vu_fd_watch->fd, true, | ||
412 | - io_read, NULL, NULL, | ||
413 | - opaque); | ||
414 | - } | ||
415 | - } | ||
416 | + server->ctx = NULL; | ||
417 | } | ||
418 | |||
419 | - | ||
420 | bool vhost_user_server_start(VuServer *server, | ||
421 | SocketAddress *socket_addr, | ||
422 | AioContext *ctx, | ||
423 | @@ -XXX,XX +XXX,XX @@ bool vhost_user_server_start(VuServer *server, | ||
424 | const VuDevIface *vu_iface, | ||
425 | Error **errp) | ||
426 | { | ||
427 | + QEMUBH *bh; | ||
428 | QIONetListener *listener = qio_net_listener_new(); | ||
429 | if (qio_net_listener_open_sync(listener, socket_addr, 1, | ||
430 | errp) < 0) { | ||
431 | @@ -XXX,XX +XXX,XX @@ bool vhost_user_server_start(VuServer *server, | ||
432 | return false; | ||
433 | } | ||
434 | |||
435 | + bh = qemu_bh_new(restart_listener_bh, server); | ||
436 | + | ||
437 | /* zero out unspecified fields */ | ||
438 | *server = (VuServer) { | ||
439 | .listener = listener, | ||
440 | + .restart_listener_bh = bh, | ||
441 | .vu_iface = vu_iface, | ||
442 | .max_queues = max_queues, | ||
443 | .ctx = ctx, | ||
101 | -- | 444 | -- |
102 | 2.40.1 | 445 | 2.26.2 |
446 | diff view generated by jsdifflib |
New patch | |||
---|---|---|---|
1 | Propagate the flush return value since errors are possible. | ||
1 | 2 | ||
3 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
4 | Message-id: 20200924151549.913737-11-stefanha@redhat.com | ||
5 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
6 | --- | ||
7 | block/export/vhost-user-blk-server.c | 11 +++++++---- | ||
8 | 1 file changed, 7 insertions(+), 4 deletions(-) | ||
9 | |||
10 | diff --git a/block/export/vhost-user-blk-server.c b/block/export/vhost-user-blk-server.c | ||
11 | index XXXXXXX..XXXXXXX 100644 | ||
12 | --- a/block/export/vhost-user-blk-server.c | ||
13 | +++ b/block/export/vhost-user-blk-server.c | ||
14 | @@ -XXX,XX +XXX,XX @@ vu_block_discard_write_zeroes(VuBlockReq *req, struct iovec *iov, | ||
15 | return -EINVAL; | ||
16 | } | ||
17 | |||
18 | -static void coroutine_fn vu_block_flush(VuBlockReq *req) | ||
19 | +static int coroutine_fn vu_block_flush(VuBlockReq *req) | ||
20 | { | ||
21 | VuBlockDev *vdev_blk = get_vu_block_device_by_server(req->server); | ||
22 | BlockBackend *backend = vdev_blk->backend; | ||
23 | - blk_co_flush(backend); | ||
24 | + return blk_co_flush(backend); | ||
25 | } | ||
26 | |||
27 | static void coroutine_fn vu_block_virtio_process_req(void *opaque) | ||
28 | @@ -XXX,XX +XXX,XX @@ static void coroutine_fn vu_block_virtio_process_req(void *opaque) | ||
29 | break; | ||
30 | } | ||
31 | case VIRTIO_BLK_T_FLUSH: | ||
32 | - vu_block_flush(req); | ||
33 | - req->in->status = VIRTIO_BLK_S_OK; | ||
34 | + if (vu_block_flush(req) == 0) { | ||
35 | + req->in->status = VIRTIO_BLK_S_OK; | ||
36 | + } else { | ||
37 | + req->in->status = VIRTIO_BLK_S_IOERR; | ||
38 | + } | ||
39 | break; | ||
40 | case VIRTIO_BLK_T_GET_ID: { | ||
41 | size_t size = MIN(iov_size(&elem->in_sg[0], in_num), | ||
42 | -- | ||
43 | 2.26.2 | ||
44 | diff view generated by jsdifflib |
1 | From: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | 1 | Use the new QAPI block exports API instead of defining our own QOM |
---|---|---|---|
2 | 2 | objects. | |
3 | Replace the way we use mutex in parallels_co_check() for simplier | 3 | |
4 | and less error prone code. | 4 | This is a large change because the lifecycle of VuBlockDev needs to |
5 | 5 | follow BlockExportDriver. QOM properties are replaced by QAPI options | |
6 | Signed-off-by: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | 6 | objects. |
7 | Reviewed-by: Denis V. Lunev <den@openvz.org> | 7 | |
8 | Message-Id: <20230424093147.197643-12-alexander.ivanov@virtuozzo.com> | 8 | VuBlockDev is renamed VuBlkExport and contains a BlockExport field. |
9 | Reviewed-by: Hanna Czenczek <hreitz@redhat.com> | 9 | Several fields can be dropped since BlockExport already has equivalents. |
10 | Signed-off-by: Hanna Czenczek <hreitz@redhat.com> | 10 | |
11 | The file names and meson build integration will be adjusted in a future | ||
12 | patch. libvhost-user should probably be built as a static library that | ||
13 | is linked into QEMU instead of as a .c file that results in duplicate | ||
14 | compilation. | ||
15 | |||
16 | The new command-line syntax is: | ||
17 | |||
18 | $ qemu-storage-daemon \ | ||
19 | --blockdev file,node-name=drive0,filename=test.img \ | ||
20 | --export vhost-user-blk,node-name=drive0,id=export0,unix-socket=/tmp/vhost-user-blk.sock | ||
21 | |||
22 | Note that unix-socket is optional because we may wish to accept chardevs | ||
23 | too in the future. | ||
24 | |||
25 | Markus noted that supported address families are not explicit in the | ||
26 | QAPI schema. It is unlikely that support for more address families will | ||
27 | be added since file descriptor passing is required and few address | ||
28 | families support it. If a new address family needs to be added, then the | ||
29 | QAPI 'features' syntax can be used to advertize them. | ||
30 | |||
31 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
32 | Acked-by: Markus Armbruster <armbru@redhat.com> | ||
33 | Message-id: 20200924151549.913737-12-stefanha@redhat.com | ||
34 | [Skip test on big-endian host architectures because this device doesn't | ||
35 | support them yet (as already mentioned in a code comment). | ||
36 | --Stefan] | ||
37 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
11 | --- | 38 | --- |
12 | block/parallels.c | 33 ++++++++++++++------------------- | 39 | qapi/block-export.json | 21 +- |
13 | 1 file changed, 14 insertions(+), 19 deletions(-) | 40 | block/export/vhost-user-blk-server.h | 23 +- |
14 | 41 | block/export/export.c | 6 + | |
15 | diff --git a/block/parallels.c b/block/parallels.c | 42 | block/export/vhost-user-blk-server.c | 452 +++++++-------------------- |
43 | util/vhost-user-server.c | 10 +- | ||
44 | block/export/meson.build | 1 + | ||
45 | block/meson.build | 1 - | ||
46 | 7 files changed, 156 insertions(+), 358 deletions(-) | ||
47 | |||
48 | diff --git a/qapi/block-export.json b/qapi/block-export.json | ||
16 | index XXXXXXX..XXXXXXX 100644 | 49 | index XXXXXXX..XXXXXXX 100644 |
17 | --- a/block/parallels.c | 50 | --- a/qapi/block-export.json |
18 | +++ b/block/parallels.c | 51 | +++ b/qapi/block-export.json |
19 | @@ -XXX,XX +XXX,XX @@ parallels_co_check(BlockDriverState *bs, BdrvCheckResult *res, | 52 | @@ -XXX,XX +XXX,XX @@ |
20 | BDRVParallelsState *s = bs->opaque; | 53 | 'data': { '*name': 'str', '*description': 'str', |
21 | int ret; | 54 | '*bitmap': 'str' } } |
22 | 55 | ||
23 | - qemu_co_mutex_lock(&s->lock); | 56 | +## |
24 | + WITH_QEMU_LOCK_GUARD(&s->lock) { | 57 | +# @BlockExportOptionsVhostUserBlk: |
25 | + parallels_check_unclean(bs, res, fix); | 58 | +# |
26 | 59 | +# A vhost-user-blk block export. | |
27 | - parallels_check_unclean(bs, res, fix); | 60 | +# |
28 | + ret = parallels_check_outside_image(bs, res, fix); | 61 | +# @addr: The vhost-user socket on which to listen. Both 'unix' and 'fd' |
29 | + if (ret < 0) { | 62 | +# SocketAddress types are supported. Passed fds must be UNIX domain |
30 | + return ret; | 63 | +# sockets. |
64 | +# @logical-block-size: Logical block size in bytes. Defaults to 512 bytes. | ||
65 | +# | ||
66 | +# Since: 5.2 | ||
67 | +## | ||
68 | +{ 'struct': 'BlockExportOptionsVhostUserBlk', | ||
69 | + 'data': { 'addr': 'SocketAddress', '*logical-block-size': 'size' } } | ||
70 | + | ||
71 | ## | ||
72 | # @NbdServerAddOptions: | ||
73 | # | ||
74 | @@ -XXX,XX +XXX,XX @@ | ||
75 | # An enumeration of block export types | ||
76 | # | ||
77 | # @nbd: NBD export | ||
78 | +# @vhost-user-blk: vhost-user-blk export (since 5.2) | ||
79 | # | ||
80 | # Since: 4.2 | ||
81 | ## | ||
82 | { 'enum': 'BlockExportType', | ||
83 | - 'data': [ 'nbd' ] } | ||
84 | + 'data': [ 'nbd', 'vhost-user-blk' ] } | ||
85 | |||
86 | ## | ||
87 | # @BlockExportOptions: | ||
88 | @@ -XXX,XX +XXX,XX @@ | ||
89 | '*writethrough': 'bool' }, | ||
90 | 'discriminator': 'type', | ||
91 | 'data': { | ||
92 | - 'nbd': 'BlockExportOptionsNbd' | ||
93 | + 'nbd': 'BlockExportOptionsNbd', | ||
94 | + 'vhost-user-blk': 'BlockExportOptionsVhostUserBlk' | ||
95 | } } | ||
96 | |||
97 | ## | ||
98 | diff --git a/block/export/vhost-user-blk-server.h b/block/export/vhost-user-blk-server.h | ||
99 | index XXXXXXX..XXXXXXX 100644 | ||
100 | --- a/block/export/vhost-user-blk-server.h | ||
101 | +++ b/block/export/vhost-user-blk-server.h | ||
102 | @@ -XXX,XX +XXX,XX @@ | ||
103 | |||
104 | #ifndef VHOST_USER_BLK_SERVER_H | ||
105 | #define VHOST_USER_BLK_SERVER_H | ||
106 | -#include "util/vhost-user-server.h" | ||
107 | |||
108 | -typedef struct VuBlockDev VuBlockDev; | ||
109 | -#define TYPE_VHOST_USER_BLK_SERVER "vhost-user-blk-server" | ||
110 | -#define VHOST_USER_BLK_SERVER(obj) \ | ||
111 | - OBJECT_CHECK(VuBlockDev, obj, TYPE_VHOST_USER_BLK_SERVER) | ||
112 | +#include "block/export.h" | ||
113 | |||
114 | -/* vhost user block device */ | ||
115 | -struct VuBlockDev { | ||
116 | - Object parent_obj; | ||
117 | - char *node_name; | ||
118 | - SocketAddress *addr; | ||
119 | - AioContext *ctx; | ||
120 | - VuServer vu_server; | ||
121 | - bool running; | ||
122 | - uint32_t blk_size; | ||
123 | - BlockBackend *backend; | ||
124 | - QIOChannelSocket *sioc; | ||
125 | - QTAILQ_ENTRY(VuBlockDev) next; | ||
126 | - struct virtio_blk_config blkcfg; | ||
127 | - bool writable; | ||
128 | -}; | ||
129 | +/* For block/export/export.c */ | ||
130 | +extern const BlockExportDriver blk_exp_vhost_user_blk; | ||
131 | |||
132 | #endif /* VHOST_USER_BLK_SERVER_H */ | ||
133 | diff --git a/block/export/export.c b/block/export/export.c | ||
134 | index XXXXXXX..XXXXXXX 100644 | ||
135 | --- a/block/export/export.c | ||
136 | +++ b/block/export/export.c | ||
137 | @@ -XXX,XX +XXX,XX @@ | ||
138 | #include "sysemu/block-backend.h" | ||
139 | #include "block/export.h" | ||
140 | #include "block/nbd.h" | ||
141 | +#if CONFIG_LINUX | ||
142 | +#include "block/export/vhost-user-blk-server.h" | ||
143 | +#endif | ||
144 | #include "qapi/error.h" | ||
145 | #include "qapi/qapi-commands-block-export.h" | ||
146 | #include "qapi/qapi-events-block-export.h" | ||
147 | @@ -XXX,XX +XXX,XX @@ | ||
148 | |||
149 | static const BlockExportDriver *blk_exp_drivers[] = { | ||
150 | &blk_exp_nbd, | ||
151 | +#if CONFIG_LINUX | ||
152 | + &blk_exp_vhost_user_blk, | ||
153 | +#endif | ||
154 | }; | ||
155 | |||
156 | /* Only accessed from the main thread */ | ||
157 | diff --git a/block/export/vhost-user-blk-server.c b/block/export/vhost-user-blk-server.c | ||
158 | index XXXXXXX..XXXXXXX 100644 | ||
159 | --- a/block/export/vhost-user-blk-server.c | ||
160 | +++ b/block/export/vhost-user-blk-server.c | ||
161 | @@ -XXX,XX +XXX,XX @@ | ||
162 | */ | ||
163 | #include "qemu/osdep.h" | ||
164 | #include "block/block.h" | ||
165 | +#include "contrib/libvhost-user/libvhost-user.h" | ||
166 | +#include "standard-headers/linux/virtio_blk.h" | ||
167 | +#include "util/vhost-user-server.h" | ||
168 | #include "vhost-user-blk-server.h" | ||
169 | #include "qapi/error.h" | ||
170 | #include "qom/object_interfaces.h" | ||
171 | @@ -XXX,XX +XXX,XX @@ struct virtio_blk_inhdr { | ||
172 | unsigned char status; | ||
173 | }; | ||
174 | |||
175 | -typedef struct VuBlockReq { | ||
176 | +typedef struct VuBlkReq { | ||
177 | VuVirtqElement elem; | ||
178 | int64_t sector_num; | ||
179 | size_t size; | ||
180 | @@ -XXX,XX +XXX,XX @@ typedef struct VuBlockReq { | ||
181 | struct virtio_blk_outhdr out; | ||
182 | VuServer *server; | ||
183 | struct VuVirtq *vq; | ||
184 | -} VuBlockReq; | ||
185 | +} VuBlkReq; | ||
186 | |||
187 | -static void vu_block_req_complete(VuBlockReq *req) | ||
188 | +/* vhost user block device */ | ||
189 | +typedef struct { | ||
190 | + BlockExport export; | ||
191 | + VuServer vu_server; | ||
192 | + uint32_t blk_size; | ||
193 | + QIOChannelSocket *sioc; | ||
194 | + struct virtio_blk_config blkcfg; | ||
195 | + bool writable; | ||
196 | +} VuBlkExport; | ||
197 | + | ||
198 | +static void vu_blk_req_complete(VuBlkReq *req) | ||
199 | { | ||
200 | VuDev *vu_dev = &req->server->vu_dev; | ||
201 | |||
202 | @@ -XXX,XX +XXX,XX @@ static void vu_block_req_complete(VuBlockReq *req) | ||
203 | free(req); | ||
204 | } | ||
205 | |||
206 | -static VuBlockDev *get_vu_block_device_by_server(VuServer *server) | ||
207 | -{ | ||
208 | - return container_of(server, VuBlockDev, vu_server); | ||
209 | -} | ||
210 | - | ||
211 | static int coroutine_fn | ||
212 | -vu_block_discard_write_zeroes(VuBlockReq *req, struct iovec *iov, | ||
213 | - uint32_t iovcnt, uint32_t type) | ||
214 | +vu_blk_discard_write_zeroes(BlockBackend *blk, struct iovec *iov, | ||
215 | + uint32_t iovcnt, uint32_t type) | ||
216 | { | ||
217 | struct virtio_blk_discard_write_zeroes desc; | ||
218 | ssize_t size = iov_to_buf(iov, iovcnt, 0, &desc, sizeof(desc)); | ||
219 | @@ -XXX,XX +XXX,XX @@ vu_block_discard_write_zeroes(VuBlockReq *req, struct iovec *iov, | ||
220 | return -EINVAL; | ||
221 | } | ||
222 | |||
223 | - VuBlockDev *vdev_blk = get_vu_block_device_by_server(req->server); | ||
224 | uint64_t range[2] = { le64_to_cpu(desc.sector) << 9, | ||
225 | le32_to_cpu(desc.num_sectors) << 9 }; | ||
226 | if (type == VIRTIO_BLK_T_DISCARD) { | ||
227 | - if (blk_co_pdiscard(vdev_blk->backend, range[0], range[1]) == 0) { | ||
228 | + if (blk_co_pdiscard(blk, range[0], range[1]) == 0) { | ||
229 | return 0; | ||
230 | } | ||
231 | } else if (type == VIRTIO_BLK_T_WRITE_ZEROES) { | ||
232 | - if (blk_co_pwrite_zeroes(vdev_blk->backend, | ||
233 | - range[0], range[1], 0) == 0) { | ||
234 | + if (blk_co_pwrite_zeroes(blk, range[0], range[1], 0) == 0) { | ||
235 | return 0; | ||
236 | } | ||
237 | } | ||
238 | @@ -XXX,XX +XXX,XX @@ vu_block_discard_write_zeroes(VuBlockReq *req, struct iovec *iov, | ||
239 | return -EINVAL; | ||
240 | } | ||
241 | |||
242 | -static int coroutine_fn vu_block_flush(VuBlockReq *req) | ||
243 | +static void coroutine_fn vu_blk_virtio_process_req(void *opaque) | ||
244 | { | ||
245 | - VuBlockDev *vdev_blk = get_vu_block_device_by_server(req->server); | ||
246 | - BlockBackend *backend = vdev_blk->backend; | ||
247 | - return blk_co_flush(backend); | ||
248 | -} | ||
249 | - | ||
250 | -static void coroutine_fn vu_block_virtio_process_req(void *opaque) | ||
251 | -{ | ||
252 | - VuBlockReq *req = opaque; | ||
253 | + VuBlkReq *req = opaque; | ||
254 | VuServer *server = req->server; | ||
255 | VuVirtqElement *elem = &req->elem; | ||
256 | uint32_t type; | ||
257 | |||
258 | - VuBlockDev *vdev_blk = get_vu_block_device_by_server(server); | ||
259 | - BlockBackend *backend = vdev_blk->backend; | ||
260 | + VuBlkExport *vexp = container_of(server, VuBlkExport, vu_server); | ||
261 | + BlockBackend *blk = vexp->export.blk; | ||
262 | |||
263 | struct iovec *in_iov = elem->in_sg; | ||
264 | struct iovec *out_iov = elem->out_sg; | ||
265 | @@ -XXX,XX +XXX,XX @@ static void coroutine_fn vu_block_virtio_process_req(void *opaque) | ||
266 | bool is_write = type & VIRTIO_BLK_T_OUT; | ||
267 | req->sector_num = le64_to_cpu(req->out.sector); | ||
268 | |||
269 | - int64_t offset = req->sector_num * vdev_blk->blk_size; | ||
270 | + if (is_write && !vexp->writable) { | ||
271 | + req->in->status = VIRTIO_BLK_S_IOERR; | ||
272 | + break; | ||
31 | + } | 273 | + } |
32 | 274 | + | |
33 | - ret = parallels_check_outside_image(bs, res, fix); | 275 | + int64_t offset = req->sector_num * vexp->blk_size; |
276 | QEMUIOVector qiov; | ||
277 | if (is_write) { | ||
278 | qemu_iovec_init_external(&qiov, out_iov, out_num); | ||
279 | - ret = blk_co_pwritev(backend, offset, qiov.size, | ||
280 | - &qiov, 0); | ||
281 | + ret = blk_co_pwritev(blk, offset, qiov.size, &qiov, 0); | ||
282 | } else { | ||
283 | qemu_iovec_init_external(&qiov, in_iov, in_num); | ||
284 | - ret = blk_co_preadv(backend, offset, qiov.size, | ||
285 | - &qiov, 0); | ||
286 | + ret = blk_co_preadv(blk, offset, qiov.size, &qiov, 0); | ||
287 | } | ||
288 | if (ret >= 0) { | ||
289 | req->in->status = VIRTIO_BLK_S_OK; | ||
290 | @@ -XXX,XX +XXX,XX @@ static void coroutine_fn vu_block_virtio_process_req(void *opaque) | ||
291 | break; | ||
292 | } | ||
293 | case VIRTIO_BLK_T_FLUSH: | ||
294 | - if (vu_block_flush(req) == 0) { | ||
295 | + if (blk_co_flush(blk) == 0) { | ||
296 | req->in->status = VIRTIO_BLK_S_OK; | ||
297 | } else { | ||
298 | req->in->status = VIRTIO_BLK_S_IOERR; | ||
299 | @@ -XXX,XX +XXX,XX @@ static void coroutine_fn vu_block_virtio_process_req(void *opaque) | ||
300 | case VIRTIO_BLK_T_DISCARD: | ||
301 | case VIRTIO_BLK_T_WRITE_ZEROES: { | ||
302 | int rc; | ||
303 | - rc = vu_block_discard_write_zeroes(req, &elem->out_sg[1], | ||
304 | - out_num, type); | ||
305 | + | ||
306 | + if (!vexp->writable) { | ||
307 | + req->in->status = VIRTIO_BLK_S_IOERR; | ||
308 | + break; | ||
309 | + } | ||
310 | + | ||
311 | + rc = vu_blk_discard_write_zeroes(blk, &elem->out_sg[1], out_num, type); | ||
312 | if (rc == 0) { | ||
313 | req->in->status = VIRTIO_BLK_S_OK; | ||
314 | } else { | ||
315 | @@ -XXX,XX +XXX,XX @@ static void coroutine_fn vu_block_virtio_process_req(void *opaque) | ||
316 | break; | ||
317 | } | ||
318 | |||
319 | - vu_block_req_complete(req); | ||
320 | + vu_blk_req_complete(req); | ||
321 | return; | ||
322 | |||
323 | err: | ||
324 | - free(elem); | ||
325 | + free(req); | ||
326 | } | ||
327 | |||
328 | -static void vu_block_process_vq(VuDev *vu_dev, int idx) | ||
329 | +static void vu_blk_process_vq(VuDev *vu_dev, int idx) | ||
330 | { | ||
331 | VuServer *server = container_of(vu_dev, VuServer, vu_dev); | ||
332 | VuVirtq *vq = vu_get_queue(vu_dev, idx); | ||
333 | |||
334 | while (1) { | ||
335 | - VuBlockReq *req; | ||
336 | + VuBlkReq *req; | ||
337 | |||
338 | - req = vu_queue_pop(vu_dev, vq, sizeof(VuBlockReq)); | ||
339 | + req = vu_queue_pop(vu_dev, vq, sizeof(VuBlkReq)); | ||
340 | if (!req) { | ||
341 | break; | ||
342 | } | ||
343 | @@ -XXX,XX +XXX,XX @@ static void vu_block_process_vq(VuDev *vu_dev, int idx) | ||
344 | req->vq = vq; | ||
345 | |||
346 | Coroutine *co = | ||
347 | - qemu_coroutine_create(vu_block_virtio_process_req, req); | ||
348 | + qemu_coroutine_create(vu_blk_virtio_process_req, req); | ||
349 | qemu_coroutine_enter(co); | ||
350 | } | ||
351 | } | ||
352 | |||
353 | -static void vu_block_queue_set_started(VuDev *vu_dev, int idx, bool started) | ||
354 | +static void vu_blk_queue_set_started(VuDev *vu_dev, int idx, bool started) | ||
355 | { | ||
356 | VuVirtq *vq; | ||
357 | |||
358 | assert(vu_dev); | ||
359 | |||
360 | vq = vu_get_queue(vu_dev, idx); | ||
361 | - vu_set_queue_handler(vu_dev, vq, started ? vu_block_process_vq : NULL); | ||
362 | + vu_set_queue_handler(vu_dev, vq, started ? vu_blk_process_vq : NULL); | ||
363 | } | ||
364 | |||
365 | -static uint64_t vu_block_get_features(VuDev *dev) | ||
366 | +static uint64_t vu_blk_get_features(VuDev *dev) | ||
367 | { | ||
368 | uint64_t features; | ||
369 | VuServer *server = container_of(dev, VuServer, vu_dev); | ||
370 | - VuBlockDev *vdev_blk = get_vu_block_device_by_server(server); | ||
371 | + VuBlkExport *vexp = container_of(server, VuBlkExport, vu_server); | ||
372 | features = 1ull << VIRTIO_BLK_F_SIZE_MAX | | ||
373 | 1ull << VIRTIO_BLK_F_SEG_MAX | | ||
374 | 1ull << VIRTIO_BLK_F_TOPOLOGY | | ||
375 | @@ -XXX,XX +XXX,XX @@ static uint64_t vu_block_get_features(VuDev *dev) | ||
376 | 1ull << VIRTIO_RING_F_EVENT_IDX | | ||
377 | 1ull << VHOST_USER_F_PROTOCOL_FEATURES; | ||
378 | |||
379 | - if (!vdev_blk->writable) { | ||
380 | + if (!vexp->writable) { | ||
381 | features |= 1ull << VIRTIO_BLK_F_RO; | ||
382 | } | ||
383 | |||
384 | return features; | ||
385 | } | ||
386 | |||
387 | -static uint64_t vu_block_get_protocol_features(VuDev *dev) | ||
388 | +static uint64_t vu_blk_get_protocol_features(VuDev *dev) | ||
389 | { | ||
390 | return 1ull << VHOST_USER_PROTOCOL_F_CONFIG | | ||
391 | 1ull << VHOST_USER_PROTOCOL_F_INFLIGHT_SHMFD; | ||
392 | } | ||
393 | |||
394 | static int | ||
395 | -vu_block_get_config(VuDev *vu_dev, uint8_t *config, uint32_t len) | ||
396 | +vu_blk_get_config(VuDev *vu_dev, uint8_t *config, uint32_t len) | ||
397 | { | ||
398 | + /* TODO blkcfg must be little-endian for VIRTIO 1.0 */ | ||
399 | VuServer *server = container_of(vu_dev, VuServer, vu_dev); | ||
400 | - VuBlockDev *vdev_blk = get_vu_block_device_by_server(server); | ||
401 | - memcpy(config, &vdev_blk->blkcfg, len); | ||
402 | - | ||
403 | + VuBlkExport *vexp = container_of(server, VuBlkExport, vu_server); | ||
404 | + memcpy(config, &vexp->blkcfg, len); | ||
405 | return 0; | ||
406 | } | ||
407 | |||
408 | static int | ||
409 | -vu_block_set_config(VuDev *vu_dev, const uint8_t *data, | ||
410 | +vu_blk_set_config(VuDev *vu_dev, const uint8_t *data, | ||
411 | uint32_t offset, uint32_t size, uint32_t flags) | ||
412 | { | ||
413 | VuServer *server = container_of(vu_dev, VuServer, vu_dev); | ||
414 | - VuBlockDev *vdev_blk = get_vu_block_device_by_server(server); | ||
415 | + VuBlkExport *vexp = container_of(server, VuBlkExport, vu_server); | ||
416 | uint8_t wce; | ||
417 | |||
418 | /* don't support live migration */ | ||
419 | @@ -XXX,XX +XXX,XX @@ vu_block_set_config(VuDev *vu_dev, const uint8_t *data, | ||
420 | } | ||
421 | |||
422 | wce = *data; | ||
423 | - vdev_blk->blkcfg.wce = wce; | ||
424 | - blk_set_enable_write_cache(vdev_blk->backend, wce); | ||
425 | + vexp->blkcfg.wce = wce; | ||
426 | + blk_set_enable_write_cache(vexp->export.blk, wce); | ||
427 | return 0; | ||
428 | } | ||
429 | |||
430 | @@ -XXX,XX +XXX,XX @@ vu_block_set_config(VuDev *vu_dev, const uint8_t *data, | ||
431 | * of vu_process_message. | ||
432 | * | ||
433 | */ | ||
434 | -static int vu_block_process_msg(VuDev *dev, VhostUserMsg *vmsg, int *do_reply) | ||
435 | +static int vu_blk_process_msg(VuDev *dev, VhostUserMsg *vmsg, int *do_reply) | ||
436 | { | ||
437 | if (vmsg->request == VHOST_USER_NONE) { | ||
438 | dev->panic(dev, "disconnect"); | ||
439 | @@ -XXX,XX +XXX,XX @@ static int vu_block_process_msg(VuDev *dev, VhostUserMsg *vmsg, int *do_reply) | ||
440 | return false; | ||
441 | } | ||
442 | |||
443 | -static const VuDevIface vu_block_iface = { | ||
444 | - .get_features = vu_block_get_features, | ||
445 | - .queue_set_started = vu_block_queue_set_started, | ||
446 | - .get_protocol_features = vu_block_get_protocol_features, | ||
447 | - .get_config = vu_block_get_config, | ||
448 | - .set_config = vu_block_set_config, | ||
449 | - .process_msg = vu_block_process_msg, | ||
450 | +static const VuDevIface vu_blk_iface = { | ||
451 | + .get_features = vu_blk_get_features, | ||
452 | + .queue_set_started = vu_blk_queue_set_started, | ||
453 | + .get_protocol_features = vu_blk_get_protocol_features, | ||
454 | + .get_config = vu_blk_get_config, | ||
455 | + .set_config = vu_blk_set_config, | ||
456 | + .process_msg = vu_blk_process_msg, | ||
457 | }; | ||
458 | |||
459 | static void blk_aio_attached(AioContext *ctx, void *opaque) | ||
460 | { | ||
461 | - VuBlockDev *vub_dev = opaque; | ||
462 | - vhost_user_server_attach_aio_context(&vub_dev->vu_server, ctx); | ||
463 | + VuBlkExport *vexp = opaque; | ||
464 | + vhost_user_server_attach_aio_context(&vexp->vu_server, ctx); | ||
465 | } | ||
466 | |||
467 | static void blk_aio_detach(void *opaque) | ||
468 | { | ||
469 | - VuBlockDev *vub_dev = opaque; | ||
470 | - vhost_user_server_detach_aio_context(&vub_dev->vu_server); | ||
471 | + VuBlkExport *vexp = opaque; | ||
472 | + vhost_user_server_detach_aio_context(&vexp->vu_server); | ||
473 | } | ||
474 | |||
475 | static void | ||
476 | -vu_block_initialize_config(BlockDriverState *bs, | ||
477 | +vu_blk_initialize_config(BlockDriverState *bs, | ||
478 | struct virtio_blk_config *config, uint32_t blk_size) | ||
479 | { | ||
480 | config->capacity = bdrv_getlength(bs) >> BDRV_SECTOR_BITS; | ||
481 | @@ -XXX,XX +XXX,XX @@ vu_block_initialize_config(BlockDriverState *bs, | ||
482 | config->max_write_zeroes_seg = 1; | ||
483 | } | ||
484 | |||
485 | -static VuBlockDev *vu_block_init(VuBlockDev *vu_block_device, Error **errp) | ||
486 | +static void vu_blk_exp_request_shutdown(BlockExport *exp) | ||
487 | { | ||
488 | + VuBlkExport *vexp = container_of(exp, VuBlkExport, export); | ||
489 | |||
490 | - BlockBackend *blk; | ||
491 | - Error *local_error = NULL; | ||
492 | - const char *node_name = vu_block_device->node_name; | ||
493 | - bool writable = vu_block_device->writable; | ||
494 | - uint64_t perm = BLK_PERM_CONSISTENT_READ; | ||
495 | - int ret; | ||
496 | - | ||
497 | - AioContext *ctx; | ||
498 | - | ||
499 | - BlockDriverState *bs = bdrv_lookup_bs(node_name, node_name, &local_error); | ||
500 | - | ||
501 | - if (!bs) { | ||
502 | - error_propagate(errp, local_error); | ||
503 | - return NULL; | ||
504 | - } | ||
505 | - | ||
506 | - if (bdrv_is_read_only(bs)) { | ||
507 | - writable = false; | ||
508 | - } | ||
509 | - | ||
510 | - if (writable) { | ||
511 | - perm |= BLK_PERM_WRITE; | ||
512 | - } | ||
513 | - | ||
514 | - ctx = bdrv_get_aio_context(bs); | ||
515 | - aio_context_acquire(ctx); | ||
516 | - bdrv_invalidate_cache(bs, NULL); | ||
517 | - aio_context_release(ctx); | ||
518 | - | ||
519 | - /* | ||
520 | - * Don't allow resize while the vhost user server is running, | ||
521 | - * otherwise we don't care what happens with the node. | ||
522 | - */ | ||
523 | - blk = blk_new(bdrv_get_aio_context(bs), perm, | ||
524 | - BLK_PERM_CONSISTENT_READ | BLK_PERM_WRITE_UNCHANGED | | ||
525 | - BLK_PERM_WRITE | BLK_PERM_GRAPH_MOD); | ||
526 | - ret = blk_insert_bs(blk, bs, errp); | ||
527 | - | ||
34 | - if (ret < 0) { | 528 | - if (ret < 0) { |
529 | - goto fail; | ||
530 | - } | ||
531 | - | ||
532 | - blk_set_enable_write_cache(blk, false); | ||
533 | - | ||
534 | - blk_set_allow_aio_context_change(blk, true); | ||
535 | - | ||
536 | - vu_block_device->blkcfg.wce = 0; | ||
537 | - vu_block_device->backend = blk; | ||
538 | - if (!vu_block_device->blk_size) { | ||
539 | - vu_block_device->blk_size = BDRV_SECTOR_SIZE; | ||
540 | - } | ||
541 | - vu_block_device->blkcfg.blk_size = vu_block_device->blk_size; | ||
542 | - blk_set_guest_block_size(blk, vu_block_device->blk_size); | ||
543 | - vu_block_initialize_config(bs, &vu_block_device->blkcfg, | ||
544 | - vu_block_device->blk_size); | ||
545 | - return vu_block_device; | ||
546 | - | ||
547 | -fail: | ||
548 | - blk_unref(blk); | ||
549 | - return NULL; | ||
550 | -} | ||
551 | - | ||
552 | -static void vu_block_deinit(VuBlockDev *vu_block_device) | ||
553 | -{ | ||
554 | - if (vu_block_device->backend) { | ||
555 | - blk_remove_aio_context_notifier(vu_block_device->backend, blk_aio_attached, | ||
556 | - blk_aio_detach, vu_block_device); | ||
557 | - } | ||
558 | - | ||
559 | - blk_unref(vu_block_device->backend); | ||
560 | -} | ||
561 | - | ||
562 | -static void vhost_user_blk_server_stop(VuBlockDev *vu_block_device) | ||
563 | -{ | ||
564 | - vhost_user_server_stop(&vu_block_device->vu_server); | ||
565 | - vu_block_deinit(vu_block_device); | ||
566 | -} | ||
567 | - | ||
568 | -static void vhost_user_blk_server_start(VuBlockDev *vu_block_device, | ||
569 | - Error **errp) | ||
570 | -{ | ||
571 | - AioContext *ctx; | ||
572 | - SocketAddress *addr = vu_block_device->addr; | ||
573 | - | ||
574 | - if (!vu_block_init(vu_block_device, errp)) { | ||
575 | - return; | ||
576 | - } | ||
577 | - | ||
578 | - ctx = bdrv_get_aio_context(blk_bs(vu_block_device->backend)); | ||
579 | - | ||
580 | - if (!vhost_user_server_start(&vu_block_device->vu_server, addr, ctx, | ||
581 | - VHOST_USER_BLK_MAX_QUEUES, &vu_block_iface, | ||
582 | - errp)) { | ||
583 | - goto error; | ||
584 | - } | ||
585 | - | ||
586 | - blk_add_aio_context_notifier(vu_block_device->backend, blk_aio_attached, | ||
587 | - blk_aio_detach, vu_block_device); | ||
588 | - vu_block_device->running = true; | ||
589 | - return; | ||
590 | - | ||
591 | - error: | ||
592 | - vu_block_deinit(vu_block_device); | ||
593 | -} | ||
594 | - | ||
595 | -static bool vu_prop_modifiable(VuBlockDev *vus, Error **errp) | ||
596 | -{ | ||
597 | - if (vus->running) { | ||
598 | - error_setg(errp, "The property can't be modified " | ||
599 | - "while the server is running"); | ||
600 | - return false; | ||
601 | - } | ||
602 | - return true; | ||
603 | -} | ||
604 | - | ||
605 | -static void vu_set_node_name(Object *obj, const char *value, Error **errp) | ||
606 | -{ | ||
607 | - VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj); | ||
608 | - | ||
609 | - if (!vu_prop_modifiable(vus, errp)) { | ||
610 | - return; | ||
611 | - } | ||
612 | - | ||
613 | - if (vus->node_name) { | ||
614 | - g_free(vus->node_name); | ||
615 | - } | ||
616 | - | ||
617 | - vus->node_name = g_strdup(value); | ||
618 | -} | ||
619 | - | ||
620 | -static char *vu_get_node_name(Object *obj, Error **errp) | ||
621 | -{ | ||
622 | - VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj); | ||
623 | - return g_strdup(vus->node_name); | ||
624 | -} | ||
625 | - | ||
626 | -static void free_socket_addr(SocketAddress *addr) | ||
627 | -{ | ||
628 | - g_free(addr->u.q_unix.path); | ||
629 | - g_free(addr); | ||
630 | -} | ||
631 | - | ||
632 | -static void vu_set_unix_socket(Object *obj, const char *value, | ||
633 | - Error **errp) | ||
634 | -{ | ||
635 | - VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj); | ||
636 | - | ||
637 | - if (!vu_prop_modifiable(vus, errp)) { | ||
638 | - return; | ||
639 | - } | ||
640 | - | ||
641 | - if (vus->addr) { | ||
642 | - free_socket_addr(vus->addr); | ||
643 | - } | ||
644 | - | ||
645 | - SocketAddress *addr = g_new0(SocketAddress, 1); | ||
646 | - addr->type = SOCKET_ADDRESS_TYPE_UNIX; | ||
647 | - addr->u.q_unix.path = g_strdup(value); | ||
648 | - vus->addr = addr; | ||
649 | + vhost_user_server_stop(&vexp->vu_server); | ||
650 | } | ||
651 | |||
652 | -static char *vu_get_unix_socket(Object *obj, Error **errp) | ||
653 | +static int vu_blk_exp_create(BlockExport *exp, BlockExportOptions *opts, | ||
654 | + Error **errp) | ||
655 | { | ||
656 | - VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj); | ||
657 | - return g_strdup(vus->addr->u.q_unix.path); | ||
658 | -} | ||
659 | - | ||
660 | -static bool vu_get_block_writable(Object *obj, Error **errp) | ||
661 | -{ | ||
662 | - VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj); | ||
663 | - return vus->writable; | ||
664 | -} | ||
665 | - | ||
666 | -static void vu_set_block_writable(Object *obj, bool value, Error **errp) | ||
667 | -{ | ||
668 | - VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj); | ||
669 | - | ||
670 | - if (!vu_prop_modifiable(vus, errp)) { | ||
671 | - return; | ||
672 | - } | ||
673 | - | ||
674 | - vus->writable = value; | ||
675 | -} | ||
676 | - | ||
677 | -static void vu_get_blk_size(Object *obj, Visitor *v, const char *name, | ||
678 | - void *opaque, Error **errp) | ||
679 | -{ | ||
680 | - VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj); | ||
681 | - uint32_t value = vus->blk_size; | ||
682 | - | ||
683 | - visit_type_uint32(v, name, &value, errp); | ||
684 | -} | ||
685 | - | ||
686 | -static void vu_set_blk_size(Object *obj, Visitor *v, const char *name, | ||
687 | - void *opaque, Error **errp) | ||
688 | -{ | ||
689 | - VuBlockDev *vus = VHOST_USER_BLK_SERVER(obj); | ||
690 | - | ||
691 | + VuBlkExport *vexp = container_of(exp, VuBlkExport, export); | ||
692 | + BlockExportOptionsVhostUserBlk *vu_opts = &opts->u.vhost_user_blk; | ||
693 | Error *local_err = NULL; | ||
694 | - uint32_t value; | ||
695 | + uint64_t logical_block_size; | ||
696 | |||
697 | - if (!vu_prop_modifiable(vus, errp)) { | ||
698 | - return; | ||
699 | - } | ||
700 | + vexp->writable = opts->writable; | ||
701 | + vexp->blkcfg.wce = 0; | ||
702 | |||
703 | - visit_type_uint32(v, name, &value, &local_err); | ||
704 | - if (local_err) { | ||
35 | - goto out; | 705 | - goto out; |
36 | - } | 706 | + if (vu_opts->has_logical_block_size) { |
37 | + ret = parallels_check_leak(bs, res, fix); | 707 | + logical_block_size = vu_opts->logical_block_size; |
38 | + if (ret < 0) { | 708 | + } else { |
39 | + return ret; | 709 | + logical_block_size = BDRV_SECTOR_SIZE; |
40 | + } | 710 | } |
41 | 711 | - | |
42 | - ret = parallels_check_leak(bs, res, fix); | 712 | - check_block_size(object_get_typename(obj), name, value, &local_err); |
43 | - if (ret < 0) { | 713 | + check_block_size(exp->id, "logical-block-size", logical_block_size, |
714 | + &local_err); | ||
715 | if (local_err) { | ||
44 | - goto out; | 716 | - goto out; |
45 | + parallels_collect_statistics(bs, res, fix); | 717 | + error_propagate(errp, local_err); |
718 | + return -EINVAL; | ||
719 | + } | ||
720 | + vexp->blk_size = logical_block_size; | ||
721 | + blk_set_guest_block_size(exp->blk, logical_block_size); | ||
722 | + vu_blk_initialize_config(blk_bs(exp->blk), &vexp->blkcfg, | ||
723 | + logical_block_size); | ||
724 | + | ||
725 | + blk_set_allow_aio_context_change(exp->blk, true); | ||
726 | + blk_add_aio_context_notifier(exp->blk, blk_aio_attached, blk_aio_detach, | ||
727 | + vexp); | ||
728 | + | ||
729 | + if (!vhost_user_server_start(&vexp->vu_server, vu_opts->addr, exp->ctx, | ||
730 | + VHOST_USER_BLK_MAX_QUEUES, &vu_blk_iface, | ||
731 | + errp)) { | ||
732 | + blk_remove_aio_context_notifier(exp->blk, blk_aio_attached, | ||
733 | + blk_aio_detach, vexp); | ||
734 | + return -EADDRNOTAVAIL; | ||
46 | } | 735 | } |
47 | 736 | ||
48 | - parallels_collect_statistics(bs, res, fix); | 737 | - vus->blk_size = value; |
49 | - | 738 | - |
50 | -out: | 739 | -out: |
51 | - qemu_co_mutex_unlock(&s->lock); | 740 | - error_propagate(errp, local_err); |
52 | - | 741 | -} |
53 | - if (ret == 0) { | 742 | - |
54 | - ret = bdrv_co_flush(bs); | 743 | -static void vhost_user_blk_server_instance_finalize(Object *obj) |
55 | - if (ret < 0) { | 744 | -{ |
56 | - res->check_errors++; | 745 | - VuBlockDev *vub = VHOST_USER_BLK_SERVER(obj); |
57 | - } | 746 | - |
58 | + ret = bdrv_co_flush(bs); | 747 | - vhost_user_blk_server_stop(vub); |
59 | + if (ret < 0) { | 748 | - |
60 | + res->check_errors++; | 749 | - /* |
61 | } | 750 | - * Unlike object_property_add_str, object_class_property_add_str |
62 | 751 | - * doesn't have a release method. Thus manual memory freeing is | |
63 | return ret; | 752 | - * needed. |
753 | - */ | ||
754 | - free_socket_addr(vub->addr); | ||
755 | - g_free(vub->node_name); | ||
756 | -} | ||
757 | - | ||
758 | -static void vhost_user_blk_server_complete(UserCreatable *obj, Error **errp) | ||
759 | -{ | ||
760 | - VuBlockDev *vub = VHOST_USER_BLK_SERVER(obj); | ||
761 | - | ||
762 | - vhost_user_blk_server_start(vub, errp); | ||
763 | + return 0; | ||
764 | } | ||
765 | |||
766 | -static void vhost_user_blk_server_class_init(ObjectClass *klass, | ||
767 | - void *class_data) | ||
768 | +static void vu_blk_exp_delete(BlockExport *exp) | ||
769 | { | ||
770 | - UserCreatableClass *ucc = USER_CREATABLE_CLASS(klass); | ||
771 | - ucc->complete = vhost_user_blk_server_complete; | ||
772 | - | ||
773 | - object_class_property_add_bool(klass, "writable", | ||
774 | - vu_get_block_writable, | ||
775 | - vu_set_block_writable); | ||
776 | - | ||
777 | - object_class_property_add_str(klass, "node-name", | ||
778 | - vu_get_node_name, | ||
779 | - vu_set_node_name); | ||
780 | - | ||
781 | - object_class_property_add_str(klass, "unix-socket", | ||
782 | - vu_get_unix_socket, | ||
783 | - vu_set_unix_socket); | ||
784 | + VuBlkExport *vexp = container_of(exp, VuBlkExport, export); | ||
785 | |||
786 | - object_class_property_add(klass, "logical-block-size", "uint32", | ||
787 | - vu_get_blk_size, vu_set_blk_size, | ||
788 | - NULL, NULL); | ||
789 | + blk_remove_aio_context_notifier(exp->blk, blk_aio_attached, blk_aio_detach, | ||
790 | + vexp); | ||
791 | } | ||
792 | |||
793 | -static const TypeInfo vhost_user_blk_server_info = { | ||
794 | - .name = TYPE_VHOST_USER_BLK_SERVER, | ||
795 | - .parent = TYPE_OBJECT, | ||
796 | - .instance_size = sizeof(VuBlockDev), | ||
797 | - .instance_finalize = vhost_user_blk_server_instance_finalize, | ||
798 | - .class_init = vhost_user_blk_server_class_init, | ||
799 | - .interfaces = (InterfaceInfo[]) { | ||
800 | - {TYPE_USER_CREATABLE}, | ||
801 | - {} | ||
802 | - }, | ||
803 | +const BlockExportDriver blk_exp_vhost_user_blk = { | ||
804 | + .type = BLOCK_EXPORT_TYPE_VHOST_USER_BLK, | ||
805 | + .instance_size = sizeof(VuBlkExport), | ||
806 | + .create = vu_blk_exp_create, | ||
807 | + .delete = vu_blk_exp_delete, | ||
808 | + .request_shutdown = vu_blk_exp_request_shutdown, | ||
809 | }; | ||
810 | - | ||
811 | -static void vhost_user_blk_server_register_types(void) | ||
812 | -{ | ||
813 | - type_register_static(&vhost_user_blk_server_info); | ||
814 | -} | ||
815 | - | ||
816 | -type_init(vhost_user_blk_server_register_types) | ||
817 | diff --git a/util/vhost-user-server.c b/util/vhost-user-server.c | ||
818 | index XXXXXXX..XXXXXXX 100644 | ||
819 | --- a/util/vhost-user-server.c | ||
820 | +++ b/util/vhost-user-server.c | ||
821 | @@ -XXX,XX +XXX,XX @@ bool vhost_user_server_start(VuServer *server, | ||
822 | Error **errp) | ||
823 | { | ||
824 | QEMUBH *bh; | ||
825 | - QIONetListener *listener = qio_net_listener_new(); | ||
826 | + QIONetListener *listener; | ||
827 | + | ||
828 | + if (socket_addr->type != SOCKET_ADDRESS_TYPE_UNIX && | ||
829 | + socket_addr->type != SOCKET_ADDRESS_TYPE_FD) { | ||
830 | + error_setg(errp, "Only socket address types 'unix' and 'fd' are supported"); | ||
831 | + return false; | ||
832 | + } | ||
833 | + | ||
834 | + listener = qio_net_listener_new(); | ||
835 | if (qio_net_listener_open_sync(listener, socket_addr, 1, | ||
836 | errp) < 0) { | ||
837 | object_unref(OBJECT(listener)); | ||
838 | diff --git a/block/export/meson.build b/block/export/meson.build | ||
839 | index XXXXXXX..XXXXXXX 100644 | ||
840 | --- a/block/export/meson.build | ||
841 | +++ b/block/export/meson.build | ||
842 | @@ -1 +1,2 @@ | ||
843 | block_ss.add(files('export.c')) | ||
844 | +block_ss.add(when: 'CONFIG_LINUX', if_true: files('vhost-user-blk-server.c', '../../contrib/libvhost-user/libvhost-user.c')) | ||
845 | diff --git a/block/meson.build b/block/meson.build | ||
846 | index XXXXXXX..XXXXXXX 100644 | ||
847 | --- a/block/meson.build | ||
848 | +++ b/block/meson.build | ||
849 | @@ -XXX,XX +XXX,XX @@ block_ss.add(when: 'CONFIG_WIN32', if_true: files('file-win32.c', 'win32-aio.c') | ||
850 | block_ss.add(when: 'CONFIG_POSIX', if_true: [files('file-posix.c'), coref, iokit]) | ||
851 | block_ss.add(when: 'CONFIG_LIBISCSI', if_true: files('iscsi-opts.c')) | ||
852 | block_ss.add(when: 'CONFIG_LINUX', if_true: files('nvme.c')) | ||
853 | -block_ss.add(when: 'CONFIG_LINUX', if_true: files('export/vhost-user-blk-server.c', '../contrib/libvhost-user/libvhost-user.c')) | ||
854 | block_ss.add(when: 'CONFIG_REPLICATION', if_true: files('replication.c')) | ||
855 | block_ss.add(when: 'CONFIG_SHEEPDOG', if_true: files('sheepdog.c')) | ||
856 | block_ss.add(when: ['CONFIG_LINUX_AIO', libaio], if_true: files('linux-aio.c')) | ||
64 | -- | 857 | -- |
65 | 2.40.1 | 858 | 2.26.2 |
859 | diff view generated by jsdifflib |
New patch | |||
---|---|---|---|
1 | Headers used by other subsystems are located in include/. Also add the | ||
2 | vhost-user-server and vhost-user-blk-server headers to MAINTAINERS. | ||
1 | 3 | ||
4 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
5 | Message-id: 20200924151549.913737-13-stefanha@redhat.com | ||
6 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
7 | --- | ||
8 | MAINTAINERS | 4 +++- | ||
9 | {util => include/qemu}/vhost-user-server.h | 0 | ||
10 | block/export/vhost-user-blk-server.c | 2 +- | ||
11 | util/vhost-user-server.c | 2 +- | ||
12 | 4 files changed, 5 insertions(+), 3 deletions(-) | ||
13 | rename {util => include/qemu}/vhost-user-server.h (100%) | ||
14 | |||
15 | diff --git a/MAINTAINERS b/MAINTAINERS | ||
16 | index XXXXXXX..XXXXXXX 100644 | ||
17 | --- a/MAINTAINERS | ||
18 | +++ b/MAINTAINERS | ||
19 | @@ -XXX,XX +XXX,XX @@ Vhost-user block device backend server | ||
20 | M: Coiby Xu <Coiby.Xu@gmail.com> | ||
21 | S: Maintained | ||
22 | F: block/export/vhost-user-blk-server.c | ||
23 | -F: util/vhost-user-server.c | ||
24 | +F: block/export/vhost-user-blk-server.h | ||
25 | +F: include/qemu/vhost-user-server.h | ||
26 | F: tests/qtest/libqos/vhost-user-blk.c | ||
27 | +F: util/vhost-user-server.c | ||
28 | |||
29 | Replication | ||
30 | M: Wen Congyang <wencongyang2@huawei.com> | ||
31 | diff --git a/util/vhost-user-server.h b/include/qemu/vhost-user-server.h | ||
32 | similarity index 100% | ||
33 | rename from util/vhost-user-server.h | ||
34 | rename to include/qemu/vhost-user-server.h | ||
35 | diff --git a/block/export/vhost-user-blk-server.c b/block/export/vhost-user-blk-server.c | ||
36 | index XXXXXXX..XXXXXXX 100644 | ||
37 | --- a/block/export/vhost-user-blk-server.c | ||
38 | +++ b/block/export/vhost-user-blk-server.c | ||
39 | @@ -XXX,XX +XXX,XX @@ | ||
40 | #include "block/block.h" | ||
41 | #include "contrib/libvhost-user/libvhost-user.h" | ||
42 | #include "standard-headers/linux/virtio_blk.h" | ||
43 | -#include "util/vhost-user-server.h" | ||
44 | +#include "qemu/vhost-user-server.h" | ||
45 | #include "vhost-user-blk-server.h" | ||
46 | #include "qapi/error.h" | ||
47 | #include "qom/object_interfaces.h" | ||
48 | diff --git a/util/vhost-user-server.c b/util/vhost-user-server.c | ||
49 | index XXXXXXX..XXXXXXX 100644 | ||
50 | --- a/util/vhost-user-server.c | ||
51 | +++ b/util/vhost-user-server.c | ||
52 | @@ -XXX,XX +XXX,XX @@ | ||
53 | */ | ||
54 | #include "qemu/osdep.h" | ||
55 | #include "qemu/main-loop.h" | ||
56 | +#include "qemu/vhost-user-server.h" | ||
57 | #include "block/aio-wait.h" | ||
58 | -#include "vhost-user-server.h" | ||
59 | |||
60 | /* | ||
61 | * Theory of operation: | ||
62 | -- | ||
63 | 2.26.2 | ||
64 | diff view generated by jsdifflib |
New patch | |||
---|---|---|---|
1 | Don't compile contrib/libvhost-user/libvhost-user.c again. Instead build | ||
2 | the static library once and then reuse it throughout QEMU. | ||
1 | 3 | ||
4 | Also switch from CONFIG_LINUX to CONFIG_VHOST_USER, which is what the | ||
5 | vhost-user tools (vhost-user-gpu, etc) do. | ||
6 | |||
7 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
8 | Message-id: 20200924151549.913737-14-stefanha@redhat.com | ||
9 | [Added CONFIG_LINUX again because libvhost-user doesn't build on macOS. | ||
10 | --Stefan] | ||
11 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
12 | --- | ||
13 | block/export/export.c | 8 ++++---- | ||
14 | block/export/meson.build | 2 +- | ||
15 | contrib/libvhost-user/meson.build | 1 + | ||
16 | meson.build | 6 +++++- | ||
17 | util/meson.build | 4 +++- | ||
18 | 5 files changed, 14 insertions(+), 7 deletions(-) | ||
19 | |||
20 | diff --git a/block/export/export.c b/block/export/export.c | ||
21 | index XXXXXXX..XXXXXXX 100644 | ||
22 | --- a/block/export/export.c | ||
23 | +++ b/block/export/export.c | ||
24 | @@ -XXX,XX +XXX,XX @@ | ||
25 | #include "sysemu/block-backend.h" | ||
26 | #include "block/export.h" | ||
27 | #include "block/nbd.h" | ||
28 | -#if CONFIG_LINUX | ||
29 | -#include "block/export/vhost-user-blk-server.h" | ||
30 | -#endif | ||
31 | #include "qapi/error.h" | ||
32 | #include "qapi/qapi-commands-block-export.h" | ||
33 | #include "qapi/qapi-events-block-export.h" | ||
34 | #include "qemu/id.h" | ||
35 | +#ifdef CONFIG_VHOST_USER | ||
36 | +#include "vhost-user-blk-server.h" | ||
37 | +#endif | ||
38 | |||
39 | static const BlockExportDriver *blk_exp_drivers[] = { | ||
40 | &blk_exp_nbd, | ||
41 | -#if CONFIG_LINUX | ||
42 | +#ifdef CONFIG_VHOST_USER | ||
43 | &blk_exp_vhost_user_blk, | ||
44 | #endif | ||
45 | }; | ||
46 | diff --git a/block/export/meson.build b/block/export/meson.build | ||
47 | index XXXXXXX..XXXXXXX 100644 | ||
48 | --- a/block/export/meson.build | ||
49 | +++ b/block/export/meson.build | ||
50 | @@ -XXX,XX +XXX,XX @@ | ||
51 | block_ss.add(files('export.c')) | ||
52 | -block_ss.add(when: 'CONFIG_LINUX', if_true: files('vhost-user-blk-server.c', '../../contrib/libvhost-user/libvhost-user.c')) | ||
53 | +block_ss.add(when: ['CONFIG_LINUX', 'CONFIG_VHOST_USER'], if_true: files('vhost-user-blk-server.c')) | ||
54 | diff --git a/contrib/libvhost-user/meson.build b/contrib/libvhost-user/meson.build | ||
55 | index XXXXXXX..XXXXXXX 100644 | ||
56 | --- a/contrib/libvhost-user/meson.build | ||
57 | +++ b/contrib/libvhost-user/meson.build | ||
58 | @@ -XXX,XX +XXX,XX @@ | ||
59 | libvhost_user = static_library('vhost-user', | ||
60 | files('libvhost-user.c', 'libvhost-user-glib.c'), | ||
61 | build_by_default: false) | ||
62 | +vhost_user = declare_dependency(link_with: libvhost_user) | ||
63 | diff --git a/meson.build b/meson.build | ||
64 | index XXXXXXX..XXXXXXX 100644 | ||
65 | --- a/meson.build | ||
66 | +++ b/meson.build | ||
67 | @@ -XXX,XX +XXX,XX @@ trace_events_subdirs += [ | ||
68 | 'util', | ||
69 | ] | ||
70 | |||
71 | +vhost_user = not_found | ||
72 | +if 'CONFIG_VHOST_USER' in config_host | ||
73 | + subdir('contrib/libvhost-user') | ||
74 | +endif | ||
75 | + | ||
76 | subdir('qapi') | ||
77 | subdir('qobject') | ||
78 | subdir('stubs') | ||
79 | @@ -XXX,XX +XXX,XX @@ if have_tools | ||
80 | install: true) | ||
81 | |||
82 | if 'CONFIG_VHOST_USER' in config_host | ||
83 | - subdir('contrib/libvhost-user') | ||
84 | subdir('contrib/vhost-user-blk') | ||
85 | subdir('contrib/vhost-user-gpu') | ||
86 | subdir('contrib/vhost-user-input') | ||
87 | diff --git a/util/meson.build b/util/meson.build | ||
88 | index XXXXXXX..XXXXXXX 100644 | ||
89 | --- a/util/meson.build | ||
90 | +++ b/util/meson.build | ||
91 | @@ -XXX,XX +XXX,XX @@ if have_block | ||
92 | util_ss.add(files('main-loop.c')) | ||
93 | util_ss.add(files('nvdimm-utils.c')) | ||
94 | util_ss.add(files('qemu-coroutine.c', 'qemu-coroutine-lock.c', 'qemu-coroutine-io.c')) | ||
95 | - util_ss.add(when: 'CONFIG_LINUX', if_true: files('vhost-user-server.c')) | ||
96 | + util_ss.add(when: ['CONFIG_LINUX', 'CONFIG_VHOST_USER'], if_true: [ | ||
97 | + files('vhost-user-server.c'), vhost_user | ||
98 | + ]) | ||
99 | util_ss.add(files('block-helpers.c')) | ||
100 | util_ss.add(files('qemu-coroutine-sleep.c')) | ||
101 | util_ss.add(files('qemu-co-shared-resource.c')) | ||
102 | -- | ||
103 | 2.26.2 | ||
104 | diff view generated by jsdifflib |
New patch | |||
---|---|---|---|
1 | Introduce libblkdev.fa to avoid recompiling blockdev_ss twice. | ||
1 | 2 | ||
3 | Suggested-by: Paolo Bonzini <pbonzini@redhat.com> | ||
4 | Reviewed-by: Paolo Bonzini <pbonzini@redhat.com> | ||
5 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
6 | Message-id: 20200929125516.186715-3-stefanha@redhat.com | ||
7 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
8 | --- | ||
9 | meson.build | 12 ++++++++++-- | ||
10 | storage-daemon/meson.build | 3 +-- | ||
11 | 2 files changed, 11 insertions(+), 4 deletions(-) | ||
12 | |||
13 | diff --git a/meson.build b/meson.build | ||
14 | index XXXXXXX..XXXXXXX 100644 | ||
15 | --- a/meson.build | ||
16 | +++ b/meson.build | ||
17 | @@ -XXX,XX +XXX,XX @@ blockdev_ss.add(files( | ||
18 | # os-win32.c does not | ||
19 | blockdev_ss.add(when: 'CONFIG_POSIX', if_true: files('os-posix.c')) | ||
20 | softmmu_ss.add(when: 'CONFIG_WIN32', if_true: [files('os-win32.c')]) | ||
21 | -softmmu_ss.add_all(blockdev_ss) | ||
22 | |||
23 | common_ss.add(files('cpus-common.c')) | ||
24 | |||
25 | @@ -XXX,XX +XXX,XX @@ block = declare_dependency(link_whole: [libblock], | ||
26 | link_args: '@block.syms', | ||
27 | dependencies: [crypto, io]) | ||
28 | |||
29 | +blockdev_ss = blockdev_ss.apply(config_host, strict: false) | ||
30 | +libblockdev = static_library('blockdev', blockdev_ss.sources() + genh, | ||
31 | + dependencies: blockdev_ss.dependencies(), | ||
32 | + name_suffix: 'fa', | ||
33 | + build_by_default: false) | ||
34 | + | ||
35 | +blockdev = declare_dependency(link_whole: [libblockdev], | ||
36 | + dependencies: [block]) | ||
37 | + | ||
38 | qmp_ss = qmp_ss.apply(config_host, strict: false) | ||
39 | libqmp = static_library('qmp', qmp_ss.sources() + genh, | ||
40 | dependencies: qmp_ss.dependencies(), | ||
41 | @@ -XXX,XX +XXX,XX @@ foreach m : block_mods + softmmu_mods | ||
42 | install_dir: config_host['qemu_moddir']) | ||
43 | endforeach | ||
44 | |||
45 | -softmmu_ss.add(authz, block, chardev, crypto, io, qmp) | ||
46 | +softmmu_ss.add(authz, blockdev, chardev, crypto, io, qmp) | ||
47 | common_ss.add(qom, qemuutil) | ||
48 | |||
49 | common_ss.add_all(when: 'CONFIG_SOFTMMU', if_true: [softmmu_ss]) | ||
50 | diff --git a/storage-daemon/meson.build b/storage-daemon/meson.build | ||
51 | index XXXXXXX..XXXXXXX 100644 | ||
52 | --- a/storage-daemon/meson.build | ||
53 | +++ b/storage-daemon/meson.build | ||
54 | @@ -XXX,XX +XXX,XX @@ | ||
55 | qsd_ss = ss.source_set() | ||
56 | qsd_ss.add(files('qemu-storage-daemon.c')) | ||
57 | -qsd_ss.add(block, chardev, qmp, qom, qemuutil) | ||
58 | -qsd_ss.add_all(blockdev_ss) | ||
59 | +qsd_ss.add(blockdev, chardev, qmp, qom, qemuutil) | ||
60 | |||
61 | subdir('qapi') | ||
62 | |||
63 | -- | ||
64 | 2.26.2 | ||
65 | diff view generated by jsdifflib |
1 | Test that even vectored IO requests with 1024 vector elements that are | 1 | Block exports are used by softmmu, qemu-storage-daemon, and qemu-nbd. |
---|---|---|---|
2 | not aligned to the device's request alignment will succeed. | 2 | They are not used by other programs and are not otherwise needed in |
3 | libblock. | ||
3 | 4 | ||
5 | Undo the recent move of blockdev-nbd.c from blockdev_ss into block_ss. | ||
6 | Since bdrv_close_all() (libblock) calls blk_exp_close_all() | ||
7 | (libblockdev) a stub function is required.. | ||
8 | |||
9 | Make qemu-nbd.c use signal handling utility functions instead of | ||
10 | duplicating the code. This helps because os-posix.c is in libblockdev | ||
11 | and it depends on a qemu_system_killed() symbol that qemu-nbd.c lacks. | ||
12 | Once we use the signal handling utility functions we also end up | ||
13 | providing the necessary symbol. | ||
14 | |||
15 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
16 | Reviewed-by: Paolo Bonzini <pbonzini@redhat.com> | ||
4 | Reviewed-by: Eric Blake <eblake@redhat.com> | 17 | Reviewed-by: Eric Blake <eblake@redhat.com> |
5 | Reviewed-by: Vladimir Sementsov-Ogievskiy <vsementsov@yandex-team.ru> | 18 | Message-id: 20200929125516.186715-4-stefanha@redhat.com |
6 | Signed-off-by: Hanna Czenczek <hreitz@redhat.com> | 19 | [Fixed s/ndb/nbd/ typo in commit description as suggested by Eric Blake |
7 | Message-Id: <20230411173418.19549-5-hreitz@redhat.com> | 20 | --Stefan] |
21 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
8 | --- | 22 | --- |
9 | tests/qemu-iotests/tests/iov-padding | 85 ++++++++++++++++++++++++ | 23 | qemu-nbd.c | 21 ++++++++------------- |
10 | tests/qemu-iotests/tests/iov-padding.out | 59 ++++++++++++++++ | 24 | stubs/blk-exp-close-all.c | 7 +++++++ |
11 | 2 files changed, 144 insertions(+) | 25 | block/export/meson.build | 4 ++-- |
12 | create mode 100755 tests/qemu-iotests/tests/iov-padding | 26 | meson.build | 4 ++-- |
13 | create mode 100644 tests/qemu-iotests/tests/iov-padding.out | 27 | nbd/meson.build | 2 ++ |
28 | stubs/meson.build | 1 + | ||
29 | 6 files changed, 22 insertions(+), 17 deletions(-) | ||
30 | create mode 100644 stubs/blk-exp-close-all.c | ||
14 | 31 | ||
15 | diff --git a/tests/qemu-iotests/tests/iov-padding b/tests/qemu-iotests/tests/iov-padding | 32 | diff --git a/qemu-nbd.c b/qemu-nbd.c |
16 | new file mode 100755 | 33 | index XXXXXXX..XXXXXXX 100644 |
17 | index XXXXXXX..XXXXXXX | 34 | --- a/qemu-nbd.c |
18 | --- /dev/null | 35 | +++ b/qemu-nbd.c |
19 | +++ b/tests/qemu-iotests/tests/iov-padding | ||
20 | @@ -XXX,XX +XXX,XX @@ | 36 | @@ -XXX,XX +XXX,XX @@ |
21 | +#!/usr/bin/env bash | 37 | #include "qapi/error.h" |
22 | +# group: rw quick | 38 | #include "qemu/cutils.h" |
23 | +# | 39 | #include "sysemu/block-backend.h" |
24 | +# Check the interaction of request padding (to fit alignment restrictions) with | 40 | +#include "sysemu/runstate.h" /* for qemu_system_killed() prototype */ |
25 | +# vectored I/O from the guest | 41 | #include "block/block_int.h" |
26 | +# | 42 | #include "block/nbd.h" |
27 | +# Copyright Red Hat | 43 | #include "qemu/main-loop.h" |
28 | +# | 44 | @@ -XXX,XX +XXX,XX @@ QEMU_COPYRIGHT "\n" |
29 | +# This program is free software; you can redistribute it and/or modify | 45 | } |
30 | +# it under the terms of the GNU General Public License as published by | 46 | |
31 | +# the Free Software Foundation; either version 2 of the License, or | 47 | #ifdef CONFIG_POSIX |
32 | +# (at your option) any later version. | 48 | -static void termsig_handler(int signum) |
33 | +# | 49 | +/* |
34 | +# This program is distributed in the hope that it will be useful, | 50 | + * The client thread uses SIGTERM to interrupt the server. A signal |
35 | +# but WITHOUT ANY WARRANTY; without even the implied warranty of | 51 | + * handler ensures that "qemu-nbd -v -c" exits with a nice status code. |
36 | +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | 52 | + */ |
37 | +# GNU General Public License for more details. | 53 | +void qemu_system_killed(int signum, pid_t pid) |
38 | +# | 54 | { |
39 | +# You should have received a copy of the GNU General Public License | 55 | qatomic_cmpxchg(&state, RUNNING, TERMINATE); |
40 | +# along with this program. If not, see <http://www.gnu.org/licenses/>. | 56 | qemu_notify_event(); |
41 | +# | 57 | @@ -XXX,XX +XXX,XX @@ int main(int argc, char **argv) |
42 | + | 58 | BlockExportOptions *export_opts; |
43 | +seq=$(basename $0) | 59 | |
44 | +echo "QA output created by $seq" | 60 | #ifdef CONFIG_POSIX |
45 | + | 61 | - /* |
46 | +status=1 # failure is the default! | 62 | - * Exit gracefully on various signals, which includes SIGTERM used |
47 | + | 63 | - * by 'qemu-nbd -v -c'. |
48 | +_cleanup() | 64 | - */ |
49 | +{ | 65 | - struct sigaction sa_sigterm; |
50 | + _cleanup_test_img | 66 | - memset(&sa_sigterm, 0, sizeof(sa_sigterm)); |
51 | +} | 67 | - sa_sigterm.sa_handler = termsig_handler; |
52 | +trap "_cleanup; exit \$status" 0 1 2 3 15 | 68 | - sigaction(SIGTERM, &sa_sigterm, NULL); |
53 | + | 69 | - sigaction(SIGINT, &sa_sigterm, NULL); |
54 | +# get standard environment, filters and checks | 70 | - sigaction(SIGHUP, &sa_sigterm, NULL); |
55 | +cd .. | 71 | - |
56 | +. ./common.rc | 72 | - signal(SIGPIPE, SIG_IGN); |
57 | +. ./common.filter | 73 | + os_setup_early_signal_handling(); |
58 | + | 74 | + os_setup_signal_handling(); |
59 | +_supported_fmt raw | 75 | #endif |
60 | +_supported_proto file | 76 | |
61 | + | 77 | socket_init(); |
62 | +_make_test_img 1M | 78 | diff --git a/stubs/blk-exp-close-all.c b/stubs/blk-exp-close-all.c |
63 | + | ||
64 | +IMGSPEC="driver=blkdebug,align=4096,image.driver=file,image.filename=$TEST_IMG" | ||
65 | + | ||
66 | +# Four combinations: | ||
67 | +# - Offset 4096, length 1023 * 512 + 512: Fully aligned to 4k | ||
68 | +# - Offset 4096, length 1023 * 512 + 4096: Head is aligned, tail is not | ||
69 | +# - Offset 512, length 1023 * 512 + 512: Neither head nor tail are aligned | ||
70 | +# - Offset 512, length 1023 * 512 + 4096: Tail is aligned, head is not | ||
71 | +for start_offset in 4096 512; do | ||
72 | + for last_element_length in 512 4096; do | ||
73 | + length=$((1023 * 512 + $last_element_length)) | ||
74 | + | ||
75 | + echo | ||
76 | + echo "== performing 1024-element vectored requests to image (offset: $start_offset; length: $length) ==" | ||
77 | + | ||
78 | + # Fill with data for testing | ||
79 | + $QEMU_IO -c 'write -P 1 0 1M' "$TEST_IMG" | _filter_qemu_io | ||
80 | + | ||
81 | + # 1023 512-byte buffers, and then one with length $last_element_length | ||
82 | + cmd_params="-P 2 $start_offset $(yes 512 | head -n 1023 | tr '\n' ' ') $last_element_length" | ||
83 | + QEMU_IO_OPTIONS="$QEMU_IO_OPTIONS_NO_FMT" $QEMU_IO \ | ||
84 | + -c "writev $cmd_params" \ | ||
85 | + --image-opts \ | ||
86 | + "$IMGSPEC" \ | ||
87 | + | _filter_qemu_io | ||
88 | + | ||
89 | + # Read all patterns -- read the part we just wrote with writev twice, | ||
90 | + # once "normally", and once with a readv, so we see that that works, too | ||
91 | + QEMU_IO_OPTIONS="$QEMU_IO_OPTIONS_NO_FMT" $QEMU_IO \ | ||
92 | + -c "read -P 1 0 $start_offset" \ | ||
93 | + -c "read -P 2 $start_offset $length" \ | ||
94 | + -c "readv $cmd_params" \ | ||
95 | + -c "read -P 1 $((start_offset + length)) $((1024 * 1024 - length - start_offset))" \ | ||
96 | + --image-opts \ | ||
97 | + "$IMGSPEC" \ | ||
98 | + | _filter_qemu_io | ||
99 | + done | ||
100 | +done | ||
101 | + | ||
102 | +# success, all done | ||
103 | +echo "*** done" | ||
104 | +rm -f $seq.full | ||
105 | +status=0 | ||
106 | diff --git a/tests/qemu-iotests/tests/iov-padding.out b/tests/qemu-iotests/tests/iov-padding.out | ||
107 | new file mode 100644 | 79 | new file mode 100644 |
108 | index XXXXXXX..XXXXXXX | 80 | index XXXXXXX..XXXXXXX |
109 | --- /dev/null | 81 | --- /dev/null |
110 | +++ b/tests/qemu-iotests/tests/iov-padding.out | 82 | +++ b/stubs/blk-exp-close-all.c |
111 | @@ -XXX,XX +XXX,XX @@ | 83 | @@ -XXX,XX +XXX,XX @@ |
112 | +QA output created by iov-padding | 84 | +#include "qemu/osdep.h" |
113 | +Formatting 'TEST_DIR/t.IMGFMT', fmt=IMGFMT size=1048576 | 85 | +#include "block/export.h" |
114 | + | 86 | + |
115 | +== performing 1024-element vectored requests to image (offset: 4096; length: 524288) == | 87 | +/* Only used in programs that support block exports (libblockdev.fa) */ |
116 | +wrote 1048576/1048576 bytes at offset 0 | 88 | +void blk_exp_close_all(void) |
117 | +1 MiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 89 | +{ |
118 | +wrote 524288/524288 bytes at offset 4096 | 90 | +} |
119 | +512 KiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 91 | diff --git a/block/export/meson.build b/block/export/meson.build |
120 | +read 4096/4096 bytes at offset 0 | 92 | index XXXXXXX..XXXXXXX 100644 |
121 | +4 KiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 93 | --- a/block/export/meson.build |
122 | +read 524288/524288 bytes at offset 4096 | 94 | +++ b/block/export/meson.build |
123 | +512 KiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 95 | @@ -XXX,XX +XXX,XX @@ |
124 | +read 524288/524288 bytes at offset 4096 | 96 | -block_ss.add(files('export.c')) |
125 | +512 KiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 97 | -block_ss.add(when: ['CONFIG_LINUX', 'CONFIG_VHOST_USER'], if_true: files('vhost-user-blk-server.c')) |
126 | +read 520192/520192 bytes at offset 528384 | 98 | +blockdev_ss.add(files('export.c')) |
127 | +508 KiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 99 | +blockdev_ss.add(when: ['CONFIG_LINUX', 'CONFIG_VHOST_USER'], if_true: files('vhost-user-blk-server.c')) |
128 | + | 100 | diff --git a/meson.build b/meson.build |
129 | +== performing 1024-element vectored requests to image (offset: 4096; length: 527872) == | 101 | index XXXXXXX..XXXXXXX 100644 |
130 | +wrote 1048576/1048576 bytes at offset 0 | 102 | --- a/meson.build |
131 | +1 MiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 103 | +++ b/meson.build |
132 | +wrote 527872/527872 bytes at offset 4096 | 104 | @@ -XXX,XX +XXX,XX @@ subdir('dump') |
133 | +515.500 KiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 105 | |
134 | +read 4096/4096 bytes at offset 0 | 106 | block_ss.add(files( |
135 | +4 KiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 107 | 'block.c', |
136 | +read 527872/527872 bytes at offset 4096 | 108 | - 'blockdev-nbd.c', |
137 | +515.500 KiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 109 | 'blockjob.c', |
138 | +read 527872/527872 bytes at offset 4096 | 110 | 'job.c', |
139 | +515.500 KiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 111 | 'qemu-io-cmds.c', |
140 | +read 516608/516608 bytes at offset 531968 | 112 | @@ -XXX,XX +XXX,XX @@ subdir('block') |
141 | +504.500 KiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 113 | |
142 | + | 114 | blockdev_ss.add(files( |
143 | +== performing 1024-element vectored requests to image (offset: 512; length: 524288) == | 115 | 'blockdev.c', |
144 | +wrote 1048576/1048576 bytes at offset 0 | 116 | + 'blockdev-nbd.c', |
145 | +1 MiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 117 | 'iothread.c', |
146 | +wrote 524288/524288 bytes at offset 512 | 118 | 'job-qmp.c', |
147 | +512 KiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 119 | )) |
148 | +read 512/512 bytes at offset 0 | 120 | @@ -XXX,XX +XXX,XX @@ if have_tools |
149 | +512 bytes, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 121 | qemu_io = executable('qemu-io', files('qemu-io.c'), |
150 | +read 524288/524288 bytes at offset 512 | 122 | dependencies: [block, qemuutil], install: true) |
151 | +512 KiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 123 | qemu_nbd = executable('qemu-nbd', files('qemu-nbd.c'), |
152 | +read 524288/524288 bytes at offset 512 | 124 | - dependencies: [block, qemuutil], install: true) |
153 | +512 KiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 125 | + dependencies: [blockdev, qemuutil], install: true) |
154 | +read 523776/523776 bytes at offset 524800 | 126 | |
155 | +511.500 KiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 127 | subdir('storage-daemon') |
156 | + | 128 | subdir('contrib/rdmacm-mux') |
157 | +== performing 1024-element vectored requests to image (offset: 512; length: 527872) == | 129 | diff --git a/nbd/meson.build b/nbd/meson.build |
158 | +wrote 1048576/1048576 bytes at offset 0 | 130 | index XXXXXXX..XXXXXXX 100644 |
159 | +1 MiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 131 | --- a/nbd/meson.build |
160 | +wrote 527872/527872 bytes at offset 512 | 132 | +++ b/nbd/meson.build |
161 | +515.500 KiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 133 | @@ -XXX,XX +XXX,XX @@ |
162 | +read 512/512 bytes at offset 0 | 134 | block_ss.add(files( |
163 | +512 bytes, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 135 | 'client.c', |
164 | +read 527872/527872 bytes at offset 512 | 136 | 'common.c', |
165 | +515.500 KiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 137 | +)) |
166 | +read 527872/527872 bytes at offset 512 | 138 | +blockdev_ss.add(files( |
167 | +515.500 KiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 139 | 'server.c', |
168 | +read 520192/520192 bytes at offset 528384 | 140 | )) |
169 | +508 KiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | 141 | diff --git a/stubs/meson.build b/stubs/meson.build |
170 | +*** done | 142 | index XXXXXXX..XXXXXXX 100644 |
143 | --- a/stubs/meson.build | ||
144 | +++ b/stubs/meson.build | ||
145 | @@ -XXX,XX +XXX,XX @@ | ||
146 | stub_ss.add(files('arch_type.c')) | ||
147 | stub_ss.add(files('bdrv-next-monitor-owned.c')) | ||
148 | stub_ss.add(files('blk-commit-all.c')) | ||
149 | +stub_ss.add(files('blk-exp-close-all.c')) | ||
150 | stub_ss.add(files('blockdev-close-all-bdrv-states.c')) | ||
151 | stub_ss.add(files('change-state-handler.c')) | ||
152 | stub_ss.add(files('cmos.c')) | ||
171 | -- | 153 | -- |
172 | 2.40.1 | 154 | 2.26.2 |
155 | diff view generated by jsdifflib |
1 | From: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | 1 | Make it possible to specify the iothread where the export will run. By |
---|---|---|---|
2 | default the block node can be moved to other AioContexts later and the | ||
3 | export will follow. The fixed-iothread option forces strict behavior | ||
4 | that prevents changing AioContext while the export is active. See the | ||
5 | QAPI docs for details. | ||
2 | 6 | ||
3 | We will add more and more checks so we need a better code structure | 7 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
4 | in parallels_co_check. Let each check performs in a separate loop | 8 | Message-id: 20200929125516.186715-5-stefanha@redhat.com |
5 | in a separate helper. | 9 | [Fix stray '#' character in block-export.json and add missing "(since: |
10 | 5.2)" as suggested by Eric Blake. | ||
11 | --Stefan] | ||
12 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
13 | --- | ||
14 | qapi/block-export.json | 11 ++++++++++ | ||
15 | block/export/export.c | 31 +++++++++++++++++++++++++++- | ||
16 | block/export/vhost-user-blk-server.c | 5 ++++- | ||
17 | nbd/server.c | 2 -- | ||
18 | 4 files changed, 45 insertions(+), 4 deletions(-) | ||
6 | 19 | ||
7 | Signed-off-by: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | 20 | diff --git a/qapi/block-export.json b/qapi/block-export.json |
8 | Message-Id: <20230424093147.197643-10-alexander.ivanov@virtuozzo.com> | ||
9 | Reviewed-by: Hanna Czenczek <hreitz@redhat.com> | ||
10 | Signed-off-by: Hanna Czenczek <hreitz@redhat.com> | ||
11 | --- | ||
12 | block/parallels.c | 74 ++++++++++++++++++++++++++++------------------- | ||
13 | 1 file changed, 45 insertions(+), 29 deletions(-) | ||
14 | |||
15 | diff --git a/block/parallels.c b/block/parallels.c | ||
16 | index XXXXXXX..XXXXXXX 100644 | 21 | index XXXXXXX..XXXXXXX 100644 |
17 | --- a/block/parallels.c | 22 | --- a/qapi/block-export.json |
18 | +++ b/block/parallels.c | 23 | +++ b/qapi/block-export.json |
19 | @@ -XXX,XX +XXX,XX @@ parallels_check_outside_image(BlockDriverState *bs, BdrvCheckResult *res, | 24 | @@ -XXX,XX +XXX,XX @@ |
20 | } | 25 | # export before completion is signalled. (since: 5.2; |
21 | 26 | # default: false) | |
22 | static int coroutine_fn GRAPH_RDLOCK | 27 | # |
23 | -parallels_co_check(BlockDriverState *bs, BdrvCheckResult *res, | 28 | +# @iothread: The name of the iothread object where the export will run. The |
24 | - BdrvCheckMode fix) | 29 | +# default is to use the thread currently associated with the |
25 | +parallels_check_leak(BlockDriverState *bs, BdrvCheckResult *res, | 30 | +# block node. (since: 5.2) |
26 | + BdrvCheckMode fix) | 31 | +# |
32 | +# @fixed-iothread: True prevents the block node from being moved to another | ||
33 | +# thread while the export is active. If true and @iothread is | ||
34 | +# given, export creation fails if the block node cannot be | ||
35 | +# moved to the iothread. The default is false. (since: 5.2) | ||
36 | +# | ||
37 | # Since: 4.2 | ||
38 | ## | ||
39 | { 'union': 'BlockExportOptions', | ||
40 | 'base': { 'type': 'BlockExportType', | ||
41 | 'id': 'str', | ||
42 | + '*fixed-iothread': 'bool', | ||
43 | + '*iothread': 'str', | ||
44 | 'node-name': 'str', | ||
45 | '*writable': 'bool', | ||
46 | '*writethrough': 'bool' }, | ||
47 | diff --git a/block/export/export.c b/block/export/export.c | ||
48 | index XXXXXXX..XXXXXXX 100644 | ||
49 | --- a/block/export/export.c | ||
50 | +++ b/block/export/export.c | ||
51 | @@ -XXX,XX +XXX,XX @@ | ||
52 | |||
53 | #include "block/block.h" | ||
54 | #include "sysemu/block-backend.h" | ||
55 | +#include "sysemu/iothread.h" | ||
56 | #include "block/export.h" | ||
57 | #include "block/nbd.h" | ||
58 | #include "qapi/error.h" | ||
59 | @@ -XXX,XX +XXX,XX @@ static const BlockExportDriver *blk_exp_find_driver(BlockExportType type) | ||
60 | |||
61 | BlockExport *blk_exp_add(BlockExportOptions *export, Error **errp) | ||
27 | { | 62 | { |
28 | BDRVParallelsState *s = bs->opaque; | 63 | + bool fixed_iothread = export->has_fixed_iothread && export->fixed_iothread; |
29 | - int64_t size, prev_off; | 64 | const BlockExportDriver *drv; |
30 | + int64_t size; | 65 | BlockExport *exp = NULL; |
66 | BlockDriverState *bs; | ||
67 | - BlockBackend *blk; | ||
68 | + BlockBackend *blk = NULL; | ||
69 | AioContext *ctx; | ||
70 | uint64_t perm; | ||
31 | int ret; | 71 | int ret; |
32 | - uint32_t i; | 72 | @@ -XXX,XX +XXX,XX @@ BlockExport *blk_exp_add(BlockExportOptions *export, Error **errp) |
33 | 73 | ctx = bdrv_get_aio_context(bs); | |
34 | size = bdrv_getlength(bs->file->bs); | 74 | aio_context_acquire(ctx); |
35 | if (size < 0) { | 75 | |
36 | @@ -XXX,XX +XXX,XX @@ parallels_co_check(BlockDriverState *bs, BdrvCheckResult *res, | 76 | + if (export->has_iothread) { |
37 | return size; | 77 | + IOThread *iothread; |
38 | } | 78 | + AioContext *new_ctx; |
39 | |||
40 | + if (size > res->image_end_offset) { | ||
41 | + int64_t count; | ||
42 | + count = DIV_ROUND_UP(size - res->image_end_offset, s->cluster_size); | ||
43 | + fprintf(stderr, "%s space leaked at the end of the image %" PRId64 "\n", | ||
44 | + fix & BDRV_FIX_LEAKS ? "Repairing" : "ERROR", | ||
45 | + size - res->image_end_offset); | ||
46 | + res->leaks += count; | ||
47 | + if (fix & BDRV_FIX_LEAKS) { | ||
48 | + Error *local_err = NULL; | ||
49 | + | 79 | + |
50 | + /* | 80 | + iothread = iothread_by_id(export->iothread); |
51 | + * In order to really repair the image, we must shrink it. | 81 | + if (!iothread) { |
52 | + * That means we have to pass exact=true. | 82 | + error_setg(errp, "iothread \"%s\" not found", export->iothread); |
53 | + */ | 83 | + goto fail; |
54 | + ret = bdrv_co_truncate(bs->file, res->image_end_offset, true, | 84 | + } |
55 | + PREALLOC_MODE_OFF, 0, &local_err); | 85 | + |
56 | + if (ret < 0) { | 86 | + new_ctx = iothread_get_aio_context(iothread); |
57 | + error_report_err(local_err); | 87 | + |
58 | + res->check_errors++; | 88 | + ret = bdrv_try_set_aio_context(bs, new_ctx, errp); |
59 | + return ret; | 89 | + if (ret == 0) { |
60 | + } | 90 | + aio_context_release(ctx); |
61 | + res->leaks_fixed += count; | 91 | + aio_context_acquire(new_ctx); |
92 | + ctx = new_ctx; | ||
93 | + } else if (fixed_iothread) { | ||
94 | + goto fail; | ||
62 | + } | 95 | + } |
63 | + } | 96 | + } |
64 | + | 97 | + |
65 | + return 0; | 98 | /* |
66 | +} | 99 | * Block exports are used for non-shared storage migration. Make sure |
100 | * that BDRV_O_INACTIVE is cleared and the image is ready for write | ||
101 | @@ -XXX,XX +XXX,XX @@ BlockExport *blk_exp_add(BlockExportOptions *export, Error **errp) | ||
102 | } | ||
103 | |||
104 | blk = blk_new(ctx, perm, BLK_PERM_ALL); | ||
67 | + | 105 | + |
68 | +static int coroutine_fn GRAPH_RDLOCK | 106 | + if (!fixed_iothread) { |
69 | +parallels_co_check(BlockDriverState *bs, BdrvCheckResult *res, | 107 | + blk_set_allow_aio_context_change(blk, true); |
70 | + BdrvCheckMode fix) | ||
71 | +{ | ||
72 | + BDRVParallelsState *s = bs->opaque; | ||
73 | + int64_t prev_off; | ||
74 | + int ret; | ||
75 | + uint32_t i; | ||
76 | + | ||
77 | qemu_co_mutex_lock(&s->lock); | ||
78 | |||
79 | parallels_check_unclean(bs, res, fix); | ||
80 | @@ -XXX,XX +XXX,XX @@ parallels_co_check(BlockDriverState *bs, BdrvCheckResult *res, | ||
81 | goto out; | ||
82 | } | ||
83 | |||
84 | + ret = parallels_check_leak(bs, res, fix); | ||
85 | + if (ret < 0) { | ||
86 | + goto out; | ||
87 | + } | 108 | + } |
88 | + | 109 | + |
89 | res->bfi.total_clusters = s->bat_size; | 110 | ret = blk_insert_bs(blk, bs, errp); |
90 | res->bfi.compressed_clusters = 0; /* compression is not supported */ | 111 | if (ret < 0) { |
91 | 112 | goto fail; | |
92 | @@ -XXX,XX +XXX,XX @@ parallels_co_check(BlockDriverState *bs, BdrvCheckResult *res, | 113 | diff --git a/block/export/vhost-user-blk-server.c b/block/export/vhost-user-blk-server.c |
93 | prev_off = off; | 114 | index XXXXXXX..XXXXXXX 100644 |
115 | --- a/block/export/vhost-user-blk-server.c | ||
116 | +++ b/block/export/vhost-user-blk-server.c | ||
117 | @@ -XXX,XX +XXX,XX @@ static const VuDevIface vu_blk_iface = { | ||
118 | static void blk_aio_attached(AioContext *ctx, void *opaque) | ||
119 | { | ||
120 | VuBlkExport *vexp = opaque; | ||
121 | + | ||
122 | + vexp->export.ctx = ctx; | ||
123 | vhost_user_server_attach_aio_context(&vexp->vu_server, ctx); | ||
124 | } | ||
125 | |||
126 | static void blk_aio_detach(void *opaque) | ||
127 | { | ||
128 | VuBlkExport *vexp = opaque; | ||
129 | + | ||
130 | vhost_user_server_detach_aio_context(&vexp->vu_server); | ||
131 | + vexp->export.ctx = NULL; | ||
132 | } | ||
133 | |||
134 | static void | ||
135 | @@ -XXX,XX +XXX,XX @@ static int vu_blk_exp_create(BlockExport *exp, BlockExportOptions *opts, | ||
136 | vu_blk_initialize_config(blk_bs(exp->blk), &vexp->blkcfg, | ||
137 | logical_block_size); | ||
138 | |||
139 | - blk_set_allow_aio_context_change(exp->blk, true); | ||
140 | blk_add_aio_context_notifier(exp->blk, blk_aio_attached, blk_aio_detach, | ||
141 | vexp); | ||
142 | |||
143 | diff --git a/nbd/server.c b/nbd/server.c | ||
144 | index XXXXXXX..XXXXXXX 100644 | ||
145 | --- a/nbd/server.c | ||
146 | +++ b/nbd/server.c | ||
147 | @@ -XXX,XX +XXX,XX @@ static int nbd_export_create(BlockExport *blk_exp, BlockExportOptions *exp_args, | ||
148 | return ret; | ||
94 | } | 149 | } |
95 | 150 | ||
96 | - if (size > res->image_end_offset) { | 151 | - blk_set_allow_aio_context_change(blk, true); |
97 | - int64_t count; | ||
98 | - count = DIV_ROUND_UP(size - res->image_end_offset, s->cluster_size); | ||
99 | - fprintf(stderr, "%s space leaked at the end of the image %" PRId64 "\n", | ||
100 | - fix & BDRV_FIX_LEAKS ? "Repairing" : "ERROR", | ||
101 | - size - res->image_end_offset); | ||
102 | - res->leaks += count; | ||
103 | - if (fix & BDRV_FIX_LEAKS) { | ||
104 | - Error *local_err = NULL; | ||
105 | - | 152 | - |
106 | - /* | 153 | QTAILQ_INIT(&exp->clients); |
107 | - * In order to really repair the image, we must shrink it. | 154 | exp->name = g_strdup(arg->name); |
108 | - * That means we have to pass exact=true. | 155 | exp->description = g_strdup(arg->description); |
109 | - */ | ||
110 | - ret = bdrv_co_truncate(bs->file, res->image_end_offset, true, | ||
111 | - PREALLOC_MODE_OFF, 0, &local_err); | ||
112 | - if (ret < 0) { | ||
113 | - error_report_err(local_err); | ||
114 | - res->check_errors++; | ||
115 | - goto out; | ||
116 | - } | ||
117 | - res->leaks_fixed += count; | ||
118 | - } | ||
119 | - } | ||
120 | - | ||
121 | out: | ||
122 | qemu_co_mutex_unlock(&s->lock); | ||
123 | |||
124 | -- | 156 | -- |
125 | 2.40.1 | 157 | 2.26.2 |
158 | diff view generated by jsdifflib |
1 | From: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | 1 | Allow the number of queues to be configured using --export |
---|---|---|---|
2 | vhost-user-blk,num-queues=N. This setting should match the QEMU --device | ||
3 | vhost-user-blk-pci,num-queues=N setting but QEMU vhost-user-blk.c lowers | ||
4 | its own value if the vhost-user-blk backend offers fewer queues than | ||
5 | QEMU. | ||
2 | 6 | ||
3 | data_end field in BDRVParallelsState is set to the biggest offset present | 7 | The vhost-user-blk-server.c code is already capable of multi-queue. All |
4 | in BAT. If this offset is outside of the image, any further write will | 8 | virtqueue processing runs in the same AioContext. No new locking is |
5 | create the cluster at this offset and/or the image will be truncated to | 9 | needed. |
6 | this offset on close. This is definitely not correct. | ||
7 | 10 | ||
8 | Raise an error in parallels_open() if data_end points outside the image | 11 | Add the num-queues=N option and set the VIRTIO_BLK_F_MQ feature bit. |
9 | and it is not a check (let the check to repaire the image). Set data_end | 12 | Note that the feature bit only announces the presence of the num_queues |
10 | to the end of the cluster with the last correct offset. | 13 | configuration space field. It does not promise that there is more than 1 |
14 | virtqueue, so we can set it unconditionally. | ||
11 | 15 | ||
12 | Signed-off-by: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | 16 | I tested multi-queue by running a random read fio test with numjobs=4 on |
13 | Message-Id: <20230424093147.197643-2-alexander.ivanov@virtuozzo.com> | 17 | an -smp 4 guest. After the benchmark finished the guest /proc/interrupts |
14 | Reviewed-by: Hanna Czenczek <hreitz@redhat.com> | 18 | file showed activity on all 4 virtio-blk MSI-X. The /sys/block/vda/mq/ |
15 | Signed-off-by: Hanna Czenczek <hreitz@redhat.com> | 19 | directory shows that Linux blk-mq has 4 queues configured. |
20 | |||
21 | An automated test is included in the next commit. | ||
22 | |||
23 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
24 | Acked-by: Markus Armbruster <armbru@redhat.com> | ||
25 | Message-id: 20201001144604.559733-2-stefanha@redhat.com | ||
26 | [Fixed accidental tab characters as suggested by Markus Armbruster | ||
27 | --Stefan] | ||
28 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
16 | --- | 29 | --- |
17 | block/parallels.c | 17 +++++++++++++++++ | 30 | qapi/block-export.json | 10 +++++++--- |
18 | 1 file changed, 17 insertions(+) | 31 | block/export/vhost-user-blk-server.c | 24 ++++++++++++++++++------ |
32 | 2 files changed, 25 insertions(+), 9 deletions(-) | ||
19 | 33 | ||
20 | diff --git a/block/parallels.c b/block/parallels.c | 34 | diff --git a/qapi/block-export.json b/qapi/block-export.json |
21 | index XXXXXXX..XXXXXXX 100644 | 35 | index XXXXXXX..XXXXXXX 100644 |
22 | --- a/block/parallels.c | 36 | --- a/qapi/block-export.json |
23 | +++ b/block/parallels.c | 37 | +++ b/qapi/block-export.json |
24 | @@ -XXX,XX +XXX,XX @@ static int parallels_open(BlockDriverState *bs, QDict *options, int flags, | 38 | @@ -XXX,XX +XXX,XX @@ |
25 | BDRVParallelsState *s = bs->opaque; | 39 | # SocketAddress types are supported. Passed fds must be UNIX domain |
26 | ParallelsHeader ph; | 40 | # sockets. |
27 | int ret, size, i; | 41 | # @logical-block-size: Logical block size in bytes. Defaults to 512 bytes. |
28 | + int64_t file_nb_sectors; | 42 | +# @num-queues: Number of request virtqueues. Must be greater than 0. Defaults |
29 | QemuOpts *opts = NULL; | 43 | +# to 1. |
44 | # | ||
45 | # Since: 5.2 | ||
46 | ## | ||
47 | { 'struct': 'BlockExportOptionsVhostUserBlk', | ||
48 | - 'data': { 'addr': 'SocketAddress', '*logical-block-size': 'size' } } | ||
49 | + 'data': { 'addr': 'SocketAddress', | ||
50 | + '*logical-block-size': 'size', | ||
51 | + '*num-queues': 'uint16'} } | ||
52 | |||
53 | ## | ||
54 | # @NbdServerAddOptions: | ||
55 | @@ -XXX,XX +XXX,XX @@ | ||
56 | { 'union': 'BlockExportOptions', | ||
57 | 'base': { 'type': 'BlockExportType', | ||
58 | 'id': 'str', | ||
59 | - '*fixed-iothread': 'bool', | ||
60 | - '*iothread': 'str', | ||
61 | + '*fixed-iothread': 'bool', | ||
62 | + '*iothread': 'str', | ||
63 | 'node-name': 'str', | ||
64 | '*writable': 'bool', | ||
65 | '*writethrough': 'bool' }, | ||
66 | diff --git a/block/export/vhost-user-blk-server.c b/block/export/vhost-user-blk-server.c | ||
67 | index XXXXXXX..XXXXXXX 100644 | ||
68 | --- a/block/export/vhost-user-blk-server.c | ||
69 | +++ b/block/export/vhost-user-blk-server.c | ||
70 | @@ -XXX,XX +XXX,XX @@ | ||
71 | #include "util/block-helpers.h" | ||
72 | |||
73 | enum { | ||
74 | - VHOST_USER_BLK_MAX_QUEUES = 1, | ||
75 | + VHOST_USER_BLK_NUM_QUEUES_DEFAULT = 1, | ||
76 | }; | ||
77 | struct virtio_blk_inhdr { | ||
78 | unsigned char status; | ||
79 | @@ -XXX,XX +XXX,XX @@ static uint64_t vu_blk_get_features(VuDev *dev) | ||
80 | 1ull << VIRTIO_BLK_F_DISCARD | | ||
81 | 1ull << VIRTIO_BLK_F_WRITE_ZEROES | | ||
82 | 1ull << VIRTIO_BLK_F_CONFIG_WCE | | ||
83 | + 1ull << VIRTIO_BLK_F_MQ | | ||
84 | 1ull << VIRTIO_F_VERSION_1 | | ||
85 | 1ull << VIRTIO_RING_F_INDIRECT_DESC | | ||
86 | 1ull << VIRTIO_RING_F_EVENT_IDX | | ||
87 | @@ -XXX,XX +XXX,XX @@ static void blk_aio_detach(void *opaque) | ||
88 | |||
89 | static void | ||
90 | vu_blk_initialize_config(BlockDriverState *bs, | ||
91 | - struct virtio_blk_config *config, uint32_t blk_size) | ||
92 | + struct virtio_blk_config *config, | ||
93 | + uint32_t blk_size, | ||
94 | + uint16_t num_queues) | ||
95 | { | ||
96 | config->capacity = bdrv_getlength(bs) >> BDRV_SECTOR_BITS; | ||
97 | config->blk_size = blk_size; | ||
98 | @@ -XXX,XX +XXX,XX @@ vu_blk_initialize_config(BlockDriverState *bs, | ||
99 | config->seg_max = 128 - 2; | ||
100 | config->min_io_size = 1; | ||
101 | config->opt_io_size = 1; | ||
102 | - config->num_queues = VHOST_USER_BLK_MAX_QUEUES; | ||
103 | + config->num_queues = num_queues; | ||
104 | config->max_discard_sectors = 32768; | ||
105 | config->max_discard_seg = 1; | ||
106 | config->discard_sector_alignment = config->blk_size >> 9; | ||
107 | @@ -XXX,XX +XXX,XX @@ static int vu_blk_exp_create(BlockExport *exp, BlockExportOptions *opts, | ||
108 | BlockExportOptionsVhostUserBlk *vu_opts = &opts->u.vhost_user_blk; | ||
30 | Error *local_err = NULL; | 109 | Error *local_err = NULL; |
31 | char *buf; | 110 | uint64_t logical_block_size; |
32 | @@ -XXX,XX +XXX,XX @@ static int parallels_open(BlockDriverState *bs, QDict *options, int flags, | 111 | + uint16_t num_queues = VHOST_USER_BLK_NUM_QUEUES_DEFAULT; |
33 | return ret; | 112 | |
113 | vexp->writable = opts->writable; | ||
114 | vexp->blkcfg.wce = 0; | ||
115 | @@ -XXX,XX +XXX,XX @@ static int vu_blk_exp_create(BlockExport *exp, BlockExportOptions *opts, | ||
34 | } | 116 | } |
35 | 117 | vexp->blk_size = logical_block_size; | |
36 | + file_nb_sectors = bdrv_nb_sectors(bs->file->bs); | 118 | blk_set_guest_block_size(exp->blk, logical_block_size); |
37 | + if (file_nb_sectors < 0) { | 119 | + |
120 | + if (vu_opts->has_num_queues) { | ||
121 | + num_queues = vu_opts->num_queues; | ||
122 | + } | ||
123 | + if (num_queues == 0) { | ||
124 | + error_setg(errp, "num-queues must be greater than 0"); | ||
38 | + return -EINVAL; | 125 | + return -EINVAL; |
39 | + } | 126 | + } |
40 | + | 127 | + |
41 | ret = bdrv_pread(bs->file, 0, sizeof(ph), &ph, 0); | 128 | vu_blk_initialize_config(blk_bs(exp->blk), &vexp->blkcfg, |
42 | if (ret < 0) { | 129 | - logical_block_size); |
43 | goto fail; | 130 | + logical_block_size, num_queues); |
44 | @@ -XXX,XX +XXX,XX @@ static int parallels_open(BlockDriverState *bs, QDict *options, int flags, | 131 | |
45 | 132 | blk_add_aio_context_notifier(exp->blk, blk_aio_attached, blk_aio_detach, | |
46 | for (i = 0; i < s->bat_size; i++) { | 133 | vexp); |
47 | int64_t off = bat2sect(s, i); | 134 | |
48 | + if (off >= file_nb_sectors) { | 135 | if (!vhost_user_server_start(&vexp->vu_server, vu_opts->addr, exp->ctx, |
49 | + if (flags & BDRV_O_CHECK) { | 136 | - VHOST_USER_BLK_MAX_QUEUES, &vu_blk_iface, |
50 | + continue; | 137 | - errp)) { |
51 | + } | 138 | + num_queues, &vu_blk_iface, errp)) { |
52 | + error_setg(errp, "parallels: Offset %" PRIi64 " in BAT[%d] entry " | 139 | blk_remove_aio_context_notifier(exp->blk, blk_aio_attached, |
53 | + "is larger than file size (%" PRIi64 ")", | 140 | blk_aio_detach, vexp); |
54 | + off << BDRV_SECTOR_BITS, i, | 141 | return -EADDRNOTAVAIL; |
55 | + file_nb_sectors << BDRV_SECTOR_BITS); | ||
56 | + ret = -EINVAL; | ||
57 | + goto fail; | ||
58 | + } | ||
59 | if (off >= s->data_end) { | ||
60 | s->data_end = off + s->tracks; | ||
61 | } | ||
62 | -- | 142 | -- |
63 | 2.40.1 | 143 | 2.26.2 |
144 | diff view generated by jsdifflib |
1 | From: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | 1 | From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com> |
---|---|---|---|
2 | 2 | ||
3 | We will add more and more checks so we need a better code structure in | 3 | bdrv_co_block_status_above has several design problems with handling |
4 | parallels_co_check. Let each check performs in a separate loop in a | 4 | short backing files: |
5 | separate helper. | ||
6 | 5 | ||
7 | Signed-off-by: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | 6 | 1. With want_zeros=true, it may return ret with BDRV_BLOCK_ZERO but |
8 | Reviewed-by: Denis V. Lunev <den@openvz.org> | 7 | without BDRV_BLOCK_ALLOCATED flag, when actually short backing file |
9 | Message-Id: <20230424093147.197643-8-alexander.ivanov@virtuozzo.com> | 8 | which produces these after-EOF zeros is inside requested backing |
10 | Reviewed-by: Hanna Czenczek <hreitz@redhat.com> | 9 | sequence. |
11 | Signed-off-by: Hanna Czenczek <hreitz@redhat.com> | 10 | |
11 | 2. With want_zero=false, it may return pnum=0 prior to actual EOF, | ||
12 | because of EOF of short backing file. | ||
13 | |||
14 | Fix these things, making logic about short backing files clearer. | ||
15 | |||
16 | With fixed bdrv_block_status_above we also have to improve is_zero in | ||
17 | qcow2 code, otherwise iotest 154 will fail, because with this patch we | ||
18 | stop to merge zeros of different types (produced by fully unallocated | ||
19 | in the whole backing chain regions vs produced by short backing files). | ||
20 | |||
21 | Note also, that this patch leaves for another day the general problem | ||
22 | around block-status: misuse of BDRV_BLOCK_ALLOCATED as is-fs-allocated | ||
23 | vs go-to-backing. | ||
24 | |||
25 | Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com> | ||
26 | Reviewed-by: Alberto Garcia <berto@igalia.com> | ||
27 | Reviewed-by: Eric Blake <eblake@redhat.com> | ||
28 | Message-id: 20200924194003.22080-2-vsementsov@virtuozzo.com | ||
29 | [Fix s/comes/come/ as suggested by Eric Blake | ||
30 | --Stefan] | ||
31 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
12 | --- | 32 | --- |
13 | block/parallels.c | 75 +++++++++++++++++++++++++++++++---------------- | 33 | block/io.c | 68 ++++++++++++++++++++++++++++++++++++++++----------- |
14 | 1 file changed, 49 insertions(+), 26 deletions(-) | 34 | block/qcow2.c | 16 ++++++++++-- |
35 | 2 files changed, 68 insertions(+), 16 deletions(-) | ||
15 | 36 | ||
16 | diff --git a/block/parallels.c b/block/parallels.c | 37 | diff --git a/block/io.c b/block/io.c |
17 | index XXXXXXX..XXXXXXX 100644 | 38 | index XXXXXXX..XXXXXXX 100644 |
18 | --- a/block/parallels.c | 39 | --- a/block/io.c |
19 | +++ b/block/parallels.c | 40 | +++ b/block/io.c |
20 | @@ -XXX,XX +XXX,XX @@ static void parallels_check_unclean(BlockDriverState *bs, | 41 | @@ -XXX,XX +XXX,XX @@ bdrv_co_common_block_status_above(BlockDriverState *bs, |
21 | } | 42 | int64_t *map, |
22 | } | 43 | BlockDriverState **file) |
23 | 44 | { | |
24 | +static int coroutine_fn GRAPH_RDLOCK | 45 | + int ret; |
25 | +parallels_check_outside_image(BlockDriverState *bs, BdrvCheckResult *res, | 46 | BlockDriverState *p; |
26 | + BdrvCheckMode fix) | 47 | - int ret = 0; |
27 | +{ | 48 | - bool first = true; |
28 | + BDRVParallelsState *s = bs->opaque; | 49 | + int64_t eof = 0; |
29 | + uint32_t i; | 50 | |
30 | + int64_t off, high_off, size; | 51 | assert(bs != base); |
52 | - for (p = bs; p != base; p = bdrv_filter_or_cow_bs(p)) { | ||
31 | + | 53 | + |
32 | + size = bdrv_getlength(bs->file->bs); | 54 | + ret = bdrv_co_block_status(bs, want_zero, offset, bytes, pnum, map, file); |
33 | + if (size < 0) { | 55 | + if (ret < 0 || *pnum == 0 || ret & BDRV_BLOCK_ALLOCATED) { |
34 | + res->check_errors++; | 56 | + return ret; |
35 | + return size; | ||
36 | + } | 57 | + } |
37 | + | 58 | + |
38 | + high_off = 0; | 59 | + if (ret & BDRV_BLOCK_EOF) { |
39 | + for (i = 0; i < s->bat_size; i++) { | 60 | + eof = offset + *pnum; |
40 | + off = bat2sect(s, i) << BDRV_SECTOR_BITS; | ||
41 | + if (off > size) { | ||
42 | + fprintf(stderr, "%s cluster %u is outside image\n", | ||
43 | + fix & BDRV_FIX_ERRORS ? "Repairing" : "ERROR", i); | ||
44 | + res->corruptions++; | ||
45 | + if (fix & BDRV_FIX_ERRORS) { | ||
46 | + parallels_set_bat_entry(s, i, 0); | ||
47 | + res->corruptions_fixed++; | ||
48 | + } | ||
49 | + continue; | ||
50 | + } | ||
51 | + if (high_off < off) { | ||
52 | + high_off = off; | ||
53 | + } | ||
54 | + } | 61 | + } |
55 | + | 62 | + |
56 | + if (high_off == 0) { | 63 | + assert(*pnum <= bytes); |
57 | + res->image_end_offset = s->data_end << BDRV_SECTOR_BITS; | 64 | + bytes = *pnum; |
58 | + } else { | 65 | + |
59 | + res->image_end_offset = high_off + s->cluster_size; | 66 | + for (p = bdrv_filter_or_cow_bs(bs); p != base; |
60 | + s->data_end = res->image_end_offset >> BDRV_SECTOR_BITS; | 67 | + p = bdrv_filter_or_cow_bs(p)) |
68 | + { | ||
69 | ret = bdrv_co_block_status(p, want_zero, offset, bytes, pnum, map, | ||
70 | file); | ||
71 | if (ret < 0) { | ||
72 | - break; | ||
73 | + return ret; | ||
74 | } | ||
75 | - if (ret & BDRV_BLOCK_ZERO && ret & BDRV_BLOCK_EOF && !first) { | ||
76 | + if (*pnum == 0) { | ||
77 | /* | ||
78 | - * Reading beyond the end of the file continues to read | ||
79 | - * zeroes, but we can only widen the result to the | ||
80 | - * unallocated length we learned from an earlier | ||
81 | - * iteration. | ||
82 | + * The top layer deferred to this layer, and because this layer is | ||
83 | + * short, any zeroes that we synthesize beyond EOF behave as if they | ||
84 | + * were allocated at this layer. | ||
85 | + * | ||
86 | + * We don't include BDRV_BLOCK_EOF into ret, as upper layer may be | ||
87 | + * larger. We'll add BDRV_BLOCK_EOF if needed at function end, see | ||
88 | + * below. | ||
89 | */ | ||
90 | + assert(ret & BDRV_BLOCK_EOF); | ||
91 | *pnum = bytes; | ||
92 | + if (file) { | ||
93 | + *file = p; | ||
94 | + } | ||
95 | + ret = BDRV_BLOCK_ZERO | BDRV_BLOCK_ALLOCATED; | ||
96 | + break; | ||
97 | } | ||
98 | - if (ret & (BDRV_BLOCK_ZERO | BDRV_BLOCK_DATA)) { | ||
99 | + if (ret & BDRV_BLOCK_ALLOCATED) { | ||
100 | + /* | ||
101 | + * We've found the node and the status, we must break. | ||
102 | + * | ||
103 | + * Drop BDRV_BLOCK_EOF, as it's not for upper layer, which may be | ||
104 | + * larger. We'll add BDRV_BLOCK_EOF if needed at function end, see | ||
105 | + * below. | ||
106 | + */ | ||
107 | + ret &= ~BDRV_BLOCK_EOF; | ||
108 | break; | ||
109 | } | ||
110 | - /* [offset, pnum] unallocated on this layer, which could be only | ||
111 | - * the first part of [offset, bytes]. */ | ||
112 | - bytes = MIN(bytes, *pnum); | ||
113 | - first = false; | ||
114 | + | ||
115 | + /* | ||
116 | + * OK, [offset, offset + *pnum) region is unallocated on this layer, | ||
117 | + * let's continue the diving. | ||
118 | + */ | ||
119 | + assert(*pnum <= bytes); | ||
120 | + bytes = *pnum; | ||
61 | + } | 121 | + } |
62 | + | 122 | + |
63 | + return 0; | 123 | + if (offset + *pnum == eof) { |
64 | +} | 124 | + ret |= BDRV_BLOCK_EOF; |
125 | } | ||
65 | + | 126 | + |
66 | static int coroutine_fn GRAPH_RDLOCK | 127 | return ret; |
67 | parallels_co_check(BlockDriverState *bs, BdrvCheckResult *res, | 128 | } |
68 | BdrvCheckMode fix) | 129 | |
69 | { | 130 | diff --git a/block/qcow2.c b/block/qcow2.c |
70 | BDRVParallelsState *s = bs->opaque; | 131 | index XXXXXXX..XXXXXXX 100644 |
71 | - int64_t size, prev_off, high_off; | 132 | --- a/block/qcow2.c |
72 | - int ret = 0; | 133 | +++ b/block/qcow2.c |
73 | + int64_t size, prev_off; | 134 | @@ -XXX,XX +XXX,XX @@ static bool is_zero(BlockDriverState *bs, int64_t offset, int64_t bytes) |
74 | + int ret; | 135 | if (!bytes) { |
75 | uint32_t i; | 136 | return true; |
76 | 137 | } | |
77 | size = bdrv_getlength(bs->file->bs); | 138 | - res = bdrv_block_status_above(bs, NULL, offset, bytes, &nr, NULL, NULL); |
78 | @@ -XXX,XX +XXX,XX @@ parallels_co_check(BlockDriverState *bs, BdrvCheckResult *res, | 139 | - return res >= 0 && (res & BDRV_BLOCK_ZERO) && nr == bytes; |
79 | |||
80 | parallels_check_unclean(bs, res, fix); | ||
81 | |||
82 | + ret = parallels_check_outside_image(bs, res, fix); | ||
83 | + if (ret < 0) { | ||
84 | + goto out; | ||
85 | + } | ||
86 | + | 140 | + |
87 | res->bfi.total_clusters = s->bat_size; | 141 | + /* |
88 | res->bfi.compressed_clusters = 0; /* compression is not supported */ | 142 | + * bdrv_block_status_above doesn't merge different types of zeros, for |
89 | 143 | + * example, zeros which come from the region which is unallocated in | |
90 | - high_off = 0; | 144 | + * the whole backing chain, and zeros which come because of a short |
91 | prev_off = 0; | 145 | + * backing file. So, we need a loop. |
92 | for (i = 0; i < s->bat_size; i++) { | 146 | + */ |
93 | int64_t off = bat2sect(s, i) << BDRV_SECTOR_BITS; | 147 | + do { |
94 | @@ -XXX,XX +XXX,XX @@ parallels_co_check(BlockDriverState *bs, BdrvCheckResult *res, | 148 | + res = bdrv_block_status_above(bs, NULL, offset, bytes, &nr, NULL, NULL); |
95 | continue; | 149 | + offset += nr; |
96 | } | 150 | + bytes -= nr; |
97 | 151 | + } while (res >= 0 && (res & BDRV_BLOCK_ZERO) && nr && bytes); | |
98 | - /* cluster outside the image */ | 152 | + |
99 | - if (off > size) { | 153 | + return res >= 0 && (res & BDRV_BLOCK_ZERO) && bytes == 0; |
100 | - fprintf(stderr, "%s cluster %u is outside image\n", | 154 | } |
101 | - fix & BDRV_FIX_ERRORS ? "Repairing" : "ERROR", i); | 155 | |
102 | - res->corruptions++; | 156 | static coroutine_fn int qcow2_co_pwrite_zeroes(BlockDriverState *bs, |
103 | - if (fix & BDRV_FIX_ERRORS) { | ||
104 | - parallels_set_bat_entry(s, i, 0); | ||
105 | - res->corruptions_fixed++; | ||
106 | - } | ||
107 | - prev_off = 0; | ||
108 | - continue; | ||
109 | - } | ||
110 | - | ||
111 | res->bfi.allocated_clusters++; | ||
112 | - if (off > high_off) { | ||
113 | - high_off = off; | ||
114 | - } | ||
115 | |||
116 | if (prev_off != 0 && (prev_off + s->cluster_size) != off) { | ||
117 | res->bfi.fragmented_clusters++; | ||
118 | @@ -XXX,XX +XXX,XX @@ parallels_co_check(BlockDriverState *bs, BdrvCheckResult *res, | ||
119 | prev_off = off; | ||
120 | } | ||
121 | |||
122 | - if (high_off == 0) { | ||
123 | - res->image_end_offset = s->data_end << BDRV_SECTOR_BITS; | ||
124 | - } else { | ||
125 | - res->image_end_offset = high_off + s->cluster_size; | ||
126 | - s->data_end = res->image_end_offset >> BDRV_SECTOR_BITS; | ||
127 | - } | ||
128 | - | ||
129 | if (size > res->image_end_offset) { | ||
130 | int64_t count; | ||
131 | count = DIV_ROUND_UP(size - res->image_end_offset, s->cluster_size); | ||
132 | -- | 157 | -- |
133 | 2.40.1 | 158 | 2.26.2 |
159 | diff view generated by jsdifflib |
1 | We want to inline qemu_iovec_init_extended() in block/io.c for padding | 1 | From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com> |
---|---|---|---|
2 | requests, and having access to qiov_slice() is useful for this. As a | ||
3 | public function, it is renamed to qemu_iovec_slice(). | ||
4 | 2 | ||
5 | (We will need to count the number of I/O vector elements of a slice | 3 | In order to reuse bdrv_common_block_status_above in |
6 | there, and then later process this slice. Without qiov_slice(), we | 4 | bdrv_is_allocated_above, let's support include_base parameter. |
7 | would need to call qemu_iovec_subvec_niov(), and all further | ||
8 | IOV-processing functions may need to skip prefixing elements to | ||
9 | accomodate for a qiov_offset. Because qemu_iovec_subvec_niov() | ||
10 | internally calls qiov_slice(), we can just have the block/io.c code call | ||
11 | qiov_slice() itself, thus get the number of elements, and also create an | ||
12 | iovec array with the superfluous prefixing elements stripped, so the | ||
13 | following processing functions no longer need to skip them.) | ||
14 | 5 | ||
6 | Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com> | ||
7 | Reviewed-by: Alberto Garcia <berto@igalia.com> | ||
15 | Reviewed-by: Eric Blake <eblake@redhat.com> | 8 | Reviewed-by: Eric Blake <eblake@redhat.com> |
16 | Reviewed-by: Vladimir Sementsov-Ogievskiy <vsementsov@yandex-team.ru> | 9 | Message-id: 20200924194003.22080-3-vsementsov@virtuozzo.com |
17 | Signed-off-by: Hanna Czenczek <hreitz@redhat.com> | 10 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
18 | Message-Id: <20230411173418.19549-2-hreitz@redhat.com> | ||
19 | --- | 11 | --- |
20 | include/qemu/iov.h | 3 +++ | 12 | block/coroutines.h | 2 ++ |
21 | util/iov.c | 14 +++++++------- | 13 | block/io.c | 21 ++++++++++++++------- |
22 | 2 files changed, 10 insertions(+), 7 deletions(-) | 14 | 2 files changed, 16 insertions(+), 7 deletions(-) |
23 | 15 | ||
24 | diff --git a/include/qemu/iov.h b/include/qemu/iov.h | 16 | diff --git a/block/coroutines.h b/block/coroutines.h |
25 | index XXXXXXX..XXXXXXX 100644 | 17 | index XXXXXXX..XXXXXXX 100644 |
26 | --- a/include/qemu/iov.h | 18 | --- a/block/coroutines.h |
27 | +++ b/include/qemu/iov.h | 19 | +++ b/block/coroutines.h |
28 | @@ -XXX,XX +XXX,XX @@ int qemu_iovec_init_extended( | 20 | @@ -XXX,XX +XXX,XX @@ bdrv_pwritev(BdrvChild *child, int64_t offset, unsigned int bytes, |
29 | void *tail_buf, size_t tail_len); | 21 | int coroutine_fn |
30 | void qemu_iovec_init_slice(QEMUIOVector *qiov, QEMUIOVector *source, | 22 | bdrv_co_common_block_status_above(BlockDriverState *bs, |
31 | size_t offset, size_t len); | 23 | BlockDriverState *base, |
32 | +struct iovec *qemu_iovec_slice(QEMUIOVector *qiov, | 24 | + bool include_base, |
33 | + size_t offset, size_t len, | 25 | bool want_zero, |
34 | + size_t *head, size_t *tail, int *niov); | 26 | int64_t offset, |
35 | int qemu_iovec_subvec_niov(QEMUIOVector *qiov, size_t offset, size_t len); | 27 | int64_t bytes, |
36 | void qemu_iovec_add(QEMUIOVector *qiov, void *base, size_t len); | 28 | @@ -XXX,XX +XXX,XX @@ bdrv_co_common_block_status_above(BlockDriverState *bs, |
37 | void qemu_iovec_concat(QEMUIOVector *dst, | 29 | int generated_co_wrapper |
38 | diff --git a/util/iov.c b/util/iov.c | 30 | bdrv_common_block_status_above(BlockDriverState *bs, |
31 | BlockDriverState *base, | ||
32 | + bool include_base, | ||
33 | bool want_zero, | ||
34 | int64_t offset, | ||
35 | int64_t bytes, | ||
36 | diff --git a/block/io.c b/block/io.c | ||
39 | index XXXXXXX..XXXXXXX 100644 | 37 | index XXXXXXX..XXXXXXX 100644 |
40 | --- a/util/iov.c | 38 | --- a/block/io.c |
41 | +++ b/util/iov.c | 39 | +++ b/block/io.c |
42 | @@ -XXX,XX +XXX,XX @@ static struct iovec *iov_skip_offset(struct iovec *iov, size_t offset, | 40 | @@ -XXX,XX +XXX,XX @@ early_out: |
41 | int coroutine_fn | ||
42 | bdrv_co_common_block_status_above(BlockDriverState *bs, | ||
43 | BlockDriverState *base, | ||
44 | + bool include_base, | ||
45 | bool want_zero, | ||
46 | int64_t offset, | ||
47 | int64_t bytes, | ||
48 | @@ -XXX,XX +XXX,XX @@ bdrv_co_common_block_status_above(BlockDriverState *bs, | ||
49 | BlockDriverState *p; | ||
50 | int64_t eof = 0; | ||
51 | |||
52 | - assert(bs != base); | ||
53 | + assert(include_base || bs != base); | ||
54 | + assert(!include_base || base); /* Can't include NULL base */ | ||
55 | |||
56 | ret = bdrv_co_block_status(bs, want_zero, offset, bytes, pnum, map, file); | ||
57 | - if (ret < 0 || *pnum == 0 || ret & BDRV_BLOCK_ALLOCATED) { | ||
58 | + if (ret < 0 || *pnum == 0 || ret & BDRV_BLOCK_ALLOCATED || bs == base) { | ||
59 | return ret; | ||
60 | } | ||
61 | |||
62 | @@ -XXX,XX +XXX,XX @@ bdrv_co_common_block_status_above(BlockDriverState *bs, | ||
63 | assert(*pnum <= bytes); | ||
64 | bytes = *pnum; | ||
65 | |||
66 | - for (p = bdrv_filter_or_cow_bs(bs); p != base; | ||
67 | + for (p = bdrv_filter_or_cow_bs(bs); include_base || p != base; | ||
68 | p = bdrv_filter_or_cow_bs(p)) | ||
69 | { | ||
70 | ret = bdrv_co_block_status(p, want_zero, offset, bytes, pnum, map, | ||
71 | @@ -XXX,XX +XXX,XX @@ bdrv_co_common_block_status_above(BlockDriverState *bs, | ||
72 | break; | ||
73 | } | ||
74 | |||
75 | + if (p == base) { | ||
76 | + assert(include_base); | ||
77 | + break; | ||
78 | + } | ||
79 | + | ||
80 | /* | ||
81 | * OK, [offset, offset + *pnum) region is unallocated on this layer, | ||
82 | * let's continue the diving. | ||
83 | @@ -XXX,XX +XXX,XX @@ int bdrv_block_status_above(BlockDriverState *bs, BlockDriverState *base, | ||
84 | int64_t offset, int64_t bytes, int64_t *pnum, | ||
85 | int64_t *map, BlockDriverState **file) | ||
86 | { | ||
87 | - return bdrv_common_block_status_above(bs, base, true, offset, bytes, | ||
88 | + return bdrv_common_block_status_above(bs, base, false, true, offset, bytes, | ||
89 | pnum, map, file); | ||
43 | } | 90 | } |
44 | 91 | ||
45 | /* | 92 | @@ -XXX,XX +XXX,XX @@ int coroutine_fn bdrv_is_allocated(BlockDriverState *bs, int64_t offset, |
46 | - * qiov_slice | 93 | int ret; |
47 | + * qemu_iovec_slice | 94 | int64_t dummy; |
48 | * | 95 | |
49 | * Find subarray of iovec's, containing requested range. @head would | 96 | - ret = bdrv_common_block_status_above(bs, bdrv_filter_or_cow_bs(bs), false, |
50 | * be offset in first iov (returned by the function), @tail would be | 97 | - offset, bytes, pnum ? pnum : &dummy, |
51 | * count of extra bytes in last iovec (returned iov + @niov - 1). | 98 | - NULL, NULL); |
52 | */ | 99 | + ret = bdrv_common_block_status_above(bs, bs, true, false, offset, |
53 | -static struct iovec *qiov_slice(QEMUIOVector *qiov, | 100 | + bytes, pnum ? pnum : &dummy, NULL, |
54 | - size_t offset, size_t len, | 101 | + NULL); |
55 | - size_t *head, size_t *tail, int *niov) | 102 | if (ret < 0) { |
56 | +struct iovec *qemu_iovec_slice(QEMUIOVector *qiov, | 103 | return ret; |
57 | + size_t offset, size_t len, | ||
58 | + size_t *head, size_t *tail, int *niov) | ||
59 | { | ||
60 | struct iovec *iov, *end_iov; | ||
61 | |||
62 | @@ -XXX,XX +XXX,XX @@ int qemu_iovec_subvec_niov(QEMUIOVector *qiov, size_t offset, size_t len) | ||
63 | size_t head, tail; | ||
64 | int niov; | ||
65 | |||
66 | - qiov_slice(qiov, offset, len, &head, &tail, &niov); | ||
67 | + qemu_iovec_slice(qiov, offset, len, &head, &tail, &niov); | ||
68 | |||
69 | return niov; | ||
70 | } | ||
71 | @@ -XXX,XX +XXX,XX @@ int qemu_iovec_init_extended( | ||
72 | } | 104 | } |
73 | |||
74 | if (mid_len) { | ||
75 | - mid_iov = qiov_slice(mid_qiov, mid_offset, mid_len, | ||
76 | - &mid_head, &mid_tail, &mid_niov); | ||
77 | + mid_iov = qemu_iovec_slice(mid_qiov, mid_offset, mid_len, | ||
78 | + &mid_head, &mid_tail, &mid_niov); | ||
79 | } | ||
80 | |||
81 | total_niov = !!head_len + mid_niov + !!tail_len; | ||
82 | -- | 105 | -- |
83 | 2.40.1 | 106 | 2.26.2 |
107 | diff view generated by jsdifflib |
1 | From: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | 1 | From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com> |
---|---|---|---|
2 | 2 | ||
3 | Exclude out-of-image clusters from allocated and fragmented clusters | 3 | We are going to reuse bdrv_common_block_status_above in |
4 | calculation. | 4 | bdrv_is_allocated_above. bdrv_is_allocated_above may be called with |
5 | include_base == false and still bs == base (for ex. from img_rebase()). | ||
5 | 6 | ||
6 | Signed-off-by: Alexander Ivanov <alexander.ivanov@virtuozzo.com> | 7 | So, support this corner case. |
7 | Message-Id: <20230424093147.197643-9-alexander.ivanov@virtuozzo.com> | 8 | |
8 | Reviewed-by: Hanna Czenczek <hreitz@redhat.com> | 9 | Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com> |
9 | Signed-off-by: Hanna Czenczek <hreitz@redhat.com> | 10 | Reviewed-by: Kevin Wolf <kwolf@redhat.com> |
11 | Reviewed-by: Eric Blake <eblake@redhat.com> | ||
12 | Reviewed-by: Alberto Garcia <berto@igalia.com> | ||
13 | Message-id: 20200924194003.22080-4-vsementsov@virtuozzo.com | ||
14 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
10 | --- | 15 | --- |
11 | block/parallels.c | 6 +++++- | 16 | block/io.c | 6 +++++- |
12 | 1 file changed, 5 insertions(+), 1 deletion(-) | 17 | 1 file changed, 5 insertions(+), 1 deletion(-) |
13 | 18 | ||
14 | diff --git a/block/parallels.c b/block/parallels.c | 19 | diff --git a/block/io.c b/block/io.c |
15 | index XXXXXXX..XXXXXXX 100644 | 20 | index XXXXXXX..XXXXXXX 100644 |
16 | --- a/block/parallels.c | 21 | --- a/block/io.c |
17 | +++ b/block/parallels.c | 22 | +++ b/block/io.c |
18 | @@ -XXX,XX +XXX,XX @@ parallels_co_check(BlockDriverState *bs, BdrvCheckResult *res, | 23 | @@ -XXX,XX +XXX,XX @@ bdrv_co_common_block_status_above(BlockDriverState *bs, |
19 | prev_off = 0; | 24 | BlockDriverState *p; |
20 | for (i = 0; i < s->bat_size; i++) { | 25 | int64_t eof = 0; |
21 | int64_t off = bat2sect(s, i) << BDRV_SECTOR_BITS; | 26 | |
22 | - if (off == 0) { | 27 | - assert(include_base || bs != base); |
23 | + /* | 28 | assert(!include_base || base); /* Can't include NULL base */ |
24 | + * If BDRV_FIX_ERRORS is not set, out-of-image BAT entries were not | 29 | |
25 | + * fixed. Skip not allocated and out-of-image BAT entries. | 30 | + if (!include_base && bs == base) { |
26 | + */ | 31 | + *pnum = bytes; |
27 | + if (off == 0 || off + s->cluster_size > res->image_end_offset) { | 32 | + return 0; |
28 | prev_off = 0; | 33 | + } |
29 | continue; | 34 | + |
30 | } | 35 | ret = bdrv_co_block_status(bs, want_zero, offset, bytes, pnum, map, file); |
36 | if (ret < 0 || *pnum == 0 || ret & BDRV_BLOCK_ALLOCATED || bs == base) { | ||
37 | return ret; | ||
31 | -- | 38 | -- |
32 | 2.40.1 | 39 | 2.26.2 |
40 | diff view generated by jsdifflib |
1 | When processing vectored guest requests that are not aligned to the | 1 | From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com> |
---|---|---|---|
2 | storage request alignment, we pad them by adding head and/or tail | ||
3 | buffers for a read-modify-write cycle. | ||
4 | 2 | ||
5 | The guest can submit I/O vectors up to IOV_MAX (1024) in length, but | 3 | bdrv_is_allocated_above wrongly handles short backing files: it reports |
6 | with this padding, the vector can exceed that limit. As of | 4 | after-EOF space as UNALLOCATED which is wrong, as on read the data is |
7 | 4c002cef0e9abe7135d7916c51abce47f7fc1ee2 ("util/iov: make | 5 | generated on the level of short backing file (if all overlays have |
8 | qemu_iovec_init_extended() honest"), we refuse to pad vectors beyond the | 6 | unallocated areas at that place). |
9 | limit, instead returning an error to the guest. | ||
10 | 7 | ||
11 | To the guest, this appears as a random I/O error. We should not return | 8 | Reusing bdrv_common_block_status_above fixes the issue and unifies code |
12 | an I/O error to the guest when it issued a perfectly valid request. | 9 | path. |
13 | 10 | ||
14 | Before 4c002cef0e9abe7135d7916c51abce47f7fc1ee2, we just made the vector | 11 | Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com> |
15 | longer than IOV_MAX, which generally seems to work (because the guest | 12 | Reviewed-by: Eric Blake <eblake@redhat.com> |
16 | assumes a smaller alignment than we really have, file-posix's | 13 | Reviewed-by: Alberto Garcia <berto@igalia.com> |
17 | raw_co_prw() will generally see bdrv_qiov_is_aligned() return false, and | 14 | Message-id: 20200924194003.22080-5-vsementsov@virtuozzo.com |
18 | so emulate the request, so that the IOV_MAX does not matter). However, | 15 | [Fix s/has/have/ as suggested by Eric Blake. Fix s/area/areas/. |
19 | that does not seem exactly great. | 16 | --Stefan] |
20 | 17 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | |
21 | I see two ways to fix this problem: | ||
22 | 1. We split such long requests into two requests. | ||
23 | 2. We join some elements of the vector into new buffers to make it | ||
24 | shorter. | ||
25 | |||
26 | I am wary of (1), because it seems like it may have unintended side | ||
27 | effects. | ||
28 | |||
29 | (2) on the other hand seems relatively simple to implement, with | ||
30 | hopefully few side effects, so this patch does that. | ||
31 | |||
32 | To do this, the use of qemu_iovec_init_extended() in bdrv_pad_request() | ||
33 | is effectively replaced by the new function bdrv_create_padded_qiov(), | ||
34 | which not only wraps the request IOV with padding head/tail, but also | ||
35 | ensures that the resulting vector will not have more than IOV_MAX | ||
36 | elements. Putting that functionality into qemu_iovec_init_extended() is | ||
37 | infeasible because it requires allocating a bounce buffer; doing so | ||
38 | would require many more parameters (buffer alignment, how to initialize | ||
39 | the buffer, and out parameters like the buffer, its length, and the | ||
40 | original elements), which is not reasonable. | ||
41 | |||
42 | Conversely, it is not difficult to move qemu_iovec_init_extended()'s | ||
43 | functionality into bdrv_create_padded_qiov() by using public | ||
44 | qemu_iovec_* functions, so that is what this patch does. | ||
45 | |||
46 | Because bdrv_pad_request() was the only "serious" user of | ||
47 | qemu_iovec_init_extended(), the next patch will remove the latter | ||
48 | function, so the functionality is not implemented twice. | ||
49 | |||
50 | Buglink: https://bugzilla.redhat.com/show_bug.cgi?id=2141964 | ||
51 | Signed-off-by: Hanna Czenczek <hreitz@redhat.com> | ||
52 | Message-Id: <20230411173418.19549-3-hreitz@redhat.com> | ||
53 | Reviewed-by: Vladimir Sementsov-Ogievskiy <vsementsov@yandex-team.ru> | ||
54 | --- | 18 | --- |
55 | block/io.c | 166 ++++++++++++++++++++++++++++++++++++++++++++++++----- | 19 | block/io.c | 43 +++++-------------------------------------- |
56 | 1 file changed, 151 insertions(+), 15 deletions(-) | 20 | 1 file changed, 5 insertions(+), 38 deletions(-) |
57 | 21 | ||
58 | diff --git a/block/io.c b/block/io.c | 22 | diff --git a/block/io.c b/block/io.c |
59 | index XXXXXXX..XXXXXXX 100644 | 23 | index XXXXXXX..XXXXXXX 100644 |
60 | --- a/block/io.c | 24 | --- a/block/io.c |
61 | +++ b/block/io.c | 25 | +++ b/block/io.c |
62 | @@ -XXX,XX +XXX,XX @@ out: | 26 | @@ -XXX,XX +XXX,XX @@ int coroutine_fn bdrv_is_allocated(BlockDriverState *bs, int64_t offset, |
63 | * @merge_reads is true for small requests, | 27 | * at 'offset + *pnum' may return the same allocation status (in other |
64 | * if @buf_len == @head + bytes + @tail. In this case it is possible that both | 28 | * words, the result is not necessarily the maximum possible range); |
65 | * head and tail exist but @buf_len == align and @tail_buf == @buf. | 29 | * but 'pnum' will only be 0 when end of file is reached. |
66 | + * | 30 | - * |
67 | + * @write is true for write requests, false for read requests. | ||
68 | + * | ||
69 | + * If padding makes the vector too long (exceeding IOV_MAX), then we need to | ||
70 | + * merge existing vector elements into a single one. @collapse_bounce_buf acts | ||
71 | + * as the bounce buffer in such cases. @pre_collapse_qiov has the pre-collapse | ||
72 | + * I/O vector elements so for read requests, the data can be copied back after | ||
73 | + * the read is done. | ||
74 | */ | 31 | */ |
75 | typedef struct BdrvRequestPadding { | 32 | int bdrv_is_allocated_above(BlockDriverState *top, |
76 | uint8_t *buf; | 33 | BlockDriverState *base, |
77 | @@ -XXX,XX +XXX,XX @@ typedef struct BdrvRequestPadding { | 34 | bool include_base, int64_t offset, |
78 | size_t head; | 35 | int64_t bytes, int64_t *pnum) |
79 | size_t tail; | ||
80 | bool merge_reads; | ||
81 | + bool write; | ||
82 | QEMUIOVector local_qiov; | ||
83 | + | ||
84 | + uint8_t *collapse_bounce_buf; | ||
85 | + size_t collapse_len; | ||
86 | + QEMUIOVector pre_collapse_qiov; | ||
87 | } BdrvRequestPadding; | ||
88 | |||
89 | static bool bdrv_init_padding(BlockDriverState *bs, | ||
90 | int64_t offset, int64_t bytes, | ||
91 | + bool write, | ||
92 | BdrvRequestPadding *pad) | ||
93 | { | 36 | { |
94 | int64_t align = bs->bl.request_alignment; | 37 | - BlockDriverState *intermediate; |
95 | @@ -XXX,XX +XXX,XX @@ static bool bdrv_init_padding(BlockDriverState *bs, | 38 | - int ret; |
96 | pad->tail_buf = pad->buf + pad->buf_len - align; | 39 | - int64_t n = bytes; |
40 | - | ||
41 | - assert(base || !include_base); | ||
42 | - | ||
43 | - intermediate = top; | ||
44 | - while (include_base || intermediate != base) { | ||
45 | - int64_t pnum_inter; | ||
46 | - int64_t size_inter; | ||
47 | - | ||
48 | - assert(intermediate); | ||
49 | - ret = bdrv_is_allocated(intermediate, offset, bytes, &pnum_inter); | ||
50 | - if (ret < 0) { | ||
51 | - return ret; | ||
52 | - } | ||
53 | - if (ret) { | ||
54 | - *pnum = pnum_inter; | ||
55 | - return 1; | ||
56 | - } | ||
57 | - | ||
58 | - size_inter = bdrv_getlength(intermediate); | ||
59 | - if (size_inter < 0) { | ||
60 | - return size_inter; | ||
61 | - } | ||
62 | - if (n > pnum_inter && | ||
63 | - (intermediate == top || offset + pnum_inter < size_inter)) { | ||
64 | - n = pnum_inter; | ||
65 | - } | ||
66 | - | ||
67 | - if (intermediate == base) { | ||
68 | - break; | ||
69 | - } | ||
70 | - | ||
71 | - intermediate = bdrv_filter_or_cow_bs(intermediate); | ||
72 | + int ret = bdrv_common_block_status_above(top, base, include_base, false, | ||
73 | + offset, bytes, pnum, NULL, NULL); | ||
74 | + if (ret < 0) { | ||
75 | + return ret; | ||
97 | } | 76 | } |
98 | 77 | ||
99 | + pad->write = write; | 78 | - *pnum = n; |
100 | + | 79 | - return 0; |
101 | return true; | 80 | + return !!(ret & BDRV_BLOCK_ALLOCATED); |
102 | } | 81 | } |
103 | 82 | ||
104 | @@ -XXX,XX +XXX,XX @@ zero_mem: | 83 | int coroutine_fn |
105 | return 0; | ||
106 | } | ||
107 | |||
108 | -static void bdrv_padding_destroy(BdrvRequestPadding *pad) | ||
109 | +/** | ||
110 | + * Free *pad's associated buffers, and perform any necessary finalization steps. | ||
111 | + */ | ||
112 | +static void bdrv_padding_finalize(BdrvRequestPadding *pad) | ||
113 | { | ||
114 | + if (pad->collapse_bounce_buf) { | ||
115 | + if (!pad->write) { | ||
116 | + /* | ||
117 | + * If padding required elements in the vector to be collapsed into a | ||
118 | + * bounce buffer, copy the bounce buffer content back | ||
119 | + */ | ||
120 | + qemu_iovec_from_buf(&pad->pre_collapse_qiov, 0, | ||
121 | + pad->collapse_bounce_buf, pad->collapse_len); | ||
122 | + } | ||
123 | + qemu_vfree(pad->collapse_bounce_buf); | ||
124 | + qemu_iovec_destroy(&pad->pre_collapse_qiov); | ||
125 | + } | ||
126 | if (pad->buf) { | ||
127 | qemu_vfree(pad->buf); | ||
128 | qemu_iovec_destroy(&pad->local_qiov); | ||
129 | @@ -XXX,XX +XXX,XX @@ static void bdrv_padding_destroy(BdrvRequestPadding *pad) | ||
130 | memset(pad, 0, sizeof(*pad)); | ||
131 | } | ||
132 | |||
133 | +/* | ||
134 | + * Create pad->local_qiov by wrapping @iov in the padding head and tail, while | ||
135 | + * ensuring that the resulting vector will not exceed IOV_MAX elements. | ||
136 | + * | ||
137 | + * To ensure this, when necessary, the first two or three elements of @iov are | ||
138 | + * merged into pad->collapse_bounce_buf and replaced by a reference to that | ||
139 | + * bounce buffer in pad->local_qiov. | ||
140 | + * | ||
141 | + * After performing a read request, the data from the bounce buffer must be | ||
142 | + * copied back into pad->pre_collapse_qiov (e.g. by bdrv_padding_finalize()). | ||
143 | + */ | ||
144 | +static int bdrv_create_padded_qiov(BlockDriverState *bs, | ||
145 | + BdrvRequestPadding *pad, | ||
146 | + struct iovec *iov, int niov, | ||
147 | + size_t iov_offset, size_t bytes) | ||
148 | +{ | ||
149 | + int padded_niov, surplus_count, collapse_count; | ||
150 | + | ||
151 | + /* Assert this invariant */ | ||
152 | + assert(niov <= IOV_MAX); | ||
153 | + | ||
154 | + /* | ||
155 | + * Cannot pad if resulting length would exceed SIZE_MAX. Returning an error | ||
156 | + * to the guest is not ideal, but there is little else we can do. At least | ||
157 | + * this will practically never happen on 64-bit systems. | ||
158 | + */ | ||
159 | + if (SIZE_MAX - pad->head < bytes || | ||
160 | + SIZE_MAX - pad->head - bytes < pad->tail) | ||
161 | + { | ||
162 | + return -EINVAL; | ||
163 | + } | ||
164 | + | ||
165 | + /* Length of the resulting IOV if we just concatenated everything */ | ||
166 | + padded_niov = !!pad->head + niov + !!pad->tail; | ||
167 | + | ||
168 | + qemu_iovec_init(&pad->local_qiov, MIN(padded_niov, IOV_MAX)); | ||
169 | + | ||
170 | + if (pad->head) { | ||
171 | + qemu_iovec_add(&pad->local_qiov, pad->buf, pad->head); | ||
172 | + } | ||
173 | + | ||
174 | + /* | ||
175 | + * If padded_niov > IOV_MAX, we cannot just concatenate everything. | ||
176 | + * Instead, merge the first two or three elements of @iov to reduce the | ||
177 | + * number of vector elements as necessary. | ||
178 | + */ | ||
179 | + if (padded_niov > IOV_MAX) { | ||
180 | + /* | ||
181 | + * Only head and tail can have lead to the number of entries exceeding | ||
182 | + * IOV_MAX, so we can exceed it by the head and tail at most. We need | ||
183 | + * to reduce the number of elements by `surplus_count`, so we merge that | ||
184 | + * many elements plus one into one element. | ||
185 | + */ | ||
186 | + surplus_count = padded_niov - IOV_MAX; | ||
187 | + assert(surplus_count <= !!pad->head + !!pad->tail); | ||
188 | + collapse_count = surplus_count + 1; | ||
189 | + | ||
190 | + /* | ||
191 | + * Move the elements to collapse into `pad->pre_collapse_qiov`, then | ||
192 | + * advance `iov` (and associated variables) by those elements. | ||
193 | + */ | ||
194 | + qemu_iovec_init(&pad->pre_collapse_qiov, collapse_count); | ||
195 | + qemu_iovec_concat_iov(&pad->pre_collapse_qiov, iov, | ||
196 | + collapse_count, iov_offset, SIZE_MAX); | ||
197 | + iov += collapse_count; | ||
198 | + iov_offset = 0; | ||
199 | + niov -= collapse_count; | ||
200 | + bytes -= pad->pre_collapse_qiov.size; | ||
201 | + | ||
202 | + /* | ||
203 | + * Construct the bounce buffer to match the length of the to-collapse | ||
204 | + * vector elements, and for write requests, initialize it with the data | ||
205 | + * from those elements. Then add it to `pad->local_qiov`. | ||
206 | + */ | ||
207 | + pad->collapse_len = pad->pre_collapse_qiov.size; | ||
208 | + pad->collapse_bounce_buf = qemu_blockalign(bs, pad->collapse_len); | ||
209 | + if (pad->write) { | ||
210 | + qemu_iovec_to_buf(&pad->pre_collapse_qiov, 0, | ||
211 | + pad->collapse_bounce_buf, pad->collapse_len); | ||
212 | + } | ||
213 | + qemu_iovec_add(&pad->local_qiov, | ||
214 | + pad->collapse_bounce_buf, pad->collapse_len); | ||
215 | + } | ||
216 | + | ||
217 | + qemu_iovec_concat_iov(&pad->local_qiov, iov, niov, iov_offset, bytes); | ||
218 | + | ||
219 | + if (pad->tail) { | ||
220 | + qemu_iovec_add(&pad->local_qiov, | ||
221 | + pad->buf + pad->buf_len - pad->tail, pad->tail); | ||
222 | + } | ||
223 | + | ||
224 | + assert(pad->local_qiov.niov == MIN(padded_niov, IOV_MAX)); | ||
225 | + return 0; | ||
226 | +} | ||
227 | + | ||
228 | /* | ||
229 | * bdrv_pad_request | ||
230 | * | ||
231 | @@ -XXX,XX +XXX,XX @@ static void bdrv_padding_destroy(BdrvRequestPadding *pad) | ||
232 | * read of padding, bdrv_padding_rmw_read() should be called separately if | ||
233 | * needed. | ||
234 | * | ||
235 | + * @write is true for write requests, false for read requests. | ||
236 | + * | ||
237 | * Request parameters (@qiov, &qiov_offset, &offset, &bytes) are in-out: | ||
238 | * - on function start they represent original request | ||
239 | * - on failure or when padding is not needed they are unchanged | ||
240 | @@ -XXX,XX +XXX,XX @@ static void bdrv_padding_destroy(BdrvRequestPadding *pad) | ||
241 | static int bdrv_pad_request(BlockDriverState *bs, | ||
242 | QEMUIOVector **qiov, size_t *qiov_offset, | ||
243 | int64_t *offset, int64_t *bytes, | ||
244 | + bool write, | ||
245 | BdrvRequestPadding *pad, bool *padded, | ||
246 | BdrvRequestFlags *flags) | ||
247 | { | ||
248 | int ret; | ||
249 | + struct iovec *sliced_iov; | ||
250 | + int sliced_niov; | ||
251 | + size_t sliced_head, sliced_tail; | ||
252 | |||
253 | bdrv_check_qiov_request(*offset, *bytes, *qiov, *qiov_offset, &error_abort); | ||
254 | |||
255 | - if (!bdrv_init_padding(bs, *offset, *bytes, pad)) { | ||
256 | + if (!bdrv_init_padding(bs, *offset, *bytes, write, pad)) { | ||
257 | if (padded) { | ||
258 | *padded = false; | ||
259 | } | ||
260 | return 0; | ||
261 | } | ||
262 | |||
263 | - ret = qemu_iovec_init_extended(&pad->local_qiov, pad->buf, pad->head, | ||
264 | - *qiov, *qiov_offset, *bytes, | ||
265 | - pad->buf + pad->buf_len - pad->tail, | ||
266 | - pad->tail); | ||
267 | + sliced_iov = qemu_iovec_slice(*qiov, *qiov_offset, *bytes, | ||
268 | + &sliced_head, &sliced_tail, | ||
269 | + &sliced_niov); | ||
270 | + | ||
271 | + /* Guaranteed by bdrv_check_qiov_request() */ | ||
272 | + assert(*bytes <= SIZE_MAX); | ||
273 | + ret = bdrv_create_padded_qiov(bs, pad, sliced_iov, sliced_niov, | ||
274 | + sliced_head, *bytes); | ||
275 | if (ret < 0) { | ||
276 | - bdrv_padding_destroy(pad); | ||
277 | + bdrv_padding_finalize(pad); | ||
278 | return ret; | ||
279 | } | ||
280 | *bytes += pad->head + pad->tail; | ||
281 | @@ -XXX,XX +XXX,XX @@ int coroutine_fn bdrv_co_preadv_part(BdrvChild *child, | ||
282 | flags |= BDRV_REQ_COPY_ON_READ; | ||
283 | } | ||
284 | |||
285 | - ret = bdrv_pad_request(bs, &qiov, &qiov_offset, &offset, &bytes, &pad, | ||
286 | - NULL, &flags); | ||
287 | + ret = bdrv_pad_request(bs, &qiov, &qiov_offset, &offset, &bytes, false, | ||
288 | + &pad, NULL, &flags); | ||
289 | if (ret < 0) { | ||
290 | goto fail; | ||
291 | } | ||
292 | @@ -XXX,XX +XXX,XX @@ int coroutine_fn bdrv_co_preadv_part(BdrvChild *child, | ||
293 | bs->bl.request_alignment, | ||
294 | qiov, qiov_offset, flags); | ||
295 | tracked_request_end(&req); | ||
296 | - bdrv_padding_destroy(&pad); | ||
297 | + bdrv_padding_finalize(&pad); | ||
298 | |||
299 | fail: | ||
300 | bdrv_dec_in_flight(bs); | ||
301 | @@ -XXX,XX +XXX,XX @@ bdrv_co_do_zero_pwritev(BdrvChild *child, int64_t offset, int64_t bytes, | ||
302 | /* This flag doesn't make sense for padding or zero writes */ | ||
303 | flags &= ~BDRV_REQ_REGISTERED_BUF; | ||
304 | |||
305 | - padding = bdrv_init_padding(bs, offset, bytes, &pad); | ||
306 | + padding = bdrv_init_padding(bs, offset, bytes, true, &pad); | ||
307 | if (padding) { | ||
308 | assert(!(flags & BDRV_REQ_NO_WAIT)); | ||
309 | bdrv_make_request_serialising(req, align); | ||
310 | @@ -XXX,XX +XXX,XX @@ bdrv_co_do_zero_pwritev(BdrvChild *child, int64_t offset, int64_t bytes, | ||
311 | } | ||
312 | |||
313 | out: | ||
314 | - bdrv_padding_destroy(&pad); | ||
315 | + bdrv_padding_finalize(&pad); | ||
316 | |||
317 | return ret; | ||
318 | } | ||
319 | @@ -XXX,XX +XXX,XX @@ int coroutine_fn bdrv_co_pwritev_part(BdrvChild *child, | ||
320 | * bdrv_co_do_zero_pwritev() does aligning by itself, so, we do | ||
321 | * alignment only if there is no ZERO flag. | ||
322 | */ | ||
323 | - ret = bdrv_pad_request(bs, &qiov, &qiov_offset, &offset, &bytes, &pad, | ||
324 | - &padded, &flags); | ||
325 | + ret = bdrv_pad_request(bs, &qiov, &qiov_offset, &offset, &bytes, true, | ||
326 | + &pad, &padded, &flags); | ||
327 | if (ret < 0) { | ||
328 | return ret; | ||
329 | } | ||
330 | @@ -XXX,XX +XXX,XX @@ int coroutine_fn bdrv_co_pwritev_part(BdrvChild *child, | ||
331 | ret = bdrv_aligned_pwritev(child, &req, offset, bytes, align, | ||
332 | qiov, qiov_offset, flags); | ||
333 | |||
334 | - bdrv_padding_destroy(&pad); | ||
335 | + bdrv_padding_finalize(&pad); | ||
336 | |||
337 | out: | ||
338 | tracked_request_end(&req); | ||
339 | -- | 84 | -- |
340 | 2.40.1 | 85 | 2.26.2 |
86 | diff view generated by jsdifflib |
New patch | |||
---|---|---|---|
1 | From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com> | ||
1 | 2 | ||
3 | These cases are fixed by previous patches around block_status and | ||
4 | is_allocated. | ||
5 | |||
6 | Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com> | ||
7 | Reviewed-by: Eric Blake <eblake@redhat.com> | ||
8 | Reviewed-by: Alberto Garcia <berto@igalia.com> | ||
9 | Message-id: 20200924194003.22080-6-vsementsov@virtuozzo.com | ||
10 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
11 | --- | ||
12 | tests/qemu-iotests/274 | 20 +++++++++++ | ||
13 | tests/qemu-iotests/274.out | 68 ++++++++++++++++++++++++++++++++++++++ | ||
14 | 2 files changed, 88 insertions(+) | ||
15 | |||
16 | diff --git a/tests/qemu-iotests/274 b/tests/qemu-iotests/274 | ||
17 | index XXXXXXX..XXXXXXX 100755 | ||
18 | --- a/tests/qemu-iotests/274 | ||
19 | +++ b/tests/qemu-iotests/274 | ||
20 | @@ -XXX,XX +XXX,XX @@ with iotests.FilePath('base') as base, \ | ||
21 | iotests.qemu_io_log('-c', 'read -P 1 0 %d' % size_short, mid) | ||
22 | iotests.qemu_io_log('-c', 'read -P 0 %d %d' % (size_short, size_diff), mid) | ||
23 | |||
24 | + iotests.log('=== Testing qemu-img commit (top -> base) ===') | ||
25 | + | ||
26 | + create_chain() | ||
27 | + iotests.qemu_img_log('commit', '-b', base, top) | ||
28 | + iotests.img_info_log(base) | ||
29 | + iotests.qemu_io_log('-c', 'read -P 1 0 %d' % size_short, base) | ||
30 | + iotests.qemu_io_log('-c', 'read -P 0 %d %d' % (size_short, size_diff), base) | ||
31 | + | ||
32 | + iotests.log('=== Testing QMP active commit (top -> base) ===') | ||
33 | + | ||
34 | + create_chain() | ||
35 | + with create_vm() as vm: | ||
36 | + vm.launch() | ||
37 | + vm.qmp_log('block-commit', device='top', base_node='base', | ||
38 | + job_id='job0', auto_dismiss=False) | ||
39 | + vm.run_job('job0', wait=5) | ||
40 | + | ||
41 | + iotests.img_info_log(mid) | ||
42 | + iotests.qemu_io_log('-c', 'read -P 1 0 %d' % size_short, base) | ||
43 | + iotests.qemu_io_log('-c', 'read -P 0 %d %d' % (size_short, size_diff), base) | ||
44 | |||
45 | iotests.log('== Resize tests ==') | ||
46 | |||
47 | diff --git a/tests/qemu-iotests/274.out b/tests/qemu-iotests/274.out | ||
48 | index XXXXXXX..XXXXXXX 100644 | ||
49 | --- a/tests/qemu-iotests/274.out | ||
50 | +++ b/tests/qemu-iotests/274.out | ||
51 | @@ -XXX,XX +XXX,XX @@ read 1048576/1048576 bytes at offset 0 | ||
52 | read 1048576/1048576 bytes at offset 1048576 | ||
53 | 1 MiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | ||
54 | |||
55 | +=== Testing qemu-img commit (top -> base) === | ||
56 | +Formatting 'TEST_DIR/PID-base', fmt=qcow2 cluster_size=65536 extended_l2=off compression_type=zlib size=2097152 lazy_refcounts=off refcount_bits=16 | ||
57 | + | ||
58 | +Formatting 'TEST_DIR/PID-mid', fmt=qcow2 cluster_size=65536 extended_l2=off compression_type=zlib size=1048576 backing_file=TEST_DIR/PID-base backing_fmt=qcow2 lazy_refcounts=off refcount_bits=16 | ||
59 | + | ||
60 | +Formatting 'TEST_DIR/PID-top', fmt=qcow2 cluster_size=65536 extended_l2=off compression_type=zlib size=2097152 backing_file=TEST_DIR/PID-mid backing_fmt=qcow2 lazy_refcounts=off refcount_bits=16 | ||
61 | + | ||
62 | +wrote 2097152/2097152 bytes at offset 0 | ||
63 | +2 MiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | ||
64 | + | ||
65 | +Image committed. | ||
66 | + | ||
67 | +image: TEST_IMG | ||
68 | +file format: IMGFMT | ||
69 | +virtual size: 2 MiB (2097152 bytes) | ||
70 | +cluster_size: 65536 | ||
71 | +Format specific information: | ||
72 | + compat: 1.1 | ||
73 | + compression type: zlib | ||
74 | + lazy refcounts: false | ||
75 | + refcount bits: 16 | ||
76 | + corrupt: false | ||
77 | + extended l2: false | ||
78 | + | ||
79 | +read 1048576/1048576 bytes at offset 0 | ||
80 | +1 MiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | ||
81 | + | ||
82 | +read 1048576/1048576 bytes at offset 1048576 | ||
83 | +1 MiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | ||
84 | + | ||
85 | +=== Testing QMP active commit (top -> base) === | ||
86 | +Formatting 'TEST_DIR/PID-base', fmt=qcow2 cluster_size=65536 extended_l2=off compression_type=zlib size=2097152 lazy_refcounts=off refcount_bits=16 | ||
87 | + | ||
88 | +Formatting 'TEST_DIR/PID-mid', fmt=qcow2 cluster_size=65536 extended_l2=off compression_type=zlib size=1048576 backing_file=TEST_DIR/PID-base backing_fmt=qcow2 lazy_refcounts=off refcount_bits=16 | ||
89 | + | ||
90 | +Formatting 'TEST_DIR/PID-top', fmt=qcow2 cluster_size=65536 extended_l2=off compression_type=zlib size=2097152 backing_file=TEST_DIR/PID-mid backing_fmt=qcow2 lazy_refcounts=off refcount_bits=16 | ||
91 | + | ||
92 | +wrote 2097152/2097152 bytes at offset 0 | ||
93 | +2 MiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | ||
94 | + | ||
95 | +{"execute": "block-commit", "arguments": {"auto-dismiss": false, "base-node": "base", "device": "top", "job-id": "job0"}} | ||
96 | +{"return": {}} | ||
97 | +{"execute": "job-complete", "arguments": {"id": "job0"}} | ||
98 | +{"return": {}} | ||
99 | +{"data": {"device": "job0", "len": 1048576, "offset": 1048576, "speed": 0, "type": "commit"}, "event": "BLOCK_JOB_READY", "timestamp": {"microseconds": "USECS", "seconds": "SECS"}} | ||
100 | +{"data": {"device": "job0", "len": 1048576, "offset": 1048576, "speed": 0, "type": "commit"}, "event": "BLOCK_JOB_COMPLETED", "timestamp": {"microseconds": "USECS", "seconds": "SECS"}} | ||
101 | +{"execute": "job-dismiss", "arguments": {"id": "job0"}} | ||
102 | +{"return": {}} | ||
103 | +image: TEST_IMG | ||
104 | +file format: IMGFMT | ||
105 | +virtual size: 1 MiB (1048576 bytes) | ||
106 | +cluster_size: 65536 | ||
107 | +backing file: TEST_DIR/PID-base | ||
108 | +backing file format: IMGFMT | ||
109 | +Format specific information: | ||
110 | + compat: 1.1 | ||
111 | + compression type: zlib | ||
112 | + lazy refcounts: false | ||
113 | + refcount bits: 16 | ||
114 | + corrupt: false | ||
115 | + extended l2: false | ||
116 | + | ||
117 | +read 1048576/1048576 bytes at offset 0 | ||
118 | +1 MiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | ||
119 | + | ||
120 | +read 1048576/1048576 bytes at offset 1048576 | ||
121 | +1 MiB, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) | ||
122 | + | ||
123 | == Resize tests == | ||
124 | === preallocation=off === | ||
125 | Formatting 'TEST_DIR/PID-base', fmt=qcow2 cluster_size=65536 extended_l2=off compression_type=zlib size=6442450944 lazy_refcounts=off refcount_bits=16 | ||
126 | -- | ||
127 | 2.26.2 | ||
128 | diff view generated by jsdifflib |