1 | The following changes since commit 248b23735645f7cbb503d9be6f5bf825f2a603ab: | 1 | The following changes since commit 56f9e46b841c7be478ca038d8d4085d776ab4b0d: |
---|---|---|---|
2 | 2 | ||
3 | Update version for v2.10.0-rc4 release (2017-08-24 17:34:26 +0100) | 3 | Merge remote-tracking branch 'remotes/armbru/tags/pull-qapi-2017-02-20' into staging (2017-02-20 17:42:47 +0000) |
4 | 4 | ||
5 | are available in the git repository at: | 5 | are available in the git repository at: |
6 | 6 | ||
7 | git://github.com/stefanha/qemu.git tags/block-pull-request | 7 | git://github.com/stefanha/qemu.git tags/block-pull-request |
8 | 8 | ||
9 | for you to fetch changes up to 3e4c705212abfe8c9882a00beb2d1466a8a53cec: | 9 | for you to fetch changes up to a7b91d35bab97a2d3e779d0c64c9b837b52a6cf7: |
10 | 10 | ||
11 | qcow2: allocate cluster_cache/cluster_data on demand (2017-08-30 18:02:10 +0100) | 11 | coroutine-lock: make CoRwlock thread-safe and fair (2017-02-21 11:39:40 +0000) |
12 | |||
13 | ---------------------------------------------------------------- | ||
14 | Pull request | ||
15 | |||
16 | v2: | ||
17 | * Rebased to resolve scsi conflicts | ||
12 | 18 | ||
13 | ---------------------------------------------------------------- | 19 | ---------------------------------------------------------------- |
14 | 20 | ||
15 | ---------------------------------------------------------------- | 21 | Paolo Bonzini (24): |
22 | block: move AioContext, QEMUTimer, main-loop to libqemuutil | ||
23 | aio: introduce aio_co_schedule and aio_co_wake | ||
24 | block-backend: allow blk_prw from coroutine context | ||
25 | test-thread-pool: use generic AioContext infrastructure | ||
26 | io: add methods to set I/O handlers on AioContext | ||
27 | io: make qio_channel_yield aware of AioContexts | ||
28 | nbd: convert to use qio_channel_yield | ||
29 | coroutine-lock: reschedule coroutine on the AioContext it was running | ||
30 | on | ||
31 | blkdebug: reschedule coroutine on the AioContext it is running on | ||
32 | qed: introduce qed_aio_start_io and qed_aio_next_io_cb | ||
33 | aio: push aio_context_acquire/release down to dispatching | ||
34 | block: explicitly acquire aiocontext in timers that need it | ||
35 | block: explicitly acquire aiocontext in callbacks that need it | ||
36 | block: explicitly acquire aiocontext in bottom halves that need it | ||
37 | block: explicitly acquire aiocontext in aio callbacks that need it | ||
38 | aio-posix: partially inline aio_dispatch into aio_poll | ||
39 | async: remove unnecessary inc/dec pairs | ||
40 | block: document fields protected by AioContext lock | ||
41 | coroutine-lock: make CoMutex thread-safe | ||
42 | coroutine-lock: add limited spinning to CoMutex | ||
43 | test-aio-multithread: add performance comparison with thread-based | ||
44 | mutexes | ||
45 | coroutine-lock: place CoMutex before CoQueue in header | ||
46 | coroutine-lock: add mutex argument to CoQueue APIs | ||
47 | coroutine-lock: make CoRwlock thread-safe and fair | ||
16 | 48 | ||
17 | Alberto Garcia (8): | 49 | Makefile.objs | 4 - |
18 | throttle: Fix wrong variable name in the header documentation | 50 | stubs/Makefile.objs | 1 + |
19 | throttle: Update the throttle_fix_bucket() documentation | 51 | tests/Makefile.include | 19 +- |
20 | throttle: Make throttle_is_valid() a bit less verbose | 52 | util/Makefile.objs | 6 +- |
21 | throttle: Remove throttle_fix_bucket() / throttle_unfix_bucket() | 53 | block/nbd-client.h | 2 +- |
22 | throttle: Make LeakyBucket.avg and LeakyBucket.max integer types | 54 | block/qed.h | 3 + |
23 | throttle: Make burst_length 64bit and add range checks | 55 | include/block/aio.h | 38 ++- |
24 | throttle: Test the valid range of config values | 56 | include/block/block_int.h | 64 +++-- |
25 | misc: Remove unused Error variables | 57 | include/io/channel.h | 72 +++++- |
26 | 58 | include/qemu/coroutine.h | 84 ++++--- | |
27 | Dan Aloni (1): | 59 | include/qemu/coroutine_int.h | 11 +- |
28 | nvme: Fix get/set number of queues feature, again | 60 | include/sysemu/block-backend.h | 14 +- |
29 | 61 | tests/iothread.h | 25 ++ | |
30 | Eduardo Habkost (1): | 62 | block/backup.c | 2 +- |
31 | oslib-posix: Print errors before aborting on qemu_alloc_stack() | 63 | block/blkdebug.c | 9 +- |
32 | 64 | block/blkreplay.c | 2 +- | |
33 | Fred Rolland (1): | 65 | block/block-backend.c | 13 +- |
34 | qemu-doc: Add UUID support in initiator name | 66 | block/curl.c | 44 +++- |
35 | 67 | block/gluster.c | 9 +- | |
36 | Stefan Hajnoczi (4): | 68 | block/io.c | 42 +--- |
37 | scripts: add argparse module for Python 2.6 compatibility | 69 | block/iscsi.c | 15 +- |
38 | docker.py: Python 2.6 argparse compatibility | 70 | block/linux-aio.c | 10 +- |
39 | tests: migration/guestperf Python 2.6 argparse compatibility | 71 | block/mirror.c | 12 +- |
40 | qcow2: allocate cluster_cache/cluster_data on demand | 72 | block/nbd-client.c | 119 +++++---- |
41 | 73 | block/nfs.c | 9 +- | |
42 | include/qemu/throttle.h | 8 +- | 74 | block/qcow2-cluster.c | 4 +- |
43 | block/qcow.c | 12 +- | 75 | block/qed-cluster.c | 2 + |
44 | block/qcow2-cluster.c | 17 + | 76 | block/qed-table.c | 12 +- |
45 | block/qcow2.c | 20 +- | 77 | block/qed.c | 58 +++-- |
46 | dump.c | 4 +- | 78 | block/sheepdog.c | 31 +-- |
47 | hw/block/nvme.c | 4 +- | 79 | block/ssh.c | 29 +-- |
48 | tests/test-throttle.c | 80 +- | 80 | block/throttle-groups.c | 4 +- |
49 | util/oslib-posix.c | 2 + | 81 | block/win32-aio.c | 9 +- |
50 | util/throttle.c | 86 +- | 82 | dma-helpers.c | 2 + |
51 | COPYING.PYTHON | 270 ++++ | 83 | hw/9pfs/9p.c | 2 +- |
52 | qemu-doc.texi | 5 +- | 84 | hw/block/virtio-blk.c | 19 +- |
53 | scripts/argparse.py | 2406 ++++++++++++++++++++++++++++++++++++ | 85 | hw/scsi/scsi-bus.c | 2 + |
54 | tests/docker/docker.py | 4 +- | 86 | hw/scsi/scsi-disk.c | 15 ++ |
55 | tests/migration/guestperf/shell.py | 8 +- | 87 | hw/scsi/scsi-generic.c | 20 +- |
56 | 14 files changed, 2831 insertions(+), 95 deletions(-) | 88 | hw/scsi/virtio-scsi.c | 7 + |
57 | create mode 100644 COPYING.PYTHON | 89 | io/channel-command.c | 13 + |
58 | create mode 100644 scripts/argparse.py | 90 | io/channel-file.c | 11 + |
91 | io/channel-socket.c | 16 +- | ||
92 | io/channel-tls.c | 12 + | ||
93 | io/channel-watch.c | 6 + | ||
94 | io/channel.c | 97 ++++++-- | ||
95 | nbd/client.c | 2 +- | ||
96 | nbd/common.c | 9 +- | ||
97 | nbd/server.c | 94 +++----- | ||
98 | stubs/linux-aio.c | 32 +++ | ||
99 | stubs/set-fd-handler.c | 11 - | ||
100 | tests/iothread.c | 91 +++++++ | ||
101 | tests/test-aio-multithread.c | 463 ++++++++++++++++++++++++++++++++++++ | ||
102 | tests/test-thread-pool.c | 12 +- | ||
103 | aio-posix.c => util/aio-posix.c | 62 ++--- | ||
104 | aio-win32.c => util/aio-win32.c | 30 +-- | ||
105 | util/aiocb.c | 55 +++++ | ||
106 | async.c => util/async.c | 84 ++++++- | ||
107 | iohandler.c => util/iohandler.c | 0 | ||
108 | main-loop.c => util/main-loop.c | 0 | ||
109 | util/qemu-coroutine-lock.c | 254 ++++++++++++++++++-- | ||
110 | util/qemu-coroutine-sleep.c | 2 +- | ||
111 | util/qemu-coroutine.c | 8 + | ||
112 | qemu-timer.c => util/qemu-timer.c | 0 | ||
113 | thread-pool.c => util/thread-pool.c | 8 +- | ||
114 | trace-events | 11 - | ||
115 | util/trace-events | 17 +- | ||
116 | 67 files changed, 1712 insertions(+), 533 deletions(-) | ||
117 | create mode 100644 tests/iothread.h | ||
118 | create mode 100644 stubs/linux-aio.c | ||
119 | create mode 100644 tests/iothread.c | ||
120 | create mode 100644 tests/test-aio-multithread.c | ||
121 | rename aio-posix.c => util/aio-posix.c (94%) | ||
122 | rename aio-win32.c => util/aio-win32.c (95%) | ||
123 | create mode 100644 util/aiocb.c | ||
124 | rename async.c => util/async.c (82%) | ||
125 | rename iohandler.c => util/iohandler.c (100%) | ||
126 | rename main-loop.c => util/main-loop.c (100%) | ||
127 | rename qemu-timer.c => util/qemu-timer.c (100%) | ||
128 | rename thread-pool.c => util/thread-pool.c (97%) | ||
59 | 129 | ||
60 | -- | 130 | -- |
61 | 2.13.5 | 131 | 2.9.3 |
62 | 132 | ||
63 | 133 | diff view generated by jsdifflib |
New patch | |||
---|---|---|---|
1 | 1 | From: Paolo Bonzini <pbonzini@redhat.com> | |
2 | |||
3 | AioContext is fairly self contained, the only dependency is QEMUTimer but | ||
4 | that in turn doesn't need anything else. So move them out of block-obj-y | ||
5 | to avoid introducing a dependency from io/ to block-obj-y. | ||
6 | |||
7 | main-loop and its dependency iohandler also need to be moved, because | ||
8 | later in this series io/ will call iohandler_get_aio_context. | ||
9 | |||
10 | [Changed copyright "the QEMU team" to "other QEMU contributors" as | ||
11 | suggested by Daniel Berrange and agreed by Paolo. | ||
12 | --Stefan] | ||
13 | |||
14 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> | ||
15 | Reviewed-by: Fam Zheng <famz@redhat.com> | ||
16 | Message-id: 20170213135235.12274-2-pbonzini@redhat.com | ||
17 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
18 | --- | ||
19 | Makefile.objs | 4 --- | ||
20 | stubs/Makefile.objs | 1 + | ||
21 | tests/Makefile.include | 11 ++++---- | ||
22 | util/Makefile.objs | 6 +++- | ||
23 | block/io.c | 29 ------------------- | ||
24 | stubs/linux-aio.c | 32 +++++++++++++++++++++ | ||
25 | stubs/set-fd-handler.c | 11 -------- | ||
26 | aio-posix.c => util/aio-posix.c | 2 +- | ||
27 | aio-win32.c => util/aio-win32.c | 0 | ||
28 | util/aiocb.c | 55 +++++++++++++++++++++++++++++++++++++ | ||
29 | async.c => util/async.c | 3 +- | ||
30 | iohandler.c => util/iohandler.c | 0 | ||
31 | main-loop.c => util/main-loop.c | 0 | ||
32 | qemu-timer.c => util/qemu-timer.c | 0 | ||
33 | thread-pool.c => util/thread-pool.c | 2 +- | ||
34 | trace-events | 11 -------- | ||
35 | util/trace-events | 11 ++++++++ | ||
36 | 17 files changed, 114 insertions(+), 64 deletions(-) | ||
37 | create mode 100644 stubs/linux-aio.c | ||
38 | rename aio-posix.c => util/aio-posix.c (99%) | ||
39 | rename aio-win32.c => util/aio-win32.c (100%) | ||
40 | create mode 100644 util/aiocb.c | ||
41 | rename async.c => util/async.c (99%) | ||
42 | rename iohandler.c => util/iohandler.c (100%) | ||
43 | rename main-loop.c => util/main-loop.c (100%) | ||
44 | rename qemu-timer.c => util/qemu-timer.c (100%) | ||
45 | rename thread-pool.c => util/thread-pool.c (99%) | ||
46 | |||
47 | diff --git a/Makefile.objs b/Makefile.objs | ||
48 | index XXXXXXX..XXXXXXX 100644 | ||
49 | --- a/Makefile.objs | ||
50 | +++ b/Makefile.objs | ||
51 | @@ -XXX,XX +XXX,XX @@ chardev-obj-y = chardev/ | ||
52 | ####################################################################### | ||
53 | # block-obj-y is code used by both qemu system emulation and qemu-img | ||
54 | |||
55 | -block-obj-y = async.o thread-pool.o | ||
56 | block-obj-y += nbd/ | ||
57 | block-obj-y += block.o blockjob.o | ||
58 | -block-obj-y += main-loop.o iohandler.o qemu-timer.o | ||
59 | -block-obj-$(CONFIG_POSIX) += aio-posix.o | ||
60 | -block-obj-$(CONFIG_WIN32) += aio-win32.o | ||
61 | block-obj-y += block/ | ||
62 | block-obj-y += qemu-io-cmds.o | ||
63 | block-obj-$(CONFIG_REPLICATION) += replication.o | ||
64 | diff --git a/stubs/Makefile.objs b/stubs/Makefile.objs | ||
65 | index XXXXXXX..XXXXXXX 100644 | ||
66 | --- a/stubs/Makefile.objs | ||
67 | +++ b/stubs/Makefile.objs | ||
68 | @@ -XXX,XX +XXX,XX @@ stub-obj-y += get-vm-name.o | ||
69 | stub-obj-y += iothread.o | ||
70 | stub-obj-y += iothread-lock.o | ||
71 | stub-obj-y += is-daemonized.o | ||
72 | +stub-obj-$(CONFIG_LINUX_AIO) += linux-aio.o | ||
73 | stub-obj-y += machine-init-done.o | ||
74 | stub-obj-y += migr-blocker.o | ||
75 | stub-obj-y += monitor.o | ||
76 | diff --git a/tests/Makefile.include b/tests/Makefile.include | ||
77 | index XXXXXXX..XXXXXXX 100644 | ||
78 | --- a/tests/Makefile.include | ||
79 | +++ b/tests/Makefile.include | ||
80 | @@ -XXX,XX +XXX,XX @@ check-unit-y += tests/test-visitor-serialization$(EXESUF) | ||
81 | check-unit-y += tests/test-iov$(EXESUF) | ||
82 | gcov-files-test-iov-y = util/iov.c | ||
83 | check-unit-y += tests/test-aio$(EXESUF) | ||
84 | +gcov-files-test-aio-y = util/async.c util/qemu-timer.o | ||
85 | +gcov-files-test-aio-$(CONFIG_WIN32) += util/aio-win32.c | ||
86 | +gcov-files-test-aio-$(CONFIG_POSIX) += util/aio-posix.c | ||
87 | check-unit-y += tests/test-throttle$(EXESUF) | ||
88 | gcov-files-test-aio-$(CONFIG_WIN32) = aio-win32.c | ||
89 | gcov-files-test-aio-$(CONFIG_POSIX) = aio-posix.c | ||
90 | @@ -XXX,XX +XXX,XX @@ tests/check-qjson$(EXESUF): tests/check-qjson.o $(test-util-obj-y) | ||
91 | tests/check-qom-interface$(EXESUF): tests/check-qom-interface.o $(test-qom-obj-y) | ||
92 | tests/check-qom-proplist$(EXESUF): tests/check-qom-proplist.o $(test-qom-obj-y) | ||
93 | |||
94 | -tests/test-char$(EXESUF): tests/test-char.o qemu-timer.o \ | ||
95 | - $(test-util-obj-y) $(qtest-obj-y) $(test-block-obj-y) $(chardev-obj-y) | ||
96 | +tests/test-char$(EXESUF): tests/test-char.o $(test-util-obj-y) $(qtest-obj-y) $(test-io-obj-y) $(chardev-obj-y) | ||
97 | tests/test-coroutine$(EXESUF): tests/test-coroutine.o $(test-block-obj-y) | ||
98 | tests/test-aio$(EXESUF): tests/test-aio.o $(test-block-obj-y) | ||
99 | tests/test-throttle$(EXESUF): tests/test-throttle.o $(test-block-obj-y) | ||
100 | @@ -XXX,XX +XXX,XX @@ tests/test-vmstate$(EXESUF): tests/test-vmstate.o \ | ||
101 | migration/vmstate.o migration/qemu-file.o \ | ||
102 | migration/qemu-file-channel.o migration/qjson.o \ | ||
103 | $(test-io-obj-y) | ||
104 | -tests/test-timed-average$(EXESUF): tests/test-timed-average.o qemu-timer.o \ | ||
105 | - $(test-util-obj-y) | ||
106 | +tests/test-timed-average$(EXESUF): tests/test-timed-average.o $(test-util-obj-y) | ||
107 | tests/test-base64$(EXESUF): tests/test-base64.o \ | ||
108 | libqemuutil.a libqemustub.a | ||
109 | tests/ptimer-test$(EXESUF): tests/ptimer-test.o tests/ptimer-test-stubs.o hw/core/ptimer.o libqemustub.a | ||
110 | @@ -XXX,XX +XXX,XX @@ tests/usb-hcd-ehci-test$(EXESUF): tests/usb-hcd-ehci-test.o $(libqos-usb-obj-y) | ||
111 | tests/usb-hcd-xhci-test$(EXESUF): tests/usb-hcd-xhci-test.o $(libqos-usb-obj-y) | ||
112 | tests/pc-cpu-test$(EXESUF): tests/pc-cpu-test.o | ||
113 | tests/postcopy-test$(EXESUF): tests/postcopy-test.o | ||
114 | -tests/vhost-user-test$(EXESUF): tests/vhost-user-test.o qemu-timer.o \ | ||
115 | +tests/vhost-user-test$(EXESUF): tests/vhost-user-test.o $(test-util-obj-y) \ | ||
116 | $(qtest-obj-y) $(test-io-obj-y) $(libqos-virtio-obj-y) $(libqos-pc-obj-y) \ | ||
117 | $(chardev-obj-y) | ||
118 | tests/qemu-iotests/socket_scm_helper$(EXESUF): tests/qemu-iotests/socket_scm_helper.o | ||
119 | diff --git a/util/Makefile.objs b/util/Makefile.objs | ||
120 | index XXXXXXX..XXXXXXX 100644 | ||
121 | --- a/util/Makefile.objs | ||
122 | +++ b/util/Makefile.objs | ||
123 | @@ -XXX,XX +XXX,XX @@ | ||
124 | util-obj-y = osdep.o cutils.o unicode.o qemu-timer-common.o | ||
125 | util-obj-y += bufferiszero.o | ||
126 | util-obj-y += lockcnt.o | ||
127 | +util-obj-y += aiocb.o async.o thread-pool.o qemu-timer.o | ||
128 | +util-obj-y += main-loop.o iohandler.o | ||
129 | +util-obj-$(CONFIG_POSIX) += aio-posix.o | ||
130 | util-obj-$(CONFIG_POSIX) += compatfd.o | ||
131 | util-obj-$(CONFIG_POSIX) += event_notifier-posix.o | ||
132 | util-obj-$(CONFIG_POSIX) += mmap-alloc.o | ||
133 | util-obj-$(CONFIG_POSIX) += oslib-posix.o | ||
134 | util-obj-$(CONFIG_POSIX) += qemu-openpty.o | ||
135 | util-obj-$(CONFIG_POSIX) += qemu-thread-posix.o | ||
136 | -util-obj-$(CONFIG_WIN32) += event_notifier-win32.o | ||
137 | util-obj-$(CONFIG_POSIX) += memfd.o | ||
138 | +util-obj-$(CONFIG_WIN32) += aio-win32.o | ||
139 | +util-obj-$(CONFIG_WIN32) += event_notifier-win32.o | ||
140 | util-obj-$(CONFIG_WIN32) += oslib-win32.o | ||
141 | util-obj-$(CONFIG_WIN32) += qemu-thread-win32.o | ||
142 | util-obj-y += envlist.o path.o module.o | ||
143 | diff --git a/block/io.c b/block/io.c | ||
144 | index XXXXXXX..XXXXXXX 100644 | ||
145 | --- a/block/io.c | ||
146 | +++ b/block/io.c | ||
147 | @@ -XXX,XX +XXX,XX @@ BlockAIOCB *bdrv_aio_flush(BlockDriverState *bs, | ||
148 | return &acb->common; | ||
149 | } | ||
150 | |||
151 | -void *qemu_aio_get(const AIOCBInfo *aiocb_info, BlockDriverState *bs, | ||
152 | - BlockCompletionFunc *cb, void *opaque) | ||
153 | -{ | ||
154 | - BlockAIOCB *acb; | ||
155 | - | ||
156 | - acb = g_malloc(aiocb_info->aiocb_size); | ||
157 | - acb->aiocb_info = aiocb_info; | ||
158 | - acb->bs = bs; | ||
159 | - acb->cb = cb; | ||
160 | - acb->opaque = opaque; | ||
161 | - acb->refcnt = 1; | ||
162 | - return acb; | ||
163 | -} | ||
164 | - | ||
165 | -void qemu_aio_ref(void *p) | ||
166 | -{ | ||
167 | - BlockAIOCB *acb = p; | ||
168 | - acb->refcnt++; | ||
169 | -} | ||
170 | - | ||
171 | -void qemu_aio_unref(void *p) | ||
172 | -{ | ||
173 | - BlockAIOCB *acb = p; | ||
174 | - assert(acb->refcnt > 0); | ||
175 | - if (--acb->refcnt == 0) { | ||
176 | - g_free(acb); | ||
177 | - } | ||
178 | -} | ||
179 | - | ||
180 | /**************************************************************/ | ||
181 | /* Coroutine block device emulation */ | ||
182 | |||
183 | diff --git a/stubs/linux-aio.c b/stubs/linux-aio.c | ||
184 | new file mode 100644 | ||
185 | index XXXXXXX..XXXXXXX | ||
186 | --- /dev/null | ||
187 | +++ b/stubs/linux-aio.c | ||
188 | @@ -XXX,XX +XXX,XX @@ | ||
189 | +/* | ||
190 | + * Linux native AIO support. | ||
191 | + * | ||
192 | + * Copyright (C) 2009 IBM, Corp. | ||
193 | + * Copyright (C) 2009 Red Hat, Inc. | ||
194 | + * | ||
195 | + * This work is licensed under the terms of the GNU GPL, version 2 or later. | ||
196 | + * See the COPYING file in the top-level directory. | ||
197 | + */ | ||
198 | +#include "qemu/osdep.h" | ||
199 | +#include "block/aio.h" | ||
200 | +#include "block/raw-aio.h" | ||
201 | + | ||
202 | +void laio_detach_aio_context(LinuxAioState *s, AioContext *old_context) | ||
203 | +{ | ||
204 | + abort(); | ||
205 | +} | ||
206 | + | ||
207 | +void laio_attach_aio_context(LinuxAioState *s, AioContext *new_context) | ||
208 | +{ | ||
209 | + abort(); | ||
210 | +} | ||
211 | + | ||
212 | +LinuxAioState *laio_init(void) | ||
213 | +{ | ||
214 | + abort(); | ||
215 | +} | ||
216 | + | ||
217 | +void laio_cleanup(LinuxAioState *s) | ||
218 | +{ | ||
219 | + abort(); | ||
220 | +} | ||
221 | diff --git a/stubs/set-fd-handler.c b/stubs/set-fd-handler.c | ||
222 | index XXXXXXX..XXXXXXX 100644 | ||
223 | --- a/stubs/set-fd-handler.c | ||
224 | +++ b/stubs/set-fd-handler.c | ||
225 | @@ -XXX,XX +XXX,XX @@ void qemu_set_fd_handler(int fd, | ||
226 | { | ||
227 | abort(); | ||
228 | } | ||
229 | - | ||
230 | -void aio_set_fd_handler(AioContext *ctx, | ||
231 | - int fd, | ||
232 | - bool is_external, | ||
233 | - IOHandler *io_read, | ||
234 | - IOHandler *io_write, | ||
235 | - AioPollFn *io_poll, | ||
236 | - void *opaque) | ||
237 | -{ | ||
238 | - abort(); | ||
239 | -} | ||
240 | diff --git a/aio-posix.c b/util/aio-posix.c | ||
241 | similarity index 99% | ||
242 | rename from aio-posix.c | ||
243 | rename to util/aio-posix.c | ||
244 | index XXXXXXX..XXXXXXX 100644 | ||
245 | --- a/aio-posix.c | ||
246 | +++ b/util/aio-posix.c | ||
247 | @@ -XXX,XX +XXX,XX @@ | ||
248 | #include "qemu/rcu_queue.h" | ||
249 | #include "qemu/sockets.h" | ||
250 | #include "qemu/cutils.h" | ||
251 | -#include "trace-root.h" | ||
252 | +#include "trace.h" | ||
253 | #ifdef CONFIG_EPOLL_CREATE1 | ||
254 | #include <sys/epoll.h> | ||
255 | #endif | ||
256 | diff --git a/aio-win32.c b/util/aio-win32.c | ||
257 | similarity index 100% | ||
258 | rename from aio-win32.c | ||
259 | rename to util/aio-win32.c | ||
260 | diff --git a/util/aiocb.c b/util/aiocb.c | ||
261 | new file mode 100644 | ||
262 | index XXXXXXX..XXXXXXX | ||
263 | --- /dev/null | ||
264 | +++ b/util/aiocb.c | ||
265 | @@ -XXX,XX +XXX,XX @@ | ||
266 | +/* | ||
267 | + * BlockAIOCB allocation | ||
268 | + * | ||
269 | + * Copyright (c) 2003-2017 Fabrice Bellard and other QEMU contributors | ||
270 | + * | ||
271 | + * Permission is hereby granted, free of charge, to any person obtaining a copy | ||
272 | + * of this software and associated documentation files (the "Software"), to deal | ||
273 | + * in the Software without restriction, including without limitation the rights | ||
274 | + * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | ||
275 | + * copies of the Software, and to permit persons to whom the Software is | ||
276 | + * furnished to do so, subject to the following conditions: | ||
277 | + * | ||
278 | + * The above copyright notice and this permission notice shall be included in | ||
279 | + * all copies or substantial portions of the Software. | ||
280 | + * | ||
281 | + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
282 | + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
283 | + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | ||
284 | + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||
285 | + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||
286 | + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN | ||
287 | + * THE SOFTWARE. | ||
288 | + */ | ||
289 | + | ||
290 | +#include "qemu/osdep.h" | ||
291 | +#include "block/aio.h" | ||
292 | + | ||
293 | +void *qemu_aio_get(const AIOCBInfo *aiocb_info, BlockDriverState *bs, | ||
294 | + BlockCompletionFunc *cb, void *opaque) | ||
295 | +{ | ||
296 | + BlockAIOCB *acb; | ||
297 | + | ||
298 | + acb = g_malloc(aiocb_info->aiocb_size); | ||
299 | + acb->aiocb_info = aiocb_info; | ||
300 | + acb->bs = bs; | ||
301 | + acb->cb = cb; | ||
302 | + acb->opaque = opaque; | ||
303 | + acb->refcnt = 1; | ||
304 | + return acb; | ||
305 | +} | ||
306 | + | ||
307 | +void qemu_aio_ref(void *p) | ||
308 | +{ | ||
309 | + BlockAIOCB *acb = p; | ||
310 | + acb->refcnt++; | ||
311 | +} | ||
312 | + | ||
313 | +void qemu_aio_unref(void *p) | ||
314 | +{ | ||
315 | + BlockAIOCB *acb = p; | ||
316 | + assert(acb->refcnt > 0); | ||
317 | + if (--acb->refcnt == 0) { | ||
318 | + g_free(acb); | ||
319 | + } | ||
320 | +} | ||
321 | diff --git a/async.c b/util/async.c | ||
322 | similarity index 99% | ||
323 | rename from async.c | ||
324 | rename to util/async.c | ||
325 | index XXXXXXX..XXXXXXX 100644 | ||
326 | --- a/async.c | ||
327 | +++ b/util/async.c | ||
328 | @@ -XXX,XX +XXX,XX @@ | ||
329 | /* | ||
330 | - * QEMU System Emulator | ||
331 | + * Data plane event loop | ||
332 | * | ||
333 | * Copyright (c) 2003-2008 Fabrice Bellard | ||
334 | + * Copyright (c) 2009-2017 QEMU contributors | ||
335 | * | ||
336 | * Permission is hereby granted, free of charge, to any person obtaining a copy | ||
337 | * of this software and associated documentation files (the "Software"), to deal | ||
338 | diff --git a/iohandler.c b/util/iohandler.c | ||
339 | similarity index 100% | ||
340 | rename from iohandler.c | ||
341 | rename to util/iohandler.c | ||
342 | diff --git a/main-loop.c b/util/main-loop.c | ||
343 | similarity index 100% | ||
344 | rename from main-loop.c | ||
345 | rename to util/main-loop.c | ||
346 | diff --git a/qemu-timer.c b/util/qemu-timer.c | ||
347 | similarity index 100% | ||
348 | rename from qemu-timer.c | ||
349 | rename to util/qemu-timer.c | ||
350 | diff --git a/thread-pool.c b/util/thread-pool.c | ||
351 | similarity index 99% | ||
352 | rename from thread-pool.c | ||
353 | rename to util/thread-pool.c | ||
354 | index XXXXXXX..XXXXXXX 100644 | ||
355 | --- a/thread-pool.c | ||
356 | +++ b/util/thread-pool.c | ||
357 | @@ -XXX,XX +XXX,XX @@ | ||
358 | #include "qemu/queue.h" | ||
359 | #include "qemu/thread.h" | ||
360 | #include "qemu/coroutine.h" | ||
361 | -#include "trace-root.h" | ||
362 | +#include "trace.h" | ||
363 | #include "block/thread-pool.h" | ||
364 | #include "qemu/main-loop.h" | ||
365 | |||
366 | diff --git a/trace-events b/trace-events | ||
367 | index XXXXXXX..XXXXXXX 100644 | ||
368 | --- a/trace-events | ||
369 | +++ b/trace-events | ||
370 | @@ -XXX,XX +XXX,XX @@ | ||
371 | # | ||
372 | # The <format-string> should be a sprintf()-compatible format string. | ||
373 | |||
374 | -# aio-posix.c | ||
375 | -run_poll_handlers_begin(void *ctx, int64_t max_ns) "ctx %p max_ns %"PRId64 | ||
376 | -run_poll_handlers_end(void *ctx, bool progress) "ctx %p progress %d" | ||
377 | -poll_shrink(void *ctx, int64_t old, int64_t new) "ctx %p old %"PRId64" new %"PRId64 | ||
378 | -poll_grow(void *ctx, int64_t old, int64_t new) "ctx %p old %"PRId64" new %"PRId64 | ||
379 | - | ||
380 | -# thread-pool.c | ||
381 | -thread_pool_submit(void *pool, void *req, void *opaque) "pool %p req %p opaque %p" | ||
382 | -thread_pool_complete(void *pool, void *req, void *opaque, int ret) "pool %p req %p opaque %p ret %d" | ||
383 | -thread_pool_cancel(void *req, void *opaque) "req %p opaque %p" | ||
384 | - | ||
385 | # ioport.c | ||
386 | cpu_in(unsigned int addr, char size, unsigned int val) "addr %#x(%c) value %u" | ||
387 | cpu_out(unsigned int addr, char size, unsigned int val) "addr %#x(%c) value %u" | ||
388 | diff --git a/util/trace-events b/util/trace-events | ||
389 | index XXXXXXX..XXXXXXX 100644 | ||
390 | --- a/util/trace-events | ||
391 | +++ b/util/trace-events | ||
392 | @@ -XXX,XX +XXX,XX @@ | ||
393 | # See docs/tracing.txt for syntax documentation. | ||
394 | |||
395 | +# util/aio-posix.c | ||
396 | +run_poll_handlers_begin(void *ctx, int64_t max_ns) "ctx %p max_ns %"PRId64 | ||
397 | +run_poll_handlers_end(void *ctx, bool progress) "ctx %p progress %d" | ||
398 | +poll_shrink(void *ctx, int64_t old, int64_t new) "ctx %p old %"PRId64" new %"PRId64 | ||
399 | +poll_grow(void *ctx, int64_t old, int64_t new) "ctx %p old %"PRId64" new %"PRId64 | ||
400 | + | ||
401 | +# util/thread-pool.c | ||
402 | +thread_pool_submit(void *pool, void *req, void *opaque) "pool %p req %p opaque %p" | ||
403 | +thread_pool_complete(void *pool, void *req, void *opaque, int ret) "pool %p req %p opaque %p ret %d" | ||
404 | +thread_pool_cancel(void *req, void *opaque) "req %p opaque %p" | ||
405 | + | ||
406 | # util/buffer.c | ||
407 | buffer_resize(const char *buf, size_t olen, size_t len) "%s: old %zd, new %zd" | ||
408 | buffer_move_empty(const char *buf, size_t len, const char *from) "%s: %zd bytes from %s" | ||
409 | -- | ||
410 | 2.9.3 | ||
411 | |||
412 | diff view generated by jsdifflib |
1 | The minimum Python version supported by QEMU is 2.6. The argparse | 1 | From: Paolo Bonzini <pbonzini@redhat.com> |
---|---|---|---|
2 | standard library module was only added in Python 2.7. Many scripts | 2 | |
3 | would like to use argparse because it supports command-line | 3 | aio_co_wake provides the infrastructure to start a coroutine on a "home" |
4 | sub-commands. | 4 | AioContext. It will be used by CoMutex and CoQueue, so that coroutines |
5 | 5 | don't jump from one context to another when they go to sleep on a | |
6 | This patch adds argparse. See the top of argparse.py for details. | 6 | mutex or waitqueue. However, it can also be used as a more efficient |
7 | 7 | alternative to one-shot bottom halves, and saves the effort of tracking | |
8 | Suggested-by: Daniel P. Berrange <berrange@redhat.com> | 8 | which AioContext a coroutine is running on. |
9 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | 9 | |
10 | Acked-by: John Snow <jsnow@redhat.com> | 10 | aio_co_schedule is the part of aio_co_wake that starts a coroutine |
11 | Message-id: 20170825155732.15665-2-stefanha@redhat.com | 11 | on a remove AioContext, but it is also useful to implement e.g. |
12 | bdrv_set_aio_context callbacks. | ||
13 | |||
14 | The implementation of aio_co_schedule is based on a lock-free | ||
15 | multiple-producer, single-consumer queue. The multiple producers use | ||
16 | cmpxchg to add to a LIFO stack. The consumer (a per-AioContext bottom | ||
17 | half) grabs all items added so far, inverts the list to make it FIFO, | ||
18 | and goes through it one item at a time until it's empty. The data | ||
19 | structure was inspired by OSv, which uses it in the very code we'll | ||
20 | "port" to QEMU for the thread-safe CoMutex. | ||
21 | |||
22 | Most of the new code is really tests. | ||
23 | |||
24 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> | ||
25 | Reviewed-by: Fam Zheng <famz@redhat.com> | ||
26 | Message-id: 20170213135235.12274-3-pbonzini@redhat.com | ||
12 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | 27 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
13 | --- | 28 | --- |
14 | COPYING.PYTHON | 270 ++++++ | 29 | tests/Makefile.include | 8 +- |
15 | scripts/argparse.py | 2406 +++++++++++++++++++++++++++++++++++++++++++++++++++ | 30 | include/block/aio.h | 32 +++++++ |
16 | 2 files changed, 2676 insertions(+) | 31 | include/qemu/coroutine_int.h | 11 ++- |
17 | create mode 100644 COPYING.PYTHON | 32 | tests/iothread.h | 25 +++++ |
18 | create mode 100644 scripts/argparse.py | 33 | tests/iothread.c | 91 ++++++++++++++++++ |
19 | 34 | tests/test-aio-multithread.c | 213 +++++++++++++++++++++++++++++++++++++++++++ | |
20 | diff --git a/COPYING.PYTHON b/COPYING.PYTHON | 35 | util/async.c | 65 +++++++++++++ |
36 | util/qemu-coroutine.c | 8 ++ | ||
37 | util/trace-events | 4 + | ||
38 | 9 files changed, 453 insertions(+), 4 deletions(-) | ||
39 | create mode 100644 tests/iothread.h | ||
40 | create mode 100644 tests/iothread.c | ||
41 | create mode 100644 tests/test-aio-multithread.c | ||
42 | |||
43 | diff --git a/tests/Makefile.include b/tests/Makefile.include | ||
44 | index XXXXXXX..XXXXXXX 100644 | ||
45 | --- a/tests/Makefile.include | ||
46 | +++ b/tests/Makefile.include | ||
47 | @@ -XXX,XX +XXX,XX @@ check-unit-y += tests/test-aio$(EXESUF) | ||
48 | gcov-files-test-aio-y = util/async.c util/qemu-timer.o | ||
49 | gcov-files-test-aio-$(CONFIG_WIN32) += util/aio-win32.c | ||
50 | gcov-files-test-aio-$(CONFIG_POSIX) += util/aio-posix.c | ||
51 | +check-unit-y += tests/test-aio-multithread$(EXESUF) | ||
52 | +gcov-files-test-aio-multithread-y = $(gcov-files-test-aio-y) | ||
53 | +gcov-files-test-aio-multithread-y += util/qemu-coroutine.c tests/iothread.c | ||
54 | check-unit-y += tests/test-throttle$(EXESUF) | ||
55 | -gcov-files-test-aio-$(CONFIG_WIN32) = aio-win32.c | ||
56 | -gcov-files-test-aio-$(CONFIG_POSIX) = aio-posix.c | ||
57 | check-unit-y += tests/test-thread-pool$(EXESUF) | ||
58 | gcov-files-test-thread-pool-y = thread-pool.c | ||
59 | gcov-files-test-hbitmap-y = util/hbitmap.c | ||
60 | @@ -XXX,XX +XXX,XX @@ test-qapi-obj-y = tests/test-qapi-visit.o tests/test-qapi-types.o \ | ||
61 | $(test-qom-obj-y) | ||
62 | test-crypto-obj-y = $(crypto-obj-y) $(test-qom-obj-y) | ||
63 | test-io-obj-y = $(io-obj-y) $(test-crypto-obj-y) | ||
64 | -test-block-obj-y = $(block-obj-y) $(test-io-obj-y) | ||
65 | +test-block-obj-y = $(block-obj-y) $(test-io-obj-y) tests/iothread.o | ||
66 | |||
67 | tests/check-qint$(EXESUF): tests/check-qint.o $(test-util-obj-y) | ||
68 | tests/check-qstring$(EXESUF): tests/check-qstring.o $(test-util-obj-y) | ||
69 | @@ -XXX,XX +XXX,XX @@ tests/check-qom-proplist$(EXESUF): tests/check-qom-proplist.o $(test-qom-obj-y) | ||
70 | tests/test-char$(EXESUF): tests/test-char.o $(test-util-obj-y) $(qtest-obj-y) $(test-io-obj-y) $(chardev-obj-y) | ||
71 | tests/test-coroutine$(EXESUF): tests/test-coroutine.o $(test-block-obj-y) | ||
72 | tests/test-aio$(EXESUF): tests/test-aio.o $(test-block-obj-y) | ||
73 | +tests/test-aio-multithread$(EXESUF): tests/test-aio-multithread.o $(test-block-obj-y) | ||
74 | tests/test-throttle$(EXESUF): tests/test-throttle.o $(test-block-obj-y) | ||
75 | tests/test-blockjob$(EXESUF): tests/test-blockjob.o $(test-block-obj-y) $(test-util-obj-y) | ||
76 | tests/test-blockjob-txn$(EXESUF): tests/test-blockjob-txn.o $(test-block-obj-y) $(test-util-obj-y) | ||
77 | diff --git a/include/block/aio.h b/include/block/aio.h | ||
78 | index XXXXXXX..XXXXXXX 100644 | ||
79 | --- a/include/block/aio.h | ||
80 | +++ b/include/block/aio.h | ||
81 | @@ -XXX,XX +XXX,XX @@ typedef void QEMUBHFunc(void *opaque); | ||
82 | typedef bool AioPollFn(void *opaque); | ||
83 | typedef void IOHandler(void *opaque); | ||
84 | |||
85 | +struct Coroutine; | ||
86 | struct ThreadPool; | ||
87 | struct LinuxAioState; | ||
88 | |||
89 | @@ -XXX,XX +XXX,XX @@ struct AioContext { | ||
90 | bool notified; | ||
91 | EventNotifier notifier; | ||
92 | |||
93 | + QSLIST_HEAD(, Coroutine) scheduled_coroutines; | ||
94 | + QEMUBH *co_schedule_bh; | ||
95 | + | ||
96 | /* Thread pool for performing work and receiving completion callbacks. | ||
97 | * Has its own locking. | ||
98 | */ | ||
99 | @@ -XXX,XX +XXX,XX @@ static inline bool aio_node_check(AioContext *ctx, bool is_external) | ||
100 | } | ||
101 | |||
102 | /** | ||
103 | + * aio_co_schedule: | ||
104 | + * @ctx: the aio context | ||
105 | + * @co: the coroutine | ||
106 | + * | ||
107 | + * Start a coroutine on a remote AioContext. | ||
108 | + * | ||
109 | + * The coroutine must not be entered by anyone else while aio_co_schedule() | ||
110 | + * is active. In addition the coroutine must have yielded unless ctx | ||
111 | + * is the context in which the coroutine is running (i.e. the value of | ||
112 | + * qemu_get_current_aio_context() from the coroutine itself). | ||
113 | + */ | ||
114 | +void aio_co_schedule(AioContext *ctx, struct Coroutine *co); | ||
115 | + | ||
116 | +/** | ||
117 | + * aio_co_wake: | ||
118 | + * @co: the coroutine | ||
119 | + * | ||
120 | + * Restart a coroutine on the AioContext where it was running last, thus | ||
121 | + * preventing coroutines from jumping from one context to another when they | ||
122 | + * go to sleep. | ||
123 | + * | ||
124 | + * aio_co_wake may be executed either in coroutine or non-coroutine | ||
125 | + * context. The coroutine must not be entered by anyone else while | ||
126 | + * aio_co_wake() is active. | ||
127 | + */ | ||
128 | +void aio_co_wake(struct Coroutine *co); | ||
129 | + | ||
130 | +/** | ||
131 | * Return the AioContext whose event loop runs in the current thread. | ||
132 | * | ||
133 | * If called from an IOThread this will be the IOThread's AioContext. If | ||
134 | diff --git a/include/qemu/coroutine_int.h b/include/qemu/coroutine_int.h | ||
135 | index XXXXXXX..XXXXXXX 100644 | ||
136 | --- a/include/qemu/coroutine_int.h | ||
137 | +++ b/include/qemu/coroutine_int.h | ||
138 | @@ -XXX,XX +XXX,XX @@ struct Coroutine { | ||
139 | CoroutineEntry *entry; | ||
140 | void *entry_arg; | ||
141 | Coroutine *caller; | ||
142 | + | ||
143 | + /* Only used when the coroutine has terminated. */ | ||
144 | QSLIST_ENTRY(Coroutine) pool_next; | ||
145 | + | ||
146 | size_t locks_held; | ||
147 | |||
148 | - /* Coroutines that should be woken up when we yield or terminate */ | ||
149 | + /* Coroutines that should be woken up when we yield or terminate. | ||
150 | + * Only used when the coroutine is running. | ||
151 | + */ | ||
152 | QSIMPLEQ_HEAD(, Coroutine) co_queue_wakeup; | ||
153 | + | ||
154 | + /* Only used when the coroutine has yielded. */ | ||
155 | + AioContext *ctx; | ||
156 | QSIMPLEQ_ENTRY(Coroutine) co_queue_next; | ||
157 | + QSLIST_ENTRY(Coroutine) co_scheduled_next; | ||
158 | }; | ||
159 | |||
160 | Coroutine *qemu_coroutine_new(void); | ||
161 | diff --git a/tests/iothread.h b/tests/iothread.h | ||
21 | new file mode 100644 | 162 | new file mode 100644 |
22 | index XXXXXXX..XXXXXXX | 163 | index XXXXXXX..XXXXXXX |
23 | --- /dev/null | 164 | --- /dev/null |
24 | +++ b/COPYING.PYTHON | 165 | +++ b/tests/iothread.h |
25 | @@ -XXX,XX +XXX,XX @@ | 166 | @@ -XXX,XX +XXX,XX @@ |
26 | +A. HISTORY OF THE SOFTWARE | 167 | +/* |
27 | +========================== | 168 | + * Event loop thread implementation for unit tests |
28 | + | 169 | + * |
29 | +Python was created in the early 1990s by Guido van Rossum at Stichting | 170 | + * Copyright Red Hat Inc., 2013, 2016 |
30 | +Mathematisch Centrum (CWI, see http://www.cwi.nl) in the Netherlands | 171 | + * |
31 | +as a successor of a language called ABC. Guido remains Python's | 172 | + * Authors: |
32 | +principal author, although it includes many contributions from others. | 173 | + * Stefan Hajnoczi <stefanha@redhat.com> |
33 | + | 174 | + * Paolo Bonzini <pbonzini@redhat.com> |
34 | +In 1995, Guido continued his work on Python at the Corporation for | 175 | + * |
35 | +National Research Initiatives (CNRI, see http://www.cnri.reston.va.us) | 176 | + * This work is licensed under the terms of the GNU GPL, version 2 or later. |
36 | +in Reston, Virginia where he released several versions of the | 177 | + * See the COPYING file in the top-level directory. |
37 | +software. | 178 | + */ |
38 | + | 179 | +#ifndef TEST_IOTHREAD_H |
39 | +In May 2000, Guido and the Python core development team moved to | 180 | +#define TEST_IOTHREAD_H |
40 | +BeOpen.com to form the BeOpen PythonLabs team. In October of the same | 181 | + |
41 | +year, the PythonLabs team moved to Digital Creations (now Zope | 182 | +#include "block/aio.h" |
42 | +Corporation, see http://www.zope.com). In 2001, the Python Software | 183 | +#include "qemu/thread.h" |
43 | +Foundation (PSF, see http://www.python.org/psf/) was formed, a | 184 | + |
44 | +non-profit organization created specifically to own Python-related | 185 | +typedef struct IOThread IOThread; |
45 | +Intellectual Property. Zope Corporation is a sponsoring member of | 186 | + |
46 | +the PSF. | 187 | +IOThread *iothread_new(void); |
47 | + | 188 | +void iothread_join(IOThread *iothread); |
48 | +All Python releases are Open Source (see http://www.opensource.org for | 189 | +AioContext *iothread_get_aio_context(IOThread *iothread); |
49 | +the Open Source Definition). Historically, most, but not all, Python | 190 | + |
50 | +releases have also been GPL-compatible; the table below summarizes | 191 | +#endif |
51 | +the various releases. | 192 | diff --git a/tests/iothread.c b/tests/iothread.c |
52 | + | ||
53 | + Release Derived Year Owner GPL- | ||
54 | + from compatible? (1) | ||
55 | + | ||
56 | + 0.9.0 thru 1.2 1991-1995 CWI yes | ||
57 | + 1.3 thru 1.5.2 1.2 1995-1999 CNRI yes | ||
58 | + 1.6 1.5.2 2000 CNRI no | ||
59 | + 2.0 1.6 2000 BeOpen.com no | ||
60 | + 1.6.1 1.6 2001 CNRI yes (2) | ||
61 | + 2.1 2.0+1.6.1 2001 PSF no | ||
62 | + 2.0.1 2.0+1.6.1 2001 PSF yes | ||
63 | + 2.1.1 2.1+2.0.1 2001 PSF yes | ||
64 | + 2.2 2.1.1 2001 PSF yes | ||
65 | + 2.1.2 2.1.1 2002 PSF yes | ||
66 | + 2.1.3 2.1.2 2002 PSF yes | ||
67 | + 2.2.1 2.2 2002 PSF yes | ||
68 | + 2.2.2 2.2.1 2002 PSF yes | ||
69 | + 2.2.3 2.2.2 2003 PSF yes | ||
70 | + 2.3 2.2.2 2002-2003 PSF yes | ||
71 | + 2.3.1 2.3 2002-2003 PSF yes | ||
72 | + 2.3.2 2.3.1 2002-2003 PSF yes | ||
73 | + 2.3.3 2.3.2 2002-2003 PSF yes | ||
74 | + 2.3.4 2.3.3 2004 PSF yes | ||
75 | + 2.3.5 2.3.4 2005 PSF yes | ||
76 | + 2.4 2.3 2004 PSF yes | ||
77 | + 2.4.1 2.4 2005 PSF yes | ||
78 | + 2.4.2 2.4.1 2005 PSF yes | ||
79 | + 2.4.3 2.4.2 2006 PSF yes | ||
80 | + 2.5 2.4 2006 PSF yes | ||
81 | + 2.7 2.6 2010 PSF yes | ||
82 | + | ||
83 | +Footnotes: | ||
84 | + | ||
85 | +(1) GPL-compatible doesn't mean that we're distributing Python under | ||
86 | + the GPL. All Python licenses, unlike the GPL, let you distribute | ||
87 | + a modified version without making your changes open source. The | ||
88 | + GPL-compatible licenses make it possible to combine Python with | ||
89 | + other software that is released under the GPL; the others don't. | ||
90 | + | ||
91 | +(2) According to Richard Stallman, 1.6.1 is not GPL-compatible, | ||
92 | + because its license has a choice of law clause. According to | ||
93 | + CNRI, however, Stallman's lawyer has told CNRI's lawyer that 1.6.1 | ||
94 | + is "not incompatible" with the GPL. | ||
95 | + | ||
96 | +Thanks to the many outside volunteers who have worked under Guido's | ||
97 | +direction to make these releases possible. | ||
98 | + | ||
99 | + | ||
100 | +B. TERMS AND CONDITIONS FOR ACCESSING OR OTHERWISE USING PYTHON | ||
101 | +=============================================================== | ||
102 | + | ||
103 | +PYTHON SOFTWARE FOUNDATION LICENSE VERSION 2 | ||
104 | +-------------------------------------------- | ||
105 | + | ||
106 | +1. This LICENSE AGREEMENT is between the Python Software Foundation | ||
107 | +("PSF"), and the Individual or Organization ("Licensee") accessing and | ||
108 | +otherwise using this software ("Python") in source or binary form and | ||
109 | +its associated documentation. | ||
110 | + | ||
111 | +2. Subject to the terms and conditions of this License Agreement, PSF | ||
112 | +hereby grants Licensee a nonexclusive, royalty-free, world-wide | ||
113 | +license to reproduce, analyze, test, perform and/or display publicly, | ||
114 | +prepare derivative works, distribute, and otherwise use Python | ||
115 | +alone or in any derivative version, provided, however, that PSF's | ||
116 | +License Agreement and PSF's notice of copyright, i.e., "Copyright (c) | ||
117 | +2001, 2002, 2003, 2004, 2005, 2006 Python Software Foundation; All Rights | ||
118 | +Reserved" are retained in Python alone or in any derivative version | ||
119 | +prepared by Licensee. | ||
120 | + | ||
121 | +3. In the event Licensee prepares a derivative work that is based on | ||
122 | +or incorporates Python or any part thereof, and wants to make | ||
123 | +the derivative work available to others as provided herein, then | ||
124 | +Licensee hereby agrees to include in any such work a brief summary of | ||
125 | +the changes made to Python. | ||
126 | + | ||
127 | +4. PSF is making Python available to Licensee on an "AS IS" | ||
128 | +basis. PSF MAKES NO REPRESENTATIONS OR WARRANTIES, EXPRESS OR | ||
129 | +IMPLIED. BY WAY OF EXAMPLE, BUT NOT LIMITATION, PSF MAKES NO AND | ||
130 | +DISCLAIMS ANY REPRESENTATION OR WARRANTY OF MERCHANTABILITY OR FITNESS | ||
131 | +FOR ANY PARTICULAR PURPOSE OR THAT THE USE OF PYTHON WILL NOT | ||
132 | +INFRINGE ANY THIRD PARTY RIGHTS. | ||
133 | + | ||
134 | +5. PSF SHALL NOT BE LIABLE TO LICENSEE OR ANY OTHER USERS OF PYTHON | ||
135 | +FOR ANY INCIDENTAL, SPECIAL, OR CONSEQUENTIAL DAMAGES OR LOSS AS | ||
136 | +A RESULT OF MODIFYING, DISTRIBUTING, OR OTHERWISE USING PYTHON, | ||
137 | +OR ANY DERIVATIVE THEREOF, EVEN IF ADVISED OF THE POSSIBILITY THEREOF. | ||
138 | + | ||
139 | +6. This License Agreement will automatically terminate upon a material | ||
140 | +breach of its terms and conditions. | ||
141 | + | ||
142 | +7. Nothing in this License Agreement shall be deemed to create any | ||
143 | +relationship of agency, partnership, or joint venture between PSF and | ||
144 | +Licensee. This License Agreement does not grant permission to use PSF | ||
145 | +trademarks or trade name in a trademark sense to endorse or promote | ||
146 | +products or services of Licensee, or any third party. | ||
147 | + | ||
148 | +8. By copying, installing or otherwise using Python, Licensee | ||
149 | +agrees to be bound by the terms and conditions of this License | ||
150 | +Agreement. | ||
151 | + | ||
152 | + | ||
153 | +BEOPEN.COM LICENSE AGREEMENT FOR PYTHON 2.0 | ||
154 | +------------------------------------------- | ||
155 | + | ||
156 | +BEOPEN PYTHON OPEN SOURCE LICENSE AGREEMENT VERSION 1 | ||
157 | + | ||
158 | +1. This LICENSE AGREEMENT is between BeOpen.com ("BeOpen"), having an | ||
159 | +office at 160 Saratoga Avenue, Santa Clara, CA 95051, and the | ||
160 | +Individual or Organization ("Licensee") accessing and otherwise using | ||
161 | +this software in source or binary form and its associated | ||
162 | +documentation ("the Software"). | ||
163 | + | ||
164 | +2. Subject to the terms and conditions of this BeOpen Python License | ||
165 | +Agreement, BeOpen hereby grants Licensee a non-exclusive, | ||
166 | +royalty-free, world-wide license to reproduce, analyze, test, perform | ||
167 | +and/or display publicly, prepare derivative works, distribute, and | ||
168 | +otherwise use the Software alone or in any derivative version, | ||
169 | +provided, however, that the BeOpen Python License is retained in the | ||
170 | +Software, alone or in any derivative version prepared by Licensee. | ||
171 | + | ||
172 | +3. BeOpen is making the Software available to Licensee on an "AS IS" | ||
173 | +basis. BEOPEN MAKES NO REPRESENTATIONS OR WARRANTIES, EXPRESS OR | ||
174 | +IMPLIED. BY WAY OF EXAMPLE, BUT NOT LIMITATION, BEOPEN MAKES NO AND | ||
175 | +DISCLAIMS ANY REPRESENTATION OR WARRANTY OF MERCHANTABILITY OR FITNESS | ||
176 | +FOR ANY PARTICULAR PURPOSE OR THAT THE USE OF THE SOFTWARE WILL NOT | ||
177 | +INFRINGE ANY THIRD PARTY RIGHTS. | ||
178 | + | ||
179 | +4. BEOPEN SHALL NOT BE LIABLE TO LICENSEE OR ANY OTHER USERS OF THE | ||
180 | +SOFTWARE FOR ANY INCIDENTAL, SPECIAL, OR CONSEQUENTIAL DAMAGES OR LOSS | ||
181 | +AS A RESULT OF USING, MODIFYING OR DISTRIBUTING THE SOFTWARE, OR ANY | ||
182 | +DERIVATIVE THEREOF, EVEN IF ADVISED OF THE POSSIBILITY THEREOF. | ||
183 | + | ||
184 | +5. This License Agreement will automatically terminate upon a material | ||
185 | +breach of its terms and conditions. | ||
186 | + | ||
187 | +6. This License Agreement shall be governed by and interpreted in all | ||
188 | +respects by the law of the State of California, excluding conflict of | ||
189 | +law provisions. Nothing in this License Agreement shall be deemed to | ||
190 | +create any relationship of agency, partnership, or joint venture | ||
191 | +between BeOpen and Licensee. This License Agreement does not grant | ||
192 | +permission to use BeOpen trademarks or trade names in a trademark | ||
193 | +sense to endorse or promote products or services of Licensee, or any | ||
194 | +third party. As an exception, the "BeOpen Python" logos available at | ||
195 | +http://www.pythonlabs.com/logos.html may be used according to the | ||
196 | +permissions granted on that web page. | ||
197 | + | ||
198 | +7. By copying, installing or otherwise using the software, Licensee | ||
199 | +agrees to be bound by the terms and conditions of this License | ||
200 | +Agreement. | ||
201 | + | ||
202 | + | ||
203 | +CNRI LICENSE AGREEMENT FOR PYTHON 1.6.1 | ||
204 | +--------------------------------------- | ||
205 | + | ||
206 | +1. This LICENSE AGREEMENT is between the Corporation for National | ||
207 | +Research Initiatives, having an office at 1895 Preston White Drive, | ||
208 | +Reston, VA 20191 ("CNRI"), and the Individual or Organization | ||
209 | +("Licensee") accessing and otherwise using Python 1.6.1 software in | ||
210 | +source or binary form and its associated documentation. | ||
211 | + | ||
212 | +2. Subject to the terms and conditions of this License Agreement, CNRI | ||
213 | +hereby grants Licensee a nonexclusive, royalty-free, world-wide | ||
214 | +license to reproduce, analyze, test, perform and/or display publicly, | ||
215 | +prepare derivative works, distribute, and otherwise use Python 1.6.1 | ||
216 | +alone or in any derivative version, provided, however, that CNRI's | ||
217 | +License Agreement and CNRI's notice of copyright, i.e., "Copyright (c) | ||
218 | +1995-2001 Corporation for National Research Initiatives; All Rights | ||
219 | +Reserved" are retained in Python 1.6.1 alone or in any derivative | ||
220 | +version prepared by Licensee. Alternately, in lieu of CNRI's License | ||
221 | +Agreement, Licensee may substitute the following text (omitting the | ||
222 | +quotes): "Python 1.6.1 is made available subject to the terms and | ||
223 | +conditions in CNRI's License Agreement. This Agreement together with | ||
224 | +Python 1.6.1 may be located on the Internet using the following | ||
225 | +unique, persistent identifier (known as a handle): 1895.22/1013. This | ||
226 | +Agreement may also be obtained from a proxy server on the Internet | ||
227 | +using the following URL: http://hdl.handle.net/1895.22/1013". | ||
228 | + | ||
229 | +3. In the event Licensee prepares a derivative work that is based on | ||
230 | +or incorporates Python 1.6.1 or any part thereof, and wants to make | ||
231 | +the derivative work available to others as provided herein, then | ||
232 | +Licensee hereby agrees to include in any such work a brief summary of | ||
233 | +the changes made to Python 1.6.1. | ||
234 | + | ||
235 | +4. CNRI is making Python 1.6.1 available to Licensee on an "AS IS" | ||
236 | +basis. CNRI MAKES NO REPRESENTATIONS OR WARRANTIES, EXPRESS OR | ||
237 | +IMPLIED. BY WAY OF EXAMPLE, BUT NOT LIMITATION, CNRI MAKES NO AND | ||
238 | +DISCLAIMS ANY REPRESENTATION OR WARRANTY OF MERCHANTABILITY OR FITNESS | ||
239 | +FOR ANY PARTICULAR PURPOSE OR THAT THE USE OF PYTHON 1.6.1 WILL NOT | ||
240 | +INFRINGE ANY THIRD PARTY RIGHTS. | ||
241 | + | ||
242 | +5. CNRI SHALL NOT BE LIABLE TO LICENSEE OR ANY OTHER USERS OF PYTHON | ||
243 | +1.6.1 FOR ANY INCIDENTAL, SPECIAL, OR CONSEQUENTIAL DAMAGES OR LOSS AS | ||
244 | +A RESULT OF MODIFYING, DISTRIBUTING, OR OTHERWISE USING PYTHON 1.6.1, | ||
245 | +OR ANY DERIVATIVE THEREOF, EVEN IF ADVISED OF THE POSSIBILITY THEREOF. | ||
246 | + | ||
247 | +6. This License Agreement will automatically terminate upon a material | ||
248 | +breach of its terms and conditions. | ||
249 | + | ||
250 | +7. This License Agreement shall be governed by the federal | ||
251 | +intellectual property law of the United States, including without | ||
252 | +limitation the federal copyright law, and, to the extent such | ||
253 | +U.S. federal law does not apply, by the law of the Commonwealth of | ||
254 | +Virginia, excluding Virginia's conflict of law provisions. | ||
255 | +Notwithstanding the foregoing, with regard to derivative works based | ||
256 | +on Python 1.6.1 that incorporate non-separable material that was | ||
257 | +previously distributed under the GNU General Public License (GPL), the | ||
258 | +law of the Commonwealth of Virginia shall govern this License | ||
259 | +Agreement only as to issues arising under or with respect to | ||
260 | +Paragraphs 4, 5, and 7 of this License Agreement. Nothing in this | ||
261 | +License Agreement shall be deemed to create any relationship of | ||
262 | +agency, partnership, or joint venture between CNRI and Licensee. This | ||
263 | +License Agreement does not grant permission to use CNRI trademarks or | ||
264 | +trade name in a trademark sense to endorse or promote products or | ||
265 | +services of Licensee, or any third party. | ||
266 | + | ||
267 | +8. By clicking on the "ACCEPT" button where indicated, or by copying, | ||
268 | +installing or otherwise using Python 1.6.1, Licensee agrees to be | ||
269 | +bound by the terms and conditions of this License Agreement. | ||
270 | + | ||
271 | + ACCEPT | ||
272 | + | ||
273 | + | ||
274 | +CWI LICENSE AGREEMENT FOR PYTHON 0.9.0 THROUGH 1.2 | ||
275 | +-------------------------------------------------- | ||
276 | + | ||
277 | +Copyright (c) 1991 - 1995, Stichting Mathematisch Centrum Amsterdam, | ||
278 | +The Netherlands. All rights reserved. | ||
279 | + | ||
280 | +Permission to use, copy, modify, and distribute this software and its | ||
281 | +documentation for any purpose and without fee is hereby granted, | ||
282 | +provided that the above copyright notice appear in all copies and that | ||
283 | +both that copyright notice and this permission notice appear in | ||
284 | +supporting documentation, and that the name of Stichting Mathematisch | ||
285 | +Centrum or CWI not be used in advertising or publicity pertaining to | ||
286 | +distribution of the software without specific, written prior | ||
287 | +permission. | ||
288 | + | ||
289 | +STICHTING MATHEMATISCH CENTRUM DISCLAIMS ALL WARRANTIES WITH REGARD TO | ||
290 | +THIS SOFTWARE, INCLUDING ALL IMPLIED WARRANTIES OF MERCHANTABILITY AND | ||
291 | +FITNESS, IN NO EVENT SHALL STICHTING MATHEMATISCH CENTRUM BE LIABLE | ||
292 | +FOR ANY SPECIAL, INDIRECT OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES | ||
293 | +WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN | ||
294 | +ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT | ||
295 | +OF OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. | ||
296 | diff --git a/scripts/argparse.py b/scripts/argparse.py | ||
297 | new file mode 100644 | 193 | new file mode 100644 |
298 | index XXXXXXX..XXXXXXX | 194 | index XXXXXXX..XXXXXXX |
299 | --- /dev/null | 195 | --- /dev/null |
300 | +++ b/scripts/argparse.py | 196 | +++ b/tests/iothread.c |
301 | @@ -XXX,XX +XXX,XX @@ | 197 | @@ -XXX,XX +XXX,XX @@ |
302 | +# This is a local copy of the standard library argparse module taken from PyPI. | 198 | +/* |
303 | +# It is licensed under the Python Software Foundation License. This is a | 199 | + * Event loop thread implementation for unit tests |
304 | +# fallback for Python 2.6 which does not include this module. Python 2.7+ and | 200 | + * |
305 | +# 3+ will never load this module because built-in modules are loaded before | 201 | + * Copyright Red Hat Inc., 2013, 2016 |
306 | +# anything in sys.path. | 202 | + * |
307 | +# | 203 | + * Authors: |
308 | +# If your script is not located in the same directory as this file, import it | 204 | + * Stefan Hajnoczi <stefanha@redhat.com> |
309 | +# like this: | 205 | + * Paolo Bonzini <pbonzini@redhat.com> |
310 | +# | 206 | + * |
311 | +# import os | 207 | + * This work is licensed under the terms of the GNU GPL, version 2 or later. |
312 | +# import sys | 208 | + * See the COPYING file in the top-level directory. |
313 | +# sys.path.append(os.path.join(os.path.dirname(__file__), ..., 'scripts')) | 209 | + * |
314 | +# import argparse | 210 | + */ |
315 | + | 211 | + |
316 | +# Author: Steven J. Bethard <steven.bethard@gmail.com>. | 212 | +#include "qemu/osdep.h" |
317 | +# Maintainer: Thomas Waldmann <tw@waldmann-edv.de> | 213 | +#include "qapi/error.h" |
318 | + | 214 | +#include "block/aio.h" |
319 | +"""Command-line parsing library | 215 | +#include "qemu/main-loop.h" |
320 | + | 216 | +#include "qemu/rcu.h" |
321 | +This module is an optparse-inspired command-line parsing library that: | 217 | +#include "iothread.h" |
322 | + | 218 | + |
323 | + - handles both optional and positional arguments | 219 | +struct IOThread { |
324 | + - produces highly informative usage messages | 220 | + AioContext *ctx; |
325 | + - supports parsers that dispatch to sub-parsers | 221 | + |
326 | + | 222 | + QemuThread thread; |
327 | +The following is a simple usage example that sums integers from the | 223 | + QemuMutex init_done_lock; |
328 | +command-line and writes the result to a file:: | 224 | + QemuCond init_done_cond; /* is thread initialization done? */ |
329 | + | 225 | + bool stopping; |
330 | + parser = argparse.ArgumentParser( | 226 | +}; |
331 | + description='sum the integers at the command line') | 227 | + |
332 | + parser.add_argument( | 228 | +static __thread IOThread *my_iothread; |
333 | + 'integers', metavar='int', nargs='+', type=int, | 229 | + |
334 | + help='an integer to be summed') | 230 | +AioContext *qemu_get_current_aio_context(void) |
335 | + parser.add_argument( | 231 | +{ |
336 | + '--log', default=sys.stdout, type=argparse.FileType('w'), | 232 | + return my_iothread ? my_iothread->ctx : qemu_get_aio_context(); |
337 | + help='the file where the sum should be written') | 233 | +} |
338 | + args = parser.parse_args() | 234 | + |
339 | + args.log.write('%s' % sum(args.integers)) | 235 | +static void *iothread_run(void *opaque) |
340 | + args.log.close() | 236 | +{ |
341 | + | 237 | + IOThread *iothread = opaque; |
342 | +The module contains the following public classes: | 238 | + |
343 | + | 239 | + rcu_register_thread(); |
344 | + - ArgumentParser -- The main entry point for command-line parsing. As the | 240 | + |
345 | + example above shows, the add_argument() method is used to populate | 241 | + my_iothread = iothread; |
346 | + the parser with actions for optional and positional arguments. Then | 242 | + qemu_mutex_lock(&iothread->init_done_lock); |
347 | + the parse_args() method is invoked to convert the args at the | 243 | + iothread->ctx = aio_context_new(&error_abort); |
348 | + command-line into an object with attributes. | 244 | + qemu_cond_signal(&iothread->init_done_cond); |
349 | + | 245 | + qemu_mutex_unlock(&iothread->init_done_lock); |
350 | + - ArgumentError -- The exception raised by ArgumentParser objects when | 246 | + |
351 | + there are errors with the parser's actions. Errors raised while | 247 | + while (!atomic_read(&iothread->stopping)) { |
352 | + parsing the command-line are caught by ArgumentParser and emitted | 248 | + aio_poll(iothread->ctx, true); |
353 | + as command-line messages. | 249 | + } |
354 | + | 250 | + |
355 | + - FileType -- A factory for defining types of files to be created. As the | 251 | + rcu_unregister_thread(); |
356 | + example above shows, instances of FileType are typically passed as | 252 | + return NULL; |
357 | + the type= argument of add_argument() calls. | 253 | +} |
358 | + | 254 | + |
359 | + - Action -- The base class for parser actions. Typically actions are | 255 | +void iothread_join(IOThread *iothread) |
360 | + selected by passing strings like 'store_true' or 'append_const' to | 256 | +{ |
361 | + the action= argument of add_argument(). However, for greater | 257 | + iothread->stopping = true; |
362 | + customization of ArgumentParser actions, subclasses of Action may | 258 | + aio_notify(iothread->ctx); |
363 | + be defined and passed as the action= argument. | 259 | + qemu_thread_join(&iothread->thread); |
364 | + | 260 | + qemu_cond_destroy(&iothread->init_done_cond); |
365 | + - HelpFormatter, RawDescriptionHelpFormatter, RawTextHelpFormatter, | 261 | + qemu_mutex_destroy(&iothread->init_done_lock); |
366 | + ArgumentDefaultsHelpFormatter -- Formatter classes which | 262 | + aio_context_unref(iothread->ctx); |
367 | + may be passed as the formatter_class= argument to the | 263 | + g_free(iothread); |
368 | + ArgumentParser constructor. HelpFormatter is the default, | 264 | +} |
369 | + RawDescriptionHelpFormatter and RawTextHelpFormatter tell the parser | 265 | + |
370 | + not to change the formatting for help text, and | 266 | +IOThread *iothread_new(void) |
371 | + ArgumentDefaultsHelpFormatter adds information about argument defaults | 267 | +{ |
372 | + to the help. | 268 | + IOThread *iothread = g_new0(IOThread, 1); |
373 | + | 269 | + |
374 | +All other classes in this module are considered implementation details. | 270 | + qemu_mutex_init(&iothread->init_done_lock); |
375 | +(Also note that HelpFormatter and RawDescriptionHelpFormatter are only | 271 | + qemu_cond_init(&iothread->init_done_cond); |
376 | +considered public as object names -- the API of the formatter objects is | 272 | + qemu_thread_create(&iothread->thread, NULL, iothread_run, |
377 | +still considered an implementation detail.) | 273 | + iothread, QEMU_THREAD_JOINABLE); |
378 | +""" | 274 | + |
379 | + | 275 | + /* Wait for initialization to complete */ |
380 | +__version__ = '1.4.0' # we use our own version number independant of the | 276 | + qemu_mutex_lock(&iothread->init_done_lock); |
381 | + # one in stdlib and we release this on pypi. | 277 | + while (iothread->ctx == NULL) { |
382 | + | 278 | + qemu_cond_wait(&iothread->init_done_cond, |
383 | +__external_lib__ = True # to make sure the tests really test THIS lib, | 279 | + &iothread->init_done_lock); |
384 | + # not the builtin one in Python stdlib | 280 | + } |
385 | + | 281 | + qemu_mutex_unlock(&iothread->init_done_lock); |
386 | +__all__ = [ | 282 | + return iothread; |
387 | + 'ArgumentParser', | 283 | +} |
388 | + 'ArgumentError', | 284 | + |
389 | + 'ArgumentTypeError', | 285 | +AioContext *iothread_get_aio_context(IOThread *iothread) |
390 | + 'FileType', | 286 | +{ |
391 | + 'HelpFormatter', | 287 | + return iothread->ctx; |
392 | + 'ArgumentDefaultsHelpFormatter', | 288 | +} |
393 | + 'RawDescriptionHelpFormatter', | 289 | diff --git a/tests/test-aio-multithread.c b/tests/test-aio-multithread.c |
394 | + 'RawTextHelpFormatter', | 290 | new file mode 100644 |
395 | + 'Namespace', | 291 | index XXXXXXX..XXXXXXX |
396 | + 'Action', | 292 | --- /dev/null |
397 | + 'ONE_OR_MORE', | 293 | +++ b/tests/test-aio-multithread.c |
398 | + 'OPTIONAL', | 294 | @@ -XXX,XX +XXX,XX @@ |
399 | + 'PARSER', | 295 | +/* |
400 | + 'REMAINDER', | 296 | + * AioContext multithreading tests |
401 | + 'SUPPRESS', | 297 | + * |
402 | + 'ZERO_OR_MORE', | 298 | + * Copyright Red Hat, Inc. 2016 |
403 | +] | 299 | + * |
404 | + | 300 | + * Authors: |
405 | + | 301 | + * Paolo Bonzini <pbonzini@redhat.com> |
406 | +import copy as _copy | 302 | + * |
407 | +import os as _os | 303 | + * This work is licensed under the terms of the GNU LGPL, version 2 or later. |
408 | +import re as _re | 304 | + * See the COPYING.LIB file in the top-level directory. |
409 | +import sys as _sys | 305 | + */ |
410 | +import textwrap as _textwrap | 306 | + |
411 | + | 307 | +#include "qemu/osdep.h" |
412 | +from gettext import gettext as _ | 308 | +#include <glib.h> |
413 | + | 309 | +#include "block/aio.h" |
414 | +try: | 310 | +#include "qapi/error.h" |
415 | + set | 311 | +#include "qemu/coroutine.h" |
416 | +except NameError: | 312 | +#include "qemu/thread.h" |
417 | + # for python < 2.4 compatibility (sets module is there since 2.3): | 313 | +#include "qemu/error-report.h" |
418 | + from sets import Set as set | 314 | +#include "iothread.h" |
419 | + | 315 | + |
420 | +try: | 316 | +/* AioContext management */ |
421 | + basestring | 317 | + |
422 | +except NameError: | 318 | +#define NUM_CONTEXTS 5 |
423 | + basestring = str | 319 | + |
424 | + | 320 | +static IOThread *threads[NUM_CONTEXTS]; |
425 | +try: | 321 | +static AioContext *ctx[NUM_CONTEXTS]; |
426 | + sorted | 322 | +static __thread int id = -1; |
427 | +except NameError: | 323 | + |
428 | + # for python < 2.4 compatibility: | 324 | +static QemuEvent done_event; |
429 | + def sorted(iterable, reverse=False): | 325 | + |
430 | + result = list(iterable) | 326 | +/* Run a function synchronously on a remote iothread. */ |
431 | + result.sort() | 327 | + |
432 | + if reverse: | 328 | +typedef struct CtxRunData { |
433 | + result.reverse() | 329 | + QEMUBHFunc *cb; |
434 | + return result | 330 | + void *arg; |
435 | + | 331 | +} CtxRunData; |
436 | + | 332 | + |
437 | +def _callable(obj): | 333 | +static void ctx_run_bh_cb(void *opaque) |
438 | + return hasattr(obj, '__call__') or hasattr(obj, '__bases__') | 334 | +{ |
439 | + | 335 | + CtxRunData *data = opaque; |
440 | + | 336 | + |
441 | +SUPPRESS = '==SUPPRESS==' | 337 | + data->cb(data->arg); |
442 | + | 338 | + qemu_event_set(&done_event); |
443 | +OPTIONAL = '?' | 339 | +} |
444 | +ZERO_OR_MORE = '*' | 340 | + |
445 | +ONE_OR_MORE = '+' | 341 | +static void ctx_run(int i, QEMUBHFunc *cb, void *opaque) |
446 | +PARSER = 'A...' | 342 | +{ |
447 | +REMAINDER = '...' | 343 | + CtxRunData data = { |
448 | +_UNRECOGNIZED_ARGS_ATTR = '_unrecognized_args' | 344 | + .cb = cb, |
449 | + | 345 | + .arg = opaque |
450 | +# ============================= | 346 | + }; |
451 | +# Utility functions and classes | 347 | + |
452 | +# ============================= | 348 | + qemu_event_reset(&done_event); |
453 | + | 349 | + aio_bh_schedule_oneshot(ctx[i], ctx_run_bh_cb, &data); |
454 | +class _AttributeHolder(object): | 350 | + qemu_event_wait(&done_event); |
455 | + """Abstract base class that provides __repr__. | 351 | +} |
456 | + | 352 | + |
457 | + The __repr__ method returns a string in the format:: | 353 | +/* Starting the iothreads. */ |
458 | + ClassName(attr=name, attr=name, ...) | 354 | + |
459 | + The attributes are determined either by a class-level attribute, | 355 | +static void set_id_cb(void *opaque) |
460 | + '_kwarg_names', or by inspecting the instance __dict__. | 356 | +{ |
461 | + """ | 357 | + int *i = opaque; |
462 | + | 358 | + |
463 | + def __repr__(self): | 359 | + id = *i; |
464 | + type_name = type(self).__name__ | 360 | +} |
465 | + arg_strings = [] | 361 | + |
466 | + for arg in self._get_args(): | 362 | +static void create_aio_contexts(void) |
467 | + arg_strings.append(repr(arg)) | 363 | +{ |
468 | + for name, value in self._get_kwargs(): | 364 | + int i; |
469 | + arg_strings.append('%s=%r' % (name, value)) | 365 | + |
470 | + return '%s(%s)' % (type_name, ', '.join(arg_strings)) | 366 | + for (i = 0; i < NUM_CONTEXTS; i++) { |
471 | + | 367 | + threads[i] = iothread_new(); |
472 | + def _get_kwargs(self): | 368 | + ctx[i] = iothread_get_aio_context(threads[i]); |
473 | + return sorted(self.__dict__.items()) | 369 | + } |
474 | + | 370 | + |
475 | + def _get_args(self): | 371 | + qemu_event_init(&done_event, false); |
476 | + return [] | 372 | + for (i = 0; i < NUM_CONTEXTS; i++) { |
477 | + | 373 | + ctx_run(i, set_id_cb, &i); |
478 | + | 374 | + } |
479 | +def _ensure_value(namespace, name, value): | 375 | +} |
480 | + if getattr(namespace, name, None) is None: | 376 | + |
481 | + setattr(namespace, name, value) | 377 | +/* Stopping the iothreads. */ |
482 | + return getattr(namespace, name) | 378 | + |
483 | + | 379 | +static void join_aio_contexts(void) |
484 | + | 380 | +{ |
485 | +# =============== | 381 | + int i; |
486 | +# Formatting Help | 382 | + |
487 | +# =============== | 383 | + for (i = 0; i < NUM_CONTEXTS; i++) { |
488 | + | 384 | + aio_context_ref(ctx[i]); |
489 | +class HelpFormatter(object): | 385 | + } |
490 | + """Formatter for generating usage messages and argument help strings. | 386 | + for (i = 0; i < NUM_CONTEXTS; i++) { |
491 | + | 387 | + iothread_join(threads[i]); |
492 | + Only the name of this class is considered a public API. All the methods | 388 | + } |
493 | + provided by the class are considered an implementation detail. | 389 | + for (i = 0; i < NUM_CONTEXTS; i++) { |
494 | + """ | 390 | + aio_context_unref(ctx[i]); |
495 | + | 391 | + } |
496 | + def __init__(self, | 392 | + qemu_event_destroy(&done_event); |
497 | + prog, | 393 | +} |
498 | + indent_increment=2, | 394 | + |
499 | + max_help_position=24, | 395 | +/* Basic test for the stuff above. */ |
500 | + width=None): | 396 | + |
501 | + | 397 | +static void test_lifecycle(void) |
502 | + # default setting for width | 398 | +{ |
503 | + if width is None: | 399 | + create_aio_contexts(); |
504 | + try: | 400 | + join_aio_contexts(); |
505 | + width = int(_os.environ['COLUMNS']) | 401 | +} |
506 | + except (KeyError, ValueError): | 402 | + |
507 | + width = 80 | 403 | +/* aio_co_schedule test. */ |
508 | + width -= 2 | 404 | + |
509 | + | 405 | +static Coroutine *to_schedule[NUM_CONTEXTS]; |
510 | + self._prog = prog | 406 | + |
511 | + self._indent_increment = indent_increment | 407 | +static bool now_stopping; |
512 | + self._max_help_position = max_help_position | 408 | + |
513 | + self._width = width | 409 | +static int count_retry; |
514 | + | 410 | +static int count_here; |
515 | + self._current_indent = 0 | 411 | +static int count_other; |
516 | + self._level = 0 | 412 | + |
517 | + self._action_max_length = 0 | 413 | +static bool schedule_next(int n) |
518 | + | 414 | +{ |
519 | + self._root_section = self._Section(self, None) | 415 | + Coroutine *co; |
520 | + self._current_section = self._root_section | 416 | + |
521 | + | 417 | + co = atomic_xchg(&to_schedule[n], NULL); |
522 | + self._whitespace_matcher = _re.compile(r'\s+') | 418 | + if (!co) { |
523 | + self._long_break_matcher = _re.compile(r'\n\n\n+') | 419 | + atomic_inc(&count_retry); |
524 | + | 420 | + return false; |
525 | + # =============================== | 421 | + } |
526 | + # Section and indentation methods | 422 | + |
527 | + # =============================== | 423 | + if (n == id) { |
528 | + def _indent(self): | 424 | + atomic_inc(&count_here); |
529 | + self._current_indent += self._indent_increment | 425 | + } else { |
530 | + self._level += 1 | 426 | + atomic_inc(&count_other); |
531 | + | 427 | + } |
532 | + def _dedent(self): | 428 | + |
533 | + self._current_indent -= self._indent_increment | 429 | + aio_co_schedule(ctx[n], co); |
534 | + assert self._current_indent >= 0, 'Indent decreased below 0.' | 430 | + return true; |
535 | + self._level -= 1 | 431 | +} |
536 | + | 432 | + |
537 | + class _Section(object): | 433 | +static void finish_cb(void *opaque) |
538 | + | 434 | +{ |
539 | + def __init__(self, formatter, parent, heading=None): | 435 | + schedule_next(id); |
540 | + self.formatter = formatter | 436 | +} |
541 | + self.parent = parent | 437 | + |
542 | + self.heading = heading | 438 | +static coroutine_fn void test_multi_co_schedule_entry(void *opaque) |
543 | + self.items = [] | 439 | +{ |
544 | + | 440 | + g_assert(to_schedule[id] == NULL); |
545 | + def format_help(self): | 441 | + atomic_mb_set(&to_schedule[id], qemu_coroutine_self()); |
546 | + # format the indented section | 442 | + |
547 | + if self.parent is not None: | 443 | + while (!atomic_mb_read(&now_stopping)) { |
548 | + self.formatter._indent() | 444 | + int n; |
549 | + join = self.formatter._join_parts | 445 | + |
550 | + for func, args in self.items: | 446 | + n = g_test_rand_int_range(0, NUM_CONTEXTS); |
551 | + func(*args) | 447 | + schedule_next(n); |
552 | + item_help = join([func(*args) for func, args in self.items]) | 448 | + qemu_coroutine_yield(); |
553 | + if self.parent is not None: | 449 | + |
554 | + self.formatter._dedent() | 450 | + g_assert(to_schedule[id] == NULL); |
555 | + | 451 | + atomic_mb_set(&to_schedule[id], qemu_coroutine_self()); |
556 | + # return nothing if the section was empty | 452 | + } |
557 | + if not item_help: | 453 | +} |
558 | + return '' | 454 | + |
559 | + | 455 | + |
560 | + # add the heading if the section was non-empty | 456 | +static void test_multi_co_schedule(int seconds) |
561 | + if self.heading is not SUPPRESS and self.heading is not None: | 457 | +{ |
562 | + current_indent = self.formatter._current_indent | 458 | + int i; |
563 | + heading = '%*s%s:\n' % (current_indent, '', self.heading) | 459 | + |
564 | + else: | 460 | + count_here = count_other = count_retry = 0; |
565 | + heading = '' | 461 | + now_stopping = false; |
566 | + | 462 | + |
567 | + # join the section-initial newline, the heading and the help | 463 | + create_aio_contexts(); |
568 | + return join(['\n', heading, item_help, '\n']) | 464 | + for (i = 0; i < NUM_CONTEXTS; i++) { |
569 | + | 465 | + Coroutine *co1 = qemu_coroutine_create(test_multi_co_schedule_entry, NULL); |
570 | + def _add_item(self, func, args): | 466 | + aio_co_schedule(ctx[i], co1); |
571 | + self._current_section.items.append((func, args)) | 467 | + } |
572 | + | 468 | + |
573 | + # ======================== | 469 | + g_usleep(seconds * 1000000); |
574 | + # Message building methods | 470 | + |
575 | + # ======================== | 471 | + atomic_mb_set(&now_stopping, true); |
576 | + def start_section(self, heading): | 472 | + for (i = 0; i < NUM_CONTEXTS; i++) { |
577 | + self._indent() | 473 | + ctx_run(i, finish_cb, NULL); |
578 | + section = self._Section(self, self._current_section, heading) | 474 | + to_schedule[i] = NULL; |
579 | + self._add_item(section.format_help, []) | 475 | + } |
580 | + self._current_section = section | 476 | + |
581 | + | 477 | + join_aio_contexts(); |
582 | + def end_section(self): | 478 | + g_test_message("scheduled %d, queued %d, retry %d, total %d\n", |
583 | + self._current_section = self._current_section.parent | 479 | + count_other, count_here, count_retry, |
584 | + self._dedent() | 480 | + count_here + count_other + count_retry); |
585 | + | 481 | +} |
586 | + def add_text(self, text): | 482 | + |
587 | + if text is not SUPPRESS and text is not None: | 483 | +static void test_multi_co_schedule_1(void) |
588 | + self._add_item(self._format_text, [text]) | 484 | +{ |
589 | + | 485 | + test_multi_co_schedule(1); |
590 | + def add_usage(self, usage, actions, groups, prefix=None): | 486 | +} |
591 | + if usage is not SUPPRESS: | 487 | + |
592 | + args = usage, actions, groups, prefix | 488 | +static void test_multi_co_schedule_10(void) |
593 | + self._add_item(self._format_usage, args) | 489 | +{ |
594 | + | 490 | + test_multi_co_schedule(10); |
595 | + def add_argument(self, action): | 491 | +} |
596 | + if action.help is not SUPPRESS: | 492 | + |
597 | + | 493 | +/* End of tests. */ |
598 | + # find all invocations | 494 | + |
599 | + get_invocation = self._format_action_invocation | 495 | +int main(int argc, char **argv) |
600 | + invocations = [get_invocation(action)] | 496 | +{ |
601 | + for subaction in self._iter_indented_subactions(action): | 497 | + init_clocks(); |
602 | + invocations.append(get_invocation(subaction)) | 498 | + |
603 | + | 499 | + g_test_init(&argc, &argv, NULL); |
604 | + # update the maximum item length | 500 | + g_test_add_func("/aio/multi/lifecycle", test_lifecycle); |
605 | + invocation_length = max([len(s) for s in invocations]) | 501 | + if (g_test_quick()) { |
606 | + action_length = invocation_length + self._current_indent | 502 | + g_test_add_func("/aio/multi/schedule", test_multi_co_schedule_1); |
607 | + self._action_max_length = max(self._action_max_length, | 503 | + } else { |
608 | + action_length) | 504 | + g_test_add_func("/aio/multi/schedule", test_multi_co_schedule_10); |
609 | + | 505 | + } |
610 | + # add the item to the list | 506 | + return g_test_run(); |
611 | + self._add_item(self._format_action, [action]) | 507 | +} |
612 | + | 508 | diff --git a/util/async.c b/util/async.c |
613 | + def add_arguments(self, actions): | 509 | index XXXXXXX..XXXXXXX 100644 |
614 | + for action in actions: | 510 | --- a/util/async.c |
615 | + self.add_argument(action) | 511 | +++ b/util/async.c |
616 | + | 512 | @@ -XXX,XX +XXX,XX @@ |
617 | + # ======================= | 513 | #include "qemu/main-loop.h" |
618 | + # Help-formatting methods | 514 | #include "qemu/atomic.h" |
619 | + # ======================= | 515 | #include "block/raw-aio.h" |
620 | + def format_help(self): | 516 | +#include "qemu/coroutine_int.h" |
621 | + help = self._root_section.format_help() | 517 | +#include "trace.h" |
622 | + if help: | 518 | |
623 | + help = self._long_break_matcher.sub('\n\n', help) | 519 | /***********************************************************/ |
624 | + help = help.strip('\n') + '\n' | 520 | /* bottom halves (can be seen as timers which expire ASAP) */ |
625 | + return help | 521 | @@ -XXX,XX +XXX,XX @@ aio_ctx_finalize(GSource *source) |
626 | + | 522 | } |
627 | + def _join_parts(self, part_strings): | 523 | #endif |
628 | + return ''.join([part | 524 | |
629 | + for part in part_strings | 525 | + assert(QSLIST_EMPTY(&ctx->scheduled_coroutines)); |
630 | + if part and part is not SUPPRESS]) | 526 | + qemu_bh_delete(ctx->co_schedule_bh); |
631 | + | 527 | + |
632 | + def _format_usage(self, usage, actions, groups, prefix): | 528 | qemu_lockcnt_lock(&ctx->list_lock); |
633 | + if prefix is None: | 529 | assert(!qemu_lockcnt_count(&ctx->list_lock)); |
634 | + prefix = _('usage: ') | 530 | while (ctx->first_bh) { |
635 | + | 531 | @@ -XXX,XX +XXX,XX @@ static bool event_notifier_poll(void *opaque) |
636 | + # if usage is specified, use that | 532 | return atomic_read(&ctx->notified); |
637 | + if usage is not None: | 533 | } |
638 | + usage = usage % dict(prog=self._prog) | 534 | |
639 | + | 535 | +static void co_schedule_bh_cb(void *opaque) |
640 | + # if no optionals or positionals are available, usage is just prog | 536 | +{ |
641 | + elif usage is None and not actions: | 537 | + AioContext *ctx = opaque; |
642 | + usage = '%(prog)s' % dict(prog=self._prog) | 538 | + QSLIST_HEAD(, Coroutine) straight, reversed; |
643 | + | 539 | + |
644 | + # if optionals and positionals are available, calculate usage | 540 | + QSLIST_MOVE_ATOMIC(&reversed, &ctx->scheduled_coroutines); |
645 | + elif usage is None: | 541 | + QSLIST_INIT(&straight); |
646 | + prog = '%(prog)s' % dict(prog=self._prog) | 542 | + |
647 | + | 543 | + while (!QSLIST_EMPTY(&reversed)) { |
648 | + # split optionals from positionals | 544 | + Coroutine *co = QSLIST_FIRST(&reversed); |
649 | + optionals = [] | 545 | + QSLIST_REMOVE_HEAD(&reversed, co_scheduled_next); |
650 | + positionals = [] | 546 | + QSLIST_INSERT_HEAD(&straight, co, co_scheduled_next); |
651 | + for action in actions: | 547 | + } |
652 | + if action.option_strings: | 548 | + |
653 | + optionals.append(action) | 549 | + while (!QSLIST_EMPTY(&straight)) { |
654 | + else: | 550 | + Coroutine *co = QSLIST_FIRST(&straight); |
655 | + positionals.append(action) | 551 | + QSLIST_REMOVE_HEAD(&straight, co_scheduled_next); |
656 | + | 552 | + trace_aio_co_schedule_bh_cb(ctx, co); |
657 | + # build full usage string | 553 | + qemu_coroutine_enter(co); |
658 | + format = self._format_actions_usage | 554 | + } |
659 | + action_usage = format(optionals + positionals, groups) | 555 | +} |
660 | + usage = ' '.join([s for s in [prog, action_usage] if s]) | 556 | + |
661 | + | 557 | AioContext *aio_context_new(Error **errp) |
662 | + # wrap the usage parts if it's too long | 558 | { |
663 | + text_width = self._width - self._current_indent | 559 | int ret; |
664 | + if len(prefix) + len(usage) > text_width: | 560 | @@ -XXX,XX +XXX,XX @@ AioContext *aio_context_new(Error **errp) |
665 | + | 561 | } |
666 | + # break usage into wrappable parts | 562 | g_source_set_can_recurse(&ctx->source, true); |
667 | + part_regexp = r'\(.*?\)+|\[.*?\]+|\S+' | 563 | qemu_lockcnt_init(&ctx->list_lock); |
668 | + opt_usage = format(optionals, groups) | 564 | + |
669 | + pos_usage = format(positionals, groups) | 565 | + ctx->co_schedule_bh = aio_bh_new(ctx, co_schedule_bh_cb, ctx); |
670 | + opt_parts = _re.findall(part_regexp, opt_usage) | 566 | + QSLIST_INIT(&ctx->scheduled_coroutines); |
671 | + pos_parts = _re.findall(part_regexp, pos_usage) | 567 | + |
672 | + assert ' '.join(opt_parts) == opt_usage | 568 | aio_set_event_notifier(ctx, &ctx->notifier, |
673 | + assert ' '.join(pos_parts) == pos_usage | 569 | false, |
674 | + | 570 | (EventNotifierHandler *) |
675 | + # helper for wrapping lines | 571 | @@ -XXX,XX +XXX,XX @@ fail: |
676 | + def get_lines(parts, indent, prefix=None): | 572 | return NULL; |
677 | + lines = [] | 573 | } |
678 | + line = [] | 574 | |
679 | + if prefix is not None: | 575 | +void aio_co_schedule(AioContext *ctx, Coroutine *co) |
680 | + line_len = len(prefix) - 1 | 576 | +{ |
681 | + else: | 577 | + trace_aio_co_schedule(ctx, co); |
682 | + line_len = len(indent) - 1 | 578 | + QSLIST_INSERT_HEAD_ATOMIC(&ctx->scheduled_coroutines, |
683 | + for part in parts: | 579 | + co, co_scheduled_next); |
684 | + if line_len + 1 + len(part) > text_width: | 580 | + qemu_bh_schedule(ctx->co_schedule_bh); |
685 | + lines.append(indent + ' '.join(line)) | 581 | +} |
686 | + line = [] | 582 | + |
687 | + line_len = len(indent) - 1 | 583 | +void aio_co_wake(struct Coroutine *co) |
688 | + line.append(part) | 584 | +{ |
689 | + line_len += len(part) + 1 | 585 | + AioContext *ctx; |
690 | + if line: | 586 | + |
691 | + lines.append(indent + ' '.join(line)) | 587 | + /* Read coroutine before co->ctx. Matches smp_wmb in |
692 | + if prefix is not None: | 588 | + * qemu_coroutine_enter. |
693 | + lines[0] = lines[0][len(indent):] | 589 | + */ |
694 | + return lines | 590 | + smp_read_barrier_depends(); |
695 | + | 591 | + ctx = atomic_read(&co->ctx); |
696 | + # if prog is short, follow it with optionals or positionals | 592 | + |
697 | + if len(prefix) + len(prog) <= 0.75 * text_width: | 593 | + if (ctx != qemu_get_current_aio_context()) { |
698 | + indent = ' ' * (len(prefix) + len(prog) + 1) | 594 | + aio_co_schedule(ctx, co); |
699 | + if opt_parts: | 595 | + return; |
700 | + lines = get_lines([prog] + opt_parts, indent, prefix) | 596 | + } |
701 | + lines.extend(get_lines(pos_parts, indent)) | 597 | + |
702 | + elif pos_parts: | 598 | + if (qemu_in_coroutine()) { |
703 | + lines = get_lines([prog] + pos_parts, indent, prefix) | 599 | + Coroutine *self = qemu_coroutine_self(); |
704 | + else: | 600 | + assert(self != co); |
705 | + lines = [prog] | 601 | + QSIMPLEQ_INSERT_TAIL(&self->co_queue_wakeup, co, co_queue_next); |
706 | + | 602 | + } else { |
707 | + # if prog is long, put it on its own line | 603 | + aio_context_acquire(ctx); |
708 | + else: | 604 | + qemu_coroutine_enter(co); |
709 | + indent = ' ' * len(prefix) | 605 | + aio_context_release(ctx); |
710 | + parts = opt_parts + pos_parts | 606 | + } |
711 | + lines = get_lines(parts, indent) | 607 | +} |
712 | + if len(lines) > 1: | 608 | + |
713 | + lines = [] | 609 | void aio_context_ref(AioContext *ctx) |
714 | + lines.extend(get_lines(opt_parts, indent)) | 610 | { |
715 | + lines.extend(get_lines(pos_parts, indent)) | 611 | g_source_ref(&ctx->source); |
716 | + lines = [prog] + lines | 612 | diff --git a/util/qemu-coroutine.c b/util/qemu-coroutine.c |
717 | + | 613 | index XXXXXXX..XXXXXXX 100644 |
718 | + # join lines into usage | 614 | --- a/util/qemu-coroutine.c |
719 | + usage = '\n'.join(lines) | 615 | +++ b/util/qemu-coroutine.c |
720 | + | 616 | @@ -XXX,XX +XXX,XX @@ |
721 | + # prefix with 'usage:' | 617 | #include "qemu/atomic.h" |
722 | + return '%s%s\n\n' % (prefix, usage) | 618 | #include "qemu/coroutine.h" |
723 | + | 619 | #include "qemu/coroutine_int.h" |
724 | + def _format_actions_usage(self, actions, groups): | 620 | +#include "block/aio.h" |
725 | + # find group indices and identify actions in groups | 621 | |
726 | + group_actions = set() | 622 | enum { |
727 | + inserts = {} | 623 | POOL_BATCH_SIZE = 64, |
728 | + for group in groups: | 624 | @@ -XXX,XX +XXX,XX @@ void qemu_coroutine_enter(Coroutine *co) |
729 | + try: | 625 | } |
730 | + start = actions.index(group._group_actions[0]) | 626 | |
731 | + except ValueError: | 627 | co->caller = self; |
732 | + continue | 628 | + co->ctx = qemu_get_current_aio_context(); |
733 | + else: | 629 | + |
734 | + end = start + len(group._group_actions) | 630 | + /* Store co->ctx before anything that stores co. Matches |
735 | + if actions[start:end] == group._group_actions: | 631 | + * barrier in aio_co_wake. |
736 | + for action in group._group_actions: | 632 | + */ |
737 | + group_actions.add(action) | 633 | + smp_wmb(); |
738 | + if not group.required: | 634 | + |
739 | + if start in inserts: | 635 | ret = qemu_coroutine_switch(self, co, COROUTINE_ENTER); |
740 | + inserts[start] += ' [' | 636 | |
741 | + else: | 637 | qemu_co_queue_run_restart(co); |
742 | + inserts[start] = '[' | 638 | diff --git a/util/trace-events b/util/trace-events |
743 | + inserts[end] = ']' | 639 | index XXXXXXX..XXXXXXX 100644 |
744 | + else: | 640 | --- a/util/trace-events |
745 | + if start in inserts: | 641 | +++ b/util/trace-events |
746 | + inserts[start] += ' (' | 642 | @@ -XXX,XX +XXX,XX @@ run_poll_handlers_end(void *ctx, bool progress) "ctx %p progress %d" |
747 | + else: | 643 | poll_shrink(void *ctx, int64_t old, int64_t new) "ctx %p old %"PRId64" new %"PRId64 |
748 | + inserts[start] = '(' | 644 | poll_grow(void *ctx, int64_t old, int64_t new) "ctx %p old %"PRId64" new %"PRId64 |
749 | + inserts[end] = ')' | 645 | |
750 | + for i in range(start + 1, end): | 646 | +# util/async.c |
751 | + inserts[i] = '|' | 647 | +aio_co_schedule(void *ctx, void *co) "ctx %p co %p" |
752 | + | 648 | +aio_co_schedule_bh_cb(void *ctx, void *co) "ctx %p co %p" |
753 | + # collect all actions format strings | 649 | + |
754 | + parts = [] | 650 | # util/thread-pool.c |
755 | + for i, action in enumerate(actions): | 651 | thread_pool_submit(void *pool, void *req, void *opaque) "pool %p req %p opaque %p" |
756 | + | 652 | thread_pool_complete(void *pool, void *req, void *opaque, int ret) "pool %p req %p opaque %p ret %d" |
757 | + # suppressed arguments are marked with None | ||
758 | + # remove | separators for suppressed arguments | ||
759 | + if action.help is SUPPRESS: | ||
760 | + parts.append(None) | ||
761 | + if inserts.get(i) == '|': | ||
762 | + inserts.pop(i) | ||
763 | + elif inserts.get(i + 1) == '|': | ||
764 | + inserts.pop(i + 1) | ||
765 | + | ||
766 | + # produce all arg strings | ||
767 | + elif not action.option_strings: | ||
768 | + part = self._format_args(action, action.dest) | ||
769 | + | ||
770 | + # if it's in a group, strip the outer [] | ||
771 | + if action in group_actions: | ||
772 | + if part[0] == '[' and part[-1] == ']': | ||
773 | + part = part[1:-1] | ||
774 | + | ||
775 | + # add the action string to the list | ||
776 | + parts.append(part) | ||
777 | + | ||
778 | + # produce the first way to invoke the option in brackets | ||
779 | + else: | ||
780 | + option_string = action.option_strings[0] | ||
781 | + | ||
782 | + # if the Optional doesn't take a value, format is: | ||
783 | + # -s or --long | ||
784 | + if action.nargs == 0: | ||
785 | + part = '%s' % option_string | ||
786 | + | ||
787 | + # if the Optional takes a value, format is: | ||
788 | + # -s ARGS or --long ARGS | ||
789 | + else: | ||
790 | + default = action.dest.upper() | ||
791 | + args_string = self._format_args(action, default) | ||
792 | + part = '%s %s' % (option_string, args_string) | ||
793 | + | ||
794 | + # make it look optional if it's not required or in a group | ||
795 | + if not action.required and action not in group_actions: | ||
796 | + part = '[%s]' % part | ||
797 | + | ||
798 | + # add the action string to the list | ||
799 | + parts.append(part) | ||
800 | + | ||
801 | + # insert things at the necessary indices | ||
802 | + for i in sorted(inserts, reverse=True): | ||
803 | + parts[i:i] = [inserts[i]] | ||
804 | + | ||
805 | + # join all the action items with spaces | ||
806 | + text = ' '.join([item for item in parts if item is not None]) | ||
807 | + | ||
808 | + # clean up separators for mutually exclusive groups | ||
809 | + open = r'[\[(]' | ||
810 | + close = r'[\])]' | ||
811 | + text = _re.sub(r'(%s) ' % open, r'\1', text) | ||
812 | + text = _re.sub(r' (%s)' % close, r'\1', text) | ||
813 | + text = _re.sub(r'%s *%s' % (open, close), r'', text) | ||
814 | + text = _re.sub(r'\(([^|]*)\)', r'\1', text) | ||
815 | + text = text.strip() | ||
816 | + | ||
817 | + # return the text | ||
818 | + return text | ||
819 | + | ||
820 | + def _format_text(self, text): | ||
821 | + if '%(prog)' in text: | ||
822 | + text = text % dict(prog=self._prog) | ||
823 | + text_width = self._width - self._current_indent | ||
824 | + indent = ' ' * self._current_indent | ||
825 | + return self._fill_text(text, text_width, indent) + '\n\n' | ||
826 | + | ||
827 | + def _format_action(self, action): | ||
828 | + # determine the required width and the entry label | ||
829 | + help_position = min(self._action_max_length + 2, | ||
830 | + self._max_help_position) | ||
831 | + help_width = self._width - help_position | ||
832 | + action_width = help_position - self._current_indent - 2 | ||
833 | + action_header = self._format_action_invocation(action) | ||
834 | + | ||
835 | + # ho nelp; start on same line and add a final newline | ||
836 | + if not action.help: | ||
837 | + tup = self._current_indent, '', action_header | ||
838 | + action_header = '%*s%s\n' % tup | ||
839 | + | ||
840 | + # short action name; start on the same line and pad two spaces | ||
841 | + elif len(action_header) <= action_width: | ||
842 | + tup = self._current_indent, '', action_width, action_header | ||
843 | + action_header = '%*s%-*s ' % tup | ||
844 | + indent_first = 0 | ||
845 | + | ||
846 | + # long action name; start on the next line | ||
847 | + else: | ||
848 | + tup = self._current_indent, '', action_header | ||
849 | + action_header = '%*s%s\n' % tup | ||
850 | + indent_first = help_position | ||
851 | + | ||
852 | + # collect the pieces of the action help | ||
853 | + parts = [action_header] | ||
854 | + | ||
855 | + # if there was help for the action, add lines of help text | ||
856 | + if action.help: | ||
857 | + help_text = self._expand_help(action) | ||
858 | + help_lines = self._split_lines(help_text, help_width) | ||
859 | + parts.append('%*s%s\n' % (indent_first, '', help_lines[0])) | ||
860 | + for line in help_lines[1:]: | ||
861 | + parts.append('%*s%s\n' % (help_position, '', line)) | ||
862 | + | ||
863 | + # or add a newline if the description doesn't end with one | ||
864 | + elif not action_header.endswith('\n'): | ||
865 | + parts.append('\n') | ||
866 | + | ||
867 | + # if there are any sub-actions, add their help as well | ||
868 | + for subaction in self._iter_indented_subactions(action): | ||
869 | + parts.append(self._format_action(subaction)) | ||
870 | + | ||
871 | + # return a single string | ||
872 | + return self._join_parts(parts) | ||
873 | + | ||
874 | + def _format_action_invocation(self, action): | ||
875 | + if not action.option_strings: | ||
876 | + metavar, = self._metavar_formatter(action, action.dest)(1) | ||
877 | + return metavar | ||
878 | + | ||
879 | + else: | ||
880 | + parts = [] | ||
881 | + | ||
882 | + # if the Optional doesn't take a value, format is: | ||
883 | + # -s, --long | ||
884 | + if action.nargs == 0: | ||
885 | + parts.extend(action.option_strings) | ||
886 | + | ||
887 | + # if the Optional takes a value, format is: | ||
888 | + # -s ARGS, --long ARGS | ||
889 | + else: | ||
890 | + default = action.dest.upper() | ||
891 | + args_string = self._format_args(action, default) | ||
892 | + for option_string in action.option_strings: | ||
893 | + parts.append('%s %s' % (option_string, args_string)) | ||
894 | + | ||
895 | + return ', '.join(parts) | ||
896 | + | ||
897 | + def _metavar_formatter(self, action, default_metavar): | ||
898 | + if action.metavar is not None: | ||
899 | + result = action.metavar | ||
900 | + elif action.choices is not None: | ||
901 | + choice_strs = [str(choice) for choice in action.choices] | ||
902 | + result = '{%s}' % ','.join(choice_strs) | ||
903 | + else: | ||
904 | + result = default_metavar | ||
905 | + | ||
906 | + def format(tuple_size): | ||
907 | + if isinstance(result, tuple): | ||
908 | + return result | ||
909 | + else: | ||
910 | + return (result, ) * tuple_size | ||
911 | + return format | ||
912 | + | ||
913 | + def _format_args(self, action, default_metavar): | ||
914 | + get_metavar = self._metavar_formatter(action, default_metavar) | ||
915 | + if action.nargs is None: | ||
916 | + result = '%s' % get_metavar(1) | ||
917 | + elif action.nargs == OPTIONAL: | ||
918 | + result = '[%s]' % get_metavar(1) | ||
919 | + elif action.nargs == ZERO_OR_MORE: | ||
920 | + result = '[%s [%s ...]]' % get_metavar(2) | ||
921 | + elif action.nargs == ONE_OR_MORE: | ||
922 | + result = '%s [%s ...]' % get_metavar(2) | ||
923 | + elif action.nargs == REMAINDER: | ||
924 | + result = '...' | ||
925 | + elif action.nargs == PARSER: | ||
926 | + result = '%s ...' % get_metavar(1) | ||
927 | + else: | ||
928 | + formats = ['%s' for _ in range(action.nargs)] | ||
929 | + result = ' '.join(formats) % get_metavar(action.nargs) | ||
930 | + return result | ||
931 | + | ||
932 | + def _expand_help(self, action): | ||
933 | + params = dict(vars(action), prog=self._prog) | ||
934 | + for name in list(params): | ||
935 | + if params[name] is SUPPRESS: | ||
936 | + del params[name] | ||
937 | + for name in list(params): | ||
938 | + if hasattr(params[name], '__name__'): | ||
939 | + params[name] = params[name].__name__ | ||
940 | + if params.get('choices') is not None: | ||
941 | + choices_str = ', '.join([str(c) for c in params['choices']]) | ||
942 | + params['choices'] = choices_str | ||
943 | + return self._get_help_string(action) % params | ||
944 | + | ||
945 | + def _iter_indented_subactions(self, action): | ||
946 | + try: | ||
947 | + get_subactions = action._get_subactions | ||
948 | + except AttributeError: | ||
949 | + pass | ||
950 | + else: | ||
951 | + self._indent() | ||
952 | + for subaction in get_subactions(): | ||
953 | + yield subaction | ||
954 | + self._dedent() | ||
955 | + | ||
956 | + def _split_lines(self, text, width): | ||
957 | + text = self._whitespace_matcher.sub(' ', text).strip() | ||
958 | + return _textwrap.wrap(text, width) | ||
959 | + | ||
960 | + def _fill_text(self, text, width, indent): | ||
961 | + text = self._whitespace_matcher.sub(' ', text).strip() | ||
962 | + return _textwrap.fill(text, width, initial_indent=indent, | ||
963 | + subsequent_indent=indent) | ||
964 | + | ||
965 | + def _get_help_string(self, action): | ||
966 | + return action.help | ||
967 | + | ||
968 | + | ||
969 | +class RawDescriptionHelpFormatter(HelpFormatter): | ||
970 | + """Help message formatter which retains any formatting in descriptions. | ||
971 | + | ||
972 | + Only the name of this class is considered a public API. All the methods | ||
973 | + provided by the class are considered an implementation detail. | ||
974 | + """ | ||
975 | + | ||
976 | + def _fill_text(self, text, width, indent): | ||
977 | + return ''.join([indent + line for line in text.splitlines(True)]) | ||
978 | + | ||
979 | + | ||
980 | +class RawTextHelpFormatter(RawDescriptionHelpFormatter): | ||
981 | + """Help message formatter which retains formatting of all help text. | ||
982 | + | ||
983 | + Only the name of this class is considered a public API. All the methods | ||
984 | + provided by the class are considered an implementation detail. | ||
985 | + """ | ||
986 | + | ||
987 | + def _split_lines(self, text, width): | ||
988 | + return text.splitlines() | ||
989 | + | ||
990 | + | ||
991 | +class ArgumentDefaultsHelpFormatter(HelpFormatter): | ||
992 | + """Help message formatter which adds default values to argument help. | ||
993 | + | ||
994 | + Only the name of this class is considered a public API. All the methods | ||
995 | + provided by the class are considered an implementation detail. | ||
996 | + """ | ||
997 | + | ||
998 | + def _get_help_string(self, action): | ||
999 | + help = action.help | ||
1000 | + if '%(default)' not in action.help: | ||
1001 | + if action.default is not SUPPRESS: | ||
1002 | + defaulting_nargs = [OPTIONAL, ZERO_OR_MORE] | ||
1003 | + if action.option_strings or action.nargs in defaulting_nargs: | ||
1004 | + help += ' (default: %(default)s)' | ||
1005 | + return help | ||
1006 | + | ||
1007 | + | ||
1008 | +# ===================== | ||
1009 | +# Options and Arguments | ||
1010 | +# ===================== | ||
1011 | + | ||
1012 | +def _get_action_name(argument): | ||
1013 | + if argument is None: | ||
1014 | + return None | ||
1015 | + elif argument.option_strings: | ||
1016 | + return '/'.join(argument.option_strings) | ||
1017 | + elif argument.metavar not in (None, SUPPRESS): | ||
1018 | + return argument.metavar | ||
1019 | + elif argument.dest not in (None, SUPPRESS): | ||
1020 | + return argument.dest | ||
1021 | + else: | ||
1022 | + return None | ||
1023 | + | ||
1024 | + | ||
1025 | +class ArgumentError(Exception): | ||
1026 | + """An error from creating or using an argument (optional or positional). | ||
1027 | + | ||
1028 | + The string value of this exception is the message, augmented with | ||
1029 | + information about the argument that caused it. | ||
1030 | + """ | ||
1031 | + | ||
1032 | + def __init__(self, argument, message): | ||
1033 | + self.argument_name = _get_action_name(argument) | ||
1034 | + self.message = message | ||
1035 | + | ||
1036 | + def __str__(self): | ||
1037 | + if self.argument_name is None: | ||
1038 | + format = '%(message)s' | ||
1039 | + else: | ||
1040 | + format = 'argument %(argument_name)s: %(message)s' | ||
1041 | + return format % dict(message=self.message, | ||
1042 | + argument_name=self.argument_name) | ||
1043 | + | ||
1044 | + | ||
1045 | +class ArgumentTypeError(Exception): | ||
1046 | + """An error from trying to convert a command line string to a type.""" | ||
1047 | + pass | ||
1048 | + | ||
1049 | + | ||
1050 | +# ============== | ||
1051 | +# Action classes | ||
1052 | +# ============== | ||
1053 | + | ||
1054 | +class Action(_AttributeHolder): | ||
1055 | + """Information about how to convert command line strings to Python objects. | ||
1056 | + | ||
1057 | + Action objects are used by an ArgumentParser to represent the information | ||
1058 | + needed to parse a single argument from one or more strings from the | ||
1059 | + command line. The keyword arguments to the Action constructor are also | ||
1060 | + all attributes of Action instances. | ||
1061 | + | ||
1062 | + Keyword Arguments: | ||
1063 | + | ||
1064 | + - option_strings -- A list of command-line option strings which | ||
1065 | + should be associated with this action. | ||
1066 | + | ||
1067 | + - dest -- The name of the attribute to hold the created object(s) | ||
1068 | + | ||
1069 | + - nargs -- The number of command-line arguments that should be | ||
1070 | + consumed. By default, one argument will be consumed and a single | ||
1071 | + value will be produced. Other values include: | ||
1072 | + - N (an integer) consumes N arguments (and produces a list) | ||
1073 | + - '?' consumes zero or one arguments | ||
1074 | + - '*' consumes zero or more arguments (and produces a list) | ||
1075 | + - '+' consumes one or more arguments (and produces a list) | ||
1076 | + Note that the difference between the default and nargs=1 is that | ||
1077 | + with the default, a single value will be produced, while with | ||
1078 | + nargs=1, a list containing a single value will be produced. | ||
1079 | + | ||
1080 | + - const -- The value to be produced if the option is specified and the | ||
1081 | + option uses an action that takes no values. | ||
1082 | + | ||
1083 | + - default -- The value to be produced if the option is not specified. | ||
1084 | + | ||
1085 | + - type -- The type which the command-line arguments should be converted | ||
1086 | + to, should be one of 'string', 'int', 'float', 'complex' or a | ||
1087 | + callable object that accepts a single string argument. If None, | ||
1088 | + 'string' is assumed. | ||
1089 | + | ||
1090 | + - choices -- A container of values that should be allowed. If not None, | ||
1091 | + after a command-line argument has been converted to the appropriate | ||
1092 | + type, an exception will be raised if it is not a member of this | ||
1093 | + collection. | ||
1094 | + | ||
1095 | + - required -- True if the action must always be specified at the | ||
1096 | + command line. This is only meaningful for optional command-line | ||
1097 | + arguments. | ||
1098 | + | ||
1099 | + - help -- The help string describing the argument. | ||
1100 | + | ||
1101 | + - metavar -- The name to be used for the option's argument with the | ||
1102 | + help string. If None, the 'dest' value will be used as the name. | ||
1103 | + """ | ||
1104 | + | ||
1105 | + def __init__(self, | ||
1106 | + option_strings, | ||
1107 | + dest, | ||
1108 | + nargs=None, | ||
1109 | + const=None, | ||
1110 | + default=None, | ||
1111 | + type=None, | ||
1112 | + choices=None, | ||
1113 | + required=False, | ||
1114 | + help=None, | ||
1115 | + metavar=None): | ||
1116 | + self.option_strings = option_strings | ||
1117 | + self.dest = dest | ||
1118 | + self.nargs = nargs | ||
1119 | + self.const = const | ||
1120 | + self.default = default | ||
1121 | + self.type = type | ||
1122 | + self.choices = choices | ||
1123 | + self.required = required | ||
1124 | + self.help = help | ||
1125 | + self.metavar = metavar | ||
1126 | + | ||
1127 | + def _get_kwargs(self): | ||
1128 | + names = [ | ||
1129 | + 'option_strings', | ||
1130 | + 'dest', | ||
1131 | + 'nargs', | ||
1132 | + 'const', | ||
1133 | + 'default', | ||
1134 | + 'type', | ||
1135 | + 'choices', | ||
1136 | + 'help', | ||
1137 | + 'metavar', | ||
1138 | + ] | ||
1139 | + return [(name, getattr(self, name)) for name in names] | ||
1140 | + | ||
1141 | + def __call__(self, parser, namespace, values, option_string=None): | ||
1142 | + raise NotImplementedError(_('.__call__() not defined')) | ||
1143 | + | ||
1144 | + | ||
1145 | +class _StoreAction(Action): | ||
1146 | + | ||
1147 | + def __init__(self, | ||
1148 | + option_strings, | ||
1149 | + dest, | ||
1150 | + nargs=None, | ||
1151 | + const=None, | ||
1152 | + default=None, | ||
1153 | + type=None, | ||
1154 | + choices=None, | ||
1155 | + required=False, | ||
1156 | + help=None, | ||
1157 | + metavar=None): | ||
1158 | + if nargs == 0: | ||
1159 | + raise ValueError('nargs for store actions must be > 0; if you ' | ||
1160 | + 'have nothing to store, actions such as store ' | ||
1161 | + 'true or store const may be more appropriate') | ||
1162 | + if const is not None and nargs != OPTIONAL: | ||
1163 | + raise ValueError('nargs must be %r to supply const' % OPTIONAL) | ||
1164 | + super(_StoreAction, self).__init__( | ||
1165 | + option_strings=option_strings, | ||
1166 | + dest=dest, | ||
1167 | + nargs=nargs, | ||
1168 | + const=const, | ||
1169 | + default=default, | ||
1170 | + type=type, | ||
1171 | + choices=choices, | ||
1172 | + required=required, | ||
1173 | + help=help, | ||
1174 | + metavar=metavar) | ||
1175 | + | ||
1176 | + def __call__(self, parser, namespace, values, option_string=None): | ||
1177 | + setattr(namespace, self.dest, values) | ||
1178 | + | ||
1179 | + | ||
1180 | +class _StoreConstAction(Action): | ||
1181 | + | ||
1182 | + def __init__(self, | ||
1183 | + option_strings, | ||
1184 | + dest, | ||
1185 | + const, | ||
1186 | + default=None, | ||
1187 | + required=False, | ||
1188 | + help=None, | ||
1189 | + metavar=None): | ||
1190 | + super(_StoreConstAction, self).__init__( | ||
1191 | + option_strings=option_strings, | ||
1192 | + dest=dest, | ||
1193 | + nargs=0, | ||
1194 | + const=const, | ||
1195 | + default=default, | ||
1196 | + required=required, | ||
1197 | + help=help) | ||
1198 | + | ||
1199 | + def __call__(self, parser, namespace, values, option_string=None): | ||
1200 | + setattr(namespace, self.dest, self.const) | ||
1201 | + | ||
1202 | + | ||
1203 | +class _StoreTrueAction(_StoreConstAction): | ||
1204 | + | ||
1205 | + def __init__(self, | ||
1206 | + option_strings, | ||
1207 | + dest, | ||
1208 | + default=False, | ||
1209 | + required=False, | ||
1210 | + help=None): | ||
1211 | + super(_StoreTrueAction, self).__init__( | ||
1212 | + option_strings=option_strings, | ||
1213 | + dest=dest, | ||
1214 | + const=True, | ||
1215 | + default=default, | ||
1216 | + required=required, | ||
1217 | + help=help) | ||
1218 | + | ||
1219 | + | ||
1220 | +class _StoreFalseAction(_StoreConstAction): | ||
1221 | + | ||
1222 | + def __init__(self, | ||
1223 | + option_strings, | ||
1224 | + dest, | ||
1225 | + default=True, | ||
1226 | + required=False, | ||
1227 | + help=None): | ||
1228 | + super(_StoreFalseAction, self).__init__( | ||
1229 | + option_strings=option_strings, | ||
1230 | + dest=dest, | ||
1231 | + const=False, | ||
1232 | + default=default, | ||
1233 | + required=required, | ||
1234 | + help=help) | ||
1235 | + | ||
1236 | + | ||
1237 | +class _AppendAction(Action): | ||
1238 | + | ||
1239 | + def __init__(self, | ||
1240 | + option_strings, | ||
1241 | + dest, | ||
1242 | + nargs=None, | ||
1243 | + const=None, | ||
1244 | + default=None, | ||
1245 | + type=None, | ||
1246 | + choices=None, | ||
1247 | + required=False, | ||
1248 | + help=None, | ||
1249 | + metavar=None): | ||
1250 | + if nargs == 0: | ||
1251 | + raise ValueError('nargs for append actions must be > 0; if arg ' | ||
1252 | + 'strings are not supplying the value to append, ' | ||
1253 | + 'the append const action may be more appropriate') | ||
1254 | + if const is not None and nargs != OPTIONAL: | ||
1255 | + raise ValueError('nargs must be %r to supply const' % OPTIONAL) | ||
1256 | + super(_AppendAction, self).__init__( | ||
1257 | + option_strings=option_strings, | ||
1258 | + dest=dest, | ||
1259 | + nargs=nargs, | ||
1260 | + const=const, | ||
1261 | + default=default, | ||
1262 | + type=type, | ||
1263 | + choices=choices, | ||
1264 | + required=required, | ||
1265 | + help=help, | ||
1266 | + metavar=metavar) | ||
1267 | + | ||
1268 | + def __call__(self, parser, namespace, values, option_string=None): | ||
1269 | + items = _copy.copy(_ensure_value(namespace, self.dest, [])) | ||
1270 | + items.append(values) | ||
1271 | + setattr(namespace, self.dest, items) | ||
1272 | + | ||
1273 | + | ||
1274 | +class _AppendConstAction(Action): | ||
1275 | + | ||
1276 | + def __init__(self, | ||
1277 | + option_strings, | ||
1278 | + dest, | ||
1279 | + const, | ||
1280 | + default=None, | ||
1281 | + required=False, | ||
1282 | + help=None, | ||
1283 | + metavar=None): | ||
1284 | + super(_AppendConstAction, self).__init__( | ||
1285 | + option_strings=option_strings, | ||
1286 | + dest=dest, | ||
1287 | + nargs=0, | ||
1288 | + const=const, | ||
1289 | + default=default, | ||
1290 | + required=required, | ||
1291 | + help=help, | ||
1292 | + metavar=metavar) | ||
1293 | + | ||
1294 | + def __call__(self, parser, namespace, values, option_string=None): | ||
1295 | + items = _copy.copy(_ensure_value(namespace, self.dest, [])) | ||
1296 | + items.append(self.const) | ||
1297 | + setattr(namespace, self.dest, items) | ||
1298 | + | ||
1299 | + | ||
1300 | +class _CountAction(Action): | ||
1301 | + | ||
1302 | + def __init__(self, | ||
1303 | + option_strings, | ||
1304 | + dest, | ||
1305 | + default=None, | ||
1306 | + required=False, | ||
1307 | + help=None): | ||
1308 | + super(_CountAction, self).__init__( | ||
1309 | + option_strings=option_strings, | ||
1310 | + dest=dest, | ||
1311 | + nargs=0, | ||
1312 | + default=default, | ||
1313 | + required=required, | ||
1314 | + help=help) | ||
1315 | + | ||
1316 | + def __call__(self, parser, namespace, values, option_string=None): | ||
1317 | + new_count = _ensure_value(namespace, self.dest, 0) + 1 | ||
1318 | + setattr(namespace, self.dest, new_count) | ||
1319 | + | ||
1320 | + | ||
1321 | +class _HelpAction(Action): | ||
1322 | + | ||
1323 | + def __init__(self, | ||
1324 | + option_strings, | ||
1325 | + dest=SUPPRESS, | ||
1326 | + default=SUPPRESS, | ||
1327 | + help=None): | ||
1328 | + super(_HelpAction, self).__init__( | ||
1329 | + option_strings=option_strings, | ||
1330 | + dest=dest, | ||
1331 | + default=default, | ||
1332 | + nargs=0, | ||
1333 | + help=help) | ||
1334 | + | ||
1335 | + def __call__(self, parser, namespace, values, option_string=None): | ||
1336 | + parser.print_help() | ||
1337 | + parser.exit() | ||
1338 | + | ||
1339 | + | ||
1340 | +class _VersionAction(Action): | ||
1341 | + | ||
1342 | + def __init__(self, | ||
1343 | + option_strings, | ||
1344 | + version=None, | ||
1345 | + dest=SUPPRESS, | ||
1346 | + default=SUPPRESS, | ||
1347 | + help="show program's version number and exit"): | ||
1348 | + super(_VersionAction, self).__init__( | ||
1349 | + option_strings=option_strings, | ||
1350 | + dest=dest, | ||
1351 | + default=default, | ||
1352 | + nargs=0, | ||
1353 | + help=help) | ||
1354 | + self.version = version | ||
1355 | + | ||
1356 | + def __call__(self, parser, namespace, values, option_string=None): | ||
1357 | + version = self.version | ||
1358 | + if version is None: | ||
1359 | + version = parser.version | ||
1360 | + formatter = parser._get_formatter() | ||
1361 | + formatter.add_text(version) | ||
1362 | + parser.exit(message=formatter.format_help()) | ||
1363 | + | ||
1364 | + | ||
1365 | +class _SubParsersAction(Action): | ||
1366 | + | ||
1367 | + class _ChoicesPseudoAction(Action): | ||
1368 | + | ||
1369 | + def __init__(self, name, aliases, help): | ||
1370 | + metavar = dest = name | ||
1371 | + if aliases: | ||
1372 | + metavar += ' (%s)' % ', '.join(aliases) | ||
1373 | + sup = super(_SubParsersAction._ChoicesPseudoAction, self) | ||
1374 | + sup.__init__(option_strings=[], dest=dest, help=help, | ||
1375 | + metavar=metavar) | ||
1376 | + | ||
1377 | + def __init__(self, | ||
1378 | + option_strings, | ||
1379 | + prog, | ||
1380 | + parser_class, | ||
1381 | + dest=SUPPRESS, | ||
1382 | + help=None, | ||
1383 | + metavar=None): | ||
1384 | + | ||
1385 | + self._prog_prefix = prog | ||
1386 | + self._parser_class = parser_class | ||
1387 | + self._name_parser_map = {} | ||
1388 | + self._choices_actions = [] | ||
1389 | + | ||
1390 | + super(_SubParsersAction, self).__init__( | ||
1391 | + option_strings=option_strings, | ||
1392 | + dest=dest, | ||
1393 | + nargs=PARSER, | ||
1394 | + choices=self._name_parser_map, | ||
1395 | + help=help, | ||
1396 | + metavar=metavar) | ||
1397 | + | ||
1398 | + def add_parser(self, name, **kwargs): | ||
1399 | + # set prog from the existing prefix | ||
1400 | + if kwargs.get('prog') is None: | ||
1401 | + kwargs['prog'] = '%s %s' % (self._prog_prefix, name) | ||
1402 | + | ||
1403 | + aliases = kwargs.pop('aliases', ()) | ||
1404 | + | ||
1405 | + # create a pseudo-action to hold the choice help | ||
1406 | + if 'help' in kwargs: | ||
1407 | + help = kwargs.pop('help') | ||
1408 | + choice_action = self._ChoicesPseudoAction(name, aliases, help) | ||
1409 | + self._choices_actions.append(choice_action) | ||
1410 | + | ||
1411 | + # create the parser and add it to the map | ||
1412 | + parser = self._parser_class(**kwargs) | ||
1413 | + self._name_parser_map[name] = parser | ||
1414 | + | ||
1415 | + # make parser available under aliases also | ||
1416 | + for alias in aliases: | ||
1417 | + self._name_parser_map[alias] = parser | ||
1418 | + | ||
1419 | + return parser | ||
1420 | + | ||
1421 | + def _get_subactions(self): | ||
1422 | + return self._choices_actions | ||
1423 | + | ||
1424 | + def __call__(self, parser, namespace, values, option_string=None): | ||
1425 | + parser_name = values[0] | ||
1426 | + arg_strings = values[1:] | ||
1427 | + | ||
1428 | + # set the parser name if requested | ||
1429 | + if self.dest is not SUPPRESS: | ||
1430 | + setattr(namespace, self.dest, parser_name) | ||
1431 | + | ||
1432 | + # select the parser | ||
1433 | + try: | ||
1434 | + parser = self._name_parser_map[parser_name] | ||
1435 | + except KeyError: | ||
1436 | + tup = parser_name, ', '.join(self._name_parser_map) | ||
1437 | + msg = _('unknown parser %r (choices: %s)' % tup) | ||
1438 | + raise ArgumentError(self, msg) | ||
1439 | + | ||
1440 | + # parse all the remaining options into the namespace | ||
1441 | + # store any unrecognized options on the object, so that the top | ||
1442 | + # level parser can decide what to do with them | ||
1443 | + namespace, arg_strings = parser.parse_known_args(arg_strings, namespace) | ||
1444 | + if arg_strings: | ||
1445 | + vars(namespace).setdefault(_UNRECOGNIZED_ARGS_ATTR, []) | ||
1446 | + getattr(namespace, _UNRECOGNIZED_ARGS_ATTR).extend(arg_strings) | ||
1447 | + | ||
1448 | + | ||
1449 | +# ============== | ||
1450 | +# Type classes | ||
1451 | +# ============== | ||
1452 | + | ||
1453 | +class FileType(object): | ||
1454 | + """Factory for creating file object types | ||
1455 | + | ||
1456 | + Instances of FileType are typically passed as type= arguments to the | ||
1457 | + ArgumentParser add_argument() method. | ||
1458 | + | ||
1459 | + Keyword Arguments: | ||
1460 | + - mode -- A string indicating how the file is to be opened. Accepts the | ||
1461 | + same values as the builtin open() function. | ||
1462 | + - bufsize -- The file's desired buffer size. Accepts the same values as | ||
1463 | + the builtin open() function. | ||
1464 | + """ | ||
1465 | + | ||
1466 | + def __init__(self, mode='r', bufsize=None): | ||
1467 | + self._mode = mode | ||
1468 | + self._bufsize = bufsize | ||
1469 | + | ||
1470 | + def __call__(self, string): | ||
1471 | + # the special argument "-" means sys.std{in,out} | ||
1472 | + if string == '-': | ||
1473 | + if 'r' in self._mode: | ||
1474 | + return _sys.stdin | ||
1475 | + elif 'w' in self._mode: | ||
1476 | + return _sys.stdout | ||
1477 | + else: | ||
1478 | + msg = _('argument "-" with mode %r' % self._mode) | ||
1479 | + raise ValueError(msg) | ||
1480 | + | ||
1481 | + try: | ||
1482 | + # all other arguments are used as file names | ||
1483 | + if self._bufsize: | ||
1484 | + return open(string, self._mode, self._bufsize) | ||
1485 | + else: | ||
1486 | + return open(string, self._mode) | ||
1487 | + except IOError: | ||
1488 | + err = _sys.exc_info()[1] | ||
1489 | + message = _("can't open '%s': %s") | ||
1490 | + raise ArgumentTypeError(message % (string, err)) | ||
1491 | + | ||
1492 | + def __repr__(self): | ||
1493 | + args = [self._mode, self._bufsize] | ||
1494 | + args_str = ', '.join([repr(arg) for arg in args if arg is not None]) | ||
1495 | + return '%s(%s)' % (type(self).__name__, args_str) | ||
1496 | + | ||
1497 | +# =========================== | ||
1498 | +# Optional and Positional Parsing | ||
1499 | +# =========================== | ||
1500 | + | ||
1501 | +class Namespace(_AttributeHolder): | ||
1502 | + """Simple object for storing attributes. | ||
1503 | + | ||
1504 | + Implements equality by attribute names and values, and provides a simple | ||
1505 | + string representation. | ||
1506 | + """ | ||
1507 | + | ||
1508 | + def __init__(self, **kwargs): | ||
1509 | + for name in kwargs: | ||
1510 | + setattr(self, name, kwargs[name]) | ||
1511 | + | ||
1512 | + __hash__ = None | ||
1513 | + | ||
1514 | + def __eq__(self, other): | ||
1515 | + return vars(self) == vars(other) | ||
1516 | + | ||
1517 | + def __ne__(self, other): | ||
1518 | + return not (self == other) | ||
1519 | + | ||
1520 | + def __contains__(self, key): | ||
1521 | + return key in self.__dict__ | ||
1522 | + | ||
1523 | + | ||
1524 | +class _ActionsContainer(object): | ||
1525 | + | ||
1526 | + def __init__(self, | ||
1527 | + description, | ||
1528 | + prefix_chars, | ||
1529 | + argument_default, | ||
1530 | + conflict_handler): | ||
1531 | + super(_ActionsContainer, self).__init__() | ||
1532 | + | ||
1533 | + self.description = description | ||
1534 | + self.argument_default = argument_default | ||
1535 | + self.prefix_chars = prefix_chars | ||
1536 | + self.conflict_handler = conflict_handler | ||
1537 | + | ||
1538 | + # set up registries | ||
1539 | + self._registries = {} | ||
1540 | + | ||
1541 | + # register actions | ||
1542 | + self.register('action', None, _StoreAction) | ||
1543 | + self.register('action', 'store', _StoreAction) | ||
1544 | + self.register('action', 'store_const', _StoreConstAction) | ||
1545 | + self.register('action', 'store_true', _StoreTrueAction) | ||
1546 | + self.register('action', 'store_false', _StoreFalseAction) | ||
1547 | + self.register('action', 'append', _AppendAction) | ||
1548 | + self.register('action', 'append_const', _AppendConstAction) | ||
1549 | + self.register('action', 'count', _CountAction) | ||
1550 | + self.register('action', 'help', _HelpAction) | ||
1551 | + self.register('action', 'version', _VersionAction) | ||
1552 | + self.register('action', 'parsers', _SubParsersAction) | ||
1553 | + | ||
1554 | + # raise an exception if the conflict handler is invalid | ||
1555 | + self._get_handler() | ||
1556 | + | ||
1557 | + # action storage | ||
1558 | + self._actions = [] | ||
1559 | + self._option_string_actions = {} | ||
1560 | + | ||
1561 | + # groups | ||
1562 | + self._action_groups = [] | ||
1563 | + self._mutually_exclusive_groups = [] | ||
1564 | + | ||
1565 | + # defaults storage | ||
1566 | + self._defaults = {} | ||
1567 | + | ||
1568 | + # determines whether an "option" looks like a negative number | ||
1569 | + self._negative_number_matcher = _re.compile(r'^-\d+$|^-\d*\.\d+$') | ||
1570 | + | ||
1571 | + # whether or not there are any optionals that look like negative | ||
1572 | + # numbers -- uses a list so it can be shared and edited | ||
1573 | + self._has_negative_number_optionals = [] | ||
1574 | + | ||
1575 | + # ==================== | ||
1576 | + # Registration methods | ||
1577 | + # ==================== | ||
1578 | + def register(self, registry_name, value, object): | ||
1579 | + registry = self._registries.setdefault(registry_name, {}) | ||
1580 | + registry[value] = object | ||
1581 | + | ||
1582 | + def _registry_get(self, registry_name, value, default=None): | ||
1583 | + return self._registries[registry_name].get(value, default) | ||
1584 | + | ||
1585 | + # ================================== | ||
1586 | + # Namespace default accessor methods | ||
1587 | + # ================================== | ||
1588 | + def set_defaults(self, **kwargs): | ||
1589 | + self._defaults.update(kwargs) | ||
1590 | + | ||
1591 | + # if these defaults match any existing arguments, replace | ||
1592 | + # the previous default on the object with the new one | ||
1593 | + for action in self._actions: | ||
1594 | + if action.dest in kwargs: | ||
1595 | + action.default = kwargs[action.dest] | ||
1596 | + | ||
1597 | + def get_default(self, dest): | ||
1598 | + for action in self._actions: | ||
1599 | + if action.dest == dest and action.default is not None: | ||
1600 | + return action.default | ||
1601 | + return self._defaults.get(dest, None) | ||
1602 | + | ||
1603 | + | ||
1604 | + # ======================= | ||
1605 | + # Adding argument actions | ||
1606 | + # ======================= | ||
1607 | + def add_argument(self, *args, **kwargs): | ||
1608 | + """ | ||
1609 | + add_argument(dest, ..., name=value, ...) | ||
1610 | + add_argument(option_string, option_string, ..., name=value, ...) | ||
1611 | + """ | ||
1612 | + | ||
1613 | + # if no positional args are supplied or only one is supplied and | ||
1614 | + # it doesn't look like an option string, parse a positional | ||
1615 | + # argument | ||
1616 | + chars = self.prefix_chars | ||
1617 | + if not args or len(args) == 1 and args[0][0] not in chars: | ||
1618 | + if args and 'dest' in kwargs: | ||
1619 | + raise ValueError('dest supplied twice for positional argument') | ||
1620 | + kwargs = self._get_positional_kwargs(*args, **kwargs) | ||
1621 | + | ||
1622 | + # otherwise, we're adding an optional argument | ||
1623 | + else: | ||
1624 | + kwargs = self._get_optional_kwargs(*args, **kwargs) | ||
1625 | + | ||
1626 | + # if no default was supplied, use the parser-level default | ||
1627 | + if 'default' not in kwargs: | ||
1628 | + dest = kwargs['dest'] | ||
1629 | + if dest in self._defaults: | ||
1630 | + kwargs['default'] = self._defaults[dest] | ||
1631 | + elif self.argument_default is not None: | ||
1632 | + kwargs['default'] = self.argument_default | ||
1633 | + | ||
1634 | + # create the action object, and add it to the parser | ||
1635 | + action_class = self._pop_action_class(kwargs) | ||
1636 | + if not _callable(action_class): | ||
1637 | + raise ValueError('unknown action "%s"' % action_class) | ||
1638 | + action = action_class(**kwargs) | ||
1639 | + | ||
1640 | + # raise an error if the action type is not callable | ||
1641 | + type_func = self._registry_get('type', action.type, action.type) | ||
1642 | + if not _callable(type_func): | ||
1643 | + raise ValueError('%r is not callable' % type_func) | ||
1644 | + | ||
1645 | + return self._add_action(action) | ||
1646 | + | ||
1647 | + def add_argument_group(self, *args, **kwargs): | ||
1648 | + group = _ArgumentGroup(self, *args, **kwargs) | ||
1649 | + self._action_groups.append(group) | ||
1650 | + return group | ||
1651 | + | ||
1652 | + def add_mutually_exclusive_group(self, **kwargs): | ||
1653 | + group = _MutuallyExclusiveGroup(self, **kwargs) | ||
1654 | + self._mutually_exclusive_groups.append(group) | ||
1655 | + return group | ||
1656 | + | ||
1657 | + def _add_action(self, action): | ||
1658 | + # resolve any conflicts | ||
1659 | + self._check_conflict(action) | ||
1660 | + | ||
1661 | + # add to actions list | ||
1662 | + self._actions.append(action) | ||
1663 | + action.container = self | ||
1664 | + | ||
1665 | + # index the action by any option strings it has | ||
1666 | + for option_string in action.option_strings: | ||
1667 | + self._option_string_actions[option_string] = action | ||
1668 | + | ||
1669 | + # set the flag if any option strings look like negative numbers | ||
1670 | + for option_string in action.option_strings: | ||
1671 | + if self._negative_number_matcher.match(option_string): | ||
1672 | + if not self._has_negative_number_optionals: | ||
1673 | + self._has_negative_number_optionals.append(True) | ||
1674 | + | ||
1675 | + # return the created action | ||
1676 | + return action | ||
1677 | + | ||
1678 | + def _remove_action(self, action): | ||
1679 | + self._actions.remove(action) | ||
1680 | + | ||
1681 | + def _add_container_actions(self, container): | ||
1682 | + # collect groups by titles | ||
1683 | + title_group_map = {} | ||
1684 | + for group in self._action_groups: | ||
1685 | + if group.title in title_group_map: | ||
1686 | + msg = _('cannot merge actions - two groups are named %r') | ||
1687 | + raise ValueError(msg % (group.title)) | ||
1688 | + title_group_map[group.title] = group | ||
1689 | + | ||
1690 | + # map each action to its group | ||
1691 | + group_map = {} | ||
1692 | + for group in container._action_groups: | ||
1693 | + | ||
1694 | + # if a group with the title exists, use that, otherwise | ||
1695 | + # create a new group matching the container's group | ||
1696 | + if group.title not in title_group_map: | ||
1697 | + title_group_map[group.title] = self.add_argument_group( | ||
1698 | + title=group.title, | ||
1699 | + description=group.description, | ||
1700 | + conflict_handler=group.conflict_handler) | ||
1701 | + | ||
1702 | + # map the actions to their new group | ||
1703 | + for action in group._group_actions: | ||
1704 | + group_map[action] = title_group_map[group.title] | ||
1705 | + | ||
1706 | + # add container's mutually exclusive groups | ||
1707 | + # NOTE: if add_mutually_exclusive_group ever gains title= and | ||
1708 | + # description= then this code will need to be expanded as above | ||
1709 | + for group in container._mutually_exclusive_groups: | ||
1710 | + mutex_group = self.add_mutually_exclusive_group( | ||
1711 | + required=group.required) | ||
1712 | + | ||
1713 | + # map the actions to their new mutex group | ||
1714 | + for action in group._group_actions: | ||
1715 | + group_map[action] = mutex_group | ||
1716 | + | ||
1717 | + # add all actions to this container or their group | ||
1718 | + for action in container._actions: | ||
1719 | + group_map.get(action, self)._add_action(action) | ||
1720 | + | ||
1721 | + def _get_positional_kwargs(self, dest, **kwargs): | ||
1722 | + # make sure required is not specified | ||
1723 | + if 'required' in kwargs: | ||
1724 | + msg = _("'required' is an invalid argument for positionals") | ||
1725 | + raise TypeError(msg) | ||
1726 | + | ||
1727 | + # mark positional arguments as required if at least one is | ||
1728 | + # always required | ||
1729 | + if kwargs.get('nargs') not in [OPTIONAL, ZERO_OR_MORE]: | ||
1730 | + kwargs['required'] = True | ||
1731 | + if kwargs.get('nargs') == ZERO_OR_MORE and 'default' not in kwargs: | ||
1732 | + kwargs['required'] = True | ||
1733 | + | ||
1734 | + # return the keyword arguments with no option strings | ||
1735 | + return dict(kwargs, dest=dest, option_strings=[]) | ||
1736 | + | ||
1737 | + def _get_optional_kwargs(self, *args, **kwargs): | ||
1738 | + # determine short and long option strings | ||
1739 | + option_strings = [] | ||
1740 | + long_option_strings = [] | ||
1741 | + for option_string in args: | ||
1742 | + # error on strings that don't start with an appropriate prefix | ||
1743 | + if not option_string[0] in self.prefix_chars: | ||
1744 | + msg = _('invalid option string %r: ' | ||
1745 | + 'must start with a character %r') | ||
1746 | + tup = option_string, self.prefix_chars | ||
1747 | + raise ValueError(msg % tup) | ||
1748 | + | ||
1749 | + # strings starting with two prefix characters are long options | ||
1750 | + option_strings.append(option_string) | ||
1751 | + if option_string[0] in self.prefix_chars: | ||
1752 | + if len(option_string) > 1: | ||
1753 | + if option_string[1] in self.prefix_chars: | ||
1754 | + long_option_strings.append(option_string) | ||
1755 | + | ||
1756 | + # infer destination, '--foo-bar' -> 'foo_bar' and '-x' -> 'x' | ||
1757 | + dest = kwargs.pop('dest', None) | ||
1758 | + if dest is None: | ||
1759 | + if long_option_strings: | ||
1760 | + dest_option_string = long_option_strings[0] | ||
1761 | + else: | ||
1762 | + dest_option_string = option_strings[0] | ||
1763 | + dest = dest_option_string.lstrip(self.prefix_chars) | ||
1764 | + if not dest: | ||
1765 | + msg = _('dest= is required for options like %r') | ||
1766 | + raise ValueError(msg % option_string) | ||
1767 | + dest = dest.replace('-', '_') | ||
1768 | + | ||
1769 | + # return the updated keyword arguments | ||
1770 | + return dict(kwargs, dest=dest, option_strings=option_strings) | ||
1771 | + | ||
1772 | + def _pop_action_class(self, kwargs, default=None): | ||
1773 | + action = kwargs.pop('action', default) | ||
1774 | + return self._registry_get('action', action, action) | ||
1775 | + | ||
1776 | + def _get_handler(self): | ||
1777 | + # determine function from conflict handler string | ||
1778 | + handler_func_name = '_handle_conflict_%s' % self.conflict_handler | ||
1779 | + try: | ||
1780 | + return getattr(self, handler_func_name) | ||
1781 | + except AttributeError: | ||
1782 | + msg = _('invalid conflict_resolution value: %r') | ||
1783 | + raise ValueError(msg % self.conflict_handler) | ||
1784 | + | ||
1785 | + def _check_conflict(self, action): | ||
1786 | + | ||
1787 | + # find all options that conflict with this option | ||
1788 | + confl_optionals = [] | ||
1789 | + for option_string in action.option_strings: | ||
1790 | + if option_string in self._option_string_actions: | ||
1791 | + confl_optional = self._option_string_actions[option_string] | ||
1792 | + confl_optionals.append((option_string, confl_optional)) | ||
1793 | + | ||
1794 | + # resolve any conflicts | ||
1795 | + if confl_optionals: | ||
1796 | + conflict_handler = self._get_handler() | ||
1797 | + conflict_handler(action, confl_optionals) | ||
1798 | + | ||
1799 | + def _handle_conflict_error(self, action, conflicting_actions): | ||
1800 | + message = _('conflicting option string(s): %s') | ||
1801 | + conflict_string = ', '.join([option_string | ||
1802 | + for option_string, action | ||
1803 | + in conflicting_actions]) | ||
1804 | + raise ArgumentError(action, message % conflict_string) | ||
1805 | + | ||
1806 | + def _handle_conflict_resolve(self, action, conflicting_actions): | ||
1807 | + | ||
1808 | + # remove all conflicting options | ||
1809 | + for option_string, action in conflicting_actions: | ||
1810 | + | ||
1811 | + # remove the conflicting option | ||
1812 | + action.option_strings.remove(option_string) | ||
1813 | + self._option_string_actions.pop(option_string, None) | ||
1814 | + | ||
1815 | + # if the option now has no option string, remove it from the | ||
1816 | + # container holding it | ||
1817 | + if not action.option_strings: | ||
1818 | + action.container._remove_action(action) | ||
1819 | + | ||
1820 | + | ||
1821 | +class _ArgumentGroup(_ActionsContainer): | ||
1822 | + | ||
1823 | + def __init__(self, container, title=None, description=None, **kwargs): | ||
1824 | + # add any missing keyword arguments by checking the container | ||
1825 | + update = kwargs.setdefault | ||
1826 | + update('conflict_handler', container.conflict_handler) | ||
1827 | + update('prefix_chars', container.prefix_chars) | ||
1828 | + update('argument_default', container.argument_default) | ||
1829 | + super_init = super(_ArgumentGroup, self).__init__ | ||
1830 | + super_init(description=description, **kwargs) | ||
1831 | + | ||
1832 | + # group attributes | ||
1833 | + self.title = title | ||
1834 | + self._group_actions = [] | ||
1835 | + | ||
1836 | + # share most attributes with the container | ||
1837 | + self._registries = container._registries | ||
1838 | + self._actions = container._actions | ||
1839 | + self._option_string_actions = container._option_string_actions | ||
1840 | + self._defaults = container._defaults | ||
1841 | + self._has_negative_number_optionals = \ | ||
1842 | + container._has_negative_number_optionals | ||
1843 | + | ||
1844 | + def _add_action(self, action): | ||
1845 | + action = super(_ArgumentGroup, self)._add_action(action) | ||
1846 | + self._group_actions.append(action) | ||
1847 | + return action | ||
1848 | + | ||
1849 | + def _remove_action(self, action): | ||
1850 | + super(_ArgumentGroup, self)._remove_action(action) | ||
1851 | + self._group_actions.remove(action) | ||
1852 | + | ||
1853 | + | ||
1854 | +class _MutuallyExclusiveGroup(_ArgumentGroup): | ||
1855 | + | ||
1856 | + def __init__(self, container, required=False): | ||
1857 | + super(_MutuallyExclusiveGroup, self).__init__(container) | ||
1858 | + self.required = required | ||
1859 | + self._container = container | ||
1860 | + | ||
1861 | + def _add_action(self, action): | ||
1862 | + if action.required: | ||
1863 | + msg = _('mutually exclusive arguments must be optional') | ||
1864 | + raise ValueError(msg) | ||
1865 | + action = self._container._add_action(action) | ||
1866 | + self._group_actions.append(action) | ||
1867 | + return action | ||
1868 | + | ||
1869 | + def _remove_action(self, action): | ||
1870 | + self._container._remove_action(action) | ||
1871 | + self._group_actions.remove(action) | ||
1872 | + | ||
1873 | + | ||
1874 | +class ArgumentParser(_AttributeHolder, _ActionsContainer): | ||
1875 | + """Object for parsing command line strings into Python objects. | ||
1876 | + | ||
1877 | + Keyword Arguments: | ||
1878 | + - prog -- The name of the program (default: sys.argv[0]) | ||
1879 | + - usage -- A usage message (default: auto-generated from arguments) | ||
1880 | + - description -- A description of what the program does | ||
1881 | + - epilog -- Text following the argument descriptions | ||
1882 | + - parents -- Parsers whose arguments should be copied into this one | ||
1883 | + - formatter_class -- HelpFormatter class for printing help messages | ||
1884 | + - prefix_chars -- Characters that prefix optional arguments | ||
1885 | + - fromfile_prefix_chars -- Characters that prefix files containing | ||
1886 | + additional arguments | ||
1887 | + - argument_default -- The default value for all arguments | ||
1888 | + - conflict_handler -- String indicating how to handle conflicts | ||
1889 | + - add_help -- Add a -h/-help option | ||
1890 | + """ | ||
1891 | + | ||
1892 | + def __init__(self, | ||
1893 | + prog=None, | ||
1894 | + usage=None, | ||
1895 | + description=None, | ||
1896 | + epilog=None, | ||
1897 | + version=None, | ||
1898 | + parents=[], | ||
1899 | + formatter_class=HelpFormatter, | ||
1900 | + prefix_chars='-', | ||
1901 | + fromfile_prefix_chars=None, | ||
1902 | + argument_default=None, | ||
1903 | + conflict_handler='error', | ||
1904 | + add_help=True): | ||
1905 | + | ||
1906 | + if version is not None: | ||
1907 | + import warnings | ||
1908 | + warnings.warn( | ||
1909 | + """The "version" argument to ArgumentParser is deprecated. """ | ||
1910 | + """Please use """ | ||
1911 | + """"add_argument(..., action='version', version="N", ...)" """ | ||
1912 | + """instead""", DeprecationWarning) | ||
1913 | + | ||
1914 | + superinit = super(ArgumentParser, self).__init__ | ||
1915 | + superinit(description=description, | ||
1916 | + prefix_chars=prefix_chars, | ||
1917 | + argument_default=argument_default, | ||
1918 | + conflict_handler=conflict_handler) | ||
1919 | + | ||
1920 | + # default setting for prog | ||
1921 | + if prog is None: | ||
1922 | + prog = _os.path.basename(_sys.argv[0]) | ||
1923 | + | ||
1924 | + self.prog = prog | ||
1925 | + self.usage = usage | ||
1926 | + self.epilog = epilog | ||
1927 | + self.version = version | ||
1928 | + self.formatter_class = formatter_class | ||
1929 | + self.fromfile_prefix_chars = fromfile_prefix_chars | ||
1930 | + self.add_help = add_help | ||
1931 | + | ||
1932 | + add_group = self.add_argument_group | ||
1933 | + self._positionals = add_group(_('positional arguments')) | ||
1934 | + self._optionals = add_group(_('optional arguments')) | ||
1935 | + self._subparsers = None | ||
1936 | + | ||
1937 | + # register types | ||
1938 | + def identity(string): | ||
1939 | + return string | ||
1940 | + self.register('type', None, identity) | ||
1941 | + | ||
1942 | + # add help and version arguments if necessary | ||
1943 | + # (using explicit default to override global argument_default) | ||
1944 | + if '-' in prefix_chars: | ||
1945 | + default_prefix = '-' | ||
1946 | + else: | ||
1947 | + default_prefix = prefix_chars[0] | ||
1948 | + if self.add_help: | ||
1949 | + self.add_argument( | ||
1950 | + default_prefix+'h', default_prefix*2+'help', | ||
1951 | + action='help', default=SUPPRESS, | ||
1952 | + help=_('show this help message and exit')) | ||
1953 | + if self.version: | ||
1954 | + self.add_argument( | ||
1955 | + default_prefix+'v', default_prefix*2+'version', | ||
1956 | + action='version', default=SUPPRESS, | ||
1957 | + version=self.version, | ||
1958 | + help=_("show program's version number and exit")) | ||
1959 | + | ||
1960 | + # add parent arguments and defaults | ||
1961 | + for parent in parents: | ||
1962 | + self._add_container_actions(parent) | ||
1963 | + try: | ||
1964 | + defaults = parent._defaults | ||
1965 | + except AttributeError: | ||
1966 | + pass | ||
1967 | + else: | ||
1968 | + self._defaults.update(defaults) | ||
1969 | + | ||
1970 | + # ======================= | ||
1971 | + # Pretty __repr__ methods | ||
1972 | + # ======================= | ||
1973 | + def _get_kwargs(self): | ||
1974 | + names = [ | ||
1975 | + 'prog', | ||
1976 | + 'usage', | ||
1977 | + 'description', | ||
1978 | + 'version', | ||
1979 | + 'formatter_class', | ||
1980 | + 'conflict_handler', | ||
1981 | + 'add_help', | ||
1982 | + ] | ||
1983 | + return [(name, getattr(self, name)) for name in names] | ||
1984 | + | ||
1985 | + # ================================== | ||
1986 | + # Optional/Positional adding methods | ||
1987 | + # ================================== | ||
1988 | + def add_subparsers(self, **kwargs): | ||
1989 | + if self._subparsers is not None: | ||
1990 | + self.error(_('cannot have multiple subparser arguments')) | ||
1991 | + | ||
1992 | + # add the parser class to the arguments if it's not present | ||
1993 | + kwargs.setdefault('parser_class', type(self)) | ||
1994 | + | ||
1995 | + if 'title' in kwargs or 'description' in kwargs: | ||
1996 | + title = _(kwargs.pop('title', 'subcommands')) | ||
1997 | + description = _(kwargs.pop('description', None)) | ||
1998 | + self._subparsers = self.add_argument_group(title, description) | ||
1999 | + else: | ||
2000 | + self._subparsers = self._positionals | ||
2001 | + | ||
2002 | + # prog defaults to the usage message of this parser, skipping | ||
2003 | + # optional arguments and with no "usage:" prefix | ||
2004 | + if kwargs.get('prog') is None: | ||
2005 | + formatter = self._get_formatter() | ||
2006 | + positionals = self._get_positional_actions() | ||
2007 | + groups = self._mutually_exclusive_groups | ||
2008 | + formatter.add_usage(self.usage, positionals, groups, '') | ||
2009 | + kwargs['prog'] = formatter.format_help().strip() | ||
2010 | + | ||
2011 | + # create the parsers action and add it to the positionals list | ||
2012 | + parsers_class = self._pop_action_class(kwargs, 'parsers') | ||
2013 | + action = parsers_class(option_strings=[], **kwargs) | ||
2014 | + self._subparsers._add_action(action) | ||
2015 | + | ||
2016 | + # return the created parsers action | ||
2017 | + return action | ||
2018 | + | ||
2019 | + def _add_action(self, action): | ||
2020 | + if action.option_strings: | ||
2021 | + self._optionals._add_action(action) | ||
2022 | + else: | ||
2023 | + self._positionals._add_action(action) | ||
2024 | + return action | ||
2025 | + | ||
2026 | + def _get_optional_actions(self): | ||
2027 | + return [action | ||
2028 | + for action in self._actions | ||
2029 | + if action.option_strings] | ||
2030 | + | ||
2031 | + def _get_positional_actions(self): | ||
2032 | + return [action | ||
2033 | + for action in self._actions | ||
2034 | + if not action.option_strings] | ||
2035 | + | ||
2036 | + # ===================================== | ||
2037 | + # Command line argument parsing methods | ||
2038 | + # ===================================== | ||
2039 | + def parse_args(self, args=None, namespace=None): | ||
2040 | + args, argv = self.parse_known_args(args, namespace) | ||
2041 | + if argv: | ||
2042 | + msg = _('unrecognized arguments: %s') | ||
2043 | + self.error(msg % ' '.join(argv)) | ||
2044 | + return args | ||
2045 | + | ||
2046 | + def parse_known_args(self, args=None, namespace=None): | ||
2047 | + # args default to the system args | ||
2048 | + if args is None: | ||
2049 | + args = _sys.argv[1:] | ||
2050 | + | ||
2051 | + # default Namespace built from parser defaults | ||
2052 | + if namespace is None: | ||
2053 | + namespace = Namespace() | ||
2054 | + | ||
2055 | + # add any action defaults that aren't present | ||
2056 | + for action in self._actions: | ||
2057 | + if action.dest is not SUPPRESS: | ||
2058 | + if not hasattr(namespace, action.dest): | ||
2059 | + if action.default is not SUPPRESS: | ||
2060 | + setattr(namespace, action.dest, action.default) | ||
2061 | + | ||
2062 | + # add any parser defaults that aren't present | ||
2063 | + for dest in self._defaults: | ||
2064 | + if not hasattr(namespace, dest): | ||
2065 | + setattr(namespace, dest, self._defaults[dest]) | ||
2066 | + | ||
2067 | + # parse the arguments and exit if there are any errors | ||
2068 | + try: | ||
2069 | + namespace, args = self._parse_known_args(args, namespace) | ||
2070 | + if hasattr(namespace, _UNRECOGNIZED_ARGS_ATTR): | ||
2071 | + args.extend(getattr(namespace, _UNRECOGNIZED_ARGS_ATTR)) | ||
2072 | + delattr(namespace, _UNRECOGNIZED_ARGS_ATTR) | ||
2073 | + return namespace, args | ||
2074 | + except ArgumentError: | ||
2075 | + err = _sys.exc_info()[1] | ||
2076 | + self.error(str(err)) | ||
2077 | + | ||
2078 | + def _parse_known_args(self, arg_strings, namespace): | ||
2079 | + # replace arg strings that are file references | ||
2080 | + if self.fromfile_prefix_chars is not None: | ||
2081 | + arg_strings = self._read_args_from_files(arg_strings) | ||
2082 | + | ||
2083 | + # map all mutually exclusive arguments to the other arguments | ||
2084 | + # they can't occur with | ||
2085 | + action_conflicts = {} | ||
2086 | + for mutex_group in self._mutually_exclusive_groups: | ||
2087 | + group_actions = mutex_group._group_actions | ||
2088 | + for i, mutex_action in enumerate(mutex_group._group_actions): | ||
2089 | + conflicts = action_conflicts.setdefault(mutex_action, []) | ||
2090 | + conflicts.extend(group_actions[:i]) | ||
2091 | + conflicts.extend(group_actions[i + 1:]) | ||
2092 | + | ||
2093 | + # find all option indices, and determine the arg_string_pattern | ||
2094 | + # which has an 'O' if there is an option at an index, | ||
2095 | + # an 'A' if there is an argument, or a '-' if there is a '--' | ||
2096 | + option_string_indices = {} | ||
2097 | + arg_string_pattern_parts = [] | ||
2098 | + arg_strings_iter = iter(arg_strings) | ||
2099 | + for i, arg_string in enumerate(arg_strings_iter): | ||
2100 | + | ||
2101 | + # all args after -- are non-options | ||
2102 | + if arg_string == '--': | ||
2103 | + arg_string_pattern_parts.append('-') | ||
2104 | + for arg_string in arg_strings_iter: | ||
2105 | + arg_string_pattern_parts.append('A') | ||
2106 | + | ||
2107 | + # otherwise, add the arg to the arg strings | ||
2108 | + # and note the index if it was an option | ||
2109 | + else: | ||
2110 | + option_tuple = self._parse_optional(arg_string) | ||
2111 | + if option_tuple is None: | ||
2112 | + pattern = 'A' | ||
2113 | + else: | ||
2114 | + option_string_indices[i] = option_tuple | ||
2115 | + pattern = 'O' | ||
2116 | + arg_string_pattern_parts.append(pattern) | ||
2117 | + | ||
2118 | + # join the pieces together to form the pattern | ||
2119 | + arg_strings_pattern = ''.join(arg_string_pattern_parts) | ||
2120 | + | ||
2121 | + # converts arg strings to the appropriate and then takes the action | ||
2122 | + seen_actions = set() | ||
2123 | + seen_non_default_actions = set() | ||
2124 | + | ||
2125 | + def take_action(action, argument_strings, option_string=None): | ||
2126 | + seen_actions.add(action) | ||
2127 | + argument_values = self._get_values(action, argument_strings) | ||
2128 | + | ||
2129 | + # error if this argument is not allowed with other previously | ||
2130 | + # seen arguments, assuming that actions that use the default | ||
2131 | + # value don't really count as "present" | ||
2132 | + if argument_values is not action.default: | ||
2133 | + seen_non_default_actions.add(action) | ||
2134 | + for conflict_action in action_conflicts.get(action, []): | ||
2135 | + if conflict_action in seen_non_default_actions: | ||
2136 | + msg = _('not allowed with argument %s') | ||
2137 | + action_name = _get_action_name(conflict_action) | ||
2138 | + raise ArgumentError(action, msg % action_name) | ||
2139 | + | ||
2140 | + # take the action if we didn't receive a SUPPRESS value | ||
2141 | + # (e.g. from a default) | ||
2142 | + if argument_values is not SUPPRESS: | ||
2143 | + action(self, namespace, argument_values, option_string) | ||
2144 | + | ||
2145 | + # function to convert arg_strings into an optional action | ||
2146 | + def consume_optional(start_index): | ||
2147 | + | ||
2148 | + # get the optional identified at this index | ||
2149 | + option_tuple = option_string_indices[start_index] | ||
2150 | + action, option_string, explicit_arg = option_tuple | ||
2151 | + | ||
2152 | + # identify additional optionals in the same arg string | ||
2153 | + # (e.g. -xyz is the same as -x -y -z if no args are required) | ||
2154 | + match_argument = self._match_argument | ||
2155 | + action_tuples = [] | ||
2156 | + while True: | ||
2157 | + | ||
2158 | + # if we found no optional action, skip it | ||
2159 | + if action is None: | ||
2160 | + extras.append(arg_strings[start_index]) | ||
2161 | + return start_index + 1 | ||
2162 | + | ||
2163 | + # if there is an explicit argument, try to match the | ||
2164 | + # optional's string arguments to only this | ||
2165 | + if explicit_arg is not None: | ||
2166 | + arg_count = match_argument(action, 'A') | ||
2167 | + | ||
2168 | + # if the action is a single-dash option and takes no | ||
2169 | + # arguments, try to parse more single-dash options out | ||
2170 | + # of the tail of the option string | ||
2171 | + chars = self.prefix_chars | ||
2172 | + if arg_count == 0 and option_string[1] not in chars: | ||
2173 | + action_tuples.append((action, [], option_string)) | ||
2174 | + char = option_string[0] | ||
2175 | + option_string = char + explicit_arg[0] | ||
2176 | + new_explicit_arg = explicit_arg[1:] or None | ||
2177 | + optionals_map = self._option_string_actions | ||
2178 | + if option_string in optionals_map: | ||
2179 | + action = optionals_map[option_string] | ||
2180 | + explicit_arg = new_explicit_arg | ||
2181 | + else: | ||
2182 | + msg = _('ignored explicit argument %r') | ||
2183 | + raise ArgumentError(action, msg % explicit_arg) | ||
2184 | + | ||
2185 | + # if the action expect exactly one argument, we've | ||
2186 | + # successfully matched the option; exit the loop | ||
2187 | + elif arg_count == 1: | ||
2188 | + stop = start_index + 1 | ||
2189 | + args = [explicit_arg] | ||
2190 | + action_tuples.append((action, args, option_string)) | ||
2191 | + break | ||
2192 | + | ||
2193 | + # error if a double-dash option did not use the | ||
2194 | + # explicit argument | ||
2195 | + else: | ||
2196 | + msg = _('ignored explicit argument %r') | ||
2197 | + raise ArgumentError(action, msg % explicit_arg) | ||
2198 | + | ||
2199 | + # if there is no explicit argument, try to match the | ||
2200 | + # optional's string arguments with the following strings | ||
2201 | + # if successful, exit the loop | ||
2202 | + else: | ||
2203 | + start = start_index + 1 | ||
2204 | + selected_patterns = arg_strings_pattern[start:] | ||
2205 | + arg_count = match_argument(action, selected_patterns) | ||
2206 | + stop = start + arg_count | ||
2207 | + args = arg_strings[start:stop] | ||
2208 | + action_tuples.append((action, args, option_string)) | ||
2209 | + break | ||
2210 | + | ||
2211 | + # add the Optional to the list and return the index at which | ||
2212 | + # the Optional's string args stopped | ||
2213 | + assert action_tuples | ||
2214 | + for action, args, option_string in action_tuples: | ||
2215 | + take_action(action, args, option_string) | ||
2216 | + return stop | ||
2217 | + | ||
2218 | + # the list of Positionals left to be parsed; this is modified | ||
2219 | + # by consume_positionals() | ||
2220 | + positionals = self._get_positional_actions() | ||
2221 | + | ||
2222 | + # function to convert arg_strings into positional actions | ||
2223 | + def consume_positionals(start_index): | ||
2224 | + # match as many Positionals as possible | ||
2225 | + match_partial = self._match_arguments_partial | ||
2226 | + selected_pattern = arg_strings_pattern[start_index:] | ||
2227 | + arg_counts = match_partial(positionals, selected_pattern) | ||
2228 | + | ||
2229 | + # slice off the appropriate arg strings for each Positional | ||
2230 | + # and add the Positional and its args to the list | ||
2231 | + for action, arg_count in zip(positionals, arg_counts): | ||
2232 | + args = arg_strings[start_index: start_index + arg_count] | ||
2233 | + start_index += arg_count | ||
2234 | + take_action(action, args) | ||
2235 | + | ||
2236 | + # slice off the Positionals that we just parsed and return the | ||
2237 | + # index at which the Positionals' string args stopped | ||
2238 | + positionals[:] = positionals[len(arg_counts):] | ||
2239 | + return start_index | ||
2240 | + | ||
2241 | + # consume Positionals and Optionals alternately, until we have | ||
2242 | + # passed the last option string | ||
2243 | + extras = [] | ||
2244 | + start_index = 0 | ||
2245 | + if option_string_indices: | ||
2246 | + max_option_string_index = max(option_string_indices) | ||
2247 | + else: | ||
2248 | + max_option_string_index = -1 | ||
2249 | + while start_index <= max_option_string_index: | ||
2250 | + | ||
2251 | + # consume any Positionals preceding the next option | ||
2252 | + next_option_string_index = min([ | ||
2253 | + index | ||
2254 | + for index in option_string_indices | ||
2255 | + if index >= start_index]) | ||
2256 | + if start_index != next_option_string_index: | ||
2257 | + positionals_end_index = consume_positionals(start_index) | ||
2258 | + | ||
2259 | + # only try to parse the next optional if we didn't consume | ||
2260 | + # the option string during the positionals parsing | ||
2261 | + if positionals_end_index > start_index: | ||
2262 | + start_index = positionals_end_index | ||
2263 | + continue | ||
2264 | + else: | ||
2265 | + start_index = positionals_end_index | ||
2266 | + | ||
2267 | + # if we consumed all the positionals we could and we're not | ||
2268 | + # at the index of an option string, there were extra arguments | ||
2269 | + if start_index not in option_string_indices: | ||
2270 | + strings = arg_strings[start_index:next_option_string_index] | ||
2271 | + extras.extend(strings) | ||
2272 | + start_index = next_option_string_index | ||
2273 | + | ||
2274 | + # consume the next optional and any arguments for it | ||
2275 | + start_index = consume_optional(start_index) | ||
2276 | + | ||
2277 | + # consume any positionals following the last Optional | ||
2278 | + stop_index = consume_positionals(start_index) | ||
2279 | + | ||
2280 | + # if we didn't consume all the argument strings, there were extras | ||
2281 | + extras.extend(arg_strings[stop_index:]) | ||
2282 | + | ||
2283 | + # if we didn't use all the Positional objects, there were too few | ||
2284 | + # arg strings supplied. | ||
2285 | + if positionals: | ||
2286 | + self.error(_('too few arguments')) | ||
2287 | + | ||
2288 | + # make sure all required actions were present, and convert defaults. | ||
2289 | + for action in self._actions: | ||
2290 | + if action not in seen_actions: | ||
2291 | + if action.required: | ||
2292 | + name = _get_action_name(action) | ||
2293 | + self.error(_('argument %s is required') % name) | ||
2294 | + else: | ||
2295 | + # Convert action default now instead of doing it before | ||
2296 | + # parsing arguments to avoid calling convert functions | ||
2297 | + # twice (which may fail) if the argument was given, but | ||
2298 | + # only if it was defined already in the namespace | ||
2299 | + if (action.default is not None and | ||
2300 | + isinstance(action.default, basestring) and | ||
2301 | + hasattr(namespace, action.dest) and | ||
2302 | + action.default is getattr(namespace, action.dest)): | ||
2303 | + setattr(namespace, action.dest, | ||
2304 | + self._get_value(action, action.default)) | ||
2305 | + | ||
2306 | + # make sure all required groups had one option present | ||
2307 | + for group in self._mutually_exclusive_groups: | ||
2308 | + if group.required: | ||
2309 | + for action in group._group_actions: | ||
2310 | + if action in seen_non_default_actions: | ||
2311 | + break | ||
2312 | + | ||
2313 | + # if no actions were used, report the error | ||
2314 | + else: | ||
2315 | + names = [_get_action_name(action) | ||
2316 | + for action in group._group_actions | ||
2317 | + if action.help is not SUPPRESS] | ||
2318 | + msg = _('one of the arguments %s is required') | ||
2319 | + self.error(msg % ' '.join(names)) | ||
2320 | + | ||
2321 | + # return the updated namespace and the extra arguments | ||
2322 | + return namespace, extras | ||
2323 | + | ||
2324 | + def _read_args_from_files(self, arg_strings): | ||
2325 | + # expand arguments referencing files | ||
2326 | + new_arg_strings = [] | ||
2327 | + for arg_string in arg_strings: | ||
2328 | + | ||
2329 | + # for regular arguments, just add them back into the list | ||
2330 | + if arg_string[0] not in self.fromfile_prefix_chars: | ||
2331 | + new_arg_strings.append(arg_string) | ||
2332 | + | ||
2333 | + # replace arguments referencing files with the file content | ||
2334 | + else: | ||
2335 | + try: | ||
2336 | + args_file = open(arg_string[1:]) | ||
2337 | + try: | ||
2338 | + arg_strings = [] | ||
2339 | + for arg_line in args_file.read().splitlines(): | ||
2340 | + for arg in self.convert_arg_line_to_args(arg_line): | ||
2341 | + arg_strings.append(arg) | ||
2342 | + arg_strings = self._read_args_from_files(arg_strings) | ||
2343 | + new_arg_strings.extend(arg_strings) | ||
2344 | + finally: | ||
2345 | + args_file.close() | ||
2346 | + except IOError: | ||
2347 | + err = _sys.exc_info()[1] | ||
2348 | + self.error(str(err)) | ||
2349 | + | ||
2350 | + # return the modified argument list | ||
2351 | + return new_arg_strings | ||
2352 | + | ||
2353 | + def convert_arg_line_to_args(self, arg_line): | ||
2354 | + return [arg_line] | ||
2355 | + | ||
2356 | + def _match_argument(self, action, arg_strings_pattern): | ||
2357 | + # match the pattern for this action to the arg strings | ||
2358 | + nargs_pattern = self._get_nargs_pattern(action) | ||
2359 | + match = _re.match(nargs_pattern, arg_strings_pattern) | ||
2360 | + | ||
2361 | + # raise an exception if we weren't able to find a match | ||
2362 | + if match is None: | ||
2363 | + nargs_errors = { | ||
2364 | + None: _('expected one argument'), | ||
2365 | + OPTIONAL: _('expected at most one argument'), | ||
2366 | + ONE_OR_MORE: _('expected at least one argument'), | ||
2367 | + } | ||
2368 | + default = _('expected %s argument(s)') % action.nargs | ||
2369 | + msg = nargs_errors.get(action.nargs, default) | ||
2370 | + raise ArgumentError(action, msg) | ||
2371 | + | ||
2372 | + # return the number of arguments matched | ||
2373 | + return len(match.group(1)) | ||
2374 | + | ||
2375 | + def _match_arguments_partial(self, actions, arg_strings_pattern): | ||
2376 | + # progressively shorten the actions list by slicing off the | ||
2377 | + # final actions until we find a match | ||
2378 | + result = [] | ||
2379 | + for i in range(len(actions), 0, -1): | ||
2380 | + actions_slice = actions[:i] | ||
2381 | + pattern = ''.join([self._get_nargs_pattern(action) | ||
2382 | + for action in actions_slice]) | ||
2383 | + match = _re.match(pattern, arg_strings_pattern) | ||
2384 | + if match is not None: | ||
2385 | + result.extend([len(string) for string in match.groups()]) | ||
2386 | + break | ||
2387 | + | ||
2388 | + # return the list of arg string counts | ||
2389 | + return result | ||
2390 | + | ||
2391 | + def _parse_optional(self, arg_string): | ||
2392 | + # if it's an empty string, it was meant to be a positional | ||
2393 | + if not arg_string: | ||
2394 | + return None | ||
2395 | + | ||
2396 | + # if it doesn't start with a prefix, it was meant to be positional | ||
2397 | + if not arg_string[0] in self.prefix_chars: | ||
2398 | + return None | ||
2399 | + | ||
2400 | + # if the option string is present in the parser, return the action | ||
2401 | + if arg_string in self._option_string_actions: | ||
2402 | + action = self._option_string_actions[arg_string] | ||
2403 | + return action, arg_string, None | ||
2404 | + | ||
2405 | + # if it's just a single character, it was meant to be positional | ||
2406 | + if len(arg_string) == 1: | ||
2407 | + return None | ||
2408 | + | ||
2409 | + # if the option string before the "=" is present, return the action | ||
2410 | + if '=' in arg_string: | ||
2411 | + option_string, explicit_arg = arg_string.split('=', 1) | ||
2412 | + if option_string in self._option_string_actions: | ||
2413 | + action = self._option_string_actions[option_string] | ||
2414 | + return action, option_string, explicit_arg | ||
2415 | + | ||
2416 | + # search through all possible prefixes of the option string | ||
2417 | + # and all actions in the parser for possible interpretations | ||
2418 | + option_tuples = self._get_option_tuples(arg_string) | ||
2419 | + | ||
2420 | + # if multiple actions match, the option string was ambiguous | ||
2421 | + if len(option_tuples) > 1: | ||
2422 | + options = ', '.join([option_string | ||
2423 | + for action, option_string, explicit_arg in option_tuples]) | ||
2424 | + tup = arg_string, options | ||
2425 | + self.error(_('ambiguous option: %s could match %s') % tup) | ||
2426 | + | ||
2427 | + # if exactly one action matched, this segmentation is good, | ||
2428 | + # so return the parsed action | ||
2429 | + elif len(option_tuples) == 1: | ||
2430 | + option_tuple, = option_tuples | ||
2431 | + return option_tuple | ||
2432 | + | ||
2433 | + # if it was not found as an option, but it looks like a negative | ||
2434 | + # number, it was meant to be positional | ||
2435 | + # unless there are negative-number-like options | ||
2436 | + if self._negative_number_matcher.match(arg_string): | ||
2437 | + if not self._has_negative_number_optionals: | ||
2438 | + return None | ||
2439 | + | ||
2440 | + # if it contains a space, it was meant to be a positional | ||
2441 | + if ' ' in arg_string: | ||
2442 | + return None | ||
2443 | + | ||
2444 | + # it was meant to be an optional but there is no such option | ||
2445 | + # in this parser (though it might be a valid option in a subparser) | ||
2446 | + return None, arg_string, None | ||
2447 | + | ||
2448 | + def _get_option_tuples(self, option_string): | ||
2449 | + result = [] | ||
2450 | + | ||
2451 | + # option strings starting with two prefix characters are only | ||
2452 | + # split at the '=' | ||
2453 | + chars = self.prefix_chars | ||
2454 | + if option_string[0] in chars and option_string[1] in chars: | ||
2455 | + if '=' in option_string: | ||
2456 | + option_prefix, explicit_arg = option_string.split('=', 1) | ||
2457 | + else: | ||
2458 | + option_prefix = option_string | ||
2459 | + explicit_arg = None | ||
2460 | + for option_string in self._option_string_actions: | ||
2461 | + if option_string.startswith(option_prefix): | ||
2462 | + action = self._option_string_actions[option_string] | ||
2463 | + tup = action, option_string, explicit_arg | ||
2464 | + result.append(tup) | ||
2465 | + | ||
2466 | + # single character options can be concatenated with their arguments | ||
2467 | + # but multiple character options always have to have their argument | ||
2468 | + # separate | ||
2469 | + elif option_string[0] in chars and option_string[1] not in chars: | ||
2470 | + option_prefix = option_string | ||
2471 | + explicit_arg = None | ||
2472 | + short_option_prefix = option_string[:2] | ||
2473 | + short_explicit_arg = option_string[2:] | ||
2474 | + | ||
2475 | + for option_string in self._option_string_actions: | ||
2476 | + if option_string == short_option_prefix: | ||
2477 | + action = self._option_string_actions[option_string] | ||
2478 | + tup = action, option_string, short_explicit_arg | ||
2479 | + result.append(tup) | ||
2480 | + elif option_string.startswith(option_prefix): | ||
2481 | + action = self._option_string_actions[option_string] | ||
2482 | + tup = action, option_string, explicit_arg | ||
2483 | + result.append(tup) | ||
2484 | + | ||
2485 | + # shouldn't ever get here | ||
2486 | + else: | ||
2487 | + self.error(_('unexpected option string: %s') % option_string) | ||
2488 | + | ||
2489 | + # return the collected option tuples | ||
2490 | + return result | ||
2491 | + | ||
2492 | + def _get_nargs_pattern(self, action): | ||
2493 | + # in all examples below, we have to allow for '--' args | ||
2494 | + # which are represented as '-' in the pattern | ||
2495 | + nargs = action.nargs | ||
2496 | + | ||
2497 | + # the default (None) is assumed to be a single argument | ||
2498 | + if nargs is None: | ||
2499 | + nargs_pattern = '(-*A-*)' | ||
2500 | + | ||
2501 | + # allow zero or one arguments | ||
2502 | + elif nargs == OPTIONAL: | ||
2503 | + nargs_pattern = '(-*A?-*)' | ||
2504 | + | ||
2505 | + # allow zero or more arguments | ||
2506 | + elif nargs == ZERO_OR_MORE: | ||
2507 | + nargs_pattern = '(-*[A-]*)' | ||
2508 | + | ||
2509 | + # allow one or more arguments | ||
2510 | + elif nargs == ONE_OR_MORE: | ||
2511 | + nargs_pattern = '(-*A[A-]*)' | ||
2512 | + | ||
2513 | + # allow any number of options or arguments | ||
2514 | + elif nargs == REMAINDER: | ||
2515 | + nargs_pattern = '([-AO]*)' | ||
2516 | + | ||
2517 | + # allow one argument followed by any number of options or arguments | ||
2518 | + elif nargs == PARSER: | ||
2519 | + nargs_pattern = '(-*A[-AO]*)' | ||
2520 | + | ||
2521 | + # all others should be integers | ||
2522 | + else: | ||
2523 | + nargs_pattern = '(-*%s-*)' % '-*'.join('A' * nargs) | ||
2524 | + | ||
2525 | + # if this is an optional action, -- is not allowed | ||
2526 | + if action.option_strings: | ||
2527 | + nargs_pattern = nargs_pattern.replace('-*', '') | ||
2528 | + nargs_pattern = nargs_pattern.replace('-', '') | ||
2529 | + | ||
2530 | + # return the pattern | ||
2531 | + return nargs_pattern | ||
2532 | + | ||
2533 | + # ======================== | ||
2534 | + # Value conversion methods | ||
2535 | + # ======================== | ||
2536 | + def _get_values(self, action, arg_strings): | ||
2537 | + # for everything but PARSER args, strip out '--' | ||
2538 | + if action.nargs not in [PARSER, REMAINDER]: | ||
2539 | + arg_strings = [s for s in arg_strings if s != '--'] | ||
2540 | + | ||
2541 | + # optional argument produces a default when not present | ||
2542 | + if not arg_strings and action.nargs == OPTIONAL: | ||
2543 | + if action.option_strings: | ||
2544 | + value = action.const | ||
2545 | + else: | ||
2546 | + value = action.default | ||
2547 | + if isinstance(value, basestring): | ||
2548 | + value = self._get_value(action, value) | ||
2549 | + self._check_value(action, value) | ||
2550 | + | ||
2551 | + # when nargs='*' on a positional, if there were no command-line | ||
2552 | + # args, use the default if it is anything other than None | ||
2553 | + elif (not arg_strings and action.nargs == ZERO_OR_MORE and | ||
2554 | + not action.option_strings): | ||
2555 | + if action.default is not None: | ||
2556 | + value = action.default | ||
2557 | + else: | ||
2558 | + value = arg_strings | ||
2559 | + self._check_value(action, value) | ||
2560 | + | ||
2561 | + # single argument or optional argument produces a single value | ||
2562 | + elif len(arg_strings) == 1 and action.nargs in [None, OPTIONAL]: | ||
2563 | + arg_string, = arg_strings | ||
2564 | + value = self._get_value(action, arg_string) | ||
2565 | + self._check_value(action, value) | ||
2566 | + | ||
2567 | + # REMAINDER arguments convert all values, checking none | ||
2568 | + elif action.nargs == REMAINDER: | ||
2569 | + value = [self._get_value(action, v) for v in arg_strings] | ||
2570 | + | ||
2571 | + # PARSER arguments convert all values, but check only the first | ||
2572 | + elif action.nargs == PARSER: | ||
2573 | + value = [self._get_value(action, v) for v in arg_strings] | ||
2574 | + self._check_value(action, value[0]) | ||
2575 | + | ||
2576 | + # all other types of nargs produce a list | ||
2577 | + else: | ||
2578 | + value = [self._get_value(action, v) for v in arg_strings] | ||
2579 | + for v in value: | ||
2580 | + self._check_value(action, v) | ||
2581 | + | ||
2582 | + # return the converted value | ||
2583 | + return value | ||
2584 | + | ||
2585 | + def _get_value(self, action, arg_string): | ||
2586 | + type_func = self._registry_get('type', action.type, action.type) | ||
2587 | + if not _callable(type_func): | ||
2588 | + msg = _('%r is not callable') | ||
2589 | + raise ArgumentError(action, msg % type_func) | ||
2590 | + | ||
2591 | + # convert the value to the appropriate type | ||
2592 | + try: | ||
2593 | + result = type_func(arg_string) | ||
2594 | + | ||
2595 | + # ArgumentTypeErrors indicate errors | ||
2596 | + except ArgumentTypeError: | ||
2597 | + name = getattr(action.type, '__name__', repr(action.type)) | ||
2598 | + msg = str(_sys.exc_info()[1]) | ||
2599 | + raise ArgumentError(action, msg) | ||
2600 | + | ||
2601 | + # TypeErrors or ValueErrors also indicate errors | ||
2602 | + except (TypeError, ValueError): | ||
2603 | + name = getattr(action.type, '__name__', repr(action.type)) | ||
2604 | + msg = _('invalid %s value: %r') | ||
2605 | + raise ArgumentError(action, msg % (name, arg_string)) | ||
2606 | + | ||
2607 | + # return the converted value | ||
2608 | + return result | ||
2609 | + | ||
2610 | + def _check_value(self, action, value): | ||
2611 | + # converted value must be one of the choices (if specified) | ||
2612 | + if action.choices is not None and value not in action.choices: | ||
2613 | + tup = value, ', '.join(map(repr, action.choices)) | ||
2614 | + msg = _('invalid choice: %r (choose from %s)') % tup | ||
2615 | + raise ArgumentError(action, msg) | ||
2616 | + | ||
2617 | + # ======================= | ||
2618 | + # Help-formatting methods | ||
2619 | + # ======================= | ||
2620 | + def format_usage(self): | ||
2621 | + formatter = self._get_formatter() | ||
2622 | + formatter.add_usage(self.usage, self._actions, | ||
2623 | + self._mutually_exclusive_groups) | ||
2624 | + return formatter.format_help() | ||
2625 | + | ||
2626 | + def format_help(self): | ||
2627 | + formatter = self._get_formatter() | ||
2628 | + | ||
2629 | + # usage | ||
2630 | + formatter.add_usage(self.usage, self._actions, | ||
2631 | + self._mutually_exclusive_groups) | ||
2632 | + | ||
2633 | + # description | ||
2634 | + formatter.add_text(self.description) | ||
2635 | + | ||
2636 | + # positionals, optionals and user-defined groups | ||
2637 | + for action_group in self._action_groups: | ||
2638 | + formatter.start_section(action_group.title) | ||
2639 | + formatter.add_text(action_group.description) | ||
2640 | + formatter.add_arguments(action_group._group_actions) | ||
2641 | + formatter.end_section() | ||
2642 | + | ||
2643 | + # epilog | ||
2644 | + formatter.add_text(self.epilog) | ||
2645 | + | ||
2646 | + # determine help from format above | ||
2647 | + return formatter.format_help() | ||
2648 | + | ||
2649 | + def format_version(self): | ||
2650 | + import warnings | ||
2651 | + warnings.warn( | ||
2652 | + 'The format_version method is deprecated -- the "version" ' | ||
2653 | + 'argument to ArgumentParser is no longer supported.', | ||
2654 | + DeprecationWarning) | ||
2655 | + formatter = self._get_formatter() | ||
2656 | + formatter.add_text(self.version) | ||
2657 | + return formatter.format_help() | ||
2658 | + | ||
2659 | + def _get_formatter(self): | ||
2660 | + return self.formatter_class(prog=self.prog) | ||
2661 | + | ||
2662 | + # ===================== | ||
2663 | + # Help-printing methods | ||
2664 | + # ===================== | ||
2665 | + def print_usage(self, file=None): | ||
2666 | + if file is None: | ||
2667 | + file = _sys.stdout | ||
2668 | + self._print_message(self.format_usage(), file) | ||
2669 | + | ||
2670 | + def print_help(self, file=None): | ||
2671 | + if file is None: | ||
2672 | + file = _sys.stdout | ||
2673 | + self._print_message(self.format_help(), file) | ||
2674 | + | ||
2675 | + def print_version(self, file=None): | ||
2676 | + import warnings | ||
2677 | + warnings.warn( | ||
2678 | + 'The print_version method is deprecated -- the "version" ' | ||
2679 | + 'argument to ArgumentParser is no longer supported.', | ||
2680 | + DeprecationWarning) | ||
2681 | + self._print_message(self.format_version(), file) | ||
2682 | + | ||
2683 | + def _print_message(self, message, file=None): | ||
2684 | + if message: | ||
2685 | + if file is None: | ||
2686 | + file = _sys.stderr | ||
2687 | + file.write(message) | ||
2688 | + | ||
2689 | + # =============== | ||
2690 | + # Exiting methods | ||
2691 | + # =============== | ||
2692 | + def exit(self, status=0, message=None): | ||
2693 | + if message: | ||
2694 | + self._print_message(message, _sys.stderr) | ||
2695 | + _sys.exit(status) | ||
2696 | + | ||
2697 | + def error(self, message): | ||
2698 | + """error(message: string) | ||
2699 | + | ||
2700 | + Prints a usage message incorporating the message to stderr and | ||
2701 | + exits. | ||
2702 | + | ||
2703 | + If you override this in a subclass, it should not return -- it | ||
2704 | + should either exit or raise an exception. | ||
2705 | + """ | ||
2706 | + self.print_usage(_sys.stderr) | ||
2707 | + self.exit(2, _('%s: error: %s\n') % (self.prog, message)) | ||
2708 | -- | 653 | -- |
2709 | 2.13.5 | 654 | 2.9.3 |
2710 | 655 | ||
2711 | 656 | diff view generated by jsdifflib |
New patch | |||
---|---|---|---|
1 | From: Paolo Bonzini <pbonzini@redhat.com> | ||
1 | 2 | ||
3 | qcow2_create2 calls this. Do not run a nested event loop, as that | ||
4 | breaks when aio_co_wake tries to queue the coroutine on the co_queue_wakeup | ||
5 | list of the currently running one. | ||
6 | |||
7 | Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
8 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> | ||
9 | Reviewed-by: Fam Zheng <famz@redhat.com> | ||
10 | Message-id: 20170213135235.12274-4-pbonzini@redhat.com | ||
11 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
12 | --- | ||
13 | block/block-backend.c | 12 ++++++++---- | ||
14 | 1 file changed, 8 insertions(+), 4 deletions(-) | ||
15 | |||
16 | diff --git a/block/block-backend.c b/block/block-backend.c | ||
17 | index XXXXXXX..XXXXXXX 100644 | ||
18 | --- a/block/block-backend.c | ||
19 | +++ b/block/block-backend.c | ||
20 | @@ -XXX,XX +XXX,XX @@ static int blk_prw(BlockBackend *blk, int64_t offset, uint8_t *buf, | ||
21 | { | ||
22 | QEMUIOVector qiov; | ||
23 | struct iovec iov; | ||
24 | - Coroutine *co; | ||
25 | BlkRwCo rwco; | ||
26 | |||
27 | iov = (struct iovec) { | ||
28 | @@ -XXX,XX +XXX,XX @@ static int blk_prw(BlockBackend *blk, int64_t offset, uint8_t *buf, | ||
29 | .ret = NOT_DONE, | ||
30 | }; | ||
31 | |||
32 | - co = qemu_coroutine_create(co_entry, &rwco); | ||
33 | - qemu_coroutine_enter(co); | ||
34 | - BDRV_POLL_WHILE(blk_bs(blk), rwco.ret == NOT_DONE); | ||
35 | + if (qemu_in_coroutine()) { | ||
36 | + /* Fast-path if already in coroutine context */ | ||
37 | + co_entry(&rwco); | ||
38 | + } else { | ||
39 | + Coroutine *co = qemu_coroutine_create(co_entry, &rwco); | ||
40 | + qemu_coroutine_enter(co); | ||
41 | + BDRV_POLL_WHILE(blk_bs(blk), rwco.ret == NOT_DONE); | ||
42 | + } | ||
43 | |||
44 | return rwco.ret; | ||
45 | } | ||
46 | -- | ||
47 | 2.9.3 | ||
48 | |||
49 | diff view generated by jsdifflib |
New patch | |||
---|---|---|---|
1 | From: Paolo Bonzini <pbonzini@redhat.com> | ||
1 | 2 | ||
3 | Once the thread pool starts using aio_co_wake, it will also need | ||
4 | qemu_get_current_aio_context(). Make test-thread-pool create | ||
5 | an AioContext with qemu_init_main_loop, so that stubs/iothread.c | ||
6 | and tests/iothread.c can provide the rest. | ||
7 | |||
8 | Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
9 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> | ||
10 | Reviewed-by: Fam Zheng <famz@redhat.com> | ||
11 | Message-id: 20170213135235.12274-5-pbonzini@redhat.com | ||
12 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
13 | --- | ||
14 | tests/test-thread-pool.c | 12 +++--------- | ||
15 | 1 file changed, 3 insertions(+), 9 deletions(-) | ||
16 | |||
17 | diff --git a/tests/test-thread-pool.c b/tests/test-thread-pool.c | ||
18 | index XXXXXXX..XXXXXXX 100644 | ||
19 | --- a/tests/test-thread-pool.c | ||
20 | +++ b/tests/test-thread-pool.c | ||
21 | @@ -XXX,XX +XXX,XX @@ | ||
22 | #include "qapi/error.h" | ||
23 | #include "qemu/timer.h" | ||
24 | #include "qemu/error-report.h" | ||
25 | +#include "qemu/main-loop.h" | ||
26 | |||
27 | static AioContext *ctx; | ||
28 | static ThreadPool *pool; | ||
29 | @@ -XXX,XX +XXX,XX @@ static void test_cancel_async(void) | ||
30 | int main(int argc, char **argv) | ||
31 | { | ||
32 | int ret; | ||
33 | - Error *local_error = NULL; | ||
34 | |||
35 | - init_clocks(); | ||
36 | - | ||
37 | - ctx = aio_context_new(&local_error); | ||
38 | - if (!ctx) { | ||
39 | - error_reportf_err(local_error, "Failed to create AIO Context: "); | ||
40 | - exit(1); | ||
41 | - } | ||
42 | + qemu_init_main_loop(&error_abort); | ||
43 | + ctx = qemu_get_current_aio_context(); | ||
44 | pool = aio_get_thread_pool(ctx); | ||
45 | |||
46 | g_test_init(&argc, &argv, NULL); | ||
47 | @@ -XXX,XX +XXX,XX @@ int main(int argc, char **argv) | ||
48 | |||
49 | ret = g_test_run(); | ||
50 | |||
51 | - aio_context_unref(ctx); | ||
52 | return ret; | ||
53 | } | ||
54 | -- | ||
55 | 2.9.3 | ||
56 | |||
57 | diff view generated by jsdifflib |
New patch | |||
---|---|---|---|
1 | 1 | From: Paolo Bonzini <pbonzini@redhat.com> | |
2 | |||
3 | This is in preparation for making qio_channel_yield work on | ||
4 | AioContexts other than the main one. | ||
5 | |||
6 | Reviewed-by: Daniel P. Berrange <berrange@redhat.com> | ||
7 | Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
8 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> | ||
9 | Reviewed-by: Fam Zheng <famz@redhat.com> | ||
10 | Message-id: 20170213135235.12274-6-pbonzini@redhat.com | ||
11 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
12 | --- | ||
13 | include/io/channel.h | 25 +++++++++++++++++++++++++ | ||
14 | io/channel-command.c | 13 +++++++++++++ | ||
15 | io/channel-file.c | 11 +++++++++++ | ||
16 | io/channel-socket.c | 16 +++++++++++----- | ||
17 | io/channel-tls.c | 12 ++++++++++++ | ||
18 | io/channel-watch.c | 6 ++++++ | ||
19 | io/channel.c | 11 +++++++++++ | ||
20 | 7 files changed, 89 insertions(+), 5 deletions(-) | ||
21 | |||
22 | diff --git a/include/io/channel.h b/include/io/channel.h | ||
23 | index XXXXXXX..XXXXXXX 100644 | ||
24 | --- a/include/io/channel.h | ||
25 | +++ b/include/io/channel.h | ||
26 | @@ -XXX,XX +XXX,XX @@ | ||
27 | |||
28 | #include "qemu-common.h" | ||
29 | #include "qom/object.h" | ||
30 | +#include "block/aio.h" | ||
31 | |||
32 | #define TYPE_QIO_CHANNEL "qio-channel" | ||
33 | #define QIO_CHANNEL(obj) \ | ||
34 | @@ -XXX,XX +XXX,XX @@ struct QIOChannelClass { | ||
35 | off_t offset, | ||
36 | int whence, | ||
37 | Error **errp); | ||
38 | + void (*io_set_aio_fd_handler)(QIOChannel *ioc, | ||
39 | + AioContext *ctx, | ||
40 | + IOHandler *io_read, | ||
41 | + IOHandler *io_write, | ||
42 | + void *opaque); | ||
43 | }; | ||
44 | |||
45 | /* General I/O handling functions */ | ||
46 | @@ -XXX,XX +XXX,XX @@ void qio_channel_yield(QIOChannel *ioc, | ||
47 | void qio_channel_wait(QIOChannel *ioc, | ||
48 | GIOCondition condition); | ||
49 | |||
50 | +/** | ||
51 | + * qio_channel_set_aio_fd_handler: | ||
52 | + * @ioc: the channel object | ||
53 | + * @ctx: the AioContext to set the handlers on | ||
54 | + * @io_read: the read handler | ||
55 | + * @io_write: the write handler | ||
56 | + * @opaque: the opaque value passed to the handler | ||
57 | + * | ||
58 | + * This is used internally by qio_channel_yield(). It can | ||
59 | + * be used by channel implementations to forward the handlers | ||
60 | + * to another channel (e.g. from #QIOChannelTLS to the | ||
61 | + * underlying socket). | ||
62 | + */ | ||
63 | +void qio_channel_set_aio_fd_handler(QIOChannel *ioc, | ||
64 | + AioContext *ctx, | ||
65 | + IOHandler *io_read, | ||
66 | + IOHandler *io_write, | ||
67 | + void *opaque); | ||
68 | + | ||
69 | #endif /* QIO_CHANNEL_H */ | ||
70 | diff --git a/io/channel-command.c b/io/channel-command.c | ||
71 | index XXXXXXX..XXXXXXX 100644 | ||
72 | --- a/io/channel-command.c | ||
73 | +++ b/io/channel-command.c | ||
74 | @@ -XXX,XX +XXX,XX @@ static int qio_channel_command_close(QIOChannel *ioc, | ||
75 | } | ||
76 | |||
77 | |||
78 | +static void qio_channel_command_set_aio_fd_handler(QIOChannel *ioc, | ||
79 | + AioContext *ctx, | ||
80 | + IOHandler *io_read, | ||
81 | + IOHandler *io_write, | ||
82 | + void *opaque) | ||
83 | +{ | ||
84 | + QIOChannelCommand *cioc = QIO_CHANNEL_COMMAND(ioc); | ||
85 | + aio_set_fd_handler(ctx, cioc->readfd, false, io_read, NULL, NULL, opaque); | ||
86 | + aio_set_fd_handler(ctx, cioc->writefd, false, NULL, io_write, NULL, opaque); | ||
87 | +} | ||
88 | + | ||
89 | + | ||
90 | static GSource *qio_channel_command_create_watch(QIOChannel *ioc, | ||
91 | GIOCondition condition) | ||
92 | { | ||
93 | @@ -XXX,XX +XXX,XX @@ static void qio_channel_command_class_init(ObjectClass *klass, | ||
94 | ioc_klass->io_set_blocking = qio_channel_command_set_blocking; | ||
95 | ioc_klass->io_close = qio_channel_command_close; | ||
96 | ioc_klass->io_create_watch = qio_channel_command_create_watch; | ||
97 | + ioc_klass->io_set_aio_fd_handler = qio_channel_command_set_aio_fd_handler; | ||
98 | } | ||
99 | |||
100 | static const TypeInfo qio_channel_command_info = { | ||
101 | diff --git a/io/channel-file.c b/io/channel-file.c | ||
102 | index XXXXXXX..XXXXXXX 100644 | ||
103 | --- a/io/channel-file.c | ||
104 | +++ b/io/channel-file.c | ||
105 | @@ -XXX,XX +XXX,XX @@ static int qio_channel_file_close(QIOChannel *ioc, | ||
106 | } | ||
107 | |||
108 | |||
109 | +static void qio_channel_file_set_aio_fd_handler(QIOChannel *ioc, | ||
110 | + AioContext *ctx, | ||
111 | + IOHandler *io_read, | ||
112 | + IOHandler *io_write, | ||
113 | + void *opaque) | ||
114 | +{ | ||
115 | + QIOChannelFile *fioc = QIO_CHANNEL_FILE(ioc); | ||
116 | + aio_set_fd_handler(ctx, fioc->fd, false, io_read, io_write, NULL, opaque); | ||
117 | +} | ||
118 | + | ||
119 | static GSource *qio_channel_file_create_watch(QIOChannel *ioc, | ||
120 | GIOCondition condition) | ||
121 | { | ||
122 | @@ -XXX,XX +XXX,XX @@ static void qio_channel_file_class_init(ObjectClass *klass, | ||
123 | ioc_klass->io_seek = qio_channel_file_seek; | ||
124 | ioc_klass->io_close = qio_channel_file_close; | ||
125 | ioc_klass->io_create_watch = qio_channel_file_create_watch; | ||
126 | + ioc_klass->io_set_aio_fd_handler = qio_channel_file_set_aio_fd_handler; | ||
127 | } | ||
128 | |||
129 | static const TypeInfo qio_channel_file_info = { | ||
130 | diff --git a/io/channel-socket.c b/io/channel-socket.c | ||
131 | index XXXXXXX..XXXXXXX 100644 | ||
132 | --- a/io/channel-socket.c | ||
133 | +++ b/io/channel-socket.c | ||
134 | @@ -XXX,XX +XXX,XX @@ qio_channel_socket_set_blocking(QIOChannel *ioc, | ||
135 | qemu_set_block(sioc->fd); | ||
136 | } else { | ||
137 | qemu_set_nonblock(sioc->fd); | ||
138 | -#ifdef WIN32 | ||
139 | - WSAEventSelect(sioc->fd, ioc->event, | ||
140 | - FD_READ | FD_ACCEPT | FD_CLOSE | | ||
141 | - FD_CONNECT | FD_WRITE | FD_OOB); | ||
142 | -#endif | ||
143 | } | ||
144 | return 0; | ||
145 | } | ||
146 | @@ -XXX,XX +XXX,XX @@ qio_channel_socket_shutdown(QIOChannel *ioc, | ||
147 | return 0; | ||
148 | } | ||
149 | |||
150 | +static void qio_channel_socket_set_aio_fd_handler(QIOChannel *ioc, | ||
151 | + AioContext *ctx, | ||
152 | + IOHandler *io_read, | ||
153 | + IOHandler *io_write, | ||
154 | + void *opaque) | ||
155 | +{ | ||
156 | + QIOChannelSocket *sioc = QIO_CHANNEL_SOCKET(ioc); | ||
157 | + aio_set_fd_handler(ctx, sioc->fd, false, io_read, io_write, NULL, opaque); | ||
158 | +} | ||
159 | + | ||
160 | static GSource *qio_channel_socket_create_watch(QIOChannel *ioc, | ||
161 | GIOCondition condition) | ||
162 | { | ||
163 | @@ -XXX,XX +XXX,XX @@ static void qio_channel_socket_class_init(ObjectClass *klass, | ||
164 | ioc_klass->io_set_cork = qio_channel_socket_set_cork; | ||
165 | ioc_klass->io_set_delay = qio_channel_socket_set_delay; | ||
166 | ioc_klass->io_create_watch = qio_channel_socket_create_watch; | ||
167 | + ioc_klass->io_set_aio_fd_handler = qio_channel_socket_set_aio_fd_handler; | ||
168 | } | ||
169 | |||
170 | static const TypeInfo qio_channel_socket_info = { | ||
171 | diff --git a/io/channel-tls.c b/io/channel-tls.c | ||
172 | index XXXXXXX..XXXXXXX 100644 | ||
173 | --- a/io/channel-tls.c | ||
174 | +++ b/io/channel-tls.c | ||
175 | @@ -XXX,XX +XXX,XX @@ static int qio_channel_tls_close(QIOChannel *ioc, | ||
176 | return qio_channel_close(tioc->master, errp); | ||
177 | } | ||
178 | |||
179 | +static void qio_channel_tls_set_aio_fd_handler(QIOChannel *ioc, | ||
180 | + AioContext *ctx, | ||
181 | + IOHandler *io_read, | ||
182 | + IOHandler *io_write, | ||
183 | + void *opaque) | ||
184 | +{ | ||
185 | + QIOChannelTLS *tioc = QIO_CHANNEL_TLS(ioc); | ||
186 | + | ||
187 | + qio_channel_set_aio_fd_handler(tioc->master, ctx, io_read, io_write, opaque); | ||
188 | +} | ||
189 | + | ||
190 | static GSource *qio_channel_tls_create_watch(QIOChannel *ioc, | ||
191 | GIOCondition condition) | ||
192 | { | ||
193 | @@ -XXX,XX +XXX,XX @@ static void qio_channel_tls_class_init(ObjectClass *klass, | ||
194 | ioc_klass->io_close = qio_channel_tls_close; | ||
195 | ioc_klass->io_shutdown = qio_channel_tls_shutdown; | ||
196 | ioc_klass->io_create_watch = qio_channel_tls_create_watch; | ||
197 | + ioc_klass->io_set_aio_fd_handler = qio_channel_tls_set_aio_fd_handler; | ||
198 | } | ||
199 | |||
200 | static const TypeInfo qio_channel_tls_info = { | ||
201 | diff --git a/io/channel-watch.c b/io/channel-watch.c | ||
202 | index XXXXXXX..XXXXXXX 100644 | ||
203 | --- a/io/channel-watch.c | ||
204 | +++ b/io/channel-watch.c | ||
205 | @@ -XXX,XX +XXX,XX @@ GSource *qio_channel_create_socket_watch(QIOChannel *ioc, | ||
206 | GSource *source; | ||
207 | QIOChannelSocketSource *ssource; | ||
208 | |||
209 | +#ifdef WIN32 | ||
210 | + WSAEventSelect(socket, ioc->event, | ||
211 | + FD_READ | FD_ACCEPT | FD_CLOSE | | ||
212 | + FD_CONNECT | FD_WRITE | FD_OOB); | ||
213 | +#endif | ||
214 | + | ||
215 | source = g_source_new(&qio_channel_socket_source_funcs, | ||
216 | sizeof(QIOChannelSocketSource)); | ||
217 | ssource = (QIOChannelSocketSource *)source; | ||
218 | diff --git a/io/channel.c b/io/channel.c | ||
219 | index XXXXXXX..XXXXXXX 100644 | ||
220 | --- a/io/channel.c | ||
221 | +++ b/io/channel.c | ||
222 | @@ -XXX,XX +XXX,XX @@ GSource *qio_channel_create_watch(QIOChannel *ioc, | ||
223 | } | ||
224 | |||
225 | |||
226 | +void qio_channel_set_aio_fd_handler(QIOChannel *ioc, | ||
227 | + AioContext *ctx, | ||
228 | + IOHandler *io_read, | ||
229 | + IOHandler *io_write, | ||
230 | + void *opaque) | ||
231 | +{ | ||
232 | + QIOChannelClass *klass = QIO_CHANNEL_GET_CLASS(ioc); | ||
233 | + | ||
234 | + klass->io_set_aio_fd_handler(ioc, ctx, io_read, io_write, opaque); | ||
235 | +} | ||
236 | + | ||
237 | guint qio_channel_add_watch(QIOChannel *ioc, | ||
238 | GIOCondition condition, | ||
239 | QIOChannelFunc func, | ||
240 | -- | ||
241 | 2.9.3 | ||
242 | |||
243 | diff view generated by jsdifflib |
New patch | |||
---|---|---|---|
1 | 1 | From: Paolo Bonzini <pbonzini@redhat.com> | |
2 | |||
3 | Support separate coroutines for reading and writing, and place the | ||
4 | read/write handlers on the AioContext that the QIOChannel is registered | ||
5 | with. | ||
6 | |||
7 | Reviewed-by: Daniel P. Berrange <berrange@redhat.com> | ||
8 | Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
9 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> | ||
10 | Reviewed-by: Fam Zheng <famz@redhat.com> | ||
11 | Message-id: 20170213135235.12274-7-pbonzini@redhat.com | ||
12 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
13 | --- | ||
14 | include/io/channel.h | 47 ++++++++++++++++++++++++++-- | ||
15 | io/channel.c | 86 +++++++++++++++++++++++++++++++++++++++------------- | ||
16 | 2 files changed, 109 insertions(+), 24 deletions(-) | ||
17 | |||
18 | diff --git a/include/io/channel.h b/include/io/channel.h | ||
19 | index XXXXXXX..XXXXXXX 100644 | ||
20 | --- a/include/io/channel.h | ||
21 | +++ b/include/io/channel.h | ||
22 | @@ -XXX,XX +XXX,XX @@ | ||
23 | |||
24 | #include "qemu-common.h" | ||
25 | #include "qom/object.h" | ||
26 | +#include "qemu/coroutine.h" | ||
27 | #include "block/aio.h" | ||
28 | |||
29 | #define TYPE_QIO_CHANNEL "qio-channel" | ||
30 | @@ -XXX,XX +XXX,XX @@ struct QIOChannel { | ||
31 | Object parent; | ||
32 | unsigned int features; /* bitmask of QIOChannelFeatures */ | ||
33 | char *name; | ||
34 | + AioContext *ctx; | ||
35 | + Coroutine *read_coroutine; | ||
36 | + Coroutine *write_coroutine; | ||
37 | #ifdef _WIN32 | ||
38 | HANDLE event; /* For use with GSource on Win32 */ | ||
39 | #endif | ||
40 | @@ -XXX,XX +XXX,XX @@ guint qio_channel_add_watch(QIOChannel *ioc, | ||
41 | |||
42 | |||
43 | /** | ||
44 | + * qio_channel_attach_aio_context: | ||
45 | + * @ioc: the channel object | ||
46 | + * @ctx: the #AioContext to set the handlers on | ||
47 | + * | ||
48 | + * Request that qio_channel_yield() sets I/O handlers on | ||
49 | + * the given #AioContext. If @ctx is %NULL, qio_channel_yield() | ||
50 | + * uses QEMU's main thread event loop. | ||
51 | + * | ||
52 | + * You can move a #QIOChannel from one #AioContext to another even if | ||
53 | + * I/O handlers are set for a coroutine. However, #QIOChannel provides | ||
54 | + * no synchronization between the calls to qio_channel_yield() and | ||
55 | + * qio_channel_attach_aio_context(). | ||
56 | + * | ||
57 | + * Therefore you should first call qio_channel_detach_aio_context() | ||
58 | + * to ensure that the coroutine is not entered concurrently. Then, | ||
59 | + * while the coroutine has yielded, call qio_channel_attach_aio_context(), | ||
60 | + * and then aio_co_schedule() to place the coroutine on the new | ||
61 | + * #AioContext. The calls to qio_channel_detach_aio_context() | ||
62 | + * and qio_channel_attach_aio_context() should be protected with | ||
63 | + * aio_context_acquire() and aio_context_release(). | ||
64 | + */ | ||
65 | +void qio_channel_attach_aio_context(QIOChannel *ioc, | ||
66 | + AioContext *ctx); | ||
67 | + | ||
68 | +/** | ||
69 | + * qio_channel_detach_aio_context: | ||
70 | + * @ioc: the channel object | ||
71 | + * | ||
72 | + * Disable any I/O handlers set by qio_channel_yield(). With the | ||
73 | + * help of aio_co_schedule(), this allows moving a coroutine that was | ||
74 | + * paused by qio_channel_yield() to another context. | ||
75 | + */ | ||
76 | +void qio_channel_detach_aio_context(QIOChannel *ioc); | ||
77 | + | ||
78 | +/** | ||
79 | * qio_channel_yield: | ||
80 | * @ioc: the channel object | ||
81 | * @condition: the I/O condition to wait for | ||
82 | * | ||
83 | - * Yields execution from the current coroutine until | ||
84 | - * the condition indicated by @condition becomes | ||
85 | - * available. | ||
86 | + * Yields execution from the current coroutine until the condition | ||
87 | + * indicated by @condition becomes available. @condition must | ||
88 | + * be either %G_IO_IN or %G_IO_OUT; it cannot contain both. In | ||
89 | + * addition, no two coroutine can be waiting on the same condition | ||
90 | + * and channel at the same time. | ||
91 | * | ||
92 | * This must only be called from coroutine context | ||
93 | */ | ||
94 | diff --git a/io/channel.c b/io/channel.c | ||
95 | index XXXXXXX..XXXXXXX 100644 | ||
96 | --- a/io/channel.c | ||
97 | +++ b/io/channel.c | ||
98 | @@ -XXX,XX +XXX,XX @@ | ||
99 | #include "qemu/osdep.h" | ||
100 | #include "io/channel.h" | ||
101 | #include "qapi/error.h" | ||
102 | -#include "qemu/coroutine.h" | ||
103 | +#include "qemu/main-loop.h" | ||
104 | |||
105 | bool qio_channel_has_feature(QIOChannel *ioc, | ||
106 | QIOChannelFeature feature) | ||
107 | @@ -XXX,XX +XXX,XX @@ off_t qio_channel_io_seek(QIOChannel *ioc, | ||
108 | } | ||
109 | |||
110 | |||
111 | -typedef struct QIOChannelYieldData QIOChannelYieldData; | ||
112 | -struct QIOChannelYieldData { | ||
113 | - QIOChannel *ioc; | ||
114 | - Coroutine *co; | ||
115 | -}; | ||
116 | +static void qio_channel_set_aio_fd_handlers(QIOChannel *ioc); | ||
117 | |||
118 | +static void qio_channel_restart_read(void *opaque) | ||
119 | +{ | ||
120 | + QIOChannel *ioc = opaque; | ||
121 | + Coroutine *co = ioc->read_coroutine; | ||
122 | + | ||
123 | + ioc->read_coroutine = NULL; | ||
124 | + qio_channel_set_aio_fd_handlers(ioc); | ||
125 | + aio_co_wake(co); | ||
126 | +} | ||
127 | |||
128 | -static gboolean qio_channel_yield_enter(QIOChannel *ioc, | ||
129 | - GIOCondition condition, | ||
130 | - gpointer opaque) | ||
131 | +static void qio_channel_restart_write(void *opaque) | ||
132 | { | ||
133 | - QIOChannelYieldData *data = opaque; | ||
134 | - qemu_coroutine_enter(data->co); | ||
135 | - return FALSE; | ||
136 | + QIOChannel *ioc = opaque; | ||
137 | + Coroutine *co = ioc->write_coroutine; | ||
138 | + | ||
139 | + ioc->write_coroutine = NULL; | ||
140 | + qio_channel_set_aio_fd_handlers(ioc); | ||
141 | + aio_co_wake(co); | ||
142 | } | ||
143 | |||
144 | +static void qio_channel_set_aio_fd_handlers(QIOChannel *ioc) | ||
145 | +{ | ||
146 | + IOHandler *rd_handler = NULL, *wr_handler = NULL; | ||
147 | + AioContext *ctx; | ||
148 | + | ||
149 | + if (ioc->read_coroutine) { | ||
150 | + rd_handler = qio_channel_restart_read; | ||
151 | + } | ||
152 | + if (ioc->write_coroutine) { | ||
153 | + wr_handler = qio_channel_restart_write; | ||
154 | + } | ||
155 | + | ||
156 | + ctx = ioc->ctx ? ioc->ctx : iohandler_get_aio_context(); | ||
157 | + qio_channel_set_aio_fd_handler(ioc, ctx, rd_handler, wr_handler, ioc); | ||
158 | +} | ||
159 | + | ||
160 | +void qio_channel_attach_aio_context(QIOChannel *ioc, | ||
161 | + AioContext *ctx) | ||
162 | +{ | ||
163 | + AioContext *old_ctx; | ||
164 | + if (ioc->ctx == ctx) { | ||
165 | + return; | ||
166 | + } | ||
167 | + | ||
168 | + old_ctx = ioc->ctx ? ioc->ctx : iohandler_get_aio_context(); | ||
169 | + qio_channel_set_aio_fd_handler(ioc, old_ctx, NULL, NULL, NULL); | ||
170 | + ioc->ctx = ctx; | ||
171 | + qio_channel_set_aio_fd_handlers(ioc); | ||
172 | +} | ||
173 | + | ||
174 | +void qio_channel_detach_aio_context(QIOChannel *ioc) | ||
175 | +{ | ||
176 | + ioc->read_coroutine = NULL; | ||
177 | + ioc->write_coroutine = NULL; | ||
178 | + qio_channel_set_aio_fd_handlers(ioc); | ||
179 | + ioc->ctx = NULL; | ||
180 | +} | ||
181 | |||
182 | void coroutine_fn qio_channel_yield(QIOChannel *ioc, | ||
183 | GIOCondition condition) | ||
184 | { | ||
185 | - QIOChannelYieldData data; | ||
186 | - | ||
187 | assert(qemu_in_coroutine()); | ||
188 | - data.ioc = ioc; | ||
189 | - data.co = qemu_coroutine_self(); | ||
190 | - qio_channel_add_watch(ioc, | ||
191 | - condition, | ||
192 | - qio_channel_yield_enter, | ||
193 | - &data, | ||
194 | - NULL); | ||
195 | + if (condition == G_IO_IN) { | ||
196 | + assert(!ioc->read_coroutine); | ||
197 | + ioc->read_coroutine = qemu_coroutine_self(); | ||
198 | + } else if (condition == G_IO_OUT) { | ||
199 | + assert(!ioc->write_coroutine); | ||
200 | + ioc->write_coroutine = qemu_coroutine_self(); | ||
201 | + } else { | ||
202 | + abort(); | ||
203 | + } | ||
204 | + qio_channel_set_aio_fd_handlers(ioc); | ||
205 | qemu_coroutine_yield(); | ||
206 | } | ||
207 | |||
208 | -- | ||
209 | 2.9.3 | ||
210 | |||
211 | diff view generated by jsdifflib |
New patch | |||
---|---|---|---|
1 | From: Paolo Bonzini <pbonzini@redhat.com> | ||
1 | 2 | ||
3 | In the client, read the reply headers from a coroutine, switching the | ||
4 | read side between the "read header" coroutine and the I/O coroutine that | ||
5 | reads the body of the reply. | ||
6 | |||
7 | In the server, if the server can read more requests it will create a new | ||
8 | "read request" coroutine as soon as a request has been read. Otherwise, | ||
9 | the new coroutine is created in nbd_request_put. | ||
10 | |||
11 | Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
12 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> | ||
13 | Reviewed-by: Fam Zheng <famz@redhat.com> | ||
14 | Reviewed-by: Daniel P. Berrange <berrange@redhat.com> | ||
15 | Message-id: 20170213135235.12274-8-pbonzini@redhat.com | ||
16 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
17 | --- | ||
18 | block/nbd-client.h | 2 +- | ||
19 | block/nbd-client.c | 117 ++++++++++++++++++++++++----------------------------- | ||
20 | nbd/client.c | 2 +- | ||
21 | nbd/common.c | 9 +---- | ||
22 | nbd/server.c | 94 +++++++++++++----------------------------- | ||
23 | 5 files changed, 83 insertions(+), 141 deletions(-) | ||
24 | |||
25 | diff --git a/block/nbd-client.h b/block/nbd-client.h | ||
26 | index XXXXXXX..XXXXXXX 100644 | ||
27 | --- a/block/nbd-client.h | ||
28 | +++ b/block/nbd-client.h | ||
29 | @@ -XXX,XX +XXX,XX @@ typedef struct NBDClientSession { | ||
30 | |||
31 | CoMutex send_mutex; | ||
32 | CoQueue free_sema; | ||
33 | - Coroutine *send_coroutine; | ||
34 | + Coroutine *read_reply_co; | ||
35 | int in_flight; | ||
36 | |||
37 | Coroutine *recv_coroutine[MAX_NBD_REQUESTS]; | ||
38 | diff --git a/block/nbd-client.c b/block/nbd-client.c | ||
39 | index XXXXXXX..XXXXXXX 100644 | ||
40 | --- a/block/nbd-client.c | ||
41 | +++ b/block/nbd-client.c | ||
42 | @@ -XXX,XX +XXX,XX @@ | ||
43 | #define HANDLE_TO_INDEX(bs, handle) ((handle) ^ ((uint64_t)(intptr_t)bs)) | ||
44 | #define INDEX_TO_HANDLE(bs, index) ((index) ^ ((uint64_t)(intptr_t)bs)) | ||
45 | |||
46 | -static void nbd_recv_coroutines_enter_all(NBDClientSession *s) | ||
47 | +static void nbd_recv_coroutines_enter_all(BlockDriverState *bs) | ||
48 | { | ||
49 | + NBDClientSession *s = nbd_get_client_session(bs); | ||
50 | int i; | ||
51 | |||
52 | for (i = 0; i < MAX_NBD_REQUESTS; i++) { | ||
53 | @@ -XXX,XX +XXX,XX @@ static void nbd_recv_coroutines_enter_all(NBDClientSession *s) | ||
54 | qemu_coroutine_enter(s->recv_coroutine[i]); | ||
55 | } | ||
56 | } | ||
57 | + BDRV_POLL_WHILE(bs, s->read_reply_co); | ||
58 | } | ||
59 | |||
60 | static void nbd_teardown_connection(BlockDriverState *bs) | ||
61 | @@ -XXX,XX +XXX,XX @@ static void nbd_teardown_connection(BlockDriverState *bs) | ||
62 | qio_channel_shutdown(client->ioc, | ||
63 | QIO_CHANNEL_SHUTDOWN_BOTH, | ||
64 | NULL); | ||
65 | - nbd_recv_coroutines_enter_all(client); | ||
66 | + nbd_recv_coroutines_enter_all(bs); | ||
67 | |||
68 | nbd_client_detach_aio_context(bs); | ||
69 | object_unref(OBJECT(client->sioc)); | ||
70 | @@ -XXX,XX +XXX,XX @@ static void nbd_teardown_connection(BlockDriverState *bs) | ||
71 | client->ioc = NULL; | ||
72 | } | ||
73 | |||
74 | -static void nbd_reply_ready(void *opaque) | ||
75 | +static coroutine_fn void nbd_read_reply_entry(void *opaque) | ||
76 | { | ||
77 | - BlockDriverState *bs = opaque; | ||
78 | - NBDClientSession *s = nbd_get_client_session(bs); | ||
79 | + NBDClientSession *s = opaque; | ||
80 | uint64_t i; | ||
81 | int ret; | ||
82 | |||
83 | - if (!s->ioc) { /* Already closed */ | ||
84 | - return; | ||
85 | - } | ||
86 | - | ||
87 | - if (s->reply.handle == 0) { | ||
88 | - /* No reply already in flight. Fetch a header. It is possible | ||
89 | - * that another thread has done the same thing in parallel, so | ||
90 | - * the socket is not readable anymore. | ||
91 | - */ | ||
92 | + for (;;) { | ||
93 | + assert(s->reply.handle == 0); | ||
94 | ret = nbd_receive_reply(s->ioc, &s->reply); | ||
95 | - if (ret == -EAGAIN) { | ||
96 | - return; | ||
97 | - } | ||
98 | if (ret < 0) { | ||
99 | - s->reply.handle = 0; | ||
100 | - goto fail; | ||
101 | + break; | ||
102 | } | ||
103 | - } | ||
104 | |||
105 | - /* There's no need for a mutex on the receive side, because the | ||
106 | - * handler acts as a synchronization point and ensures that only | ||
107 | - * one coroutine is called until the reply finishes. */ | ||
108 | - i = HANDLE_TO_INDEX(s, s->reply.handle); | ||
109 | - if (i >= MAX_NBD_REQUESTS) { | ||
110 | - goto fail; | ||
111 | - } | ||
112 | + /* There's no need for a mutex on the receive side, because the | ||
113 | + * handler acts as a synchronization point and ensures that only | ||
114 | + * one coroutine is called until the reply finishes. | ||
115 | + */ | ||
116 | + i = HANDLE_TO_INDEX(s, s->reply.handle); | ||
117 | + if (i >= MAX_NBD_REQUESTS || !s->recv_coroutine[i]) { | ||
118 | + break; | ||
119 | + } | ||
120 | |||
121 | - if (s->recv_coroutine[i]) { | ||
122 | - qemu_coroutine_enter(s->recv_coroutine[i]); | ||
123 | - return; | ||
124 | + /* We're woken up by the recv_coroutine itself. Note that there | ||
125 | + * is no race between yielding and reentering read_reply_co. This | ||
126 | + * is because: | ||
127 | + * | ||
128 | + * - if recv_coroutine[i] runs on the same AioContext, it is only | ||
129 | + * entered after we yield | ||
130 | + * | ||
131 | + * - if recv_coroutine[i] runs on a different AioContext, reentering | ||
132 | + * read_reply_co happens through a bottom half, which can only | ||
133 | + * run after we yield. | ||
134 | + */ | ||
135 | + aio_co_wake(s->recv_coroutine[i]); | ||
136 | + qemu_coroutine_yield(); | ||
137 | } | ||
138 | - | ||
139 | -fail: | ||
140 | - nbd_teardown_connection(bs); | ||
141 | -} | ||
142 | - | ||
143 | -static void nbd_restart_write(void *opaque) | ||
144 | -{ | ||
145 | - BlockDriverState *bs = opaque; | ||
146 | - | ||
147 | - qemu_coroutine_enter(nbd_get_client_session(bs)->send_coroutine); | ||
148 | + s->read_reply_co = NULL; | ||
149 | } | ||
150 | |||
151 | static int nbd_co_send_request(BlockDriverState *bs, | ||
152 | @@ -XXX,XX +XXX,XX @@ static int nbd_co_send_request(BlockDriverState *bs, | ||
153 | QEMUIOVector *qiov) | ||
154 | { | ||
155 | NBDClientSession *s = nbd_get_client_session(bs); | ||
156 | - AioContext *aio_context; | ||
157 | int rc, ret, i; | ||
158 | |||
159 | qemu_co_mutex_lock(&s->send_mutex); | ||
160 | @@ -XXX,XX +XXX,XX @@ static int nbd_co_send_request(BlockDriverState *bs, | ||
161 | return -EPIPE; | ||
162 | } | ||
163 | |||
164 | - s->send_coroutine = qemu_coroutine_self(); | ||
165 | - aio_context = bdrv_get_aio_context(bs); | ||
166 | - | ||
167 | - aio_set_fd_handler(aio_context, s->sioc->fd, false, | ||
168 | - nbd_reply_ready, nbd_restart_write, NULL, bs); | ||
169 | if (qiov) { | ||
170 | qio_channel_set_cork(s->ioc, true); | ||
171 | rc = nbd_send_request(s->ioc, request); | ||
172 | @@ -XXX,XX +XXX,XX @@ static int nbd_co_send_request(BlockDriverState *bs, | ||
173 | } else { | ||
174 | rc = nbd_send_request(s->ioc, request); | ||
175 | } | ||
176 | - aio_set_fd_handler(aio_context, s->sioc->fd, false, | ||
177 | - nbd_reply_ready, NULL, NULL, bs); | ||
178 | - s->send_coroutine = NULL; | ||
179 | qemu_co_mutex_unlock(&s->send_mutex); | ||
180 | return rc; | ||
181 | } | ||
182 | @@ -XXX,XX +XXX,XX @@ static void nbd_co_receive_reply(NBDClientSession *s, | ||
183 | { | ||
184 | int ret; | ||
185 | |||
186 | - /* Wait until we're woken up by the read handler. TODO: perhaps | ||
187 | - * peek at the next reply and avoid yielding if it's ours? */ | ||
188 | + /* Wait until we're woken up by nbd_read_reply_entry. */ | ||
189 | qemu_coroutine_yield(); | ||
190 | *reply = s->reply; | ||
191 | if (reply->handle != request->handle || | ||
192 | @@ -XXX,XX +XXX,XX @@ static void nbd_coroutine_start(NBDClientSession *s, | ||
193 | /* s->recv_coroutine[i] is set as soon as we get the send_lock. */ | ||
194 | } | ||
195 | |||
196 | -static void nbd_coroutine_end(NBDClientSession *s, | ||
197 | +static void nbd_coroutine_end(BlockDriverState *bs, | ||
198 | NBDRequest *request) | ||
199 | { | ||
200 | + NBDClientSession *s = nbd_get_client_session(bs); | ||
201 | int i = HANDLE_TO_INDEX(s, request->handle); | ||
202 | + | ||
203 | s->recv_coroutine[i] = NULL; | ||
204 | - if (s->in_flight-- == MAX_NBD_REQUESTS) { | ||
205 | - qemu_co_queue_next(&s->free_sema); | ||
206 | + s->in_flight--; | ||
207 | + qemu_co_queue_next(&s->free_sema); | ||
208 | + | ||
209 | + /* Kick the read_reply_co to get the next reply. */ | ||
210 | + if (s->read_reply_co) { | ||
211 | + aio_co_wake(s->read_reply_co); | ||
212 | } | ||
213 | } | ||
214 | |||
215 | @@ -XXX,XX +XXX,XX @@ int nbd_client_co_preadv(BlockDriverState *bs, uint64_t offset, | ||
216 | } else { | ||
217 | nbd_co_receive_reply(client, &request, &reply, qiov); | ||
218 | } | ||
219 | - nbd_coroutine_end(client, &request); | ||
220 | + nbd_coroutine_end(bs, &request); | ||
221 | return -reply.error; | ||
222 | } | ||
223 | |||
224 | @@ -XXX,XX +XXX,XX @@ int nbd_client_co_pwritev(BlockDriverState *bs, uint64_t offset, | ||
225 | } else { | ||
226 | nbd_co_receive_reply(client, &request, &reply, NULL); | ||
227 | } | ||
228 | - nbd_coroutine_end(client, &request); | ||
229 | + nbd_coroutine_end(bs, &request); | ||
230 | return -reply.error; | ||
231 | } | ||
232 | |||
233 | @@ -XXX,XX +XXX,XX @@ int nbd_client_co_pwrite_zeroes(BlockDriverState *bs, int64_t offset, | ||
234 | } else { | ||
235 | nbd_co_receive_reply(client, &request, &reply, NULL); | ||
236 | } | ||
237 | - nbd_coroutine_end(client, &request); | ||
238 | + nbd_coroutine_end(bs, &request); | ||
239 | return -reply.error; | ||
240 | } | ||
241 | |||
242 | @@ -XXX,XX +XXX,XX @@ int nbd_client_co_flush(BlockDriverState *bs) | ||
243 | } else { | ||
244 | nbd_co_receive_reply(client, &request, &reply, NULL); | ||
245 | } | ||
246 | - nbd_coroutine_end(client, &request); | ||
247 | + nbd_coroutine_end(bs, &request); | ||
248 | return -reply.error; | ||
249 | } | ||
250 | |||
251 | @@ -XXX,XX +XXX,XX @@ int nbd_client_co_pdiscard(BlockDriverState *bs, int64_t offset, int count) | ||
252 | } else { | ||
253 | nbd_co_receive_reply(client, &request, &reply, NULL); | ||
254 | } | ||
255 | - nbd_coroutine_end(client, &request); | ||
256 | + nbd_coroutine_end(bs, &request); | ||
257 | return -reply.error; | ||
258 | |||
259 | } | ||
260 | |||
261 | void nbd_client_detach_aio_context(BlockDriverState *bs) | ||
262 | { | ||
263 | - aio_set_fd_handler(bdrv_get_aio_context(bs), | ||
264 | - nbd_get_client_session(bs)->sioc->fd, | ||
265 | - false, NULL, NULL, NULL, NULL); | ||
266 | + NBDClientSession *client = nbd_get_client_session(bs); | ||
267 | + qio_channel_detach_aio_context(QIO_CHANNEL(client->sioc)); | ||
268 | } | ||
269 | |||
270 | void nbd_client_attach_aio_context(BlockDriverState *bs, | ||
271 | AioContext *new_context) | ||
272 | { | ||
273 | - aio_set_fd_handler(new_context, nbd_get_client_session(bs)->sioc->fd, | ||
274 | - false, nbd_reply_ready, NULL, NULL, bs); | ||
275 | + NBDClientSession *client = nbd_get_client_session(bs); | ||
276 | + qio_channel_attach_aio_context(QIO_CHANNEL(client->sioc), new_context); | ||
277 | + aio_co_schedule(new_context, client->read_reply_co); | ||
278 | } | ||
279 | |||
280 | void nbd_client_close(BlockDriverState *bs) | ||
281 | @@ -XXX,XX +XXX,XX @@ int nbd_client_init(BlockDriverState *bs, | ||
282 | /* Now that we're connected, set the socket to be non-blocking and | ||
283 | * kick the reply mechanism. */ | ||
284 | qio_channel_set_blocking(QIO_CHANNEL(sioc), false, NULL); | ||
285 | - | ||
286 | + client->read_reply_co = qemu_coroutine_create(nbd_read_reply_entry, client); | ||
287 | nbd_client_attach_aio_context(bs, bdrv_get_aio_context(bs)); | ||
288 | |||
289 | logout("Established connection with NBD server\n"); | ||
290 | diff --git a/nbd/client.c b/nbd/client.c | ||
291 | index XXXXXXX..XXXXXXX 100644 | ||
292 | --- a/nbd/client.c | ||
293 | +++ b/nbd/client.c | ||
294 | @@ -XXX,XX +XXX,XX @@ ssize_t nbd_receive_reply(QIOChannel *ioc, NBDReply *reply) | ||
295 | ssize_t ret; | ||
296 | |||
297 | ret = read_sync(ioc, buf, sizeof(buf)); | ||
298 | - if (ret < 0) { | ||
299 | + if (ret <= 0) { | ||
300 | return ret; | ||
301 | } | ||
302 | |||
303 | diff --git a/nbd/common.c b/nbd/common.c | ||
304 | index XXXXXXX..XXXXXXX 100644 | ||
305 | --- a/nbd/common.c | ||
306 | +++ b/nbd/common.c | ||
307 | @@ -XXX,XX +XXX,XX @@ ssize_t nbd_wr_syncv(QIOChannel *ioc, | ||
308 | } | ||
309 | if (len == QIO_CHANNEL_ERR_BLOCK) { | ||
310 | if (qemu_in_coroutine()) { | ||
311 | - /* XXX figure out if we can create a variant on | ||
312 | - * qio_channel_yield() that works with AIO contexts | ||
313 | - * and consider using that in this branch */ | ||
314 | - qemu_coroutine_yield(); | ||
315 | - } else if (done) { | ||
316 | - /* XXX this is needed by nbd_reply_ready. */ | ||
317 | - qio_channel_wait(ioc, | ||
318 | - do_read ? G_IO_IN : G_IO_OUT); | ||
319 | + qio_channel_yield(ioc, do_read ? G_IO_IN : G_IO_OUT); | ||
320 | } else { | ||
321 | return -EAGAIN; | ||
322 | } | ||
323 | diff --git a/nbd/server.c b/nbd/server.c | ||
324 | index XXXXXXX..XXXXXXX 100644 | ||
325 | --- a/nbd/server.c | ||
326 | +++ b/nbd/server.c | ||
327 | @@ -XXX,XX +XXX,XX @@ struct NBDClient { | ||
328 | CoMutex send_lock; | ||
329 | Coroutine *send_coroutine; | ||
330 | |||
331 | - bool can_read; | ||
332 | - | ||
333 | QTAILQ_ENTRY(NBDClient) next; | ||
334 | int nb_requests; | ||
335 | bool closing; | ||
336 | @@ -XXX,XX +XXX,XX @@ struct NBDClient { | ||
337 | |||
338 | /* That's all folks */ | ||
339 | |||
340 | -static void nbd_set_handlers(NBDClient *client); | ||
341 | -static void nbd_unset_handlers(NBDClient *client); | ||
342 | -static void nbd_update_can_read(NBDClient *client); | ||
343 | +static void nbd_client_receive_next_request(NBDClient *client); | ||
344 | |||
345 | static gboolean nbd_negotiate_continue(QIOChannel *ioc, | ||
346 | GIOCondition condition, | ||
347 | @@ -XXX,XX +XXX,XX @@ void nbd_client_put(NBDClient *client) | ||
348 | */ | ||
349 | assert(client->closing); | ||
350 | |||
351 | - nbd_unset_handlers(client); | ||
352 | + qio_channel_detach_aio_context(client->ioc); | ||
353 | object_unref(OBJECT(client->sioc)); | ||
354 | object_unref(OBJECT(client->ioc)); | ||
355 | if (client->tlscreds) { | ||
356 | @@ -XXX,XX +XXX,XX @@ static NBDRequestData *nbd_request_get(NBDClient *client) | ||
357 | |||
358 | assert(client->nb_requests <= MAX_NBD_REQUESTS - 1); | ||
359 | client->nb_requests++; | ||
360 | - nbd_update_can_read(client); | ||
361 | |||
362 | req = g_new0(NBDRequestData, 1); | ||
363 | nbd_client_get(client); | ||
364 | @@ -XXX,XX +XXX,XX @@ static void nbd_request_put(NBDRequestData *req) | ||
365 | g_free(req); | ||
366 | |||
367 | client->nb_requests--; | ||
368 | - nbd_update_can_read(client); | ||
369 | + nbd_client_receive_next_request(client); | ||
370 | + | ||
371 | nbd_client_put(client); | ||
372 | } | ||
373 | |||
374 | @@ -XXX,XX +XXX,XX @@ static void blk_aio_attached(AioContext *ctx, void *opaque) | ||
375 | exp->ctx = ctx; | ||
376 | |||
377 | QTAILQ_FOREACH(client, &exp->clients, next) { | ||
378 | - nbd_set_handlers(client); | ||
379 | + qio_channel_attach_aio_context(client->ioc, ctx); | ||
380 | + if (client->recv_coroutine) { | ||
381 | + aio_co_schedule(ctx, client->recv_coroutine); | ||
382 | + } | ||
383 | + if (client->send_coroutine) { | ||
384 | + aio_co_schedule(ctx, client->send_coroutine); | ||
385 | + } | ||
386 | } | ||
387 | } | ||
388 | |||
389 | @@ -XXX,XX +XXX,XX @@ static void blk_aio_detach(void *opaque) | ||
390 | TRACE("Export %s: Detaching clients from AIO context %p\n", exp->name, exp->ctx); | ||
391 | |||
392 | QTAILQ_FOREACH(client, &exp->clients, next) { | ||
393 | - nbd_unset_handlers(client); | ||
394 | + qio_channel_detach_aio_context(client->ioc); | ||
395 | } | ||
396 | |||
397 | exp->ctx = NULL; | ||
398 | @@ -XXX,XX +XXX,XX @@ static ssize_t nbd_co_send_reply(NBDRequestData *req, NBDReply *reply, | ||
399 | g_assert(qemu_in_coroutine()); | ||
400 | qemu_co_mutex_lock(&client->send_lock); | ||
401 | client->send_coroutine = qemu_coroutine_self(); | ||
402 | - nbd_set_handlers(client); | ||
403 | |||
404 | if (!len) { | ||
405 | rc = nbd_send_reply(client->ioc, reply); | ||
406 | @@ -XXX,XX +XXX,XX @@ static ssize_t nbd_co_send_reply(NBDRequestData *req, NBDReply *reply, | ||
407 | } | ||
408 | |||
409 | client->send_coroutine = NULL; | ||
410 | - nbd_set_handlers(client); | ||
411 | qemu_co_mutex_unlock(&client->send_lock); | ||
412 | return rc; | ||
413 | } | ||
414 | @@ -XXX,XX +XXX,XX @@ static ssize_t nbd_co_receive_request(NBDRequestData *req, | ||
415 | ssize_t rc; | ||
416 | |||
417 | g_assert(qemu_in_coroutine()); | ||
418 | - client->recv_coroutine = qemu_coroutine_self(); | ||
419 | - nbd_update_can_read(client); | ||
420 | - | ||
421 | + assert(client->recv_coroutine == qemu_coroutine_self()); | ||
422 | rc = nbd_receive_request(client->ioc, request); | ||
423 | if (rc < 0) { | ||
424 | if (rc != -EAGAIN) { | ||
425 | @@ -XXX,XX +XXX,XX @@ static ssize_t nbd_co_receive_request(NBDRequestData *req, | ||
426 | |||
427 | out: | ||
428 | client->recv_coroutine = NULL; | ||
429 | - nbd_update_can_read(client); | ||
430 | + nbd_client_receive_next_request(client); | ||
431 | |||
432 | return rc; | ||
433 | } | ||
434 | |||
435 | -static void nbd_trip(void *opaque) | ||
436 | +/* Owns a reference to the NBDClient passed as opaque. */ | ||
437 | +static coroutine_fn void nbd_trip(void *opaque) | ||
438 | { | ||
439 | NBDClient *client = opaque; | ||
440 | NBDExport *exp = client->exp; | ||
441 | NBDRequestData *req; | ||
442 | - NBDRequest request; | ||
443 | + NBDRequest request = { 0 }; /* GCC thinks it can be used uninitialized */ | ||
444 | NBDReply reply; | ||
445 | ssize_t ret; | ||
446 | int flags; | ||
447 | |||
448 | TRACE("Reading request."); | ||
449 | if (client->closing) { | ||
450 | + nbd_client_put(client); | ||
451 | return; | ||
452 | } | ||
453 | |||
454 | @@ -XXX,XX +XXX,XX @@ static void nbd_trip(void *opaque) | ||
455 | |||
456 | done: | ||
457 | nbd_request_put(req); | ||
458 | + nbd_client_put(client); | ||
459 | return; | ||
460 | |||
461 | out: | ||
462 | nbd_request_put(req); | ||
463 | client_close(client); | ||
464 | + nbd_client_put(client); | ||
465 | } | ||
466 | |||
467 | -static void nbd_read(void *opaque) | ||
468 | +static void nbd_client_receive_next_request(NBDClient *client) | ||
469 | { | ||
470 | - NBDClient *client = opaque; | ||
471 | - | ||
472 | - if (client->recv_coroutine) { | ||
473 | - qemu_coroutine_enter(client->recv_coroutine); | ||
474 | - } else { | ||
475 | - qemu_coroutine_enter(qemu_coroutine_create(nbd_trip, client)); | ||
476 | - } | ||
477 | -} | ||
478 | - | ||
479 | -static void nbd_restart_write(void *opaque) | ||
480 | -{ | ||
481 | - NBDClient *client = opaque; | ||
482 | - | ||
483 | - qemu_coroutine_enter(client->send_coroutine); | ||
484 | -} | ||
485 | - | ||
486 | -static void nbd_set_handlers(NBDClient *client) | ||
487 | -{ | ||
488 | - if (client->exp && client->exp->ctx) { | ||
489 | - aio_set_fd_handler(client->exp->ctx, client->sioc->fd, true, | ||
490 | - client->can_read ? nbd_read : NULL, | ||
491 | - client->send_coroutine ? nbd_restart_write : NULL, | ||
492 | - NULL, client); | ||
493 | - } | ||
494 | -} | ||
495 | - | ||
496 | -static void nbd_unset_handlers(NBDClient *client) | ||
497 | -{ | ||
498 | - if (client->exp && client->exp->ctx) { | ||
499 | - aio_set_fd_handler(client->exp->ctx, client->sioc->fd, true, NULL, | ||
500 | - NULL, NULL, NULL); | ||
501 | - } | ||
502 | -} | ||
503 | - | ||
504 | -static void nbd_update_can_read(NBDClient *client) | ||
505 | -{ | ||
506 | - bool can_read = client->recv_coroutine || | ||
507 | - client->nb_requests < MAX_NBD_REQUESTS; | ||
508 | - | ||
509 | - if (can_read != client->can_read) { | ||
510 | - client->can_read = can_read; | ||
511 | - nbd_set_handlers(client); | ||
512 | - | ||
513 | - /* There is no need to invoke aio_notify(), since aio_set_fd_handler() | ||
514 | - * in nbd_set_handlers() will have taken care of that */ | ||
515 | + if (!client->recv_coroutine && client->nb_requests < MAX_NBD_REQUESTS) { | ||
516 | + nbd_client_get(client); | ||
517 | + client->recv_coroutine = qemu_coroutine_create(nbd_trip, client); | ||
518 | + aio_co_schedule(client->exp->ctx, client->recv_coroutine); | ||
519 | } | ||
520 | } | ||
521 | |||
522 | @@ -XXX,XX +XXX,XX @@ static coroutine_fn void nbd_co_client_start(void *opaque) | ||
523 | goto out; | ||
524 | } | ||
525 | qemu_co_mutex_init(&client->send_lock); | ||
526 | - nbd_set_handlers(client); | ||
527 | |||
528 | if (exp) { | ||
529 | QTAILQ_INSERT_TAIL(&exp->clients, client, next); | ||
530 | } | ||
531 | + | ||
532 | + nbd_client_receive_next_request(client); | ||
533 | + | ||
534 | out: | ||
535 | g_free(data); | ||
536 | } | ||
537 | @@ -XXX,XX +XXX,XX @@ void nbd_client_new(NBDExport *exp, | ||
538 | object_ref(OBJECT(client->sioc)); | ||
539 | client->ioc = QIO_CHANNEL(sioc); | ||
540 | object_ref(OBJECT(client->ioc)); | ||
541 | - client->can_read = true; | ||
542 | client->close = close_fn; | ||
543 | |||
544 | data->client = client; | ||
545 | -- | ||
546 | 2.9.3 | ||
547 | |||
548 | diff view generated by jsdifflib |
1 | Add the scripts/ directory to sys.path so Python 2.6 will be able to | 1 | From: Paolo Bonzini <pbonzini@redhat.com> |
---|---|---|---|
2 | import argparse. | ||
3 | 2 | ||
4 | Cc: Daniel P. Berrange <berrange@redhat.com> | 3 | As a small step towards the introduction of multiqueue, we want |
5 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | 4 | coroutines to remain on the same AioContext that started them, |
6 | Acked-by: John Snow <jsnow@redhat.com> | 5 | unless they are moved explicitly with e.g. aio_co_schedule. This patch |
7 | Acked-by: Fam Zheng <famz@redhat.com> | 6 | avoids that coroutines switch AioContext when they use a CoMutex. |
8 | Message-id: 20170825155732.15665-4-stefanha@redhat.com | 7 | For now it does not make much of a difference, because the CoMutex |
8 | is not thread-safe and the AioContext itself is used to protect the | ||
9 | CoMutex from concurrent access. However, this is going to change. | ||
10 | |||
11 | Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
12 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> | ||
13 | Reviewed-by: Fam Zheng <famz@redhat.com> | ||
14 | Reviewed-by: Daniel P. Berrange <berrange@redhat.com> | ||
15 | Message-id: 20170213135235.12274-9-pbonzini@redhat.com | ||
9 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | 16 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
10 | --- | 17 | --- |
11 | tests/migration/guestperf/shell.py | 8 +++++--- | 18 | util/qemu-coroutine-lock.c | 5 ++--- |
12 | 1 file changed, 5 insertions(+), 3 deletions(-) | 19 | util/trace-events | 1 - |
20 | 2 files changed, 2 insertions(+), 4 deletions(-) | ||
13 | 21 | ||
14 | diff --git a/tests/migration/guestperf/shell.py b/tests/migration/guestperf/shell.py | 22 | diff --git a/util/qemu-coroutine-lock.c b/util/qemu-coroutine-lock.c |
15 | index XXXXXXX..XXXXXXX 100644 | 23 | index XXXXXXX..XXXXXXX 100644 |
16 | --- a/tests/migration/guestperf/shell.py | 24 | --- a/util/qemu-coroutine-lock.c |
17 | +++ b/tests/migration/guestperf/shell.py | 25 | +++ b/util/qemu-coroutine-lock.c |
18 | @@ -XXX,XX +XXX,XX @@ | 26 | @@ -XXX,XX +XXX,XX @@ |
19 | # | 27 | #include "qemu/coroutine.h" |
20 | 28 | #include "qemu/coroutine_int.h" | |
21 | 29 | #include "qemu/queue.h" | |
22 | -import argparse | 30 | +#include "block/aio.h" |
23 | -import fnmatch | 31 | #include "trace.h" |
24 | import os | 32 | |
25 | import os.path | 33 | void qemu_co_queue_init(CoQueue *queue) |
26 | -import platform | 34 | @@ -XXX,XX +XXX,XX @@ void qemu_co_queue_run_restart(Coroutine *co) |
27 | import sys | 35 | |
28 | +sys.path.append(os.path.join(os.path.dirname(__file__), | 36 | static bool qemu_co_queue_do_restart(CoQueue *queue, bool single) |
29 | + '..', '..', '..', 'scripts')) | 37 | { |
30 | +import argparse | 38 | - Coroutine *self = qemu_coroutine_self(); |
31 | +import fnmatch | 39 | Coroutine *next; |
32 | +import platform | 40 | |
33 | 41 | if (QSIMPLEQ_EMPTY(&queue->entries)) { | |
34 | from guestperf.hardware import Hardware | 42 | @@ -XXX,XX +XXX,XX @@ static bool qemu_co_queue_do_restart(CoQueue *queue, bool single) |
35 | from guestperf.engine import Engine | 43 | |
44 | while ((next = QSIMPLEQ_FIRST(&queue->entries)) != NULL) { | ||
45 | QSIMPLEQ_REMOVE_HEAD(&queue->entries, co_queue_next); | ||
46 | - QSIMPLEQ_INSERT_TAIL(&self->co_queue_wakeup, next, co_queue_next); | ||
47 | - trace_qemu_co_queue_next(next); | ||
48 | + aio_co_wake(next); | ||
49 | if (single) { | ||
50 | break; | ||
51 | } | ||
52 | diff --git a/util/trace-events b/util/trace-events | ||
53 | index XXXXXXX..XXXXXXX 100644 | ||
54 | --- a/util/trace-events | ||
55 | +++ b/util/trace-events | ||
56 | @@ -XXX,XX +XXX,XX @@ qemu_coroutine_terminate(void *co) "self %p" | ||
57 | |||
58 | # util/qemu-coroutine-lock.c | ||
59 | qemu_co_queue_run_restart(void *co) "co %p" | ||
60 | -qemu_co_queue_next(void *nxt) "next %p" | ||
61 | qemu_co_mutex_lock_entry(void *mutex, void *self) "mutex %p self %p" | ||
62 | qemu_co_mutex_lock_return(void *mutex, void *self) "mutex %p self %p" | ||
63 | qemu_co_mutex_unlock_entry(void *mutex, void *self) "mutex %p self %p" | ||
36 | -- | 64 | -- |
37 | 2.13.5 | 65 | 2.9.3 |
38 | 66 | ||
39 | 67 | diff view generated by jsdifflib |
1 | From: Eduardo Habkost <ehabkost@redhat.com> | 1 | From: Paolo Bonzini <pbonzini@redhat.com> |
---|---|---|---|
2 | 2 | ||
3 | If QEMU is running on a system that's out of memory and mmap() | 3 | Keep the coroutine on the same AioContext. Without this change, |
4 | fails, QEMU aborts with no error message at all, making it hard | 4 | there would be a race between yielding the coroutine and reentering it. |
5 | to debug the reason for the failure. | 5 | While the race cannot happen now, because the code only runs from a single |
6 | AioContext, this will change with multiqueue support in the block layer. | ||
6 | 7 | ||
7 | Add perror() calls that will print error information before | 8 | While doing the change, replace custom bottom half with aio_co_schedule. |
8 | aborting. | ||
9 | 9 | ||
10 | Signed-off-by: Eduardo Habkost <ehabkost@redhat.com> | 10 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> |
11 | Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | 11 | Reviewed-by: Fam Zheng <famz@redhat.com> |
12 | Tested-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | 12 | Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> |
13 | Message-id: 20170829212053.6003-1-ehabkost@redhat.com | 13 | Reviewed-by: Daniel P. Berrange <berrange@redhat.com> |
14 | Message-id: 20170213135235.12274-10-pbonzini@redhat.com | ||
14 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | 15 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
15 | --- | 16 | --- |
16 | util/oslib-posix.c | 2 ++ | 17 | block/blkdebug.c | 9 +-------- |
17 | 1 file changed, 2 insertions(+) | 18 | 1 file changed, 1 insertion(+), 8 deletions(-) |
18 | 19 | ||
19 | diff --git a/util/oslib-posix.c b/util/oslib-posix.c | 20 | diff --git a/block/blkdebug.c b/block/blkdebug.c |
20 | index XXXXXXX..XXXXXXX 100644 | 21 | index XXXXXXX..XXXXXXX 100644 |
21 | --- a/util/oslib-posix.c | 22 | --- a/block/blkdebug.c |
22 | +++ b/util/oslib-posix.c | 23 | +++ b/block/blkdebug.c |
23 | @@ -XXX,XX +XXX,XX @@ void *qemu_alloc_stack(size_t *sz) | 24 | @@ -XXX,XX +XXX,XX @@ out: |
24 | ptr = mmap(NULL, *sz, PROT_READ | PROT_WRITE, | 25 | return ret; |
25 | MAP_PRIVATE | MAP_ANONYMOUS, -1, 0); | 26 | } |
26 | if (ptr == MAP_FAILED) { | 27 | |
27 | + perror("failed to allocate memory for stack"); | 28 | -static void error_callback_bh(void *opaque) |
28 | abort(); | 29 | -{ |
30 | - Coroutine *co = opaque; | ||
31 | - qemu_coroutine_enter(co); | ||
32 | -} | ||
33 | - | ||
34 | static int inject_error(BlockDriverState *bs, BlkdebugRule *rule) | ||
35 | { | ||
36 | BDRVBlkdebugState *s = bs->opaque; | ||
37 | @@ -XXX,XX +XXX,XX @@ static int inject_error(BlockDriverState *bs, BlkdebugRule *rule) | ||
29 | } | 38 | } |
30 | 39 | ||
31 | @@ -XXX,XX +XXX,XX @@ void *qemu_alloc_stack(size_t *sz) | 40 | if (!immediately) { |
32 | guardpage = ptr; | 41 | - aio_bh_schedule_oneshot(bdrv_get_aio_context(bs), error_callback_bh, |
33 | #endif | 42 | - qemu_coroutine_self()); |
34 | if (mprotect(guardpage, pagesz, PROT_NONE) != 0) { | 43 | + aio_co_schedule(qemu_get_current_aio_context(), qemu_coroutine_self()); |
35 | + perror("failed to set up stack guard page"); | 44 | qemu_coroutine_yield(); |
36 | abort(); | ||
37 | } | 45 | } |
38 | 46 | ||
39 | -- | 47 | -- |
40 | 2.13.5 | 48 | 2.9.3 |
41 | 49 | ||
42 | 50 | diff view generated by jsdifflib |
1 | From: Alberto Garcia <berto@igalia.com> | 1 | From: Paolo Bonzini <pbonzini@redhat.com> |
---|---|---|---|
2 | 2 | ||
3 | Signed-off-by: Alberto Garcia <berto@igalia.com> | 3 | qed_aio_start_io and qed_aio_next_io will not have to acquire/release |
4 | Message-id: a57dd6274e1b6dc9c28769fec4c7ea543be5c5e3.1503580370.git.berto@igalia.com | 4 | the AioContext, while qed_aio_next_io_cb will. Split the functionality |
5 | and gain a little type-safety in the process. | ||
6 | |||
7 | Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
8 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> | ||
9 | Reviewed-by: Fam Zheng <famz@redhat.com> | ||
10 | Reviewed-by: Daniel P. Berrange <berrange@redhat.com> | ||
11 | Message-id: 20170213135235.12274-11-pbonzini@redhat.com | ||
5 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | 12 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
6 | --- | 13 | --- |
7 | tests/test-throttle.c | 77 +++++++++++++++++++++++++++++++++++++++++++++++++++ | 14 | block/qed.c | 39 +++++++++++++++++++++++++-------------- |
8 | 1 file changed, 77 insertions(+) | 15 | 1 file changed, 25 insertions(+), 14 deletions(-) |
9 | 16 | ||
10 | diff --git a/tests/test-throttle.c b/tests/test-throttle.c | 17 | diff --git a/block/qed.c b/block/qed.c |
11 | index XXXXXXX..XXXXXXX 100644 | 18 | index XXXXXXX..XXXXXXX 100644 |
12 | --- a/tests/test-throttle.c | 19 | --- a/block/qed.c |
13 | +++ b/tests/test-throttle.c | 20 | +++ b/block/qed.c |
14 | @@ -XXX,XX +XXX,XX @@ static void test_is_valid(void) | 21 | @@ -XXX,XX +XXX,XX @@ static CachedL2Table *qed_new_l2_table(BDRVQEDState *s) |
15 | test_is_valid_for_value(1, true); | 22 | return l2_table; |
16 | } | 23 | } |
17 | 24 | ||
18 | +static void test_ranges(void) | 25 | -static void qed_aio_next_io(void *opaque, int ret); |
26 | +static void qed_aio_next_io(QEDAIOCB *acb, int ret); | ||
27 | + | ||
28 | +static void qed_aio_start_io(QEDAIOCB *acb) | ||
19 | +{ | 29 | +{ |
20 | + int i; | 30 | + qed_aio_next_io(acb, 0); |
21 | + | ||
22 | + for (i = 0; i < BUCKETS_COUNT; i++) { | ||
23 | + LeakyBucket *b = &cfg.buckets[i]; | ||
24 | + throttle_config_init(&cfg); | ||
25 | + | ||
26 | + /* avg = 0 means throttling is disabled, but the config is valid */ | ||
27 | + b->avg = 0; | ||
28 | + g_assert(throttle_is_valid(&cfg, NULL)); | ||
29 | + g_assert(!throttle_enabled(&cfg)); | ||
30 | + | ||
31 | + /* These are valid configurations (values <= THROTTLE_VALUE_MAX) */ | ||
32 | + b->avg = 1; | ||
33 | + g_assert(throttle_is_valid(&cfg, NULL)); | ||
34 | + | ||
35 | + b->avg = THROTTLE_VALUE_MAX; | ||
36 | + g_assert(throttle_is_valid(&cfg, NULL)); | ||
37 | + | ||
38 | + b->avg = THROTTLE_VALUE_MAX; | ||
39 | + b->max = THROTTLE_VALUE_MAX; | ||
40 | + g_assert(throttle_is_valid(&cfg, NULL)); | ||
41 | + | ||
42 | + /* Values over THROTTLE_VALUE_MAX are not allowed */ | ||
43 | + b->avg = THROTTLE_VALUE_MAX + 1; | ||
44 | + g_assert(!throttle_is_valid(&cfg, NULL)); | ||
45 | + | ||
46 | + b->avg = THROTTLE_VALUE_MAX; | ||
47 | + b->max = THROTTLE_VALUE_MAX + 1; | ||
48 | + g_assert(!throttle_is_valid(&cfg, NULL)); | ||
49 | + | ||
50 | + /* burst_length must be between 1 and THROTTLE_VALUE_MAX */ | ||
51 | + b->avg = 1; | ||
52 | + b->max = 1; | ||
53 | + b->burst_length = 0; | ||
54 | + g_assert(!throttle_is_valid(&cfg, NULL)); | ||
55 | + | ||
56 | + b->avg = 1; | ||
57 | + b->max = 1; | ||
58 | + b->burst_length = 1; | ||
59 | + g_assert(throttle_is_valid(&cfg, NULL)); | ||
60 | + | ||
61 | + b->avg = 1; | ||
62 | + b->max = 1; | ||
63 | + b->burst_length = THROTTLE_VALUE_MAX; | ||
64 | + g_assert(throttle_is_valid(&cfg, NULL)); | ||
65 | + | ||
66 | + b->avg = 1; | ||
67 | + b->max = 1; | ||
68 | + b->burst_length = THROTTLE_VALUE_MAX + 1; | ||
69 | + g_assert(!throttle_is_valid(&cfg, NULL)); | ||
70 | + | ||
71 | + /* burst_length * max cannot exceed THROTTLE_VALUE_MAX */ | ||
72 | + b->avg = 1; | ||
73 | + b->max = 2; | ||
74 | + b->burst_length = THROTTLE_VALUE_MAX / 2; | ||
75 | + g_assert(throttle_is_valid(&cfg, NULL)); | ||
76 | + | ||
77 | + b->avg = 1; | ||
78 | + b->max = 3; | ||
79 | + b->burst_length = THROTTLE_VALUE_MAX / 2; | ||
80 | + g_assert(!throttle_is_valid(&cfg, NULL)); | ||
81 | + | ||
82 | + b->avg = 1; | ||
83 | + b->max = THROTTLE_VALUE_MAX; | ||
84 | + b->burst_length = 1; | ||
85 | + g_assert(throttle_is_valid(&cfg, NULL)); | ||
86 | + | ||
87 | + b->avg = 1; | ||
88 | + b->max = THROTTLE_VALUE_MAX; | ||
89 | + b->burst_length = 2; | ||
90 | + g_assert(!throttle_is_valid(&cfg, NULL)); | ||
91 | + } | ||
92 | +} | 31 | +} |
93 | + | 32 | + |
94 | static void test_max_is_missing_limit(void) | 33 | +static void qed_aio_next_io_cb(void *opaque, int ret) |
34 | +{ | ||
35 | + QEDAIOCB *acb = opaque; | ||
36 | + | ||
37 | + qed_aio_next_io(acb, ret); | ||
38 | +} | ||
39 | |||
40 | static void qed_plug_allocating_write_reqs(BDRVQEDState *s) | ||
95 | { | 41 | { |
96 | int i; | 42 | @@ -XXX,XX +XXX,XX @@ static void qed_unplug_allocating_write_reqs(BDRVQEDState *s) |
97 | @@ -XXX,XX +XXX,XX @@ int main(int argc, char **argv) | 43 | |
98 | g_test_add_func("/throttle/config/enabled", test_enabled); | 44 | acb = QSIMPLEQ_FIRST(&s->allocating_write_reqs); |
99 | g_test_add_func("/throttle/config/conflicting", test_conflicting_config); | 45 | if (acb) { |
100 | g_test_add_func("/throttle/config/is_valid", test_is_valid); | 46 | - qed_aio_next_io(acb, 0); |
101 | + g_test_add_func("/throttle/config/ranges", test_ranges); | 47 | + qed_aio_start_io(acb); |
102 | g_test_add_func("/throttle/config/max", test_max_is_missing_limit); | 48 | } |
103 | g_test_add_func("/throttle/config/iops_size", | 49 | } |
104 | test_iops_size_is_missing_limit); | 50 | |
51 | @@ -XXX,XX +XXX,XX @@ static void qed_aio_complete(QEDAIOCB *acb, int ret) | ||
52 | QSIMPLEQ_REMOVE_HEAD(&s->allocating_write_reqs, next); | ||
53 | acb = QSIMPLEQ_FIRST(&s->allocating_write_reqs); | ||
54 | if (acb) { | ||
55 | - qed_aio_next_io(acb, 0); | ||
56 | + qed_aio_start_io(acb); | ||
57 | } else if (s->header.features & QED_F_NEED_CHECK) { | ||
58 | qed_start_need_check_timer(s); | ||
59 | } | ||
60 | @@ -XXX,XX +XXX,XX @@ static void qed_commit_l2_update(void *opaque, int ret) | ||
61 | acb->request.l2_table = qed_find_l2_cache_entry(&s->l2_cache, l2_offset); | ||
62 | assert(acb->request.l2_table != NULL); | ||
63 | |||
64 | - qed_aio_next_io(opaque, ret); | ||
65 | + qed_aio_next_io(acb, ret); | ||
66 | } | ||
67 | |||
68 | /** | ||
69 | @@ -XXX,XX +XXX,XX @@ static void qed_aio_write_l2_update(QEDAIOCB *acb, int ret, uint64_t offset) | ||
70 | if (need_alloc) { | ||
71 | /* Write out the whole new L2 table */ | ||
72 | qed_write_l2_table(s, &acb->request, 0, s->table_nelems, true, | ||
73 | - qed_aio_write_l1_update, acb); | ||
74 | + qed_aio_write_l1_update, acb); | ||
75 | } else { | ||
76 | /* Write out only the updated part of the L2 table */ | ||
77 | qed_write_l2_table(s, &acb->request, index, acb->cur_nclusters, false, | ||
78 | - qed_aio_next_io, acb); | ||
79 | + qed_aio_next_io_cb, acb); | ||
80 | } | ||
81 | return; | ||
82 | |||
83 | @@ -XXX,XX +XXX,XX @@ static void qed_aio_write_main(void *opaque, int ret) | ||
84 | } | ||
85 | |||
86 | if (acb->find_cluster_ret == QED_CLUSTER_FOUND) { | ||
87 | - next_fn = qed_aio_next_io; | ||
88 | + next_fn = qed_aio_next_io_cb; | ||
89 | } else { | ||
90 | if (s->bs->backing) { | ||
91 | next_fn = qed_aio_write_flush_before_l2_update; | ||
92 | @@ -XXX,XX +XXX,XX @@ static void qed_aio_write_alloc(QEDAIOCB *acb, size_t len) | ||
93 | if (acb->flags & QED_AIOCB_ZERO) { | ||
94 | /* Skip ahead if the clusters are already zero */ | ||
95 | if (acb->find_cluster_ret == QED_CLUSTER_ZERO) { | ||
96 | - qed_aio_next_io(acb, 0); | ||
97 | + qed_aio_start_io(acb); | ||
98 | return; | ||
99 | } | ||
100 | |||
101 | @@ -XXX,XX +XXX,XX @@ static void qed_aio_read_data(void *opaque, int ret, | ||
102 | /* Handle zero cluster and backing file reads */ | ||
103 | if (ret == QED_CLUSTER_ZERO) { | ||
104 | qemu_iovec_memset(&acb->cur_qiov, 0, 0, acb->cur_qiov.size); | ||
105 | - qed_aio_next_io(acb, 0); | ||
106 | + qed_aio_start_io(acb); | ||
107 | return; | ||
108 | } else if (ret != QED_CLUSTER_FOUND) { | ||
109 | qed_read_backing_file(s, acb->cur_pos, &acb->cur_qiov, | ||
110 | - &acb->backing_qiov, qed_aio_next_io, acb); | ||
111 | + &acb->backing_qiov, qed_aio_next_io_cb, acb); | ||
112 | return; | ||
113 | } | ||
114 | |||
115 | BLKDBG_EVENT(bs->file, BLKDBG_READ_AIO); | ||
116 | bdrv_aio_readv(bs->file, offset / BDRV_SECTOR_SIZE, | ||
117 | &acb->cur_qiov, acb->cur_qiov.size / BDRV_SECTOR_SIZE, | ||
118 | - qed_aio_next_io, acb); | ||
119 | + qed_aio_next_io_cb, acb); | ||
120 | return; | ||
121 | |||
122 | err: | ||
123 | @@ -XXX,XX +XXX,XX @@ err: | ||
124 | /** | ||
125 | * Begin next I/O or complete the request | ||
126 | */ | ||
127 | -static void qed_aio_next_io(void *opaque, int ret) | ||
128 | +static void qed_aio_next_io(QEDAIOCB *acb, int ret) | ||
129 | { | ||
130 | - QEDAIOCB *acb = opaque; | ||
131 | BDRVQEDState *s = acb_to_s(acb); | ||
132 | QEDFindClusterFunc *io_fn = (acb->flags & QED_AIOCB_WRITE) ? | ||
133 | qed_aio_write_data : qed_aio_read_data; | ||
134 | @@ -XXX,XX +XXX,XX @@ static BlockAIOCB *qed_aio_setup(BlockDriverState *bs, | ||
135 | qemu_iovec_init(&acb->cur_qiov, qiov->niov); | ||
136 | |||
137 | /* Start request */ | ||
138 | - qed_aio_next_io(acb, 0); | ||
139 | + qed_aio_start_io(acb); | ||
140 | return &acb->common; | ||
141 | } | ||
142 | |||
105 | -- | 143 | -- |
106 | 2.13.5 | 144 | 2.9.3 |
107 | 145 | ||
108 | 146 | diff view generated by jsdifflib |
1 | Add the scripts/ directory to sys.path so Python 2.6 will be able to | 1 | From: Paolo Bonzini <pbonzini@redhat.com> |
---|---|---|---|
2 | import argparse. | ||
3 | 2 | ||
4 | Cc: Fam Zheng <famz@redhat.com> | 3 | The AioContext data structures are now protected by list_lock and/or |
5 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | 4 | they are walked with FOREACH_RCU primitives. There is no need anymore |
6 | Acked-by: John Snow <jsnow@redhat.com> | 5 | to acquire the AioContext for the entire duration of aio_dispatch. |
7 | Acked-by: Fam Zheng <famz@redhat.com> | 6 | Instead, just acquire it before and after invoking the callbacks. |
8 | Message-id: 20170825155732.15665-3-stefanha@redhat.com | 7 | The next step is then to push it further down. |
8 | |||
9 | Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
10 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> | ||
11 | Reviewed-by: Fam Zheng <famz@redhat.com> | ||
12 | Reviewed-by: Daniel P. Berrange <berrange@redhat.com> | ||
13 | Message-id: 20170213135235.12274-12-pbonzini@redhat.com | ||
9 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | 14 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
10 | --- | 15 | --- |
11 | tests/docker/docker.py | 4 +++- | 16 | util/aio-posix.c | 25 +++++++++++-------------- |
12 | 1 file changed, 3 insertions(+), 1 deletion(-) | 17 | util/aio-win32.c | 15 +++++++-------- |
18 | util/async.c | 2 ++ | ||
19 | 3 files changed, 20 insertions(+), 22 deletions(-) | ||
13 | 20 | ||
14 | diff --git a/tests/docker/docker.py b/tests/docker/docker.py | 21 | diff --git a/util/aio-posix.c b/util/aio-posix.c |
15 | index XXXXXXX..XXXXXXX 100755 | 22 | index XXXXXXX..XXXXXXX 100644 |
16 | --- a/tests/docker/docker.py | 23 | --- a/util/aio-posix.c |
17 | +++ b/tests/docker/docker.py | 24 | +++ b/util/aio-posix.c |
18 | @@ -XXX,XX +XXX,XX @@ | 25 | @@ -XXX,XX +XXX,XX @@ static bool aio_dispatch_handlers(AioContext *ctx) |
19 | 26 | (revents & (G_IO_IN | G_IO_HUP | G_IO_ERR)) && | |
20 | import os | 27 | aio_node_check(ctx, node->is_external) && |
21 | import sys | 28 | node->io_read) { |
22 | +sys.path.append(os.path.join(os.path.dirname(__file__), | 29 | + aio_context_acquire(ctx); |
23 | + '..', '..', 'scripts')) | 30 | node->io_read(node->opaque); |
24 | +import argparse | 31 | + aio_context_release(ctx); |
25 | import subprocess | 32 | |
26 | import json | 33 | /* aio_notify() does not count as progress */ |
27 | import hashlib | 34 | if (node->opaque != &ctx->notifier) { |
28 | import atexit | 35 | @@ -XXX,XX +XXX,XX @@ static bool aio_dispatch_handlers(AioContext *ctx) |
29 | import uuid | 36 | (revents & (G_IO_OUT | G_IO_ERR)) && |
30 | -import argparse | 37 | aio_node_check(ctx, node->is_external) && |
31 | import tempfile | 38 | node->io_write) { |
32 | import re | 39 | + aio_context_acquire(ctx); |
33 | import signal | 40 | node->io_write(node->opaque); |
41 | + aio_context_release(ctx); | ||
42 | progress = true; | ||
43 | } | ||
44 | |||
45 | @@ -XXX,XX +XXX,XX @@ bool aio_dispatch(AioContext *ctx, bool dispatch_fds) | ||
46 | } | ||
47 | |||
48 | /* Run our timers */ | ||
49 | + aio_context_acquire(ctx); | ||
50 | progress |= timerlistgroup_run_timers(&ctx->tlg); | ||
51 | + aio_context_release(ctx); | ||
52 | |||
53 | return progress; | ||
54 | } | ||
55 | @@ -XXX,XX +XXX,XX @@ bool aio_poll(AioContext *ctx, bool blocking) | ||
56 | int64_t timeout; | ||
57 | int64_t start = 0; | ||
58 | |||
59 | - aio_context_acquire(ctx); | ||
60 | - progress = false; | ||
61 | - | ||
62 | /* aio_notify can avoid the expensive event_notifier_set if | ||
63 | * everything (file descriptors, bottom halves, timers) will | ||
64 | * be re-evaluated before the next blocking poll(). This is | ||
65 | @@ -XXX,XX +XXX,XX @@ bool aio_poll(AioContext *ctx, bool blocking) | ||
66 | start = qemu_clock_get_ns(QEMU_CLOCK_REALTIME); | ||
67 | } | ||
68 | |||
69 | - if (try_poll_mode(ctx, blocking)) { | ||
70 | - progress = true; | ||
71 | - } else { | ||
72 | + aio_context_acquire(ctx); | ||
73 | + progress = try_poll_mode(ctx, blocking); | ||
74 | + aio_context_release(ctx); | ||
75 | + | ||
76 | + if (!progress) { | ||
77 | assert(npfd == 0); | ||
78 | |||
79 | /* fill pollfds */ | ||
80 | @@ -XXX,XX +XXX,XX @@ bool aio_poll(AioContext *ctx, bool blocking) | ||
81 | timeout = blocking ? aio_compute_timeout(ctx) : 0; | ||
82 | |||
83 | /* wait until next event */ | ||
84 | - if (timeout) { | ||
85 | - aio_context_release(ctx); | ||
86 | - } | ||
87 | if (aio_epoll_check_poll(ctx, pollfds, npfd, timeout)) { | ||
88 | AioHandler epoll_handler; | ||
89 | |||
90 | @@ -XXX,XX +XXX,XX @@ bool aio_poll(AioContext *ctx, bool blocking) | ||
91 | } else { | ||
92 | ret = qemu_poll_ns(pollfds, npfd, timeout); | ||
93 | } | ||
94 | - if (timeout) { | ||
95 | - aio_context_acquire(ctx); | ||
96 | - } | ||
97 | } | ||
98 | |||
99 | if (blocking) { | ||
100 | @@ -XXX,XX +XXX,XX @@ bool aio_poll(AioContext *ctx, bool blocking) | ||
101 | progress = true; | ||
102 | } | ||
103 | |||
104 | - aio_context_release(ctx); | ||
105 | - | ||
106 | return progress; | ||
107 | } | ||
108 | |||
109 | diff --git a/util/aio-win32.c b/util/aio-win32.c | ||
110 | index XXXXXXX..XXXXXXX 100644 | ||
111 | --- a/util/aio-win32.c | ||
112 | +++ b/util/aio-win32.c | ||
113 | @@ -XXX,XX +XXX,XX @@ static bool aio_dispatch_handlers(AioContext *ctx, HANDLE event) | ||
114 | (revents || event_notifier_get_handle(node->e) == event) && | ||
115 | node->io_notify) { | ||
116 | node->pfd.revents = 0; | ||
117 | + aio_context_acquire(ctx); | ||
118 | node->io_notify(node->e); | ||
119 | + aio_context_release(ctx); | ||
120 | |||
121 | /* aio_notify() does not count as progress */ | ||
122 | if (node->e != &ctx->notifier) { | ||
123 | @@ -XXX,XX +XXX,XX @@ static bool aio_dispatch_handlers(AioContext *ctx, HANDLE event) | ||
124 | (node->io_read || node->io_write)) { | ||
125 | node->pfd.revents = 0; | ||
126 | if ((revents & G_IO_IN) && node->io_read) { | ||
127 | + aio_context_acquire(ctx); | ||
128 | node->io_read(node->opaque); | ||
129 | + aio_context_release(ctx); | ||
130 | progress = true; | ||
131 | } | ||
132 | if ((revents & G_IO_OUT) && node->io_write) { | ||
133 | + aio_context_acquire(ctx); | ||
134 | node->io_write(node->opaque); | ||
135 | + aio_context_release(ctx); | ||
136 | progress = true; | ||
137 | } | ||
138 | |||
139 | @@ -XXX,XX +XXX,XX @@ bool aio_poll(AioContext *ctx, bool blocking) | ||
140 | int count; | ||
141 | int timeout; | ||
142 | |||
143 | - aio_context_acquire(ctx); | ||
144 | progress = false; | ||
145 | |||
146 | /* aio_notify can avoid the expensive event_notifier_set if | ||
147 | @@ -XXX,XX +XXX,XX @@ bool aio_poll(AioContext *ctx, bool blocking) | ||
148 | |||
149 | timeout = blocking && !have_select_revents | ||
150 | ? qemu_timeout_ns_to_ms(aio_compute_timeout(ctx)) : 0; | ||
151 | - if (timeout) { | ||
152 | - aio_context_release(ctx); | ||
153 | - } | ||
154 | ret = WaitForMultipleObjects(count, events, FALSE, timeout); | ||
155 | if (blocking) { | ||
156 | assert(first); | ||
157 | atomic_sub(&ctx->notify_me, 2); | ||
158 | } | ||
159 | - if (timeout) { | ||
160 | - aio_context_acquire(ctx); | ||
161 | - } | ||
162 | |||
163 | if (first) { | ||
164 | aio_notify_accept(ctx); | ||
165 | @@ -XXX,XX +XXX,XX @@ bool aio_poll(AioContext *ctx, bool blocking) | ||
166 | progress |= aio_dispatch_handlers(ctx, event); | ||
167 | } while (count > 0); | ||
168 | |||
169 | + aio_context_acquire(ctx); | ||
170 | progress |= timerlistgroup_run_timers(&ctx->tlg); | ||
171 | - | ||
172 | aio_context_release(ctx); | ||
173 | return progress; | ||
174 | } | ||
175 | diff --git a/util/async.c b/util/async.c | ||
176 | index XXXXXXX..XXXXXXX 100644 | ||
177 | --- a/util/async.c | ||
178 | +++ b/util/async.c | ||
179 | @@ -XXX,XX +XXX,XX @@ int aio_bh_poll(AioContext *ctx) | ||
180 | ret = 1; | ||
181 | } | ||
182 | bh->idle = 0; | ||
183 | + aio_context_acquire(ctx); | ||
184 | aio_bh_call(bh); | ||
185 | + aio_context_release(ctx); | ||
186 | } | ||
187 | if (bh->deleted) { | ||
188 | deleted = true; | ||
34 | -- | 189 | -- |
35 | 2.13.5 | 190 | 2.9.3 |
36 | 191 | ||
37 | 192 | diff view generated by jsdifflib |
New patch | |||
---|---|---|---|
1 | 1 | From: Paolo Bonzini <pbonzini@redhat.com> | |
2 | |||
3 | Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
4 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> | ||
5 | Reviewed-by: Fam Zheng <famz@redhat.com> | ||
6 | Reviewed-by: Daniel P. Berrange <berrange@redhat.com> | ||
7 | Message-id: 20170213135235.12274-13-pbonzini@redhat.com | ||
8 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
9 | --- | ||
10 | block/qed.h | 3 +++ | ||
11 | block/curl.c | 2 ++ | ||
12 | block/io.c | 5 +++++ | ||
13 | block/iscsi.c | 8 ++++++-- | ||
14 | block/null.c | 4 ++++ | ||
15 | block/qed.c | 12 ++++++++++++ | ||
16 | block/throttle-groups.c | 2 ++ | ||
17 | util/aio-posix.c | 2 -- | ||
18 | util/aio-win32.c | 2 -- | ||
19 | util/qemu-coroutine-sleep.c | 2 +- | ||
20 | 10 files changed, 35 insertions(+), 7 deletions(-) | ||
21 | |||
22 | diff --git a/block/qed.h b/block/qed.h | ||
23 | index XXXXXXX..XXXXXXX 100644 | ||
24 | --- a/block/qed.h | ||
25 | +++ b/block/qed.h | ||
26 | @@ -XXX,XX +XXX,XX @@ enum { | ||
27 | */ | ||
28 | typedef void QEDFindClusterFunc(void *opaque, int ret, uint64_t offset, size_t len); | ||
29 | |||
30 | +void qed_acquire(BDRVQEDState *s); | ||
31 | +void qed_release(BDRVQEDState *s); | ||
32 | + | ||
33 | /** | ||
34 | * Generic callback for chaining async callbacks | ||
35 | */ | ||
36 | diff --git a/block/curl.c b/block/curl.c | ||
37 | index XXXXXXX..XXXXXXX 100644 | ||
38 | --- a/block/curl.c | ||
39 | +++ b/block/curl.c | ||
40 | @@ -XXX,XX +XXX,XX @@ static void curl_multi_timeout_do(void *arg) | ||
41 | return; | ||
42 | } | ||
43 | |||
44 | + aio_context_acquire(s->aio_context); | ||
45 | curl_multi_socket_action(s->multi, CURL_SOCKET_TIMEOUT, 0, &running); | ||
46 | |||
47 | curl_multi_check_completion(s); | ||
48 | + aio_context_release(s->aio_context); | ||
49 | #else | ||
50 | abort(); | ||
51 | #endif | ||
52 | diff --git a/block/io.c b/block/io.c | ||
53 | index XXXXXXX..XXXXXXX 100644 | ||
54 | --- a/block/io.c | ||
55 | +++ b/block/io.c | ||
56 | @@ -XXX,XX +XXX,XX @@ void bdrv_aio_cancel(BlockAIOCB *acb) | ||
57 | if (acb->aiocb_info->get_aio_context) { | ||
58 | aio_poll(acb->aiocb_info->get_aio_context(acb), true); | ||
59 | } else if (acb->bs) { | ||
60 | + /* qemu_aio_ref and qemu_aio_unref are not thread-safe, so | ||
61 | + * assert that we're not using an I/O thread. Thread-safe | ||
62 | + * code should use bdrv_aio_cancel_async exclusively. | ||
63 | + */ | ||
64 | + assert(bdrv_get_aio_context(acb->bs) == qemu_get_aio_context()); | ||
65 | aio_poll(bdrv_get_aio_context(acb->bs), true); | ||
66 | } else { | ||
67 | abort(); | ||
68 | diff --git a/block/iscsi.c b/block/iscsi.c | ||
69 | index XXXXXXX..XXXXXXX 100644 | ||
70 | --- a/block/iscsi.c | ||
71 | +++ b/block/iscsi.c | ||
72 | @@ -XXX,XX +XXX,XX @@ static void iscsi_retry_timer_expired(void *opaque) | ||
73 | struct IscsiTask *iTask = opaque; | ||
74 | iTask->complete = 1; | ||
75 | if (iTask->co) { | ||
76 | - qemu_coroutine_enter(iTask->co); | ||
77 | + aio_co_wake(iTask->co); | ||
78 | } | ||
79 | } | ||
80 | |||
81 | @@ -XXX,XX +XXX,XX @@ static void iscsi_nop_timed_event(void *opaque) | ||
82 | { | ||
83 | IscsiLun *iscsilun = opaque; | ||
84 | |||
85 | + aio_context_acquire(iscsilun->aio_context); | ||
86 | if (iscsi_get_nops_in_flight(iscsilun->iscsi) >= MAX_NOP_FAILURES) { | ||
87 | error_report("iSCSI: NOP timeout. Reconnecting..."); | ||
88 | iscsilun->request_timed_out = true; | ||
89 | } else if (iscsi_nop_out_async(iscsilun->iscsi, NULL, NULL, 0, NULL) != 0) { | ||
90 | error_report("iSCSI: failed to sent NOP-Out. Disabling NOP messages."); | ||
91 | - return; | ||
92 | + goto out; | ||
93 | } | ||
94 | |||
95 | timer_mod(iscsilun->nop_timer, qemu_clock_get_ms(QEMU_CLOCK_REALTIME) + NOP_INTERVAL); | ||
96 | iscsi_set_events(iscsilun); | ||
97 | + | ||
98 | +out: | ||
99 | + aio_context_release(iscsilun->aio_context); | ||
100 | } | ||
101 | |||
102 | static void iscsi_readcapacity_sync(IscsiLun *iscsilun, Error **errp) | ||
103 | diff --git a/block/null.c b/block/null.c | ||
104 | index XXXXXXX..XXXXXXX 100644 | ||
105 | --- a/block/null.c | ||
106 | +++ b/block/null.c | ||
107 | @@ -XXX,XX +XXX,XX @@ static void null_bh_cb(void *opaque) | ||
108 | static void null_timer_cb(void *opaque) | ||
109 | { | ||
110 | NullAIOCB *acb = opaque; | ||
111 | + AioContext *ctx = bdrv_get_aio_context(acb->common.bs); | ||
112 | + | ||
113 | + aio_context_acquire(ctx); | ||
114 | acb->common.cb(acb->common.opaque, 0); | ||
115 | + aio_context_release(ctx); | ||
116 | timer_deinit(&acb->timer); | ||
117 | qemu_aio_unref(acb); | ||
118 | } | ||
119 | diff --git a/block/qed.c b/block/qed.c | ||
120 | index XXXXXXX..XXXXXXX 100644 | ||
121 | --- a/block/qed.c | ||
122 | +++ b/block/qed.c | ||
123 | @@ -XXX,XX +XXX,XX @@ static void qed_need_check_timer_cb(void *opaque) | ||
124 | |||
125 | trace_qed_need_check_timer_cb(s); | ||
126 | |||
127 | + qed_acquire(s); | ||
128 | qed_plug_allocating_write_reqs(s); | ||
129 | |||
130 | /* Ensure writes are on disk before clearing flag */ | ||
131 | bdrv_aio_flush(s->bs->file->bs, qed_clear_need_check, s); | ||
132 | + qed_release(s); | ||
133 | +} | ||
134 | + | ||
135 | +void qed_acquire(BDRVQEDState *s) | ||
136 | +{ | ||
137 | + aio_context_acquire(bdrv_get_aio_context(s->bs)); | ||
138 | +} | ||
139 | + | ||
140 | +void qed_release(BDRVQEDState *s) | ||
141 | +{ | ||
142 | + aio_context_release(bdrv_get_aio_context(s->bs)); | ||
143 | } | ||
144 | |||
145 | static void qed_start_need_check_timer(BDRVQEDState *s) | ||
146 | diff --git a/block/throttle-groups.c b/block/throttle-groups.c | ||
147 | index XXXXXXX..XXXXXXX 100644 | ||
148 | --- a/block/throttle-groups.c | ||
149 | +++ b/block/throttle-groups.c | ||
150 | @@ -XXX,XX +XXX,XX @@ static void timer_cb(BlockBackend *blk, bool is_write) | ||
151 | qemu_mutex_unlock(&tg->lock); | ||
152 | |||
153 | /* Run the request that was waiting for this timer */ | ||
154 | + aio_context_acquire(blk_get_aio_context(blk)); | ||
155 | empty_queue = !qemu_co_enter_next(&blkp->throttled_reqs[is_write]); | ||
156 | + aio_context_release(blk_get_aio_context(blk)); | ||
157 | |||
158 | /* If the request queue was empty then we have to take care of | ||
159 | * scheduling the next one */ | ||
160 | diff --git a/util/aio-posix.c b/util/aio-posix.c | ||
161 | index XXXXXXX..XXXXXXX 100644 | ||
162 | --- a/util/aio-posix.c | ||
163 | +++ b/util/aio-posix.c | ||
164 | @@ -XXX,XX +XXX,XX @@ bool aio_dispatch(AioContext *ctx, bool dispatch_fds) | ||
165 | } | ||
166 | |||
167 | /* Run our timers */ | ||
168 | - aio_context_acquire(ctx); | ||
169 | progress |= timerlistgroup_run_timers(&ctx->tlg); | ||
170 | - aio_context_release(ctx); | ||
171 | |||
172 | return progress; | ||
173 | } | ||
174 | diff --git a/util/aio-win32.c b/util/aio-win32.c | ||
175 | index XXXXXXX..XXXXXXX 100644 | ||
176 | --- a/util/aio-win32.c | ||
177 | +++ b/util/aio-win32.c | ||
178 | @@ -XXX,XX +XXX,XX @@ bool aio_poll(AioContext *ctx, bool blocking) | ||
179 | progress |= aio_dispatch_handlers(ctx, event); | ||
180 | } while (count > 0); | ||
181 | |||
182 | - aio_context_acquire(ctx); | ||
183 | progress |= timerlistgroup_run_timers(&ctx->tlg); | ||
184 | - aio_context_release(ctx); | ||
185 | return progress; | ||
186 | } | ||
187 | |||
188 | diff --git a/util/qemu-coroutine-sleep.c b/util/qemu-coroutine-sleep.c | ||
189 | index XXXXXXX..XXXXXXX 100644 | ||
190 | --- a/util/qemu-coroutine-sleep.c | ||
191 | +++ b/util/qemu-coroutine-sleep.c | ||
192 | @@ -XXX,XX +XXX,XX @@ static void co_sleep_cb(void *opaque) | ||
193 | { | ||
194 | CoSleepCB *sleep_cb = opaque; | ||
195 | |||
196 | - qemu_coroutine_enter(sleep_cb->co); | ||
197 | + aio_co_wake(sleep_cb->co); | ||
198 | } | ||
199 | |||
200 | void coroutine_fn co_aio_sleep_ns(AioContext *ctx, QEMUClockType type, | ||
201 | -- | ||
202 | 2.9.3 | ||
203 | |||
204 | diff view generated by jsdifflib |
1 | From: Alberto Garcia <berto@igalia.com> | 1 | From: Paolo Bonzini <pbonzini@redhat.com> |
---|---|---|---|
2 | 2 | ||
3 | The throttling code can change internally the value of bkt->max if it | 3 | This covers both file descriptor callbacks and polling callbacks, |
4 | hasn't been set by the user. The problem with this is that if we want | 4 | since they execute related code. |
5 | to retrieve the original value we have to undo this change first. This | ||
6 | is ugly and unnecessary: this patch removes the throttle_fix_bucket() | ||
7 | and throttle_unfix_bucket() functions completely and moves the logic | ||
8 | to throttle_compute_wait(). | ||
9 | 5 | ||
10 | Signed-off-by: Alberto Garcia <berto@igalia.com> | 6 | Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> |
11 | Reviewed-by: Manos Pitsidianakis <el13635@mail.ntua.gr> | 7 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> |
12 | Message-id: 5b0b9e1ac6eb208d709eddc7b09e7669a523bff3.1503580370.git.berto@igalia.com | 8 | Reviewed-by: Fam Zheng <famz@redhat.com> |
9 | Reviewed-by: Daniel P. Berrange <berrange@redhat.com> | ||
10 | Message-id: 20170213135235.12274-14-pbonzini@redhat.com | ||
13 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | 11 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
14 | --- | 12 | --- |
15 | util/throttle.c | 62 +++++++++++++++++++++------------------------------------ | 13 | block/curl.c | 16 +++++++++++++--- |
16 | 1 file changed, 23 insertions(+), 39 deletions(-) | 14 | block/iscsi.c | 4 ++++ |
15 | block/linux-aio.c | 4 ++++ | ||
16 | block/nfs.c | 6 ++++++ | ||
17 | block/sheepdog.c | 29 +++++++++++++++-------------- | ||
18 | block/ssh.c | 29 +++++++++-------------------- | ||
19 | block/win32-aio.c | 10 ++++++---- | ||
20 | hw/block/virtio-blk.c | 5 ++++- | ||
21 | hw/scsi/virtio-scsi.c | 7 +++++++ | ||
22 | util/aio-posix.c | 7 ------- | ||
23 | util/aio-win32.c | 6 ------ | ||
24 | 11 files changed, 68 insertions(+), 55 deletions(-) | ||
17 | 25 | ||
18 | diff --git a/util/throttle.c b/util/throttle.c | 26 | diff --git a/block/curl.c b/block/curl.c |
19 | index XXXXXXX..XXXXXXX 100644 | 27 | index XXXXXXX..XXXXXXX 100644 |
20 | --- a/util/throttle.c | 28 | --- a/block/curl.c |
21 | +++ b/util/throttle.c | 29 | +++ b/block/curl.c |
22 | @@ -XXX,XX +XXX,XX @@ static int64_t throttle_do_compute_wait(double limit, double extra) | 30 | @@ -XXX,XX +XXX,XX @@ static void curl_multi_check_completion(BDRVCURLState *s) |
23 | int64_t throttle_compute_wait(LeakyBucket *bkt) | 31 | } |
24 | { | 32 | } |
25 | double extra; /* the number of extra units blocking the io */ | 33 | |
26 | + double bucket_size; /* I/O before throttling to bkt->avg */ | 34 | -static void curl_multi_do(void *arg) |
27 | + double burst_bucket_size; /* Before throttling to bkt->max */ | 35 | +static void curl_multi_do_locked(CURLState *s) |
28 | 36 | { | |
29 | if (!bkt->avg) { | 37 | - CURLState *s = (CURLState *)arg; |
30 | return 0; | 38 | CURLSocket *socket, *next_socket; |
31 | } | 39 | int running; |
32 | 40 | int r; | |
33 | - /* If the bucket is full then we have to wait */ | 41 | @@ -XXX,XX +XXX,XX @@ static void curl_multi_do(void *arg) |
34 | - extra = bkt->level - bkt->max * bkt->burst_length; | 42 | } |
35 | + if (!bkt->max) { | 43 | } |
36 | + /* If bkt->max is 0 we still want to allow short bursts of I/O | 44 | |
37 | + * from the guest, otherwise every other request will be throttled | 45 | +static void curl_multi_do(void *arg) |
38 | + * and performance will suffer considerably. */ | 46 | +{ |
39 | + bucket_size = bkt->avg / 10; | 47 | + CURLState *s = (CURLState *)arg; |
40 | + burst_bucket_size = 0; | 48 | + |
41 | + } else { | 49 | + aio_context_acquire(s->s->aio_context); |
42 | + /* If we have a burst limit then we have to wait until all I/O | 50 | + curl_multi_do_locked(s); |
43 | + * at burst rate has finished before throttling to bkt->avg */ | 51 | + aio_context_release(s->s->aio_context); |
44 | + bucket_size = bkt->max * bkt->burst_length; | 52 | +} |
45 | + burst_bucket_size = bkt->max / 10; | 53 | + |
46 | + } | 54 | static void curl_multi_read(void *arg) |
47 | + | 55 | { |
48 | + /* If the main bucket is full then we have to wait */ | 56 | CURLState *s = (CURLState *)arg; |
49 | + extra = bkt->level - bucket_size; | 57 | |
50 | if (extra > 0) { | 58 | - curl_multi_do(arg); |
51 | return throttle_do_compute_wait(bkt->avg, extra); | 59 | + aio_context_acquire(s->s->aio_context); |
52 | } | 60 | + curl_multi_do_locked(s); |
53 | 61 | curl_multi_check_completion(s->s); | |
54 | - /* If the bucket is not full yet we have to make sure that we | 62 | + aio_context_release(s->s->aio_context); |
55 | - * fulfill the goal of bkt->max units per second. */ | 63 | } |
56 | + /* If the main bucket is not full yet we still have to check the | 64 | |
57 | + * burst bucket in order to enforce the burst limit */ | 65 | static void curl_multi_timeout_do(void *arg) |
58 | if (bkt->burst_length > 1) { | 66 | diff --git a/block/iscsi.c b/block/iscsi.c |
59 | - /* We use 1/10 of the max value to smooth the throttling. | 67 | index XXXXXXX..XXXXXXX 100644 |
60 | - * See throttle_fix_bucket() for more details. */ | 68 | --- a/block/iscsi.c |
61 | - extra = bkt->burst_level - bkt->max / 10; | 69 | +++ b/block/iscsi.c |
62 | + extra = bkt->burst_level - burst_bucket_size; | 70 | @@ -XXX,XX +XXX,XX @@ iscsi_process_read(void *arg) |
63 | if (extra > 0) { | 71 | IscsiLun *iscsilun = arg; |
64 | return throttle_do_compute_wait(bkt->max, extra); | 72 | struct iscsi_context *iscsi = iscsilun->iscsi; |
65 | } | 73 | |
66 | @@ -XXX,XX +XXX,XX @@ bool throttle_is_valid(ThrottleConfig *cfg, Error **errp) | 74 | + aio_context_acquire(iscsilun->aio_context); |
75 | iscsi_service(iscsi, POLLIN); | ||
76 | iscsi_set_events(iscsilun); | ||
77 | + aio_context_release(iscsilun->aio_context); | ||
78 | } | ||
79 | |||
80 | static void | ||
81 | @@ -XXX,XX +XXX,XX @@ iscsi_process_write(void *arg) | ||
82 | IscsiLun *iscsilun = arg; | ||
83 | struct iscsi_context *iscsi = iscsilun->iscsi; | ||
84 | |||
85 | + aio_context_acquire(iscsilun->aio_context); | ||
86 | iscsi_service(iscsi, POLLOUT); | ||
87 | iscsi_set_events(iscsilun); | ||
88 | + aio_context_release(iscsilun->aio_context); | ||
89 | } | ||
90 | |||
91 | static int64_t sector_lun2qemu(int64_t sector, IscsiLun *iscsilun) | ||
92 | diff --git a/block/linux-aio.c b/block/linux-aio.c | ||
93 | index XXXXXXX..XXXXXXX 100644 | ||
94 | --- a/block/linux-aio.c | ||
95 | +++ b/block/linux-aio.c | ||
96 | @@ -XXX,XX +XXX,XX @@ static void qemu_laio_completion_cb(EventNotifier *e) | ||
97 | LinuxAioState *s = container_of(e, LinuxAioState, e); | ||
98 | |||
99 | if (event_notifier_test_and_clear(&s->e)) { | ||
100 | + aio_context_acquire(s->aio_context); | ||
101 | qemu_laio_process_completions_and_submit(s); | ||
102 | + aio_context_release(s->aio_context); | ||
103 | } | ||
104 | } | ||
105 | |||
106 | @@ -XXX,XX +XXX,XX @@ static bool qemu_laio_poll_cb(void *opaque) | ||
107 | return false; | ||
108 | } | ||
109 | |||
110 | + aio_context_acquire(s->aio_context); | ||
111 | qemu_laio_process_completions_and_submit(s); | ||
112 | + aio_context_release(s->aio_context); | ||
67 | return true; | 113 | return true; |
68 | } | 114 | } |
69 | 115 | ||
70 | -/* fix bucket parameters */ | 116 | diff --git a/block/nfs.c b/block/nfs.c |
71 | -static void throttle_fix_bucket(LeakyBucket *bkt) | 117 | index XXXXXXX..XXXXXXX 100644 |
118 | --- a/block/nfs.c | ||
119 | +++ b/block/nfs.c | ||
120 | @@ -XXX,XX +XXX,XX @@ static void nfs_set_events(NFSClient *client) | ||
121 | static void nfs_process_read(void *arg) | ||
122 | { | ||
123 | NFSClient *client = arg; | ||
124 | + | ||
125 | + aio_context_acquire(client->aio_context); | ||
126 | nfs_service(client->context, POLLIN); | ||
127 | nfs_set_events(client); | ||
128 | + aio_context_release(client->aio_context); | ||
129 | } | ||
130 | |||
131 | static void nfs_process_write(void *arg) | ||
132 | { | ||
133 | NFSClient *client = arg; | ||
134 | + | ||
135 | + aio_context_acquire(client->aio_context); | ||
136 | nfs_service(client->context, POLLOUT); | ||
137 | nfs_set_events(client); | ||
138 | + aio_context_release(client->aio_context); | ||
139 | } | ||
140 | |||
141 | static void nfs_co_init_task(BlockDriverState *bs, NFSRPC *task) | ||
142 | diff --git a/block/sheepdog.c b/block/sheepdog.c | ||
143 | index XXXXXXX..XXXXXXX 100644 | ||
144 | --- a/block/sheepdog.c | ||
145 | +++ b/block/sheepdog.c | ||
146 | @@ -XXX,XX +XXX,XX @@ static coroutine_fn int send_co_req(int sockfd, SheepdogReq *hdr, void *data, | ||
147 | return ret; | ||
148 | } | ||
149 | |||
150 | -static void restart_co_req(void *opaque) | ||
72 | -{ | 151 | -{ |
73 | - double min; | 152 | - Coroutine *co = opaque; |
74 | - | 153 | - |
75 | - /* zero bucket level */ | 154 | - qemu_coroutine_enter(co); |
76 | - bkt->level = bkt->burst_level = 0; | ||
77 | - | ||
78 | - /* If bkt->max is 0 we still want to allow short bursts of I/O | ||
79 | - * from the guest, otherwise every other request will be throttled | ||
80 | - * and performance will suffer considerably. */ | ||
81 | - min = bkt->avg / 10; | ||
82 | - if (bkt->avg && !bkt->max) { | ||
83 | - bkt->max = min; | ||
84 | - } | ||
85 | -} | 155 | -} |
86 | - | 156 | - |
87 | -/* undo internal bucket parameter changes (see throttle_fix_bucket()) */ | 157 | typedef struct SheepdogReqCo { |
88 | -static void throttle_unfix_bucket(LeakyBucket *bkt) | 158 | int sockfd; |
89 | -{ | 159 | BlockDriverState *bs; |
90 | - if (bkt->max < bkt->avg) { | 160 | @@ -XXX,XX +XXX,XX @@ typedef struct SheepdogReqCo { |
91 | - bkt->max = 0; | 161 | unsigned int *rlen; |
92 | - } | 162 | int ret; |
163 | bool finished; | ||
164 | + Coroutine *co; | ||
165 | } SheepdogReqCo; | ||
166 | |||
167 | +static void restart_co_req(void *opaque) | ||
168 | +{ | ||
169 | + SheepdogReqCo *srco = opaque; | ||
170 | + | ||
171 | + aio_co_wake(srco->co); | ||
172 | +} | ||
173 | + | ||
174 | static coroutine_fn void do_co_req(void *opaque) | ||
175 | { | ||
176 | int ret; | ||
177 | - Coroutine *co; | ||
178 | SheepdogReqCo *srco = opaque; | ||
179 | int sockfd = srco->sockfd; | ||
180 | SheepdogReq *hdr = srco->hdr; | ||
181 | @@ -XXX,XX +XXX,XX @@ static coroutine_fn void do_co_req(void *opaque) | ||
182 | unsigned int *wlen = srco->wlen; | ||
183 | unsigned int *rlen = srco->rlen; | ||
184 | |||
185 | - co = qemu_coroutine_self(); | ||
186 | + srco->co = qemu_coroutine_self(); | ||
187 | aio_set_fd_handler(srco->aio_context, sockfd, false, | ||
188 | - NULL, restart_co_req, NULL, co); | ||
189 | + NULL, restart_co_req, NULL, srco); | ||
190 | |||
191 | ret = send_co_req(sockfd, hdr, data, wlen); | ||
192 | if (ret < 0) { | ||
193 | @@ -XXX,XX +XXX,XX @@ static coroutine_fn void do_co_req(void *opaque) | ||
194 | } | ||
195 | |||
196 | aio_set_fd_handler(srco->aio_context, sockfd, false, | ||
197 | - restart_co_req, NULL, NULL, co); | ||
198 | + restart_co_req, NULL, NULL, srco); | ||
199 | |||
200 | ret = qemu_co_recv(sockfd, hdr, sizeof(*hdr)); | ||
201 | if (ret != sizeof(*hdr)) { | ||
202 | @@ -XXX,XX +XXX,XX @@ out: | ||
203 | aio_set_fd_handler(srco->aio_context, sockfd, false, | ||
204 | NULL, NULL, NULL, NULL); | ||
205 | |||
206 | + srco->co = NULL; | ||
207 | srco->ret = ret; | ||
208 | srco->finished = true; | ||
209 | if (srco->bs) { | ||
210 | @@ -XXX,XX +XXX,XX @@ static void coroutine_fn aio_read_response(void *opaque) | ||
211 | * We've finished all requests which belong to the AIOCB, so | ||
212 | * we can switch back to sd_co_readv/writev now. | ||
213 | */ | ||
214 | - qemu_coroutine_enter(acb->coroutine); | ||
215 | + aio_co_wake(acb->coroutine); | ||
216 | } | ||
217 | |||
218 | return; | ||
219 | @@ -XXX,XX +XXX,XX @@ static void co_read_response(void *opaque) | ||
220 | s->co_recv = qemu_coroutine_create(aio_read_response, opaque); | ||
221 | } | ||
222 | |||
223 | - qemu_coroutine_enter(s->co_recv); | ||
224 | + aio_co_wake(s->co_recv); | ||
225 | } | ||
226 | |||
227 | static void co_write_request(void *opaque) | ||
228 | { | ||
229 | BDRVSheepdogState *s = opaque; | ||
230 | |||
231 | - qemu_coroutine_enter(s->co_send); | ||
232 | + aio_co_wake(s->co_send); | ||
233 | } | ||
234 | |||
235 | /* | ||
236 | diff --git a/block/ssh.c b/block/ssh.c | ||
237 | index XXXXXXX..XXXXXXX 100644 | ||
238 | --- a/block/ssh.c | ||
239 | +++ b/block/ssh.c | ||
240 | @@ -XXX,XX +XXX,XX @@ static void restart_coroutine(void *opaque) | ||
241 | |||
242 | DPRINTF("co=%p", co); | ||
243 | |||
244 | - qemu_coroutine_enter(co); | ||
245 | + aio_co_wake(co); | ||
246 | } | ||
247 | |||
248 | -static coroutine_fn void set_fd_handler(BDRVSSHState *s, BlockDriverState *bs) | ||
249 | +/* A non-blocking call returned EAGAIN, so yield, ensuring the | ||
250 | + * handlers are set up so that we'll be rescheduled when there is an | ||
251 | + * interesting event on the socket. | ||
252 | + */ | ||
253 | +static coroutine_fn void co_yield(BDRVSSHState *s, BlockDriverState *bs) | ||
254 | { | ||
255 | int r; | ||
256 | IOHandler *rd_handler = NULL, *wr_handler = NULL; | ||
257 | @@ -XXX,XX +XXX,XX @@ static coroutine_fn void set_fd_handler(BDRVSSHState *s, BlockDriverState *bs) | ||
258 | |||
259 | aio_set_fd_handler(bdrv_get_aio_context(bs), s->sock, | ||
260 | false, rd_handler, wr_handler, NULL, co); | ||
93 | -} | 261 | -} |
94 | - | 262 | - |
95 | /* Used to configure the throttle | 263 | -static coroutine_fn void clear_fd_handler(BDRVSSHState *s, |
96 | * | 264 | - BlockDriverState *bs) |
97 | * @ts: the throttle state we are working on | 265 | -{ |
98 | @@ -XXX,XX +XXX,XX @@ void throttle_config(ThrottleState *ts, | 266 | - DPRINTF("s->sock=%d", s->sock); |
99 | 267 | - aio_set_fd_handler(bdrv_get_aio_context(bs), s->sock, | |
100 | ts->cfg = *cfg; | 268 | - false, NULL, NULL, NULL, NULL); |
101 | 269 | -} | |
102 | + /* Zero bucket level */ | ||
103 | for (i = 0; i < BUCKETS_COUNT; i++) { | ||
104 | - throttle_fix_bucket(&ts->cfg.buckets[i]); | ||
105 | + ts->cfg.buckets[i].level = 0; | ||
106 | + ts->cfg.buckets[i].burst_level = 0; | ||
107 | } | ||
108 | |||
109 | ts->previous_leak = qemu_clock_get_ns(clock_type); | ||
110 | @@ -XXX,XX +XXX,XX @@ void throttle_config(ThrottleState *ts, | ||
111 | */ | ||
112 | void throttle_get_config(ThrottleState *ts, ThrottleConfig *cfg) | ||
113 | { | ||
114 | - int i; | ||
115 | - | 270 | - |
116 | *cfg = ts->cfg; | 271 | -/* A non-blocking call returned EAGAIN, so yield, ensuring the |
272 | - * handlers are set up so that we'll be rescheduled when there is an | ||
273 | - * interesting event on the socket. | ||
274 | - */ | ||
275 | -static coroutine_fn void co_yield(BDRVSSHState *s, BlockDriverState *bs) | ||
276 | -{ | ||
277 | - set_fd_handler(s, bs); | ||
278 | qemu_coroutine_yield(); | ||
279 | - clear_fd_handler(s, bs); | ||
280 | + DPRINTF("s->sock=%d - back", s->sock); | ||
281 | + aio_set_fd_handler(bdrv_get_aio_context(bs), s->sock, false, | ||
282 | + NULL, NULL, NULL, NULL); | ||
283 | } | ||
284 | |||
285 | /* SFTP has a function `libssh2_sftp_seek64' which seeks to a position | ||
286 | diff --git a/block/win32-aio.c b/block/win32-aio.c | ||
287 | index XXXXXXX..XXXXXXX 100644 | ||
288 | --- a/block/win32-aio.c | ||
289 | +++ b/block/win32-aio.c | ||
290 | @@ -XXX,XX +XXX,XX @@ struct QEMUWin32AIOState { | ||
291 | HANDLE hIOCP; | ||
292 | EventNotifier e; | ||
293 | int count; | ||
294 | - bool is_aio_context_attached; | ||
295 | + AioContext *aio_ctx; | ||
296 | }; | ||
297 | |||
298 | typedef struct QEMUWin32AIOCB { | ||
299 | @@ -XXX,XX +XXX,XX @@ static void win32_aio_process_completion(QEMUWin32AIOState *s, | ||
300 | } | ||
301 | |||
302 | |||
303 | + aio_context_acquire(s->aio_ctx); | ||
304 | waiocb->common.cb(waiocb->common.opaque, ret); | ||
305 | + aio_context_release(s->aio_ctx); | ||
306 | qemu_aio_unref(waiocb); | ||
307 | } | ||
308 | |||
309 | @@ -XXX,XX +XXX,XX @@ void win32_aio_detach_aio_context(QEMUWin32AIOState *aio, | ||
310 | AioContext *old_context) | ||
311 | { | ||
312 | aio_set_event_notifier(old_context, &aio->e, false, NULL, NULL); | ||
313 | - aio->is_aio_context_attached = false; | ||
314 | + aio->aio_ctx = NULL; | ||
315 | } | ||
316 | |||
317 | void win32_aio_attach_aio_context(QEMUWin32AIOState *aio, | ||
318 | AioContext *new_context) | ||
319 | { | ||
320 | - aio->is_aio_context_attached = true; | ||
321 | + aio->aio_ctx = new_context; | ||
322 | aio_set_event_notifier(new_context, &aio->e, false, | ||
323 | win32_aio_completion_cb, NULL); | ||
324 | } | ||
325 | @@ -XXX,XX +XXX,XX @@ out_free_state: | ||
326 | |||
327 | void win32_aio_cleanup(QEMUWin32AIOState *aio) | ||
328 | { | ||
329 | - assert(!aio->is_aio_context_attached); | ||
330 | + assert(!aio->aio_ctx); | ||
331 | CloseHandle(aio->hIOCP); | ||
332 | event_notifier_cleanup(&aio->e); | ||
333 | g_free(aio); | ||
334 | diff --git a/hw/block/virtio-blk.c b/hw/block/virtio-blk.c | ||
335 | index XXXXXXX..XXXXXXX 100644 | ||
336 | --- a/hw/block/virtio-blk.c | ||
337 | +++ b/hw/block/virtio-blk.c | ||
338 | @@ -XXX,XX +XXX,XX @@ static void virtio_blk_ioctl_complete(void *opaque, int status) | ||
339 | { | ||
340 | VirtIOBlockIoctlReq *ioctl_req = opaque; | ||
341 | VirtIOBlockReq *req = ioctl_req->req; | ||
342 | - VirtIODevice *vdev = VIRTIO_DEVICE(req->dev); | ||
343 | + VirtIOBlock *s = req->dev; | ||
344 | + VirtIODevice *vdev = VIRTIO_DEVICE(s); | ||
345 | struct virtio_scsi_inhdr *scsi; | ||
346 | struct sg_io_hdr *hdr; | ||
347 | |||
348 | @@ -XXX,XX +XXX,XX @@ bool virtio_blk_handle_vq(VirtIOBlock *s, VirtQueue *vq) | ||
349 | MultiReqBuffer mrb = {}; | ||
350 | bool progress = false; | ||
351 | |||
352 | + aio_context_acquire(blk_get_aio_context(s->blk)); | ||
353 | blk_io_plug(s->blk); | ||
354 | |||
355 | do { | ||
356 | @@ -XXX,XX +XXX,XX @@ bool virtio_blk_handle_vq(VirtIOBlock *s, VirtQueue *vq) | ||
357 | } | ||
358 | |||
359 | blk_io_unplug(s->blk); | ||
360 | + aio_context_release(blk_get_aio_context(s->blk)); | ||
361 | return progress; | ||
362 | } | ||
363 | |||
364 | diff --git a/hw/scsi/virtio-scsi.c b/hw/scsi/virtio-scsi.c | ||
365 | index XXXXXXX..XXXXXXX 100644 | ||
366 | --- a/hw/scsi/virtio-scsi.c | ||
367 | +++ b/hw/scsi/virtio-scsi.c | ||
368 | @@ -XXX,XX +XXX,XX @@ bool virtio_scsi_handle_ctrl_vq(VirtIOSCSI *s, VirtQueue *vq) | ||
369 | VirtIOSCSIReq *req; | ||
370 | bool progress = false; | ||
371 | |||
372 | + virtio_scsi_acquire(s); | ||
373 | while ((req = virtio_scsi_pop_req(s, vq))) { | ||
374 | progress = true; | ||
375 | virtio_scsi_handle_ctrl_req(s, req); | ||
376 | } | ||
377 | + virtio_scsi_release(s); | ||
378 | return progress; | ||
379 | } | ||
380 | |||
381 | @@ -XXX,XX +XXX,XX @@ bool virtio_scsi_handle_cmd_vq(VirtIOSCSI *s, VirtQueue *vq) | ||
382 | |||
383 | QTAILQ_HEAD(, VirtIOSCSIReq) reqs = QTAILQ_HEAD_INITIALIZER(reqs); | ||
384 | |||
385 | + virtio_scsi_acquire(s); | ||
386 | do { | ||
387 | virtio_queue_set_notification(vq, 0); | ||
388 | |||
389 | @@ -XXX,XX +XXX,XX @@ bool virtio_scsi_handle_cmd_vq(VirtIOSCSI *s, VirtQueue *vq) | ||
390 | QTAILQ_FOREACH_SAFE(req, &reqs, next, next) { | ||
391 | virtio_scsi_handle_cmd_req_submit(s, req); | ||
392 | } | ||
393 | + virtio_scsi_release(s); | ||
394 | return progress; | ||
395 | } | ||
396 | |||
397 | @@ -XXX,XX +XXX,XX @@ out: | ||
398 | |||
399 | bool virtio_scsi_handle_event_vq(VirtIOSCSI *s, VirtQueue *vq) | ||
400 | { | ||
401 | + virtio_scsi_acquire(s); | ||
402 | if (s->events_dropped) { | ||
403 | virtio_scsi_push_event(s, NULL, VIRTIO_SCSI_T_NO_EVENT, 0); | ||
404 | + virtio_scsi_release(s); | ||
405 | return true; | ||
406 | } | ||
407 | + virtio_scsi_release(s); | ||
408 | return false; | ||
409 | } | ||
410 | |||
411 | diff --git a/util/aio-posix.c b/util/aio-posix.c | ||
412 | index XXXXXXX..XXXXXXX 100644 | ||
413 | --- a/util/aio-posix.c | ||
414 | +++ b/util/aio-posix.c | ||
415 | @@ -XXX,XX +XXX,XX @@ static bool aio_dispatch_handlers(AioContext *ctx) | ||
416 | (revents & (G_IO_IN | G_IO_HUP | G_IO_ERR)) && | ||
417 | aio_node_check(ctx, node->is_external) && | ||
418 | node->io_read) { | ||
419 | - aio_context_acquire(ctx); | ||
420 | node->io_read(node->opaque); | ||
421 | - aio_context_release(ctx); | ||
422 | |||
423 | /* aio_notify() does not count as progress */ | ||
424 | if (node->opaque != &ctx->notifier) { | ||
425 | @@ -XXX,XX +XXX,XX @@ static bool aio_dispatch_handlers(AioContext *ctx) | ||
426 | (revents & (G_IO_OUT | G_IO_ERR)) && | ||
427 | aio_node_check(ctx, node->is_external) && | ||
428 | node->io_write) { | ||
429 | - aio_context_acquire(ctx); | ||
430 | node->io_write(node->opaque); | ||
431 | - aio_context_release(ctx); | ||
432 | progress = true; | ||
433 | } | ||
434 | |||
435 | @@ -XXX,XX +XXX,XX @@ bool aio_poll(AioContext *ctx, bool blocking) | ||
436 | start = qemu_clock_get_ns(QEMU_CLOCK_REALTIME); | ||
437 | } | ||
438 | |||
439 | - aio_context_acquire(ctx); | ||
440 | progress = try_poll_mode(ctx, blocking); | ||
441 | - aio_context_release(ctx); | ||
117 | - | 442 | - |
118 | - for (i = 0; i < BUCKETS_COUNT; i++) { | 443 | if (!progress) { |
119 | - throttle_unfix_bucket(&cfg->buckets[i]); | 444 | assert(npfd == 0); |
120 | - } | 445 | |
121 | } | 446 | diff --git a/util/aio-win32.c b/util/aio-win32.c |
122 | 447 | index XXXXXXX..XXXXXXX 100644 | |
448 | --- a/util/aio-win32.c | ||
449 | +++ b/util/aio-win32.c | ||
450 | @@ -XXX,XX +XXX,XX @@ static bool aio_dispatch_handlers(AioContext *ctx, HANDLE event) | ||
451 | (revents || event_notifier_get_handle(node->e) == event) && | ||
452 | node->io_notify) { | ||
453 | node->pfd.revents = 0; | ||
454 | - aio_context_acquire(ctx); | ||
455 | node->io_notify(node->e); | ||
456 | - aio_context_release(ctx); | ||
457 | |||
458 | /* aio_notify() does not count as progress */ | ||
459 | if (node->e != &ctx->notifier) { | ||
460 | @@ -XXX,XX +XXX,XX @@ static bool aio_dispatch_handlers(AioContext *ctx, HANDLE event) | ||
461 | (node->io_read || node->io_write)) { | ||
462 | node->pfd.revents = 0; | ||
463 | if ((revents & G_IO_IN) && node->io_read) { | ||
464 | - aio_context_acquire(ctx); | ||
465 | node->io_read(node->opaque); | ||
466 | - aio_context_release(ctx); | ||
467 | progress = true; | ||
468 | } | ||
469 | if ((revents & G_IO_OUT) && node->io_write) { | ||
470 | - aio_context_acquire(ctx); | ||
471 | node->io_write(node->opaque); | ||
472 | - aio_context_release(ctx); | ||
473 | progress = true; | ||
474 | } | ||
123 | 475 | ||
124 | -- | 476 | -- |
125 | 2.13.5 | 477 | 2.9.3 |
126 | 478 | ||
127 | 479 | diff view generated by jsdifflib |
1 | From: Alberto Garcia <berto@igalia.com> | 1 | From: Paolo Bonzini <pbonzini@redhat.com> |
---|---|---|---|
2 | 2 | ||
3 | LeakyBucket.burst_length is defined as an unsigned integer but the | 3 | Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> |
4 | code never checks for overflows and it only makes sure that the value | 4 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> |
5 | is not 0. | 5 | Reviewed-by: Fam Zheng <famz@redhat.com> |
6 | 6 | Reviewed-by: Daniel P. Berrange <berrange@redhat.com> | |
7 | In practice this means that the user can set something like | 7 | Message-id: 20170213135235.12274-15-pbonzini@redhat.com |
8 | throttling.iops-total-max-length=4294967300 despite being larger than | ||
9 | UINT_MAX and the final value after casting to unsigned int will be 4. | ||
10 | |||
11 | This patch changes the data type to uint64_t. This does not increase | ||
12 | the storage size of LeakyBucket, and allows us to assign the value | ||
13 | directly from qemu_opt_get_number() or BlockIOThrottle and then do the | ||
14 | checks directly in throttle_is_valid(). | ||
15 | |||
16 | The value of burst_length does not have a specific upper limit, | ||
17 | but since the bucket size is defined by max * burst_length we have | ||
18 | to prevent overflows. Instead of going for UINT64_MAX or something | ||
19 | similar this patch reuses THROTTLE_VALUE_MAX, which allows I/O bursts | ||
20 | of 1 GiB/s for 10 days in a row. | ||
21 | |||
22 | Signed-off-by: Alberto Garcia <berto@igalia.com> | ||
23 | Message-id: 1b2e3049803f71cafb2e1fa1be4fb47147a0d398.1503580370.git.berto@igalia.com | ||
24 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | 8 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
25 | --- | 9 | --- |
26 | include/qemu/throttle.h | 2 +- | 10 | block/archipelago.c | 3 +++ |
27 | util/throttle.c | 5 +++++ | 11 | block/blkreplay.c | 2 +- |
28 | 2 files changed, 6 insertions(+), 1 deletion(-) | 12 | block/block-backend.c | 6 ++++++ |
13 | block/curl.c | 26 ++++++++++++++++++-------- | ||
14 | block/gluster.c | 9 +-------- | ||
15 | block/io.c | 6 +++++- | ||
16 | block/iscsi.c | 6 +++++- | ||
17 | block/linux-aio.c | 15 +++++++++------ | ||
18 | block/nfs.c | 3 ++- | ||
19 | block/null.c | 4 ++++ | ||
20 | block/qed.c | 3 +++ | ||
21 | block/rbd.c | 4 ++++ | ||
22 | dma-helpers.c | 2 ++ | ||
23 | hw/block/virtio-blk.c | 2 ++ | ||
24 | hw/scsi/scsi-bus.c | 2 ++ | ||
25 | util/async.c | 4 ++-- | ||
26 | util/thread-pool.c | 2 ++ | ||
27 | 17 files changed, 71 insertions(+), 28 deletions(-) | ||
29 | 28 | ||
30 | diff --git a/include/qemu/throttle.h b/include/qemu/throttle.h | 29 | diff --git a/block/archipelago.c b/block/archipelago.c |
31 | index XXXXXXX..XXXXXXX 100644 | 30 | index XXXXXXX..XXXXXXX 100644 |
32 | --- a/include/qemu/throttle.h | 31 | --- a/block/archipelago.c |
33 | +++ b/include/qemu/throttle.h | 32 | +++ b/block/archipelago.c |
34 | @@ -XXX,XX +XXX,XX @@ typedef struct LeakyBucket { | 33 | @@ -XXX,XX +XXX,XX @@ static void qemu_archipelago_complete_aio(void *opaque) |
35 | uint64_t max; /* leaky bucket max burst in units */ | 34 | { |
36 | double level; /* bucket level in units */ | 35 | AIORequestData *reqdata = (AIORequestData *) opaque; |
37 | double burst_level; /* bucket level in units (for computing bursts) */ | 36 | ArchipelagoAIOCB *aio_cb = (ArchipelagoAIOCB *) reqdata->aio_cb; |
38 | - unsigned burst_length; /* max length of the burst period, in seconds */ | 37 | + AioContext *ctx = bdrv_get_aio_context(aio_cb->common.bs); |
39 | + uint64_t burst_length; /* max length of the burst period, in seconds */ | 38 | |
40 | } LeakyBucket; | 39 | + aio_context_acquire(ctx); |
41 | 40 | aio_cb->common.cb(aio_cb->common.opaque, aio_cb->ret); | |
42 | /* The following structure is used to configure a ThrottleState | 41 | + aio_context_release(ctx); |
43 | diff --git a/util/throttle.c b/util/throttle.c | 42 | aio_cb->status = 0; |
44 | index XXXXXXX..XXXXXXX 100644 | 43 | |
45 | --- a/util/throttle.c | 44 | qemu_aio_unref(aio_cb); |
46 | +++ b/util/throttle.c | 45 | diff --git a/block/blkreplay.c b/block/blkreplay.c |
47 | @@ -XXX,XX +XXX,XX @@ bool throttle_is_valid(ThrottleConfig *cfg, Error **errp) | 46 | index XXXXXXX..XXXXXXX 100755 |
48 | return false; | 47 | --- a/block/blkreplay.c |
48 | +++ b/block/blkreplay.c | ||
49 | @@ -XXX,XX +XXX,XX @@ static int64_t blkreplay_getlength(BlockDriverState *bs) | ||
50 | static void blkreplay_bh_cb(void *opaque) | ||
51 | { | ||
52 | Request *req = opaque; | ||
53 | - qemu_coroutine_enter(req->co); | ||
54 | + aio_co_wake(req->co); | ||
55 | qemu_bh_delete(req->bh); | ||
56 | g_free(req); | ||
57 | } | ||
58 | diff --git a/block/block-backend.c b/block/block-backend.c | ||
59 | index XXXXXXX..XXXXXXX 100644 | ||
60 | --- a/block/block-backend.c | ||
61 | +++ b/block/block-backend.c | ||
62 | @@ -XXX,XX +XXX,XX @@ int blk_make_zero(BlockBackend *blk, BdrvRequestFlags flags) | ||
63 | static void error_callback_bh(void *opaque) | ||
64 | { | ||
65 | struct BlockBackendAIOCB *acb = opaque; | ||
66 | + AioContext *ctx = bdrv_get_aio_context(acb->common.bs); | ||
67 | |||
68 | bdrv_dec_in_flight(acb->common.bs); | ||
69 | + aio_context_acquire(ctx); | ||
70 | acb->common.cb(acb->common.opaque, acb->ret); | ||
71 | + aio_context_release(ctx); | ||
72 | qemu_aio_unref(acb); | ||
73 | } | ||
74 | |||
75 | @@ -XXX,XX +XXX,XX @@ static void blk_aio_complete(BlkAioEmAIOCB *acb) | ||
76 | static void blk_aio_complete_bh(void *opaque) | ||
77 | { | ||
78 | BlkAioEmAIOCB *acb = opaque; | ||
79 | + AioContext *ctx = bdrv_get_aio_context(acb->common.bs); | ||
80 | |||
81 | assert(acb->has_returned); | ||
82 | + aio_context_acquire(ctx); | ||
83 | blk_aio_complete(acb); | ||
84 | + aio_context_release(ctx); | ||
85 | } | ||
86 | |||
87 | static BlockAIOCB *blk_aio_prwv(BlockBackend *blk, int64_t offset, int bytes, | ||
88 | diff --git a/block/curl.c b/block/curl.c | ||
89 | index XXXXXXX..XXXXXXX 100644 | ||
90 | --- a/block/curl.c | ||
91 | +++ b/block/curl.c | ||
92 | @@ -XXX,XX +XXX,XX @@ static void curl_readv_bh_cb(void *p) | ||
93 | { | ||
94 | CURLState *state; | ||
95 | int running; | ||
96 | + int ret = -EINPROGRESS; | ||
97 | |||
98 | CURLAIOCB *acb = p; | ||
99 | - BDRVCURLState *s = acb->common.bs->opaque; | ||
100 | + BlockDriverState *bs = acb->common.bs; | ||
101 | + BDRVCURLState *s = bs->opaque; | ||
102 | + AioContext *ctx = bdrv_get_aio_context(bs); | ||
103 | |||
104 | size_t start = acb->sector_num * BDRV_SECTOR_SIZE; | ||
105 | size_t end; | ||
106 | |||
107 | + aio_context_acquire(ctx); | ||
108 | + | ||
109 | // In case we have the requested data already (e.g. read-ahead), | ||
110 | // we can just call the callback and be done. | ||
111 | switch (curl_find_buf(s, start, acb->nb_sectors * BDRV_SECTOR_SIZE, acb)) { | ||
112 | @@ -XXX,XX +XXX,XX @@ static void curl_readv_bh_cb(void *p) | ||
113 | qemu_aio_unref(acb); | ||
114 | // fall through | ||
115 | case FIND_RET_WAIT: | ||
116 | - return; | ||
117 | + goto out; | ||
118 | default: | ||
119 | break; | ||
120 | } | ||
121 | @@ -XXX,XX +XXX,XX @@ static void curl_readv_bh_cb(void *p) | ||
122 | // No cache found, so let's start a new request | ||
123 | state = curl_init_state(acb->common.bs, s); | ||
124 | if (!state) { | ||
125 | - acb->common.cb(acb->common.opaque, -EIO); | ||
126 | - qemu_aio_unref(acb); | ||
127 | - return; | ||
128 | + ret = -EIO; | ||
129 | + goto out; | ||
130 | } | ||
131 | |||
132 | acb->start = 0; | ||
133 | @@ -XXX,XX +XXX,XX @@ static void curl_readv_bh_cb(void *p) | ||
134 | state->orig_buf = g_try_malloc(state->buf_len); | ||
135 | if (state->buf_len && state->orig_buf == NULL) { | ||
136 | curl_clean_state(state); | ||
137 | - acb->common.cb(acb->common.opaque, -ENOMEM); | ||
138 | - qemu_aio_unref(acb); | ||
139 | - return; | ||
140 | + ret = -ENOMEM; | ||
141 | + goto out; | ||
142 | } | ||
143 | state->acb[0] = acb; | ||
144 | |||
145 | @@ -XXX,XX +XXX,XX @@ static void curl_readv_bh_cb(void *p) | ||
146 | |||
147 | /* Tell curl it needs to kick things off */ | ||
148 | curl_multi_socket_action(s->multi, CURL_SOCKET_TIMEOUT, 0, &running); | ||
149 | + | ||
150 | +out: | ||
151 | + if (ret != -EINPROGRESS) { | ||
152 | + acb->common.cb(acb->common.opaque, ret); | ||
153 | + qemu_aio_unref(acb); | ||
154 | + } | ||
155 | + aio_context_release(ctx); | ||
156 | } | ||
157 | |||
158 | static BlockAIOCB *curl_aio_readv(BlockDriverState *bs, | ||
159 | diff --git a/block/gluster.c b/block/gluster.c | ||
160 | index XXXXXXX..XXXXXXX 100644 | ||
161 | --- a/block/gluster.c | ||
162 | +++ b/block/gluster.c | ||
163 | @@ -XXX,XX +XXX,XX @@ static struct glfs *qemu_gluster_init(BlockdevOptionsGluster *gconf, | ||
164 | return qemu_gluster_glfs_init(gconf, errp); | ||
165 | } | ||
166 | |||
167 | -static void qemu_gluster_complete_aio(void *opaque) | ||
168 | -{ | ||
169 | - GlusterAIOCB *acb = (GlusterAIOCB *)opaque; | ||
170 | - | ||
171 | - qemu_coroutine_enter(acb->coroutine); | ||
172 | -} | ||
173 | - | ||
174 | /* | ||
175 | * AIO callback routine called from GlusterFS thread. | ||
176 | */ | ||
177 | @@ -XXX,XX +XXX,XX @@ static void gluster_finish_aiocb(struct glfs_fd *fd, ssize_t ret, void *arg) | ||
178 | acb->ret = -EIO; /* Partial read/write - fail it */ | ||
179 | } | ||
180 | |||
181 | - aio_bh_schedule_oneshot(acb->aio_context, qemu_gluster_complete_aio, acb); | ||
182 | + aio_co_schedule(acb->aio_context, acb->coroutine); | ||
183 | } | ||
184 | |||
185 | static void qemu_gluster_parse_flags(int bdrv_flags, int *open_flags) | ||
186 | diff --git a/block/io.c b/block/io.c | ||
187 | index XXXXXXX..XXXXXXX 100644 | ||
188 | --- a/block/io.c | ||
189 | +++ b/block/io.c | ||
190 | @@ -XXX,XX +XXX,XX @@ static void bdrv_co_drain_bh_cb(void *opaque) | ||
191 | bdrv_dec_in_flight(bs); | ||
192 | bdrv_drained_begin(bs); | ||
193 | data->done = true; | ||
194 | - qemu_coroutine_enter(co); | ||
195 | + aio_co_wake(co); | ||
196 | } | ||
197 | |||
198 | static void coroutine_fn bdrv_co_yield_to_drain(BlockDriverState *bs) | ||
199 | @@ -XXX,XX +XXX,XX @@ static void bdrv_co_complete(BlockAIOCBCoroutine *acb) | ||
200 | static void bdrv_co_em_bh(void *opaque) | ||
201 | { | ||
202 | BlockAIOCBCoroutine *acb = opaque; | ||
203 | + BlockDriverState *bs = acb->common.bs; | ||
204 | + AioContext *ctx = bdrv_get_aio_context(bs); | ||
205 | |||
206 | assert(!acb->need_bh); | ||
207 | + aio_context_acquire(ctx); | ||
208 | bdrv_co_complete(acb); | ||
209 | + aio_context_release(ctx); | ||
210 | } | ||
211 | |||
212 | static void bdrv_co_maybe_schedule_bh(BlockAIOCBCoroutine *acb) | ||
213 | diff --git a/block/iscsi.c b/block/iscsi.c | ||
214 | index XXXXXXX..XXXXXXX 100644 | ||
215 | --- a/block/iscsi.c | ||
216 | +++ b/block/iscsi.c | ||
217 | @@ -XXX,XX +XXX,XX @@ static void | ||
218 | iscsi_bh_cb(void *p) | ||
219 | { | ||
220 | IscsiAIOCB *acb = p; | ||
221 | + AioContext *ctx = bdrv_get_aio_context(acb->common.bs); | ||
222 | |||
223 | qemu_bh_delete(acb->bh); | ||
224 | |||
225 | g_free(acb->buf); | ||
226 | acb->buf = NULL; | ||
227 | |||
228 | + aio_context_acquire(ctx); | ||
229 | acb->common.cb(acb->common.opaque, acb->status); | ||
230 | + aio_context_release(ctx); | ||
231 | |||
232 | if (acb->task != NULL) { | ||
233 | scsi_free_scsi_task(acb->task); | ||
234 | @@ -XXX,XX +XXX,XX @@ iscsi_schedule_bh(IscsiAIOCB *acb) | ||
235 | static void iscsi_co_generic_bh_cb(void *opaque) | ||
236 | { | ||
237 | struct IscsiTask *iTask = opaque; | ||
238 | + | ||
239 | iTask->complete = 1; | ||
240 | - qemu_coroutine_enter(iTask->co); | ||
241 | + aio_co_wake(iTask->co); | ||
242 | } | ||
243 | |||
244 | static void iscsi_retry_timer_expired(void *opaque) | ||
245 | diff --git a/block/linux-aio.c b/block/linux-aio.c | ||
246 | index XXXXXXX..XXXXXXX 100644 | ||
247 | --- a/block/linux-aio.c | ||
248 | +++ b/block/linux-aio.c | ||
249 | @@ -XXX,XX +XXX,XX @@ struct LinuxAioState { | ||
250 | io_context_t ctx; | ||
251 | EventNotifier e; | ||
252 | |||
253 | - /* io queue for submit at batch */ | ||
254 | + /* io queue for submit at batch. Protected by AioContext lock. */ | ||
255 | LaioQueue io_q; | ||
256 | |||
257 | - /* I/O completion processing */ | ||
258 | + /* I/O completion processing. Only runs in I/O thread. */ | ||
259 | QEMUBH *completion_bh; | ||
260 | int event_idx; | ||
261 | int event_max; | ||
262 | @@ -XXX,XX +XXX,XX @@ static inline ssize_t io_event_ret(struct io_event *ev) | ||
263 | */ | ||
264 | static void qemu_laio_process_completion(struct qemu_laiocb *laiocb) | ||
265 | { | ||
266 | + LinuxAioState *s = laiocb->ctx; | ||
267 | int ret; | ||
268 | |||
269 | ret = laiocb->ret; | ||
270 | @@ -XXX,XX +XXX,XX @@ static void qemu_laio_process_completion(struct qemu_laiocb *laiocb) | ||
271 | } | ||
272 | |||
273 | laiocb->ret = ret; | ||
274 | + aio_context_acquire(s->aio_context); | ||
275 | if (laiocb->co) { | ||
276 | /* If the coroutine is already entered it must be in ioq_submit() and | ||
277 | * will notice laio->ret has been filled in when it eventually runs | ||
278 | @@ -XXX,XX +XXX,XX @@ static void qemu_laio_process_completion(struct qemu_laiocb *laiocb) | ||
279 | laiocb->common.cb(laiocb->common.opaque, ret); | ||
280 | qemu_aio_unref(laiocb); | ||
281 | } | ||
282 | + aio_context_release(s->aio_context); | ||
283 | } | ||
284 | |||
285 | /** | ||
286 | @@ -XXX,XX +XXX,XX @@ static void qemu_laio_process_completions(LinuxAioState *s) | ||
287 | static void qemu_laio_process_completions_and_submit(LinuxAioState *s) | ||
288 | { | ||
289 | qemu_laio_process_completions(s); | ||
290 | + | ||
291 | + aio_context_acquire(s->aio_context); | ||
292 | if (!s->io_q.plugged && !QSIMPLEQ_EMPTY(&s->io_q.pending)) { | ||
293 | ioq_submit(s); | ||
294 | } | ||
295 | + aio_context_release(s->aio_context); | ||
296 | } | ||
297 | |||
298 | static void qemu_laio_completion_bh(void *opaque) | ||
299 | @@ -XXX,XX +XXX,XX @@ static void qemu_laio_completion_cb(EventNotifier *e) | ||
300 | LinuxAioState *s = container_of(e, LinuxAioState, e); | ||
301 | |||
302 | if (event_notifier_test_and_clear(&s->e)) { | ||
303 | - aio_context_acquire(s->aio_context); | ||
304 | qemu_laio_process_completions_and_submit(s); | ||
305 | - aio_context_release(s->aio_context); | ||
306 | } | ||
307 | } | ||
308 | |||
309 | @@ -XXX,XX +XXX,XX @@ static bool qemu_laio_poll_cb(void *opaque) | ||
310 | return false; | ||
311 | } | ||
312 | |||
313 | - aio_context_acquire(s->aio_context); | ||
314 | qemu_laio_process_completions_and_submit(s); | ||
315 | - aio_context_release(s->aio_context); | ||
316 | return true; | ||
317 | } | ||
318 | |||
319 | @@ -XXX,XX +XXX,XX @@ void laio_detach_aio_context(LinuxAioState *s, AioContext *old_context) | ||
320 | { | ||
321 | aio_set_event_notifier(old_context, &s->e, false, NULL, NULL); | ||
322 | qemu_bh_delete(s->completion_bh); | ||
323 | + s->aio_context = NULL; | ||
324 | } | ||
325 | |||
326 | void laio_attach_aio_context(LinuxAioState *s, AioContext *new_context) | ||
327 | diff --git a/block/nfs.c b/block/nfs.c | ||
328 | index XXXXXXX..XXXXXXX 100644 | ||
329 | --- a/block/nfs.c | ||
330 | +++ b/block/nfs.c | ||
331 | @@ -XXX,XX +XXX,XX @@ static void nfs_co_init_task(BlockDriverState *bs, NFSRPC *task) | ||
332 | static void nfs_co_generic_bh_cb(void *opaque) | ||
333 | { | ||
334 | NFSRPC *task = opaque; | ||
335 | + | ||
336 | task->complete = 1; | ||
337 | - qemu_coroutine_enter(task->co); | ||
338 | + aio_co_wake(task->co); | ||
339 | } | ||
340 | |||
341 | static void | ||
342 | diff --git a/block/null.c b/block/null.c | ||
343 | index XXXXXXX..XXXXXXX 100644 | ||
344 | --- a/block/null.c | ||
345 | +++ b/block/null.c | ||
346 | @@ -XXX,XX +XXX,XX @@ static const AIOCBInfo null_aiocb_info = { | ||
347 | static void null_bh_cb(void *opaque) | ||
348 | { | ||
349 | NullAIOCB *acb = opaque; | ||
350 | + AioContext *ctx = bdrv_get_aio_context(acb->common.bs); | ||
351 | + | ||
352 | + aio_context_acquire(ctx); | ||
353 | acb->common.cb(acb->common.opaque, 0); | ||
354 | + aio_context_release(ctx); | ||
355 | qemu_aio_unref(acb); | ||
356 | } | ||
357 | |||
358 | diff --git a/block/qed.c b/block/qed.c | ||
359 | index XXXXXXX..XXXXXXX 100644 | ||
360 | --- a/block/qed.c | ||
361 | +++ b/block/qed.c | ||
362 | @@ -XXX,XX +XXX,XX @@ static void qed_update_l2_table(BDRVQEDState *s, QEDTable *table, int index, | ||
363 | static void qed_aio_complete_bh(void *opaque) | ||
364 | { | ||
365 | QEDAIOCB *acb = opaque; | ||
366 | + BDRVQEDState *s = acb_to_s(acb); | ||
367 | BlockCompletionFunc *cb = acb->common.cb; | ||
368 | void *user_opaque = acb->common.opaque; | ||
369 | int ret = acb->bh_ret; | ||
370 | @@ -XXX,XX +XXX,XX @@ static void qed_aio_complete_bh(void *opaque) | ||
371 | qemu_aio_unref(acb); | ||
372 | |||
373 | /* Invoke callback */ | ||
374 | + qed_acquire(s); | ||
375 | cb(user_opaque, ret); | ||
376 | + qed_release(s); | ||
377 | } | ||
378 | |||
379 | static void qed_aio_complete(QEDAIOCB *acb, int ret) | ||
380 | diff --git a/block/rbd.c b/block/rbd.c | ||
381 | index XXXXXXX..XXXXXXX 100644 | ||
382 | --- a/block/rbd.c | ||
383 | +++ b/block/rbd.c | ||
384 | @@ -XXX,XX +XXX,XX @@ shutdown: | ||
385 | static void qemu_rbd_complete_aio(RADOSCB *rcb) | ||
386 | { | ||
387 | RBDAIOCB *acb = rcb->acb; | ||
388 | + AioContext *ctx = bdrv_get_aio_context(acb->common.bs); | ||
389 | int64_t r; | ||
390 | |||
391 | r = rcb->ret; | ||
392 | @@ -XXX,XX +XXX,XX @@ static void qemu_rbd_complete_aio(RADOSCB *rcb) | ||
393 | qemu_iovec_from_buf(acb->qiov, 0, acb->bounce, acb->qiov->size); | ||
394 | } | ||
395 | qemu_vfree(acb->bounce); | ||
396 | + | ||
397 | + aio_context_acquire(ctx); | ||
398 | acb->common.cb(acb->common.opaque, (acb->ret > 0 ? 0 : acb->ret)); | ||
399 | + aio_context_release(ctx); | ||
400 | |||
401 | qemu_aio_unref(acb); | ||
402 | } | ||
403 | diff --git a/dma-helpers.c b/dma-helpers.c | ||
404 | index XXXXXXX..XXXXXXX 100644 | ||
405 | --- a/dma-helpers.c | ||
406 | +++ b/dma-helpers.c | ||
407 | @@ -XXX,XX +XXX,XX @@ static void dma_blk_cb(void *opaque, int ret) | ||
408 | QEMU_ALIGN_DOWN(dbs->iov.size, dbs->align)); | ||
409 | } | ||
410 | |||
411 | + aio_context_acquire(dbs->ctx); | ||
412 | dbs->acb = dbs->io_func(dbs->offset, &dbs->iov, | ||
413 | dma_blk_cb, dbs, dbs->io_func_opaque); | ||
414 | + aio_context_release(dbs->ctx); | ||
415 | assert(dbs->acb); | ||
416 | } | ||
417 | |||
418 | diff --git a/hw/block/virtio-blk.c b/hw/block/virtio-blk.c | ||
419 | index XXXXXXX..XXXXXXX 100644 | ||
420 | --- a/hw/block/virtio-blk.c | ||
421 | +++ b/hw/block/virtio-blk.c | ||
422 | @@ -XXX,XX +XXX,XX @@ static void virtio_blk_dma_restart_bh(void *opaque) | ||
423 | |||
424 | s->rq = NULL; | ||
425 | |||
426 | + aio_context_acquire(blk_get_aio_context(s->conf.conf.blk)); | ||
427 | while (req) { | ||
428 | VirtIOBlockReq *next = req->next; | ||
429 | if (virtio_blk_handle_request(req, &mrb)) { | ||
430 | @@ -XXX,XX +XXX,XX @@ static void virtio_blk_dma_restart_bh(void *opaque) | ||
431 | if (mrb.num_reqs) { | ||
432 | virtio_blk_submit_multireq(s->blk, &mrb); | ||
433 | } | ||
434 | + aio_context_release(blk_get_aio_context(s->conf.conf.blk)); | ||
435 | } | ||
436 | |||
437 | static void virtio_blk_dma_restart_cb(void *opaque, int running, | ||
438 | diff --git a/hw/scsi/scsi-bus.c b/hw/scsi/scsi-bus.c | ||
439 | index XXXXXXX..XXXXXXX 100644 | ||
440 | --- a/hw/scsi/scsi-bus.c | ||
441 | +++ b/hw/scsi/scsi-bus.c | ||
442 | @@ -XXX,XX +XXX,XX @@ static void scsi_dma_restart_bh(void *opaque) | ||
443 | qemu_bh_delete(s->bh); | ||
444 | s->bh = NULL; | ||
445 | |||
446 | + aio_context_acquire(blk_get_aio_context(s->conf.blk)); | ||
447 | QTAILQ_FOREACH_SAFE(req, &s->requests, next, next) { | ||
448 | scsi_req_ref(req); | ||
449 | if (req->retry) { | ||
450 | @@ -XXX,XX +XXX,XX @@ static void scsi_dma_restart_bh(void *opaque) | ||
49 | } | 451 | } |
50 | 452 | scsi_req_unref(req); | |
51 | + if (bkt->max && bkt->burst_length > THROTTLE_VALUE_MAX / bkt->max) { | 453 | } |
52 | + error_setg(errp, "burst length too high for this burst rate"); | 454 | + aio_context_release(blk_get_aio_context(s->conf.blk)); |
53 | + return false; | 455 | } |
54 | + } | 456 | |
55 | + | 457 | void scsi_req_retry(SCSIRequest *req) |
56 | if (bkt->max && !bkt->avg) { | 458 | diff --git a/util/async.c b/util/async.c |
57 | error_setg(errp, "bps_max/iops_max require corresponding" | 459 | index XXXXXXX..XXXXXXX 100644 |
58 | " bps/iops values"); | 460 | --- a/util/async.c |
461 | +++ b/util/async.c | ||
462 | @@ -XXX,XX +XXX,XX @@ int aio_bh_poll(AioContext *ctx) | ||
463 | ret = 1; | ||
464 | } | ||
465 | bh->idle = 0; | ||
466 | - aio_context_acquire(ctx); | ||
467 | aio_bh_call(bh); | ||
468 | - aio_context_release(ctx); | ||
469 | } | ||
470 | if (bh->deleted) { | ||
471 | deleted = true; | ||
472 | @@ -XXX,XX +XXX,XX @@ static void co_schedule_bh_cb(void *opaque) | ||
473 | Coroutine *co = QSLIST_FIRST(&straight); | ||
474 | QSLIST_REMOVE_HEAD(&straight, co_scheduled_next); | ||
475 | trace_aio_co_schedule_bh_cb(ctx, co); | ||
476 | + aio_context_acquire(ctx); | ||
477 | qemu_coroutine_enter(co); | ||
478 | + aio_context_release(ctx); | ||
479 | } | ||
480 | } | ||
481 | |||
482 | diff --git a/util/thread-pool.c b/util/thread-pool.c | ||
483 | index XXXXXXX..XXXXXXX 100644 | ||
484 | --- a/util/thread-pool.c | ||
485 | +++ b/util/thread-pool.c | ||
486 | @@ -XXX,XX +XXX,XX @@ static void thread_pool_completion_bh(void *opaque) | ||
487 | ThreadPool *pool = opaque; | ||
488 | ThreadPoolElement *elem, *next; | ||
489 | |||
490 | + aio_context_acquire(pool->ctx); | ||
491 | restart: | ||
492 | QLIST_FOREACH_SAFE(elem, &pool->head, all, next) { | ||
493 | if (elem->state != THREAD_DONE) { | ||
494 | @@ -XXX,XX +XXX,XX @@ restart: | ||
495 | qemu_aio_unref(elem); | ||
496 | } | ||
497 | } | ||
498 | + aio_context_release(pool->ctx); | ||
499 | } | ||
500 | |||
501 | static void thread_pool_cancel(BlockAIOCB *acb) | ||
59 | -- | 502 | -- |
60 | 2.13.5 | 503 | 2.9.3 |
61 | 504 | ||
62 | 505 | diff view generated by jsdifflib |
1 | From: Alberto Garcia <berto@igalia.com> | 1 | From: Paolo Bonzini <pbonzini@redhat.com> |
---|---|---|---|
2 | 2 | ||
3 | The way the throttling algorithm works is that requests start being | 3 | Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> |
4 | throttled once the bucket level exceeds the burst limit. When we get | 4 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> |
5 | there the bucket leaks at the level set by the user (bkt->avg), and | 5 | Reviewed-by: Fam Zheng <famz@redhat.com> |
6 | that leak rate is what prevents guest I/O from exceeding the desired | 6 | Reviewed-by: Daniel P. Berrange <berrange@redhat.com> |
7 | limit. | 7 | Message-id: 20170213135235.12274-16-pbonzini@redhat.com |
8 | |||
9 | If we don't allow bursts (i.e. bkt->max == 0) then we can start | ||
10 | throttling requests immediately. The problem with keeping the | ||
11 | threshold at 0 is that it only allows one request at a time, and as | ||
12 | soon as there's a bit of I/O from the guest every other request will | ||
13 | be throttled and performance will suffer considerably. That can even | ||
14 | make the guest unable to reach the throttle limit if that limit is | ||
15 | high enough, and that happens regardless of the block scheduler used | ||
16 | by the guest. | ||
17 | |||
18 | Increasing that threshold gives flexibility to the guest, allowing it | ||
19 | to perform short bursts of I/O before being throttled. Increasing the | ||
20 | threshold too much does not make a difference in the long run (because | ||
21 | it's the leak rate what defines the actual throughput) but it does | ||
22 | allow the guest to perform longer initial bursts and exceed the | ||
23 | throttle limit for a short while. | ||
24 | |||
25 | A burst value of bkt->avg / 10 allows the guest to perform 100ms' | ||
26 | worth of I/O at the target rate without being throttled. | ||
27 | |||
28 | Signed-off-by: Alberto Garcia <berto@igalia.com> | ||
29 | Message-id: 31aae6645f0d1fbf3860fb2b528b757236f0c0a7.1503580370.git.berto@igalia.com | ||
30 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | 8 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
31 | --- | 9 | --- |
32 | util/throttle.c | 11 +++-------- | 10 | block/archipelago.c | 3 --- |
33 | 1 file changed, 3 insertions(+), 8 deletions(-) | 11 | block/block-backend.c | 7 ------- |
12 | block/curl.c | 2 +- | ||
13 | block/io.c | 6 +----- | ||
14 | block/iscsi.c | 3 --- | ||
15 | block/linux-aio.c | 5 +---- | ||
16 | block/mirror.c | 12 +++++++++--- | ||
17 | block/null.c | 8 -------- | ||
18 | block/qed-cluster.c | 2 ++ | ||
19 | block/qed-table.c | 12 ++++++++++-- | ||
20 | block/qed.c | 4 ++-- | ||
21 | block/rbd.c | 4 ---- | ||
22 | block/win32-aio.c | 3 --- | ||
23 | hw/block/virtio-blk.c | 12 +++++++++++- | ||
24 | hw/scsi/scsi-disk.c | 15 +++++++++++++++ | ||
25 | hw/scsi/scsi-generic.c | 20 +++++++++++++++++--- | ||
26 | util/thread-pool.c | 4 +++- | ||
27 | 17 files changed, 72 insertions(+), 50 deletions(-) | ||
34 | 28 | ||
35 | diff --git a/util/throttle.c b/util/throttle.c | 29 | diff --git a/block/archipelago.c b/block/archipelago.c |
36 | index XXXXXXX..XXXXXXX 100644 | 30 | index XXXXXXX..XXXXXXX 100644 |
37 | --- a/util/throttle.c | 31 | --- a/block/archipelago.c |
38 | +++ b/util/throttle.c | 32 | +++ b/block/archipelago.c |
39 | @@ -XXX,XX +XXX,XX @@ static void throttle_fix_bucket(LeakyBucket *bkt) | 33 | @@ -XXX,XX +XXX,XX @@ static void qemu_archipelago_complete_aio(void *opaque) |
40 | /* zero bucket level */ | 34 | { |
41 | bkt->level = bkt->burst_level = 0; | 35 | AIORequestData *reqdata = (AIORequestData *) opaque; |
42 | 36 | ArchipelagoAIOCB *aio_cb = (ArchipelagoAIOCB *) reqdata->aio_cb; | |
43 | - /* The following is done to cope with the Linux CFQ block scheduler | 37 | - AioContext *ctx = bdrv_get_aio_context(aio_cb->common.bs); |
44 | - * which regroup reads and writes by block of 100ms in the guest. | 38 | |
45 | - * When they are two process one making reads and one making writes cfq | 39 | - aio_context_acquire(ctx); |
46 | - * make a pattern looking like the following: | 40 | aio_cb->common.cb(aio_cb->common.opaque, aio_cb->ret); |
47 | - * WWWWWWWWWWWRRRRRRRRRRRRRRWWWWWWWWWWWWWwRRRRRRRRRRRRRRRRR | 41 | - aio_context_release(ctx); |
48 | - * Having a max burst value of 100ms of the average will help smooth the | 42 | aio_cb->status = 0; |
49 | - * throttling | 43 | |
50 | - */ | 44 | qemu_aio_unref(aio_cb); |
51 | + /* If bkt->max is 0 we still want to allow short bursts of I/O | 45 | diff --git a/block/block-backend.c b/block/block-backend.c |
52 | + * from the guest, otherwise every other request will be throttled | 46 | index XXXXXXX..XXXXXXX 100644 |
53 | + * and performance will suffer considerably. */ | 47 | --- a/block/block-backend.c |
54 | min = bkt->avg / 10; | 48 | +++ b/block/block-backend.c |
55 | if (bkt->avg && !bkt->max) { | 49 | @@ -XXX,XX +XXX,XX @@ int blk_make_zero(BlockBackend *blk, BdrvRequestFlags flags) |
56 | bkt->max = min; | 50 | static void error_callback_bh(void *opaque) |
51 | { | ||
52 | struct BlockBackendAIOCB *acb = opaque; | ||
53 | - AioContext *ctx = bdrv_get_aio_context(acb->common.bs); | ||
54 | |||
55 | bdrv_dec_in_flight(acb->common.bs); | ||
56 | - aio_context_acquire(ctx); | ||
57 | acb->common.cb(acb->common.opaque, acb->ret); | ||
58 | - aio_context_release(ctx); | ||
59 | qemu_aio_unref(acb); | ||
60 | } | ||
61 | |||
62 | @@ -XXX,XX +XXX,XX @@ static void blk_aio_complete(BlkAioEmAIOCB *acb) | ||
63 | static void blk_aio_complete_bh(void *opaque) | ||
64 | { | ||
65 | BlkAioEmAIOCB *acb = opaque; | ||
66 | - AioContext *ctx = bdrv_get_aio_context(acb->common.bs); | ||
67 | - | ||
68 | assert(acb->has_returned); | ||
69 | - aio_context_acquire(ctx); | ||
70 | blk_aio_complete(acb); | ||
71 | - aio_context_release(ctx); | ||
72 | } | ||
73 | |||
74 | static BlockAIOCB *blk_aio_prwv(BlockBackend *blk, int64_t offset, int bytes, | ||
75 | diff --git a/block/curl.c b/block/curl.c | ||
76 | index XXXXXXX..XXXXXXX 100644 | ||
77 | --- a/block/curl.c | ||
78 | +++ b/block/curl.c | ||
79 | @@ -XXX,XX +XXX,XX @@ static void curl_readv_bh_cb(void *p) | ||
80 | curl_multi_socket_action(s->multi, CURL_SOCKET_TIMEOUT, 0, &running); | ||
81 | |||
82 | out: | ||
83 | + aio_context_release(ctx); | ||
84 | if (ret != -EINPROGRESS) { | ||
85 | acb->common.cb(acb->common.opaque, ret); | ||
86 | qemu_aio_unref(acb); | ||
87 | } | ||
88 | - aio_context_release(ctx); | ||
89 | } | ||
90 | |||
91 | static BlockAIOCB *curl_aio_readv(BlockDriverState *bs, | ||
92 | diff --git a/block/io.c b/block/io.c | ||
93 | index XXXXXXX..XXXXXXX 100644 | ||
94 | --- a/block/io.c | ||
95 | +++ b/block/io.c | ||
96 | @@ -XXX,XX +XXX,XX @@ static void bdrv_co_io_em_complete(void *opaque, int ret) | ||
97 | CoroutineIOCompletion *co = opaque; | ||
98 | |||
99 | co->ret = ret; | ||
100 | - qemu_coroutine_enter(co->coroutine); | ||
101 | + aio_co_wake(co->coroutine); | ||
102 | } | ||
103 | |||
104 | static int coroutine_fn bdrv_driver_preadv(BlockDriverState *bs, | ||
105 | @@ -XXX,XX +XXX,XX @@ static void bdrv_co_complete(BlockAIOCBCoroutine *acb) | ||
106 | static void bdrv_co_em_bh(void *opaque) | ||
107 | { | ||
108 | BlockAIOCBCoroutine *acb = opaque; | ||
109 | - BlockDriverState *bs = acb->common.bs; | ||
110 | - AioContext *ctx = bdrv_get_aio_context(bs); | ||
111 | |||
112 | assert(!acb->need_bh); | ||
113 | - aio_context_acquire(ctx); | ||
114 | bdrv_co_complete(acb); | ||
115 | - aio_context_release(ctx); | ||
116 | } | ||
117 | |||
118 | static void bdrv_co_maybe_schedule_bh(BlockAIOCBCoroutine *acb) | ||
119 | diff --git a/block/iscsi.c b/block/iscsi.c | ||
120 | index XXXXXXX..XXXXXXX 100644 | ||
121 | --- a/block/iscsi.c | ||
122 | +++ b/block/iscsi.c | ||
123 | @@ -XXX,XX +XXX,XX @@ static void | ||
124 | iscsi_bh_cb(void *p) | ||
125 | { | ||
126 | IscsiAIOCB *acb = p; | ||
127 | - AioContext *ctx = bdrv_get_aio_context(acb->common.bs); | ||
128 | |||
129 | qemu_bh_delete(acb->bh); | ||
130 | |||
131 | g_free(acb->buf); | ||
132 | acb->buf = NULL; | ||
133 | |||
134 | - aio_context_acquire(ctx); | ||
135 | acb->common.cb(acb->common.opaque, acb->status); | ||
136 | - aio_context_release(ctx); | ||
137 | |||
138 | if (acb->task != NULL) { | ||
139 | scsi_free_scsi_task(acb->task); | ||
140 | diff --git a/block/linux-aio.c b/block/linux-aio.c | ||
141 | index XXXXXXX..XXXXXXX 100644 | ||
142 | --- a/block/linux-aio.c | ||
143 | +++ b/block/linux-aio.c | ||
144 | @@ -XXX,XX +XXX,XX @@ static inline ssize_t io_event_ret(struct io_event *ev) | ||
145 | */ | ||
146 | static void qemu_laio_process_completion(struct qemu_laiocb *laiocb) | ||
147 | { | ||
148 | - LinuxAioState *s = laiocb->ctx; | ||
149 | int ret; | ||
150 | |||
151 | ret = laiocb->ret; | ||
152 | @@ -XXX,XX +XXX,XX @@ static void qemu_laio_process_completion(struct qemu_laiocb *laiocb) | ||
153 | } | ||
154 | |||
155 | laiocb->ret = ret; | ||
156 | - aio_context_acquire(s->aio_context); | ||
157 | if (laiocb->co) { | ||
158 | /* If the coroutine is already entered it must be in ioq_submit() and | ||
159 | * will notice laio->ret has been filled in when it eventually runs | ||
160 | @@ -XXX,XX +XXX,XX @@ static void qemu_laio_process_completion(struct qemu_laiocb *laiocb) | ||
161 | * that! | ||
162 | */ | ||
163 | if (!qemu_coroutine_entered(laiocb->co)) { | ||
164 | - qemu_coroutine_enter(laiocb->co); | ||
165 | + aio_co_wake(laiocb->co); | ||
166 | } | ||
167 | } else { | ||
168 | laiocb->common.cb(laiocb->common.opaque, ret); | ||
169 | qemu_aio_unref(laiocb); | ||
170 | } | ||
171 | - aio_context_release(s->aio_context); | ||
172 | } | ||
173 | |||
174 | /** | ||
175 | diff --git a/block/mirror.c b/block/mirror.c | ||
176 | index XXXXXXX..XXXXXXX 100644 | ||
177 | --- a/block/mirror.c | ||
178 | +++ b/block/mirror.c | ||
179 | @@ -XXX,XX +XXX,XX @@ static void mirror_write_complete(void *opaque, int ret) | ||
180 | { | ||
181 | MirrorOp *op = opaque; | ||
182 | MirrorBlockJob *s = op->s; | ||
183 | + | ||
184 | + aio_context_acquire(blk_get_aio_context(s->common.blk)); | ||
185 | if (ret < 0) { | ||
186 | BlockErrorAction action; | ||
187 | |||
188 | @@ -XXX,XX +XXX,XX @@ static void mirror_write_complete(void *opaque, int ret) | ||
189 | } | ||
190 | } | ||
191 | mirror_iteration_done(op, ret); | ||
192 | + aio_context_release(blk_get_aio_context(s->common.blk)); | ||
193 | } | ||
194 | |||
195 | static void mirror_read_complete(void *opaque, int ret) | ||
196 | { | ||
197 | MirrorOp *op = opaque; | ||
198 | MirrorBlockJob *s = op->s; | ||
199 | + | ||
200 | + aio_context_acquire(blk_get_aio_context(s->common.blk)); | ||
201 | if (ret < 0) { | ||
202 | BlockErrorAction action; | ||
203 | |||
204 | @@ -XXX,XX +XXX,XX @@ static void mirror_read_complete(void *opaque, int ret) | ||
205 | } | ||
206 | |||
207 | mirror_iteration_done(op, ret); | ||
208 | - return; | ||
209 | + } else { | ||
210 | + blk_aio_pwritev(s->target, op->sector_num * BDRV_SECTOR_SIZE, &op->qiov, | ||
211 | + 0, mirror_write_complete, op); | ||
212 | } | ||
213 | - blk_aio_pwritev(s->target, op->sector_num * BDRV_SECTOR_SIZE, &op->qiov, | ||
214 | - 0, mirror_write_complete, op); | ||
215 | + aio_context_release(blk_get_aio_context(s->common.blk)); | ||
216 | } | ||
217 | |||
218 | static inline void mirror_clip_sectors(MirrorBlockJob *s, | ||
219 | diff --git a/block/null.c b/block/null.c | ||
220 | index XXXXXXX..XXXXXXX 100644 | ||
221 | --- a/block/null.c | ||
222 | +++ b/block/null.c | ||
223 | @@ -XXX,XX +XXX,XX @@ static const AIOCBInfo null_aiocb_info = { | ||
224 | static void null_bh_cb(void *opaque) | ||
225 | { | ||
226 | NullAIOCB *acb = opaque; | ||
227 | - AioContext *ctx = bdrv_get_aio_context(acb->common.bs); | ||
228 | - | ||
229 | - aio_context_acquire(ctx); | ||
230 | acb->common.cb(acb->common.opaque, 0); | ||
231 | - aio_context_release(ctx); | ||
232 | qemu_aio_unref(acb); | ||
233 | } | ||
234 | |||
235 | static void null_timer_cb(void *opaque) | ||
236 | { | ||
237 | NullAIOCB *acb = opaque; | ||
238 | - AioContext *ctx = bdrv_get_aio_context(acb->common.bs); | ||
239 | - | ||
240 | - aio_context_acquire(ctx); | ||
241 | acb->common.cb(acb->common.opaque, 0); | ||
242 | - aio_context_release(ctx); | ||
243 | timer_deinit(&acb->timer); | ||
244 | qemu_aio_unref(acb); | ||
245 | } | ||
246 | diff --git a/block/qed-cluster.c b/block/qed-cluster.c | ||
247 | index XXXXXXX..XXXXXXX 100644 | ||
248 | --- a/block/qed-cluster.c | ||
249 | +++ b/block/qed-cluster.c | ||
250 | @@ -XXX,XX +XXX,XX @@ static void qed_find_cluster_cb(void *opaque, int ret) | ||
251 | unsigned int index; | ||
252 | unsigned int n; | ||
253 | |||
254 | + qed_acquire(s); | ||
255 | if (ret) { | ||
256 | goto out; | ||
257 | } | ||
258 | @@ -XXX,XX +XXX,XX @@ static void qed_find_cluster_cb(void *opaque, int ret) | ||
259 | |||
260 | out: | ||
261 | find_cluster_cb->cb(find_cluster_cb->opaque, ret, offset, len); | ||
262 | + qed_release(s); | ||
263 | g_free(find_cluster_cb); | ||
264 | } | ||
265 | |||
266 | diff --git a/block/qed-table.c b/block/qed-table.c | ||
267 | index XXXXXXX..XXXXXXX 100644 | ||
268 | --- a/block/qed-table.c | ||
269 | +++ b/block/qed-table.c | ||
270 | @@ -XXX,XX +XXX,XX @@ static void qed_read_table_cb(void *opaque, int ret) | ||
271 | { | ||
272 | QEDReadTableCB *read_table_cb = opaque; | ||
273 | QEDTable *table = read_table_cb->table; | ||
274 | + BDRVQEDState *s = read_table_cb->s; | ||
275 | int noffsets = read_table_cb->qiov.size / sizeof(uint64_t); | ||
276 | int i; | ||
277 | |||
278 | @@ -XXX,XX +XXX,XX @@ static void qed_read_table_cb(void *opaque, int ret) | ||
279 | } | ||
280 | |||
281 | /* Byteswap offsets */ | ||
282 | + qed_acquire(s); | ||
283 | for (i = 0; i < noffsets; i++) { | ||
284 | table->offsets[i] = le64_to_cpu(table->offsets[i]); | ||
285 | } | ||
286 | + qed_release(s); | ||
287 | |||
288 | out: | ||
289 | /* Completion */ | ||
290 | - trace_qed_read_table_cb(read_table_cb->s, read_table_cb->table, ret); | ||
291 | + trace_qed_read_table_cb(s, read_table_cb->table, ret); | ||
292 | gencb_complete(&read_table_cb->gencb, ret); | ||
293 | } | ||
294 | |||
295 | @@ -XXX,XX +XXX,XX @@ typedef struct { | ||
296 | static void qed_write_table_cb(void *opaque, int ret) | ||
297 | { | ||
298 | QEDWriteTableCB *write_table_cb = opaque; | ||
299 | + BDRVQEDState *s = write_table_cb->s; | ||
300 | |||
301 | - trace_qed_write_table_cb(write_table_cb->s, | ||
302 | + trace_qed_write_table_cb(s, | ||
303 | write_table_cb->orig_table, | ||
304 | write_table_cb->flush, | ||
305 | ret); | ||
306 | @@ -XXX,XX +XXX,XX @@ static void qed_write_table_cb(void *opaque, int ret) | ||
307 | if (write_table_cb->flush) { | ||
308 | /* We still need to flush first */ | ||
309 | write_table_cb->flush = false; | ||
310 | + qed_acquire(s); | ||
311 | bdrv_aio_flush(write_table_cb->s->bs, qed_write_table_cb, | ||
312 | write_table_cb); | ||
313 | + qed_release(s); | ||
314 | return; | ||
315 | } | ||
316 | |||
317 | @@ -XXX,XX +XXX,XX @@ static void qed_read_l2_table_cb(void *opaque, int ret) | ||
318 | CachedL2Table *l2_table = request->l2_table; | ||
319 | uint64_t l2_offset = read_l2_table_cb->l2_offset; | ||
320 | |||
321 | + qed_acquire(s); | ||
322 | if (ret) { | ||
323 | /* can't trust loaded L2 table anymore */ | ||
324 | qed_unref_l2_cache_entry(l2_table); | ||
325 | @@ -XXX,XX +XXX,XX @@ static void qed_read_l2_table_cb(void *opaque, int ret) | ||
326 | request->l2_table = qed_find_l2_cache_entry(&s->l2_cache, l2_offset); | ||
327 | assert(request->l2_table != NULL); | ||
328 | } | ||
329 | + qed_release(s); | ||
330 | |||
331 | gencb_complete(&read_l2_table_cb->gencb, ret); | ||
332 | } | ||
333 | diff --git a/block/qed.c b/block/qed.c | ||
334 | index XXXXXXX..XXXXXXX 100644 | ||
335 | --- a/block/qed.c | ||
336 | +++ b/block/qed.c | ||
337 | @@ -XXX,XX +XXX,XX @@ static void qed_is_allocated_cb(void *opaque, int ret, uint64_t offset, size_t l | ||
338 | } | ||
339 | |||
340 | if (cb->co) { | ||
341 | - qemu_coroutine_enter(cb->co); | ||
342 | + aio_co_wake(cb->co); | ||
343 | } | ||
344 | } | ||
345 | |||
346 | @@ -XXX,XX +XXX,XX @@ static void coroutine_fn qed_co_pwrite_zeroes_cb(void *opaque, int ret) | ||
347 | cb->done = true; | ||
348 | cb->ret = ret; | ||
349 | if (cb->co) { | ||
350 | - qemu_coroutine_enter(cb->co); | ||
351 | + aio_co_wake(cb->co); | ||
352 | } | ||
353 | } | ||
354 | |||
355 | diff --git a/block/rbd.c b/block/rbd.c | ||
356 | index XXXXXXX..XXXXXXX 100644 | ||
357 | --- a/block/rbd.c | ||
358 | +++ b/block/rbd.c | ||
359 | @@ -XXX,XX +XXX,XX @@ shutdown: | ||
360 | static void qemu_rbd_complete_aio(RADOSCB *rcb) | ||
361 | { | ||
362 | RBDAIOCB *acb = rcb->acb; | ||
363 | - AioContext *ctx = bdrv_get_aio_context(acb->common.bs); | ||
364 | int64_t r; | ||
365 | |||
366 | r = rcb->ret; | ||
367 | @@ -XXX,XX +XXX,XX @@ static void qemu_rbd_complete_aio(RADOSCB *rcb) | ||
368 | qemu_iovec_from_buf(acb->qiov, 0, acb->bounce, acb->qiov->size); | ||
369 | } | ||
370 | qemu_vfree(acb->bounce); | ||
371 | - | ||
372 | - aio_context_acquire(ctx); | ||
373 | acb->common.cb(acb->common.opaque, (acb->ret > 0 ? 0 : acb->ret)); | ||
374 | - aio_context_release(ctx); | ||
375 | |||
376 | qemu_aio_unref(acb); | ||
377 | } | ||
378 | diff --git a/block/win32-aio.c b/block/win32-aio.c | ||
379 | index XXXXXXX..XXXXXXX 100644 | ||
380 | --- a/block/win32-aio.c | ||
381 | +++ b/block/win32-aio.c | ||
382 | @@ -XXX,XX +XXX,XX @@ static void win32_aio_process_completion(QEMUWin32AIOState *s, | ||
383 | qemu_vfree(waiocb->buf); | ||
384 | } | ||
385 | |||
386 | - | ||
387 | - aio_context_acquire(s->aio_ctx); | ||
388 | waiocb->common.cb(waiocb->common.opaque, ret); | ||
389 | - aio_context_release(s->aio_ctx); | ||
390 | qemu_aio_unref(waiocb); | ||
391 | } | ||
392 | |||
393 | diff --git a/hw/block/virtio-blk.c b/hw/block/virtio-blk.c | ||
394 | index XXXXXXX..XXXXXXX 100644 | ||
395 | --- a/hw/block/virtio-blk.c | ||
396 | +++ b/hw/block/virtio-blk.c | ||
397 | @@ -XXX,XX +XXX,XX @@ static int virtio_blk_handle_rw_error(VirtIOBlockReq *req, int error, | ||
398 | static void virtio_blk_rw_complete(void *opaque, int ret) | ||
399 | { | ||
400 | VirtIOBlockReq *next = opaque; | ||
401 | + VirtIOBlock *s = next->dev; | ||
402 | |||
403 | + aio_context_acquire(blk_get_aio_context(s->conf.conf.blk)); | ||
404 | while (next) { | ||
405 | VirtIOBlockReq *req = next; | ||
406 | next = req->mr_next; | ||
407 | @@ -XXX,XX +XXX,XX @@ static void virtio_blk_rw_complete(void *opaque, int ret) | ||
408 | block_acct_done(blk_get_stats(req->dev->blk), &req->acct); | ||
409 | virtio_blk_free_request(req); | ||
410 | } | ||
411 | + aio_context_release(blk_get_aio_context(s->conf.conf.blk)); | ||
412 | } | ||
413 | |||
414 | static void virtio_blk_flush_complete(void *opaque, int ret) | ||
415 | { | ||
416 | VirtIOBlockReq *req = opaque; | ||
417 | + VirtIOBlock *s = req->dev; | ||
418 | |||
419 | + aio_context_acquire(blk_get_aio_context(s->conf.conf.blk)); | ||
420 | if (ret) { | ||
421 | if (virtio_blk_handle_rw_error(req, -ret, 0)) { | ||
422 | - return; | ||
423 | + goto out; | ||
424 | } | ||
425 | } | ||
426 | |||
427 | virtio_blk_req_complete(req, VIRTIO_BLK_S_OK); | ||
428 | block_acct_done(blk_get_stats(req->dev->blk), &req->acct); | ||
429 | virtio_blk_free_request(req); | ||
430 | + | ||
431 | +out: | ||
432 | + aio_context_release(blk_get_aio_context(s->conf.conf.blk)); | ||
433 | } | ||
434 | |||
435 | #ifdef __linux__ | ||
436 | @@ -XXX,XX +XXX,XX @@ static void virtio_blk_ioctl_complete(void *opaque, int status) | ||
437 | virtio_stl_p(vdev, &scsi->data_len, hdr->dxfer_len); | ||
438 | |||
439 | out: | ||
440 | + aio_context_acquire(blk_get_aio_context(s->conf.conf.blk)); | ||
441 | virtio_blk_req_complete(req, status); | ||
442 | virtio_blk_free_request(req); | ||
443 | + aio_context_release(blk_get_aio_context(s->conf.conf.blk)); | ||
444 | g_free(ioctl_req); | ||
445 | } | ||
446 | |||
447 | diff --git a/hw/scsi/scsi-disk.c b/hw/scsi/scsi-disk.c | ||
448 | index XXXXXXX..XXXXXXX 100644 | ||
449 | --- a/hw/scsi/scsi-disk.c | ||
450 | +++ b/hw/scsi/scsi-disk.c | ||
451 | @@ -XXX,XX +XXX,XX @@ static void scsi_aio_complete(void *opaque, int ret) | ||
452 | |||
453 | assert(r->req.aiocb != NULL); | ||
454 | r->req.aiocb = NULL; | ||
455 | + aio_context_acquire(blk_get_aio_context(s->qdev.conf.blk)); | ||
456 | if (scsi_disk_req_check_error(r, ret, true)) { | ||
457 | goto done; | ||
458 | } | ||
459 | @@ -XXX,XX +XXX,XX @@ static void scsi_aio_complete(void *opaque, int ret) | ||
460 | scsi_req_complete(&r->req, GOOD); | ||
461 | |||
462 | done: | ||
463 | + aio_context_release(blk_get_aio_context(s->qdev.conf.blk)); | ||
464 | scsi_req_unref(&r->req); | ||
465 | } | ||
466 | |||
467 | @@ -XXX,XX +XXX,XX @@ static void scsi_dma_complete(void *opaque, int ret) | ||
468 | assert(r->req.aiocb != NULL); | ||
469 | r->req.aiocb = NULL; | ||
470 | |||
471 | + aio_context_acquire(blk_get_aio_context(s->qdev.conf.blk)); | ||
472 | if (ret < 0) { | ||
473 | block_acct_failed(blk_get_stats(s->qdev.conf.blk), &r->acct); | ||
474 | } else { | ||
475 | block_acct_done(blk_get_stats(s->qdev.conf.blk), &r->acct); | ||
476 | } | ||
477 | scsi_dma_complete_noio(r, ret); | ||
478 | + aio_context_release(blk_get_aio_context(s->qdev.conf.blk)); | ||
479 | } | ||
480 | |||
481 | static void scsi_read_complete(void * opaque, int ret) | ||
482 | @@ -XXX,XX +XXX,XX @@ static void scsi_read_complete(void * opaque, int ret) | ||
483 | |||
484 | assert(r->req.aiocb != NULL); | ||
485 | r->req.aiocb = NULL; | ||
486 | + aio_context_acquire(blk_get_aio_context(s->qdev.conf.blk)); | ||
487 | if (scsi_disk_req_check_error(r, ret, true)) { | ||
488 | goto done; | ||
489 | } | ||
490 | @@ -XXX,XX +XXX,XX @@ static void scsi_read_complete(void * opaque, int ret) | ||
491 | |||
492 | done: | ||
493 | scsi_req_unref(&r->req); | ||
494 | + aio_context_release(blk_get_aio_context(s->qdev.conf.blk)); | ||
495 | } | ||
496 | |||
497 | /* Actually issue a read to the block device. */ | ||
498 | @@ -XXX,XX +XXX,XX @@ static void scsi_do_read_cb(void *opaque, int ret) | ||
499 | assert (r->req.aiocb != NULL); | ||
500 | r->req.aiocb = NULL; | ||
501 | |||
502 | + aio_context_acquire(blk_get_aio_context(s->qdev.conf.blk)); | ||
503 | if (ret < 0) { | ||
504 | block_acct_failed(blk_get_stats(s->qdev.conf.blk), &r->acct); | ||
505 | } else { | ||
506 | block_acct_done(blk_get_stats(s->qdev.conf.blk), &r->acct); | ||
507 | } | ||
508 | scsi_do_read(opaque, ret); | ||
509 | + aio_context_release(blk_get_aio_context(s->qdev.conf.blk)); | ||
510 | } | ||
511 | |||
512 | /* Read more data from scsi device into buffer. */ | ||
513 | @@ -XXX,XX +XXX,XX @@ static void scsi_write_complete(void * opaque, int ret) | ||
514 | assert (r->req.aiocb != NULL); | ||
515 | r->req.aiocb = NULL; | ||
516 | |||
517 | + aio_context_acquire(blk_get_aio_context(s->qdev.conf.blk)); | ||
518 | if (ret < 0) { | ||
519 | block_acct_failed(blk_get_stats(s->qdev.conf.blk), &r->acct); | ||
520 | } else { | ||
521 | block_acct_done(blk_get_stats(s->qdev.conf.blk), &r->acct); | ||
522 | } | ||
523 | scsi_write_complete_noio(r, ret); | ||
524 | + aio_context_release(blk_get_aio_context(s->qdev.conf.blk)); | ||
525 | } | ||
526 | |||
527 | static void scsi_write_data(SCSIRequest *req) | ||
528 | @@ -XXX,XX +XXX,XX @@ static void scsi_unmap_complete(void *opaque, int ret) | ||
529 | { | ||
530 | UnmapCBData *data = opaque; | ||
531 | SCSIDiskReq *r = data->r; | ||
532 | + SCSIDiskState *s = DO_UPCAST(SCSIDiskState, qdev, r->req.dev); | ||
533 | |||
534 | assert(r->req.aiocb != NULL); | ||
535 | r->req.aiocb = NULL; | ||
536 | |||
537 | + aio_context_acquire(blk_get_aio_context(s->qdev.conf.blk)); | ||
538 | scsi_unmap_complete_noio(data, ret); | ||
539 | + aio_context_release(blk_get_aio_context(s->qdev.conf.blk)); | ||
540 | } | ||
541 | |||
542 | static void scsi_disk_emulate_unmap(SCSIDiskReq *r, uint8_t *inbuf) | ||
543 | @@ -XXX,XX +XXX,XX @@ static void scsi_write_same_complete(void *opaque, int ret) | ||
544 | |||
545 | assert(r->req.aiocb != NULL); | ||
546 | r->req.aiocb = NULL; | ||
547 | + aio_context_acquire(blk_get_aio_context(s->qdev.conf.blk)); | ||
548 | if (scsi_disk_req_check_error(r, ret, true)) { | ||
549 | goto done; | ||
550 | } | ||
551 | @@ -XXX,XX +XXX,XX @@ done: | ||
552 | scsi_req_unref(&r->req); | ||
553 | qemu_vfree(data->iov.iov_base); | ||
554 | g_free(data); | ||
555 | + aio_context_release(blk_get_aio_context(s->qdev.conf.blk)); | ||
556 | } | ||
557 | |||
558 | static void scsi_disk_emulate_write_same(SCSIDiskReq *r, uint8_t *inbuf) | ||
559 | diff --git a/hw/scsi/scsi-generic.c b/hw/scsi/scsi-generic.c | ||
560 | index XXXXXXX..XXXXXXX 100644 | ||
561 | --- a/hw/scsi/scsi-generic.c | ||
562 | +++ b/hw/scsi/scsi-generic.c | ||
563 | @@ -XXX,XX +XXX,XX @@ done: | ||
564 | static void scsi_command_complete(void *opaque, int ret) | ||
565 | { | ||
566 | SCSIGenericReq *r = (SCSIGenericReq *)opaque; | ||
567 | + SCSIDevice *s = r->req.dev; | ||
568 | |||
569 | assert(r->req.aiocb != NULL); | ||
570 | r->req.aiocb = NULL; | ||
571 | + | ||
572 | + aio_context_acquire(blk_get_aio_context(s->conf.blk)); | ||
573 | scsi_command_complete_noio(r, ret); | ||
574 | + aio_context_release(blk_get_aio_context(s->conf.blk)); | ||
575 | } | ||
576 | |||
577 | static int execute_command(BlockBackend *blk, | ||
578 | @@ -XXX,XX +XXX,XX @@ static void scsi_read_complete(void * opaque, int ret) | ||
579 | assert(r->req.aiocb != NULL); | ||
580 | r->req.aiocb = NULL; | ||
581 | |||
582 | + aio_context_acquire(blk_get_aio_context(s->conf.blk)); | ||
583 | + | ||
584 | if (ret || r->req.io_canceled) { | ||
585 | scsi_command_complete_noio(r, ret); | ||
586 | - return; | ||
587 | + goto done; | ||
588 | } | ||
589 | |||
590 | len = r->io_header.dxfer_len - r->io_header.resid; | ||
591 | @@ -XXX,XX +XXX,XX @@ static void scsi_read_complete(void * opaque, int ret) | ||
592 | r->len = -1; | ||
593 | if (len == 0) { | ||
594 | scsi_command_complete_noio(r, 0); | ||
595 | - return; | ||
596 | + goto done; | ||
597 | } | ||
598 | |||
599 | /* Snoop READ CAPACITY output to set the blocksize. */ | ||
600 | @@ -XXX,XX +XXX,XX @@ static void scsi_read_complete(void * opaque, int ret) | ||
601 | } | ||
602 | scsi_req_data(&r->req, len); | ||
603 | scsi_req_unref(&r->req); | ||
604 | + | ||
605 | +done: | ||
606 | + aio_context_release(blk_get_aio_context(s->conf.blk)); | ||
607 | } | ||
608 | |||
609 | /* Read more data from scsi device into buffer. */ | ||
610 | @@ -XXX,XX +XXX,XX @@ static void scsi_write_complete(void * opaque, int ret) | ||
611 | assert(r->req.aiocb != NULL); | ||
612 | r->req.aiocb = NULL; | ||
613 | |||
614 | + aio_context_acquire(blk_get_aio_context(s->conf.blk)); | ||
615 | + | ||
616 | if (ret || r->req.io_canceled) { | ||
617 | scsi_command_complete_noio(r, ret); | ||
618 | - return; | ||
619 | + goto done; | ||
620 | } | ||
621 | |||
622 | if (r->req.cmd.buf[0] == MODE_SELECT && r->req.cmd.buf[4] == 12 && | ||
623 | @@ -XXX,XX +XXX,XX @@ static void scsi_write_complete(void * opaque, int ret) | ||
624 | } | ||
625 | |||
626 | scsi_command_complete_noio(r, ret); | ||
627 | + | ||
628 | +done: | ||
629 | + aio_context_release(blk_get_aio_context(s->conf.blk)); | ||
630 | } | ||
631 | |||
632 | /* Write data to a scsi device. Returns nonzero on failure. | ||
633 | diff --git a/util/thread-pool.c b/util/thread-pool.c | ||
634 | index XXXXXXX..XXXXXXX 100644 | ||
635 | --- a/util/thread-pool.c | ||
636 | +++ b/util/thread-pool.c | ||
637 | @@ -XXX,XX +XXX,XX @@ restart: | ||
638 | */ | ||
639 | qemu_bh_schedule(pool->completion_bh); | ||
640 | |||
641 | + aio_context_release(pool->ctx); | ||
642 | elem->common.cb(elem->common.opaque, elem->ret); | ||
643 | + aio_context_acquire(pool->ctx); | ||
644 | qemu_aio_unref(elem); | ||
645 | goto restart; | ||
646 | } else { | ||
647 | @@ -XXX,XX +XXX,XX @@ static void thread_pool_co_cb(void *opaque, int ret) | ||
648 | ThreadPoolCo *co = opaque; | ||
649 | |||
650 | co->ret = ret; | ||
651 | - qemu_coroutine_enter(co->co); | ||
652 | + aio_co_wake(co->co); | ||
653 | } | ||
654 | |||
655 | int coroutine_fn thread_pool_submit_co(ThreadPool *pool, ThreadPoolFunc *func, | ||
57 | -- | 656 | -- |
58 | 2.13.5 | 657 | 2.9.3 |
59 | 658 | ||
60 | 659 | diff view generated by jsdifflib |
1 | From: Alberto Garcia <berto@igalia.com> | 1 | From: Paolo Bonzini <pbonzini@redhat.com> |
---|---|---|---|
2 | 2 | ||
3 | There's a few cases which we're passing an Error pointer to a function | 3 | This patch prepares for the removal of unnecessary lockcnt inc/dec pairs. |
4 | only to discard it immediately afterwards without checking it. In | 4 | Extract the dispatching loop for file descriptor handlers into a new |
5 | these cases we can simply remove the variable and pass NULL instead. | 5 | function aio_dispatch_handlers, and then inline aio_dispatch into |
6 | aio_poll. | ||
6 | 7 | ||
7 | Signed-off-by: Alberto Garcia <berto@igalia.com> | 8 | aio_dispatch can now become void. |
8 | Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org> | 9 | |
9 | Reviewed-by: Eric Blake <eblake@redhat.com> | 10 | Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> |
10 | Message-id: 20170829120836.16091-1-berto@igalia.com | 11 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> |
12 | Reviewed-by: Fam Zheng <famz@redhat.com> | ||
13 | Reviewed-by: Daniel P. Berrange <berrange@redhat.com> | ||
14 | Message-id: 20170213135235.12274-17-pbonzini@redhat.com | ||
11 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | 15 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
12 | --- | 16 | --- |
13 | block/qcow.c | 12 +++--------- | 17 | include/block/aio.h | 6 +----- |
14 | block/qcow2.c | 8 ++------ | 18 | util/aio-posix.c | 44 ++++++++++++++------------------------------ |
15 | dump.c | 4 +--- | 19 | util/aio-win32.c | 13 ++++--------- |
16 | 3 files changed, 6 insertions(+), 18 deletions(-) | 20 | util/async.c | 2 +- |
21 | 4 files changed, 20 insertions(+), 45 deletions(-) | ||
17 | 22 | ||
18 | diff --git a/block/qcow.c b/block/qcow.c | 23 | diff --git a/include/block/aio.h b/include/block/aio.h |
19 | index XXXXXXX..XXXXXXX 100644 | 24 | index XXXXXXX..XXXXXXX 100644 |
20 | --- a/block/qcow.c | 25 | --- a/include/block/aio.h |
21 | +++ b/block/qcow.c | 26 | +++ b/include/block/aio.h |
22 | @@ -XXX,XX +XXX,XX @@ static uint64_t get_cluster_offset(BlockDriverState *bs, | 27 | @@ -XXX,XX +XXX,XX @@ bool aio_pending(AioContext *ctx); |
23 | start_sect = (offset & ~(s->cluster_size - 1)) >> 9; | 28 | /* Dispatch any pending callbacks from the GSource attached to the AioContext. |
24 | for(i = 0; i < s->cluster_sectors; i++) { | 29 | * |
25 | if (i < n_start || i >= n_end) { | 30 | * This is used internally in the implementation of the GSource. |
26 | - Error *err = NULL; | 31 | - * |
27 | memset(s->cluster_data, 0x00, 512); | 32 | - * @dispatch_fds: true to process fds, false to skip them |
28 | if (qcrypto_block_encrypt(s->crypto, start_sect + i, | 33 | - * (can be used as an optimization by callers that know there |
29 | s->cluster_data, | 34 | - * are no fds ready) |
30 | BDRV_SECTOR_SIZE, | 35 | */ |
31 | - &err) < 0) { | 36 | -bool aio_dispatch(AioContext *ctx, bool dispatch_fds); |
32 | - error_free(err); | 37 | +void aio_dispatch(AioContext *ctx); |
33 | + NULL) < 0) { | 38 | |
34 | errno = EIO; | 39 | /* Progress in completing AIO work to occur. This can issue new pending |
35 | return -1; | 40 | * aio as a result of executing I/O completion or bh callbacks. |
36 | } | 41 | diff --git a/util/aio-posix.c b/util/aio-posix.c |
37 | @@ -XXX,XX +XXX,XX @@ static coroutine_fn int qcow_co_readv(BlockDriverState *bs, int64_t sector_num, | 42 | index XXXXXXX..XXXXXXX 100644 |
38 | QEMUIOVector hd_qiov; | 43 | --- a/util/aio-posix.c |
39 | uint8_t *buf; | 44 | +++ b/util/aio-posix.c |
40 | void *orig_buf; | 45 | @@ -XXX,XX +XXX,XX @@ static bool aio_dispatch_handlers(AioContext *ctx) |
41 | - Error *err = NULL; | 46 | AioHandler *node, *tmp; |
42 | 47 | bool progress = false; | |
43 | if (qiov->niov > 1) { | 48 | |
44 | buf = orig_buf = qemu_try_blockalign(bs, qiov->size); | 49 | - /* |
45 | @@ -XXX,XX +XXX,XX @@ static coroutine_fn int qcow_co_readv(BlockDriverState *bs, int64_t sector_num, | 50 | - * We have to walk very carefully in case aio_set_fd_handler is |
46 | if (bs->encrypted) { | 51 | - * called while we're walking. |
47 | assert(s->crypto); | 52 | - */ |
48 | if (qcrypto_block_decrypt(s->crypto, sector_num, buf, | 53 | - qemu_lockcnt_inc(&ctx->list_lock); |
49 | - n * BDRV_SECTOR_SIZE, &err) < 0) { | 54 | - |
50 | + n * BDRV_SECTOR_SIZE, NULL) < 0) { | 55 | QLIST_FOREACH_SAFE_RCU(node, &ctx->aio_handlers, node, tmp) { |
51 | goto fail; | 56 | int revents; |
52 | } | 57 | |
53 | } | 58 | @@ -XXX,XX +XXX,XX @@ static bool aio_dispatch_handlers(AioContext *ctx) |
54 | @@ -XXX,XX +XXX,XX @@ done: | 59 | } |
55 | return ret; | 60 | } |
56 | 61 | ||
57 | fail: | 62 | - qemu_lockcnt_dec(&ctx->list_lock); |
58 | - error_free(err); | 63 | return progress; |
59 | ret = -EIO; | ||
60 | goto done; | ||
61 | } | 64 | } |
62 | @@ -XXX,XX +XXX,XX @@ static coroutine_fn int qcow_co_writev(BlockDriverState *bs, int64_t sector_num, | 65 | |
63 | break; | 66 | -/* |
64 | } | 67 | - * Note that dispatch_fds == false has the side-effect of post-poning the |
65 | if (bs->encrypted) { | 68 | - * freeing of deleted handlers. |
66 | - Error *err = NULL; | 69 | - */ |
67 | assert(s->crypto); | 70 | -bool aio_dispatch(AioContext *ctx, bool dispatch_fds) |
68 | if (qcrypto_block_encrypt(s->crypto, sector_num, buf, | 71 | +void aio_dispatch(AioContext *ctx) |
69 | - n * BDRV_SECTOR_SIZE, &err) < 0) { | 72 | { |
70 | - error_free(err); | 73 | - bool progress; |
71 | + n * BDRV_SECTOR_SIZE, NULL) < 0) { | 74 | + aio_bh_poll(ctx); |
72 | ret = -EIO; | 75 | |
73 | break; | 76 | - /* |
74 | } | 77 | - * If there are callbacks left that have been queued, we need to call them. |
75 | diff --git a/block/qcow2.c b/block/qcow2.c | 78 | - * Do not call select in this case, because it is possible that the caller |
79 | - * does not need a complete flush (as is the case for aio_poll loops). | ||
80 | - */ | ||
81 | - progress = aio_bh_poll(ctx); | ||
82 | + qemu_lockcnt_inc(&ctx->list_lock); | ||
83 | + aio_dispatch_handlers(ctx); | ||
84 | + qemu_lockcnt_dec(&ctx->list_lock); | ||
85 | |||
86 | - if (dispatch_fds) { | ||
87 | - progress |= aio_dispatch_handlers(ctx); | ||
88 | - } | ||
89 | - | ||
90 | - /* Run our timers */ | ||
91 | - progress |= timerlistgroup_run_timers(&ctx->tlg); | ||
92 | - | ||
93 | - return progress; | ||
94 | + timerlistgroup_run_timers(&ctx->tlg); | ||
95 | } | ||
96 | |||
97 | /* These thread-local variables are used only in a small part of aio_poll | ||
98 | @@ -XXX,XX +XXX,XX @@ bool aio_poll(AioContext *ctx, bool blocking) | ||
99 | npfd = 0; | ||
100 | qemu_lockcnt_dec(&ctx->list_lock); | ||
101 | |||
102 | - /* Run dispatch even if there were no readable fds to run timers */ | ||
103 | - if (aio_dispatch(ctx, ret > 0)) { | ||
104 | - progress = true; | ||
105 | + progress |= aio_bh_poll(ctx); | ||
106 | + | ||
107 | + if (ret > 0) { | ||
108 | + qemu_lockcnt_inc(&ctx->list_lock); | ||
109 | + progress |= aio_dispatch_handlers(ctx); | ||
110 | + qemu_lockcnt_dec(&ctx->list_lock); | ||
111 | } | ||
112 | |||
113 | + progress |= timerlistgroup_run_timers(&ctx->tlg); | ||
114 | + | ||
115 | return progress; | ||
116 | } | ||
117 | |||
118 | diff --git a/util/aio-win32.c b/util/aio-win32.c | ||
76 | index XXXXXXX..XXXXXXX 100644 | 119 | index XXXXXXX..XXXXXXX 100644 |
77 | --- a/block/qcow2.c | 120 | --- a/util/aio-win32.c |
78 | +++ b/block/qcow2.c | 121 | +++ b/util/aio-win32.c |
79 | @@ -XXX,XX +XXX,XX @@ static coroutine_fn int qcow2_co_preadv(BlockDriverState *bs, uint64_t offset, | 122 | @@ -XXX,XX +XXX,XX @@ static bool aio_dispatch_handlers(AioContext *ctx, HANDLE event) |
80 | assert(s->crypto); | 123 | return progress; |
81 | assert((offset & (BDRV_SECTOR_SIZE - 1)) == 0); | 124 | } |
82 | assert((cur_bytes & (BDRV_SECTOR_SIZE - 1)) == 0); | 125 | |
83 | - Error *err = NULL; | 126 | -bool aio_dispatch(AioContext *ctx, bool dispatch_fds) |
84 | if (qcrypto_block_decrypt(s->crypto, | 127 | +void aio_dispatch(AioContext *ctx) |
85 | (s->crypt_physical_offset ? | 128 | { |
86 | cluster_offset + offset_in_cluster : | 129 | - bool progress; |
87 | offset) >> BDRV_SECTOR_BITS, | 130 | - |
88 | cluster_data, | 131 | - progress = aio_bh_poll(ctx); |
89 | cur_bytes, | 132 | - if (dispatch_fds) { |
90 | - &err) < 0) { | 133 | - progress |= aio_dispatch_handlers(ctx, INVALID_HANDLE_VALUE); |
91 | - error_free(err); | 134 | - } |
92 | + NULL) < 0) { | 135 | - progress |= timerlistgroup_run_timers(&ctx->tlg); |
93 | ret = -EIO; | 136 | - return progress; |
94 | goto fail; | 137 | + aio_bh_poll(ctx); |
95 | } | 138 | + aio_dispatch_handlers(ctx, INVALID_HANDLE_VALUE); |
96 | @@ -XXX,XX +XXX,XX @@ static coroutine_fn int qcow2_co_pwritev(BlockDriverState *bs, uint64_t offset, | 139 | + timerlistgroup_run_timers(&ctx->tlg); |
97 | qemu_iovec_concat(&hd_qiov, qiov, bytes_done, cur_bytes); | 140 | } |
98 | 141 | ||
99 | if (bs->encrypted) { | 142 | bool aio_poll(AioContext *ctx, bool blocking) |
100 | - Error *err = NULL; | 143 | diff --git a/util/async.c b/util/async.c |
101 | assert(s->crypto); | ||
102 | if (!cluster_data) { | ||
103 | cluster_data = qemu_try_blockalign(bs->file->bs, | ||
104 | @@ -XXX,XX +XXX,XX @@ static coroutine_fn int qcow2_co_pwritev(BlockDriverState *bs, uint64_t offset, | ||
105 | cluster_offset + offset_in_cluster : | ||
106 | offset) >> BDRV_SECTOR_BITS, | ||
107 | cluster_data, | ||
108 | - cur_bytes, &err) < 0) { | ||
109 | - error_free(err); | ||
110 | + cur_bytes, NULL) < 0) { | ||
111 | ret = -EIO; | ||
112 | goto fail; | ||
113 | } | ||
114 | diff --git a/dump.c b/dump.c | ||
115 | index XXXXXXX..XXXXXXX 100644 | 144 | index XXXXXXX..XXXXXXX 100644 |
116 | --- a/dump.c | 145 | --- a/util/async.c |
117 | +++ b/dump.c | 146 | +++ b/util/async.c |
118 | @@ -XXX,XX +XXX,XX @@ static void dump_process(DumpState *s, Error **errp) | 147 | @@ -XXX,XX +XXX,XX @@ aio_ctx_dispatch(GSource *source, |
119 | 148 | AioContext *ctx = (AioContext *) source; | |
120 | static void *dump_thread(void *data) | 149 | |
121 | { | 150 | assert(callback == NULL); |
122 | - Error *err = NULL; | 151 | - aio_dispatch(ctx, true); |
123 | DumpState *s = (DumpState *)data; | 152 | + aio_dispatch(ctx); |
124 | - dump_process(s, &err); | 153 | return true; |
125 | - error_free(err); | ||
126 | + dump_process(s, NULL); | ||
127 | return NULL; | ||
128 | } | 154 | } |
129 | 155 | ||
130 | -- | 156 | -- |
131 | 2.13.5 | 157 | 2.9.3 |
132 | 158 | ||
133 | 159 | diff view generated by jsdifflib |
1 | From: Alberto Garcia <berto@igalia.com> | 1 | From: Paolo Bonzini <pbonzini@redhat.com> |
---|---|---|---|
2 | 2 | ||
3 | Both the throttling limits set with the throttling.iops-* and | 3 | Pull the increment/decrement pair out of aio_bh_poll and into the |
4 | throttling.bps-* options and their QMP equivalents defined in the | 4 | callers. |
5 | BlockIOThrottle struct are integer values. | ||
6 | 5 | ||
7 | Those limits are also reported in the BlockDeviceInfo struct and they | 6 | Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> |
8 | are integers there as well. | 7 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> |
9 | 8 | Reviewed-by: Fam Zheng <famz@redhat.com> | |
10 | Therefore there's no reason to store them internally as double and do | 9 | Reviewed-by: Daniel P. Berrange <berrange@redhat.com> |
11 | the conversion everytime we're setting or querying them, so this patch | 10 | Message-id: 20170213135235.12274-18-pbonzini@redhat.com |
12 | uses uint64_t for those types. Let's also use an unsigned type because | ||
13 | we don't allow negative values anyway. | ||
14 | |||
15 | LeakyBucket.level and LeakyBucket.burst_level do however remain double | ||
16 | because their value changes depending on the fraction of time elapsed | ||
17 | since the previous I/O operation. | ||
18 | |||
19 | Signed-off-by: Alberto Garcia <berto@igalia.com> | ||
20 | Message-id: f29b840422767b5be2c41c2dfdbbbf6c5f8fedf8.1503580370.git.berto@igalia.com | ||
21 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | 11 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
22 | --- | 12 | --- |
23 | include/qemu/throttle.h | 4 ++-- | 13 | util/aio-posix.c | 8 +++----- |
24 | tests/test-throttle.c | 3 ++- | 14 | util/aio-win32.c | 8 ++++---- |
25 | util/throttle.c | 7 +++---- | 15 | util/async.c | 12 ++++++------ |
26 | 3 files changed, 7 insertions(+), 7 deletions(-) | 16 | 3 files changed, 13 insertions(+), 15 deletions(-) |
27 | 17 | ||
28 | diff --git a/include/qemu/throttle.h b/include/qemu/throttle.h | 18 | diff --git a/util/aio-posix.c b/util/aio-posix.c |
29 | index XXXXXXX..XXXXXXX 100644 | 19 | index XXXXXXX..XXXXXXX 100644 |
30 | --- a/include/qemu/throttle.h | 20 | --- a/util/aio-posix.c |
31 | +++ b/include/qemu/throttle.h | 21 | +++ b/util/aio-posix.c |
32 | @@ -XXX,XX +XXX,XX @@ typedef enum { | 22 | @@ -XXX,XX +XXX,XX @@ static bool aio_dispatch_handlers(AioContext *ctx) |
33 | */ | 23 | |
34 | 24 | void aio_dispatch(AioContext *ctx) | |
35 | typedef struct LeakyBucket { | 25 | { |
36 | - double avg; /* average goal in units per second */ | 26 | + qemu_lockcnt_inc(&ctx->list_lock); |
37 | - double max; /* leaky bucket max burst in units */ | 27 | aio_bh_poll(ctx); |
38 | + uint64_t avg; /* average goal in units per second */ | 28 | - |
39 | + uint64_t max; /* leaky bucket max burst in units */ | 29 | - qemu_lockcnt_inc(&ctx->list_lock); |
40 | double level; /* bucket level in units */ | 30 | aio_dispatch_handlers(ctx); |
41 | double burst_level; /* bucket level in units (for computing bursts) */ | 31 | qemu_lockcnt_dec(&ctx->list_lock); |
42 | unsigned burst_length; /* max length of the burst period, in seconds */ | 32 | |
43 | diff --git a/tests/test-throttle.c b/tests/test-throttle.c | 33 | @@ -XXX,XX +XXX,XX @@ bool aio_poll(AioContext *ctx, bool blocking) |
34 | } | ||
35 | |||
36 | npfd = 0; | ||
37 | - qemu_lockcnt_dec(&ctx->list_lock); | ||
38 | |||
39 | progress |= aio_bh_poll(ctx); | ||
40 | |||
41 | if (ret > 0) { | ||
42 | - qemu_lockcnt_inc(&ctx->list_lock); | ||
43 | progress |= aio_dispatch_handlers(ctx); | ||
44 | - qemu_lockcnt_dec(&ctx->list_lock); | ||
45 | } | ||
46 | |||
47 | + qemu_lockcnt_dec(&ctx->list_lock); | ||
48 | + | ||
49 | progress |= timerlistgroup_run_timers(&ctx->tlg); | ||
50 | |||
51 | return progress; | ||
52 | diff --git a/util/aio-win32.c b/util/aio-win32.c | ||
44 | index XXXXXXX..XXXXXXX 100644 | 53 | index XXXXXXX..XXXXXXX 100644 |
45 | --- a/tests/test-throttle.c | 54 | --- a/util/aio-win32.c |
46 | +++ b/tests/test-throttle.c | 55 | +++ b/util/aio-win32.c |
47 | @@ -XXX,XX +XXX,XX @@ static void test_enabled(void) | 56 | @@ -XXX,XX +XXX,XX @@ static bool aio_dispatch_handlers(AioContext *ctx, HANDLE event) |
48 | for (i = 0; i < BUCKETS_COUNT; i++) { | 57 | bool progress = false; |
49 | throttle_config_init(&cfg); | 58 | AioHandler *tmp; |
50 | set_cfg_value(false, i, 150); | 59 | |
51 | + g_assert(throttle_is_valid(&cfg, NULL)); | 60 | - qemu_lockcnt_inc(&ctx->list_lock); |
52 | g_assert(throttle_enabled(&cfg)); | 61 | - |
62 | /* | ||
63 | * We have to walk very carefully in case aio_set_fd_handler is | ||
64 | * called while we're walking. | ||
65 | @@ -XXX,XX +XXX,XX @@ static bool aio_dispatch_handlers(AioContext *ctx, HANDLE event) | ||
66 | } | ||
53 | } | 67 | } |
54 | 68 | ||
55 | for (i = 0; i < BUCKETS_COUNT; i++) { | 69 | - qemu_lockcnt_dec(&ctx->list_lock); |
56 | throttle_config_init(&cfg); | 70 | return progress; |
57 | set_cfg_value(false, i, -150); | 71 | } |
58 | - g_assert(!throttle_enabled(&cfg)); | 72 | |
59 | + g_assert(!throttle_is_valid(&cfg, NULL)); | 73 | void aio_dispatch(AioContext *ctx) |
74 | { | ||
75 | + qemu_lockcnt_inc(&ctx->list_lock); | ||
76 | aio_bh_poll(ctx); | ||
77 | aio_dispatch_handlers(ctx, INVALID_HANDLE_VALUE); | ||
78 | + qemu_lockcnt_dec(&ctx->list_lock); | ||
79 | timerlistgroup_run_timers(&ctx->tlg); | ||
80 | } | ||
81 | |||
82 | @@ -XXX,XX +XXX,XX @@ bool aio_poll(AioContext *ctx, bool blocking) | ||
83 | } | ||
60 | } | 84 | } |
85 | |||
86 | - qemu_lockcnt_dec(&ctx->list_lock); | ||
87 | first = true; | ||
88 | |||
89 | /* ctx->notifier is always registered. */ | ||
90 | @@ -XXX,XX +XXX,XX @@ bool aio_poll(AioContext *ctx, bool blocking) | ||
91 | progress |= aio_dispatch_handlers(ctx, event); | ||
92 | } while (count > 0); | ||
93 | |||
94 | + qemu_lockcnt_dec(&ctx->list_lock); | ||
95 | + | ||
96 | progress |= timerlistgroup_run_timers(&ctx->tlg); | ||
97 | return progress; | ||
61 | } | 98 | } |
62 | 99 | diff --git a/util/async.c b/util/async.c | |
63 | diff --git a/util/throttle.c b/util/throttle.c | ||
64 | index XXXXXXX..XXXXXXX 100644 | 100 | index XXXXXXX..XXXXXXX 100644 |
65 | --- a/util/throttle.c | 101 | --- a/util/async.c |
66 | +++ b/util/throttle.c | 102 | +++ b/util/async.c |
67 | @@ -XXX,XX +XXX,XX @@ int64_t throttle_compute_wait(LeakyBucket *bkt) | 103 | @@ -XXX,XX +XXX,XX @@ void aio_bh_call(QEMUBH *bh) |
68 | /* If bkt->max is 0 we still want to allow short bursts of I/O | 104 | bh->cb(bh->opaque); |
69 | * from the guest, otherwise every other request will be throttled | 105 | } |
70 | * and performance will suffer considerably. */ | 106 | |
71 | - bucket_size = bkt->avg / 10; | 107 | -/* Multiple occurrences of aio_bh_poll cannot be called concurrently */ |
72 | + bucket_size = (double) bkt->avg / 10; | 108 | +/* Multiple occurrences of aio_bh_poll cannot be called concurrently. |
73 | burst_bucket_size = 0; | 109 | + * The count in ctx->list_lock is incremented before the call, and is |
74 | } else { | 110 | + * not affected by the call. |
75 | /* If we have a burst limit then we have to wait until all I/O | 111 | + */ |
76 | * at burst rate has finished before throttling to bkt->avg */ | 112 | int aio_bh_poll(AioContext *ctx) |
77 | bucket_size = bkt->max * bkt->burst_length; | 113 | { |
78 | - burst_bucket_size = bkt->max / 10; | 114 | QEMUBH *bh, **bhp, *next; |
79 | + burst_bucket_size = (double) bkt->max / 10; | 115 | int ret; |
116 | bool deleted = false; | ||
117 | |||
118 | - qemu_lockcnt_inc(&ctx->list_lock); | ||
119 | - | ||
120 | ret = 0; | ||
121 | for (bh = atomic_rcu_read(&ctx->first_bh); bh; bh = next) { | ||
122 | next = atomic_rcu_read(&bh->next); | ||
123 | @@ -XXX,XX +XXX,XX @@ int aio_bh_poll(AioContext *ctx) | ||
124 | |||
125 | /* remove deleted bhs */ | ||
126 | if (!deleted) { | ||
127 | - qemu_lockcnt_dec(&ctx->list_lock); | ||
128 | return ret; | ||
80 | } | 129 | } |
81 | 130 | ||
82 | /* If the main bucket is full then we have to wait */ | 131 | - if (qemu_lockcnt_dec_and_lock(&ctx->list_lock)) { |
83 | @@ -XXX,XX +XXX,XX @@ bool throttle_is_valid(ThrottleConfig *cfg, Error **errp) | 132 | + if (qemu_lockcnt_dec_if_lock(&ctx->list_lock)) { |
84 | 133 | bhp = &ctx->first_bh; | |
85 | for (i = 0; i < BUCKETS_COUNT; i++) { | 134 | while (*bhp) { |
86 | LeakyBucket *bkt = &cfg->buckets[i]; | 135 | bh = *bhp; |
87 | - if (bkt->avg < 0 || bkt->max < 0 || | 136 | @@ -XXX,XX +XXX,XX @@ int aio_bh_poll(AioContext *ctx) |
88 | - bkt->avg > THROTTLE_VALUE_MAX || bkt->max > THROTTLE_VALUE_MAX) { | 137 | bhp = &bh->next; |
89 | + if (bkt->avg > THROTTLE_VALUE_MAX || bkt->max > THROTTLE_VALUE_MAX) { | 138 | } |
90 | error_setg(errp, "bps/iops/max values must be within [0, %lld]", | 139 | } |
91 | THROTTLE_VALUE_MAX); | 140 | - qemu_lockcnt_unlock(&ctx->list_lock); |
92 | return false; | 141 | + qemu_lockcnt_inc_and_unlock(&ctx->list_lock); |
142 | } | ||
143 | return ret; | ||
144 | } | ||
93 | -- | 145 | -- |
94 | 2.13.5 | 146 | 2.9.3 |
95 | 147 | ||
96 | 148 | diff view generated by jsdifflib |
1 | From: Alberto Garcia <berto@igalia.com> | 1 | From: Paolo Bonzini <pbonzini@redhat.com> |
---|---|---|---|
2 | 2 | ||
3 | The level of the burst bucket is stored in bkt.burst_level, not | 3 | Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com> |
4 | bkt.burst_length. | 4 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> |
5 | 5 | Reviewed-by: Fam Zheng <famz@redhat.com> | |
6 | Signed-off-by: Alberto Garcia <berto@igalia.com> | 6 | Reviewed-by: Daniel P. Berrange <berrange@redhat.com> |
7 | Reviewed-by: Manos Pitsidianakis <el13635@mail.ntua.gr> | 7 | Message-id: 20170213135235.12274-19-pbonzini@redhat.com |
8 | Message-id: 49aab2711d02f285567f3b3b13a113847af33812.1503580370.git.berto@igalia.com | ||
9 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | 8 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
10 | --- | 9 | --- |
11 | include/qemu/throttle.h | 2 +- | 10 | include/block/block_int.h | 64 +++++++++++++++++++++++++----------------- |
12 | 1 file changed, 1 insertion(+), 1 deletion(-) | 11 | include/sysemu/block-backend.h | 14 ++++++--- |
12 | 2 files changed, 49 insertions(+), 29 deletions(-) | ||
13 | 13 | ||
14 | diff --git a/include/qemu/throttle.h b/include/qemu/throttle.h | 14 | diff --git a/include/block/block_int.h b/include/block/block_int.h |
15 | index XXXXXXX..XXXXXXX 100644 | 15 | index XXXXXXX..XXXXXXX 100644 |
16 | --- a/include/qemu/throttle.h | 16 | --- a/include/block/block_int.h |
17 | +++ b/include/qemu/throttle.h | 17 | +++ b/include/block/block_int.h |
18 | @@ -XXX,XX +XXX,XX @@ typedef enum { | 18 | @@ -XXX,XX +XXX,XX @@ struct BdrvChild { |
19 | * - The bkt.avg rate does not apply until the bucket is full, | 19 | * copied as well. |
20 | * allowing the user to do bursts until then. The I/O limit during | 20 | */ |
21 | * bursts is bkt.max. To enforce this limit we keep an additional | 21 | struct BlockDriverState { |
22 | - * bucket in bkt.burst_length that leaks at a rate of bkt.max units | 22 | - int64_t total_sectors; /* if we are reading a disk image, give its |
23 | + * bucket in bkt.burst_level that leaks at a rate of bkt.max units | 23 | - size in sectors */ |
24 | * per second. | 24 | + /* Protected by big QEMU lock or read-only after opening. No special |
25 | * | 25 | + * locking needed during I/O... |
26 | * - Because of all of the above, the user can perform I/O at a | 26 | + */ |
27 | int open_flags; /* flags used to open the file, re-used for re-open */ | ||
28 | bool read_only; /* if true, the media is read only */ | ||
29 | bool encrypted; /* if true, the media is encrypted */ | ||
30 | @@ -XXX,XX +XXX,XX @@ struct BlockDriverState { | ||
31 | bool sg; /* if true, the device is a /dev/sg* */ | ||
32 | bool probed; /* if true, format was probed rather than specified */ | ||
33 | |||
34 | - int copy_on_read; /* if nonzero, copy read backing sectors into image. | ||
35 | - note this is a reference count */ | ||
36 | - | ||
37 | - CoQueue flush_queue; /* Serializing flush queue */ | ||
38 | - bool active_flush_req; /* Flush request in flight? */ | ||
39 | - unsigned int write_gen; /* Current data generation */ | ||
40 | - unsigned int flushed_gen; /* Flushed write generation */ | ||
41 | - | ||
42 | BlockDriver *drv; /* NULL means no media */ | ||
43 | void *opaque; | ||
44 | |||
45 | @@ -XXX,XX +XXX,XX @@ struct BlockDriverState { | ||
46 | BdrvChild *backing; | ||
47 | BdrvChild *file; | ||
48 | |||
49 | - /* Callback before write request is processed */ | ||
50 | - NotifierWithReturnList before_write_notifiers; | ||
51 | - | ||
52 | - /* number of in-flight requests; overall and serialising */ | ||
53 | - unsigned int in_flight; | ||
54 | - unsigned int serialising_in_flight; | ||
55 | - | ||
56 | - bool wakeup; | ||
57 | - | ||
58 | - /* Offset after the highest byte written to */ | ||
59 | - uint64_t wr_highest_offset; | ||
60 | - | ||
61 | /* I/O Limits */ | ||
62 | BlockLimits bl; | ||
63 | |||
64 | @@ -XXX,XX +XXX,XX @@ struct BlockDriverState { | ||
65 | QTAILQ_ENTRY(BlockDriverState) bs_list; | ||
66 | /* element of the list of monitor-owned BDS */ | ||
67 | QTAILQ_ENTRY(BlockDriverState) monitor_list; | ||
68 | - QLIST_HEAD(, BdrvDirtyBitmap) dirty_bitmaps; | ||
69 | int refcnt; | ||
70 | |||
71 | - QLIST_HEAD(, BdrvTrackedRequest) tracked_requests; | ||
72 | - | ||
73 | /* operation blockers */ | ||
74 | QLIST_HEAD(, BdrvOpBlocker) op_blockers[BLOCK_OP_TYPE_MAX]; | ||
75 | |||
76 | @@ -XXX,XX +XXX,XX @@ struct BlockDriverState { | ||
77 | /* The error object in use for blocking operations on backing_hd */ | ||
78 | Error *backing_blocker; | ||
79 | |||
80 | + /* Protected by AioContext lock */ | ||
81 | + | ||
82 | + /* If true, copy read backing sectors into image. Can be >1 if more | ||
83 | + * than one client has requested copy-on-read. | ||
84 | + */ | ||
85 | + int copy_on_read; | ||
86 | + | ||
87 | + /* If we are reading a disk image, give its size in sectors. | ||
88 | + * Generally read-only; it is written to by load_vmstate and save_vmstate, | ||
89 | + * but the block layer is quiescent during those. | ||
90 | + */ | ||
91 | + int64_t total_sectors; | ||
92 | + | ||
93 | + /* Callback before write request is processed */ | ||
94 | + NotifierWithReturnList before_write_notifiers; | ||
95 | + | ||
96 | + /* number of in-flight requests; overall and serialising */ | ||
97 | + unsigned int in_flight; | ||
98 | + unsigned int serialising_in_flight; | ||
99 | + | ||
100 | + bool wakeup; | ||
101 | + | ||
102 | + /* Offset after the highest byte written to */ | ||
103 | + uint64_t wr_highest_offset; | ||
104 | + | ||
105 | /* threshold limit for writes, in bytes. "High water mark". */ | ||
106 | uint64_t write_threshold_offset; | ||
107 | NotifierWithReturn write_threshold_notifier; | ||
108 | @@ -XXX,XX +XXX,XX @@ struct BlockDriverState { | ||
109 | /* counter for nested bdrv_io_plug */ | ||
110 | unsigned io_plugged; | ||
111 | |||
112 | + QLIST_HEAD(, BdrvTrackedRequest) tracked_requests; | ||
113 | + CoQueue flush_queue; /* Serializing flush queue */ | ||
114 | + bool active_flush_req; /* Flush request in flight? */ | ||
115 | + unsigned int write_gen; /* Current data generation */ | ||
116 | + unsigned int flushed_gen; /* Flushed write generation */ | ||
117 | + | ||
118 | + QLIST_HEAD(, BdrvDirtyBitmap) dirty_bitmaps; | ||
119 | + | ||
120 | + /* do we need to tell the quest if we have a volatile write cache? */ | ||
121 | + int enable_write_cache; | ||
122 | + | ||
123 | int quiesce_counter; | ||
124 | }; | ||
125 | |||
126 | diff --git a/include/sysemu/block-backend.h b/include/sysemu/block-backend.h | ||
127 | index XXXXXXX..XXXXXXX 100644 | ||
128 | --- a/include/sysemu/block-backend.h | ||
129 | +++ b/include/sysemu/block-backend.h | ||
130 | @@ -XXX,XX +XXX,XX @@ typedef struct BlockDevOps { | ||
131 | * fields that must be public. This is in particular for QLIST_ENTRY() and | ||
132 | * friends so that BlockBackends can be kept in lists outside block-backend.c */ | ||
133 | typedef struct BlockBackendPublic { | ||
134 | - /* I/O throttling. | ||
135 | - * throttle_state tells us if this BlockBackend has I/O limits configured. | ||
136 | - * io_limits_disabled tells us if they are currently being enforced */ | ||
137 | + /* I/O throttling has its own locking, but also some fields are | ||
138 | + * protected by the AioContext lock. | ||
139 | + */ | ||
140 | + | ||
141 | + /* Protected by AioContext lock. */ | ||
142 | CoQueue throttled_reqs[2]; | ||
143 | + | ||
144 | + /* Nonzero if the I/O limits are currently being ignored; generally | ||
145 | + * it is zero. */ | ||
146 | unsigned int io_limits_disabled; | ||
147 | |||
148 | /* The following fields are protected by the ThrottleGroup lock. | ||
149 | - * See the ThrottleGroup documentation for details. */ | ||
150 | + * See the ThrottleGroup documentation for details. | ||
151 | + * throttle_state tells us if I/O limits are configured. */ | ||
152 | ThrottleState *throttle_state; | ||
153 | ThrottleTimers throttle_timers; | ||
154 | unsigned pending_reqs[2]; | ||
27 | -- | 155 | -- |
28 | 2.13.5 | 156 | 2.9.3 |
29 | 157 | ||
30 | 158 | diff view generated by jsdifflib |
New patch | |||
---|---|---|---|
1 | 1 | From: Paolo Bonzini <pbonzini@redhat.com> | |
2 | |||
3 | This uses the lock-free mutex described in the paper '"Blocking without | ||
4 | Locking", or LFTHREADS: A lock-free thread library' by Gidenstam and | ||
5 | Papatriantafilou. The same technique is used in OSv, and in fact | ||
6 | the code is essentially a conversion to C of OSv's code. | ||
7 | |||
8 | [Added missing coroutine_fn in tests/test-aio-multithread.c. | ||
9 | --Stefan] | ||
10 | |||
11 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> | ||
12 | Reviewed-by: Fam Zheng <famz@redhat.com> | ||
13 | Message-id: 20170213181244.16297-2-pbonzini@redhat.com | ||
14 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
15 | --- | ||
16 | include/qemu/coroutine.h | 17 ++++- | ||
17 | tests/test-aio-multithread.c | 86 ++++++++++++++++++++++++ | ||
18 | util/qemu-coroutine-lock.c | 155 ++++++++++++++++++++++++++++++++++++++++--- | ||
19 | util/trace-events | 1 + | ||
20 | 4 files changed, 246 insertions(+), 13 deletions(-) | ||
21 | |||
22 | diff --git a/include/qemu/coroutine.h b/include/qemu/coroutine.h | ||
23 | index XXXXXXX..XXXXXXX 100644 | ||
24 | --- a/include/qemu/coroutine.h | ||
25 | +++ b/include/qemu/coroutine.h | ||
26 | @@ -XXX,XX +XXX,XX @@ bool qemu_co_queue_empty(CoQueue *queue); | ||
27 | /** | ||
28 | * Provides a mutex that can be used to synchronise coroutines | ||
29 | */ | ||
30 | +struct CoWaitRecord; | ||
31 | typedef struct CoMutex { | ||
32 | - bool locked; | ||
33 | + /* Count of pending lockers; 0 for a free mutex, 1 for an | ||
34 | + * uncontended mutex. | ||
35 | + */ | ||
36 | + unsigned locked; | ||
37 | + | ||
38 | + /* A queue of waiters. Elements are added atomically in front of | ||
39 | + * from_push. to_pop is only populated, and popped from, by whoever | ||
40 | + * is in charge of the next wakeup. This can be an unlocker or, | ||
41 | + * through the handoff protocol, a locker that is about to go to sleep. | ||
42 | + */ | ||
43 | + QSLIST_HEAD(, CoWaitRecord) from_push, to_pop; | ||
44 | + | ||
45 | + unsigned handoff, sequence; | ||
46 | + | ||
47 | Coroutine *holder; | ||
48 | - CoQueue queue; | ||
49 | } CoMutex; | ||
50 | |||
51 | /** | ||
52 | diff --git a/tests/test-aio-multithread.c b/tests/test-aio-multithread.c | ||
53 | index XXXXXXX..XXXXXXX 100644 | ||
54 | --- a/tests/test-aio-multithread.c | ||
55 | +++ b/tests/test-aio-multithread.c | ||
56 | @@ -XXX,XX +XXX,XX @@ static void test_multi_co_schedule_10(void) | ||
57 | test_multi_co_schedule(10); | ||
58 | } | ||
59 | |||
60 | +/* CoMutex thread-safety. */ | ||
61 | + | ||
62 | +static uint32_t atomic_counter; | ||
63 | +static uint32_t running; | ||
64 | +static uint32_t counter; | ||
65 | +static CoMutex comutex; | ||
66 | + | ||
67 | +static void coroutine_fn test_multi_co_mutex_entry(void *opaque) | ||
68 | +{ | ||
69 | + while (!atomic_mb_read(&now_stopping)) { | ||
70 | + qemu_co_mutex_lock(&comutex); | ||
71 | + counter++; | ||
72 | + qemu_co_mutex_unlock(&comutex); | ||
73 | + | ||
74 | + /* Increase atomic_counter *after* releasing the mutex. Otherwise | ||
75 | + * there is a chance (it happens about 1 in 3 runs) that the iothread | ||
76 | + * exits before the coroutine is woken up, causing a spurious | ||
77 | + * assertion failure. | ||
78 | + */ | ||
79 | + atomic_inc(&atomic_counter); | ||
80 | + } | ||
81 | + atomic_dec(&running); | ||
82 | +} | ||
83 | + | ||
84 | +static void test_multi_co_mutex(int threads, int seconds) | ||
85 | +{ | ||
86 | + int i; | ||
87 | + | ||
88 | + qemu_co_mutex_init(&comutex); | ||
89 | + counter = 0; | ||
90 | + atomic_counter = 0; | ||
91 | + now_stopping = false; | ||
92 | + | ||
93 | + create_aio_contexts(); | ||
94 | + assert(threads <= NUM_CONTEXTS); | ||
95 | + running = threads; | ||
96 | + for (i = 0; i < threads; i++) { | ||
97 | + Coroutine *co1 = qemu_coroutine_create(test_multi_co_mutex_entry, NULL); | ||
98 | + aio_co_schedule(ctx[i], co1); | ||
99 | + } | ||
100 | + | ||
101 | + g_usleep(seconds * 1000000); | ||
102 | + | ||
103 | + atomic_mb_set(&now_stopping, true); | ||
104 | + while (running > 0) { | ||
105 | + g_usleep(100000); | ||
106 | + } | ||
107 | + | ||
108 | + join_aio_contexts(); | ||
109 | + g_test_message("%d iterations/second\n", counter / seconds); | ||
110 | + g_assert_cmpint(counter, ==, atomic_counter); | ||
111 | +} | ||
112 | + | ||
113 | +/* Testing with NUM_CONTEXTS threads focuses on the queue. The mutex however | ||
114 | + * is too contended (and the threads spend too much time in aio_poll) | ||
115 | + * to actually stress the handoff protocol. | ||
116 | + */ | ||
117 | +static void test_multi_co_mutex_1(void) | ||
118 | +{ | ||
119 | + test_multi_co_mutex(NUM_CONTEXTS, 1); | ||
120 | +} | ||
121 | + | ||
122 | +static void test_multi_co_mutex_10(void) | ||
123 | +{ | ||
124 | + test_multi_co_mutex(NUM_CONTEXTS, 10); | ||
125 | +} | ||
126 | + | ||
127 | +/* Testing with fewer threads stresses the handoff protocol too. Still, the | ||
128 | + * case where the locker _can_ pick up a handoff is very rare, happening | ||
129 | + * about 10 times in 1 million, so increase the runtime a bit compared to | ||
130 | + * other "quick" testcases that only run for 1 second. | ||
131 | + */ | ||
132 | +static void test_multi_co_mutex_2_3(void) | ||
133 | +{ | ||
134 | + test_multi_co_mutex(2, 3); | ||
135 | +} | ||
136 | + | ||
137 | +static void test_multi_co_mutex_2_30(void) | ||
138 | +{ | ||
139 | + test_multi_co_mutex(2, 30); | ||
140 | +} | ||
141 | + | ||
142 | /* End of tests. */ | ||
143 | |||
144 | int main(int argc, char **argv) | ||
145 | @@ -XXX,XX +XXX,XX @@ int main(int argc, char **argv) | ||
146 | g_test_add_func("/aio/multi/lifecycle", test_lifecycle); | ||
147 | if (g_test_quick()) { | ||
148 | g_test_add_func("/aio/multi/schedule", test_multi_co_schedule_1); | ||
149 | + g_test_add_func("/aio/multi/mutex/contended", test_multi_co_mutex_1); | ||
150 | + g_test_add_func("/aio/multi/mutex/handoff", test_multi_co_mutex_2_3); | ||
151 | } else { | ||
152 | g_test_add_func("/aio/multi/schedule", test_multi_co_schedule_10); | ||
153 | + g_test_add_func("/aio/multi/mutex/contended", test_multi_co_mutex_10); | ||
154 | + g_test_add_func("/aio/multi/mutex/handoff", test_multi_co_mutex_2_30); | ||
155 | } | ||
156 | return g_test_run(); | ||
157 | } | ||
158 | diff --git a/util/qemu-coroutine-lock.c b/util/qemu-coroutine-lock.c | ||
159 | index XXXXXXX..XXXXXXX 100644 | ||
160 | --- a/util/qemu-coroutine-lock.c | ||
161 | +++ b/util/qemu-coroutine-lock.c | ||
162 | @@ -XXX,XX +XXX,XX @@ | ||
163 | * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | ||
164 | * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN | ||
165 | * THE SOFTWARE. | ||
166 | + * | ||
167 | + * The lock-free mutex implementation is based on OSv | ||
168 | + * (core/lfmutex.cc, include/lockfree/mutex.hh). | ||
169 | + * Copyright (C) 2013 Cloudius Systems, Ltd. | ||
170 | */ | ||
171 | |||
172 | #include "qemu/osdep.h" | ||
173 | @@ -XXX,XX +XXX,XX @@ bool qemu_co_queue_empty(CoQueue *queue) | ||
174 | return QSIMPLEQ_FIRST(&queue->entries) == NULL; | ||
175 | } | ||
176 | |||
177 | +/* The wait records are handled with a multiple-producer, single-consumer | ||
178 | + * lock-free queue. There cannot be two concurrent pop_waiter() calls | ||
179 | + * because pop_waiter() can only be called while mutex->handoff is zero. | ||
180 | + * This can happen in three cases: | ||
181 | + * - in qemu_co_mutex_unlock, before the hand-off protocol has started. | ||
182 | + * In this case, qemu_co_mutex_lock will see mutex->handoff == 0 and | ||
183 | + * not take part in the handoff. | ||
184 | + * - in qemu_co_mutex_lock, if it steals the hand-off responsibility from | ||
185 | + * qemu_co_mutex_unlock. In this case, qemu_co_mutex_unlock will fail | ||
186 | + * the cmpxchg (it will see either 0 or the next sequence value) and | ||
187 | + * exit. The next hand-off cannot begin until qemu_co_mutex_lock has | ||
188 | + * woken up someone. | ||
189 | + * - in qemu_co_mutex_unlock, if it takes the hand-off token itself. | ||
190 | + * In this case another iteration starts with mutex->handoff == 0; | ||
191 | + * a concurrent qemu_co_mutex_lock will fail the cmpxchg, and | ||
192 | + * qemu_co_mutex_unlock will go back to case (1). | ||
193 | + * | ||
194 | + * The following functions manage this queue. | ||
195 | + */ | ||
196 | +typedef struct CoWaitRecord { | ||
197 | + Coroutine *co; | ||
198 | + QSLIST_ENTRY(CoWaitRecord) next; | ||
199 | +} CoWaitRecord; | ||
200 | + | ||
201 | +static void push_waiter(CoMutex *mutex, CoWaitRecord *w) | ||
202 | +{ | ||
203 | + w->co = qemu_coroutine_self(); | ||
204 | + QSLIST_INSERT_HEAD_ATOMIC(&mutex->from_push, w, next); | ||
205 | +} | ||
206 | + | ||
207 | +static void move_waiters(CoMutex *mutex) | ||
208 | +{ | ||
209 | + QSLIST_HEAD(, CoWaitRecord) reversed; | ||
210 | + QSLIST_MOVE_ATOMIC(&reversed, &mutex->from_push); | ||
211 | + while (!QSLIST_EMPTY(&reversed)) { | ||
212 | + CoWaitRecord *w = QSLIST_FIRST(&reversed); | ||
213 | + QSLIST_REMOVE_HEAD(&reversed, next); | ||
214 | + QSLIST_INSERT_HEAD(&mutex->to_pop, w, next); | ||
215 | + } | ||
216 | +} | ||
217 | + | ||
218 | +static CoWaitRecord *pop_waiter(CoMutex *mutex) | ||
219 | +{ | ||
220 | + CoWaitRecord *w; | ||
221 | + | ||
222 | + if (QSLIST_EMPTY(&mutex->to_pop)) { | ||
223 | + move_waiters(mutex); | ||
224 | + if (QSLIST_EMPTY(&mutex->to_pop)) { | ||
225 | + return NULL; | ||
226 | + } | ||
227 | + } | ||
228 | + w = QSLIST_FIRST(&mutex->to_pop); | ||
229 | + QSLIST_REMOVE_HEAD(&mutex->to_pop, next); | ||
230 | + return w; | ||
231 | +} | ||
232 | + | ||
233 | +static bool has_waiters(CoMutex *mutex) | ||
234 | +{ | ||
235 | + return QSLIST_EMPTY(&mutex->to_pop) || QSLIST_EMPTY(&mutex->from_push); | ||
236 | +} | ||
237 | + | ||
238 | void qemu_co_mutex_init(CoMutex *mutex) | ||
239 | { | ||
240 | memset(mutex, 0, sizeof(*mutex)); | ||
241 | - qemu_co_queue_init(&mutex->queue); | ||
242 | } | ||
243 | |||
244 | -void coroutine_fn qemu_co_mutex_lock(CoMutex *mutex) | ||
245 | +static void coroutine_fn qemu_co_mutex_lock_slowpath(CoMutex *mutex) | ||
246 | { | ||
247 | Coroutine *self = qemu_coroutine_self(); | ||
248 | + CoWaitRecord w; | ||
249 | + unsigned old_handoff; | ||
250 | |||
251 | trace_qemu_co_mutex_lock_entry(mutex, self); | ||
252 | + w.co = self; | ||
253 | + push_waiter(mutex, &w); | ||
254 | |||
255 | - while (mutex->locked) { | ||
256 | - qemu_co_queue_wait(&mutex->queue); | ||
257 | + /* This is the "Responsibility Hand-Off" protocol; a lock() picks from | ||
258 | + * a concurrent unlock() the responsibility of waking somebody up. | ||
259 | + */ | ||
260 | + old_handoff = atomic_mb_read(&mutex->handoff); | ||
261 | + if (old_handoff && | ||
262 | + has_waiters(mutex) && | ||
263 | + atomic_cmpxchg(&mutex->handoff, old_handoff, 0) == old_handoff) { | ||
264 | + /* There can be no concurrent pops, because there can be only | ||
265 | + * one active handoff at a time. | ||
266 | + */ | ||
267 | + CoWaitRecord *to_wake = pop_waiter(mutex); | ||
268 | + Coroutine *co = to_wake->co; | ||
269 | + if (co == self) { | ||
270 | + /* We got the lock ourselves! */ | ||
271 | + assert(to_wake == &w); | ||
272 | + return; | ||
273 | + } | ||
274 | + | ||
275 | + aio_co_wake(co); | ||
276 | } | ||
277 | |||
278 | - mutex->locked = true; | ||
279 | - mutex->holder = self; | ||
280 | - self->locks_held++; | ||
281 | - | ||
282 | + qemu_coroutine_yield(); | ||
283 | trace_qemu_co_mutex_lock_return(mutex, self); | ||
284 | } | ||
285 | |||
286 | +void coroutine_fn qemu_co_mutex_lock(CoMutex *mutex) | ||
287 | +{ | ||
288 | + Coroutine *self = qemu_coroutine_self(); | ||
289 | + | ||
290 | + if (atomic_fetch_inc(&mutex->locked) == 0) { | ||
291 | + /* Uncontended. */ | ||
292 | + trace_qemu_co_mutex_lock_uncontended(mutex, self); | ||
293 | + } else { | ||
294 | + qemu_co_mutex_lock_slowpath(mutex); | ||
295 | + } | ||
296 | + mutex->holder = self; | ||
297 | + self->locks_held++; | ||
298 | +} | ||
299 | + | ||
300 | void coroutine_fn qemu_co_mutex_unlock(CoMutex *mutex) | ||
301 | { | ||
302 | Coroutine *self = qemu_coroutine_self(); | ||
303 | |||
304 | trace_qemu_co_mutex_unlock_entry(mutex, self); | ||
305 | |||
306 | - assert(mutex->locked == true); | ||
307 | + assert(mutex->locked); | ||
308 | assert(mutex->holder == self); | ||
309 | assert(qemu_in_coroutine()); | ||
310 | |||
311 | - mutex->locked = false; | ||
312 | mutex->holder = NULL; | ||
313 | self->locks_held--; | ||
314 | - qemu_co_queue_next(&mutex->queue); | ||
315 | + if (atomic_fetch_dec(&mutex->locked) == 1) { | ||
316 | + /* No waiting qemu_co_mutex_lock(). Pfew, that was easy! */ | ||
317 | + return; | ||
318 | + } | ||
319 | + | ||
320 | + for (;;) { | ||
321 | + CoWaitRecord *to_wake = pop_waiter(mutex); | ||
322 | + unsigned our_handoff; | ||
323 | + | ||
324 | + if (to_wake) { | ||
325 | + Coroutine *co = to_wake->co; | ||
326 | + aio_co_wake(co); | ||
327 | + break; | ||
328 | + } | ||
329 | + | ||
330 | + /* Some concurrent lock() is in progress (we know this because | ||
331 | + * mutex->locked was >1) but it hasn't yet put itself on the wait | ||
332 | + * queue. Pick a sequence number for the handoff protocol (not 0). | ||
333 | + */ | ||
334 | + if (++mutex->sequence == 0) { | ||
335 | + mutex->sequence = 1; | ||
336 | + } | ||
337 | + | ||
338 | + our_handoff = mutex->sequence; | ||
339 | + atomic_mb_set(&mutex->handoff, our_handoff); | ||
340 | + if (!has_waiters(mutex)) { | ||
341 | + /* The concurrent lock has not added itself yet, so it | ||
342 | + * will be able to pick our handoff. | ||
343 | + */ | ||
344 | + break; | ||
345 | + } | ||
346 | + | ||
347 | + /* Try to do the handoff protocol ourselves; if somebody else has | ||
348 | + * already taken it, however, we're done and they're responsible. | ||
349 | + */ | ||
350 | + if (atomic_cmpxchg(&mutex->handoff, our_handoff, 0) != our_handoff) { | ||
351 | + break; | ||
352 | + } | ||
353 | + } | ||
354 | |||
355 | trace_qemu_co_mutex_unlock_return(mutex, self); | ||
356 | } | ||
357 | diff --git a/util/trace-events b/util/trace-events | ||
358 | index XXXXXXX..XXXXXXX 100644 | ||
359 | --- a/util/trace-events | ||
360 | +++ b/util/trace-events | ||
361 | @@ -XXX,XX +XXX,XX @@ qemu_coroutine_terminate(void *co) "self %p" | ||
362 | |||
363 | # util/qemu-coroutine-lock.c | ||
364 | qemu_co_queue_run_restart(void *co) "co %p" | ||
365 | +qemu_co_mutex_lock_uncontended(void *mutex, void *self) "mutex %p self %p" | ||
366 | qemu_co_mutex_lock_entry(void *mutex, void *self) "mutex %p self %p" | ||
367 | qemu_co_mutex_lock_return(void *mutex, void *self) "mutex %p self %p" | ||
368 | qemu_co_mutex_unlock_entry(void *mutex, void *self) "mutex %p self %p" | ||
369 | -- | ||
370 | 2.9.3 | ||
371 | |||
372 | diff view generated by jsdifflib |
1 | From: Alberto Garcia <berto@igalia.com> | 1 | From: Paolo Bonzini <pbonzini@redhat.com> |
---|---|---|---|
2 | 2 | ||
3 | Use a pointer to the bucket instead of repeating cfg->buckets[i] all | 3 | Running a very small critical section on pthread_mutex_t and CoMutex |
4 | the time. This makes the code more concise and will help us expand the | 4 | shows that pthread_mutex_t is much faster because it doesn't actually |
5 | checks later and save a few line breaks. | 5 | go to sleep. What happens is that the critical section is shorter |
6 | than the latency of entering the kernel and thus FUTEX_WAIT always | ||
7 | fails. With CoMutex there is no such latency but you still want to | ||
8 | avoid wait and wakeup. So introduce it artificially. | ||
6 | 9 | ||
7 | Signed-off-by: Alberto Garcia <berto@igalia.com> | 10 | This only works with one waiters; because CoMutex is fair, it will |
8 | Message-id: 763ffc40a26b17d54cf93f5a999e4656049fcf0c.1503580370.git.berto@igalia.com | 11 | always have more waits and wakeups than a pthread_mutex_t. |
12 | |||
13 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> | ||
14 | Reviewed-by: Fam Zheng <famz@redhat.com> | ||
15 | Message-id: 20170213181244.16297-3-pbonzini@redhat.com | ||
9 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | 16 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
10 | --- | 17 | --- |
11 | util/throttle.c | 15 +++++++-------- | 18 | include/qemu/coroutine.h | 5 +++++ |
12 | 1 file changed, 7 insertions(+), 8 deletions(-) | 19 | util/qemu-coroutine-lock.c | 51 ++++++++++++++++++++++++++++++++++++++++------ |
20 | util/qemu-coroutine.c | 2 +- | ||
21 | 3 files changed, 51 insertions(+), 7 deletions(-) | ||
13 | 22 | ||
14 | diff --git a/util/throttle.c b/util/throttle.c | 23 | diff --git a/include/qemu/coroutine.h b/include/qemu/coroutine.h |
15 | index XXXXXXX..XXXXXXX 100644 | 24 | index XXXXXXX..XXXXXXX 100644 |
16 | --- a/util/throttle.c | 25 | --- a/include/qemu/coroutine.h |
17 | +++ b/util/throttle.c | 26 | +++ b/include/qemu/coroutine.h |
18 | @@ -XXX,XX +XXX,XX @@ bool throttle_is_valid(ThrottleConfig *cfg, Error **errp) | 27 | @@ -XXX,XX +XXX,XX @@ typedef struct CoMutex { |
28 | */ | ||
29 | unsigned locked; | ||
30 | |||
31 | + /* Context that is holding the lock. Useful to avoid spinning | ||
32 | + * when two coroutines on the same AioContext try to get the lock. :) | ||
33 | + */ | ||
34 | + AioContext *ctx; | ||
35 | + | ||
36 | /* A queue of waiters. Elements are added atomically in front of | ||
37 | * from_push. to_pop is only populated, and popped from, by whoever | ||
38 | * is in charge of the next wakeup. This can be an unlocker or, | ||
39 | diff --git a/util/qemu-coroutine-lock.c b/util/qemu-coroutine-lock.c | ||
40 | index XXXXXXX..XXXXXXX 100644 | ||
41 | --- a/util/qemu-coroutine-lock.c | ||
42 | +++ b/util/qemu-coroutine-lock.c | ||
43 | @@ -XXX,XX +XXX,XX @@ | ||
44 | #include "qemu-common.h" | ||
45 | #include "qemu/coroutine.h" | ||
46 | #include "qemu/coroutine_int.h" | ||
47 | +#include "qemu/processor.h" | ||
48 | #include "qemu/queue.h" | ||
49 | #include "block/aio.h" | ||
50 | #include "trace.h" | ||
51 | @@ -XXX,XX +XXX,XX @@ void qemu_co_mutex_init(CoMutex *mutex) | ||
52 | memset(mutex, 0, sizeof(*mutex)); | ||
53 | } | ||
54 | |||
55 | -static void coroutine_fn qemu_co_mutex_lock_slowpath(CoMutex *mutex) | ||
56 | +static void coroutine_fn qemu_co_mutex_wake(CoMutex *mutex, Coroutine *co) | ||
57 | +{ | ||
58 | + /* Read co before co->ctx; pairs with smp_wmb() in | ||
59 | + * qemu_coroutine_enter(). | ||
60 | + */ | ||
61 | + smp_read_barrier_depends(); | ||
62 | + mutex->ctx = co->ctx; | ||
63 | + aio_co_wake(co); | ||
64 | +} | ||
65 | + | ||
66 | +static void coroutine_fn qemu_co_mutex_lock_slowpath(AioContext *ctx, | ||
67 | + CoMutex *mutex) | ||
68 | { | ||
69 | Coroutine *self = qemu_coroutine_self(); | ||
70 | CoWaitRecord w; | ||
71 | @@ -XXX,XX +XXX,XX @@ static void coroutine_fn qemu_co_mutex_lock_slowpath(CoMutex *mutex) | ||
72 | if (co == self) { | ||
73 | /* We got the lock ourselves! */ | ||
74 | assert(to_wake == &w); | ||
75 | + mutex->ctx = ctx; | ||
76 | return; | ||
77 | } | ||
78 | |||
79 | - aio_co_wake(co); | ||
80 | + qemu_co_mutex_wake(mutex, co); | ||
19 | } | 81 | } |
20 | 82 | ||
21 | for (i = 0; i < BUCKETS_COUNT; i++) { | 83 | qemu_coroutine_yield(); |
22 | - if (cfg->buckets[i].avg < 0 || | 84 | @@ -XXX,XX +XXX,XX @@ static void coroutine_fn qemu_co_mutex_lock_slowpath(CoMutex *mutex) |
23 | - cfg->buckets[i].max < 0 || | 85 | |
24 | - cfg->buckets[i].avg > THROTTLE_VALUE_MAX || | 86 | void coroutine_fn qemu_co_mutex_lock(CoMutex *mutex) |
25 | - cfg->buckets[i].max > THROTTLE_VALUE_MAX) { | 87 | { |
26 | + LeakyBucket *bkt = &cfg->buckets[i]; | 88 | + AioContext *ctx = qemu_get_current_aio_context(); |
27 | + if (bkt->avg < 0 || bkt->max < 0 || | 89 | Coroutine *self = qemu_coroutine_self(); |
28 | + bkt->avg > THROTTLE_VALUE_MAX || bkt->max > THROTTLE_VALUE_MAX) { | 90 | + int waiters, i; |
29 | error_setg(errp, "bps/iops/max values must be within [0, %lld]", | 91 | |
30 | THROTTLE_VALUE_MAX); | 92 | - if (atomic_fetch_inc(&mutex->locked) == 0) { |
31 | return false; | 93 | + /* Running a very small critical section on pthread_mutex_t and CoMutex |
94 | + * shows that pthread_mutex_t is much faster because it doesn't actually | ||
95 | + * go to sleep. What happens is that the critical section is shorter | ||
96 | + * than the latency of entering the kernel and thus FUTEX_WAIT always | ||
97 | + * fails. With CoMutex there is no such latency but you still want to | ||
98 | + * avoid wait and wakeup. So introduce it artificially. | ||
99 | + */ | ||
100 | + i = 0; | ||
101 | +retry_fast_path: | ||
102 | + waiters = atomic_cmpxchg(&mutex->locked, 0, 1); | ||
103 | + if (waiters != 0) { | ||
104 | + while (waiters == 1 && ++i < 1000) { | ||
105 | + if (atomic_read(&mutex->ctx) == ctx) { | ||
106 | + break; | ||
107 | + } | ||
108 | + if (atomic_read(&mutex->locked) == 0) { | ||
109 | + goto retry_fast_path; | ||
110 | + } | ||
111 | + cpu_relax(); | ||
112 | + } | ||
113 | + waiters = atomic_fetch_inc(&mutex->locked); | ||
114 | + } | ||
115 | + | ||
116 | + if (waiters == 0) { | ||
117 | /* Uncontended. */ | ||
118 | trace_qemu_co_mutex_lock_uncontended(mutex, self); | ||
119 | + mutex->ctx = ctx; | ||
120 | } else { | ||
121 | - qemu_co_mutex_lock_slowpath(mutex); | ||
122 | + qemu_co_mutex_lock_slowpath(ctx, mutex); | ||
123 | } | ||
124 | mutex->holder = self; | ||
125 | self->locks_held++; | ||
126 | @@ -XXX,XX +XXX,XX @@ void coroutine_fn qemu_co_mutex_unlock(CoMutex *mutex) | ||
127 | assert(mutex->holder == self); | ||
128 | assert(qemu_in_coroutine()); | ||
129 | |||
130 | + mutex->ctx = NULL; | ||
131 | mutex->holder = NULL; | ||
132 | self->locks_held--; | ||
133 | if (atomic_fetch_dec(&mutex->locked) == 1) { | ||
134 | @@ -XXX,XX +XXX,XX @@ void coroutine_fn qemu_co_mutex_unlock(CoMutex *mutex) | ||
135 | unsigned our_handoff; | ||
136 | |||
137 | if (to_wake) { | ||
138 | - Coroutine *co = to_wake->co; | ||
139 | - aio_co_wake(co); | ||
140 | + qemu_co_mutex_wake(mutex, to_wake->co); | ||
141 | break; | ||
32 | } | 142 | } |
33 | 143 | ||
34 | - if (!cfg->buckets[i].burst_length) { | 144 | diff --git a/util/qemu-coroutine.c b/util/qemu-coroutine.c |
35 | + if (!bkt->burst_length) { | 145 | index XXXXXXX..XXXXXXX 100644 |
36 | error_setg(errp, "the burst length cannot be 0"); | 146 | --- a/util/qemu-coroutine.c |
37 | return false; | 147 | +++ b/util/qemu-coroutine.c |
38 | } | 148 | @@ -XXX,XX +XXX,XX @@ void qemu_coroutine_enter(Coroutine *co) |
39 | 149 | co->ctx = qemu_get_current_aio_context(); | |
40 | - if (cfg->buckets[i].burst_length > 1 && !cfg->buckets[i].max) { | 150 | |
41 | + if (bkt->burst_length > 1 && !bkt->max) { | 151 | /* Store co->ctx before anything that stores co. Matches |
42 | error_setg(errp, "burst length set without burst rate"); | 152 | - * barrier in aio_co_wake. |
43 | return false; | 153 | + * barrier in aio_co_wake and qemu_co_mutex_wake. |
44 | } | 154 | */ |
45 | 155 | smp_wmb(); | |
46 | - if (cfg->buckets[i].max && !cfg->buckets[i].avg) { | 156 | |
47 | + if (bkt->max && !bkt->avg) { | ||
48 | error_setg(errp, "bps_max/iops_max require corresponding" | ||
49 | " bps/iops values"); | ||
50 | return false; | ||
51 | } | ||
52 | |||
53 | - if (cfg->buckets[i].max && cfg->buckets[i].max < cfg->buckets[i].avg) { | ||
54 | + if (bkt->max && bkt->max < bkt->avg) { | ||
55 | error_setg(errp, "bps_max/iops_max cannot be lower than bps/iops"); | ||
56 | return false; | ||
57 | } | ||
58 | -- | 157 | -- |
59 | 2.13.5 | 158 | 2.9.3 |
60 | 159 | ||
61 | 160 | diff view generated by jsdifflib |
New patch | |||
---|---|---|---|
1 | 1 | From: Paolo Bonzini <pbonzini@redhat.com> | |
2 | |||
3 | Add two implementations of the same benchmark as the previous patch, | ||
4 | but using pthreads. One uses a normal QemuMutex, the other is Linux | ||
5 | only and implements a fair mutex based on MCS locks and futexes. | ||
6 | This shows that the slower performance of the 5-thread case is due to | ||
7 | the fairness of CoMutex, rather than to coroutines. If fairness does | ||
8 | not matter, as is the case with two threads, CoMutex can actually be | ||
9 | faster than pthreads. | ||
10 | |||
11 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> | ||
12 | Reviewed-by: Fam Zheng <famz@redhat.com> | ||
13 | Message-id: 20170213181244.16297-4-pbonzini@redhat.com | ||
14 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
15 | --- | ||
16 | tests/test-aio-multithread.c | 164 +++++++++++++++++++++++++++++++++++++++++++ | ||
17 | 1 file changed, 164 insertions(+) | ||
18 | |||
19 | diff --git a/tests/test-aio-multithread.c b/tests/test-aio-multithread.c | ||
20 | index XXXXXXX..XXXXXXX 100644 | ||
21 | --- a/tests/test-aio-multithread.c | ||
22 | +++ b/tests/test-aio-multithread.c | ||
23 | @@ -XXX,XX +XXX,XX @@ static void test_multi_co_mutex_2_30(void) | ||
24 | test_multi_co_mutex(2, 30); | ||
25 | } | ||
26 | |||
27 | +/* Same test with fair mutexes, for performance comparison. */ | ||
28 | + | ||
29 | +#ifdef CONFIG_LINUX | ||
30 | +#include "qemu/futex.h" | ||
31 | + | ||
32 | +/* The nodes for the mutex reside in this structure (on which we try to avoid | ||
33 | + * false sharing). The head of the mutex is in the "mutex_head" variable. | ||
34 | + */ | ||
35 | +static struct { | ||
36 | + int next, locked; | ||
37 | + int padding[14]; | ||
38 | +} nodes[NUM_CONTEXTS] __attribute__((__aligned__(64))); | ||
39 | + | ||
40 | +static int mutex_head = -1; | ||
41 | + | ||
42 | +static void mcs_mutex_lock(void) | ||
43 | +{ | ||
44 | + int prev; | ||
45 | + | ||
46 | + nodes[id].next = -1; | ||
47 | + nodes[id].locked = 1; | ||
48 | + prev = atomic_xchg(&mutex_head, id); | ||
49 | + if (prev != -1) { | ||
50 | + atomic_set(&nodes[prev].next, id); | ||
51 | + qemu_futex_wait(&nodes[id].locked, 1); | ||
52 | + } | ||
53 | +} | ||
54 | + | ||
55 | +static void mcs_mutex_unlock(void) | ||
56 | +{ | ||
57 | + int next; | ||
58 | + if (nodes[id].next == -1) { | ||
59 | + if (atomic_read(&mutex_head) == id && | ||
60 | + atomic_cmpxchg(&mutex_head, id, -1) == id) { | ||
61 | + /* Last item in the list, exit. */ | ||
62 | + return; | ||
63 | + } | ||
64 | + while (atomic_read(&nodes[id].next) == -1) { | ||
65 | + /* mcs_mutex_lock did the xchg, but has not updated | ||
66 | + * nodes[prev].next yet. | ||
67 | + */ | ||
68 | + } | ||
69 | + } | ||
70 | + | ||
71 | + /* Wake up the next in line. */ | ||
72 | + next = nodes[id].next; | ||
73 | + nodes[next].locked = 0; | ||
74 | + qemu_futex_wake(&nodes[next].locked, 1); | ||
75 | +} | ||
76 | + | ||
77 | +static void test_multi_fair_mutex_entry(void *opaque) | ||
78 | +{ | ||
79 | + while (!atomic_mb_read(&now_stopping)) { | ||
80 | + mcs_mutex_lock(); | ||
81 | + counter++; | ||
82 | + mcs_mutex_unlock(); | ||
83 | + atomic_inc(&atomic_counter); | ||
84 | + } | ||
85 | + atomic_dec(&running); | ||
86 | +} | ||
87 | + | ||
88 | +static void test_multi_fair_mutex(int threads, int seconds) | ||
89 | +{ | ||
90 | + int i; | ||
91 | + | ||
92 | + assert(mutex_head == -1); | ||
93 | + counter = 0; | ||
94 | + atomic_counter = 0; | ||
95 | + now_stopping = false; | ||
96 | + | ||
97 | + create_aio_contexts(); | ||
98 | + assert(threads <= NUM_CONTEXTS); | ||
99 | + running = threads; | ||
100 | + for (i = 0; i < threads; i++) { | ||
101 | + Coroutine *co1 = qemu_coroutine_create(test_multi_fair_mutex_entry, NULL); | ||
102 | + aio_co_schedule(ctx[i], co1); | ||
103 | + } | ||
104 | + | ||
105 | + g_usleep(seconds * 1000000); | ||
106 | + | ||
107 | + atomic_mb_set(&now_stopping, true); | ||
108 | + while (running > 0) { | ||
109 | + g_usleep(100000); | ||
110 | + } | ||
111 | + | ||
112 | + join_aio_contexts(); | ||
113 | + g_test_message("%d iterations/second\n", counter / seconds); | ||
114 | + g_assert_cmpint(counter, ==, atomic_counter); | ||
115 | +} | ||
116 | + | ||
117 | +static void test_multi_fair_mutex_1(void) | ||
118 | +{ | ||
119 | + test_multi_fair_mutex(NUM_CONTEXTS, 1); | ||
120 | +} | ||
121 | + | ||
122 | +static void test_multi_fair_mutex_10(void) | ||
123 | +{ | ||
124 | + test_multi_fair_mutex(NUM_CONTEXTS, 10); | ||
125 | +} | ||
126 | +#endif | ||
127 | + | ||
128 | +/* Same test with pthread mutexes, for performance comparison and | ||
129 | + * portability. */ | ||
130 | + | ||
131 | +static QemuMutex mutex; | ||
132 | + | ||
133 | +static void test_multi_mutex_entry(void *opaque) | ||
134 | +{ | ||
135 | + while (!atomic_mb_read(&now_stopping)) { | ||
136 | + qemu_mutex_lock(&mutex); | ||
137 | + counter++; | ||
138 | + qemu_mutex_unlock(&mutex); | ||
139 | + atomic_inc(&atomic_counter); | ||
140 | + } | ||
141 | + atomic_dec(&running); | ||
142 | +} | ||
143 | + | ||
144 | +static void test_multi_mutex(int threads, int seconds) | ||
145 | +{ | ||
146 | + int i; | ||
147 | + | ||
148 | + qemu_mutex_init(&mutex); | ||
149 | + counter = 0; | ||
150 | + atomic_counter = 0; | ||
151 | + now_stopping = false; | ||
152 | + | ||
153 | + create_aio_contexts(); | ||
154 | + assert(threads <= NUM_CONTEXTS); | ||
155 | + running = threads; | ||
156 | + for (i = 0; i < threads; i++) { | ||
157 | + Coroutine *co1 = qemu_coroutine_create(test_multi_mutex_entry, NULL); | ||
158 | + aio_co_schedule(ctx[i], co1); | ||
159 | + } | ||
160 | + | ||
161 | + g_usleep(seconds * 1000000); | ||
162 | + | ||
163 | + atomic_mb_set(&now_stopping, true); | ||
164 | + while (running > 0) { | ||
165 | + g_usleep(100000); | ||
166 | + } | ||
167 | + | ||
168 | + join_aio_contexts(); | ||
169 | + g_test_message("%d iterations/second\n", counter / seconds); | ||
170 | + g_assert_cmpint(counter, ==, atomic_counter); | ||
171 | +} | ||
172 | + | ||
173 | +static void test_multi_mutex_1(void) | ||
174 | +{ | ||
175 | + test_multi_mutex(NUM_CONTEXTS, 1); | ||
176 | +} | ||
177 | + | ||
178 | +static void test_multi_mutex_10(void) | ||
179 | +{ | ||
180 | + test_multi_mutex(NUM_CONTEXTS, 10); | ||
181 | +} | ||
182 | + | ||
183 | /* End of tests. */ | ||
184 | |||
185 | int main(int argc, char **argv) | ||
186 | @@ -XXX,XX +XXX,XX @@ int main(int argc, char **argv) | ||
187 | g_test_add_func("/aio/multi/schedule", test_multi_co_schedule_1); | ||
188 | g_test_add_func("/aio/multi/mutex/contended", test_multi_co_mutex_1); | ||
189 | g_test_add_func("/aio/multi/mutex/handoff", test_multi_co_mutex_2_3); | ||
190 | +#ifdef CONFIG_LINUX | ||
191 | + g_test_add_func("/aio/multi/mutex/mcs", test_multi_fair_mutex_1); | ||
192 | +#endif | ||
193 | + g_test_add_func("/aio/multi/mutex/pthread", test_multi_mutex_1); | ||
194 | } else { | ||
195 | g_test_add_func("/aio/multi/schedule", test_multi_co_schedule_10); | ||
196 | g_test_add_func("/aio/multi/mutex/contended", test_multi_co_mutex_10); | ||
197 | g_test_add_func("/aio/multi/mutex/handoff", test_multi_co_mutex_2_30); | ||
198 | +#ifdef CONFIG_LINUX | ||
199 | + g_test_add_func("/aio/multi/mutex/mcs", test_multi_fair_mutex_10); | ||
200 | +#endif | ||
201 | + g_test_add_func("/aio/multi/mutex/pthread", test_multi_mutex_10); | ||
202 | } | ||
203 | return g_test_run(); | ||
204 | } | ||
205 | -- | ||
206 | 2.9.3 | ||
207 | |||
208 | diff view generated by jsdifflib |
1 | From: Fred Rolland <rollandf@gmail.com> | 1 | From: Paolo Bonzini <pbonzini@redhat.com> |
---|---|---|---|
2 | 2 | ||
3 | Update doc with the usage of UUID for initiator name. | 3 | This will avoid forward references in the next patch. It is also |
4 | more logical because CoQueue is not anymore the basic primitive. | ||
4 | 5 | ||
5 | Related-To: https://bugzilla.redhat.com/1006468 | 6 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> |
6 | Signed-off-by: Fred Rolland <frolland@redhat.com> | 7 | Reviewed-by: Fam Zheng <famz@redhat.com> |
7 | Message-id: 20170823084830.30500-1-frolland@redhat.com | 8 | Message-id: 20170213181244.16297-5-pbonzini@redhat.com |
8 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | 9 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
9 | --- | 10 | --- |
10 | qemu-doc.texi | 5 +++-- | 11 | include/qemu/coroutine.h | 89 ++++++++++++++++++++++++------------------------ |
11 | 1 file changed, 3 insertions(+), 2 deletions(-) | 12 | 1 file changed, 44 insertions(+), 45 deletions(-) |
12 | 13 | ||
13 | diff --git a/qemu-doc.texi b/qemu-doc.texi | 14 | diff --git a/include/qemu/coroutine.h b/include/qemu/coroutine.h |
14 | index XXXXXXX..XXXXXXX 100644 | 15 | index XXXXXXX..XXXXXXX 100644 |
15 | --- a/qemu-doc.texi | 16 | --- a/include/qemu/coroutine.h |
16 | +++ b/qemu-doc.texi | 17 | +++ b/include/qemu/coroutine.h |
17 | @@ -XXX,XX +XXX,XX @@ in a configuration file provided via '-readconfig' or directly on the | 18 | @@ -XXX,XX +XXX,XX @@ bool qemu_in_coroutine(void); |
18 | command line. | 19 | */ |
19 | 20 | bool qemu_coroutine_entered(Coroutine *co); | |
20 | If the initiator-name is not specified qemu will use a default name | ||
21 | -of 'iqn.2008-11.org.linux-kvm[:<name>'] where <name> is the name of the | ||
22 | +of 'iqn.2008-11.org.linux-kvm[:<uuid>'] where <uuid> is the UUID of the | ||
23 | +virtual machine. If the UUID is not specified qemu will use | ||
24 | +'iqn.2008-11.org.linux-kvm[:<name>'] where <name> is the name of the | ||
25 | virtual machine. | ||
26 | 21 | ||
27 | - | 22 | - |
28 | @example | 23 | -/** |
29 | Setting a specific initiator name to use when logging in to the target | 24 | - * CoQueues are a mechanism to queue coroutines in order to continue executing |
30 | -iscsi initiator-name=iqn.qemu.test:my-initiator | 25 | - * them later. They provide the fundamental primitives on which coroutine locks |
26 | - * are built. | ||
27 | - */ | ||
28 | -typedef struct CoQueue { | ||
29 | - QSIMPLEQ_HEAD(, Coroutine) entries; | ||
30 | -} CoQueue; | ||
31 | - | ||
32 | -/** | ||
33 | - * Initialise a CoQueue. This must be called before any other operation is used | ||
34 | - * on the CoQueue. | ||
35 | - */ | ||
36 | -void qemu_co_queue_init(CoQueue *queue); | ||
37 | - | ||
38 | -/** | ||
39 | - * Adds the current coroutine to the CoQueue and transfers control to the | ||
40 | - * caller of the coroutine. | ||
41 | - */ | ||
42 | -void coroutine_fn qemu_co_queue_wait(CoQueue *queue); | ||
43 | - | ||
44 | -/** | ||
45 | - * Restarts the next coroutine in the CoQueue and removes it from the queue. | ||
46 | - * | ||
47 | - * Returns true if a coroutine was restarted, false if the queue is empty. | ||
48 | - */ | ||
49 | -bool coroutine_fn qemu_co_queue_next(CoQueue *queue); | ||
50 | - | ||
51 | -/** | ||
52 | - * Restarts all coroutines in the CoQueue and leaves the queue empty. | ||
53 | - */ | ||
54 | -void coroutine_fn qemu_co_queue_restart_all(CoQueue *queue); | ||
55 | - | ||
56 | -/** | ||
57 | - * Enter the next coroutine in the queue | ||
58 | - */ | ||
59 | -bool qemu_co_enter_next(CoQueue *queue); | ||
60 | - | ||
61 | -/** | ||
62 | - * Checks if the CoQueue is empty. | ||
63 | - */ | ||
64 | -bool qemu_co_queue_empty(CoQueue *queue); | ||
65 | - | ||
66 | - | ||
67 | /** | ||
68 | * Provides a mutex that can be used to synchronise coroutines | ||
69 | */ | ||
70 | @@ -XXX,XX +XXX,XX @@ void coroutine_fn qemu_co_mutex_lock(CoMutex *mutex); | ||
71 | */ | ||
72 | void coroutine_fn qemu_co_mutex_unlock(CoMutex *mutex); | ||
73 | |||
74 | + | ||
75 | +/** | ||
76 | + * CoQueues are a mechanism to queue coroutines in order to continue executing | ||
77 | + * them later. | ||
78 | + */ | ||
79 | +typedef struct CoQueue { | ||
80 | + QSIMPLEQ_HEAD(, Coroutine) entries; | ||
81 | +} CoQueue; | ||
82 | + | ||
83 | +/** | ||
84 | + * Initialise a CoQueue. This must be called before any other operation is used | ||
85 | + * on the CoQueue. | ||
86 | + */ | ||
87 | +void qemu_co_queue_init(CoQueue *queue); | ||
88 | + | ||
89 | +/** | ||
90 | + * Adds the current coroutine to the CoQueue and transfers control to the | ||
91 | + * caller of the coroutine. | ||
92 | + */ | ||
93 | +void coroutine_fn qemu_co_queue_wait(CoQueue *queue); | ||
94 | + | ||
95 | +/** | ||
96 | + * Restarts the next coroutine in the CoQueue and removes it from the queue. | ||
97 | + * | ||
98 | + * Returns true if a coroutine was restarted, false if the queue is empty. | ||
99 | + */ | ||
100 | +bool coroutine_fn qemu_co_queue_next(CoQueue *queue); | ||
101 | + | ||
102 | +/** | ||
103 | + * Restarts all coroutines in the CoQueue and leaves the queue empty. | ||
104 | + */ | ||
105 | +void coroutine_fn qemu_co_queue_restart_all(CoQueue *queue); | ||
106 | + | ||
107 | +/** | ||
108 | + * Enter the next coroutine in the queue | ||
109 | + */ | ||
110 | +bool qemu_co_enter_next(CoQueue *queue); | ||
111 | + | ||
112 | +/** | ||
113 | + * Checks if the CoQueue is empty. | ||
114 | + */ | ||
115 | +bool qemu_co_queue_empty(CoQueue *queue); | ||
116 | + | ||
117 | + | ||
118 | typedef struct CoRwlock { | ||
119 | bool writer; | ||
120 | int reader; | ||
31 | -- | 121 | -- |
32 | 2.13.5 | 122 | 2.9.3 |
33 | 123 | ||
34 | 124 | diff view generated by jsdifflib |
1 | Most qcow2 files are uncompressed so it is wasteful to allocate (32 + 1) | 1 | From: Paolo Bonzini <pbonzini@redhat.com> |
---|---|---|---|
2 | * cluster_size + 512 bytes upfront. Allocate s->cluster_cache and | 2 | |
3 | s->cluster_data when the first read operation is performance on a | 3 | All that CoQueue needs in order to become thread-safe is help |
4 | compressed cluster. | 4 | from an external mutex. Add this to the API. |
5 | 5 | ||
6 | The buffers are freed in .bdrv_close(). .bdrv_open() no longer has any | 6 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> |
7 | code paths that can allocate these buffers, so remove the free functions | 7 | Reviewed-by: Fam Zheng <famz@redhat.com> |
8 | in the error code path. | 8 | Message-id: 20170213181244.16297-6-pbonzini@redhat.com |
9 | |||
10 | This patch can result in significant memory savings when many qcow2 | ||
11 | disks are attached or backing file chains are long: | ||
12 | |||
13 | Before 12.81% (1,023,193,088B) | ||
14 | After 5.36% (393,893,888B) | ||
15 | |||
16 | Reported-by: Alexey Kardashevskiy <aik@ozlabs.ru> | ||
17 | Tested-by: Alexey Kardashevskiy <aik@ozlabs.ru> | ||
18 | Reviewed-by: Eric Blake <eblake@redhat.com> | ||
19 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | ||
20 | Message-id: 20170821135530.32344-1-stefanha@redhat.com | ||
21 | Cc: Kevin Wolf <kwolf@redhat.com> | ||
22 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | 9 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
23 | --- | 10 | --- |
24 | block/qcow2-cluster.c | 17 +++++++++++++++++ | 11 | include/qemu/coroutine.h | 8 +++++--- |
25 | block/qcow2.c | 12 ------------ | 12 | block/backup.c | 2 +- |
26 | 2 files changed, 17 insertions(+), 12 deletions(-) | 13 | block/io.c | 4 ++-- |
27 | 14 | block/nbd-client.c | 2 +- | |
15 | block/qcow2-cluster.c | 4 +--- | ||
16 | block/sheepdog.c | 2 +- | ||
17 | block/throttle-groups.c | 2 +- | ||
18 | hw/9pfs/9p.c | 2 +- | ||
19 | util/qemu-coroutine-lock.c | 24 +++++++++++++++++++++--- | ||
20 | 9 files changed, 34 insertions(+), 16 deletions(-) | ||
21 | |||
22 | diff --git a/include/qemu/coroutine.h b/include/qemu/coroutine.h | ||
23 | index XXXXXXX..XXXXXXX 100644 | ||
24 | --- a/include/qemu/coroutine.h | ||
25 | +++ b/include/qemu/coroutine.h | ||
26 | @@ -XXX,XX +XXX,XX @@ void coroutine_fn qemu_co_mutex_unlock(CoMutex *mutex); | ||
27 | |||
28 | /** | ||
29 | * CoQueues are a mechanism to queue coroutines in order to continue executing | ||
30 | - * them later. | ||
31 | + * them later. They are similar to condition variables, but they need help | ||
32 | + * from an external mutex in order to maintain thread-safety. | ||
33 | */ | ||
34 | typedef struct CoQueue { | ||
35 | QSIMPLEQ_HEAD(, Coroutine) entries; | ||
36 | @@ -XXX,XX +XXX,XX @@ void qemu_co_queue_init(CoQueue *queue); | ||
37 | |||
38 | /** | ||
39 | * Adds the current coroutine to the CoQueue and transfers control to the | ||
40 | - * caller of the coroutine. | ||
41 | + * caller of the coroutine. The mutex is unlocked during the wait and | ||
42 | + * locked again afterwards. | ||
43 | */ | ||
44 | -void coroutine_fn qemu_co_queue_wait(CoQueue *queue); | ||
45 | +void coroutine_fn qemu_co_queue_wait(CoQueue *queue, CoMutex *mutex); | ||
46 | |||
47 | /** | ||
48 | * Restarts the next coroutine in the CoQueue and removes it from the queue. | ||
49 | diff --git a/block/backup.c b/block/backup.c | ||
50 | index XXXXXXX..XXXXXXX 100644 | ||
51 | --- a/block/backup.c | ||
52 | +++ b/block/backup.c | ||
53 | @@ -XXX,XX +XXX,XX @@ static void coroutine_fn wait_for_overlapping_requests(BackupBlockJob *job, | ||
54 | retry = false; | ||
55 | QLIST_FOREACH(req, &job->inflight_reqs, list) { | ||
56 | if (end > req->start && start < req->end) { | ||
57 | - qemu_co_queue_wait(&req->wait_queue); | ||
58 | + qemu_co_queue_wait(&req->wait_queue, NULL); | ||
59 | retry = true; | ||
60 | break; | ||
61 | } | ||
62 | diff --git a/block/io.c b/block/io.c | ||
63 | index XXXXXXX..XXXXXXX 100644 | ||
64 | --- a/block/io.c | ||
65 | +++ b/block/io.c | ||
66 | @@ -XXX,XX +XXX,XX @@ static bool coroutine_fn wait_serialising_requests(BdrvTrackedRequest *self) | ||
67 | * (instead of producing a deadlock in the former case). */ | ||
68 | if (!req->waiting_for) { | ||
69 | self->waiting_for = req; | ||
70 | - qemu_co_queue_wait(&req->wait_queue); | ||
71 | + qemu_co_queue_wait(&req->wait_queue, NULL); | ||
72 | self->waiting_for = NULL; | ||
73 | retry = true; | ||
74 | waited = true; | ||
75 | @@ -XXX,XX +XXX,XX @@ int coroutine_fn bdrv_co_flush(BlockDriverState *bs) | ||
76 | |||
77 | /* Wait until any previous flushes are completed */ | ||
78 | while (bs->active_flush_req) { | ||
79 | - qemu_co_queue_wait(&bs->flush_queue); | ||
80 | + qemu_co_queue_wait(&bs->flush_queue, NULL); | ||
81 | } | ||
82 | |||
83 | bs->active_flush_req = true; | ||
84 | diff --git a/block/nbd-client.c b/block/nbd-client.c | ||
85 | index XXXXXXX..XXXXXXX 100644 | ||
86 | --- a/block/nbd-client.c | ||
87 | +++ b/block/nbd-client.c | ||
88 | @@ -XXX,XX +XXX,XX @@ static void nbd_coroutine_start(NBDClientSession *s, | ||
89 | /* Poor man semaphore. The free_sema is locked when no other request | ||
90 | * can be accepted, and unlocked after receiving one reply. */ | ||
91 | if (s->in_flight == MAX_NBD_REQUESTS) { | ||
92 | - qemu_co_queue_wait(&s->free_sema); | ||
93 | + qemu_co_queue_wait(&s->free_sema, NULL); | ||
94 | assert(s->in_flight < MAX_NBD_REQUESTS); | ||
95 | } | ||
96 | s->in_flight++; | ||
28 | diff --git a/block/qcow2-cluster.c b/block/qcow2-cluster.c | 97 | diff --git a/block/qcow2-cluster.c b/block/qcow2-cluster.c |
29 | index XXXXXXX..XXXXXXX 100644 | 98 | index XXXXXXX..XXXXXXX 100644 |
30 | --- a/block/qcow2-cluster.c | 99 | --- a/block/qcow2-cluster.c |
31 | +++ b/block/qcow2-cluster.c | 100 | +++ b/block/qcow2-cluster.c |
32 | @@ -XXX,XX +XXX,XX @@ int qcow2_decompress_cluster(BlockDriverState *bs, uint64_t cluster_offset) | 101 | @@ -XXX,XX +XXX,XX @@ static int handle_dependencies(BlockDriverState *bs, uint64_t guest_offset, |
33 | nb_csectors = ((cluster_offset >> s->csize_shift) & s->csize_mask) + 1; | 102 | if (bytes == 0) { |
34 | sector_offset = coffset & 511; | 103 | /* Wait for the dependency to complete. We need to recheck |
35 | csize = nb_csectors * 512 - sector_offset; | 104 | * the free/allocated clusters when we continue. */ |
105 | - qemu_co_mutex_unlock(&s->lock); | ||
106 | - qemu_co_queue_wait(&old_alloc->dependent_requests); | ||
107 | - qemu_co_mutex_lock(&s->lock); | ||
108 | + qemu_co_queue_wait(&old_alloc->dependent_requests, &s->lock); | ||
109 | return -EAGAIN; | ||
110 | } | ||
111 | } | ||
112 | diff --git a/block/sheepdog.c b/block/sheepdog.c | ||
113 | index XXXXXXX..XXXXXXX 100644 | ||
114 | --- a/block/sheepdog.c | ||
115 | +++ b/block/sheepdog.c | ||
116 | @@ -XXX,XX +XXX,XX @@ static void wait_for_overlapping_aiocb(BDRVSheepdogState *s, SheepdogAIOCB *acb) | ||
117 | retry: | ||
118 | QLIST_FOREACH(cb, &s->inflight_aiocb_head, aiocb_siblings) { | ||
119 | if (AIOCBOverlapping(acb, cb)) { | ||
120 | - qemu_co_queue_wait(&s->overlapping_queue); | ||
121 | + qemu_co_queue_wait(&s->overlapping_queue, NULL); | ||
122 | goto retry; | ||
123 | } | ||
124 | } | ||
125 | diff --git a/block/throttle-groups.c b/block/throttle-groups.c | ||
126 | index XXXXXXX..XXXXXXX 100644 | ||
127 | --- a/block/throttle-groups.c | ||
128 | +++ b/block/throttle-groups.c | ||
129 | @@ -XXX,XX +XXX,XX @@ void coroutine_fn throttle_group_co_io_limits_intercept(BlockBackend *blk, | ||
130 | if (must_wait || blkp->pending_reqs[is_write]) { | ||
131 | blkp->pending_reqs[is_write]++; | ||
132 | qemu_mutex_unlock(&tg->lock); | ||
133 | - qemu_co_queue_wait(&blkp->throttled_reqs[is_write]); | ||
134 | + qemu_co_queue_wait(&blkp->throttled_reqs[is_write], NULL); | ||
135 | qemu_mutex_lock(&tg->lock); | ||
136 | blkp->pending_reqs[is_write]--; | ||
137 | } | ||
138 | diff --git a/hw/9pfs/9p.c b/hw/9pfs/9p.c | ||
139 | index XXXXXXX..XXXXXXX 100644 | ||
140 | --- a/hw/9pfs/9p.c | ||
141 | +++ b/hw/9pfs/9p.c | ||
142 | @@ -XXX,XX +XXX,XX @@ static void coroutine_fn v9fs_flush(void *opaque) | ||
143 | /* | ||
144 | * Wait for pdu to complete. | ||
145 | */ | ||
146 | - qemu_co_queue_wait(&cancel_pdu->complete); | ||
147 | + qemu_co_queue_wait(&cancel_pdu->complete, NULL); | ||
148 | cancel_pdu->cancelled = 0; | ||
149 | pdu_free(cancel_pdu); | ||
150 | } | ||
151 | diff --git a/util/qemu-coroutine-lock.c b/util/qemu-coroutine-lock.c | ||
152 | index XXXXXXX..XXXXXXX 100644 | ||
153 | --- a/util/qemu-coroutine-lock.c | ||
154 | +++ b/util/qemu-coroutine-lock.c | ||
155 | @@ -XXX,XX +XXX,XX @@ void qemu_co_queue_init(CoQueue *queue) | ||
156 | QSIMPLEQ_INIT(&queue->entries); | ||
157 | } | ||
158 | |||
159 | -void coroutine_fn qemu_co_queue_wait(CoQueue *queue) | ||
160 | +void coroutine_fn qemu_co_queue_wait(CoQueue *queue, CoMutex *mutex) | ||
161 | { | ||
162 | Coroutine *self = qemu_coroutine_self(); | ||
163 | QSIMPLEQ_INSERT_TAIL(&queue->entries, self, co_queue_next); | ||
36 | + | 164 | + |
37 | + /* Allocate buffers on first decompress operation, most images are | 165 | + if (mutex) { |
38 | + * uncompressed and the memory overhead can be avoided. The buffers | 166 | + qemu_co_mutex_unlock(mutex); |
39 | + * are freed in .bdrv_close(). | 167 | + } |
40 | + */ | ||
41 | + if (!s->cluster_data) { | ||
42 | + /* one more sector for decompressed data alignment */ | ||
43 | + s->cluster_data = qemu_try_blockalign(bs->file->bs, | ||
44 | + QCOW_MAX_CRYPT_CLUSTERS * s->cluster_size + 512); | ||
45 | + if (!s->cluster_data) { | ||
46 | + return -ENOMEM; | ||
47 | + } | ||
48 | + } | ||
49 | + if (!s->cluster_cache) { | ||
50 | + s->cluster_cache = g_malloc(s->cluster_size); | ||
51 | + } | ||
52 | + | 168 | + |
53 | BLKDBG_EVENT(bs->file, BLKDBG_READ_COMPRESSED); | 169 | + /* There is no race condition here. Other threads will call |
54 | ret = bdrv_read(bs->file, coffset >> 9, s->cluster_data, | 170 | + * aio_co_schedule on our AioContext, which can reenter this |
55 | nb_csectors); | 171 | + * coroutine but only after this yield and after the main loop |
56 | diff --git a/block/qcow2.c b/block/qcow2.c | 172 | + * has gone through the next iteration. |
57 | index XXXXXXX..XXXXXXX 100644 | 173 | + */ |
58 | --- a/block/qcow2.c | 174 | qemu_coroutine_yield(); |
59 | +++ b/block/qcow2.c | 175 | assert(qemu_in_coroutine()); |
60 | @@ -XXX,XX +XXX,XX @@ static int qcow2_do_open(BlockDriverState *bs, QDict *options, int flags, | 176 | + |
61 | goto fail; | 177 | + /* TODO: OSv implements wait morphing here, where the wakeup |
62 | } | 178 | + * primitive automatically places the woken coroutine on the |
63 | 179 | + * mutex's queue. This avoids the thundering herd effect. | |
64 | - s->cluster_cache = g_malloc(s->cluster_size); | 180 | + */ |
65 | - /* one more sector for decompressed data alignment */ | 181 | + if (mutex) { |
66 | - s->cluster_data = qemu_try_blockalign(bs->file->bs, QCOW_MAX_CRYPT_CLUSTERS | 182 | + qemu_co_mutex_lock(mutex); |
67 | - * s->cluster_size + 512); | 183 | + } |
68 | - if (s->cluster_data == NULL) { | 184 | } |
69 | - error_setg(errp, "Could not allocate temporary cluster buffer"); | 185 | |
70 | - ret = -ENOMEM; | 186 | /** |
71 | - goto fail; | 187 | @@ -XXX,XX +XXX,XX @@ void qemu_co_rwlock_rdlock(CoRwlock *lock) |
72 | - } | 188 | Coroutine *self = qemu_coroutine_self(); |
73 | - | 189 | |
74 | s->cluster_cache_offset = -1; | 190 | while (lock->writer) { |
75 | s->flags = flags; | 191 | - qemu_co_queue_wait(&lock->queue); |
76 | 192 | + qemu_co_queue_wait(&lock->queue, NULL); | |
77 | @@ -XXX,XX +XXX,XX @@ static int qcow2_do_open(BlockDriverState *bs, QDict *options, int flags, | 193 | } |
78 | if (s->refcount_block_cache) { | 194 | lock->reader++; |
79 | qcow2_cache_destroy(bs, s->refcount_block_cache); | 195 | self->locks_held++; |
80 | } | 196 | @@ -XXX,XX +XXX,XX @@ void qemu_co_rwlock_wrlock(CoRwlock *lock) |
81 | - g_free(s->cluster_cache); | 197 | Coroutine *self = qemu_coroutine_self(); |
82 | - qemu_vfree(s->cluster_data); | 198 | |
83 | qcrypto_block_free(s->crypto); | 199 | while (lock->writer || lock->reader) { |
84 | qapi_free_QCryptoBlockOpenOptions(s->crypto_opts); | 200 | - qemu_co_queue_wait(&lock->queue); |
85 | return ret; | 201 | + qemu_co_queue_wait(&lock->queue, NULL); |
202 | } | ||
203 | lock->writer = true; | ||
204 | self->locks_held++; | ||
86 | -- | 205 | -- |
87 | 2.13.5 | 206 | 2.9.3 |
88 | 207 | ||
89 | 208 | diff view generated by jsdifflib |
1 | From: Dan Aloni <dan@kernelim.com> | 1 | From: Paolo Bonzini <pbonzini@redhat.com> |
---|---|---|---|
2 | 2 | ||
3 | The number of queues that should be return by the admin command should: | 3 | This adds a CoMutex around the existing CoQueue. Because the write-side |
4 | can just take CoMutex, the old "writer" field is not necessary anymore. | ||
5 | Instead of removing it altogether, count the number of pending writers | ||
6 | during a read-side critical section and forbid further readers from | ||
7 | entering. | ||
4 | 8 | ||
5 | 1) Only mention the number of non-admin queues. | 9 | Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> |
6 | 2) It is zero-based, meaning that '0 == one non-admin queue', | 10 | Reviewed-by: Fam Zheng <famz@redhat.com> |
7 | '1 == two non-admin queues', and so forth. | 11 | Message-id: 20170213181244.16297-7-pbonzini@redhat.com |
8 | |||
9 | Because our `num_queues` means the number of queues _plus_ the admin | ||
10 | queue, then the right calculation for the number returned from the admin | ||
11 | command is `num_queues - 2`, combining the two requirements mentioned. | ||
12 | |||
13 | The issue was discovered by reducing num_queues from 64 to 8 and running | ||
14 | a Linux VM with an SMP parameter larger than that (e.g. 22). It tries to | ||
15 | utilize all queues, and therefore fails with an invalid queue number | ||
16 | when trying to queue I/Os on the last queue. | ||
17 | |||
18 | Signed-off-by: Dan Aloni <dan@kernelim.com> | ||
19 | CC: Alex Friedman <alex@e8storage.com> | ||
20 | CC: Keith Busch <keith.busch@intel.com> | ||
21 | CC: Stefan Hajnoczi <stefanha@redhat.com> | ||
22 | Reviewed-by: Keith Busch <keith.busch@intel.com> | ||
23 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | 12 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
24 | --- | 13 | --- |
25 | hw/block/nvme.c | 4 ++-- | 14 | include/qemu/coroutine.h | 3 ++- |
26 | 1 file changed, 2 insertions(+), 2 deletions(-) | 15 | util/qemu-coroutine-lock.c | 35 ++++++++++++++++++++++++----------- |
16 | 2 files changed, 26 insertions(+), 12 deletions(-) | ||
27 | 17 | ||
28 | diff --git a/hw/block/nvme.c b/hw/block/nvme.c | 18 | diff --git a/include/qemu/coroutine.h b/include/qemu/coroutine.h |
29 | index XXXXXXX..XXXXXXX 100644 | 19 | index XXXXXXX..XXXXXXX 100644 |
30 | --- a/hw/block/nvme.c | 20 | --- a/include/qemu/coroutine.h |
31 | +++ b/hw/block/nvme.c | 21 | +++ b/include/qemu/coroutine.h |
32 | @@ -XXX,XX +XXX,XX @@ static uint16_t nvme_get_feature(NvmeCtrl *n, NvmeCmd *cmd, NvmeRequest *req) | 22 | @@ -XXX,XX +XXX,XX @@ bool qemu_co_queue_empty(CoQueue *queue); |
33 | result = blk_enable_write_cache(n->conf.blk); | 23 | |
34 | break; | 24 | |
35 | case NVME_NUMBER_OF_QUEUES: | 25 | typedef struct CoRwlock { |
36 | - result = cpu_to_le32((n->num_queues - 1) | ((n->num_queues - 1) << 16)); | 26 | - bool writer; |
37 | + result = cpu_to_le32((n->num_queues - 2) | ((n->num_queues - 2) << 16)); | 27 | + int pending_writer; |
38 | break; | 28 | int reader; |
39 | default: | 29 | + CoMutex mutex; |
40 | return NVME_INVALID_FIELD | NVME_DNR; | 30 | CoQueue queue; |
41 | @@ -XXX,XX +XXX,XX @@ static uint16_t nvme_set_feature(NvmeCtrl *n, NvmeCmd *cmd, NvmeRequest *req) | 31 | } CoRwlock; |
42 | break; | 32 | |
43 | case NVME_NUMBER_OF_QUEUES: | 33 | diff --git a/util/qemu-coroutine-lock.c b/util/qemu-coroutine-lock.c |
44 | req->cqe.result = | 34 | index XXXXXXX..XXXXXXX 100644 |
45 | - cpu_to_le32((n->num_queues - 1) | ((n->num_queues - 1) << 16)); | 35 | --- a/util/qemu-coroutine-lock.c |
46 | + cpu_to_le32((n->num_queues - 2) | ((n->num_queues - 2) << 16)); | 36 | +++ b/util/qemu-coroutine-lock.c |
47 | break; | 37 | @@ -XXX,XX +XXX,XX @@ void qemu_co_rwlock_init(CoRwlock *lock) |
48 | default: | 38 | { |
49 | return NVME_INVALID_FIELD | NVME_DNR; | 39 | memset(lock, 0, sizeof(*lock)); |
40 | qemu_co_queue_init(&lock->queue); | ||
41 | + qemu_co_mutex_init(&lock->mutex); | ||
42 | } | ||
43 | |||
44 | void qemu_co_rwlock_rdlock(CoRwlock *lock) | ||
45 | { | ||
46 | Coroutine *self = qemu_coroutine_self(); | ||
47 | |||
48 | - while (lock->writer) { | ||
49 | - qemu_co_queue_wait(&lock->queue, NULL); | ||
50 | + qemu_co_mutex_lock(&lock->mutex); | ||
51 | + /* For fairness, wait if a writer is in line. */ | ||
52 | + while (lock->pending_writer) { | ||
53 | + qemu_co_queue_wait(&lock->queue, &lock->mutex); | ||
54 | } | ||
55 | lock->reader++; | ||
56 | + qemu_co_mutex_unlock(&lock->mutex); | ||
57 | + | ||
58 | + /* The rest of the read-side critical section is run without the mutex. */ | ||
59 | self->locks_held++; | ||
60 | } | ||
61 | |||
62 | @@ -XXX,XX +XXX,XX @@ void qemu_co_rwlock_unlock(CoRwlock *lock) | ||
63 | Coroutine *self = qemu_coroutine_self(); | ||
64 | |||
65 | assert(qemu_in_coroutine()); | ||
66 | - if (lock->writer) { | ||
67 | - lock->writer = false; | ||
68 | + if (!lock->reader) { | ||
69 | + /* The critical section started in qemu_co_rwlock_wrlock. */ | ||
70 | qemu_co_queue_restart_all(&lock->queue); | ||
71 | } else { | ||
72 | + self->locks_held--; | ||
73 | + | ||
74 | + qemu_co_mutex_lock(&lock->mutex); | ||
75 | lock->reader--; | ||
76 | assert(lock->reader >= 0); | ||
77 | /* Wakeup only one waiting writer */ | ||
78 | @@ -XXX,XX +XXX,XX @@ void qemu_co_rwlock_unlock(CoRwlock *lock) | ||
79 | qemu_co_queue_next(&lock->queue); | ||
80 | } | ||
81 | } | ||
82 | - self->locks_held--; | ||
83 | + qemu_co_mutex_unlock(&lock->mutex); | ||
84 | } | ||
85 | |||
86 | void qemu_co_rwlock_wrlock(CoRwlock *lock) | ||
87 | { | ||
88 | - Coroutine *self = qemu_coroutine_self(); | ||
89 | - | ||
90 | - while (lock->writer || lock->reader) { | ||
91 | - qemu_co_queue_wait(&lock->queue, NULL); | ||
92 | + qemu_co_mutex_lock(&lock->mutex); | ||
93 | + lock->pending_writer++; | ||
94 | + while (lock->reader) { | ||
95 | + qemu_co_queue_wait(&lock->queue, &lock->mutex); | ||
96 | } | ||
97 | - lock->writer = true; | ||
98 | - self->locks_held++; | ||
99 | + lock->pending_writer--; | ||
100 | + | ||
101 | + /* The rest of the write-side critical section is run with | ||
102 | + * the mutex taken, so that lock->reader remains zero. | ||
103 | + * There is no need to update self->locks_held. | ||
104 | + */ | ||
105 | } | ||
50 | -- | 106 | -- |
51 | 2.13.5 | 107 | 2.9.3 |
52 | 108 | ||
53 | 109 | diff view generated by jsdifflib |