1 | The following changes since commit ea6abffa8a08d832feb759d359d5b935e3087cf7: | 1 | The following changes since commit afc9fcde55296b83f659de9da3cdf044812a6eeb: |
---|---|---|---|
2 | 2 | ||
3 | Update version for v3.0.0-rc1 release (2018-07-17 18:15:19 +0100) | 3 | Merge remote-tracking branch 'remotes/mst/tags/for_upstream' into staging (2021-10-20 06:10:51 -0700) |
4 | 4 | ||
5 | are available in the Git repository at: | 5 | are available in the Git repository at: |
6 | 6 | ||
7 | git://github.com/stefanha/qemu.git tags/block-pull-request | 7 | https://gitlab.com/stefanha/qemu.git tags/block-pull-request |
8 | 8 | ||
9 | for you to fetch changes up to 6fccbb475bc6effc313ee9481726a1748b6dae57: | 9 | for you to fetch changes up to 4b2b3d2653f255ef4259a7689af1956536565901: |
10 | 10 | ||
11 | throttle-groups: fix hang when group member leaves (2018-07-19 13:08:26 +0100) | 11 | coroutine: resize pool periodically instead of limiting size (2021-10-21 18:40:07 +0100) |
12 | 12 | ||
13 | ---------------------------------------------------------------- | 13 | ---------------------------------------------------------------- |
14 | Pull request | 14 | Pull request |
15 | 15 | ||
16 | This fix prevents hangs when a drive leaves a throttling group. | 16 | Performance optimization when guest applications submit a lot of parallel I/O. |
17 | This has also been found to improve clang SafeStack performance. | ||
17 | 18 | ||
18 | ---------------------------------------------------------------- | 19 | ---------------------------------------------------------------- |
19 | 20 | ||
20 | Stefan Hajnoczi (1): | 21 | Stefan Hajnoczi (1): |
21 | throttle-groups: fix hang when group member leaves | 22 | coroutine: resize pool periodically instead of limiting size |
22 | 23 | ||
23 | block/throttle-groups.c | 4 ++++ | 24 | include/qemu/coroutine-pool-timer.h | 36 ++++++++++++++++ |
24 | 1 file changed, 4 insertions(+) | 25 | include/qemu/coroutine.h | 7 ++++ |
26 | iothread.c | 6 +++ | ||
27 | util/coroutine-pool-timer.c | 35 ++++++++++++++++ | ||
28 | util/main-loop.c | 5 +++ | ||
29 | util/qemu-coroutine.c | 64 ++++++++++++++++------------- | ||
30 | util/meson.build | 1 + | ||
31 | 7 files changed, 125 insertions(+), 29 deletions(-) | ||
32 | create mode 100644 include/qemu/coroutine-pool-timer.h | ||
33 | create mode 100644 util/coroutine-pool-timer.c | ||
25 | 34 | ||
26 | -- | 35 | -- |
27 | 2.17.1 | 36 | 2.31.1 |
28 | 37 | ||
29 | 38 | ||
39 | diff view generated by jsdifflib |
1 | Throttle groups consist of members sharing one throttling state | 1 | It was reported that enabling SafeStack reduces IOPS significantly |
---|---|---|---|
2 | (including bps/iops limits). Round-robin scheduling is used to ensure | 2 | (>25%) with the following fio benchmark on virtio-blk using a NVMe host |
3 | fairness. If a group member already has a timer pending then other | 3 | block device: |
4 | groups members do not schedule their own timers. The next group member | 4 | |
5 | will have its turn when the existing timer expires. | 5 | # fio --rw=randrw --bs=4k --iodepth=64 --runtime=1m --direct=1 \ |
6 | 6 | --filename=/dev/vdb --name=job1 --ioengine=libaio --thread \ | |
7 | A hang may occur when a group member leaves while it had a timer | 7 | --group_reporting --numjobs=16 --time_based \ |
8 | scheduled. Although the code carefully removes the group member from | 8 | --output=/tmp/fio_result |
9 | the round-robin list, it does not schedule the next member. Therefore | 9 | |
10 | remaining members continue to wait for the removed member's timer to | 10 | Serge Guelton and I found that SafeStack is not really at fault, it just |
11 | expire. | 11 | increases the cost of coroutine creation. This fio workload exhausts the |
12 | 12 | coroutine pool and coroutine creation becomes a bottleneck. Previous | |
13 | This patch schedules the next request if a timer is pending. | 13 | work by Honghao Wang also pointed to excessive coroutine creation. |
14 | Unfortunately the actual bug is a race condition that I've been unable | 14 | |
15 | to capture in a test case. | 15 | Creating new coroutines is expensive due to allocating new stacks with |
16 | 16 | mmap(2) and mprotect(2). Currently there are thread-local and global | |
17 | Sometimes drive2 hangs when drive1 is removed from the throttling group: | 17 | pools that recycle old Coroutine objects and their stacks but the |
18 | 18 | hardcoded size limit of 64 for thread-local pools and 128 for the global | |
19 | $ qemu ... -drive if=none,id=drive1,cache=none,format=qcow2,file=data1.qcow2,iops=100,group=foo \ | 19 | pool is insufficient for the fio benchmark shown above. |
20 | -device virtio-blk-pci,id=virtio-blk-pci0,drive=drive1 \ | 20 | |
21 | -drive if=none,id=drive2,cache=none,format=qcow2,file=data2.qcow2,iops=10,group=foo \ | 21 | This patch changes the coroutine pool algorithm to a simple thread-local |
22 | -device virtio-blk-pci,id=virtio-blk-pci1,drive=drive2 | 22 | pool without a maximum size limit. Threads periodically shrink the pool |
23 | (guest-console1)# fio -filename /dev/vda 4k-seq-read.job | 23 | down to a size sufficient for the maximum observed number of coroutines. |
24 | (guest-console2)# fio -filename /dev/vdb 4k-seq-read.job | 24 | |
25 | (qmp) {"execute": "block_set_io_throttle", "arguments": {"device": "drive1","bps": 0,"bps_rd": 0,"bps_wr": 0,"iops": 0,"iops_rd": 0,"iops_wr": 0}} | 25 | The global pool is removed by this patch. It can help to hide the fact |
26 | 26 | that local pools are easily exhausted, but it's doesn't fix the root | |
27 | Reported-by: Nini Gu <ngu@redhat.com> | 27 | cause. I don't think there is a need for a global pool because QEMU's |
28 | threads are long-lived, so let's keep things simple. | ||
29 | |||
30 | Performance of the above fio benchmark is as follows: | ||
31 | |||
32 | Before After | ||
33 | IOPS 60k 97k | ||
34 | |||
35 | Memory usage varies over time as needed by the workload: | ||
36 | |||
37 | VSZ (KB) RSS (KB) | ||
38 | Before fio 4705248 843128 | ||
39 | During fio 5747668 (+ ~100 MB) 849280 | ||
40 | After fio 4694996 (- ~100 MB) 845184 | ||
41 | |||
42 | This confirms that coroutines are indeed being freed when no longer | ||
43 | needed. | ||
44 | |||
45 | Thanks to Serge Guelton for working on identifying the bottleneck with | ||
46 | me! | ||
47 | |||
48 | Reported-by: Tingting Mao <timao@redhat.com> | ||
28 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | 49 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
29 | Message-id: 20180704145410.794-1-stefanha@redhat.com | 50 | Message-id: 20210913153524.1190696-1-stefanha@redhat.com |
30 | RHBZ: https://bugzilla.redhat.com/show_bug.cgi?id=1535914 | 51 | Cc: Serge Guelton <sguelton@redhat.com> |
31 | Cc: Alberto Garcia <berto@igalia.com> | 52 | Cc: Honghao Wang <wanghonghao@bytedance.com> |
53 | Cc: Paolo Bonzini <pbonzini@redhat.com> | ||
54 | Cc: Daniele Buono <dbuono@linux.vnet.ibm.com> | ||
32 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> | 55 | Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> |
56 | |||
57 | [Moved atexit notifier to coroutine_delete() after GitLab CI reported a | ||
58 | memory leak in tests/unit/test-aio-multithread because the Coroutine | ||
59 | object was created in the main thread but runs in an IOThread (where | ||
60 | it's also deleted). | ||
61 | --Stefan] | ||
33 | --- | 62 | --- |
34 | block/throttle-groups.c | 4 ++++ | 63 | include/qemu/coroutine-pool-timer.h | 36 ++++++++++++++++ |
35 | 1 file changed, 4 insertions(+) | 64 | include/qemu/coroutine.h | 7 ++++ |
36 | 65 | iothread.c | 6 +++ | |
37 | diff --git a/block/throttle-groups.c b/block/throttle-groups.c | 66 | util/coroutine-pool-timer.c | 35 ++++++++++++++++ |
38 | index XXXXXXX..XXXXXXX 100644 | 67 | util/main-loop.c | 5 +++ |
39 | --- a/block/throttle-groups.c | 68 | util/qemu-coroutine.c | 64 ++++++++++++++++------------- |
40 | +++ b/block/throttle-groups.c | 69 | util/meson.build | 1 + |
41 | @@ -XXX,XX +XXX,XX @@ void throttle_group_unregister_tgm(ThrottleGroupMember *tgm) | 70 | 7 files changed, 125 insertions(+), 29 deletions(-) |
42 | 71 | create mode 100644 include/qemu/coroutine-pool-timer.h | |
43 | qemu_mutex_lock(&tg->lock); | 72 | create mode 100644 util/coroutine-pool-timer.c |
44 | for (i = 0; i < 2; i++) { | 73 | |
45 | + if (timer_pending(tgm->throttle_timers.timers[i])) { | 74 | diff --git a/include/qemu/coroutine-pool-timer.h b/include/qemu/coroutine-pool-timer.h |
46 | + tg->any_timer_armed[i] = false; | 75 | new file mode 100644 |
47 | + schedule_next_request(tgm, i); | 76 | index XXXXXXX..XXXXXXX |
77 | --- /dev/null | ||
78 | +++ b/include/qemu/coroutine-pool-timer.h | ||
79 | @@ -XXX,XX +XXX,XX @@ | ||
80 | +/* | ||
81 | + * QEMU coroutine pool timer | ||
82 | + * | ||
83 | + * Copyright (c) 2021 Red Hat, Inc. | ||
84 | + * | ||
85 | + * SPDX-License-Identifier: LGPL-2.1-or-later | ||
86 | + * | ||
87 | + * This work is licensed under the terms of the GNU LGPL, version 2.1 or later. | ||
88 | + * See the COPYING.LIB file in the top-level directory. | ||
89 | + * | ||
90 | + */ | ||
91 | +#ifndef COROUTINE_POOL_TIMER_H | ||
92 | +#define COROUTINE_POOL_TIMER_H | ||
93 | + | ||
94 | +#include "qemu/osdep.h" | ||
95 | +#include "block/aio.h" | ||
96 | + | ||
97 | +/** | ||
98 | + * A timer that periodically resizes this thread's coroutine pool, freeing | ||
99 | + * memory if there are too many unused coroutines. | ||
100 | + * | ||
101 | + * Threads that make heavy use of coroutines should use this. Failure to resize | ||
102 | + * the coroutine pool can lead to large amounts of memory sitting idle and | ||
103 | + * never being used after the first time. | ||
104 | + */ | ||
105 | +typedef struct { | ||
106 | + QEMUTimer *timer; | ||
107 | +} CoroutinePoolTimer; | ||
108 | + | ||
109 | +/* Call this before the thread runs the AioContext */ | ||
110 | +void coroutine_pool_timer_init(CoroutinePoolTimer *pt, AioContext *ctx); | ||
111 | + | ||
112 | +/* Call this before the AioContext from the init function is destroyed */ | ||
113 | +void coroutine_pool_timer_cleanup(CoroutinePoolTimer *pt); | ||
114 | + | ||
115 | +#endif /* COROUTINE_POOL_TIMER_H */ | ||
116 | diff --git a/include/qemu/coroutine.h b/include/qemu/coroutine.h | ||
117 | index XXXXXXX..XXXXXXX 100644 | ||
118 | --- a/include/qemu/coroutine.h | ||
119 | +++ b/include/qemu/coroutine.h | ||
120 | @@ -XXX,XX +XXX,XX @@ bool qemu_in_coroutine(void); | ||
121 | */ | ||
122 | bool qemu_coroutine_entered(Coroutine *co); | ||
123 | |||
124 | +/** | ||
125 | + * Optionally call this function periodically to shrink the thread-local pool | ||
126 | + * down. Spiky workloads can create many coroutines and then never reach that | ||
127 | + * level again. Shrinking the pool reclaims memory in this case. | ||
128 | + */ | ||
129 | +void qemu_coroutine_pool_periodic_resize(void); | ||
130 | + | ||
131 | /** | ||
132 | * Provides a mutex that can be used to synchronise coroutines | ||
133 | */ | ||
134 | diff --git a/iothread.c b/iothread.c | ||
135 | index XXXXXXX..XXXXXXX 100644 | ||
136 | --- a/iothread.c | ||
137 | +++ b/iothread.c | ||
138 | @@ -XXX,XX +XXX,XX @@ | ||
139 | #include "qemu/error-report.h" | ||
140 | #include "qemu/rcu.h" | ||
141 | #include "qemu/main-loop.h" | ||
142 | +#include "qemu/coroutine-pool-timer.h" | ||
143 | |||
144 | typedef ObjectClass IOThreadClass; | ||
145 | |||
146 | @@ -XXX,XX +XXX,XX @@ DECLARE_CLASS_CHECKERS(IOThreadClass, IOTHREAD, | ||
147 | static void *iothread_run(void *opaque) | ||
148 | { | ||
149 | IOThread *iothread = opaque; | ||
150 | + CoroutinePoolTimer co_pool_timer; | ||
151 | |||
152 | rcu_register_thread(); | ||
153 | /* | ||
154 | @@ -XXX,XX +XXX,XX @@ static void *iothread_run(void *opaque) | ||
155 | iothread->thread_id = qemu_get_thread_id(); | ||
156 | qemu_sem_post(&iothread->init_done_sem); | ||
157 | |||
158 | + coroutine_pool_timer_init(&co_pool_timer, iothread->ctx); | ||
159 | + | ||
160 | while (iothread->running) { | ||
161 | /* | ||
162 | * Note: from functional-wise the g_main_loop_run() below can | ||
163 | @@ -XXX,XX +XXX,XX @@ static void *iothread_run(void *opaque) | ||
164 | } | ||
165 | } | ||
166 | |||
167 | + coroutine_pool_timer_cleanup(&co_pool_timer); | ||
168 | + | ||
169 | g_main_context_pop_thread_default(iothread->worker_context); | ||
170 | rcu_unregister_thread(); | ||
171 | return NULL; | ||
172 | diff --git a/util/coroutine-pool-timer.c b/util/coroutine-pool-timer.c | ||
173 | new file mode 100644 | ||
174 | index XXXXXXX..XXXXXXX | ||
175 | --- /dev/null | ||
176 | +++ b/util/coroutine-pool-timer.c | ||
177 | @@ -XXX,XX +XXX,XX @@ | ||
178 | +/* | ||
179 | + * QEMU coroutine pool timer | ||
180 | + * | ||
181 | + * Copyright (c) 2021 Red Hat, Inc. | ||
182 | + * | ||
183 | + * SPDX-License-Identifier: LGPL-2.1-or-later | ||
184 | + * | ||
185 | + * This work is licensed under the terms of the GNU LGPL, version 2.1 or later. | ||
186 | + * See the COPYING.LIB file in the top-level directory. | ||
187 | + * | ||
188 | + */ | ||
189 | +#include "qemu/coroutine-pool-timer.h" | ||
190 | + | ||
191 | +static void coroutine_pool_timer_cb(void *opaque) | ||
192 | +{ | ||
193 | + CoroutinePoolTimer *pt = opaque; | ||
194 | + int64_t expiry_time_ns = qemu_clock_get_ns(QEMU_CLOCK_REALTIME) + | ||
195 | + 15 * NANOSECONDS_PER_SECOND; | ||
196 | + | ||
197 | + qemu_coroutine_pool_periodic_resize(); | ||
198 | + timer_mod(pt->timer, expiry_time_ns); | ||
199 | +} | ||
200 | + | ||
201 | +void coroutine_pool_timer_init(CoroutinePoolTimer *pt, AioContext *ctx) | ||
202 | +{ | ||
203 | + pt->timer = aio_timer_new(ctx, QEMU_CLOCK_REALTIME, SCALE_NS, | ||
204 | + coroutine_pool_timer_cb, pt); | ||
205 | + coroutine_pool_timer_cb(pt); | ||
206 | +} | ||
207 | + | ||
208 | +void coroutine_pool_timer_cleanup(CoroutinePoolTimer *pt) | ||
209 | +{ | ||
210 | + timer_free(pt->timer); | ||
211 | + pt->timer = NULL; | ||
212 | +} | ||
213 | diff --git a/util/main-loop.c b/util/main-loop.c | ||
214 | index XXXXXXX..XXXXXXX 100644 | ||
215 | --- a/util/main-loop.c | ||
216 | +++ b/util/main-loop.c | ||
217 | @@ -XXX,XX +XXX,XX @@ | ||
218 | #include "qemu/error-report.h" | ||
219 | #include "qemu/queue.h" | ||
220 | #include "qemu/compiler.h" | ||
221 | +#include "qemu/coroutine-pool-timer.h" | ||
222 | |||
223 | #ifndef _WIN32 | ||
224 | #include <sys/wait.h> | ||
225 | @@ -XXX,XX +XXX,XX @@ static int qemu_signal_init(Error **errp) | ||
226 | |||
227 | static AioContext *qemu_aio_context; | ||
228 | static QEMUBH *qemu_notify_bh; | ||
229 | +static CoroutinePoolTimer main_loop_co_pool_timer; | ||
230 | |||
231 | static void notify_event_cb(void *opaque) | ||
232 | { | ||
233 | @@ -XXX,XX +XXX,XX @@ int qemu_init_main_loop(Error **errp) | ||
234 | g_source_set_name(src, "io-handler"); | ||
235 | g_source_attach(src, NULL); | ||
236 | g_source_unref(src); | ||
237 | + | ||
238 | + coroutine_pool_timer_init(&main_loop_co_pool_timer, qemu_aio_context); | ||
239 | + | ||
240 | return 0; | ||
241 | } | ||
242 | |||
243 | diff --git a/util/qemu-coroutine.c b/util/qemu-coroutine.c | ||
244 | index XXXXXXX..XXXXXXX 100644 | ||
245 | --- a/util/qemu-coroutine.c | ||
246 | +++ b/util/qemu-coroutine.c | ||
247 | @@ -XXX,XX +XXX,XX @@ | ||
248 | #include "block/aio.h" | ||
249 | |||
250 | enum { | ||
251 | - POOL_BATCH_SIZE = 64, | ||
252 | + /* | ||
253 | + * qemu_coroutine_pool_periodic_resize() keeps at least this many | ||
254 | + * coroutines around. | ||
255 | + */ | ||
256 | + ALLOC_POOL_MIN = 64, | ||
257 | }; | ||
258 | |||
259 | + | ||
260 | /** Free list to speed up creation */ | ||
261 | -static QSLIST_HEAD(, Coroutine) release_pool = QSLIST_HEAD_INITIALIZER(pool); | ||
262 | -static unsigned int release_pool_size; | ||
263 | static __thread QSLIST_HEAD(, Coroutine) alloc_pool = QSLIST_HEAD_INITIALIZER(pool); | ||
264 | static __thread unsigned int alloc_pool_size; | ||
265 | +static __thread unsigned int num_coroutines; | ||
266 | +static __thread unsigned int max_coroutines_this_slice; | ||
267 | static __thread Notifier coroutine_pool_cleanup_notifier; | ||
268 | |||
269 | static void coroutine_pool_cleanup(Notifier *n, void *value) | ||
270 | @@ -XXX,XX +XXX,XX @@ Coroutine *qemu_coroutine_create(CoroutineEntry *entry, void *opaque) | ||
271 | |||
272 | if (CONFIG_COROUTINE_POOL) { | ||
273 | co = QSLIST_FIRST(&alloc_pool); | ||
274 | - if (!co) { | ||
275 | - if (release_pool_size > POOL_BATCH_SIZE) { | ||
276 | - /* Slow path; a good place to register the destructor, too. */ | ||
277 | - if (!coroutine_pool_cleanup_notifier.notify) { | ||
278 | - coroutine_pool_cleanup_notifier.notify = coroutine_pool_cleanup; | ||
279 | - qemu_thread_atexit_add(&coroutine_pool_cleanup_notifier); | ||
280 | - } | ||
281 | - | ||
282 | - /* This is not exact; there could be a little skew between | ||
283 | - * release_pool_size and the actual size of release_pool. But | ||
284 | - * it is just a heuristic, it does not need to be perfect. | ||
285 | - */ | ||
286 | - alloc_pool_size = qatomic_xchg(&release_pool_size, 0); | ||
287 | - QSLIST_MOVE_ATOMIC(&alloc_pool, &release_pool); | ||
288 | - co = QSLIST_FIRST(&alloc_pool); | ||
289 | - } | ||
290 | - } | ||
291 | if (co) { | ||
292 | QSLIST_REMOVE_HEAD(&alloc_pool, pool_next); | ||
293 | alloc_pool_size--; | ||
294 | } | ||
295 | + | ||
296 | + num_coroutines++; | ||
297 | + if (num_coroutines > max_coroutines_this_slice) { | ||
298 | + max_coroutines_this_slice = num_coroutines; | ||
48 | + } | 299 | + } |
49 | if (tg->tokens[i] == tgm) { | 300 | } |
50 | token = throttle_group_next_tgm(tgm); | 301 | |
51 | /* Take care of the case where this is the last tgm in the group */ | 302 | if (!co) { |
303 | @@ -XXX,XX +XXX,XX @@ static void coroutine_delete(Coroutine *co) | ||
304 | co->caller = NULL; | ||
305 | |||
306 | if (CONFIG_COROUTINE_POOL) { | ||
307 | - if (release_pool_size < POOL_BATCH_SIZE * 2) { | ||
308 | - QSLIST_INSERT_HEAD_ATOMIC(&release_pool, co, pool_next); | ||
309 | - qatomic_inc(&release_pool_size); | ||
310 | - return; | ||
311 | - } | ||
312 | - if (alloc_pool_size < POOL_BATCH_SIZE) { | ||
313 | - QSLIST_INSERT_HEAD(&alloc_pool, co, pool_next); | ||
314 | - alloc_pool_size++; | ||
315 | - return; | ||
316 | + if (!coroutine_pool_cleanup_notifier.notify) { | ||
317 | + coroutine_pool_cleanup_notifier.notify = coroutine_pool_cleanup; | ||
318 | + qemu_thread_atexit_add(&coroutine_pool_cleanup_notifier); | ||
319 | } | ||
320 | + | ||
321 | + num_coroutines--; | ||
322 | + QSLIST_INSERT_HEAD(&alloc_pool, co, pool_next); | ||
323 | + alloc_pool_size++; | ||
324 | + return; | ||
325 | } | ||
326 | |||
327 | qemu_coroutine_delete(co); | ||
328 | } | ||
329 | |||
330 | +void qemu_coroutine_pool_periodic_resize(void) | ||
331 | +{ | ||
332 | + unsigned pool_size_target = | ||
333 | + MAX(ALLOC_POOL_MIN, max_coroutines_this_slice) - num_coroutines; | ||
334 | + max_coroutines_this_slice = num_coroutines; | ||
335 | + | ||
336 | + while (alloc_pool_size > pool_size_target) { | ||
337 | + Coroutine *co = QSLIST_FIRST(&alloc_pool); | ||
338 | + QSLIST_REMOVE_HEAD(&alloc_pool, pool_next); | ||
339 | + qemu_coroutine_delete(co); | ||
340 | + alloc_pool_size--; | ||
341 | + } | ||
342 | +} | ||
343 | + | ||
344 | void qemu_aio_coroutine_enter(AioContext *ctx, Coroutine *co) | ||
345 | { | ||
346 | QSIMPLEQ_HEAD(, Coroutine) pending = QSIMPLEQ_HEAD_INITIALIZER(pending); | ||
347 | diff --git a/util/meson.build b/util/meson.build | ||
348 | index XXXXXXX..XXXXXXX 100644 | ||
349 | --- a/util/meson.build | ||
350 | +++ b/util/meson.build | ||
351 | @@ -XXX,XX +XXX,XX @@ if have_block | ||
352 | util_ss.add(files('buffer.c')) | ||
353 | util_ss.add(files('bufferiszero.c')) | ||
354 | util_ss.add(files('coroutine-@0@.c'.format(config_host['CONFIG_COROUTINE_BACKEND']))) | ||
355 | + util_ss.add(files('coroutine-pool-timer.c')) | ||
356 | util_ss.add(files('hbitmap.c')) | ||
357 | util_ss.add(files('hexdump.c')) | ||
358 | util_ss.add(files('iova-tree.c')) | ||
52 | -- | 359 | -- |
53 | 2.17.1 | 360 | 2.31.1 |
54 | 361 | ||
55 | 362 | diff view generated by jsdifflib |