Throttling group members are restarted on config change and by
bdrv_drained_begin(). Pending timers should be cancelled before
restarting the queue, otherwise requests see that tg->any_timer_armed[]
is already set and do not schedule a timer.
For example, a hang occurs at least since QEMU 2.10.0 with -drive
iops=100 because no further timers will be scheduled:
(guest)$ dd if=/dev/zero of=/dev/vdb oflag=direct count=1000
(qemu) stop
(qemu) cont
...I/O is stuck...
This can be fixed by calling throttle_group_detach_aio_context() from a
bdrv_drained_begin/end() region. This way timers are quiesced properly,
requests are drained, and other throttle group members are scheduled, if
necessary.
Reported-by: Yongxue Hong <yhong@redhat.com>
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
---
v3:
* Different approach this time, based on Berto's insight that another
TGM may need to be scheduled now that our TGM's timers have been
detached.
I realized the problem isn't the detach operation, it's the restart
operation that runs before detach. Timers shouldn't be left active
across restart.
After fixing restart no change is necessary in detach, but I decided
to add assertions to make it clear that this function must be called
in a quiesced environment.
block/block-backend.c | 2 ++
block/throttle-groups.c | 28 ++++++++++++++++++++++++++++
2 files changed, 30 insertions(+)
diff --git a/block/block-backend.c b/block/block-backend.c
index 45d9101be3..da2f6c0f8a 100644
--- a/block/block-backend.c
+++ b/block/block-backend.c
@@ -1748,8 +1748,10 @@ void blk_set_aio_context(BlockBackend *blk, AioContext *new_context)
if (bs) {
if (tgm->throttle_state) {
+ bdrv_drained_begin(bs);
throttle_group_detach_aio_context(tgm);
throttle_group_attach_aio_context(tgm, new_context);
+ bdrv_drained_end(bs);
}
bdrv_set_aio_context(bs, new_context);
}
diff --git a/block/throttle-groups.c b/block/throttle-groups.c
index 6ba992c8d7..c13530695a 100644
--- a/block/throttle-groups.c
+++ b/block/throttle-groups.c
@@ -420,6 +420,23 @@ static void throttle_group_restart_queue(ThrottleGroupMember *tgm, bool is_write
void throttle_group_restart_tgm(ThrottleGroupMember *tgm)
{
if (tgm->throttle_state) {
+ ThrottleState *ts = tgm->throttle_state;
+ ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts);
+ ThrottleTimers *tt = &tgm->throttle_timers;
+
+ /* Cancel pending timers */
+ qemu_mutex_lock(&tg->lock);
+ if (timer_pending(tt->timers[0])) {
+ timer_del(tt->timers[0]);
+ tg->any_timer_armed[0] = false;
+ }
+ if (timer_pending(tt->timers[1])) {
+ timer_del(tt->timers[1]);
+ tg->any_timer_armed[1] = false;
+ }
+ qemu_mutex_unlock(&tg->lock);
+
+ /* This also schedules the next request in other TGMs, if necessary */
throttle_group_restart_queue(tgm, 0);
throttle_group_restart_queue(tgm, 1);
}
@@ -592,6 +609,17 @@ void throttle_group_attach_aio_context(ThrottleGroupMember *tgm,
void throttle_group_detach_aio_context(ThrottleGroupMember *tgm)
{
ThrottleTimers *tt = &tgm->throttle_timers;
+
+ /* Requests must have been drained */
+ assert(tgm->pending_reqs[0] == 0);
+ assert(tgm->pending_reqs[1] == 0);
+ assert(qemu_co_queue_empty(&tgm->throttled_reqs[0]));
+ assert(qemu_co_queue_empty(&tgm->throttled_reqs[1]));
+
+ /* Timers must be disabled */
+ assert(!timer_pending(tt->timers[0]));
+ assert(!timer_pending(tt->timers[1]));
+
throttle_timers_detach_aio_context(tt);
tgm->aio_context = NULL;
}
--
2.13.5
On 09/25/2017 08:57 AM, Stefan Hajnoczi wrote: > Throttling group members are restarted on config change and by > bdrv_drained_begin(). Pending timers should be cancelled before > restarting the queue, otherwise requests see that tg->any_timer_armed[] > is already set and do not schedule a timer. > > For example, a hang occurs at least since QEMU 2.10.0 with -drive > iops=100 because no further timers will be scheduled: > > (guest)$ dd if=/dev/zero of=/dev/vdb oflag=direct count=1000 > (qemu) stop > (qemu) cont > ...I/O is stuck... > > This can be fixed by calling throttle_group_detach_aio_context() from a > bdrv_drained_begin/end() region. This way timers are quiesced properly, > requests are drained, and other throttle group members are scheduled, if > necessary. > > Reported-by: Yongxue Hong <yhong@redhat.com> > Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com> > --- Reviewed-by: Eric Blake <eblake@redhat.com> -- Eric Blake, Principal Software Engineer Red Hat, Inc. +1-919-301-3266 Virtualization: qemu.org | libvirt.org
On Mon 25 Sep 2017 03:57:35 PM CEST, Stefan Hajnoczi <stefanha@redhat.com> wrote: Hey Stefan, sorry for the late reply but I've been looking a this for a while trying to understand it. > v3: > * Different approach this time, based on Berto's insight that another > TGM may need to be scheduled now that our TGM's timers have been > detached. > > I realized the problem isn't the detach operation, it's the restart > operation that runs before detach. Timers shouldn't be left active > across restart. I'm not sure if I follow this paragraph, why is the problem in the restart operation? I may have overlooked something after all the recent changes in the throttling code, but the original point of restart was to ensure that -after e.g. destroying and creating a new set of timers for a drive- at least one throttled request would wake up so all I/O could continue normally. Restart itself doesn't care about timers in the ThrottleGroupMember, it just takes a coroutine from the queue and restarts it. If there would be already a timer set in the same tgm it wouldn't cause any trouble, after all the timer callback simply restarts the queue. > void throttle_group_restart_tgm(ThrottleGroupMember *tgm) > { > if (tgm->throttle_state) { > + ThrottleState *ts = tgm->throttle_state; > + ThrottleGroup *tg = container_of(ts, ThrottleGroup, ts); > + ThrottleTimers *tt = &tgm->throttle_timers; > + > + /* Cancel pending timers */ > + qemu_mutex_lock(&tg->lock); > + if (timer_pending(tt->timers[0])) { > + timer_del(tt->timers[0]); > + tg->any_timer_armed[0] = false; > + } > + if (timer_pending(tt->timers[1])) { > + timer_del(tt->timers[1]); > + tg->any_timer_armed[1] = false; > + } > + qemu_mutex_unlock(&tg->lock); > + > + /* This also schedules the next request in other TGMs, if necessary */ > throttle_group_restart_queue(tgm, 0); > throttle_group_restart_queue(tgm, 1); > } I think if you cancel the timers _after_ throttle_group_restart_queue() it would work just the same, because as I said earlier that function doesn't touch tgm->throttle_timers. And if you cancel the timers after throttle_group_restart_queue() then you get something similar to what you had in patch v2 (minus the bdrv_drained_begin/end calls). Berto
© 2016 - 2024 Red Hat, Inc.