arch/x86/xen/mmu_pv.c | 4 ++++ 1 file changed, 4 insertions(+)
Function xen_pin_page calls xen_pte_lock, which in turn grab page
table lock (ptlock). When locking, xen_pte_lock expect mm->page_table_lock
to be held before grabbing ptlock, but this does not happen when pinning
is caused by xen_mm_pin_all.
This commit addresses lockdep warning below, which shows up when
suspending a Xen VM.
[ 3680.658422] Freezing user space processes
[ 3680.660156] Freezing user space processes completed (elapsed 0.001 seconds)
[ 3680.660182] OOM killer disabled.
[ 3680.660192] Freezing remaining freezable tasks
[ 3680.661485] Freezing remaining freezable tasks completed (elapsed 0.001 seconds)
[ 3680.685254]
[ 3680.685265] ==================================
[ 3680.685269] WARNING: Nested lock was not taken
[ 3680.685274] 6.12.0+ #16 Tainted: G W
[ 3680.685279] ----------------------------------
[ 3680.685283] migration/0/19 is trying to lock:
[ 3680.685288] ffff88800bac33c0 (ptlock_ptr(ptdesc)#2){+.+.}-{3:3}, at: xen_pin_page+0x175/0x1d0
[ 3680.685303]
[ 3680.685303] but this task is not holding:
[ 3680.685308] init_mm.page_table_lock
[ 3680.685311]
[ 3680.685311] stack backtrace:
[ 3680.685316] CPU: 0 UID: 0 PID: 19 Comm: migration/0 Tainted: G W 6.12.0+ #16
[ 3680.685324] Tainted: [W]=WARN
[ 3680.685328] Stopper: multi_cpu_stop+0x0/0x120 <- __stop_cpus.constprop.0+0x8c/0xd0
[ 3680.685339] Call Trace:
[ 3680.685344] <TASK>
[ 3680.685347] dump_stack_lvl+0x77/0xb0
[ 3680.685356] __lock_acquire+0x917/0x2310
[ 3680.685364] lock_acquire+0xce/0x2c0
[ 3680.685369] ? xen_pin_page+0x175/0x1d0
[ 3680.685373] _raw_spin_lock_nest_lock+0x2f/0x70
[ 3680.685381] ? xen_pin_page+0x175/0x1d0
[ 3680.685386] xen_pin_page+0x175/0x1d0
[ 3680.685390] ? __pfx_xen_pin_page+0x10/0x10
[ 3680.685394] __xen_pgd_walk+0x233/0x2c0
[ 3680.685401] ? stop_one_cpu+0x91/0x100
[ 3680.685405] __xen_pgd_pin+0x5d/0x250
[ 3680.685410] xen_mm_pin_all+0x70/0xa0
[ 3680.685415] xen_pv_pre_suspend+0xf/0x280
[ 3680.685420] xen_suspend+0x57/0x1a0
[ 3680.685428] multi_cpu_stop+0x6b/0x120
[ 3680.685432] ? update_cpumasks_hier+0x7c/0xa60
[ 3680.685439] ? __pfx_multi_cpu_stop+0x10/0x10
[ 3680.685443] cpu_stopper_thread+0x8c/0x140
[ 3680.685448] ? smpboot_thread_fn+0x20/0x1f0
[ 3680.685454] ? __pfx_smpboot_thread_fn+0x10/0x10
[ 3680.685458] smpboot_thread_fn+0xed/0x1f0
[ 3680.685462] kthread+0xde/0x110
[ 3680.685467] ? __pfx_kthread+0x10/0x10
[ 3680.685471] ret_from_fork+0x2f/0x50
[ 3680.685478] ? __pfx_kthread+0x10/0x10
[ 3680.685482] ret_from_fork_asm+0x1a/0x30
[ 3680.685489] </TASK>
[ 3680.685491]
[ 3680.685491] other info that might help us debug this:
[ 3680.685497] 1 lock held by migration/0/19:
[ 3680.685500] #0: ffffffff8284df38 (pgd_lock){+.+.}-{3:3}, at: xen_mm_pin_all+0x14/0xa0
[ 3680.685512]
[ 3680.685512] stack backtrace:
[ 3680.685518] CPU: 0 UID: 0 PID: 19 Comm: migration/0 Tainted: G W 6.12.0+ #16
[ 3680.685528] Tainted: [W]=WARN
[ 3680.685531] Stopper: multi_cpu_stop+0x0/0x120 <- __stop_cpus.constprop.0+0x8c/0xd0
[ 3680.685538] Call Trace:
[ 3680.685541] <TASK>
[ 3680.685544] dump_stack_lvl+0x77/0xb0
[ 3680.685549] __lock_acquire+0x93c/0x2310
[ 3680.685554] lock_acquire+0xce/0x2c0
[ 3680.685558] ? xen_pin_page+0x175/0x1d0
[ 3680.685562] _raw_spin_lock_nest_lock+0x2f/0x70
[ 3680.685568] ? xen_pin_page+0x175/0x1d0
[ 3680.685572] xen_pin_page+0x175/0x1d0
[ 3680.685578] ? __pfx_xen_pin_page+0x10/0x10
[ 3680.685582] __xen_pgd_walk+0x233/0x2c0
[ 3680.685588] ? stop_one_cpu+0x91/0x100
[ 3680.685592] __xen_pgd_pin+0x5d/0x250
[ 3680.685596] xen_mm_pin_all+0x70/0xa0
[ 3680.685600] xen_pv_pre_suspend+0xf/0x280
[ 3680.685607] xen_suspend+0x57/0x1a0
[ 3680.685611] multi_cpu_stop+0x6b/0x120
[ 3680.685615] ? update_cpumasks_hier+0x7c/0xa60
[ 3680.685620] ? __pfx_multi_cpu_stop+0x10/0x10
[ 3680.685625] cpu_stopper_thread+0x8c/0x140
[ 3680.685629] ? smpboot_thread_fn+0x20/0x1f0
[ 3680.685634] ? __pfx_smpboot_thread_fn+0x10/0x10
[ 3680.685638] smpboot_thread_fn+0xed/0x1f0
[ 3680.685642] kthread+0xde/0x110
[ 3680.685645] ? __pfx_kthread+0x10/0x10
[ 3680.685649] ret_from_fork+0x2f/0x50
[ 3680.685654] ? __pfx_kthread+0x10/0x10
[ 3680.685657] ret_from_fork_asm+0x1a/0x30
[ 3680.685662] </TASK>
[ 3680.685267] xen:grant_table: Grant tables using version 1 layout
[ 3680.685921] OOM killer enabled.
[ 3680.685934] Restarting tasks ... done.
Signed-off-by: Maksym Planeta <maksym@exostellar.io>
---
arch/x86/xen/mmu_pv.c | 4 ++++
1 file changed, 4 insertions(+)
diff --git a/arch/x86/xen/mmu_pv.c b/arch/x86/xen/mmu_pv.c
index 55a4996d0c04..2c70cd35e72c 100644
--- a/arch/x86/xen/mmu_pv.c
+++ b/arch/x86/xen/mmu_pv.c
@@ -781,6 +781,7 @@ void xen_mm_pin_all(void)
{
struct page *page;
+ spin_lock(&init_mm.page_table_lock);
spin_lock(&pgd_lock);
list_for_each_entry(page, &pgd_list, lru) {
@@ -791,6 +792,7 @@ void xen_mm_pin_all(void)
}
spin_unlock(&pgd_lock);
+ spin_unlock(&init_mm.page_table_lock);
}
static void __init xen_mark_pinned(struct mm_struct *mm, struct page *page,
@@ -887,6 +889,7 @@ void xen_mm_unpin_all(void)
{
struct page *page;
+ spin_lock(&init_mm.page_table_lock);
spin_lock(&pgd_lock);
list_for_each_entry(page, &pgd_list, lru) {
@@ -898,6 +901,7 @@ void xen_mm_unpin_all(void)
}
spin_unlock(&pgd_lock);
+ spin_unlock(&init_mm.page_table_lock);
}
static void xen_enter_mmap(struct mm_struct *mm)
--
2.42.0
On 04.12.24 11:35, Maksym Planeta wrote:
> Function xen_pin_page calls xen_pte_lock, which in turn grab page
> table lock (ptlock). When locking, xen_pte_lock expect mm->page_table_lock
> to be held before grabbing ptlock, but this does not happen when pinning
> is caused by xen_mm_pin_all.
>
> This commit addresses lockdep warning below, which shows up when
> suspending a Xen VM.
>
> [ 3680.658422] Freezing user space processes
> [ 3680.660156] Freezing user space processes completed (elapsed 0.001 seconds)
> [ 3680.660182] OOM killer disabled.
> [ 3680.660192] Freezing remaining freezable tasks
> [ 3680.661485] Freezing remaining freezable tasks completed (elapsed 0.001 seconds)
> [ 3680.685254]
> [ 3680.685265] ==================================
> [ 3680.685269] WARNING: Nested lock was not taken
> [ 3680.685274] 6.12.0+ #16 Tainted: G W
> [ 3680.685279] ----------------------------------
> [ 3680.685283] migration/0/19 is trying to lock:
> [ 3680.685288] ffff88800bac33c0 (ptlock_ptr(ptdesc)#2){+.+.}-{3:3}, at: xen_pin_page+0x175/0x1d0
> [ 3680.685303]
> [ 3680.685303] but this task is not holding:
> [ 3680.685308] init_mm.page_table_lock
> [ 3680.685311]
> [ 3680.685311] stack backtrace:
> [ 3680.685316] CPU: 0 UID: 0 PID: 19 Comm: migration/0 Tainted: G W 6.12.0+ #16
> [ 3680.685324] Tainted: [W]=WARN
> [ 3680.685328] Stopper: multi_cpu_stop+0x0/0x120 <- __stop_cpus.constprop.0+0x8c/0xd0
> [ 3680.685339] Call Trace:
> [ 3680.685344] <TASK>
> [ 3680.685347] dump_stack_lvl+0x77/0xb0
> [ 3680.685356] __lock_acquire+0x917/0x2310
> [ 3680.685364] lock_acquire+0xce/0x2c0
> [ 3680.685369] ? xen_pin_page+0x175/0x1d0
> [ 3680.685373] _raw_spin_lock_nest_lock+0x2f/0x70
> [ 3680.685381] ? xen_pin_page+0x175/0x1d0
> [ 3680.685386] xen_pin_page+0x175/0x1d0
> [ 3680.685390] ? __pfx_xen_pin_page+0x10/0x10
> [ 3680.685394] __xen_pgd_walk+0x233/0x2c0
> [ 3680.685401] ? stop_one_cpu+0x91/0x100
> [ 3680.685405] __xen_pgd_pin+0x5d/0x250
> [ 3680.685410] xen_mm_pin_all+0x70/0xa0
> [ 3680.685415] xen_pv_pre_suspend+0xf/0x280
> [ 3680.685420] xen_suspend+0x57/0x1a0
> [ 3680.685428] multi_cpu_stop+0x6b/0x120
> [ 3680.685432] ? update_cpumasks_hier+0x7c/0xa60
> [ 3680.685439] ? __pfx_multi_cpu_stop+0x10/0x10
> [ 3680.685443] cpu_stopper_thread+0x8c/0x140
> [ 3680.685448] ? smpboot_thread_fn+0x20/0x1f0
> [ 3680.685454] ? __pfx_smpboot_thread_fn+0x10/0x10
> [ 3680.685458] smpboot_thread_fn+0xed/0x1f0
> [ 3680.685462] kthread+0xde/0x110
> [ 3680.685467] ? __pfx_kthread+0x10/0x10
> [ 3680.685471] ret_from_fork+0x2f/0x50
> [ 3680.685478] ? __pfx_kthread+0x10/0x10
> [ 3680.685482] ret_from_fork_asm+0x1a/0x30
> [ 3680.685489] </TASK>
> [ 3680.685491]
> [ 3680.685491] other info that might help us debug this:
> [ 3680.685497] 1 lock held by migration/0/19:
> [ 3680.685500] #0: ffffffff8284df38 (pgd_lock){+.+.}-{3:3}, at: xen_mm_pin_all+0x14/0xa0
> [ 3680.685512]
> [ 3680.685512] stack backtrace:
> [ 3680.685518] CPU: 0 UID: 0 PID: 19 Comm: migration/0 Tainted: G W 6.12.0+ #16
> [ 3680.685528] Tainted: [W]=WARN
> [ 3680.685531] Stopper: multi_cpu_stop+0x0/0x120 <- __stop_cpus.constprop.0+0x8c/0xd0
> [ 3680.685538] Call Trace:
> [ 3680.685541] <TASK>
> [ 3680.685544] dump_stack_lvl+0x77/0xb0
> [ 3680.685549] __lock_acquire+0x93c/0x2310
> [ 3680.685554] lock_acquire+0xce/0x2c0
> [ 3680.685558] ? xen_pin_page+0x175/0x1d0
> [ 3680.685562] _raw_spin_lock_nest_lock+0x2f/0x70
> [ 3680.685568] ? xen_pin_page+0x175/0x1d0
> [ 3680.685572] xen_pin_page+0x175/0x1d0
> [ 3680.685578] ? __pfx_xen_pin_page+0x10/0x10
> [ 3680.685582] __xen_pgd_walk+0x233/0x2c0
> [ 3680.685588] ? stop_one_cpu+0x91/0x100
> [ 3680.685592] __xen_pgd_pin+0x5d/0x250
> [ 3680.685596] xen_mm_pin_all+0x70/0xa0
> [ 3680.685600] xen_pv_pre_suspend+0xf/0x280
> [ 3680.685607] xen_suspend+0x57/0x1a0
> [ 3680.685611] multi_cpu_stop+0x6b/0x120
> [ 3680.685615] ? update_cpumasks_hier+0x7c/0xa60
> [ 3680.685620] ? __pfx_multi_cpu_stop+0x10/0x10
> [ 3680.685625] cpu_stopper_thread+0x8c/0x140
> [ 3680.685629] ? smpboot_thread_fn+0x20/0x1f0
> [ 3680.685634] ? __pfx_smpboot_thread_fn+0x10/0x10
> [ 3680.685638] smpboot_thread_fn+0xed/0x1f0
> [ 3680.685642] kthread+0xde/0x110
> [ 3680.685645] ? __pfx_kthread+0x10/0x10
> [ 3680.685649] ret_from_fork+0x2f/0x50
> [ 3680.685654] ? __pfx_kthread+0x10/0x10
> [ 3680.685657] ret_from_fork_asm+0x1a/0x30
> [ 3680.685662] </TASK>
> [ 3680.685267] xen:grant_table: Grant tables using version 1 layout
> [ 3680.685921] OOM killer enabled.
> [ 3680.685934] Restarting tasks ... done.
>
> Signed-off-by: Maksym Planeta <maksym@exostellar.io>
Reviewed-by: Juergen Gross <jgross@suse.com>
Juergen
On 12/4/24 02:35, Maksym Planeta wrote:
> Function xen_pin_page calls xen_pte_lock, which in turn grab page
> table lock (ptlock). When locking, xen_pte_lock expect mm->page_table_lock
> to be held before grabbing ptlock, but this does not happen when pinning
> is caused by xen_mm_pin_all.
In changelogs, please indicate functions with parenthesis like this:
func(). It makes it easier to tell what is a function versus a variable
or other text.
The use of init_mm here in the preexisting code is a _bit_ fishy because
these pgds *HAVE* an mm and yet xen_mm_pin_all() passes in '&init_mm'
for them. That's relevant here because locking 'init_mm' obviously
doesn't do any good for other mm's.
I have the _feeling_ it's just a big hack and this code throws caution
tot the wind because of:
> * Expected to be called in stop_machine() ("equivalent to taking
> * every spinlock in the system"), so the locking doesn't really
> * matter all that much.
So the patch here kinda doubles down on the hack and continues the theme
because "locking doesn't really matter all that much."
If so, it's not super satisfying, but it is consistent with the existing
code.
> This commit addresses lockdep warning below, which shows up when
> suspending a Xen VM.
If the comment I quote above is right, this is a _harmless_ warning
because nothing else can race here and the locking that gets added is
useless anyway, right? You seem to agree because there's no cc:stable@
or Fixes: tags.
Either way, I'm hesitant to add incorrect-in-any-other-context and
uncommented locking just to shut up lockdep. Is there no better way?
If not, can we comment it at least, please, so nobody else tries to
duplicate the locking?
On 05/12/2024 19:52, Dave Hansen wrote:
> I have the _feeling_ it's just a big hack and this code throws caution
> tot the wind because of:
>
>> * Expected to be called in stop_machine() ("equivalent to taking
>> * every spinlock in the system"), so the locking doesn't really
>> * matter all that much.
>
> So the patch here kinda doubles down on the hack and continues the theme
> because "locking doesn't really matter all that much."
>
> If so, it's not super satisfying, but it is consistent with the existing
> code.
>
I indeed could not find reasons why locking would be strictly necessary for correctness here. On the other hand a
clearly benign warning should not be triggered, especially considering that panic_on_warn may be on on some systems.
© 2016 - 2025 Red Hat, Inc.