Re: [syzbot] inconsistent lock state in sync_info_debugfs_show

From: Daniel Vetter
Date: Tue Nov 22 2022 - 09:51:28 EST


On Fri, Nov 18, 2022 at 06:46:38PM -0800, syzbot wrote:
> syzbot has found a reproducer for the following issue on:
>
> HEAD commit: 84368d882b96 Merge tag 'soc-fixes-6.1-3' of git://git.kern..
> git tree: upstream
> console+strace: https://syzkaller.appspot.com/x/log.txt?x=1670fb65880000
> kernel config: https://syzkaller.appspot.com/x/.config?x=6f4e5e9899396248
> dashboard link: https://syzkaller.appspot.com/bug?extid=007bfe0f3330f6e1e7d1
> compiler: gcc (Debian 10.2.1-6) 10.2.1 20210110, GNU ld (GNU Binutils for Debian) 2.35.2
> syz repro: https://syzkaller.appspot.com/x/repro.syz?x=164376f9880000
> C reproducer: https://syzkaller.appspot.com/x/repro.c?x=16cf0965880000
>
> Downloadable assets:
> disk image: https://storage.googleapis.com/syzbot-assets/031b6e68785d/disk-84368d88.raw.xz
> vmlinux: https://storage.googleapis.com/syzbot-assets/cff5e66b90e8/vmlinux-84368d88.xz
> kernel image: https://storage.googleapis.com/syzbot-assets/e75525784a66/bzImage-84368d88.xz
>
> IMPORTANT: if you fix the issue, please add the following tag to the commit:
> Reported-by: syzbot+007bfe0f3330f6e1e7d1@xxxxxxxxxxxxxxxxxxxxxxxxx
>
> ================================
> WARNING: inconsistent lock state
> 6.1.0-rc5-syzkaller-00144-g84368d882b96 #0 Not tainted
> --------------------------------
> inconsistent {IN-HARDIRQ-W} -> {HARDIRQ-ON-W} usage.
> syz-executor333/3645 [HC0[0]:SC0[0]:HE0:SE1] takes:
> ffffffff8d295c38 (sync_timeline_list_lock){?...}-{2:2}, at: spin_lock_irq include/linux/spinlock.h:375 [inline]
> ffffffff8d295c38 (sync_timeline_list_lock){?...}-{2:2}, at: sync_info_debugfs_show+0x31/0x200 drivers/dma-buf/sync_debug.c:147
> {IN-HARDIRQ-W} state was registered at:
> lock_acquire kernel/locking/lockdep.c:5668 [inline]
> lock_acquire+0x1e3/0x630 kernel/locking/lockdep.c:5633
> __raw_spin_lock_irqsave include/linux/spinlock_api_smp.h:110 [inline]
> _raw_spin_lock_irqsave+0x3d/0x60 kernel/locking/spinlock.c:162
> sync_timeline_debug_remove+0x29/0x1a0 drivers/dma-buf/sync_debug.c:31
> sync_timeline_free drivers/dma-buf/sw_sync.c:104 [inline]
> kref_put include/linux/kref.h:65 [inline]
> sync_timeline_put drivers/dma-buf/sw_sync.c:116 [inline]
> timeline_fence_release+0x267/0x340 drivers/dma-buf/sw_sync.c:144
> dma_fence_release+0x14b/0x690 drivers/dma-buf/dma-fence.c:559

Do we need to just generally push all dma_fence_release finalization to an
irq work to untangle this mess once and for all? Or is this not just plain
old recursion? Christian, any good ideas?
-Daniel

> kref_put include/linux/kref.h:65 [inline]
> dma_fence_put include/linux/dma-fence.h:276 [inline]
> dma_fence_array_release+0x1fa/0x2d0 drivers/dma-buf/dma-fence-array.c:120
> dma_fence_release+0x14b/0x690 drivers/dma-buf/dma-fence.c:559
> kref_put include/linux/kref.h:65 [inline]
> dma_fence_put include/linux/dma-fence.h:276 [inline]
> irq_dma_fence_array_work+0xa9/0xd0 drivers/dma-buf/dma-fence-array.c:52
> irq_work_single+0x124/0x260 kernel/irq_work.c:211
> irq_work_run_list kernel/irq_work.c:242 [inline]
> irq_work_run_list+0x91/0xc0 kernel/irq_work.c:225
> irq_work_run+0x58/0xd0 kernel/irq_work.c:251
> __sysvec_irq_work+0xce/0x4e0 arch/x86/kernel/irq_work.c:22
> sysvec_irq_work+0x92/0xc0 arch/x86/kernel/irq_work.c:17
> asm_sysvec_irq_work+0x1a/0x20 arch/x86/include/asm/idtentry.h:675
> __raw_spin_unlock_irq include/linux/spinlock_api_smp.h:160 [inline]
> _raw_spin_unlock_irq+0x29/0x50 kernel/locking/spinlock.c:202
> spin_unlock_irq include/linux/spinlock.h:400 [inline]
> sw_sync_debugfs_release+0x162/0x240 drivers/dma-buf/sw_sync.c:321
> __fput+0x27c/0xa90 fs/file_table.c:320
> task_work_run+0x16f/0x270 kernel/task_work.c:179
> ptrace_notify+0x118/0x140 kernel/signal.c:2354
> ptrace_report_syscall include/linux/ptrace.h:420 [inline]
> ptrace_report_syscall_exit include/linux/ptrace.h:482 [inline]
> syscall_exit_work kernel/entry/common.c:251 [inline]
> syscall_exit_to_user_mode_prepare+0x129/0x280 kernel/entry/common.c:278
> __syscall_exit_to_user_mode_work kernel/entry/common.c:283 [inline]
> syscall_exit_to_user_mode+0xd/0x50 kernel/entry/common.c:296
> do_syscall_64+0x46/0xb0 arch/x86/entry/common.c:86
> entry_SYSCALL_64_after_hwframe+0x63/0xcd
> irq event stamp: 638
> hardirqs last enabled at (637): [<ffffffff81cc6530>] ___slab_alloc+0xca0/0x1400 mm/slub.c:3132
> hardirqs last disabled at (638): [<ffffffff89f47e95>] __raw_spin_lock_irq include/linux/spinlock_api_smp.h:117 [inline]
> hardirqs last disabled at (638): [<ffffffff89f47e95>] _raw_spin_lock_irq+0x45/0x50 kernel/locking/spinlock.c:170
> softirqs last enabled at (538): [<ffffffff814c1d03>] invoke_softirq kernel/softirq.c:445 [inline]
> softirqs last enabled at (538): [<ffffffff814c1d03>] __irq_exit_rcu+0x123/0x180 kernel/softirq.c:650
> softirqs last disabled at (505): [<ffffffff814c1d03>] invoke_softirq kernel/softirq.c:445 [inline]
> softirqs last disabled at (505): [<ffffffff814c1d03>] __irq_exit_rcu+0x123/0x180 kernel/softirq.c:650
>
> other info that might help us debug this:
> Possible unsafe locking scenario:
>
> CPU0
> ----
> lock(sync_timeline_list_lock);
> <Interrupt>
> lock(sync_timeline_list_lock);
>
> *** DEADLOCK ***
>
> 2 locks held by syz-executor333/3645:
> #0: ffff888021f8c8b8 (&p->lock){+.+.}-{3:3}, at: seq_read_iter+0xe3/0x1280 fs/seq_file.c:182
> #1: ffffffff8d295c38 (sync_timeline_list_lock){?...}-{2:2}, at: spin_lock_irq include/linux/spinlock.h:375 [inline]
> #1: ffffffff8d295c38 (sync_timeline_list_lock){?...}-{2:2}, at: sync_info_debugfs_show+0x31/0x200 drivers/dma-buf/sync_debug.c:147
>
> stack backtrace:
> CPU: 1 PID: 3645 Comm: syz-executor333 Not tainted 6.1.0-rc5-syzkaller-00144-g84368d882b96 #0
> Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 10/26/2022
> Call Trace:
> <TASK>
> __dump_stack lib/dump_stack.c:88 [inline]
> dump_stack_lvl+0xd1/0x138 lib/dump_stack.c:106
> print_usage_bug kernel/locking/lockdep.c:3963 [inline]
> valid_state kernel/locking/lockdep.c:3975 [inline]
> mark_lock_irq kernel/locking/lockdep.c:4178 [inline]
> mark_lock.part.0.cold+0x18/0xd8 kernel/locking/lockdep.c:4634
> mark_lock kernel/locking/lockdep.c:4598 [inline]
> mark_held_locks+0x9f/0xe0 kernel/locking/lockdep.c:4236
> __trace_hardirqs_on_caller kernel/locking/lockdep.c:4254 [inline]
> lockdep_hardirqs_on_prepare kernel/locking/lockdep.c:4321 [inline]
> lockdep_hardirqs_on_prepare+0x139/0x410 kernel/locking/lockdep.c:4273
> trace_hardirqs_on+0x31/0x160 kernel/trace/trace_preemptirq.c:49
> __raw_spin_unlock_irq include/linux/spinlock_api_smp.h:159 [inline]
> _raw_spin_unlock_irq+0x23/0x50 kernel/locking/spinlock.c:202
> spin_unlock_irq include/linux/spinlock.h:400 [inline]
> sync_print_obj drivers/dma-buf/sync_debug.c:118 [inline]
> sync_info_debugfs_show+0xef/0x200 drivers/dma-buf/sync_debug.c:153
> seq_read_iter+0x4f9/0x1280 fs/seq_file.c:230
> seq_read+0x171/0x210 fs/seq_file.c:162
> vfs_read+0x25b/0x930 fs/read_write.c:468
> ksys_read+0x12b/0x250 fs/read_write.c:613
> do_syscall_x64 arch/x86/entry/common.c:50 [inline]
> do_syscall_64+0x39/0xb0 arch/x86/entry/common.c:80
> entry_SYSCALL_64_after_hwframe+0x63/0xcd
> RIP: 0033:0x7f7cb1bbb049
> Code: 28 00 00 00 75 05 48 83 c4 28 c3 e8 b1 14 00 00 90 48 89 f8 48 89 f7 48 89 d6 48 89 ca 4d 89 c2 4d 89 c8 4c 8b 4c 24 08 0f 05 <48> 3d 01 f0 ff ff 73 01 c3 48 c7 c1 c0 ff ff ff f7 d8 64 89 01 48
> RSP: 002b:00007ffd52951818 EFLAGS: 00000246 ORIG_RAX: 0000000000000000
> RAX: ffffffffffffffda RBX: 00000000000f4240 RCX: 00007f7cb1bbb049
> RDX: 0000000000002020 RSI: 0000000020002200 RDI: 0000000000000003
> RBP: 0000000000000000 R08: 0000000000000001 R09: 0000000000000001
> R10: 00007ffd52951290 R11: 0000000000000246 R12: 000000000000e32c
> R13: 00007ffd5295182c R14: 00007ffd52951840 R15: 00007ffd52951830
> </TASK>
>

--
Daniel Vetter
Software Engineer, Intel Corporation
http://blog.ffwll.ch