upstream boot error: BUG: soft lockup in smp_call_function

78 views
Skip to first unread message

syzbot

unread,
Jul 26, 2020, 3:04:21 PM7/26/20
to b...@alien8.de, h...@zytor.com, linux-...@vger.kernel.org, lu...@kernel.org, mi...@redhat.com, syzkall...@googlegroups.com, tg...@linutronix.de, x...@kernel.org
Hello,

syzbot found the following issue on:

HEAD commit: 8c26c87b Merge tag 'sound-5.8-rc7' of git://git.kernel.org..
git tree: upstream
console output: https://syzkaller.appspot.com/x/log.txt?x=15069c64900000
kernel config: https://syzkaller.appspot.com/x/.config?x=455d284425bae0bd
dashboard link: https://syzkaller.appspot.com/bug?extid=b33acc95084b5829b7c6
compiler: gcc (GCC) 10.1.0-syz 20200507

IMPORTANT: if you fix the issue, please add the following tag to the commit:
Reported-by: syzbot+b33acc...@syzkaller.appspotmail.com

random: fast init done
watchdog: BUG: soft lockup - CPU#2 stuck for 22s! [kworker/2:2:1225]
Modules linked in:
irq event stamp: 1613250
hardirqs last enabled at (1613249): [<ffffffff88000c42>] asm_sysvec_apic_timer_interrupt+0x12/0x20 arch/x86/include/asm/idtentry.h:585
hardirqs last disabled at (1613250): [<ffffffff87e592ed>] idtentry_enter_cond_rcu+0x1d/0x50 arch/x86/entry/common.c:649
softirqs last enabled at (1609304): [<ffffffff88200748>] __do_softirq+0x748/0xa60 kernel/softirq.c:319
softirqs last disabled at (1609289): [<ffffffff88000f0f>] asm_call_on_stack+0xf/0x20 arch/x86/entry/entry_64.S:711
CPU: 2 PID: 1225 Comm: kworker/2:2 Not tainted 5.8.0-rc6-syzkaller #0
Hardware name: QEMU Standard PC (Q35 + ICH9, 2009), BIOS rel-1.12.0-59-gc9ba5276e321-prebuilt.qemu.org 04/01/2014
Workqueue: events drm_fb_helper_dirty_work
RIP: 0010:csd_lock_wait kernel/smp.c:108 [inline]
RIP: 0010:smp_call_function_many_cond+0x7c4/0x990 kernel/smp.c:555
Code: 0a 00 85 ed 74 4d 48 b8 00 00 00 00 00 fc ff df 4d 89 f4 4c 89 f5 49 c1 ec 03 83 e5 07 49 01 c4 83 c5 03 e8 7e f3 0a 00 f3 90 <41> 0f b6 04 24 40 38 c5 7c 08 84 c0 0f 85 a2 00 00 00 8b 43 08 31
RSP: 0018:ffffc90004257990 EFLAGS: 00000293
RAX: 0000000000000000 RBX: ffffe8ffade098e0 RCX: ffffffff8168c938
RDX: ffff888027918000 RSI: ffffffff8168c912 RDI: 0000000000000005
RBP: 0000000000000003 R08: 0000000000000000 R09: ffffffff8aaed233
R10: 0000000000000000 R11: 0000000000000000 R12: fffff91ff5bc131d
R13: 0000000000000000 R14: ffffe8ffade098e8 R15: 0000000000000001
FS: 0000000000000000(0000) GS:ffff88802d000000(0000) knlGS:0000000000000000
CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
CR2: 00007f8bf43cc9de CR3: 0000000009a79000 CR4: 0000000000340ee0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400
Call Trace:
smp_call_function_many kernel/smp.c:577 [inline]
smp_call_function kernel/smp.c:599 [inline]
on_each_cpu+0x4a/0x240 kernel/smp.c:699
__purge_vmap_area_lazy+0x11e/0x1f20 mm/vmalloc.c:1367
try_purge_vmap_area_lazy mm/vmalloc.c:1404 [inline]
free_vmap_area_noflush+0x2bc/0x370 mm/vmalloc.c:1440
free_unmap_vmap_area mm/vmalloc.c:1453 [inline]
remove_vm_area+0x1ca/0x230 mm/vmalloc.c:2204
vm_remove_mappings mm/vmalloc.c:2231 [inline]
__vunmap+0x348/0xac0 mm/vmalloc.c:2293
vunmap+0x50/0x60 mm/vmalloc.c:2398
drm_gem_shmem_vunmap_locked drivers/gpu/drm/drm_gem_shmem_helper.c:323 [inline]
drm_gem_shmem_vunmap+0x174/0x1c0 drivers/gpu/drm/drm_gem_shmem_helper.c:340
drm_gem_vunmap+0x81/0x130 drivers/gpu/drm/drm_gem.c:1277
drm_client_buffer_vunmap+0x51/0xa0 drivers/gpu/drm/drm_client.c:340
drm_fb_helper_dirty_work+0x599/0x730 drivers/gpu/drm/drm_fb_helper.c:438
process_one_work+0x94c/0x1670 kernel/workqueue.c:2269
worker_thread+0x64c/0x1120 kernel/workqueue.c:2415
kthread+0x3b5/0x4a0 kernel/kthread.c:291
ret_from_fork+0x1f/0x30 arch/x86/entry/entry_64.S:293


---
This report is generated by a bot. It may contain errors.
See https://goo.gl/tpsmEJ for more information about syzbot.
syzbot engineers can be reached at syzk...@googlegroups.com.

syzbot will keep track of this issue. See:
https://goo.gl/tpsmEJ#status for how to communicate with syzbot.

syzbot

unread,
Feb 20, 2021, 4:28:18 AM2/20/21
to syzkall...@googlegroups.com
Auto-closing this bug as obsolete.
Crashes did not happen for a while, no reproducer and no activity.
Reply all
Reply to author
Forward
0 new messages