Hello,
syzbot found the following issue on:
HEAD commit: 6c6a6c7e211c Linux 6.1.66
git tree: linux-6.1.y
console output:
https://syzkaller.appspot.com/x/log.txt?x=119c383ae80000
kernel config:
https://syzkaller.appspot.com/x/.config?x=ae60e3c09665494a
dashboard link:
https://syzkaller.appspot.com/bug?extid=7a18eb8772c67f7582c8
compiler: Debian clang version 15.0.6, GNU ld (GNU Binutils for Debian) 2.40
userspace arch: arm64
Unfortunately, I don't have any reproducer for this issue yet.
Downloadable assets:
disk image:
https://storage.googleapis.com/syzbot-assets/a1d0b246f59f/disk-6c6a6c7e.raw.xz
vmlinux:
https://storage.googleapis.com/syzbot-assets/777cb9cf9777/vmlinux-6c6a6c7e.xz
kernel image:
https://storage.googleapis.com/syzbot-assets/4a872139cccc/Image-6c6a6c7e.gz.xz
IMPORTANT: if you fix the issue, please add the following tag to the commit:
Reported-by:
syzbot+7a18eb...@syzkaller.appspotmail.com
INFO: task kworker/1:8:4350 blocked for more than 143 seconds.
Not tainted 6.1.66-syzkaller #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/1:8 state:D stack:0 pid:4350 ppid:2 flags:0x00000008
Workqueue: events switchdev_deferred_process_work
Call trace:
__switch_to+0x320/0x754 arch/arm64/kernel/process.c:553
context_switch kernel/sched/core.c:5245 [inline]
__schedule+0xee4/0x1c98 kernel/sched/core.c:6558
schedule+0xc4/0x170 kernel/sched/core.c:6634
schedule_preempt_disabled+0x18/0x2c kernel/sched/core.c:6693
__mutex_lock_common+0xbd8/0x21a0 kernel/locking/mutex.c:679
__mutex_lock kernel/locking/mutex.c:747 [inline]
mutex_lock_nested+0x38/0x44 kernel/locking/mutex.c:799
rtnl_lock+0x20/0x2c net/core/rtnetlink.c:74
switchdev_deferred_process_work+0x14/0x28 net/switchdev/switchdev.c:75
process_one_work+0x7ac/0x1404 kernel/workqueue.c:2292
worker_thread+0x8e4/0xfec kernel/workqueue.c:2439
kthread+0x250/0x2d8 kernel/kthread.c:376
ret_from_fork+0x10/0x20 arch/arm64/kernel/entry.S:864
INFO: task kworker/0:1:15365 blocked for more than 143 seconds.
Not tainted 6.1.66-syzkaller #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/0:1 state:D stack:0 pid:15365 ppid:2 flags:0x00000008
Workqueue: ipv6_addrconf addrconf_dad_work
Call trace:
__switch_to+0x320/0x754 arch/arm64/kernel/process.c:553
context_switch kernel/sched/core.c:5245 [inline]
__schedule+0xee4/0x1c98 kernel/sched/core.c:6558
schedule+0xc4/0x170 kernel/sched/core.c:6634
schedule_preempt_disabled+0x18/0x2c kernel/sched/core.c:6693
__mutex_lock_common+0xbd8/0x21a0 kernel/locking/mutex.c:679
__mutex_lock kernel/locking/mutex.c:747 [inline]
mutex_lock_nested+0x38/0x44 kernel/locking/mutex.c:799
rtnl_lock+0x20/0x2c net/core/rtnetlink.c:74
addrconf_dad_work+0xd0/0x1390 net/ipv6/addrconf.c:4098
process_one_work+0x7ac/0x1404 kernel/workqueue.c:2292
worker_thread+0x8e4/0xfec kernel/workqueue.c:2439
kthread+0x250/0x2d8 kernel/kthread.c:376
ret_from_fork+0x10/0x20 arch/arm64/kernel/entry.S:864
INFO: task kworker/1:1:26906 blocked for more than 143 seconds.
Not tainted 6.1.66-syzkaller #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/1:1 state:D stack:0 pid:26906 ppid:2 flags:0x00000008
Workqueue: events linkwatch_event
Call trace:
__switch_to+0x320/0x754 arch/arm64/kernel/process.c:553
context_switch kernel/sched/core.c:5245 [inline]
__schedule+0xee4/0x1c98 kernel/sched/core.c:6558
schedule+0xc4/0x170 kernel/sched/core.c:6634
schedule_preempt_disabled+0x18/0x2c kernel/sched/core.c:6693
__mutex_lock_common+0xbd8/0x21a0 kernel/locking/mutex.c:679
__mutex_lock kernel/locking/mutex.c:747 [inline]
mutex_lock_nested+0x38/0x44 kernel/locking/mutex.c:799
rtnl_lock+0x20/0x2c net/core/rtnetlink.c:74
linkwatch_event+0x14/0x68 net/core/link_watch.c:263
process_one_work+0x7ac/0x1404 kernel/workqueue.c:2292
worker_thread+0x8e4/0xfec kernel/workqueue.c:2439
kthread+0x250/0x2d8 kernel/kthread.c:376
ret_from_fork+0x10/0x20 arch/arm64/kernel/entry.S:864
INFO: task kworker/1:5:29069 blocked for more than 143 seconds.
Not tainted 6.1.66-syzkaller #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/1:5 state:D stack:0 pid:29069 ppid:2 flags:0x00000008
Workqueue: ipv6_addrconf addrconf_dad_work
Call trace:
__switch_to+0x320/0x754 arch/arm64/kernel/process.c:553
context_switch kernel/sched/core.c:5245 [inline]
__schedule+0xee4/0x1c98 kernel/sched/core.c:6558
schedule+0xc4/0x170 kernel/sched/core.c:6634
schedule_preempt_disabled+0x18/0x2c kernel/sched/core.c:6693
__mutex_lock_common+0xbd8/0x21a0 kernel/locking/mutex.c:679
__mutex_lock kernel/locking/mutex.c:747 [inline]
mutex_lock_nested+0x38/0x44 kernel/locking/mutex.c:799
rtnl_lock+0x20/0x2c net/core/rtnetlink.c:74
addrconf_dad_work+0xd0/0x1390 net/ipv6/addrconf.c:4098
process_one_work+0x7ac/0x1404 kernel/workqueue.c:2292
worker_thread+0x8e4/0xfec kernel/workqueue.c:2439
kthread+0x250/0x2d8 kernel/kthread.c:376
ret_from_fork+0x10/0x20 arch/arm64/kernel/entry.S:864
INFO: task syz-executor.1:1163 blocked for more than 143 seconds.
Not tainted 6.1.66-syzkaller #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:syz-executor.1 state:D stack:0 pid:1163 ppid:1037 flags:0x00000005
Call trace:
__switch_to+0x320/0x754 arch/arm64/kernel/process.c:553
context_switch kernel/sched/core.c:5245 [inline]
__schedule+0xee4/0x1c98 kernel/sched/core.c:6558
schedule+0xc4/0x170 kernel/sched/core.c:6634
schedule_preempt_disabled+0x18/0x2c kernel/sched/core.c:6693
__mutex_lock_common+0xbd8/0x21a0 kernel/locking/mutex.c:679
__mutex_lock kernel/locking/mutex.c:747 [inline]
mutex_lock_nested+0x38/0x44 kernel/locking/mutex.c:799
rtnl_lock net/core/rtnetlink.c:74 [inline]
rtnetlink_rcv_msg+0x6e8/0xd94 net/core/rtnetlink.c:6119
netlink_rcv_skb+0x20c/0x3b8 net/netlink/af_netlink.c:2508
rtnetlink_rcv+0x28/0x38 net/core/rtnetlink.c:6140
netlink_unicast_kernel net/netlink/af_netlink.c:1326 [inline]
netlink_unicast+0x65c/0x898 net/netlink/af_netlink.c:1352
netlink_sendmsg+0x834/0xb18 net/netlink/af_netlink.c:1874
sock_sendmsg_nosec net/socket.c:716 [inline]
__sock_sendmsg net/socket.c:728 [inline]
____sys_sendmsg+0x558/0x844 net/socket.c:2499
___sys_sendmsg net/socket.c:2553 [inline]
__sys_sendmsg+0x26c/0x33c net/socket.c:2582
__do_sys_sendmsg net/socket.c:2591 [inline]
__se_sys_sendmsg net/socket.c:2589 [inline]
__arm64_sys_sendmsg+0x80/0x94 net/socket.c:2589
__invoke_syscall arch/arm64/kernel/syscall.c:38 [inline]
invoke_syscall+0x98/0x2c0 arch/arm64/kernel/syscall.c:52
el0_svc_common+0x138/0x258 arch/arm64/kernel/syscall.c:142
do_el0_svc+0x64/0x218 arch/arm64/kernel/syscall.c:206
el0_svc+0x58/0x168 arch/arm64/kernel/entry-common.c:637
el0t_64_sync_handler+0x84/0xf0 arch/arm64/kernel/entry-common.c:655
el0t_64_sync+0x18c/0x190 arch/arm64/kernel/entry.S:585
INFO: task syz-executor.1:1170 blocked for more than 143 seconds.
Not tainted 6.1.66-syzkaller #0
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:syz-executor.1 state:D stack:0 pid:1170 ppid:1037 flags:0x00000005
Call trace:
__switch_to+0x320/0x754 arch/arm64/kernel/process.c:553
context_switch kernel/sched/core.c:5245 [inline]
__schedule+0xee4/0x1c98 kernel/sched/core.c:6558
schedule+0xc4/0x170 kernel/sched/core.c:6634
schedule_preempt_disabled+0x18/0x2c kernel/sched/core.c:6693
__mutex_lock_common+0xbd8/0x21a0 kernel/locking/mutex.c:679
__mutex_lock kernel/locking/mutex.c:747 [inline]
mutex_lock_nested+0x38/0x44 kernel/locking/mutex.c:799
rtnl_lock net/core/rtnetlink.c:74 [inline]
rtnetlink_rcv_msg+0x6e8/0xd94 net/core/rtnetlink.c:6119
netlink_rcv_skb+0x20c/0x3b8 net/netlink/af_netlink.c:2508
rtnetlink_rcv+0x28/0x38 net/core/rtnetlink.c:6140
netlink_unicast_kernel net/netlink/af_netlink.c:1326 [inline]
netlink_unicast+0x65c/0x898 net/netlink/af_netlink.c:1352
netlink_sendmsg+0x834/0xb18 net/netlink/af_netlink.c:1874
sock_sendmsg_nosec net/socket.c:716 [inline]
__sock_sendmsg net/socket.c:728 [inline]
____sys_sendmsg+0x558/0x844 net/socket.c:2499
___sys_sendmsg net/socket.c:2553 [inline]
__sys_sendmsg+0x26c/0x33c net/socket.c:2582
__do_sys_sendmsg net/socket.c:2591 [inline]
__se_sys_sendmsg net/socket.c:2589 [inline]
__arm64_sys_sendmsg+0x80/0x94 net/socket.c:2589
__invoke_syscall arch/arm64/kernel/syscall.c:38 [inline]
invoke_syscall+0x98/0x2c0 arch/arm64/kernel/syscall.c:52
el0_svc_common+0x138/0x258 arch/arm64/kernel/syscall.c:142
do_el0_svc+0x64/0x218 arch/arm64/kernel/syscall.c:206
el0_svc+0x58/0x168 arch/arm64/kernel/entry-common.c:637
el0t_64_sync_handler+0x84/0xf0 arch/arm64/kernel/entry-common.c:655
el0t_64_sync+0x18c/0x190 arch/arm64/kernel/entry.S:585
Showing all locks held in the system:
1 lock held by rcu_tasks_kthre/12:
#0: ffff8000159f4e70 (rcu_tasks.tasks_gp_mutex){+.+.}-{3:3}, at: rcu_tasks_one_gp+0x44/0xcf4 kernel/rcu/tasks.h:516
1 lock held by rcu_tasks_trace/13:
#0: ffff8000159f5670 (rcu_tasks_trace.tasks_gp_mutex){+.+.}-{3:3}, at: rcu_tasks_one_gp+0x44/0xcf4 kernel/rcu/tasks.h:516
1 lock held by khungtaskd/28:
#0: ffff8000159f4ca0 (rcu_read_lock){....}-{1:2}, at: rcu_lock_acquire+0xc/0x44 include/linux/rcupdate.h:305
3 locks held by kworker/u4:2/39:
#0: ffff0000c002a138 ((wq_completion)events_power_efficient){+.+.}-{0:0}, at: process_one_work+0x664/0x1404 kernel/workqueue.c:2265
#1: ffff80001b287c20 ((reg_check_chans).work){+.+.}-{0:0}, at: process_one_work+0x6a8/0x1404 kernel/workqueue.c:2267
#2: ffff800017df8888 (rtnl_mutex){+.+.}-{3:3}, at: rtnl_lock+0x20/0x2c net/core/rtnetlink.c:74
2 locks held by getty/3989:
#0: ffff0000d7780098 (&tty->ldisc_sem){++++}-{0:0}, at: ldsem_down_read+0x3c/0x4c drivers/tty/tty_ldsem.c:340
#1: ffff80001bcb02f0 (&ldata->atomic_read_lock){+.+.}-{3:3}, at: n_tty_read+0x414/0x1214 drivers/tty/n_tty.c:2188
3 locks held by kworker/1:4/4234:
#0: ffff0000c4e56538 ((wq_completion)infiniband){+.+.}-{0:0}, at: process_one_work+0x664/0x1404 kernel/workqueue.c:2265
#1: ffff80001dd17c20 ((work_completion)(&work->work)#2){+.+.}-{0:0}, at: process_one_work+0x6a8/0x1404 kernel/workqueue.c:2267
#2: ffff800017df8888 (rtnl_mutex){+.+.}-{3:3}, at: rtnl_lock+0x20/0x2c net/core/rtnetlink.c:74
3 locks held by kworker/1:8/4350:
#0: ffff0000c0020d38 ((wq_completion)events){+.+.}-{0:0}, at: process_one_work+0x664/0x1404 kernel/workqueue.c:2265
#1: ffff80001f2e7c20 (deferred_process_work){+.+.}-{0:0}, at: process_one_work+0x6a8/0x1404 kernel/workqueue.c:2267
#2: ffff800017df8888 (rtnl_mutex){+.+.}-{3:3}, at: rtnl_lock+0x20/0x2c net/core/rtnetlink.c:74
3 locks held by kworker/0:1/15365:
#0: ffff0000d5d10d38 ((wq_completion)ipv6_addrconf){+.+.}-{0:0}, at: process_one_work+0x664/0x1404 kernel/workqueue.c:2265
#1: ffff800021427c20 ((work_completion)(&(&ifa->dad_work)->work)){+.+.}-{0:0}, at: process_one_work+0x6a8/0x1404 kernel/workqueue.c:2267
#2: ffff800017df8888 (rtnl_mutex){+.+.}-{3:3}, at: rtnl_lock+0x20/0x2c net/core/rtnetlink.c:74
5 locks held by kworker/u4:3/28463:
#0: ffff0000c0845138 ((wq_completion)netns){+.+.}-{0:0}, at: process_one_work+0x664/0x1404 kernel/workqueue.c:2265
#1: ffff80001f3f7c20 (net_cleanup_work){+.+.}-{0:0}, at: process_one_work+0x6a8/0x1404 kernel/workqueue.c:2267
#2: ffff800017deca10 (pernet_ops_rwsem){++++}-{3:3}, at: cleanup_net+0xf4/0x994 net/core/net_namespace.c:563
#3: ffff800017df8888 (rtnl_mutex){+.+.}-{3:3}, at: rtnl_lock+0x20/0x2c net/core/rtnetlink.c:74
#4: ffff8000159fa140 (rcu_state.barrier_mutex){+.+.}-{3:3}, at: rcu_barrier+0x58/0x5c4 kernel/rcu/tree.c:3986
3 locks held by kworker/1:1/26906:
#0: ffff0000c0020d38 ((wq_completion)events){+.+.}-{0:0}, at: process_one_work+0x664/0x1404 kernel/workqueue.c:2265
#1: ffff800023497c20 ((linkwatch_work).work){+.+.}-{0:0}, at: process_one_work+0x6a8/0x1404 kernel/workqueue.c:2267
#2: ffff800017df8888 (rtnl_mutex){+.+.}-{3:3}, at: rtnl_lock+0x20/0x2c net/core/rtnetlink.c:74
3 locks held by kworker/1:5/29069:
#0: ffff0000d5d10d38 ((wq_completion)ipv6_addrconf){+.+.}-{0:0}, at: process_one_work+0x664/0x1404 kernel/workqueue.c:2265
#1: ffff800024a97c20 ((work_completion)(&(&ifa->dad_work)->work)){+.+.}-{0:0}, at: process_one_work+0x6a8/0x1404 kernel/workqueue.c:2267
#2: ffff800017df8888 (rtnl_mutex){+.+.}-{3:3}, at: rtnl_lock+0x20/0x2c net/core/rtnetlink.c:74
1 lock held by syz-executor.1/1163:
#0: ffff800017df8888 (rtnl_mutex){+.+.}-{3:3}, at: rtnl_lock net/core/rtnetlink.c:74 [inline]
#0: ffff800017df8888 (rtnl_mutex){+.+.}-{3:3}, at: rtnetlink_rcv_msg+0x6e8/0xd94 net/core/rtnetlink.c:6119
1 lock held by syz-executor.1/1170:
#0: ffff800017df8888 (rtnl_mutex){+.+.}-{3:3}, at: rtnl_lock net/core/rtnetlink.c:74 [inline]
#0: ffff800017df8888 (rtnl_mutex){+.+.}-{3:3}, at: rtnetlink_rcv_msg+0x6e8/0xd94 net/core/rtnetlink.c:6119
1 lock held by syz-executor.1/1221:
#0: ffff800017df8888 (rtnl_mutex){+.+.}-{3:3}, at: rtnl_lock net/core/rtnetlink.c:74 [inline]
#0: ffff800017df8888 (rtnl_mutex){+.+.}-{3:3}, at: rtnetlink_rcv_msg+0x6e8/0xd94 net/core/rtnetlink.c:6119
1 lock held by syz-executor.1/1229:
#0: ffff800017df8888 (rtnl_mutex){+.+.}-{3:3}, at: rtnl_lock net/core/rtnetlink.c:74 [inline]
#0: ffff800017df8888 (rtnl_mutex){+.+.}-{3:3}, at: rtnetlink_rcv_msg+0x6e8/0xd94 net/core/rtnetlink.c:6119
=============================================
---
This report is generated by a bot. It may contain errors.
See
https://goo.gl/tpsmEJ for more information about syzbot.
syzbot engineers can be reached at
syzk...@googlegroups.com.
syzbot will keep track of this issue. See:
https://goo.gl/tpsmEJ#status for how to communicate with syzbot.
If the report is already addressed, let syzbot know by replying with:
#syz fix: exact-commit-title
If you want to overwrite report's subsystems, reply with:
#syz set subsystems: new-subsystem
(See the list of subsystem names on the web dashboard)
If the report is a duplicate of another one, reply with:
#syz dup: exact-subject-of-another-report
If you want to undo deduplication, reply with:
#syz undup