possible deadlock in process_measurement

9 views
Skip to first unread message

syzbot

unread,
Jan 15, 2020, 4:21:12 PM1/15/20
to syzkaller...@googlegroups.com
Hello,

syzbot found the following crash on:

HEAD commit: db5b9190 Linux 4.19.96
git tree: linux-4.19.y
console output: https://syzkaller.appspot.com/x/log.txt?x=11c60c35e00000
kernel config: https://syzkaller.appspot.com/x/.config?x=7a24894346f71303
dashboard link: https://syzkaller.appspot.com/bug?extid=607c3f5980adfa8249f0
compiler: gcc (GCC) 9.0.0 20181231 (experimental)

Unfortunately, I don't have any reproducer for this crash yet.

IMPORTANT: if you fix the bug, please add the following tag to the commit:
Reported-by: syzbot+607c3f...@syzkaller.appspotmail.com

======================================================
WARNING: possible circular locking dependency detected
4.19.96-syzkaller #0 Not tainted
------------------------------------------------------
syz-executor.2/17259 is trying to acquire lock:
000000000cb304f7 (&ovl_i_mutex_key[depth]){+.+.}, at: inode_lock
include/linux/fs.h:747 [inline]
000000000cb304f7 (&ovl_i_mutex_key[depth]){+.+.}, at:
process_measurement+0x9ae/0x1560 security/integrity/ima/ima_main.c:205

but task is already holding lock:
000000000a4a8dc1 (&sig->cred_guard_mutex){+.+.}, at:
prepare_bprm_creds+0x55/0x120 fs/exec.c:1404

which lock already depends on the new lock.


the existing dependency chain (in reverse order) is:

-> #3 (&sig->cred_guard_mutex){+.+.}:
__mutex_lock_common kernel/locking/mutex.c:925 [inline]
__mutex_lock+0xf7/0x1300 kernel/locking/mutex.c:1072
mutex_lock_killable_nested+0x16/0x20 kernel/locking/mutex.c:1102
lock_trace+0x4a/0xe0 fs/proc/base.c:402
proc_pid_syscall+0x98/0x250 fs/proc/base.c:635
proc_single_show+0xf0/0x180 fs/proc/base.c:755
seq_read+0x4ca/0x1110 fs/seq_file.c:229
do_loop_readv_writev fs/read_write.c:701 [inline]
do_loop_readv_writev fs/read_write.c:688 [inline]
do_iter_read+0x490/0x640 fs/read_write.c:925
vfs_readv+0xf0/0x160 fs/read_write.c:987
do_preadv+0x1c4/0x280 fs/read_write.c:1071
__do_sys_preadv fs/read_write.c:1121 [inline]
__se_sys_preadv fs/read_write.c:1116 [inline]
__x64_sys_preadv+0x9a/0xf0 fs/read_write.c:1116
do_syscall_64+0xfd/0x620 arch/x86/entry/common.c:293
entry_SYSCALL_64_after_hwframe+0x49/0xbe

-> #2 (&p->lock){+.+.}:
__mutex_lock_common kernel/locking/mutex.c:925 [inline]
__mutex_lock+0xf7/0x1300 kernel/locking/mutex.c:1072
mutex_lock_nested+0x16/0x20 kernel/locking/mutex.c:1087
seq_read+0x71/0x1110 fs/seq_file.c:161
proc_reg_read+0x1f8/0x2b0 fs/proc/inode.c:231
do_loop_readv_writev fs/read_write.c:701 [inline]
do_loop_readv_writev fs/read_write.c:688 [inline]
do_iter_read+0x490/0x640 fs/read_write.c:925
vfs_readv+0xf0/0x160 fs/read_write.c:987
kernel_readv fs/splice.c:362 [inline]
default_file_splice_read+0x478/0x970 fs/splice.c:417
do_splice_to+0x127/0x180 fs/splice.c:881
splice_direct_to_actor+0x2d3/0x970 fs/splice.c:959
do_splice_direct+0x1da/0x2a0 fs/splice.c:1068
do_sendfile+0x597/0xce0 fs/read_write.c:1447
__do_sys_sendfile64 fs/read_write.c:1508 [inline]
__se_sys_sendfile64 fs/read_write.c:1494 [inline]
__x64_sys_sendfile64+0x1dd/0x220 fs/read_write.c:1494
do_syscall_64+0xfd/0x620 arch/x86/entry/common.c:293
entry_SYSCALL_64_after_hwframe+0x49/0xbe

-> #1 (sb_writers#4){.+.+}:
percpu_down_read_preempt_disable include/linux/percpu-rwsem.h:36
[inline]
percpu_down_read include/linux/percpu-rwsem.h:59 [inline]
__sb_start_write+0x208/0x360 fs/super.c:1387
sb_start_write include/linux/fs.h:1578 [inline]
mnt_want_write+0x3f/0xc0 fs/namespace.c:360
ovl_want_write+0x76/0xa0 fs/overlayfs/util.c:24
ovl_setattr+0xdd/0x950 fs/overlayfs/inode.c:30
notify_change+0xad7/0xfb0 fs/attr.c:334
do_truncate+0x158/0x220 fs/open.c:63
handle_truncate fs/namei.c:3008 [inline]
do_last fs/namei.c:3424 [inline]
path_openat+0x2c27/0x45e0 fs/namei.c:3534
do_filp_open+0x1a1/0x280 fs/namei.c:3564
do_sys_open+0x3fe/0x550 fs/open.c:1088
ksys_open include/linux/syscalls.h:1276 [inline]
__do_sys_creat fs/open.c:1146 [inline]
__se_sys_creat fs/open.c:1144 [inline]
__x64_sys_creat+0x61/0x80 fs/open.c:1144
do_syscall_64+0xfd/0x620 arch/x86/entry/common.c:293
entry_SYSCALL_64_after_hwframe+0x49/0xbe

-> #0 (&ovl_i_mutex_key[depth]){+.+.}:
lock_acquire+0x16f/0x3f0 kernel/locking/lockdep.c:3903
down_write+0x38/0x90 kernel/locking/rwsem.c:70
inode_lock include/linux/fs.h:747 [inline]
process_measurement+0x9ae/0x1560
security/integrity/ima/ima_main.c:205
ima_file_check+0xc5/0x110 security/integrity/ima/ima_main.c:391
do_last fs/namei.c:3422 [inline]
path_openat+0x1130/0x45e0 fs/namei.c:3534
do_filp_open+0x1a1/0x280 fs/namei.c:3564
do_open_execat+0x140/0x660 fs/exec.c:853
__do_execve_file.isra.0+0x15a4/0x2150 fs/exec.c:1755
do_execveat_common fs/exec.c:1866 [inline]
do_execve fs/exec.c:1883 [inline]
__do_sys_execve fs/exec.c:1964 [inline]
__se_sys_execve fs/exec.c:1959 [inline]
__x64_sys_execve+0x8f/0xc0 fs/exec.c:1959
do_syscall_64+0xfd/0x620 arch/x86/entry/common.c:293
entry_SYSCALL_64_after_hwframe+0x49/0xbe

other info that might help us debug this:

Chain exists of:
&ovl_i_mutex_key[depth] --> &p->lock --> &sig->cred_guard_mutex

Possible unsafe locking scenario:

CPU0 CPU1
---- ----
lock(&sig->cred_guard_mutex);
lock(&p->lock);
lock(&sig->cred_guard_mutex);
lock(&ovl_i_mutex_key[depth]);

*** DEADLOCK ***

1 lock held by syz-executor.2/17259:
#0: 000000000a4a8dc1 (&sig->cred_guard_mutex){+.+.}, at:
prepare_bprm_creds+0x55/0x120 fs/exec.c:1404

stack backtrace:
CPU: 1 PID: 17259 Comm: syz-executor.2 Not tainted 4.19.96-syzkaller #0
Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS
Google 01/01/2011
Call Trace:
__dump_stack lib/dump_stack.c:77 [inline]
dump_stack+0x197/0x210 lib/dump_stack.c:118
print_circular_bug.isra.0.cold+0x1cc/0x28f kernel/locking/lockdep.c:1221
check_prev_add kernel/locking/lockdep.c:1861 [inline]
check_prevs_add kernel/locking/lockdep.c:1974 [inline]
validate_chain kernel/locking/lockdep.c:2415 [inline]
__lock_acquire+0x2e19/0x49c0 kernel/locking/lockdep.c:3411
lock_acquire+0x16f/0x3f0 kernel/locking/lockdep.c:3903
down_write+0x38/0x90 kernel/locking/rwsem.c:70
inode_lock include/linux/fs.h:747 [inline]
process_measurement+0x9ae/0x1560 security/integrity/ima/ima_main.c:205
ima_file_check+0xc5/0x110 security/integrity/ima/ima_main.c:391
do_last fs/namei.c:3422 [inline]
path_openat+0x1130/0x45e0 fs/namei.c:3534
do_filp_open+0x1a1/0x280 fs/namei.c:3564
do_open_execat+0x140/0x660 fs/exec.c:853
__do_execve_file.isra.0+0x15a4/0x2150 fs/exec.c:1755
do_execveat_common fs/exec.c:1866 [inline]
do_execve fs/exec.c:1883 [inline]
__do_sys_execve fs/exec.c:1964 [inline]
__se_sys_execve fs/exec.c:1959 [inline]
__x64_sys_execve+0x8f/0xc0 fs/exec.c:1959
do_syscall_64+0xfd/0x620 arch/x86/entry/common.c:293
entry_SYSCALL_64_after_hwframe+0x49/0xbe
RIP: 0033:0x45aff9
Code: ad b6 fb ff c3 66 2e 0f 1f 84 00 00 00 00 00 66 90 48 89 f8 48 89 f7
48 89 d6 48 89 ca 4d 89 c2 4d 89 c8 4c 8b 4c 24 08 0f 05 <48> 3d 01 f0 ff
ff 0f 83 7b b6 fb ff c3 66 2e 0f 1f 84 00 00 00 00
RSP: 002b:00007fded7261c78 EFLAGS: 00000246 ORIG_RAX: 000000000000003b
RAX: ffffffffffffffda RBX: 00007fded72626d4 RCX: 000000000045aff9
RDX: 0000000000000000 RSI: 0000000000000000 RDI: 0000000020000040
RBP: 000000000075bf20 R08: 0000000000000000 R09: 0000000000000000
R10: 0000000000000000 R11: 0000000000000246 R12: 00000000ffffffff
R13: 00000000000000a2 R14: 00000000004c1a35 R15: 000000000075bf2c
EXT4-fs (loop0): VFS: Can't find ext4 filesystem
EXT4-fs (loop0): VFS: Can't find ext4 filesystem
EXT4-fs (loop0): VFS: Can't find ext4 filesystem
xt_NFQUEUE: number of total queues is 0
EXT4-fs (loop0): VFS: Can't find ext4 filesystem
xt_NFQUEUE: number of total queues is 0
EXT4-fs (loop0): VFS: Can't find ext4 filesystem
net_ratelimit: 11 callbacks suppressed
protocol 88fb is buggy, dev hsr_slave_0
protocol 88fb is buggy, dev hsr_slave_1
protocol 88fb is buggy, dev hsr_slave_0
protocol 88fb is buggy, dev hsr_slave_1
protocol 88fb is buggy, dev hsr_slave_0
protocol 88fb is buggy, dev hsr_slave_1
xt_NFQUEUE: number of total queues is 0
EXT4-fs (loop0): VFS: Can't find ext4 filesystem
protocol 88fb is buggy, dev hsr_slave_0
protocol 88fb is buggy, dev hsr_slave_1
xt_NFQUEUE: number of total queues is 0
EXT4-fs (loop0): VFS: Can't find ext4 filesystem
xt_NFQUEUE: number of total queues is 0
EXT4-fs (loop0): VFS: Can't find ext4 filesystem
xt_NFQUEUE: number of total queues is 0
protocol 88fb is buggy, dev hsr_slave_0
EXT4-fs (loop0): VFS: Can't find ext4 filesystem
xt_NFQUEUE: number of total queues is 0
EXT4-fs (loop0): VFS: Can't find ext4 filesystem
EXT4-fs (loop0): VFS: Can't find ext4 filesystem
protocol 88fb is buggy, dev hsr_slave_0
xt_NFQUEUE: number of total queues is 0
EXT4-fs (loop0): VFS: Can't find ext4 filesystem
xt_NFQUEUE: number of total queues is 0
xt_NFQUEUE: number of total queues is 0
EXT4-fs (loop0): VFS: Can't find ext4 filesystem
EXT4-fs (loop0): VFS: Can't find ext4 filesystem
nfqueue_tg_check: 1 callbacks suppressed
xt_NFQUEUE: number of total queues is 0
EXT4-fs (loop0): VFS: Can't find ext4 filesystem
EXT4-fs (loop0): VFS: Can't find ext4 filesystem
xt_NFQUEUE: number of total queues is 0
EXT4-fs (loop0): VFS: Can't find ext4 filesystem
net_ratelimit: 18 callbacks suppressed
protocol 88fb is buggy, dev hsr_slave_0
EXT4-fs (loop0): VFS: Can't find ext4 filesystem
protocol 88fb is buggy, dev hsr_slave_0
protocol 88fb is buggy, dev hsr_slave_0
protocol 88fb is buggy, dev hsr_slave_1
protocol 88fb is buggy, dev hsr_slave_1
protocol 88fb is buggy, dev hsr_slave_0
protocol 88fb is buggy, dev hsr_slave_0
protocol 88fb is buggy, dev hsr_slave_1
EXT4-fs (loop0): VFS: Can't find ext4 filesystem
protocol 88fb is buggy, dev hsr_slave_0
protocol 88fb is buggy, dev hsr_slave_1
xt_NFQUEUE: number of total queues is 0
EXT4-fs (loop0): VFS: Can't find ext4 filesystem
EXT4-fs (loop0): VFS: Can't find ext4 filesystem
xt_NFQUEUE: number of total queues is 0
EXT4-fs (loop0): VFS: Can't find ext4 filesystem
xt_NFQUEUE: number of total queues is 0


---
This bug is generated by a bot. It may contain errors.
See https://goo.gl/tpsmEJ for more information about syzbot.
syzbot engineers can be reached at syzk...@googlegroups.com.

syzbot will keep track of this bug report. See:
https://goo.gl/tpsmEJ#status for how to communicate with syzbot.

syzbot

unread,
Jan 11, 2023, 5:20:48 PM1/11/23
to syzkaller...@googlegroups.com
Auto-closing this bug as obsolete.
Crashes did not happen for a while, no reproducer and no activity.
Reply all
Reply to author
Forward
0 new messages