4.1 lockdep problem
Eric Paris
eparis at redhat.com
Wed Apr 15 13:47:37 CDT 2015
Booting 4.0 my system is totally fine. Although my 4.0 (probably)
doesn't have any debug/lockdep code turned on. Booting Fedora's 4.1
this morning does cause some problems.
The first time I booted, I ran dracut -f, a lockdep popped out, and
dracut never returned...
On successive boots I see that my system boots without error, but then
the lockdep pops out when I ssh in. When I reboot, sshd actually
segfaults instead of closing properly. 4.0 kernel has no such problem.
Maybe this is yet another xfs false positive, but the segfaulting sshd
is quite strange...
[ 225.300470] ======================================================
[ 225.300507] [ INFO: possible circular locking dependency detected ]
[ 225.300543] 4.1.0-0.rc0.git1.1.fc23.x86_64 #1 Not tainted
[ 225.300579] -------------------------------------------------------
[ 225.300615] sshd/11261 is trying to acquire lock:
[ 225.300650] (&isec->lock){+.+.+.}, at: [<ffffffff813a7b35>] inode_doinit_with_dentry+0xc5/0x6a0
[ 225.300700]
but task is already holding lock:
[ 225.300771] (&mm->mmap_sem){++++++}, at: [<ffffffff8120ae7f>] vm_mmap_pgoff+0x8f/0xf0
[ 225.300817]
which lock already depends on the new lock.
[ 225.300934]
the existing dependency chain (in reverse order) is:
[ 225.301012]
-> #2 (&mm->mmap_sem){++++++}:
[ 225.301012] [<ffffffff8110d647>] lock_acquire+0xc7/0x2a0
[ 225.301012] [<ffffffff8121c57c>] might_fault+0x8c/0xb0
[ 225.301012] [<ffffffff8128f49a>] filldir+0x9a/0x130
[ 225.301012] [<ffffffffa0176606>] xfs_dir2_block_getdents.isra.12+0x1a6/0x1d0 [xfs]
[ 225.301012] [<ffffffffa0177064>] xfs_readdir+0x1a4/0x330 [xfs]
[ 225.301012] [<ffffffffa0179f7b>] xfs_file_readdir+0x2b/0x30 [xfs]
[ 225.301012] [<ffffffff8128f26a>] iterate_dir+0x9a/0x140
[ 225.301012] [<ffffffff8128f7a1>] SyS_getdents+0x91/0x120
[ 225.301012] [<ffffffff8188c8ee>] system_call_fastpath+0x12/0x76
[ 225.301012]
-> #1 (&xfs_dir_ilock_class){++++.+}:
[ 225.301012] [<ffffffff8110d647>] lock_acquire+0xc7/0x2a0
[ 225.301012] [<ffffffff81105337>] down_read_nested+0x57/0xa0
[ 225.301012] [<ffffffffa018ae12>] xfs_ilock+0xe2/0x2a0 [xfs]
[ 225.301012] [<ffffffffa018b048>] xfs_ilock_attr_map_shared+0x38/0x50 [xfs]
[ 225.301012] [<ffffffffa01289dd>] xfs_attr_get+0xbd/0x1b0 [xfs]
[ 225.301012] [<ffffffffa019b01d>] xfs_xattr_get+0x3d/0x80 [xfs]
[ 225.301012] [<ffffffff812a47df>] generic_getxattr+0x4f/0x70
[ 225.301012] [<ffffffff813a7be2>] inode_doinit_with_dentry+0x172/0x6a0
[ 225.301012] [<ffffffff813a914b>] sb_finish_set_opts+0xdb/0x260
[ 225.301012] [<ffffffff813a9861>] selinux_set_mnt_opts+0x331/0x670
[ 225.301012] [<ffffffff813ac3a7>] superblock_doinit+0x77/0xf0
[ 225.301012] [<ffffffff813ac430>] delayed_superblock_init+0x10/0x20
[ 225.301012] [<ffffffff8127af4a>] iterate_supers+0xba/0x120
[ 225.301012] [<ffffffff813b1783>] selinux_complete_init+0x33/0x40
[ 225.301012] [<ffffffff813c1a43>] security_load_policy+0x103/0x640
[ 225.301012] [<ffffffff813b32d6>] sel_write_load+0xb6/0x790
[ 225.301012] [<ffffffff81276f87>] vfs_write+0xb7/0x210
[ 225.301012] [<ffffffff81277c9c>] SyS_write+0x5c/0xd0
[ 225.301012] [<ffffffff8188c8ee>] system_call_fastpath+0x12/0x76
[ 225.301012]
-> #0 (&isec->lock){+.+.+.}:
[ 225.301012] [<ffffffff8110ca82>] __lock_acquire+0x1cb2/0x1e50
[ 225.301012] [<ffffffff8110d647>] lock_acquire+0xc7/0x2a0
[ 225.301012] [<ffffffff81888c0d>] mutex_lock_nested+0x7d/0x460
[ 225.301012] [<ffffffff813a7b35>] inode_doinit_with_dentry+0xc5/0x6a0
[ 225.301012] [<ffffffff813a812c>] selinux_d_instantiate+0x1c/0x20
[ 225.301012] [<ffffffff813a2efb>] security_d_instantiate+0x1b/0x30
[ 225.301012] [<ffffffff81292e04>] d_instantiate+0x54/0x80
[ 225.301012] [<ffffffff8120631c>] __shmem_file_setup+0xdc/0x250
[ 225.301012] [<ffffffff8120a8e8>] shmem_zero_setup+0x28/0x70
[ 225.301012] [<ffffffff812286ac>] mmap_region+0x66c/0x680
[ 225.301012] [<ffffffff812289e3>] do_mmap_pgoff+0x323/0x410
[ 225.301012] [<ffffffff8120aea0>] vm_mmap_pgoff+0xb0/0xf0
[ 225.301012] [<ffffffff81226b86>] SyS_mmap_pgoff+0x116/0x2b0
[ 225.301012] [<ffffffff810232db>] SyS_mmap+0x1b/0x30
[ 225.301012] [<ffffffff8188c8ee>] system_call_fastpath+0x12/0x76
[ 225.301012]
other info that might help us debug this:
[ 225.301012] Chain exists of:
&isec->lock --> &xfs_dir_ilock_class --> &mm->mmap_sem
[ 225.301012] Possible unsafe locking scenario:
[ 225.301012] CPU0 CPU1
[ 225.301012] ---- ----
[ 225.301012] lock(&mm->mmap_sem);
[ 225.301012] lock(&xfs_dir_ilock_class);
[ 225.301012] lock(&mm->mmap_sem);
[ 225.301012] lock(&isec->lock);
[ 225.301012]
*** DEADLOCK ***
[ 225.301012] 1 lock held by sshd/11261:
[ 225.301012] #0: (&mm->mmap_sem){++++++}, at: [<ffffffff8120ae7f>] vm_mmap_pgoff+0x8f/0xf0
[ 225.301012]
stack backtrace:
[ 225.301012] CPU: 2 PID: 11261 Comm: sshd Not tainted 4.1.0-0.rc0.git1.1.fc23.x86_64 #1
[ 225.301012] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.7.5-20140709_153950- 04/01/2014
[ 225.301012] 0000000000000000 00000000445fcd3f ffff88005bd539c8 ffffffff81883265
[ 225.301012] 0000000000000000 ffffffff82b876e0 ffff88005bd53a18 ffffffff811091a6
[ 225.301012] 00000000001d8f80 ffff88005bd53a78 0000000000000001 ffff88005a882690
[ 225.301012] Call Trace:
[ 225.301012] [<ffffffff81883265>] dump_stack+0x4c/0x65
[ 225.301012] [<ffffffff811091a6>] print_circular_bug+0x206/0x280
[ 225.301012] [<ffffffff8110ca82>] __lock_acquire+0x1cb2/0x1e50
[ 225.301012] [<ffffffff810ea1d5>] ? sched_clock_local+0x25/0x90
[ 225.301012] [<ffffffff8110d647>] lock_acquire+0xc7/0x2a0
[ 225.301012] [<ffffffff813a7b35>] ? inode_doinit_with_dentry+0xc5/0x6a0
[ 225.301012] [<ffffffff81888c0d>] mutex_lock_nested+0x7d/0x460
[ 225.301012] [<ffffffff813a7b35>] ? inode_doinit_with_dentry+0xc5/0x6a0
[ 225.301012] [<ffffffff8106aca5>] ? kvm_clock_read+0x25/0x30
[ 225.301012] [<ffffffff813a7b35>] ? inode_doinit_with_dentry+0xc5/0x6a0
[ 225.301012] [<ffffffff810ea1d5>] ? sched_clock_local+0x25/0x90
[ 225.301012] [<ffffffff813a7b35>] inode_doinit_with_dentry+0xc5/0x6a0
[ 225.301012] [<ffffffff813a812c>] selinux_d_instantiate+0x1c/0x20
[ 225.301012] [<ffffffff813a2efb>] security_d_instantiate+0x1b/0x30
[ 225.301012] [<ffffffff81292e04>] d_instantiate+0x54/0x80
[ 225.301012] [<ffffffff8120631c>] __shmem_file_setup+0xdc/0x250
[ 225.301012] [<ffffffff8120a8e8>] shmem_zero_setup+0x28/0x70
[ 225.301012] [<ffffffff812286ac>] mmap_region+0x66c/0x680
[ 225.301012] [<ffffffff812289e3>] do_mmap_pgoff+0x323/0x410
[ 225.301012] [<ffffffff8120ae7f>] ? vm_mmap_pgoff+0x8f/0xf0
[ 225.301012] [<ffffffff8120aea0>] vm_mmap_pgoff+0xb0/0xf0
[ 225.301012] [<ffffffff81226b86>] SyS_mmap_pgoff+0x116/0x2b0
[ 225.301012] [<ffffffff8128db9e>] ? SyS_fcntl+0x5de/0x760
[ 225.301012] [<ffffffff810232db>] SyS_mmap+0x1b/0x30
[ 225.301012] [<ffffffff8188c8ee>] system_call_fastpath+0x12/0x76
More information about the xfs
mailing list