Issue on kernel 2.6.35.9 vanilla
naveen yadav
yad.naveen at gmail.com
Wed Dec 22 23:19:40 CST 2010
Hi Dave,
The question you ask about log difference, the reason is I just reboot
my PC and
again reconnect my corrupted disk again,
FYI, I already send rest of info in my previous mail, If you need more
info, i will provide
Thanks
On Wed, Dec 22, 2010 at 12:30 PM, naveen yadav <yad.naveen at gmail.com> wrote:
> Thanks a lot Dave,
>
> Please find attached log for command xfs_repair -n.
>
> Thanks a lot
>
> On Wed, Dec 22, 2010 at 11:07 AM, Dave Chinner <david at fromorbit.com> wrote:
>> On Wed, Dec 22, 2010 at 10:27:16AM +0530, naveen yadav wrote:
>>> Hi Dave,
>>>
>>> Please find attached log as suggested by you.
>>>
>>> Kind regards
>>> Naveen
>>>
>>> On Wed, Dec 22, 2010 at 3:01 AM, Dave Chinner <david at fromorbit.com> wrote:
>>> > On Tue, Dec 21, 2010 at 07:41:51PM +0530, naveen yadav wrote:
>>> >> Hi all,
>>> >>
>>> >> We have one disk that got corrupted, when I connect to my PC, haveing
>>> >> kernel version(2.6.35.9).
>>> >> The Disk mount well, but when i do 'ls; command it hangs.
>>> >
>>> > ls shouldn't hang. This should return:
>>> >
>>> >> Please find the dmesg.
>>> >> /0x22 [xfs]
>>> >> [<e0dfd2c2>] xfs_da_do_buf+0x582/0x628 [xfs]
>>> >> [<e0dfd3ce>] ? xfs_da_read_buf+0x1d/0x22 [xfs]
>>> >> [<e0dfe0d3>] ? xfs_da_node_lookup_int+0x52/0x207 [xfs]
>>> >> [<e0dfd3ce>] ? xfs_da_read_buf+0x1d/0x22 [xfs]
>>> >> [<e0dfd3ce>] xfs_da_read_buf+0x1d/0x22 [xfs]
>>> >> [<e0dfe0d3>] ? xfs_da_node_lookup_int+0x52/0x207 [xfs]
>>> >> [<e0dfe0d3>] xfs_da_node_lookup_int+0x52/0x207 [xfs]
>> .....
>>> >> c2d9d000: ff ff ff ff ff ff ff ff ff ff ff ff ff ff ff ff ................
>>> >> Filesystem "sdb2": XFS internal error xfs_da_do_buf(2) at line 2113 of
>>> >> file fs/xfs/xfs_da_btree.c. Caller 0xe0dfd3ce
>>
>> This is not in your dmesg log. When did it actually happen? Before
>> the hung task timer started to trip? From your log:
>>
>> scsi 5:0:0:0: Direct-Access SanDisk Cruzer Blade 1.00 PQ: 0 ANSI: 2
>> sd 5:0:0:0: Attached scsi generic sg1 type 0
>> sd 5:0:0:0: [sdb] 15625216 512-byte logical blocks: (8.00 GB/7.45 GiB)
>> sd 5:0:0:0: [sdb] Write Protect is off
>> sd 5:0:0:0: [sdb] Mode Sense: 03 00 00 00
>> sd 5:0:0:0: [sdb] Assuming drive cache: write through
>> sd 5:0:0:0: [sdb] Assuming drive cache: write through
>> sdb: sdb1 sdb2
>> sd 5:0:0:0: [sdb] Assuming drive cache: write through
>> sd 5:0:0:0: [sdb] Attached SCSI removable disk
>> SELinux: initialized (dev sdb1, type vfat), uses genfs_contexts
>> XFS mounting filesystem sdb2
>> Starting XFS recovery on filesystem: sdb2 (logdev: internal)
>> Ending XFS recovery on filesystem: sdb2 (logdev: internal)
>> SELinux: initialized (dev sdb2, type xfs), uses xattr
>> INFO: task gvfs-gdu-volume:2311 blocked for more than 120 seconds.
>> "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
>> gvfs-gdu-volu D 00000026 0 2311 1 0x00000080
>> c6cf9b2c 00000086 a41cc623 00000026 c0a25e00 c0a25e00 c0a25e00 c0a25e00
>> d1290f54 c0a25e00 c0a25e00 000336ad 00000000 cd871c00 00000026 d1290cd0
>> 00000000 cd8d2a08 cd8d2a00 7fffffff 7fffffff c6cf9b70 c0781c43 00000000
>> Call Trace:
>> [<c0781c43>] schedule_timeout+0x1b/0x95
>> [<c07824d1>] __down_common+0x82/0xb9
>> [<e0e28ae8>] ? _xfs_buf_find+0x122/0x1b8 [xfs]
>> [<c0782567>] __down+0x17/0x19
>> [<c045827c>] down+0x27/0x37
>> [<e0e278da>] xfs_buf_lock+0x67/0x93 [xfs]
>> [<e0e28ae8>] _xfs_buf_find+0x122/0x1b8 [xfs]
>> [<e0e28bde>] xfs_buf_get+0x60/0x149 [xfs]
>> [<e0e28ce9>] xfs_buf_read+0x22/0xb0 [xfs]
>> [<e0e1ffa9>] xfs_trans_read_buf+0x53/0x2e9 [xfs]
>> [<e0dfd151>] xfs_da_do_buf+0x411/0x628 [xfs]
>> [<e0dfe0d3>] ? xfs_da_node_lookup_int+0x52/0x207 [xfs]
>> [<e0dfd3ce>] xfs_da_read_buf+0x1d/0x22 [xfs]
>> [<e0dfe0d3>] ? xfs_da_node_lookup_int+0x52/0x207 [xfs]
>> [<e0dfe0d3>] xfs_da_node_lookup_int+0x52/0x207 [xfs]
>> [<e0e03888>] xfs_dir2_node_lookup+0x5f/0xee [xfs]
>> [<e0dff26a>] xfs_dir_lookup+0xde/0x110 [xfs]
>> [<e0e22c0a>] xfs_lookup+0x50/0x9f [xfs]
>> [<e0e2c5a6>] xfs_vn_lookup+0x3e/0x76 [xfs]
>> [<c04da3b2>] do_lookup+0xc9/0x139
>> [<c04dbd59>] do_last+0x186/0x49f
>> [<c04dc415>] do_filp_open+0x1bd/0x459
>> [<c045b191>] ? timekeeping_get_ns+0x16/0x54
>> [<c05a9170>] ? might_fault+0x1e/0x20
>> [<c04e479a>] ? alloc_fd+0x58/0xbe
>> [<c04d1941>] do_sys_open+0x4d/0xe4
>> [<c047d0f4>] ? audit_syscall_entry+0x12a/0x14c
>> [<c04d1a24>] sys_open+0x23/0x2b
>> [<c0407fd8>] sysenter_do_call+0x12/0x2d
>>
>> You've got a gvfs (gnome-vfs?) process stuck waiting on a buffer
>> lock. The onyl way I can see it getting stuck here is if it the
>> buffer has not been unlocked somewhere. It's possible that it is
>> stuck on the same buffer that the corruption error came from,
>> but the corrupted buffer is unlocked in the error handling path.
>> what does `xfs_repair -n` tell about the filesytsem?
>>
>> FWIW, later on:
>>
>> ......
>> INFO: task gvfsd-trash:1891 blocked for more than 120 seconds.
>> "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
>> gvfsd-trash D 00000026 0 1891 1 0x00000080
>>
>> gvfs-trashd gets stuck on a mutex during a path walk which is
>> probably held by the above directory read.
>>
>> ....
>> INFO: task gvfs-gdu-volume:2321 blocked for more than 120 seconds.
>> "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
>> gvfs-gdu-volu D 00000026 0 2321 1 0x00000088
>>
>> As does this one.
>>
>> ....
>>
>> gvfs-gdu-volu D 00000026 0 1889 1 0x00000080
>> cda3df10 00000086 a422923c 00000026 c0a25e00 c0a25e00 c0a25e00 c0a25e00
>> cd936904 c0a25e00 c0a25e00 000b9ea1 00000000 cda24400 00000026 cd936680
>> 00ae3000 c34c01a4 c34c019c cd936680 c34c01a0 cda3df44 c0782093 c34c01ac
>> Call Trace:
>> [<c0782093>] __mutex_lock_common+0xe8/0x137
>> [<c0782113>] __mutex_lock_killable_slowpath+0x17/0x19
>> [<c0782160>] ? mutex_lock_killable+0x32/0x45
>> [<c0782160>] mutex_lock_killable+0x32/0x45
>> [<c04deb0a>] vfs_readdir+0x46/0x94
>> [<c04de814>] ? filldir64+0x0/0xf5
>> [<c04debca>] sys_getdents64+0x72/0xb2
>> [<c0407fd8>] sysenter_do_call+0x12/0x2d
>>
>> And this one, too.
>>
>> .....
>> ls D 00000000 0 2325 2044 0x00000080
>> c6d03f10 00200086 00000000 00000000 c0a25e00 c0a25e00 c0a25e00 c0a25e00
>> cdacb5c4 c0a25e00 c0a25e00 d6779fa1 00000029 00000000 00000029 cdacb340
>> 00000001 c34c01a4 c34c019c cdacb340 c34c01a0 c6d03f44 c0782093 c34c01ac
>> Call Trace:
>> [<c0782093>] __mutex_lock_common+0xe8/0x137
>> [<c0782113>] __mutex_lock_killable_slowpath+0x17/0x19
>> [<c0782160>] ? mutex_lock_killable+0x32/0x45
>> [<c0782160>] mutex_lock_killable+0x32/0x45
>> [<c04deb0a>] vfs_readdir+0x46/0x94
>> [<c04de814>] ? filldir64+0x0/0xf5
>> [<c04debca>] sys_getdents64+0x72/0xb2
>> [<c0407fd8>] sysenter_do_call+0x12/0x2d
>>
>> And finally, there is an ls process that is hung, stuck on a
>> directory mutex. Is this the one you were seeing hang rather than
>> whatever generated the corrupion report?
>>
>> Cheers,
>>
>> Dave.
>> --
>> Dave Chinner
>> david at fromorbit.com
>>
>
More information about the xfs
mailing list