On 09/23/2013 10:17 AM, Mark Tinguely wrote:
> On 09/08/13 20:33, Dave Chinner wrote:
>> From: Dave Chinner<dchinner@xxxxxxxxxx>
>>
>> CPU overhead of buffer lookups dominate most metadata intensive
>> workloads. The thing is, most such workloads are hitting a
>> relatively small number of buffers repeatedly, and so caching
>> recently hit buffers is a good idea.
>>
> ...
>
> I think this needs more testing.
>
> I get the following panic in a loop test after a few (3-8) iterations:
>
> while true
> do
> tar zxpf xfs.tar
> cd xfs
> make
> make modules
> cd ..
> rm -r xfs
> done
>
>
> BUG: unable to handle kernel paging request at ffff880831c1d218
> IP: [<ffffffffa01886c8>] _xfs_buf_find_lookaside+0x98/0xb0 [xfs]
> PGD 1c5d067 PUD 85ffe0067 PMD 85fe51067 PTE 8000000831c1d060
> Oops: 0000 [#1] SMP DEBUG_PAGEALLOC
> Modules linked in: xfs(O) e1000e exportfs libcrc32c ext3 jbd [last unloaded:
> xfs
> ]
> CPU: 0 PID: 23423 Comm: tar Tainted: G O 3.11.0-rc1+ #3
> task: ffff880837f087a0 ti: ffff880831c46000 task.ti: ffff880831c46000
> RIP: 0010:[<ffffffffa01886c8>] [<ffffffffa01886c8>]
> _xfs_buf_find_lookaside+0x9
> 8/0xb0 [xfs]
> RSP: 0018:ffff880831c47918 EFLAGS: 00010286
> RAX: ffff880831c1d200 RBX: ffff8808372e0000 RCX: 0000000000000003
> RDX: 0000000000000011 RSI: 00000000000009c0 RDI: ffff8808372e0000
> RBP: ffff880831c47938 R08: ffff8808372e0000 R09: ffff8808376e8d80
> R10: 0000000000000010 R11: 00000000000009c0 R12: 00000000000009c0
> R13: 0000000000000010 R14: 0000000000000001 R15: 00000000000009c0
> FS: 00007fa4bc51f700(0000) GS:ffff88085bc00000(0000) knlGS:0000000000000000
> CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
> CR2: ffff880831c1d218 CR3: 000000082ed00000 CR4: 00000000000007f0
> Stack:
> ffff880831c47938 ffff880831c47aa8 0000000000000010 ffff880834ab7900
> ffff880831c479b8 ffffffffa018a679 ffff8808372e00c0 ffff88082eed01a0
> 0000000000000029 ffff8808372e01f0 0000000000000000 000200015bfe1c68
> Call Trace:
> [<ffffffffa018a679>] _xfs_buf_find+0x159/0x520 [xfs]
> [<ffffffffa018aea0>] xfs_buf_get_map+0x30/0x130 [xfs]
> [<ffffffffa018afc6>] xfs_buf_read_map+0x26/0xa0 [xfs]
> [<ffffffffa01fbf5d>] xfs_trans_read_buf_map+0x16d/0x4c0 [xfs]
> [<ffffffffa01e784c>] xfs_imap_to_bp+0x6c/0x120 [xfs]
> [<ffffffffa01e7975>] xfs_iread+0x75/0x2f0 [xfs]
> [<ffffffff8114eafb>] ? inode_init_always+0xfb/0x1c0
> [<ffffffffa019311a>] xfs_iget_cache_miss+0x5a/0x1e0 [xfs]
> [<ffffffffa01933db>] xfs_iget+0x13b/0x1c0 [xfs]
> [<ffffffffa01dfaad>] xfs_ialloc+0xbd/0x860 [xfs]
> [<ffffffffa01e02e7>] xfs_dir_ialloc+0x97/0x2e0 [xfs]
> [<ffffffffa01a2308>] ? xfs_trans_reserve+0x308/0x310 [xfs]
>
> I got the same panic running xfstest 319 with the patch at:
> http://oss.sgi.com/archives/xfs/2013-09/msg00578.html
> once it hung on a xfs_buf lock before the panic.
>
> And these are the only tests that I threw at this patch.
>
> --Mark.
I got similar issues in full runs of xfstests, but I'm having
severe setup problems here and also had to adjust the patch for
32-bit x86. Thanks for reproducing the problem on 64-bit Linux.
Michael
|