[PATCH 0/3] xfs: allocation worker causes freelist buffer lock hang
Mark Tinguely
tinguely at sgi.com
Wed Sep 26 09:14:14 CDT 2012
On 09/25/12 17:01, Dave Chinner wrote:
> On Tue, Sep 25, 2012 at 10:14:16AM -0500, Mark Tinguely wrote:
<deletes>
>>>>
>>>> As a bonus, consolidating the loops into one worker actually gives a slight
>>>> performance advantage.
>>>
>>> Can you quantify it?
>>
>> I was comparing the bonnie and iozone benchmarks outputs. I will see
>> if someone can enlighten me on how to quantify those numbers.
>
> Ugh.
>
> Don't bother. Those are two of the worst offenders in the "useless
> benchmarks for regression testing" category. Yeah, they *look* like
> they give decent numbers, but I've wasted so much time looking at
> results from these benhmarks only to find they do basic things wrong
> and give numbers that vary simple because you've made a change that
> increases or decreases the CPU cache footprint of a code path.
>
> e.g. IOZone uses the same memory buffer as the source/destination of
> all it's IO, and does not touch the contents of it at all. Hence for
> small IO, the buffer stays resident in the CPU caches and gives
> unrealsitically high throughput results. Worse is the fact that CPU
> cache residency of the buffer can change according to the kernel
> code path taken, so you can get massive changes in throughput just
> by changing the layout of the code without changing any logic....
>
> IOZone can be useful if you know exactly what you are doing and
> using it to test a specific code path with a specific set of
> configurations. e.g. comparing ext3/4/xfs/btrfs on the same kernel
> and storage is fine. However, the moment you start using it to
> compare different kernels, it's a total crap shoot....
does anyone have a good benchmark XFS should use to share performance
results? A number that we can agree a series does not degrade the
filesystem..
lies, damn lies, statistics and then filesystem benchmarks?! :)
> I guess I don't understand what you mean by "loop on
> xfs_alloc_vextent()" then.
>
> The problem I see above is this:
>
> thread 1 worker 1 worker 2..max
> xfs_bmapi_write(userdata)
loops here calling xfs_bmapi_alloc()
> xfs_bmapi_allocate(user)
> xfs_alloc_vextent(user)
> wait
>
> _xfs_alloc_vextent()
> locks AGF
first loop it takes the lock
one of the next times through the above
loop it cannot get a worker. deadlock here.
I saved the xfs_bmalloca and fs_alloc_arg when
allocating a buffer to verify the paths.
> _xfs_alloc_vextent()
> blocks on AGF lock
> completes allocation
>
> <returns with AGF locked in transaction>
> xfs_bmap_add_extent_hole_real
> xfs_bmap_extents_to_btree
> xfs_alloc_vextent(user)
> wait
this does not need a worker, and since in the same
transaction all locks to the AGF buffer are recursive locks.
no wait here.
>
> <deadlock as no more workers available>
<deletes>
--Mark.
More information about the xfs
mailing list