I've tried this on a system with two x 3ware (lsi) 9750-16i4e each
having 12 x Hitachi Deskstar 1TB disks formatted as raid 6; the two
hardware raids combined as a software raid 0 [*].
xfs formatted and mounted with 'noatime,noalign,nobarrier'.
Result (seems reasonably consistent):
Operations performed: 0 Read, 127458 Write, 0 Other = 127458 Total
Read 0b Written 995.77Mb Total transferred 995.77Mb (66.337Mb/sec)
8491.11 Requests/sec executed]
This with CentOS 5.8 kernel. Note that the 17TB volume is 92% full.
# xfs_bmap test_file.0
0: [0..4963295]: 21324666648..21329629943
1: [4963296..9919871]: 22779572824..22784529399
2: [9919872..14871367]: 22769382704..22774334199
3: [14871368..16777215]: 22767476856..22769382703
[*] actually a variation of raid 0 which distributes the blocks in a
pattern which compensates for the units being much faster at their edge
than at their center, to give a flatter performance curve.
On Wed, 2012-04-25 at 10:07 +0200, Stefan Ring wrote:
> This grew out of the discussion in my other thread ("Abysmal write
> performance because of excessive seeking (allocation groups to
> blame?)") -- that should in fact have been called "Free space
> fragmentation causes excessive seeks".
> Could someone with a good hardware RAID (5 or 6, but also mirrored
> setups would be interesting) please conduct a little experiment for
> I've put up a modified sysbench here:
> <https://github.com/Ringdingcoder/sysbench>. This tries to simulate
> the write pattern I've seen with XFS. It would be really interesting
> to know how different RAID controllers cope with this.
> - Checkout (or download tarball):
> - ./configure --without-mysql && make
> - fallocate -l 8g test_file.0
> - ./sysbench/sysbench --test=fileio --max-time=15
> --max-requests=10000000 --file-num=1 --file-extra-flags=direct
> --file-total-size=8G --file-block-size=8192 --file-fsync-all=off
> --file-fsync-freq=0 --file-fsync-mode=fdatasync --num-threads=1
> --file-test-mode=ag4 run
> If you don't have fallocate, you can also use the last line with "run"
> replaced by "prepare" to create the file. Run the benchmark a few
> times to check if the numbers are somewhat stable. When doing a few
> runs in direct succession, the first one will likely be faster because
> the cache has not been loaded up yet. The interesting part of the
> output is this:
> Read 0b Written 64.516Mb Total transferred 64.516Mb (4.301Mb/sec)
> 550.53 Requests/sec executed
> That's a measurement from my troubled RAID 6 volume (SmartArray P400,
> 6x 10k disks).
> >From the other controller in this machine (RAID 1, SmartArray P410i,
> 2x 15k disks), I get:
> Read 0b Written 276.85Mb Total transferred 276.85Mb (18.447Mb/sec)
> 2361.21 Requests/sec executed
> The better result might be caused by the better controller or the RAID
> 1, with the latter reason being more likely.
> xfs mailing list
Roger Willcocks <roger@xxxxxxxxxxxxxxxx>