|To:||Alan Cox <alan@xxxxxxxxxxxxxxxxxxx>|
|Subject:||Re: Performance Characteristics of All Linux RAIDs (mdadm/bonnie++)|
|From:||Bill Davidsen <davidsen@xxxxxxx>|
|Date:||Fri, 30 May 2008 08:22:12 -0400|
|Cc:||Jens Bäckman <jens.backman@xxxxxxxxx>, Justin Piszcz <jpiszcz@xxxxxxxxxxxxxxx>, linux-kernel@xxxxxxxxxxxxxxx, linux-raid@xxxxxxxxxxxxxxx, xfs@xxxxxxxxxxx|
|Organization:||TMR Associates Inc, Schenectady NY|
|References:||<alpine.DEB.firstname.lastname@example.org> <email@example.com> <20080528195752.0cdcbc6d@core> <483DE40D.firstname.lastname@example.org> <20080529122223.462bf396@core>|
|User-agent:||Mozilla/5.0 (X11; U; Linux i686; en-US; rv:220.127.116.11) Gecko/20061105 SeaMonkey/1.0.6|
Alan Cox wrote:
I really don't think that's any part of the issue, the same memory and bridge went 4-5x faster in other read cases. The truth is that the raid-1 performance is really bad, and it's the code causing it AFAIK. If you track the actual io it seems to read one drive at a time, in order, without overlap.
IIRC Justin has readahead at 16MB and chunk at 256k. I would think that if multiple devices were used at all by the md code, that the chunk rather than stripe size would be the issue. In this case the RA seems large enough to trigger good behavior, were there are available.
Note: this testing was done with an old(er) kernel, as were all of mine. Since my one large raid array has become more mission critical I'm not comfortable playing with new kernels. The fate of big, fast, and stable machines is to slide into production use. :-(
I suppose that's not a bad way to do it, I now have faith in what I'm running.
Bill Davidsen <davidsen@xxxxxxx>
"Woe unto the statesman who makes war without a reason that will still
be valid when the war is over..." Otto von Bismark
|<Prev in Thread]||Current Thread||[Next in Thread>|