[ cc'd XFS list ]
On Tue, Jun 08, 2010 at 01:55:51PM +0400, Michael Tokarev wrote:
> I've got a.. difficult issue here, and am asking if anyone else
> has some expirence or information about it.
> Production environment (database). Machine with an Adaptec
> RAID SCSI controller, 6 drives in raid10 array, XFS filesystem
> and Oracle database on top of it (with - hopefully - proper
> Upgrading kernel from 2.6.27 to 2.6.32, and users starts screaming
> about very bad performance. Iostat reports increased I/O latencies,
> I/O time increases from ~5ms to ~30ms. Switching back to 2.6.27,
> and everything is back to normal (or, rather, usual).
> I tried testing I/O with a sample program which performs direct random
> I/O on a given device, and all speeds are actually better in .32
> compared with .27, except of random concurrent r+w test, where .27
> gives a bit more chances to reads than .32. Looking at the synthetic
> tests I'd expect .32 to be faster, but apparently it is not.
> This is only one machine here which is still running 2.6.27, all the
> rest are upgraded to 2.6.32, and I see good performance of .32 there.
> But this is also the only machine with hardware raid controller, which
> is onboard and hence not easy to get rid of, so I'm sorta forced to
> use it (I prefer software raid solution because of numerous reasons).
> One possible cause of this that comes to mind is block device write
> barriers. But I can't find when they're actually implemented.
> The most problematic issue here is that this is only one machine that
> behaves like this, and it is a production server, so I've very little
> chances to experiment with it.
> So before the next try, I'd love to have some suggestions about what
> to look for. In particular, I think it's worth the effort to look
> at write barriers, but again, I don't know how to check if they're
> actually being used.
> Anyone have suggestions for me to collect and to look at?