On Wed, Dec 04, 2013 at 06:55:05PM -0800, Mike Dacre wrote:
> Hi Folks,
> Apologies if this is the wrong place to post or if this has been answered
> I have a 16 2TB drive RAID6 array powered by an LSI 9240-4i. It has an XFS
> filesystem and has been online for over a year. It is accessed by 23
> different machines connected via Infiniband over NFS v3. I haven't had any
> major problems yet, one drive failed but it was easily replaced.
> However, today the drive suddenly stopped responding and started returning
> IO errors when any requests were made. This happened while it was being
> accessed by 5 different users, one was doing a very large rm operation (rm
> *sh on thousands on files in a directory). Also, about 30 minutes before
> we had connected the globus connect endpoint to allow easy file transfers
> to SDSC.
So, you had a drive die and at roughly the same time XFS started
reporting corruption problems and shut down? Chances are that the
drive returned garbage to XFS before died completely and that's what
XFS detected and shut down on. If you are unlucky in this situation,
the corruption can get propagated into the log by changes that are
adjacent to the corrupted region, and then you have problems with log
recovery failing because the corruption gets replayed....
> I have attached the complete log from the time it died until now.
> In the end, I successfully repaired the filesystem with `xfs_repair -L
> /dev/sda1`. However, I am nervous that some files may have been corrupted.
> Do any of you have any idea what could have caused this problem?
When corruption appears at roughly the same time a drive dies, it's
almost always caused by the drive that failed. RAID doesn't repvent
disks from returning crap to the OS because nobody configures the
arrays to do read-verify cycles that would catch such a condition.