xfs
[Top] [All Lists]

Re: XFS corruption on 3ware RAID6-volume

To: Erik Gulliksson <erik@xxxxxxxxxxxxxx>
Subject: Re: XFS corruption on 3ware RAID6-volume
From: Justin Piszcz <jpiszcz@xxxxxxxxxxxxxxx>
Date: Wed, 23 Feb 2011 10:29:15 -0500 (EST)
Cc: Emmanuel Florac <eflorac@xxxxxxxxxxxxxx>, xfs@xxxxxxxxxxx
In-reply-to: <AANLkTinsfwr5E7KkffwWOweWJLCmaLnLdtYA4g_m--b0@xxxxxxxxxxxxxx>
References: <AANLkTinWByYooMnPL7BryPowDexBeiHJdh3aVh+fdm-a@xxxxxxxxxxxxxx> <20110223154651.54f0a8dc@xxxxxxxxxxxxxxxxxxxx> <AANLkTinsfwr5E7KkffwWOweWJLCmaLnLdtYA4g_m--b0@xxxxxxxxxxxxxx>
User-agent: Alpine 2.02 (DEB 1266 2009-07-14)


On Wed, 23 Feb 2011, Erik Gulliksson wrote:

Hi Emmanuel,

Thanks for your prompt reply.

On Wed, Feb 23, 2011 at 3:46 PM, Emmanuel Florac <eflorac@xxxxxxxxxxxxxx> wrote:

What firmware version are you using?

( tw_cli /cX show firmware )

# tw_cli /c0 show firmware
/c0 Firmware Version = FE9X 4.10.00.007
The latest is:

9.5.1-9650-Upgrade.zip
9.5.2-9650-Upgrade.zip
9.5.3-9650-Upgrade.zip
9650SE_9690SA_firmware_beta_fw4.10.00.016.zip
9650SE_9690SA_firmware_beta_fw_4.10.00.019.zip <- latest




Augh. That sounds pretty bad. What does " tw_cli /cX/uY show all" look
like?

Yes, it is bad - a decision has been made to replace these disks with
"enterprise"-versions (without TLER/ERC problems etc). Tw_cli produces
this output for the volume:
This would seem to be the problem, you should go with Hiatchi next time.
You can use regular non-enterprise drives (Hiatchi) and they just work.
Seagate is a question
Samsung is a question
WD needs TLER.


# tw_cli /c0/u0 show all
/c0/u0 status = OK
/c0/u0 is not rebuilding, its current state is OK
/c0/u0 is not verifying, its current state is OK
/c0/u0 is initialized.
/c0/u0 Write Cache = on
/c0/u0 Read Cache = Intelligent
/c0/u0 volume(s) = 1
/c0/u0 name = xxx
/c0/u0 serial number = yyy
/c0/u0 Ignore ECC policy = off
/c0/u0 Auto Verify Policy = off
/c0/u0 Storsave Policy = protection
/c0/u0 Command Queuing Policy = on
/c0/u0 Rapid RAID Recovery setting = all
/c0/u0 Parity Number = 2

Unit     UnitType  Status         %RCmpl  %V/I/M  Port  Stripe  Size(GB)
------------------------------------------------------------------------
u0       RAID-6    OK             -       -       -     256K    12572.8
u0-0     DISK      OK             -       -       p12   -       1396.97
u0-1     DISK      OK             -       -       p21   -       1396.97
u0-2     DISK      OK             -       -       p14   -       1396.97
u0-3     DISK      OK             -       -       p15   -       1396.97
u0-4     DISK      OK             -       -       p16   -       1396.97
u0-5     DISK      OK             -       -       p17   -       1396.97
u0-6     DISK      OK             -       -       p18   -       1396.97
u0-7     DISK      OK             -       -       p19   -       1396.97
u0-8     DISK      OK             -       -       p20   -       1396.97
u0-9     DISK      OK             -       -       p0    -       1396.97
u0-10    DISK      OK             -       -       p22   -       1396.97
u0/v0    Volume    -              -       -       -     -       12572.8

As far as the problem at hand, I do not know of a good way to fix it unless
you had ls -lRi /raid_array output so you could map the inodes to their
original locations.  Sorry don't have a better answer..

Justin.

<Prev in Thread] Current Thread [Next in Thread>