XFS corruption on 3ware RAID6-volume

Justin Piszcz jpiszcz at lucidpixels.com
Wed Feb 23 09:29:15 CST 2011



On Wed, 23 Feb 2011, Erik Gulliksson wrote:

> Hi Emmanuel,
>
> Thanks for your prompt reply.
>
> On Wed, Feb 23, 2011 at 3:46 PM, Emmanuel Florac <eflorac at intellique.com> wrote:
>>
>> What firmware version are you using?
>>
>> ( tw_cli /cX show firmware )
>
> # tw_cli /c0 show firmware
> /c0 Firmware Version = FE9X 4.10.00.007
The latest is:

9.5.1-9650-Upgrade.zip
9.5.2-9650-Upgrade.zip
9.5.3-9650-Upgrade.zip
9650SE_9690SA_firmware_beta_fw4.10.00.016.zip
9650SE_9690SA_firmware_beta_fw_4.10.00.019.zip <- latest

>
>
>>
>> Augh. That sounds pretty bad. What does " tw_cli /cX/uY show all" look
>> like?
>
> Yes, it is bad - a decision has been made to replace these disks with
> "enterprise"-versions (without TLER/ERC problems etc). Tw_cli produces
> this output for the volume:
This would seem to be the problem, you should go with Hiatchi next time.
You can use regular non-enterprise drives (Hiatchi) and they just work.
Seagate is a question
Samsung is a question
WD needs TLER.

>
> # tw_cli /c0/u0 show all
> /c0/u0 status = OK
> /c0/u0 is not rebuilding, its current state is OK
> /c0/u0 is not verifying, its current state is OK
> /c0/u0 is initialized.
> /c0/u0 Write Cache = on
> /c0/u0 Read Cache = Intelligent
> /c0/u0 volume(s) = 1
> /c0/u0 name = xxx
> /c0/u0 serial number = yyy
> /c0/u0 Ignore ECC policy = off
> /c0/u0 Auto Verify Policy = off
> /c0/u0 Storsave Policy = protection
> /c0/u0 Command Queuing Policy = on
> /c0/u0 Rapid RAID Recovery setting = all
> /c0/u0 Parity Number = 2
>
> Unit     UnitType  Status         %RCmpl  %V/I/M  Port  Stripe  Size(GB)
> ------------------------------------------------------------------------
> u0       RAID-6    OK             -       -       -     256K    12572.8
> u0-0     DISK      OK             -       -       p12   -       1396.97
> u0-1     DISK      OK             -       -       p21   -       1396.97
> u0-2     DISK      OK             -       -       p14   -       1396.97
> u0-3     DISK      OK             -       -       p15   -       1396.97
> u0-4     DISK      OK             -       -       p16   -       1396.97
> u0-5     DISK      OK             -       -       p17   -       1396.97
> u0-6     DISK      OK             -       -       p18   -       1396.97
> u0-7     DISK      OK             -       -       p19   -       1396.97
> u0-8     DISK      OK             -       -       p20   -       1396.97
> u0-9     DISK      OK             -       -       p0    -       1396.97
> u0-10    DISK      OK             -       -       p22   -       1396.97
> u0/v0    Volume    -              -       -       -     -       12572.8

As far as the problem at hand, I do not know of a good way to fix it unless
you had ls -lRi /raid_array output so you could map the inodes to their
original locations.  Sorry don't have a better answer..

Justin.




More information about the xfs mailing list