Re: XFS corruption on 3ware RAID6-volume

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 





On Wed, 23 Feb 2011, Erik Gulliksson wrote:

Hi Emmanuel,

Thanks for your prompt reply.

On Wed, Feb 23, 2011 at 3:46 PM, Emmanuel Florac <eflorac@xxxxxxxxxxxxxx> wrote:

What firmware version are you using?

( tw_cli /cX show firmware )

# tw_cli /c0 show firmware
/c0 Firmware Version = FE9X 4.10.00.007
The latest is:

9.5.1-9650-Upgrade.zip
9.5.2-9650-Upgrade.zip
9.5.3-9650-Upgrade.zip
9650SE_9690SA_firmware_beta_fw4.10.00.016.zip
9650SE_9690SA_firmware_beta_fw_4.10.00.019.zip <- latest




Augh. That sounds pretty bad. What does " tw_cli /cX/uY show all" look
like?

Yes, it is bad - a decision has been made to replace these disks with
"enterprise"-versions (without TLER/ERC problems etc). Tw_cli produces
this output for the volume:
This would seem to be the problem, you should go with Hiatchi next time.
You can use regular non-enterprise drives (Hiatchi) and they just work.
Seagate is a question
Samsung is a question
WD needs TLER.


# tw_cli /c0/u0 show all
/c0/u0 status = OK
/c0/u0 is not rebuilding, its current state is OK
/c0/u0 is not verifying, its current state is OK
/c0/u0 is initialized.
/c0/u0 Write Cache = on
/c0/u0 Read Cache = Intelligent
/c0/u0 volume(s) = 1
/c0/u0 name = xxx
/c0/u0 serial number = yyy
/c0/u0 Ignore ECC policy = off
/c0/u0 Auto Verify Policy = off
/c0/u0 Storsave Policy = protection
/c0/u0 Command Queuing Policy = on
/c0/u0 Rapid RAID Recovery setting = all
/c0/u0 Parity Number = 2

Unit     UnitType  Status         %RCmpl  %V/I/M  Port  Stripe  Size(GB)
------------------------------------------------------------------------
u0       RAID-6    OK             -       -       -     256K    12572.8
u0-0     DISK      OK             -       -       p12   -       1396.97
u0-1     DISK      OK             -       -       p21   -       1396.97
u0-2     DISK      OK             -       -       p14   -       1396.97
u0-3     DISK      OK             -       -       p15   -       1396.97
u0-4     DISK      OK             -       -       p16   -       1396.97
u0-5     DISK      OK             -       -       p17   -       1396.97
u0-6     DISK      OK             -       -       p18   -       1396.97
u0-7     DISK      OK             -       -       p19   -       1396.97
u0-8     DISK      OK             -       -       p20   -       1396.97
u0-9     DISK      OK             -       -       p0    -       1396.97
u0-10    DISK      OK             -       -       p22   -       1396.97
u0/v0    Volume    -              -       -       -     -       12572.8

As far as the problem at hand, I do not know of a good way to fix it unless
you had ls -lRi /raid_array output so you could map the inodes to their
original locations.  Sorry don't have a better answer..

Justin.

_______________________________________________
xfs mailing list
xfs@xxxxxxxxxxx
http://oss.sgi.com/mailman/listinfo/xfs


[Index of Archives]     [Linux XFS Devel]     [Linux Filesystem Development]     [Filesystem Testing]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux