Re: isw device for volume broken after opensuse livecd boot

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Ok, now it's even stanger. I changed the sata cables to other sata
ports and now the matrix storage manager says I have 4 volumes:

0  RAID0:1   80Gb       Failed
1  RAID1:1  109.0Gb   Degraded
2  RAID0     80Gb        Failed
3  RAID1     109.0Gb   Degraded

Port
0     Hitachi        149.1GB     Member Disk(0,1)
1     Hitachi        149.1GB     Member Disk(2,3)

While researching I found that with that this "offline member" error
happens a lot to intel matrix raid users. Some just a few hours after
setup because of a bad shutdown. What does this mean? Is a UPS needed
for matrix raid?

If someone can use the metadata to investigate this, please ask me for
it until I redo the arrays.

On 9/26/07, Tiago Freitas <tiago.frt@xxxxxxxxx> wrote:
> I forgot to tell that after the errors in the opensuse live, I
> rebooted and the intel raid bios said the raid0 volume had failed and
> the raid1 was degraded. It then said the raid1 should be rebuilded in
> the OS and rebooted. In the following reboot, it said the disks were
> offline.
>
> I'm pretty sure the problem originated in the opensuse livecd, because
> both volumes were working in windows before I booted the livecd.
> Novell bug report:
> https://bugzilla.novell.com/show_bug.cgi?id=328388
> It might be related to this bug:
> https://bugs.launchpad.net/ubuntu/+bug/141435
>
>
> So it's not possible to repair the metadata? Shouldn't I dump it and
> send it to someone?
>
> When I do the matrix raid again, if I backup the metadata and
> something happens to it, can I put it back in place? How can I do
> that?
>
>
> On 9/26/07, Hall, Eric R <eric.r.hall@xxxxxxxxx> wrote:
> > Tiago -
> >
> > Sounds like you were using a 'matrix raid.'  Both sda and sdb had a two
> > different raid levels on them.
> >
> > In my experience when the OROM/Intel Matrix Storage manager reports a
> > 'member offline' it is because a setting was changed in the BIOS, a hard
> > disk was attached to the system after the BIOS had indexed it but before
> > the OROM indexed it, or the disk was responding but failed to spin up.
> >
> > Verify that the settings in the BIOS for IDE are set as "RAID" and that
> > all the sata ports are enabled.  Next, try to reseat the cables between
> > the drives and the motherboard.
> >
> > The released version of dmraid rc14 does not have the ability to
> > re-write metadata on the disks for Intel OROM.  I do not think that
> > dmraid or device-mapper caused the problem.  The OROM is still reading
> > the metadata from the disks otherwise they would not show up as a
> > 'member' in the "DISK/VOLUME Information" window.
> >
> > Currently there is no way to rebuild a set with ISW metadata in Linux.
> > You _might_ be able to recover the data from the RAID 1 by using the
> > Windows iMSM (Intel Matrix Storage Manager) software; which can be
> > downloaded via support.intel.com
> >
> > - Eric
> >
> >
> >
> >
> > >-----Original Message-----
> > >From: ataraid-list-bounces@xxxxxxxxxx [mailto:ataraid-list-
> > >bounces@xxxxxxxxxx] On Behalf Of Tiago Freitas
> > >Sent: Tuesday, September 25, 2007 4:36 PM
> > >To: ataraid-list@xxxxxxxxxx
> > >Subject: isw device for volume broken after opensuse livecd boot
> > >
> > >Hi.
> > >I had a working raid array in an Intel ICH9R controller with the two
> > >volumes working ( ok in windows and mandriva 2008 livecd).
> > >
> > >I booted the Opensuse 10.3RC2 live cd and did "dmraid -ay" and got the
> > >following errors:
> > >
> > >ERROR: isw device for volume "RAID0" broken on /dev/sda in RAID set
> > >"isw_bbcfceegje_RAID0"
> > >ERROR: isw: wrong # of devices in RAID set "isw_bbcfceegje_RAID0"
> > >[1/2] on /dev/sda
> > >ERROR: isw device for volume "RAID1" broken on /dev/sda in RAID set
> > >"isw_bbcfceegje_RAID1"
> > >ERROR: isw: wrong # of devices in RAID set "isw_bbcfceegje_RAID1"
> > >[1/2] on /dev/sda
> > >ERROR: isw device for volume "RAID0" broken on /dev/sdb in RAID set
> > >"isw_chgffgbhfa_RAID0"
> > >ERROR: isw: wrong # of devices in RAID set "isw_chgffgbhfa_RAID0"
> > >[1/2] on /dev/sdb
> > >ERROR: isw device for volume "RAID1" broken on /dev/sdb in RAID set
> > >"isw_chgffgbhfa_RAID1"
> > >ERROR: isw: wrong # of devices in RAID set "isw_chgffgbhfa_RAID1"
> > >[1/2] on /dev/sdb
> > >ERROR: no mapping possible for RAID set isw_bbcfceegje_RAID1
> > >ERROR: no mapping possible for RAID set isw_chgffgbhfa_RAID1
> > >
> > >Now in the intel bios it says the disks are offline.
> > >The livecd must have a bug that destroys all raid arrays. I think the
> > >problem is in device-mapper, because I did "dmraid --version" and the
> > >device-mapper was "unknown". In mandriva it is 4.11.0. Dmraid is rc14
> > >in both.
> > >I don't have any logs from the live boot because the network was not
> > >working.
> > >
> > >Is there any way to recover the raid volumes? I can provide you with
> > >all the files needed, just ask.
> > >
> > >Thanks
> > >
> > >Tiago Freitas
> > >
> > >_______________________________________________
> > >Ataraid-list mailing list
> > >Ataraid-list@xxxxxxxxxx
> > >https://www.redhat.com/mailman/listinfo/ataraid-list
> >
>

_______________________________________________
Ataraid-list mailing list
Ataraid-list@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/ataraid-list

[Index of Archives]     [Linux RAID]     [Linux Device Mapper]     [Linux IDE]     [Linux SCSI]     [Kernel]     [Linux Books]     [Linux Admin]     [GFS]     [RPM]     [Yosemite Campgrounds]     [AMD 64]

  Powered by Linux