Ok, now it's even stanger. I changed the sata cables to other sata ports and now the matrix storage manager says I have 4 volumes: 0 RAID0:1 80Gb Failed 1 RAID1:1 109.0Gb Degraded 2 RAID0 80Gb Failed 3 RAID1 109.0Gb Degraded Port 0 Hitachi 149.1GB Member Disk(0,1) 1 Hitachi 149.1GB Member Disk(2,3) While researching I found that with that this "offline member" error happens a lot to intel matrix raid users. Some just a few hours after setup because of a bad shutdown. What does this mean? Is a UPS needed for matrix raid? If someone can use the metadata to investigate this, please ask me for it until I redo the arrays. On 9/26/07, Tiago Freitas <tiago.frt@xxxxxxxxx> wrote: > I forgot to tell that after the errors in the opensuse live, I > rebooted and the intel raid bios said the raid0 volume had failed and > the raid1 was degraded. It then said the raid1 should be rebuilded in > the OS and rebooted. In the following reboot, it said the disks were > offline. > > I'm pretty sure the problem originated in the opensuse livecd, because > both volumes were working in windows before I booted the livecd. > Novell bug report: > https://bugzilla.novell.com/show_bug.cgi?id=328388 > It might be related to this bug: > https://bugs.launchpad.net/ubuntu/+bug/141435 > > > So it's not possible to repair the metadata? Shouldn't I dump it and > send it to someone? > > When I do the matrix raid again, if I backup the metadata and > something happens to it, can I put it back in place? How can I do > that? > > > On 9/26/07, Hall, Eric R <eric.r.hall@xxxxxxxxx> wrote: > > Tiago - > > > > Sounds like you were using a 'matrix raid.' Both sda and sdb had a two > > different raid levels on them. > > > > In my experience when the OROM/Intel Matrix Storage manager reports a > > 'member offline' it is because a setting was changed in the BIOS, a hard > > disk was attached to the system after the BIOS had indexed it but before > > the OROM indexed it, or the disk was responding but failed to spin up. > > > > Verify that the settings in the BIOS for IDE are set as "RAID" and that > > all the sata ports are enabled. Next, try to reseat the cables between > > the drives and the motherboard. > > > > The released version of dmraid rc14 does not have the ability to > > re-write metadata on the disks for Intel OROM. I do not think that > > dmraid or device-mapper caused the problem. The OROM is still reading > > the metadata from the disks otherwise they would not show up as a > > 'member' in the "DISK/VOLUME Information" window. > > > > Currently there is no way to rebuild a set with ISW metadata in Linux. > > You _might_ be able to recover the data from the RAID 1 by using the > > Windows iMSM (Intel Matrix Storage Manager) software; which can be > > downloaded via support.intel.com > > > > - Eric > > > > > > > > > > >-----Original Message----- > > >From: ataraid-list-bounces@xxxxxxxxxx [mailto:ataraid-list- > > >bounces@xxxxxxxxxx] On Behalf Of Tiago Freitas > > >Sent: Tuesday, September 25, 2007 4:36 PM > > >To: ataraid-list@xxxxxxxxxx > > >Subject: isw device for volume broken after opensuse livecd boot > > > > > >Hi. > > >I had a working raid array in an Intel ICH9R controller with the two > > >volumes working ( ok in windows and mandriva 2008 livecd). > > > > > >I booted the Opensuse 10.3RC2 live cd and did "dmraid -ay" and got the > > >following errors: > > > > > >ERROR: isw device for volume "RAID0" broken on /dev/sda in RAID set > > >"isw_bbcfceegje_RAID0" > > >ERROR: isw: wrong # of devices in RAID set "isw_bbcfceegje_RAID0" > > >[1/2] on /dev/sda > > >ERROR: isw device for volume "RAID1" broken on /dev/sda in RAID set > > >"isw_bbcfceegje_RAID1" > > >ERROR: isw: wrong # of devices in RAID set "isw_bbcfceegje_RAID1" > > >[1/2] on /dev/sda > > >ERROR: isw device for volume "RAID0" broken on /dev/sdb in RAID set > > >"isw_chgffgbhfa_RAID0" > > >ERROR: isw: wrong # of devices in RAID set "isw_chgffgbhfa_RAID0" > > >[1/2] on /dev/sdb > > >ERROR: isw device for volume "RAID1" broken on /dev/sdb in RAID set > > >"isw_chgffgbhfa_RAID1" > > >ERROR: isw: wrong # of devices in RAID set "isw_chgffgbhfa_RAID1" > > >[1/2] on /dev/sdb > > >ERROR: no mapping possible for RAID set isw_bbcfceegje_RAID1 > > >ERROR: no mapping possible for RAID set isw_chgffgbhfa_RAID1 > > > > > >Now in the intel bios it says the disks are offline. > > >The livecd must have a bug that destroys all raid arrays. I think the > > >problem is in device-mapper, because I did "dmraid --version" and the > > >device-mapper was "unknown". In mandriva it is 4.11.0. Dmraid is rc14 > > >in both. > > >I don't have any logs from the live boot because the network was not > > >working. > > > > > >Is there any way to recover the raid volumes? I can provide you with > > >all the files needed, just ask. > > > > > >Thanks > > > > > >Tiago Freitas > > > > > >_______________________________________________ > > >Ataraid-list mailing list > > >Ataraid-list@xxxxxxxxxx > > >https://www.redhat.com/mailman/listinfo/ataraid-list > > > _______________________________________________ Ataraid-list mailing list Ataraid-list@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/ataraid-list