Re: recreate superblock? (was Re: mdadm: failed to RUN_ARRAY)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Sharif,

>From the kernel log, I believe that the raid5 array now has 4 working
disks out of 6 raid disks.  My suggestion is to recreate the array.

Before recreating the MD array, I would suggest that you run mdadm
--examine /dev/hdxx (i.e. hde1, hdf1 or hdg1) and write down the
information of the original array setup. 

Some of the important Raid5 info are : Raid Devices, Total Devices,
Preferred Minor (/dev/mdx), Spare Devices, Layout, Chunk Size, order of
devices. 

When you **decide** to use mdadm -C to recreate array, mdadm will ask
you to confirm on overwriting existing md superblock.

Good luck!
Mike T.
 

> /dev/hdi is one of the bad one, but I think it is not totally gone. So I 
> was trying to force it.
> 
> [root@pep root]# mdadm -A /dev/md0 --force /dev/hdefghi]1
> mdadm: no RAID superblock on /dev/hdi1
> mdadm: /dev/hdi1 has no superblock - assembly aborted
> 
> I would appreciate some hints. Thanks.
> -Sharif
> 
> 
> Sharif Islam wrote:
> 
> > Ok, more on this. I think two of my RAID drives failed. and after  I
> > posted the email below, I did something that I am guessign I shouldn't 
> > have.
> >
> > mdadm --zero-superblock /dev/hdk1
> > mdadm --zero-superblock /dev/hdi1
> >
> > Should I try recreating the array in degrading mode.
> >
> > mdadm --create /dev/md0 --level=5 /dev/hde1 /dev/hdf1 /dev/hdg1 
> > /dev/hdh1 missing
> >
> > But I think that only works with one failed device.
> > -Sharif
> >
> >
> > Sharif Islam wrote:
> >
> >> One of my RAID 5 stopped last night.  I get this when I try to
> >> restart it:
> >> [root@pep raid]# mdadm -A /dev/md0
> >> mdadm: device 5 in /dev/md0 has wrong state in superblock, but 
> >> /dev/hdk1 seems ok
> >> mdadm: failed to RUN_ARRAY /dev/md0: Invalid argument
> >>
> >> Is my hdk1 gone. Should I try to replace that?
> >>
> >> from dmesg:
> >>
> >> Oct  9 10:36:04 pep kernel: md: export_rdev(hdf1)
> >> Oct  9 10:36:04 pep kernel:  [events: 0000000b]
> >> Oct  9 10:36:04 pep kernel: md: bind<hdf1,1>
> >> Oct  9 10:36:04 pep kernel:  [events: 0000000b]
> >> Oct  9 10:36:04 pep kernel: md: bind<hdg1,2>
> >> Oct  9 10:36:04 pep kernel:  [events: 0000000b]
> >> Oct  9 10:36:04 pep kernel: md: bind<hdh1,3>
> >> Oct  9 10:36:04 pep kernel:  [events: 00000009]
> >> Oct  9 10:36:04 pep kernel: md: bind<hdi1,4>
> >> Oct  9 10:36:04 pep kernel:  [events: 0000000a]
> >> Oct  9 10:36:04 pep kernel: md: bind<hdk1,5>
> >> Oct  9 10:36:04 pep kernel:  [events: 0000000b]
> >> Oct  9 10:36:04 pep kernel: md: bind<hde1,6>
> >> Oct  9 10:36:04 pep kernel: md: hde1's event counter: 0000000b
> >> Oct  9 10:36:04 pep kernel: md: hdk1's event counter: 0000000a
> >> Oct  9 10:36:04 pep kernel: md: hdi1's event counter: 00000009
> >> Oct  9 10:36:04 pep kernel: md: hdh1's event counter: 0000000b
> >> Oct  9 10:36:04 pep kernel: md: hdg1's event counter: 0000000b
> >> Oct  9 10:36:04 pep kernel: md: hdf1's event counter: 0000000b
> >> Oct  9 10:36:04 pep kernel: md: superblock update time inconsistency 
> >> -- using the most recent one
> >> Oct  9 10:36:04 pep kernel: md: freshest: hde1
> >> Oct  9 10:36:04 pep kernel: md: kicking non-fresh hdi1 from array!
> >> Oct  9 10:36:04 pep kernel: md: unbind<hdi1,5>
> >> Oct  9 10:36:04 pep kernel: md: export_rdev(hdi1)
> >> Oct  9 10:36:04 pep kernel: md0: removing former faulty hdi1!
> >> Oct  9 10:36:04 pep kernel: md0: kicking faulty hdk1!
> >> Oct  9 10:36:04 pep kernel: md: unbind<hdk1,4>
> >> Oct  9 10:36:04 pep kernel: md: export_rdev(hdk1)
> >> Oct  9 10:36:04 pep kernel: md: md0: raid array is not clean -- 
> >> starting background reconstruction
> >> Oct  9 10:36:04 pep kernel: md0: max total readahead window set to 2560k
> >> Oct  9 10:36:04 pep kernel: md0: 5 data-disks, max readahead per 
> >> data-disk: 512k
> >> Oct  9 10:36:04 pep kernel: raid5: device hde1 operational as raid 
> >> disk 0
> >> Oct  9 10:36:04 pep kernel: raid5: device hdh1 operational as raid 
> >> disk 3
> >> Oct  9 10:36:04 pep kernel: raid5: device hdg1 operational as raid 
> >> disk 2
> >> Oct  9 10:36:04 pep kernel: raid5: device hdf1 operational as raid 
> >> disk 1
> >> Oct  9 10:36:04 pep kernel: raid5: not enough operational devices for 
> >> md0 (2/6 failed)
> >> Oct  9 10:36:04 pep kernel: RAID5 conf printout:
> >> Oct  9 10:36:04 pep kernel:  --- rd:6 wd:4 fd:2
> >> Oct  9 10:36:04 pep kernel:  disk 0, s:0, o:1, n:0 rd:0 us:1 dev:hde1
> >> Oct  9 10:36:04 pep kernel:  disk 1, s:0, o:1, n:1 rd:1 us:1 dev:hdf1
> >> Oct  9 10:36:04 pep kernel:  disk 2, s:0, o:1, n:2 rd:2 us:1 dev:hdg1
> >> Oct  9 10:36:04 pep kernel:  disk 3, s:0, o:1, n:3 rd:3 us:1 dev:hdh1
> >> Oct  9 10:36:04 pep kernel:  disk 4, s:0, o:0, n:4 rd:4 us:1 dev:[dev 
> >> 00:00]
> >> Oct  9 10:36:04 pep kernel:  disk 5, s:0, o:0, n:5 rd:5 us:1 dev:[dev 
> >> 00:00]
> >> Oct  9 10:36:04 pep kernel: raid5: failed to run raid set md0
> >> Oct  9 10:36:04 pep kernel: md: pers->run() failed ...
> >> Oct  9 10:36:04 pep kernel: md: md0 stopped.
> >> Oct  9 10:36:04 pep kernel: md: unbind<hde1,3>
> >> Oct  9 10:36:04 pep kernel: md: export_rdev(hde1)
> >> Oct  9 10:36:04 pep kernel: md: unbind<hdh1,2>
> >> Oct  9 10:36:04 pep kernel: md: export_rdev(hdh1)
> >> Oct  9 10:36:04 pep kernel: md: unbind<hdg1,1>
> >> Oct  9 10:36:04 pep kernel: md: export_rdev(hdg1)
> >> Oct  9 10:36:04 pep kernel: md: unbind<hdf1,0>
> >> Oct  9 10:36:04 pep kernel: md: export_rdev(hdf1)
> >>
> >> [root@pep raid]# cat /proc/mdstat
> >> Personalities : [raid5]
> >> read_ahead not set
> >> unused devices: <none
> >>
> >> Thanks.
> >> -Sharif
> >>
> >>
> >
> >
> 
> -
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 

-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux