Re: Recovery/Access of imsm raid via mdadm?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Ok thanks for the tips, I am imaging the disks now and will try after
that is done. Just out of curiousity what could become corrupted by
forcing the assemble? I was under the impression that as long as I
have one member missing that the only thing that would be touched is
metadata, is that right?

thanks
chris

On Mon, Jan 14, 2013 at 9:24 AM, Dorau, Lukasz <lukasz.dorau@xxxxxxxxx> wrote:
> On Monday, January 14, 2013 3:11 PM Dorau, Lukasz <lukasz.dorau@xxxxxxxxx> wrote:
>> On Monday, January 14, 2013 1:56 AM chris <tknchris@xxxxxxxxx> wrote:
>> > [292295.923942] bio: create slab <bio-1> at 1
>> > [292295.923965] md/raid:md126: not clean -- starting background
>> > reconstruction
>> > [292295.924000] md/raid:md126: device sdb operational as raid disk 2
>> > [292295.924005] md/raid:md126: device sdc operational as raid disk 1
>> > [292295.924009] md/raid:md126: device sde operational as raid disk 0
>> > [292295.925149] md/raid:md126: allocated 4250kB
>> > [292295.927268] md/raid:md126: cannot start dirty degraded array.
>>
>> Hi
>>
>> *Remember to backup the disks you have before trying the following! *
>>
>> You can try starting dirty degraded array using:
>> #  mdadm --assemble --force ....
>>
>
> I meant adding --force option to:
> # mdadm --create --verbose --force /dev/md/Volume0 /dev/sdc missing /dev/sdb /dev/sdd --raid-devices 4 --level=5
>
> Be very careful using "--force" option, because it can cause data corruption!
>
> Lukasz
>
>
>> See also the "Boot time assembly of degraded/dirty arrays" chapter in:
>> http://www.kernel.org/doc/Documentation/md.txt
>> (you can boot with option md-mod.start_dirty_degraded=1)
>>
>> Lukasz
>>
>>
>> > [292295.929666] RAID conf printout:
>> > [292295.929677]  --- level:5 rd:4 wd:3
>> > [292295.929683]  disk 0, o:1, dev:sde
>> > [292295.929688]  disk 1, o:1, dev:sdc
>> > [292295.929693]  disk 2, o:1, dev:sdb
>> > [292295.930898] md/raid:md126: failed to run raid set.
>> > [292295.930902] md: pers->run() failed ...
>> > [292295.931079] md: md126 stopped.
>> > [292295.931096] md: unbind<sdb>
>> > [292295.944228] md: export_rdev(sdb)
>> > [292295.944267] md: unbind<sdc>
>> > [292295.958126] md: export_rdev(sdc)
>> > [292295.958167] md: unbind<sde>
>> > [292295.970902] md: export_rdev(sde)
>> > [292296.219837] device-mapper: table: 252:1: raid45: unknown target type
>> > [292296.219845] device-mapper: ioctl: error adding target to table
>> > [292296.291542] device-mapper: table: 252:1: raid45: unknown target type
>> > [292296.291548] device-mapper: ioctl: error adding target to table
>> > [292296.310926] quiet_error: 1116 callbacks suppressed
>> > [292296.310934] Buffer I/O error on device dm-0, logical block 3907022720
>> > [292296.310940] Buffer I/O error on device dm-0, logical block 3907022721
>> > [292296.310944] Buffer I/O error on device dm-0, logical block 3907022722
>> > [292296.310949] Buffer I/O error on device dm-0, logical block 3907022723
>> > [292296.310953] Buffer I/O error on device dm-0, logical block 3907022724
>> > [292296.310958] Buffer I/O error on device dm-0, logical block 3907022725
>> > [292296.310962] Buffer I/O error on device dm-0, logical block 3907022726
>> > [292296.310966] Buffer I/O error on device dm-0, logical block 3907022727
>> > [292296.310973] Buffer I/O error on device dm-0, logical block 3907022720
>> > [292296.310977] Buffer I/O error on device dm-0, logical block 3907022721
>> > [292296.319968] device-mapper: table: 252:1: raid45: unknown target type
>> > [292296.319975] device-mapper: ioctl: error adding target to table
>> >
>> > Any ideas from here? Am I up the creek without a paddle? :(
>> >
>> > thanks to everyone for all your help so far
>> > chris
>> >
>> > On Sun, Jan 13, 2013 at 4:05 PM, Dan Williams <djbw@xxxxxx> wrote:
>> > >
>> > >
>> > > On 1/13/13 11:00 AM, "chris" <tknchris@xxxxxxxxx> wrote:
>> > >
>> > >>Neil/Dave,
>> > >>
>> > >>Is it not possible to create imsm container with missing disk?
>> > >>If not, Is there any way to recreate the array with all disks but
>> > >>prevent any kind of sync which may overwrite array data?
>> > >
>> > > The example was in that link I sent:
>> > >
>> > > mdadm --create /dev/md/imsm /dev/sd[bde] -e imsm
>> > > mdadm --create /dev/md/vol0 /dev/sde missing /dev/sdb /dev/sdd -n 4 -l 5
>> > >
>> > > The first command marks all devices as spares.  The second creates the
>> > > degraded array.
>> > >
>> > > You probably want at least sdb and sdd in there since they have a copy of
>> > > the metadata.
>> > >
>> > > --
>> > > Dan
>> > >
>> > --
>> > To unsubscribe from this list: send the line "unsubscribe linux-raid" in
>> > the body of a message to majordomo@xxxxxxxxxxxxxxx
>> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux