Re: re-adding a disk to a raid1 array with bitmap

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Fri, 09 Mar 2012 17:07:42 +0100 Jes Sorensen <Jes.Sorensen@xxxxxxxxxx>
wrote:

> Neil,
> 
> I have been spinning my head over this for a bit trying to figure out
> what is the right solution to this problem.
> 
> In bedd86b7773fd97f0d708cc0c371c8963ba7ba9a you added a test to reject
> re-adding a drive to an array in some cases.
> 
> The problem I have been looking at is if one has a raid1 with a bitmap.
> Basically in the situation where we have one of the drives pulled from
> the array, then if I try to add it back, it fails like this:
> 
> [root@monkeybay ~]#  mdadm -I --run /dev/sdf5
> mdadm: failed to add /dev/sdf5 to /dev/md32: Invalid argument.
> 
> However this works:
> 
> [root@monkeybay ~]# mdadm -a /dev/md32 /dev/sdf5
> mdadm: re-added /dev/sdf5
> 
> I dug through the kernel and it shows up that the failure is due to this
> test in the above mentioned commit:
> 
> +                    rdev->raid_disk != info->raid_disk)) {
> 
> So basically when doing -I it seems the disk itself expects to be
> raid_disk = 0, whereas the kernel expects it should be raid_disk = 1.
> 
> I agree with the previous discussion that it makes sense to reject a
> drive in the normal case without a bitmap. However it seems illogical to
> me that -a works but -I should fail in this case.
> 
> What would be the right fix here? Relaxing the test in the kernel to not
> require the raid_disk numbers match up for a bitmap raid, or should
> mdadm be taught to examine the raids and set the expected disk number
> before submitting the add_new_disk ioctl?
> 
> Cheers,
> Jes

Does this patch fix it?

http://neil.brown.name/git?p=mdadm;a=commitdiff;h=69fe207ed68e560d76a592fd86af32a9d1deca25

I found it in a collection of half-forgotten patches recently and decided it
was almost certainly correct, but I didn't remember what motivated it.

It is entirely possible that it was seeing something like the problem you
mention.

The comment for that patch says "This is particularly important for getting
info.disk.state correct", but maybe it is equally important for getting 
info.disk.raid_disk correct.

So I think your last suggestion is right: "mdadm be taught to examine the
raids and set the expected disk number" - and that patch should do the trick.

(and thanks for the reminder to reply to this).

Note that you might need
   POLICY action=re-add
in mdadm.conf for this to work.  I don't think it is a given that when a
recently failed disk is found that it should always be re-added.  So if the
admin wants that it is reasonable to require the policy be explicitly stated.

But I'm not sure if this is currently enforced....


NeilBrown

Attachment: signature.asc
Description: PGP signature


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux