Re: raid6 rebuild not starting

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Mon, Dec 12, 2011 at 9:10 AM, NeilBrown <neilb@xxxxxxx> wrote:
> On Mon, 12 Dec 2011 08:42:49 +0200 Anssi Hannula <anssi.hannula@xxxxxx> wrote:
>
>> On Mon, Dec 12, 2011 at 8:24 AM, NeilBrown <neilb@xxxxxxx> wrote:
>> > On Mon, 12 Dec 2011 08:02:33 +0200 Anssi Hannula <anssi.hannula@xxxxxx> wrote:
>> >
>> >> On Mon, Dec 12, 2011 at 7:42 AM, NeilBrown <neilb@xxxxxxx> wrote:
>> >> > On Mon, 12 Dec 2011 07:22:17 +0200 Anssi Hannula <anssi.hannula@xxxxxx> wrote:
>> >> >
>> >> >> On Mon, Dec 12, 2011 at 5:01 AM, NeilBrown <neilb@xxxxxxx> wrote:
>> >> >> > On Sun, 11 Dec 2011 09:03:14 +0200 Anssi Hannula <anssi.hannula@xxxxxx> wrote:
>> >> >> >
>> >> >> >> Hi!
>> >> >> >>
>> >> >> >> After I rebooted during a raid6 rebuild, the rebuild didn't start again.
>> >> >> >> Instead, there is a flood of "RAID conf printout"s that seemingly happen
>> >> >> >> on array activity.
>> >> >> >>
>> >> >> >> All the devices show up properly in --detail and two devices are marked
>> >> >> >> as "spare rebuilding", and I can access the contents of the array just
>> >> >> >> fine, but the rebuild doesn't actually start. Is this a bug or am I
>> >> >> >> missing something? :)
>> >> >> >>
>> >> >> >> I was initially on 2.6.38.8, but also tried 3.1.4 which seems to have
>> >> >> >> the same issue. mdadm is 3.1.5.
>> >> >> >>
>> >> >> >> I'm not using start_ro and writing to the array doesn't trigger a
>> >> >> >> rebuild either.
>> >> >> >>
>> >> >> >> Attached are --examine outputs before assembly, kernel log output on
>> >> >> >> assembly, /proc/mdstat and --detail after assembly (on 3.1.4).
>> >> >> >>
>> >> >> >
>> >> >> > Thank you for the very detailed problem report.
>> >> >>
>> >> >> Thanks for the quick response :)
>> >> >>
>> >> >> > Unfortunately it is a complete mystery to me what is happening.
>> >> >> >
>> >> >> > The repeated "RAID conf printout" messages are almost certainly coming from
>> >> >> > the end of raid5_remove_disk.
>> >> >> > It is being called from remove_and_add_spares for each of the two devices
>> >> >> > that are being rebuilt.  raid5_remove_disk declines to remove them because it
>> >> >> > can keep rebuilding them.
>> >> >> >
>> >> >> > remove_and_add_spares then counts them and notes there are 2.
>> >> >> > md_check_recovery notes that this is > 0, so it should create a thread to run
>> >> >> > md_do_sync.
>> >> >> >
>> >> >> > md_do_sync should then print out a message like
>> >> >> >  md: recovery of RAID array md0
>> >> >> >
>> >> >> > but it doesn't.  So something went wrong.
>> >> >> > There are three reasons that md_do_sync might not print a message:
>> >> >> >
>> >> >> > 1/ MD_RECOVERY_DONE is set.  As only md_do_sync ever sets it, that is
>> >> >> >    unlikely, and in any case md_check_recovery clears it.
>> >> >> > 2/ mddev->ro != 0.  It is only ever set to 0, 1, or 2.  If it is 1 or 2
>> >> >> >   then we would be able to see that in /proc/mdstat as a "(readonly)"
>> >> >> >   status.  But we don't.
>> >> >> > 3/ MD_RECOVERY_INTR is set. Again, md_check_recovery clears this.  It does
>> >> >> >   get set if kthread_should_stop() returns 'true', but that should only
>> >> >> >   happen if kthread_stop() was called.  That is only called by
>> >> >> >   md_unregister_thread and I cannot see any way that could be call.
>> >> >> >
>> >> >> > So.  No idea.
>> >> >> >
>> >> >> > Are you compiling these kernels yourself?
>> >> >>
>> >> >> Nope (used Mageia kernels), but I did now (3.1.5).
>> >> >>
>> >> >> > If so, could you:
>> >> >> >  - put a printk in the top of md_do_sync to report the values of
>> >> >> >   mddev->recovery and mddev->ro
>> >> >> >  - print a message whenever md_unregister_thread is called
>> >> >> >  - in md_check_recovery, in the
>> >> >> >                if (mddev->ro) {
>> >> >> >                        /* Only thing we do on a ro array is remove
>> >> >> >                         * failed devices.
>> >> >> >                         */
>> >> >> >                        mdk_rdev_t *rdev;
>> >> >> >
>> >> >> >  in statement, print the value of mddev->ro.
>> >> >> >
>> >> >> > Then see which of those printk's fire, and what they tell us.
>> >> >>
>> >> >> Only the last one does, and mddev->ro == 0.
>> >> >>
>> >> >> For reference, attached is the used patch and resulting log output.
>> >> >>
>> >> >
>> >> > Thanks.
>> >> >
>> >> > So it isn't running md_do_sync at all. Odd.
>> >> >
>> >> > Could please add:
>> >> >  - call "WARN_ON(1);" in print_raid5_conf() so we get a stack trace and can
>> >> >    see who is calling it.
>> >> >  - print the value that remove_and_add_spares is going to return.
>> >>
>> >> Attached. As you can see, remove_and_add_spare returns 0.
>> >>
>> >> --
>> >> Anssi Hannula
>> >
>> >
>> > Please add:
>> >
>> > diff --git a/drivers/md/md.c b/drivers/md/md.c
>> > index 5c95ccb..fa56ac5 100644
>> > --- a/drivers/md/md.c
>> > +++ b/drivers/md/md.c
>> > @@ -7328,8 +7328,10 @@ static int remove_and_add_spares(mddev_t *mddev)
>> >                        }
>> >                }
>> >
>> > +       printk("degraded=%d\n", mddev->degraded);
>> >        if (mddev->degraded) {
>> >                list_for_each_entry(rdev, &mddev->disks, same_set) {
>> > +                       printk("raid_disk=%d flags=%x\n", rdev->raid_disk, rdev->flags);
>> >                        if (rdev->raid_disk >= 0 &&
>> >                            !test_bit(In_sync, &rdev->flags) &&
>> >                            !test_bit(Faulty, &rdev->flags))
>> >
>> >
>> > 'degraded' must be 2 as dmesg contains
>> >
>> > [   45.544806] md/raid:md0: raid level 6 active with 8 out of 10 devices, algorithm 2
>> >
>> > and 'degraded' is exactly the difference between '8' and '10' there.
>> >
>> > raid disks 3 and 7 must have In_sync and Faulty clear as both of them just
>> > show "spare rebuilding" in the 'detail' output.
>> >
>> > so remove_and_add_spares "must" return 2.
>> >
>> > Hopefully the above patch will help me understand which of those is wrong.
>>
>> The output is:
>> [   47.389379] md0: degraded=2
>> [   47.389380] md0: raid_disk=0 flags=4
>> [   47.389381] md0: raid_disk=-1 flags=0
>>
>> Full assemble log attached.
>>
>
> Bingo.
>
> This will fix it.   We don't really need that 'break' there, and it is a
> problem.
>
> Thanks.

Confirmed. Thanks for the quick fix :)

-- 
Anssi Hannula
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux