Re: MDADM RAID 6 Bad Superblock after reboot

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




On 10/18/2017 12:40 PM, Wols Lists wrote:
> What does --detail tell us about the array?

https://imgur.com/a/Own0W

Apologies for the imgur link, but this was the easiest way to
communicate --detail.

> Are you sure the three drives were added? SELinux has a habit of causing
> havoc. Did the available space on the array increase? Did you check?

Yeah. The array took about 17 hours to rebuild with the three drives
added (I had expected over 24 as that had been my experience with adding
the 5th drive long ago), and I had immediately started using the extra
space. The extra 8+ TB showed up in --details, as well as df, and my
guests could see the extra space.

The SELinux audit log, however, was very clear about mdadm not being
able to edit the conf. And it's true - the conf file did not have the
extra drives added. I've since audited and applied a rule to allow the
editing of the conf file, but the system is currently in permissive mode
until the array is back online. I can disable if needed.

> Firstly, make sure SELinux didn't interfere with the grow. My guess is
> the add failed because SELinux blocked it, and in reality you've still
> got a five-drive array, it just thinks it's an eight-drive array, so
> when the system rebooted it said "five drives of eight? Not enough!" and
> stopped.

I could see this being the case - on reboot, the configuration would
specify 5 drives instead of the 8. In addition, the system did not get
rebooted after the array had been grown - I just kept it running and put
it to work. Lesson learned.

> If I'm right, recovery is hopefully a simple matter of disabling
> SELinux, and re-assembling the array with either reverting the grow, or
> firing it off so it can actually run and complete.


Thanks for the vote of confidence in not losing data here. As I
mentioned above, I've set SELinux into permissive mode. I'm sort of at a
loss as to what to do next. Since sd[fgh] don't have any superblock
data, can I try to bring the array back online with 5 drives, then
re-add the last three, with the hope that they sync? There obviously
hasn't been any writes to the last three drives since the hard system
reboot, so I'd hope their event numbers would be in sync?

Thanks,

-Sean
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux