Potential data rollback/corruption after drive failure and re-appearance

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

I am testing the following scenario: a simple RAID1 md array with drives A and B. Assume that drive B fails, but the array remains operational and services IOs. After a while, machine is rebooted. After reboot drive B comes back, but now drive A becomes inaccessible. Assembling the array with both drives results in a degraded array, with a single drive B. However, B's data is the array's data at the time of drive B failure, not the latest array's data. So the data kind of rolls back in time.

Testing a similar scenario with RAID5: A,B and C drives, C drive fails, RAID5 becomes degraded but operational. After reboot B and C are accessible, but A disappears. Assembling the array fails, unless --force is given. With --force, the array comes up, but the data, of course, is corrupted.

Is this behavior intentional?

Suppose I want to protect against this by first examining the MD superblocks (--examine). I want to find the most updated drive, and check what array state it shows. Which part of "mdadm --examine" output should I use to find the most updated drive? The "Update Time" or the "Events" counter? Or perhaps something else?

Thanks,
Moshe Melnikov
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux