Re: mdXX: bitmap superblock UUID mismatch

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 14 February 2011 16:33, Reynald Borer <reynald.borer@xxxxxxxxx> wrote:
> Hi,
>
> Nice catch for the 1 bit difference, I didn't see it. My point
> regarding bitmap reconstruction was because this raid was used in a
> LVM setup. the LVM partition used two distinct raid 1 disks, and I was
> not able to start the LVM correctly without this failing raid.
>
> In the end, I was able to save my LVM by simply skipping the raid 1
> and using directly one partition. The LVM tool was clever enough to
> detect the MD bits and proposed me to remove them in order to mount
> directly the partition, and it worked fine. Thus I was able to save my
> data.
>
> Thanks for your answer though.
>
> Regards,
> Reynald
>
>
> On Thu, Jan 27, 2011 at 9:53 PM, NeilBrown <neilb@xxxxxxx> wrote:
>> On Wed, 26 Jan 2011 21:58:25 +0100 Reynald Borer <reynald.borer@xxxxxxxxx>
>> wrote:
>>
>>> Hello guys,
>>>
>>> I have been using md raids for quite a long time now and it always
>>> worked fine, until recently when I upgraded some hardware on my
>>> workstation. Unfortunately the hardware I changed proved itself to be
>>> very unstable, and I encountered a lot of hard lockups of the system
>>> while running. Those lockups recently made one of my raid 1 array
>>> fails with the infamous error message "mdXX: bitmap superblock UUID
>>> mismatch".
>>>
>>> Here is what I have found in the kernel logs when I try to activate
>>> the given raid group:
>>> -----------------
>>> md/raid1:md126: active with 2 out of 2 mirrors
>>> md126: bitmap superblock UUID mismatch
>>> md126: bitmap file superblock:
>>> Â Â Â Â Âmagic: 6d746962
>>> Â Â Â Âversion: 4
>>> Â Â Â Â Â uuid: 37102258.af9c1930.b8397fb8.eba356af
>> Â Â Â Â Â Â Â Â Â Â Â Â Â ^ this is an 'a'
>>
>>> Â Â Â Â events: 199168
>>> events cleared: 199166
>>> Â Â Â Â Âstate: 00000000
>>> Â Â Âchunksize: 524288 B
>>> Â daemon sleep: 5s
>>> Â Â Âsync size: 248075584 KB
>>> max write behind: 0
>>> md126: failed to create bitmap (-22)
>>> -----------------
>>>
>>>
>>> Such error messages are displayed each time I try to run the raid
>>> group. Content of /proc/mdstat is:
>>> -----------------
>>> md126 : inactive sdb6[0] sda6[1]
>>> Â Â Â 496151168 blocks
>>> -----------------
>>>
>>>
>>> If I try to examine both disks with mdadm -E it shows some checksum
>>> mismatch for both partitions:
>>> -----------------
>>> root@bob # mdadm -E /dev/sda6
>>> /dev/sda6:
>>> Â Â Â Â Â Magic : a92b4efc
>>> Â Â Â Â Version : 0.90.03
>>> Â Â Â Â Â ÂUUID : 37102258:bf9c1930:b8397fb8:eba356af
>> Â Â Â Â Â Â Â Â Â Â Â Â Â Â ^ this is a 'b'
>>
>> So you certainly do have some sick hardware!!!
>>
>> I suggest that you find some hardware that you can trust,
>> mount one of the two devices ( (sdb6 or sda6) ignoring the raid stuff,
>> and copy data off to the device that you trust.
>>
>> Then start again.
>>
>> NeilBrown
>>
>>
>>> Â Creation Time : Mon Aug Â7 21:06:47 2006
>>> Â Â ÂRaid Level : raid1
>>> Â Used Dev Size : 248075584 (236.58 GiB 254.03 GB)
>>> Â Â ÂArray Size : 248075584 (236.58 GiB 254.03 GB)
>>> Â ÂRaid Devices : 2
>>> Â Total Devices : 2
>>> Preferred Minor : 6
>>>
>>> Â Â Update Time : Wed Jan 12 00:12:44 2011
>>> Â Â Â Â Â State : clean
>>> ÂActive Devices : 2
>>> Working Devices : 2
>>> ÂFailed Devices : 0
>>> Â Spare Devices : 0
>>> Â Â Â ÂChecksum : e4883f8e - expected e4883e8e
>>> Â Â Â Â ÂEvents : 199168
>>>
>>>
>>>    Number  Major  Minor  RaidDevice State
>>> this   1    8    38    Â1   Âactive sync
>>>
>>> Â Â0 Â Â 0 Â Â Â 8 Â Â Â 70 Â Â Â Â0 Â Â Âactive sync
>>> Â Â1 Â Â 1 Â Â Â 8 Â Â Â 38 Â Â Â Â1 Â Â Âactive sync
>>> root@bob # mdadm -E /dev/sdb6
>>> /dev/sdb6:
>>> Â Â Â Â Â Magic : a92b4efc
>>> Â Â Â Â Version : 0.90.03
>>> Â Â Â Â Â ÂUUID : 37102258:bf9c1930:b8397fb8:eba356af
>>> Â Creation Time : Mon Aug Â7 21:06:47 2006
>>> Â Â ÂRaid Level : raid1
>>> Â Used Dev Size : 248075584 (236.58 GiB 254.03 GB)
>>> Â Â ÂArray Size : 248075584 (236.58 GiB 254.03 GB)
>>> Â ÂRaid Devices : 2
>>> Â Total Devices : 2
>>> Preferred Minor : 6
>>>
>>> Â Â Update Time : Wed Jan 12 00:12:44 2011
>>> Â Â Â Â Â State : clean
>>> ÂActive Devices : 2
>>> Working Devices : 2
>>> ÂFailed Devices : 0
>>> Â Spare Devices : 0
>>> Â Â Â ÂChecksum : e4883fac - expected e4883eac
>>> Â Â Â Â ÂEvents : 199168
>>>
>>>
>>>    Number  Major  Minor  RaidDevice State
>>> this   0    8    70    Â0   Âactive sync
>>>
>>> Â Â0 Â Â 0 Â Â Â 8 Â Â Â 70 Â Â Â Â0 Â Â Âactive sync
>>> Â Â1 Â Â 1 Â Â Â 8 Â Â Â 38 Â Â Â Â1 Â Â Âactive sync
>>> -----------------
>>>
>>>
>>> Any idea how I could try to save my raid group?
>>>
>>> Thanks in advance for your help.
>>>
>>> Best Regards,
>>> Reynald
>>> --
>>> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
>>> the body of a message to majordomo@xxxxxxxxxxxxxxx
>>> More majordomo info at Âhttp://vger.kernel.org/majordomo-info.html
>>
>>
> --
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at Âhttp://vger.kernel.org/majordomo-info.html
>

Wow, that's cool. I suppose RAID1 is simple enough for this to be
possible, though. Still, cool. :-)

// M
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux