On Sun, May 2, 2010 at 12:33 AM, christian (private) <chr.graf@xxxxxx> wrote: > Dear all, > > I am kindly asking for help please. > > The issue is the following: > I am creating an mdadm raid6-array and after successful sync all looks fine. > However after rebooting the raid6 refuses to start due to missing disks. > To my understanding this is all related to incorrect UUID's, however I am > kindly asking the experts to have alook into it: > > The error-message (similiar for all my 1TB-drives /dev/sd[abi]1) > ====================================== > keeper:~# mdadm --assemble /dev/md10 /dev/sd[abi]1 /dev/md[1-3] > mdadm: cannot open device /dev/sda1: Device or resource busy > mdadm: /dev/sda1 has no superblock - assembly aborted > > > systems basics > ========== > keeper:~# uname -a > Linux keeper 2.6.30-2-amd64 #1 SMP Fri Dec 4 03:31:05 UTC 2009 x86_64 > GNU/Linux > > distro: debian testing > > ii mdadm 3.0.3-2 > tool to administer Linux MD arrays (software > ii e2fslibs 1.41.11-1 > ext2/ext3/ext4 file system libraries > ii e2fsprogs 1.41.11-1 > ext2/ext3/ext4 file system utilities > > > > creating the array (hint, after reboot its /dev/sd[abi]1 instead of > /dev/sdb1 /dev/sdc1 /dev/sdi1) > =========================================================== > 387 mdadm --create /dev/md10 --name=10 --metadata=1.0 --level=6 > --raid-devices=6 /dev/sdb1 /dev/sdc1 /dev/sdi1 /dev/md[1-3] > > the array gets created - verifying after resync without rebooting > ========================================= > /dev/md10: > Version : 1.00 > Creation Time : Sat May 1 19:41:39 2010 > Raid Level : raid6 > Array Size : 3907039232 (3726.04 GiB 4000.81 GB) > Used Dev Size : 976759808 (931.51 GiB 1000.20 GB) > Raid Devices : 6 > Total Devices : 6 > Persistence : Superblock is persistent > > Update Time : Sun May 2 03:44:29 2010 > State : active > Active Devices : 6 > Working Devices : 6 > Failed Devices : 0 > Spare Devices : 0 > > Chunk Size : 64K > > Name : keeper:10 (local to host keeper) > *UUID : d9b9e76d:0113d173:39173474:7f43de08* > Events : 35 > > Number Major Minor RaidDevice State > 0 8 17 0 active sync /dev/sdb1 > 1 8 33 1 active sync /dev/sdc1 > 2 8 129 2 active sync /dev/sdi1 > 3 9 1 3 active sync /dev/md1 > 4 9 2 4 active sync /dev/md2 > 5 9 3 5 active sync /dev/md3 > > > keeper:~# cat /proc/mdstat > This is looking fine as well > ================== > Personalities : [raid6] [raid5] [raid4] [multipath] [faulty] [raid0] > md10 : active raid6 md3[5] md2[4] md1[3] sdi1[2] sdc1[1] sdb1[0] > 3907039232 blocks super 1.0 level 6, 64k chunk, algorithm 2 [6/6] > [UUUUUU] > > md1 : active raid0 sdf1[0] sdg1[1] > 976767872 blocks 64k chunks > > md2 : active raid0 sdh1[1] sda1[0] > 976767872 blocks 64k chunks > > md3 : active raid0 sdd1[1] sde1[0] > 976767872 blocks 64k chunks > > > To my understanding all member-drives of md10 should have the same UUID, but > after a reboot this is not any longer the case. Please notice that I created > the array with "/dev/sd[bci]1 /dev/md[1-3]" and after reboot its > /dev/sd[abi]1 and /dev/md[1-3]. > But as you can see, the UUID's are not identical anyhow: > ============================================================= > keeper:~# blkid > /dev/hda1: UUID="3A60F2845E1B3C0F" TYPE="ntfs" > /dev/hda2: UUID="e9a9b276-0530-42eb-8317-cd3999ff8b90" TYPE="swap" > /dev/hda3: UUID="738eeb2f-fae4-40a0-b0de-fc188837e44c" TYPE="ext3" > /dev/sdb1: UUID="d9b9e76d-0113-d173-3917-34747f43de08" LABEL="keeper:10" > TYPE="linux_raid_member" << looks good > /dev/sdc1: UUID="f45cfa44-1770-7e1c-3e3a-de9f30eaa984" > TYPE="linux_raid_member" LABEL="keeper:10" > /dev/sdd1: UUID="f45cfa44-1770-7e1c-3e3a-de9f30eaa984" > TYPE="linux_raid_member" LABEL="keeper:11" > /dev/sde1: UUID="0805c25a-cf77-4fb6-3e3a-de9f30eaa984" > TYPE="linux_raid_member" > /dev/sdf1: UUID="0805c25a-cf77-4fb6-3e3a-de9f30eaa984" > TYPE="linux_raid_member" > /dev/sdg1: UUID="13d13245-3c7d-8682-3e3a-de9f30eaa984" > TYPE="linux_raid_member" > /dev/sdh1: UUID="13d13245-3c7d-8682-3e3a-de9f30eaa984" > TYPE="linux_raid_member" > /dev/sdi1: UUID="d9b9e76d-0113-d173-3917-34747f43de08" > TYPE="linux_raid_member" LABEL="keeper:10" << looks good > /dev/sda1: UUID="c6f73dbb-528d-aa64-3e3a-de9f30eaa984" > TYPE="linux_raid_member" << should be *d9b9e76d:0113d173:39173474:7f43de08* > /dev/md1: UUID="058e47cd-e3ff-0fb9-3e3a-de9f30eaa984" > TYPE="linux_raid_member" << should be *d9b9e76d:0113d173:39173474:7f43de08* > /dev/md2: UUID="058e47cd-e3ff-0fb9-3e3a-de9f30eaa984" > TYPE="linux_raid_member" << should be *d9b9e76d:0113d173:39173474:7f43de08* > /dev/md3: UUID="058e47cd-e3ff-0fb9-3e3a-de9f30eaa984" > TYPE="linux_raid_member" << should be *d9b9e76d:0113d173:39173474:7f43de08* > > > Can please anyone help me digging out whats going on here? > > What I can say, that if I create 2 * raid5 (/dev/sd[abi]1 and /dev/md[1-3]) > instead of combining them into a single raid6, then all is perfect. > Maybe I should say, that /dev/md[1-3] are 2* striped 500G disks and > /dev/sd[abi]1 are 1TB drives. > > > any help is very much appreciated. The next step I am doing is manually > setting the UUID to have it matched... > > > further more I thoughts setting the UUID manually could solve the issue > > keeper:~# tune2fs -U d9b9e76d-0113-d173-3917-34747f43de08 /dev/sda1 > tune2fs 1.41.11 (14-Mar-2010) > tune2fs: Filesystem revision too high while trying to open /dev/sda1 > Couldn't find valid filesystem superblock. <<< does not look good to me > > > keeper:~# fdisk -l /dev/sda > > Disk /dev/sda: 1000.2 GB, 1000204886016 bytes > 255 heads, 63 sectors/track, 121601 cylinders > Units = cylinders of 16065 * 512 = 8225280 bytes > Disk identifier: 0x00000000 > > Device Boot Start End Blocks Id System > /dev/sda1 1 121601 976760001 83 Linux > > keeper:~# blkid /dev/sda > keeper:~# blkid /dev/sda1 > /dev/sda1: UUID="c6f73dbb-528d-aa64-3e3a-de9f30eaa984" > TYPE="linux_raid_member" > > keeper:~# tune2fs -l /dev/sda1 > tune2fs 1.41.11 (14-Mar-2010) > tune2fs: Filesystem revision too high while trying to open /dev/sda1 > Couldn't find valid filesystem superblock. > > > Any idea why the superblock is giving me trouble here? > > > many thanks > > christian > > > > > > -- > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html > I forgot to add, the blkid issue is probably related to caching. blkid -g might be sufficient blkid -p would bypass the cache entirely man blkid for more info. -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html