Re: Missing superblock on two disks in raid5

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Wol,

thank you.
I followed the overlay-procedure for all 7 disks and was able to start the array with:
mdadm --create --assume-clean --level=5 --raid-devices=7 /dev/md0 $OVERLAYS

# echo $OVERLAYS
/dev/mapper/sdd /dev/mapper/sdg /dev/mapper/sdh /dev/mapper/sdf1 /dev/mapper/sde /dev/mapper/sdi /dev/mapper/sdc

# cat /proc/mdstat
Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]
md0 : active raid5 dm-5[6] dm-10[5] dm-8[4] dm-9[3] dm-11[2] dm-7[1] dm-6[0]
      23441316864 blocks super 1.2 level 5, 512k chunk, algorithm 2 [7/7] [UUUUUUU]
      bitmap: 0/30 pages [0KB], 65536KB chunk

# mdadm --detail /dev/md0
/dev/md0:
           Version : 1.2
     Creation Time : Thu Sep 30 21:57:03 2021
        Raid Level : raid5
        Array Size : 23441316864 (22355.38 GiB 24003.91 GB)
     Used Dev Size : 3906886144 (3725.90 GiB 4000.65 GB)
      Raid Devices : 7
     Total Devices : 7
       Persistence : Superblock is persistent

     Intent Bitmap : Internal

       Update Time : Thu Sep 30 21:57:03 2021
             State : clean
    Active Devices : 7
   Working Devices : 7
    Failed Devices : 0
     Spare Devices : 0

            Layout : left-symmetric
        Chunk Size : 512K

Consistency Policy : bitmap

              Name : titan:0  (local to host titan)
              UUID : 45954f34:4559be76:61d18d85:f9e7ed80
            Events : 0

    Number   Major   Minor   RaidDevice State
       0     253        6        0      active sync   /dev/dm-6
       1     253        7        1      active sync   /dev/dm-7
       2     253       11        2      active sync   /dev/dm-11
       3     253        9        3      active sync   /dev/dm-9
       4     253        8        4      active sync   /dev/dm-8
       5     253       10        5      active sync   /dev/dm-10
       6     253        5        6      active sync   /dev/dm-5

I also examined all overlay-disks and they seem to have proper superblocks:
# mdadm -E /dev/mapper/sdc
/dev/mapper/sdc:
          Magic : a92b4efc
        Version : 1.2
    Feature Map : 0x1
     Array UUID : 45954f34:4559be76:61d18d85:f9e7ed80
           Name : titan:0  (local to host titan)
  Creation Time : Thu Sep 30 21:57:03 2021
     Raid Level : raid5
   Raid Devices : 7

 Avail Dev Size : 7813772976 (3725.90 GiB 4000.65 GB)
     Array Size : 23441316864 (22355.38 GiB 24003.91 GB)
  Used Dev Size : 7813772288 (3725.90 GiB 4000.65 GB)
    Data Offset : 264192 sectors
   Super Offset : 8 sectors
   Unused Space : before=264112 sectors, after=688 sectors
          State : clean
    Device UUID : 10d49408:d28fe8ee:58ce698e:a2d66ac0

Internal Bitmap : 8 sectors from superblock
    Update Time : Thu Sep 30 21:57:03 2021
  Bad Block Log : 512 entries available at offset 24 sectors
       Checksum : 68e54840 - correct
         Events : 0

         Layout : left-symmetric
     Chunk Size : 512K

   Device Role : Active device 6
   Array State : AAAAAAA ('A' == active, '.' == missing, 'R' == replacing)

But it looks like there is no data on it?
Initially I used LVM to create /dev/md0 as a physical volume, created a volume-group with it and a logical volume on it with all free space. Either the physical volume doesn't exist in this state, or I don't know to use it..
I tried pvs, pvscan, vgscan, dmsetup info, but it won't show up.
Do you think this is just a lvm problem now, or did I miss something?

Best regards,
Clemens

Am 30.09.2021 um 20:11 schrieb Wols Lists:
On 30/09/2021 18:21, Clemens Kiehas wrote:
Hello,

until recently I was using Unraid with 8 disks in an XFS array.
I installed Debian on a spare SSD on the same machine and started migrating disk by disk from the Unraid array to a raid5 array using a second server as temporay storage. So I switched between Debian and Unraid a lot to copy data and remove/add drives from/to arrays.

 From the beginning I always had to assemble the array without /dev/sdd, add it afterwards and let it rebuild - since the array was working fine afterwards I didn't really think much of it.

Not wise - as you've found out it was trashing your redundancy ...

Appearently Unraid always overwrote the superblock of that 1 (and later 2) disks (/dev/sdc and /dev/sdd) when I switched between the two OSs and now mdadm isn't recognizing those 2 disks and I can't assemble the array obviously. At least that's what I think happened, since file tells me that the first 32k bytes are XFS:
# losetup -o 32768 -f /dev/sdd
# file -s /dev/loop0
/dev/loop0: SGI XFS filesystem data (blksz 4096, inosz 512, v2 dirs)

That feels good, but I'm not sure how to proceed from here.

Right now mdadm only assembles 5 (instead of 7) disks as spares into an inactive array at boot:
# cat /proc/mdstat
Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]
md0 : inactive sde[6](S) sdf1[5](S) sdg[1](S) sdh[3](S) sdi[8](S)
       21487980896 blocks super 1.2
unused devices: <none>

My system:
Linux titan 5.10.0-8-amd64 #1 SMP Debian 5.10.46-4 (2021-08-03) x86_64 GNU/Linux
mdadm - v4.1 - 2018-10-01

Maybe I could try to assemble it with assume-clean and read-only?
I found some pages in the wiki but I'm not 100% sure that they will solve my problem and I don't want to make things worse.
https://raid.wiki.kernel.org/index.php/Recovering_a_damaged_RAID
https://raid.wiki.kernel.org/index.php/RAID_Recovery

The first page talks about overlays. THEY WILL PROTECT YOUR DATA.

Make sure every disk is overlayed, then try assembling the overlays with your assume-clean. You won't need read-only but do it if you like it then do.

If you can then mount and run a fsck or whatever it is over the file system and it says everything is clean, then you can redo it without the overlays and recover the array properly. If it's messed up, just tear down the overlays and you're back where you started.

If it doesn't work, come back and tell us what happened (if it does work, please let us know :-)

Cheers,
Wol

--
MatrNr: 11805432
Mobile: 06642573109
E-Mail: clemens@xxxxxxxxx
Matrix: @clemens:matrix.spacebeard.tech




[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux