Re: upgrade to jessie/newer kernel and mdadm problems

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I was leaving your case for people who know IMSM to pipe up, as I don't
have any experience with it.  But the silence is deafening :-(

That's OK, a good guy (PascalHambourg) in the french debian forum was able to help me a lot.


However, if you've been using the system in this degraded state, you
will need to do the manual assembly with only the good partitions, then
add the other partitions to rebuild each.

Yes, it was in use, but this what we done :

1. restoring the original mdadm.conf

2. modifying the DEVICE lines with this : DEVICE /dev/sdb?* /dev/sdc?*

3. updating initram: update-initramfs -u (got some errors but we ignored them, however I made two initrd to be failsafe)

4. rebooted, md126 was gone and md9 back. However all arrays had a partition marked as fail

5. rebuilded each partiton with mdadm /dev/mdX --add /dev/sdcX or mdadm /dev/mdX --add /dev/sdbX accordingly (sometimes, the failed partition was on sdb and sometimes on sdc, this scarried me as I thought I would loose everything if one drive failed).

6. rebuild was ok, I needed to remove the superblock on /dev/sdb and /dev/sdc because it looked like this (it contains Intel RAID data but my disk are softraid and we thought that was one origin of the problem):


<--------------------------------------------------------------------->
# mdadm -E /dev/sdb
mdmon: /dev/sdb is not attached to Intel(R) RAID controller.
mdmon: /dev/sdb is not attached to Intel(R) RAID controller.
/dev/sdb:
          Magic : Intel Raid ISM Cfg Sig.
        Version : 1.1.00
    Orig Family : 26b5a9e0
         Family : 26b5a9e0
     Generation : 00004db7
     Attributes : All supported
           UUID : d9cfa6d9:2a715e4f:1fbc2095:be342429
       Checksum : 261d2aed correct
    MPB Sectors : 1
          Disks : 2
   RAID Devices : 1

  Disk01 Serial : VFC100R10BE79D
          State : active
             Id : 00010000
    Usable Size : 488390862 (232.88 GiB 250.06 GB)

[raidlin]:
           UUID : 91449a9d:9242bfe9:d99bceb0:a59f9314
     RAID Level : 1
        Members : 2
          Slots : [UU]
    Failed disk : none
      This Slot : 1
     Array Size : 488390656 (232.88 GiB 250.06 GB)
   Per Dev Size : 488390656 (232.88 GiB 250.06 GB)
  Sector Offset : 0
    Num Stripes : 1907776
     Chunk Size : 64 KiB
       Reserved : 0
  Migrate State : idle
      Map State : normal
    Dirty State : dirty

  Disk00 Serial : VFC100R10BRKMD
          State : active
             Id : 00000000
    Usable Size : 488390862 (232.88 GiB 250.06 GB)
<--------------------------------------------------------------------->
# mdadm -E /dev/sdc
mdmon: /dev/sdc is not attached to Intel(R) RAID controller.
mdmon: /dev/sdc is not attached to Intel(R) RAID controller.
/dev/sdc:
          Magic : Intel Raid ISM Cfg Sig.
        Version : 1.1.00
    Orig Family : 26b5a9e0
         Family : 26b5a9e0
     Generation : 00004dbc
     Attributes : All supported
           UUID : d9cfa6d9:2a715e4f:1fbc2095:be342429
       Checksum : 261c2af2 correct
    MPB Sectors : 1
          Disks : 2
   RAID Devices : 1

  Disk00 Serial : VFC100R10BRKMD
          State : active
             Id : 00000000
    Usable Size : 488390862 (232.88 GiB 250.06 GB)

[raidlin]:
           UUID : 91449a9d:9242bfe9:d99bceb0:a59f9314
     RAID Level : 1
        Members : 2
          Slots : [UU]
    Failed disk : none
      This Slot : 0
     Array Size : 488390656 (232.88 GiB 250.06 GB)
   Per Dev Size : 488390656 (232.88 GiB 250.06 GB)
  Sector Offset : 0
    Num Stripes : 1907776
     Chunk Size : 64 KiB
       Reserved : 0
  Migrate State : idle
      Map State : normal
    Dirty State : clean

  Disk01 Serial : VFC100R10BE79D
          State : active
             Id : 00010000
    Usable Size : 488390862 (232.88 GiB 250.06 GB)
<--------------------------------------------------------------------->

7. So I rebooted into initramfs shell by editing GRUB command line to add "break" at the end of kernel line

8. stopped the raid (they was mounted accordingly to /proc/mdstat) : mdadm --stop --scan

9. removed the superblock on /dev/sdb and /dev/sdc : mdadm --zero-superblock --metadata=imsm /dev/sdb ; mdadm --zero-superblock --metadata=imsm /dev/sdc

This is what they look now :
<--------------------------------------------------------------------->
# mdadm -E /dev/sdb
/dev/sdb:
   MBR Magic : aa55
Partition[0] :       979902 sectors at           63 (type fd)
Partition[1] :      9767520 sectors at       979965 (type fd)
Partition[2] :      3903795 sectors at     10747485 (type fd)
Partition[3] :    473740785 sectors at     14651280 (type 05)
<--------------------------------------------------------------------->
# mdadm -E /dev/sdc
/dev/sdc:
   MBR Magic : aa55
Partition[0] :       979902 sectors at           63 (type fd)
Partition[1] :      9767520 sectors at       979965 (type fd)
Partition[2] :      3903795 sectors at     10747485 (type fd)
Partition[3] :    473740785 sectors at     14651280 (type 05)
<--------------------------------------------------------------------->

10. rebooted, changed back DEVICE line in mdadm.conf to DEVICE partitions

The array looks like OK now.

Anyway, Should I upgrade to superblock 1.0 (or 1.2) ? If so, can I use your method do it in initramfs shell (because my system is live with active raid arrays) ?


Full thread there (in french, sorry): https://www.debian-fr.org/mise-a-jour-vers-jessie-et-mdadm-t51945.html
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux