upgrade to jessie/newer kernel and mdadm problems

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello list,

Sorry for the long post, but I wanted to be as much informative as I can be.
Forgive my lack of knowlegde in mdadm, I know how to create it using the debian installer and few things to get information but that's all.
Forgive also my english, I'm not a native in this language.

My system has been installed in 2007, and I've upgraded it several times until this week to Debian Jessie (the latest version).

So, I've upgraded my system to jessie today (only partially with apt-get upgrade + kernel upgrade) and I faced with a problem with my RAID 1 soft.

I have two disk (/dev/sdb and /dev/sdc) which are members of the raid array
There's 10 partitions on these disks, each one is a array (ie sdb1 and sdc1).
All of these form my raids array : md0 to md9

Today , one of my md partition was missing.
Before upgrade I had partitions from md0 to md9, after reboot I'm missing md9.

my mdadm.conf before reboot looked like that:
>-------------------------------------------------------------------------------------------------------------<

   # mdadm.conf
   #
   # Please refer to mdadm.conf(5) for information about this file.
   #

   # by default, scan all partitions (/proc/partitions) for MD superblocks.
   # alternatively, specify devices to scan, using wildcards if desired.
   DEVICE partitions

   # auto-create devices with Debian standard permissions
   CREATE owner=root group=disk mode=0660 auto=yes

   # automatically tag new arrays as belonging to the local system
   HOMEHOST <system>

   # instruct the monitoring daemon where to send mail alerts
   MAILADDR root

   # definitions of existing MD arrays
   ARRAY /dev/md0 level=raid1 num-devices=2
   UUID=350e253f:863b7b04:b1617c47:b213a024
   ARRAY /dev/md1 level=raid1 num-devices=2
   UUID=086e68ed:3607317f:60b56e23:6bae62bc
   ARRAY /dev/md2 level=raid1 num-devices=2
   UUID=0f6e3ed5:aeee975a:c3647deb:763d68ce
   ARRAY /dev/md3 level=raid1 num-devices=2
   UUID=6b560fe8:f24d6f2e:8942bd3a:1903abbc
   ARRAY /dev/md4 level=raid1 num-devices=2
   UUID=462dab96:b7ca2a17:7c4aebf1:d4d7ec3b
   ARRAY /dev/md5 level=raid1 num-devices=2
   UUID=ea511351:3abc7b12:4c81e838:93dbd21a
   ARRAY /dev/md6 level=raid1 num-devices=2
   UUID=da0d76c6:91422584:dc3d6162:37ced53b
   ARRAY /dev/md7 level=raid1 num-devices=2
   UUID=387c831c:8a6d05e3:b649696c:0870b930
   ARRAY /dev/md8 level=raid1 num-devices=2
   UUID=b07c4ab4:39d0ba53:9913afa9:fd9cc323
   ARRAY /dev/md9 level=raid1 num-devices=2
   UUID=36c4edd0:a0492cc9:0cd2fce0:2745e358

   # This file was auto-generated on Wed, 07 Feb 2007 17:15:33 +0000
   # by mkconf $Id: mkconf 261 2006-11-09 13:32:35Z madduck $

>-------------------------------------------------------------------------------------------------------------<

After reboot and after a mdadm -Es, I got this:
>-------------------------------------------------------------------------------------------------------------<

   # mdadm.conf
   #
   # Please refer to mdadm.conf(5) for information about this file.
   #

   # by default, scan all partitions (/proc/partitions) for MD superblocks.
   # alternatively, specify devices to scan, using wildcards if desired.
   DEVICE partitions

   # auto-create devices with Debian standard permissions
   CREATE owner=root group=disk mode=0660 auto=yes

   # automatically tag new arrays as belonging to the local system
   HOMEHOST <system>

   # instruct the monitoring daemon where to send mail alerts
   MAILADDR root

   # definitions of existing MD arrays
   ARRAY metadata=imsm UUID=d9cfa6d9:2a715e4f:1fbc2095:be342429
   ARRAY /dev/md/raidlin container=d9cfa6d9:2a715e4f:1fbc2095:be342429
   member=0 UUID=91449a9d:9242bfe9:d99bceb0:a59f9314
   ARRAY /dev/md0 UUID=350e253f:863b7b04:b1617c47:b213a024
   ARRAY /dev/md1 UUID=086e68ed:3607317f:60b56e23:6bae62bc
   ARRAY /dev/md2 UUID=0f6e3ed5:aeee975a:c3647deb:763d68ce
   ARRAY /dev/md3 UUID=6b560fe8:f24d6f2e:8942bd3a:1903abbc
   ARRAY /dev/md4 UUID=462dab96:b7ca2a17:7c4aebf1:d4d7ec3b
   ARRAY /dev/md5 UUID=ea511351:3abc7b12:4c81e838:93dbd21a
   ARRAY /dev/md6 UUID=da0d76c6:91422584:dc3d6162:37ced53b
   ARRAY /dev/md7 UUID=387c831c:8a6d05e3:b649696c:0870b930
   ARRAY /dev/md8 UUID=b07c4ab4:39d0ba53:9913afa9:fd9cc323

>-------------------------------------------------------------------------------------------------------------<

This is what is looking my fdisk extract and /proc/mdstat:
>-------------------------------------------------------------------------------------------------------------<

   # fdisk -l /dev/sdb

   Disk /dev/sdb: 250.1 GB, 250059350016 bytes
   255 heads, 63 sectors/track, 30401 cylinders, total 488397168 sectors
   Units = sectors of 1 * 512 = 512 bytes
   Sector size (logical/physical): 512 bytes / 512 bytes
   I/O size (minimum/optimal): 512 bytes / 512 bytes
   Disk identifier: 0x0001edee

       Device Boot      Start         End      Blocks   Id  System
   /dev/sdb1   *          63      979964      489951   fd  Linux raid
   autodetect
   /dev/sdb2          979965    10747484     4883760   fd  Linux raid
   autodetect
   /dev/sdb3        10747485    14651279     1951897+  fd  Linux raid
   autodetect
   /dev/sdb4        14651280   488392064   236870392+   5  Extended
   /dev/sdb5        14651343    24418799     4883728+  fd  Linux raid
   autodetect
   /dev/sdb6        24418863    43953839     9767488+  fd  Linux raid
   autodetect
   /dev/sdb7        43953903    53721359     4883728+  fd  Linux raid
   autodetect
   /dev/sdb8        53721423    63488879     4883728+  fd  Linux raid
   autodetect
   /dev/sdb9        63488943    73256399     4883728+  fd  Linux raid
   autodetect
   /dev/sdb10       73256463    83023919     4883728+  fd  Linux raid
   autodetect
   /dev/sdb11       83023983   488392064   202684041   fd  Linux raid
   autodetect

>-------------------------------------------------------------------------------------------------------------<

   # fdisk -l /dev/sdc

   Disk /dev/sdc: 250.1 GB, 250059350016 bytes
   255 heads, 63 sectors/track, 30401 cylinders, total 488397168 sectors
   Units = sectors of 1 * 512 = 512 bytes
   Sector size (logical/physical): 512 bytes / 512 bytes
   I/O size (minimum/optimal): 512 bytes / 512 bytes
   Disk identifier: 0x000c352c

       Device Boot      Start         End      Blocks   Id  System
   /dev/sdc1   *          63      979964      489951   fd  Linux raid
   autodetect
   /dev/sdc2          979965    10747484     4883760   fd  Linux raid
   autodetect
   /dev/sdc3        10747485    14651279     1951897+  fd  Linux raid
   autodetect
   /dev/sdc4        14651280   488392064   236870392+   5  Extended
   /dev/sdc5        14651343    24418799     4883728+  fd  Linux raid
   autodetect
   /dev/sdc6        24418863    43953839     9767488+  fd  Linux raid
   autodetect
   /dev/sdc7        43953903    53721359     4883728+  fd  Linux raid
   autodetect
   /dev/sdc8        53721423    63488879     4883728+  fd  Linux raid
   autodetect
   /dev/sdc9        63488943    73256399     4883728+  fd  Linux raid
   autodetect
   /dev/sdc10       73256463    83023919     4883728+  fd  Linux raid
   autodetect
   /dev/sdc11       83023983   488392064   202684041   fd  Linux raid
   autodetect

>-------------------------------------------------------------------------------------------------------------<

   # cat /proc/mdstat
   Personalities : [raid1]
   md8 : active raid1 md126p10[0]
          4883648 blocks [2/1] [U_]

   md7 : active raid1 md126p9[0]
          4883648 blocks [2/1] [U_]

   md6 : active raid1 md126p8[0]
          4883648 blocks [2/1] [U_]

   md5 : active raid1 md126p7[0]
          4883648 blocks [2/1] [U_]

   md4 : active raid1 md126p6[0]
          9767424 blocks [2/1] [U_]

   md3 : active raid1 md126p5[0]
          4883648 blocks [2/1] [U_]

   md2 : active (auto-read-only) raid1 md126p3[0]
          1951808 blocks [2/1] [U_]

   md1 : active raid1 md126p2[0]
          4883648 blocks [2/1] [U_]

   md0 : active raid1 md126p1[0]
          489856 blocks [2/1] [U_]

   md126 : active raid1 sdc[1] sdb[0]
          244195328 blocks super external:/md127/0 [2/2] [UU]

   md127 : inactive sdc[1](S) sdb[0](S)
          6306 blocks super external:imsm

   unused devices: <none>

>-------------------------------------------------------------------------------------------------------------<


I'm not an expert at md at all, so I called a friend and we were able to find my lost md9 partition, it was md126p11.

This is what we saw in dmesg:
>-------------------------------------------------------------------------------------------------------------<

   [    1.593297] md: bind<sdb>
   [    1.598582] md: bind<sdc>
   [    1.599902] md: bind<sdb>
   [    1.600045] md: bind<sdc>
   [    1.606550] md: raid1 personality registered for level 1
   [    1.607396] md/raid1:md126: active with 2 out of 2 mirrors
   [    1.607433] md126: detected capacity change from 0 to 250056015872
   [    1.632396] md: md126 switched to read-write mode.
   [    1.669910]  md126: p1 p2 p3 p4 < p5 p6 p7 p8 p9 p10 p11 >
   [    1.670501] md126: p11 size 405368082 extends beyond EOD, truncated
   [    4.100396] md: md0 stopped.
   [    4.100919] md: bind<md126p1>
   [    4.101708] md/raid1:md0: active with 1 out of 2 mirrors
   [    4.101734] md0: detected capacity change from 0 to 501612544
   [    4.102179]  md0: unknown partition table
   [    4.224625] md: md1 stopped.
   [    4.225109] md: bind<md126p2>
   [    4.225886] md/raid1:md1: active with 1 out of 2 mirrors
   [    4.225911] md1: detected capacity change from 0 to 5000855552
   [    4.226658]  md1: unknown partition table
   [    4.420746] md: md2 stopped.
   [    4.421441] md: bind<md126p3>
   [    4.422216] md/raid1:md2: active with 1 out of 2 mirrors
   [    4.422241] md2: detected capacity change from 0 to 1998651392
   [    4.422677]  md2: unknown partition table
   [    4.595729] md: md3 stopped.
   [    4.596410] md: bind<md126p5>
   [    4.597189] md/raid1:md3: active with 1 out of 2 mirrors
   [    4.597215] md3: detected capacity change from 0 to 5000855552
   [    4.597638]  md3: unknown partition table
   [    4.668224] md: md4 stopped.
   [    4.668693] md: bind<md126p6>
   [    4.669446] md/raid1:md4: active with 1 out of 2 mirrors
   [    4.669474] md4: detected capacity change from 0 to 10001842176
   [    4.669909]  md4: unknown partition table
   [    4.783732] md: md5 stopped.
   [    4.784236] md: bind<md126p7>
   [    4.785024] md/raid1:md5: active with 1 out of 2 mirrors
   [    4.785049] md5: detected capacity change from 0 to 5000855552
   [    4.785479]  md5: unknown partition table
   [    4.970769] md: md6 stopped.
   [    4.971366] md: bind<md126p8>
   [    4.972129] md/raid1:md6: active with 1 out of 2 mirrors
   [    4.972158] md6: detected capacity change from 0 to 5000855552
   [    4.972594]  md6: unknown partition table
   [    5.137394] md: md7 stopped.
   [    5.138011] md: bind<md126p9>
   [    5.138754] md/raid1:md7: active with 1 out of 2 mirrors
   [    5.138779] md7: detected capacity change from 0 to 5000855552
   [    5.139232]  md7: unknown partition table
   [    5.329093] md: md8 stopped.
   [    5.330228] md: bind<md126p10>
   [    5.330977] md/raid1:md8: active with 1 out of 2 mirrors
   [    5.331003] md8: detected capacity change from 0 to 5000855552
   [    5.350896]  md8: unknown partition table

>-------------------------------------------------------------------------------------------------------------<

note the line "[ 1.670501] md126: p11 size 405368082 extends beyond EOD, truncated"

After force mounting /dev/md126p11, I noted these messages in dmesg:
>-------------------------------------------------------------------------------------------------------------<

   [ 4361.248369] md126p11: rw=32, want=405367936, limit=405366673
   [ 4361.248382] XFS (md126p11): Mounting V4 Filesystem
   [ 4361.657842] XFS (md126p11): Ending clean mount
   [ 4609.295281] md126p11: rw=32, want=405367936, limit=405366673
   [ 4609.295300] XFS (md126p11): Mounting V4 Filesystem
   [ 4609.513109] XFS (md126p11): Ending clean mount

>-------------------------------------------------------------------------------------------------------------<


and in daemon.log (lines correspond to the reboot after the upgrade, the time I lost my md9):
>-------------------------------------------------------------------------------------------------------------<

   Apr 29 20:26:26 shax mdadm[4735]: DeviceDisappeared event detected
   on md device /dev/md9
   Apr 29 20:26:29 shax mdadm[4735]: DegradedArray event detected on md
   device /dev/md8
   Apr 29 20:26:30 shax mdadm[4735]: DegradedArray event detected on md
   device /dev/md7
   Apr 29 20:26:31 shax mdadm[4735]: DegradedArray event detected on md
   device /dev/md6
   Apr 29 20:26:33 shax mdadm[4735]: DegradedArray event detected on md
   device /dev/md5
   Apr 29 20:26:34 shax mdadm[4735]: DegradedArray event detected on md
   device /dev/md4
   Apr 29 20:26:35 shax mdadm[4735]: DegradedArray event detected on md
   device /dev/md3
   Apr 29 20:26:35 shax mdadm[4735]: DegradedArray event detected on md
   device /dev/md2
   Apr 29 20:26:36 shax mdadm[4735]: DegradedArray event detected on md
   device /dev/md1
   Apr 29 20:26:37 shax mdadm[4735]: DegradedArray event detected on md
   device /dev/md0
   Apr 29 20:26:37 shax mdadm[4735]: NewArray event detected on md
   device /dev/md127
   Apr 29 20:26:37 shax mdadm[4735]: NewArray event detected on md
   device /dev/md126

>-------------------------------------------------------------------------------------------------------------<

output of fdisk -l /dev/md126 :

>-------------------------------------------------------------------------------------------------------------<

   fdisk -l /dev/md126

   Disk /dev/md126: 250.1 GB, 250056015872 bytes
   255 heads, 63 sectors/track, 30400 cylinders, total 488390656 sectors
   Units = sectors of 1 * 512 = 512 bytes
   Sector size (logical/physical): 512 bytes / 512 bytes
   I/O size (minimum/optimal): 512 bytes / 512 bytes
   Disk identifier: 0x000c352c

          Device Boot      Start         End      Blocks   Id System
   /dev/md126p1   *          63      979964      489951   fd  Linux
   raid autodetect
   /dev/md126p2          979965    10747484     4883760   fd  Linux
   raid autodetect
   /dev/md126p3        10747485    14651279     1951897+  fd  Linux
   raid autodetect
   /dev/md126p4        14651280   488392064   236870392+   5 Extended
   /dev/md126p5        14651343    24418799     4883728+  fd  Linux
   raid autodetect
   /dev/md126p6        24418863    43953839     9767488+  fd  Linux
   raid autodetect
   /dev/md126p7        43953903    53721359     4883728+  fd  Linux
   raid autodetect
   /dev/md126p8        53721423    63488879     4883728+  fd  Linux
   raid autodetect
   /dev/md126p9        63488943    73256399     4883728+  fd  Linux
   raid autodetect
   /dev/md126p10       73256463    83023919     4883728+  fd  Linux
   raid autodetect
   /dev/md126p11       83023983   488392064   202684041   fd  Linux
   raid autodetect

>-------------------------------------------------------------------------------------------------------------<

mdadm -E /dev/sdbX ou mdadm -E /dev/sdcX doesn't give anything :
>-------------------------------------------------------------------------------------------------------------<

   # mdadm -E /dev/sdb1
   mdadm: cannot open /dev/sdb1: No such device or address

   # mdadm -E /dev/sdb2
   mdadm: cannot open /dev/sdb2: No such device or address

   # mdadm -E /dev/sdb3
   mdadm: cannot open /dev/sdb3: No such device or address

   # mdadm -E /dev/sdb4
   mdadm: cannot open /dev/sdb4: No such device or address

   # mdadm -E /dev/sdb5
   mdadm: cannot open /dev/sdb5: No such device or address

   # mdadm -E /dev/sdb6
   mdadm: cannot open /dev/sdb6: No such device or address

   # mdadm -E /dev/sdb7
   mdadm: cannot open /dev/sdb7: No such device or address

   # mdadm -E /dev/sdb8
   mdadm: cannot open /dev/sdb8: No such device or address

   # mdadm -E /dev/sdb9
   mdadm: cannot open /dev/sdb9: No such device or address

   # mdadm -E /dev/sdb10
   mdadm: cannot open /dev/sdb10: No such device or address

   # mdadm -E /dev/sdb11
   mdadm: cannot open /dev/sdb11: No such device or address

>-------------------------------------------------------------------------------------------------------------<

mdadm -E /dev/md126pX do give something (except for md126p11):
>-------------------------------------------------------------------------------------------------------------<


   # mdadm -E /dev/md126p1
   /dev/md126p1:
              Magic : a92b4efc
            Version : 0.90.00
               UUID : 350e253f:863b7b04:b1617c47:b213a024
      Creation Time : Mon May  7 20:29:35 2007
         Raid Level : raid1
      Used Dev Size : 489856 (478.46 MiB 501.61 MB)
         Array Size : 489856 (478.46 MiB 501.61 MB)
       Raid Devices : 2
      Total Devices : 1
   Preferred Minor : 0

        Update Time : Thu Apr 30 08:00:08 2015
              State : clean
     Active Devices : 1
   Working Devices : 1
     Failed Devices : 1
      Spare Devices : 0
           Checksum : 6372ca75 - correct
             Events : 1769


          Number   Major   Minor   RaidDevice State
   this     0     259        0        0      active sync
     /dev/md/raidlin_0p1

       0     0     259        0        0      active sync
     /dev/md/raidlin_0p1
       1     1       0        0        1      faulty removed

   # mdadm -E /dev/md126p2
   /dev/md126p2:
              Magic : a92b4efc
            Version : 0.90.00
               UUID : 086e68ed:3607317f:60b56e23:6bae62bc
      Creation Time : Mon May  7 20:29:41 2007
         Raid Level : raid1
      Used Dev Size : 4883648 (4.66 GiB 5.00 GB)
         Array Size : 4883648 (4.66 GiB 5.00 GB)
       Raid Devices : 2
      Total Devices : 1
   Preferred Minor : 1

        Update Time : Thu Apr 30 20:57:21 2015
              State : clean
     Active Devices : 1
   Working Devices : 1
     Failed Devices : 1
      Spare Devices : 0
           Checksum : 4fd14bc9 - correct
             Events : 3980


          Number   Major   Minor   RaidDevice State
   this     0     259        1        0      active sync
     /dev/md/raidlin_0p2

       0     0     259        1        0      active sync
     /dev/md/raidlin_0p2
       1     1       0        0        1      faulty removed

   # mdadm -E /dev/md126p3
   /dev/md126p3:
              Magic : a92b4efc
            Version : 0.90.00
               UUID : 0f6e3ed5:aeee975a:c3647deb:763d68ce
      Creation Time : Mon May  7 20:29:48 2007
         Raid Level : raid1
      Used Dev Size : 1951808 (1906.38 MiB 1998.65 MB)
         Array Size : 1951808 (1906.38 MiB 1998.65 MB)
       Raid Devices : 2
      Total Devices : 1
   Preferred Minor : 2

        Update Time : Thu Apr 30 18:27:58 2015
              State : clean
     Active Devices : 1
   Working Devices : 1
     Failed Devices : 1
      Spare Devices : 0
           Checksum : 3cc9ac24 - correct
             Events : 1703


          Number   Major   Minor   RaidDevice State
   this     0     259        2        0      active sync
     /dev/md/raidlin_0p3

       0     0     259        2        0      active sync
     /dev/md/raidlin_0p3
       1     1       0        0        1      faulty removed

   # mdadm -E /dev/md126p4
   /dev/md126p4:
       MBR Magic : aa55
   Partition[0] :      9767457 sectors at           63 (type fd)
   Partition[1] :     19535040 sectors at      9767520 (type 05)

   # mdadm -E /dev/md126p5
   /dev/md126p5:
              Magic : a92b4efc
            Version : 0.90.00
               UUID : 6b560fe8:f24d6f2e:8942bd3a:1903abbc
      Creation Time : Mon May  7 20:29:55 2007
         Raid Level : raid1
      Used Dev Size : 4883648 (4.66 GiB 5.00 GB)
         Array Size : 4883648 (4.66 GiB 5.00 GB)
       Raid Devices : 2
      Total Devices : 1
   Preferred Minor : 3

        Update Time : Thu Apr 30 20:57:18 2015
              State : clean
     Active Devices : 1
   Working Devices : 1
     Failed Devices : 1
      Spare Devices : 0
           Checksum : 44e1e6e6 - correct
             Events : 7856


          Number   Major   Minor   RaidDevice State
   this     0     259        4        0      active sync
     /dev/md/raidlin_0p5

       0     0     259        4        0      active sync
     /dev/md/raidlin_0p5
       1     1       0        0        1      faulty removed

   # mdadm -E /dev/md126p6
   /dev/md126p6:
              Magic : a92b4efc
            Version : 0.90.00
               UUID : 462dab96:b7ca2a17:7c4aebf1:d4d7ec3b
      Creation Time : Mon May  7 20:30:02 2007
         Raid Level : raid1
      Used Dev Size : 9767424 (9.31 GiB 10.00 GB)
         Array Size : 9767424 (9.31 GiB 10.00 GB)
       Raid Devices : 2
      Total Devices : 1
   Preferred Minor : 4

        Update Time : Thu Apr 30 20:57:14 2015
              State : clean
     Active Devices : 1
   Working Devices : 1
     Failed Devices : 1
      Spare Devices : 0
           Checksum : 945d0c07 - correct
             Events : 2999


          Number   Major   Minor   RaidDevice State
   this     0     259        5        0      active sync
     /dev/md/raidlin_0p6

       0     0     259        5        0      active sync
     /dev/md/raidlin_0p6
       1     1       0        0        1      faulty removed

   # mdadm -E /dev/md126p7
   /dev/md126p7:
              Magic : a92b4efc
            Version : 0.90.00
               UUID : ea511351:3abc7b12:4c81e838:93dbd21a
      Creation Time : Mon May  7 20:30:09 2007
         Raid Level : raid1
      Used Dev Size : 4883648 (4.66 GiB 5.00 GB)
         Array Size : 4883648 (4.66 GiB 5.00 GB)
       Raid Devices : 2
      Total Devices : 1
   Preferred Minor : 5

        Update Time : Thu Apr 30 20:57:18 2015
              State : clean
     Active Devices : 1
   Working Devices : 1
     Failed Devices : 1
      Spare Devices : 0
           Checksum : 4a634da7 - correct
             Events : 8626


          Number   Major   Minor   RaidDevice State
   this     0     259        6        0      active sync
     /dev/md/raidlin_0p7

       0     0     259        6        0      active sync
     /dev/md/raidlin_0p7
       1     1       0        0        1      faulty removed

   # mdadm -E /dev/md126p8
   /dev/md126p8:
              Magic : a92b4efc
            Version : 0.90.00
               UUID : da0d76c6:91422584:dc3d6162:37ced53b
      Creation Time : Mon May  7 20:30:15 2007
         Raid Level : raid1
      Used Dev Size : 4883648 (4.66 GiB 5.00 GB)
         Array Size : 4883648 (4.66 GiB 5.00 GB)
       Raid Devices : 2
      Total Devices : 1
   Preferred Minor : 6

        Update Time : Thu Apr 30 20:57:31 2015
              State : active
     Active Devices : 1
   Working Devices : 1
     Failed Devices : 1
      Spare Devices : 0
           Checksum : c4540a0c - correct
             Events : 30082


          Number   Major   Minor   RaidDevice State
   this     0     259        7        0      active sync
     /dev/md/raidlin_0p8

       0     0     259        7        0      active sync
     /dev/md/raidlin_0p8
       1     1       0        0        1      faulty removed

   # mdadm -E /dev/md126p9
   /dev/md126p9:
              Magic : a92b4efc
            Version : 0.90.00
               UUID : 387c831c:8a6d05e3:b649696c:0870b930
      Creation Time : Mon May  7 20:30:21 2007
         Raid Level : raid1
      Used Dev Size : 4883648 (4.66 GiB 5.00 GB)
         Array Size : 4883648 (4.66 GiB 5.00 GB)
       Raid Devices : 2
      Total Devices : 1
   Preferred Minor : 7

        Update Time : Thu Apr 30 20:57:10 2015
              State : clean
     Active Devices : 1
   Working Devices : 1
     Failed Devices : 1
      Spare Devices : 0
           Checksum : c69b931a - correct
             Events : 4852


          Number   Major   Minor   RaidDevice State
   this     0     259        8        0      active sync
     /dev/md/raidlin_0p9

       0     0     259        8        0      active sync
     /dev/md/raidlin_0p9
       1     1       0        0        1      faulty removed

   # mdadm -E /dev/md126p10
   /dev/md126p10:
              Magic : a92b4efc
            Version : 0.90.00
               UUID : b07c4ab4:39d0ba53:9913afa9:fd9cc323
      Creation Time : Mon May  7 20:30:28 2007
         Raid Level : raid1
      Used Dev Size : 4883648 (4.66 GiB 5.00 GB)
         Array Size : 4883648 (4.66 GiB 5.00 GB)
       Raid Devices : 2
      Total Devices : 1
   Preferred Minor : 8

        Update Time : Thu Apr 30 20:57:28 2015
              State : clean
     Active Devices : 1
   Working Devices : 1
     Failed Devices : 1
      Spare Devices : 0
           Checksum : c5f5d015 - correct
             Events : 19271


          Number   Major   Minor   RaidDevice State
   this     0     259        9        0      active sync
     /dev/md/raidlin_0p10

       0     0     259        9        0      active sync
     /dev/md/raidlin_0p10
       1     1       0        0        1      faulty removed

   # mdadm -E /dev/md126p11
mdadm: No md superblock detected on /dev/md126p11.
>-------------------------------------------------------------------------------------------------------------<

Some more commands:
>-------------------------------------------------------------------------------------------------------------<

   # mdadm -D /dev/md126
   /dev/md126:
          Container : /dev/md/imsm0, member 0
         Raid Level : raid1
         Array Size : 244195328 (232.88 GiB 250.06 GB)
      Used Dev Size : 244195328 (232.88 GiB 250.06 GB)
       Raid Devices : 2
      Total Devices : 2

              State : active
     Active Devices : 2
   Working Devices : 2
     Failed Devices : 0
      Spare Devices : 0


               UUID : 91449a9d:9242bfe9:d99bceb0:a59f9314
        Number   Major   Minor   RaidDevice State
           1       8       32        0      active sync   /dev/sdc
           0       8       16        1      active sync   /dev/sdb

>-------------------------------------------------------------------------------------------------------------<

   # mdadm -E /dev/sdb
   mdmon: /dev/sdb is not attached to Intel(R) RAID controller.
   mdmon: /dev/sdb is not attached to Intel(R) RAID controller.
   /dev/sdb:
              Magic : Intel Raid ISM Cfg Sig.
            Version : 1.1.00
        Orig Family : 26b5a9e0
             Family : 26b5a9e0
         Generation : 00004db7
         Attributes : All supported
               UUID : d9cfa6d9:2a715e4f:1fbc2095:be342429
           Checksum : 261d2aed correct
        MPB Sectors : 1
              Disks : 2
       RAID Devices : 1

      Disk01 Serial : VFC100R10BE79D
              State : active
                 Id : 00010000
        Usable Size : 488390862 (232.88 GiB 250.06 GB)

   [raidlin]:
               UUID : 91449a9d:9242bfe9:d99bceb0:a59f9314
         RAID Level : 1
            Members : 2
              Slots : [UU]
        Failed disk : none
          This Slot : 1
         Array Size : 488390656 (232.88 GiB 250.06 GB)
       Per Dev Size : 488390656 (232.88 GiB 250.06 GB)
      Sector Offset : 0
        Num Stripes : 1907776
         Chunk Size : 64 KiB
           Reserved : 0
      Migrate State : idle
          Map State : normal
        Dirty State : dirty

      Disk00 Serial : VFC100R10BRKMD
              State : active
                 Id : 00000000
        Usable Size : 488390862 (232.88 GiB 250.06 GB)

>-------------------------------------------------------------------------------------------------------------<

   # mdadm -E /dev/sdc
   mdmon: /dev/sdc is not attached to Intel(R) RAID controller.
   mdmon: /dev/sdc is not attached to Intel(R) RAID controller.
   /dev/sdc:
              Magic : Intel Raid ISM Cfg Sig.
            Version : 1.1.00
        Orig Family : 26b5a9e0
             Family : 26b5a9e0
         Generation : 00004dbc
         Attributes : All supported
               UUID : d9cfa6d9:2a715e4f:1fbc2095:be342429
           Checksum : 261c2af2 correct
        MPB Sectors : 1
              Disks : 2
       RAID Devices : 1

      Disk00 Serial : VFC100R10BRKMD
              State : active
                 Id : 00000000
        Usable Size : 488390862 (232.88 GiB 250.06 GB)

   [raidlin]:
               UUID : 91449a9d:9242bfe9:d99bceb0:a59f9314
         RAID Level : 1
            Members : 2
              Slots : [UU]
        Failed disk : none
          This Slot : 0
         Array Size : 488390656 (232.88 GiB 250.06 GB)
       Per Dev Size : 488390656 (232.88 GiB 250.06 GB)
      Sector Offset : 0
        Num Stripes : 1907776
         Chunk Size : 64 KiB
           Reserved : 0
      Migrate State : idle
          Map State : normal
        Dirty State : clean

      Disk01 Serial : VFC100R10BE79D
              State : active
                 Id : 00010000
        Usable Size : 488390862 (232.88 GiB 250.06 GB)

>-------------------------------------------------------------------------------------------------------------<

I see something wicked in there and I don't know how to correct it...
I see at least two problems :
- mdadm is not able to detect md9
- my array does not seems fully operationa

Can you help ?

If you need more logs / command output, tell me which one.
I hope I'm not too much confusing...

kernel : Linux shax 3.16.0-4-686-pae #1 SMP Debian 3.16.7-ckt9-3~deb8u1 (2015-04-24) i686 GNU/Linux




--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux