Re: Rebuilt Array Issue

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



With a version 1 superblock, that is the way it is. Whether that may change, or at least change in the way it is displayed, is yet to be seen. Neil mentioned that he thought the display should be fixed, so the working device (#28) in your case should be listed in #26's position in the display, and the "removed" drive wouldn't be shown, to quote Neil:

It should look more like:

   Number   Major   Minor   RaidDevice State
      0       8        2        0      active sync   /dev/.static/dev/sda2
      1       8       18        1      active sync   /dev/.static/dev/sdb2
      3       8       34        2      active sync   /dev/.static/dev/sdc2

i.e. printing something that is 'removed' is pointless.  And the list
should be sorted by 'RaidDevice', not 'Number'.

Regards,
Tyler.

David M. Strang wrote:

Awhile back; with some help from Neil and the others in this mailing list -- I was able to bring my failed array back online. It's running healthy 28/28 disks -- however; I rebooted the other day. Attempted to reassemble the raid, and it would only assemble 27 of 28 disks. I had to assemble with a force and hotadd /dev/sdaa back into the raid and let it rebuild. Below is the output from my mdadm --detail /dev/md0

/dev/md0:
       Version : 01.00.01
 Creation Time : Wed Dec 31 19:00:00 1969
    Raid Level : raid5
    Array Size : 1935556992 (1845.89 GiB 1982.01 GB)
   Device Size : 71687296 (68.37 GiB 73.41 GB)
  Raid Devices : 28
 Total Devices : 28
Preferred Minor : 0
   Persistence : Superblock is persistent

   Update Time : Thu Aug 11 06:09:12 2005
         State : clean
Active Devices : 28
Working Devices : 28
Failed Devices : 0
 Spare Devices : 0

        Layout : left-asymmetric
    Chunk Size : 128K

          UUID : 4e2b6b0a8e:92e91c0c:018a4bf0:9bb74d
        Events : 259462

   Number   Major   Minor   RaidDevice State
0 8 0 0 active sync /dev/scsi/host2/bus0/target0/lun0/disc 1 8 16 1 active sync /dev/scsi/host2/bus0/target1/lun0/disc 2 8 32 2 active sync /dev/scsi/host2/bus0/target2/lun0/disc 3 8 48 3 active sync /dev/scsi/host2/bus0/target3/lun0/disc 4 8 64 4 active sync /dev/scsi/host2/bus0/target4/lun0/disc 5 8 80 5 active sync /dev/scsi/host2/bus0/target5/lun0/disc 6 8 96 6 active sync /dev/scsi/host2/bus0/target6/lun0/disc 7 8 112 7 active sync /dev/scsi/host2/bus0/target7/lun0/disc 8 8 128 8 active sync /dev/scsi/host2/bus0/target8/lun0/disc 9 8 144 9 active sync /dev/scsi/host2/bus0/target9/lun0/disc 10 8 160 10 active sync /dev/scsi/host2/bus0/target10/lun0/disc 11 8 176 11 active sync /dev/scsi/host2/bus0/target11/lun0/disc 12 8 192 12 active sync /dev/scsi/host2/bus0/target12/lun0/disc 13 8 208 13 active sync /dev/scsi/host2/bus0/target13/lun0/disc 14 8 224 14 active sync /dev/scsi/host2/bus0/target14/lun0/disc 15 8 240 15 active sync /dev/scsi/host2/bus0/target15/lun0/disc 16 65 0 16 active sync /dev/scsi/host2/bus0/target16/lun0/disc 17 65 16 17 active sync /dev/scsi/host2/bus0/target17/lun0/disc 18 65 32 18 active sync /dev/scsi/host2/bus0/target18/lun0/disc 19 65 48 19 active sync /dev/scsi/host2/bus0/target19/lun0/disc 20 65 64 20 active sync /dev/scsi/host2/bus0/target20/lun0/disc 21 65 80 21 active sync /dev/scsi/host2/bus0/target21/lun0/disc 22 65 96 22 active sync /dev/scsi/host2/bus0/target22/lun0/disc 23 65 112 23 active sync /dev/scsi/host2/bus0/target23/lun0/disc 24 65 128 24 active sync /dev/scsi/host2/bus0/target24/lun0/disc 25 65 144 25 active sync /dev/scsi/host2/bus0/target25/lun0/disc
     26       0        0        -      removed
27 65 176 27 active sync /dev/scsi/host2/bus0/target27/lun0/disc

28 65 160 26 active sync /dev/scsi/host2/bus0/target26/lun0/disc


Is it supposed to stay with number 26 as removed forever? Does number 28 ever jump back up to that spot? I shouldn't have to hotadd and allow it to resync everytime I re-assemble the raid should I?

-- David M. Strang
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux