apologies for the verbosity just adding some more info which is now making me lose hope. Using parted -l instead of fdisk gives me this: [root@lamachine ~]# parted -l Model: ATA WDC WD5000AAKS-0 (scsi) Disk /dev/sda: 500GB Sector size (logical/physical): 512B/512B Partition Table: msdos Disk Flags: Number Start End Size Type File system Flags 1 32.3kB 31.5GB 31.5GB primary raid 2 31.5GB 294GB 262GB primary ext4 raid 3 294GB 500GB 207GB extended 5 294GB 326GB 32.2GB logical 6 336GB 339GB 3644MB logical raid Model: ATA WDC WD5000AAKS-0 (scsi) Disk /dev/sdb: 500GB Sector size (logical/physical): 512B/512B Partition Table: msdos Disk Flags: Number Start End Size Type File system Flags 2 210MB 262GB 262GB primary raid 3 262GB 271GB 8389MB primary linux-swap(v1) 4 271GB 500GB 229GB extended 5 271GB 303GB 32.2GB logical 6 313GB 317GB 3644MB logical raid Error: Invalid argument during seek for read on /dev/sdc Retry/Ignore/Cancel? R Error: Invalid argument during seek for read on /dev/sdc Retry/Ignore/Cancel? I Error: The backup GPT table is corrupt, but the primary appears OK, so that will be used. OK/Cancel? O Model: ATA WDC WD30EZRX-00D (scsi) Disk /dev/sdc: 3001GB Sector size (logical/physical): 512B/4096B Partition Table: unknown Disk Flags: Model: ATA WDC WD30EZRX-00D (scsi) Disk /dev/sdd: 3001GB Sector size (logical/physical): 512B/4096B Partition Table: gpt Disk Flags: Number Start End Size File system Name Flags 1 1049kB 2199GB 2199GB raid 2 2199GB 2736GB 537GB Error: Invalid argument during seek for read on /dev/sde Retry/Ignore/Cancel? C Model: ATA WDC WD30EZRX-00D (scsi) Disk /dev/sde: 3001GB Sector size (logical/physical): 512B/4096B Partition Table: unknown Disk Flags: Model: ATA WDC WD5000AAKS-0 (scsi) Disk /dev/sdf: 500GB Sector size (logical/physical): 512B/512B Partition Table: msdos Disk Flags: Number Start End Size Type File system Flags 1 1049kB 210MB 209MB primary ext4 boot 2 210MB 31.7GB 31.5GB primary raid 3 31.7GB 294GB 262GB primary ext4 raid 4 294GB 500GB 206GB extended 5 294GB 326GB 32.2GB logical 6 336GB 340GB 3644MB logical raid Model: Linux Software RAID Array (md) Disk /dev/md2: 524GB Sector size (logical/physical): 512B/512B Partition Table: loop Disk Flags: Number Start End Size File system Flags 1 0.00B 524GB 524GB ext4 Error: /dev/md126: unrecognised disk label Model: Linux Software RAID Array (md) Disk /dev/md126: 31.5GB Sector size (logical/physical): 512B/512B Partition Table: unknown Disk Flags: Error: /dev/md127: unrecognised disk label Model: Linux Software RAID Array (md) Disk /dev/md127: 96.6GB Sector size (logical/physical): 512B/512B Partition Table: unknown Disk Flags: On 12 September 2016 at 20:41, Daniel Sanabria <sanabria.d@xxxxxxxxx> wrote: > ok, I just adjusted system time so that I can start tracking logs. > > what I'm noticing however is that fdisk -l is not giving me the expect > partitions (I was expecting at least 2 partitions in every 2.7 disk > similar to what I have in sdd): > > [root@lamachine lamachine_220315]# fdisk -l /dev/{sdc,sdd,sde} > Disk /dev/sdc: 2.7 TiB, 3000591900160 bytes, 5860531055 sectors > Units: sectors of 1 * 512 = 512 bytes > Sector size (logical/physical): 512 bytes / 4096 bytes > I/O size (minimum/optimal): 4096 bytes / 4096 bytes > Disklabel type: dos > Disk identifier: 0x00000000 > > Device Boot Start End Sectors Size Id Type > /dev/sdc1 1 4294967295 4294967295 2T ee GPT > > Partition 1 does not start on physical sector boundary. > Disk /dev/sdd: 2.7 TiB, 3000592982016 bytes, 5860533168 sectors > Units: sectors of 1 * 512 = 512 bytes > Sector size (logical/physical): 512 bytes / 4096 bytes > I/O size (minimum/optimal): 4096 bytes / 4096 bytes > Disklabel type: gpt > Disk identifier: D3233810-F552-4126-8281-7F71A4938DF9 > > Device Start End Sectors Size Type > /dev/sdd1 2048 4294969343 4294967296 2T Linux RAID > /dev/sdd2 4294969344 5343545343 1048576000 500G Linux filesystem > Disk /dev/sde: 2.7 TiB, 3000591900160 bytes, 5860531055 sectors > Units: sectors of 1 * 512 = 512 bytes > Sector size (logical/physical): 512 bytes / 4096 bytes > I/O size (minimum/optimal): 4096 bytes / 4096 bytes > Disklabel type: dos > Disk identifier: 0x00000000 > > Device Boot Start End Sectors Size Id Type > /dev/sde1 1 4294967295 4294967295 2T ee GPT > > Partition 1 does not start on physical sector boundary. > [root@lamachine lamachine_220315]# > > what could've happened here? any ideas why the partition tables ended > up like that? > > From previous information I have an idea of what the md128 and md129 > are supposed to looks like (also noticed that the device names > changed): > > # md128 and md129 details From an old command output > /dev/md128: > Version : 1.2 > Creation Time : Fri Oct 24 15:24:38 2014 > Raid Level : raid5 > Array Size : 4294705152 (4095.75 GiB 4397.78 GB) > Used Dev Size : 2147352576 (2047.88 GiB 2198.89 GB) > Raid Devices : 3 > Total Devices : 3 > Persistence : Superblock is persistent > > Intent Bitmap : Internal > > Update Time : Sun Mar 22 06:20:08 2015 > State : clean > Active Devices : 3 > Working Devices : 3 > Failed Devices : 0 > Spare Devices : 0 > > Layout : left-symmetric > Chunk Size : 512K > > Name : lamachine:128 (local to host lamachine) > UUID : f2372cb9:d3816fd6:ce86d826:882ec82e > Events : 4041 > > Number Major Minor RaidDevice State > 0 8 49 0 active sync /dev/sdd1 > 1 8 65 1 active sync /dev/sde1 > 3 8 81 2 active sync /dev/sdf1 > /dev/md129: > Version : 1.2 > Creation Time : Mon Nov 10 16:28:11 2014 > Raid Level : raid0 > Array Size : 1572470784 (1499.63 GiB 1610.21 GB) > Raid Devices : 3 > Total Devices : 3 > Persistence : Superblock is persistent > > Update Time : Mon Nov 10 16:28:11 2014 > State : clean > Active Devices : 3 > Working Devices : 3 > Failed Devices : 0 > Spare Devices : 0 > > Chunk Size : 512K > > Name : lamachine:129 (local to host lamachine) > UUID : 895dae98:d1a496de:4f590b8b:cb8ac12a > Events : 0 > > Number Major Minor RaidDevice State > 0 8 50 0 active sync /dev/sdd2 > 1 8 66 1 active sync /dev/sde2 > 2 8 82 2 active sync /dev/sdf2 > > Is there any way to recover the contents of these two arrays ? :( > > On 11 September 2016 at 21:06, Daniel Sanabria <sanabria.d@xxxxxxxxx> wrote: >> However I'm noticing that the details with this new MB are somewhat different: >> >> [root@lamachine ~]# cat /etc/mdadm.conf >> # mdadm.conf written out by anaconda >> MAILADDR root >> AUTO +imsm +1.x -all >> ARRAY /dev/md2 level=raid5 num-devices=3 >> UUID=2cff15d1:e411447b:fd5d4721:03e44022 >> ARRAY /dev/md126 level=raid10 num-devices=2 >> UUID=9af006ca:8845bbd3:bfe78010:bc810f04 >> ARRAY /dev/md127 level=raid0 num-devices=3 >> UUID=acd5374f:72628c93:6a906c4b:5f675ce5 >> ARRAY /dev/md128 metadata=1.2 spares=1 name=lamachine:128 >> UUID=f2372cb9:d3816fd6:ce86d826:882ec82e >> ARRAY /dev/md129 metadata=1.2 name=lamachine:129 >> UUID=895dae98:d1a496de:4f590b8b:cb8ac12a >> [root@lamachine ~]# mdadm --detail /dev/md1* >> /dev/md126: >> Version : 0.90 >> Creation Time : Thu Dec 3 22:12:12 2009 >> Raid Level : raid10 >> Array Size : 30719936 (29.30 GiB 31.46 GB) >> Used Dev Size : 30719936 (29.30 GiB 31.46 GB) >> Raid Devices : 2 >> Total Devices : 2 >> Preferred Minor : 126 >> Persistence : Superblock is persistent >> >> Update Time : Tue Jan 12 04:03:41 2016 >> State : clean >> Active Devices : 2 >> Working Devices : 2 >> Failed Devices : 0 >> Spare Devices : 0 >> >> Layout : near=2 >> Chunk Size : 64K >> >> UUID : 9af006ca:8845bbd3:bfe78010:bc810f04 >> Events : 0.264152 >> >> Number Major Minor RaidDevice State >> 0 8 82 0 active sync set-A /dev/sdf2 >> 1 8 1 1 active sync set-B /dev/sda1 >> /dev/md127: >> Version : 1.2 >> Creation Time : Tue Jul 26 19:00:28 2011 >> Raid Level : raid0 >> Array Size : 94367232 (90.00 GiB 96.63 GB) >> Raid Devices : 3 >> Total Devices : 3 >> Persistence : Superblock is persistent >> >> Update Time : Tue Jul 26 19:00:28 2011 >> State : clean >> Active Devices : 3 >> Working Devices : 3 >> Failed Devices : 0 >> Spare Devices : 0 >> >> Chunk Size : 512K >> >> Name : reading.homeunix.com:3 >> UUID : acd5374f:72628c93:6a906c4b:5f675ce5 >> Events : 0 >> >> Number Major Minor RaidDevice State >> 0 8 85 0 active sync /dev/sdf5 >> 1 8 21 1 active sync /dev/sdb5 >> 2 8 5 2 active sync /dev/sda5 >> /dev/md128: >> Version : 1.2 >> Raid Level : raid0 >> Total Devices : 1 >> Persistence : Superblock is persistent >> >> State : inactive >> >> Name : lamachine:128 (local to host lamachine) >> UUID : f2372cb9:d3816fd6:ce86d826:882ec82e >> Events : 4154 >> >> Number Major Minor RaidDevice >> >> - 8 49 - /dev/sdd1 >> /dev/md129: >> Version : 1.2 >> Raid Level : raid0 >> Total Devices : 1 >> Persistence : Superblock is persistent >> >> State : inactive >> >> Name : lamachine:129 (local to host lamachine) >> UUID : 895dae98:d1a496de:4f590b8b:cb8ac12a >> Events : 0 >> >> Number Major Minor RaidDevice >> >> - 8 50 - /dev/sdd2 >> [root@lamachine ~]# mdadm --detail /dev/md2* >> /dev/md2: >> Version : 0.90 >> Creation Time : Mon Feb 11 07:54:36 2013 >> Raid Level : raid5 >> Array Size : 511999872 (488.28 GiB 524.29 GB) >> Used Dev Size : 255999936 (244.14 GiB 262.14 GB) >> Raid Devices : 3 >> Total Devices : 3 >> Preferred Minor : 2 >> Persistence : Superblock is persistent >> >> Update Time : Tue Jan 12 02:31:50 2016 >> State : clean >> Active Devices : 3 >> Working Devices : 3 >> Failed Devices : 0 >> Spare Devices : 0 >> >> Layout : left-symmetric >> Chunk Size : 64K >> >> UUID : 2cff15d1:e411447b:fd5d4721:03e44022 (local to host lamachine) >> Events : 0.611 >> >> Number Major Minor RaidDevice State >> 0 8 83 0 active sync /dev/sdf3 >> 1 8 18 1 active sync /dev/sdb2 >> 2 8 2 2 active sync /dev/sda2 >> [root@lamachine ~]# cat /proc/mdstat >> Personalities : [raid10] [raid0] [raid6] [raid5] [raid4] >> md2 : active raid5 sda2[2] sdf3[0] sdb2[1] >> 511999872 blocks level 5, 64k chunk, algorithm 2 [3/3] [UUU] >> >> md127 : active raid0 sda5[2] sdf5[0] sdb5[1] >> 94367232 blocks super 1.2 512k chunks >> >> md129 : inactive sdd2[2](S) >> 524156928 blocks super 1.2 >> >> md128 : inactive sdd1[3](S) >> 2147352576 blocks super 1.2 >> >> md126 : active raid10 sdf2[0] sda1[1] >> 30719936 blocks 2 near-copies [2/2] [UU] >> >> unused devices: <none> >> [root@lamachine ~]# >> >> On 11 September 2016 at 19:48, Daniel Sanabria <sanabria.d@xxxxxxxxx> wrote: >>> ok, system up and running after MB was replaced however the arrays >>> remain inactive. >>> >>> mdadm version is: >>> mdadm - v3.3.4 - 3rd August 2015 >>> >>> Here's the output from Phil's lsdrv: >>> >>> [root@lamachine ~]# ./lsdrv >>> PCI [ahci] 00:1f.2 SATA controller: Intel Corporation C600/X79 series >>> chipset 6-Port SATA AHCI Controller (rev 06) >>> ├scsi 0:0:0:0 ATA WDC WD5000AAKS-0 {WD-WCASZ0505379} >>> │└sda 465.76g [8:0] Partitioned (dos) >>> │ ├sda1 29.30g [8:1] MD raid10,near2 (1/2) (w/ sdf2) in_sync >>> {9af006ca-8845-bbd3-bfe7-8010bc810f04} >>> │ │└md126 29.30g [9:126] MD v0.90 raid10,near2 (2) clean, 64k Chunk >>> {9af006ca:8845bbd3:bfe78010:bc810f04} >>> │ │ │ PV LVM2_member 28.03g used, 1.26g free >>> {cE4ePh-RWO8-Wgdy-YPOY-ehyC-KI6u-io1cyH} >>> │ │ └VG vg_bigblackbox 29.29g 1.26g free >>> {VWfuwI-5v2q-w8qf-FEbc-BdGW-3mKX-pZd7hR} >>> │ │ ├dm-2 7.81g [253:2] LV LogVol_opt ext4 >>> {b08d7f5e-f15f-4241-804e-edccecab6003} >>> │ │ │└Mounted as /dev/mapper/vg_bigblackbox-LogVol_opt @ /opt >>> │ │ ├dm-0 9.77g [253:0] LV LogVol_root ext4 >>> {4dabd6b0-b1a3-464d-8ed7-0aab93fab6c3} >>> │ │ │└Mounted as /dev/mapper/vg_bigblackbox-LogVol_root @ / >>> │ │ ├dm-3 1.95g [253:3] LV LogVol_tmp ext4 >>> {f6b46363-170b-4038-83bd-2c5f9f6a1973} >>> │ │ │└Mounted as /dev/mapper/vg_bigblackbox-LogVol_tmp @ /tmp >>> │ │ └dm-1 8.50g [253:1] LV LogVol_var ext4 >>> {ab165c61-3d62-4c55-8639-6c2c2bf4b021} >>> │ │ └Mounted as /dev/mapper/vg_bigblackbox-LogVol_var @ /var >>> │ ├sda2 244.14g [8:2] MD raid5 (2/3) (w/ sdb2,sdf3) in_sync >>> {2cff15d1-e411-447b-fd5d-472103e44022} >>> │ │└md2 488.28g [9:2] MD v0.90 raid5 (3) clean, 64k Chunk >>> {2cff15d1:e411447b:fd5d4721:03e44022} >>> │ │ │ ext4 {e9c1c787-496f-4e8f-b62e-35d5b1ff8311} >>> │ │ └Mounted as /dev/md2 @ /home >>> │ ├sda3 1.00k [8:3] Partitioned (dos) >>> │ ├sda5 30.00g [8:5] MD raid0 (2/3) (w/ sdb5,sdf5) in_sync >>> 'reading.homeunix.com:3' {acd5374f-7262-8c93-6a90-6c4b5f675ce5} >>> │ │└md127 90.00g [9:127] MD v1.2 raid0 (3) clean, 512k Chunk, None >>> (None) None {acd5374f:72628c93:6a906c4b:5f675ce5} >>> │ │ │ PV LVM2_member 86.00g used, 3.99g free >>> {VmsWRd-8qHt-bauf-lvAn-FC97-KyH5-gk89ox} >>> │ │ └VG libvirt_lvm 89.99g 3.99g free {t8GQck-f2Eu-iD2V-fnJQ-kBm6-QyKw-dR31PB} >>> │ │ ├dm-6 8.00g [253:6] LV builder2 Partitioned (dos) >>> │ │ ├dm-7 8.00g [253:7] LV builder3 Partitioned (dos) >>> │ │ ├dm-9 8.00g [253:9] LV builder5.3 Partitioned (dos) >>> │ │ ├dm-8 8.00g [253:8] LV builder5.6 Partitioned (dos) >>> │ │ ├dm-5 8.00g [253:5] LV centos_updt Partitioned (dos) >>> │ │ ├dm-10 16.00g [253:10] LV f22lvm Partitioned (dos) >>> │ │ └dm-4 30.00g [253:4] LV win7 Partitioned (dos) >>> │ └sda6 3.39g [8:6] Empty/Unknown >>> ├scsi 1:0:0:0 ATA WDC WD5000AAKS-0 {WD-WCASY7694185} >>> │└sdb 465.76g [8:16] Partitioned (dos) >>> │ ├sdb2 244.14g [8:18] MD raid5 (1/3) (w/ sda2,sdf3) in_sync >>> {2cff15d1-e411-447b-fd5d-472103e44022} >>> │ │└md2 488.28g [9:2] MD v0.90 raid5 (3) clean, 64k Chunk >>> {2cff15d1:e411447b:fd5d4721:03e44022} >>> │ │ ext4 {e9c1c787-496f-4e8f-b62e-35d5b1ff8311} >>> │ ├sdb3 7.81g [8:19] swap {9194f492-881a-4fc3-ac09-ca4e1cc2985a} >>> │ ├sdb4 1.00k [8:20] Partitioned (dos) >>> │ ├sdb5 30.00g [8:21] MD raid0 (1/3) (w/ sda5,sdf5) in_sync >>> 'reading.homeunix.com:3' {acd5374f-7262-8c93-6a90-6c4b5f675ce5} >>> │ │└md127 90.00g [9:127] MD v1.2 raid0 (3) clean, 512k Chunk, None >>> (None) None {acd5374f:72628c93:6a906c4b:5f675ce5} >>> │ │ PV LVM2_member 86.00g used, 3.99g free >>> {VmsWRd-8qHt-bauf-lvAn-FC97-KyH5-gk89ox} >>> │ └sdb6 3.39g [8:22] Empty/Unknown >>> ├scsi 2:x:x:x [Empty] >>> ├scsi 3:x:x:x [Empty] >>> ├scsi 4:x:x:x [Empty] >>> └scsi 5:x:x:x [Empty] >>> PCI [ahci] 0a:00.0 SATA controller: Marvell Technology Group Ltd. >>> 88SE9230 PCIe SATA 6Gb/s Controller (rev 11) >>> ├scsi 6:0:0:0 ATA WDC WD30EZRX-00D {WD-WCC4NCWT13RF} >>> │└sdc 2.73t [8:32] Partitioned (PMBR) >>> ├scsi 7:0:0:0 ATA WDC WD30EZRX-00D {WD-WCC4NPRDD6D7} >>> │└sdd 2.73t [8:48] Partitioned (gpt) >>> │ ├sdd1 2.00t [8:49] MD (none/) spare 'lamachine:128' >>> {f2372cb9-d381-6fd6-ce86-d826882ec82e} >>> │ │└md128 0.00k [9:128] MD v1.2 () inactive, None (None) None >>> {f2372cb9:d3816fd6:ce86d826:882ec82e} >>> │ │ Empty/Unknown >>> │ └sdd2 500.00g [8:50] MD (none/) spare 'lamachine:129' >>> {895dae98-d1a4-96de-4f59-0b8bcb8ac12a} >>> │ └md129 0.00k [9:129] MD v1.2 () inactive, None (None) None >>> {895dae98:d1a496de:4f590b8b:cb8ac12a} >>> │ Empty/Unknown >>> ├scsi 8:0:0:0 ATA WDC WD30EZRX-00D {WD-WCC4N1294906} >>> │└sde 2.73t [8:64] Partitioned (PMBR) >>> ├scsi 9:0:0:0 ATA WDC WD5000AAKS-0 {WD-WMAWF0085724} >>> │└sdf 465.76g [8:80] Partitioned (dos) >>> │ ├sdf1 199.00m [8:81] ext4 {4e51f903-37ca-4479-9197-fac7b2280557} >>> │ │└Mounted as /dev/sdf1 @ /boot >>> │ ├sdf2 29.30g [8:82] MD raid10,near2 (0/2) (w/ sda1) in_sync >>> {9af006ca-8845-bbd3-bfe7-8010bc810f04} >>> │ │└md126 29.30g [9:126] MD v0.90 raid10,near2 (2) clean, 64k Chunk >>> {9af006ca:8845bbd3:bfe78010:bc810f04} >>> │ │ PV LVM2_member 28.03g used, 1.26g free >>> {cE4ePh-RWO8-Wgdy-YPOY-ehyC-KI6u-io1cyH} >>> │ ├sdf3 244.14g [8:83] MD raid5 (0/3) (w/ sda2,sdb2) in_sync >>> {2cff15d1-e411-447b-fd5d-472103e44022} >>> │ │└md2 488.28g [9:2] MD v0.90 raid5 (3) clean, 64k Chunk >>> {2cff15d1:e411447b:fd5d4721:03e44022} >>> │ │ ext4 {e9c1c787-496f-4e8f-b62e-35d5b1ff8311} >>> │ ├sdf4 1.00k [8:84] Partitioned (dos) >>> │ ├sdf5 30.00g [8:85] MD raid0 (0/3) (w/ sda5,sdb5) in_sync >>> 'reading.homeunix.com:3' {acd5374f-7262-8c93-6a90-6c4b5f675ce5} >>> │ │└md127 90.00g [9:127] MD v1.2 raid0 (3) clean, 512k Chunk, None >>> (None) None {acd5374f:72628c93:6a906c4b:5f675ce5} >>> │ │ PV LVM2_member 86.00g used, 3.99g free >>> {VmsWRd-8qHt-bauf-lvAn-FC97-KyH5-gk89ox} >>> │ └sdf6 3.39g [8:86] Empty/Unknown >>> ├scsi 10:x:x:x [Empty] >>> ├scsi 11:x:x:x [Empty] >>> └scsi 12:x:x:x [Empty] >>> PCI [isci] 05:00.0 Serial Attached SCSI controller: Intel Corporation >>> C602 chipset 4-Port SATA Storage Control Unit (rev 06) >>> └scsi 14:x:x:x [Empty] >>> [root@lamachine ~]# >>> >>> Thanks in advance for any recommendations on what steps to take in >>> order to bring these arrays back online. >>> >>> Regards, >>> >>> Daniel >>> >>> >>> On 2 August 2016 at 11:45, Daniel Sanabria <sanabria.d@xxxxxxxxx> wrote: >>>> Thanks very much for the response Wol. >>>> >>>> It looks like the PSU is dead (server automatically powers off a few >>>> seconds after power on). >>>> >>>> I'm planning to order a PSU replacement to resume troubleshooting so >>>> please bear with me; maybe the PSU was degraded and couldn't power >>>> some of drives? >>>> >>>> Cheers, >>>> >>>> Daniel >>>> >>>> On 2 August 2016 at 11:17, Wols Lists <antlists@xxxxxxxxxxxxxxx> wrote: >>>>> Just a quick first response. I see md128 and md129 are both down, and >>>>> are both listed as one drive, raid0. Bit odd, that ... >>>>> >>>>> What version of mdadm are you using? One of them had a bug (3.2.3 era?) >>>>> that would split an array in two. Is it possible that you should have >>>>> one raid0 array with sdf1 and sdf2? But that's a bit of a weird setup... >>>>> >>>>> I notice also that md126 is raid10 across two drives. That's odd, too. >>>>> >>>>> How much do you know about what the setup should be, and why it was set >>>>> up that way? >>>>> >>>>> Download lspci by Phil Turmel (it requires python2.7, if your machine is >>>>> python3 a quick fix to the shebang at the start should get it to work). >>>>> Post the output from that here. >>>>> >>>>> Cheers, >>>>> Wol >>>>> >>>>> On 02/08/16 08:36, Daniel Sanabria wrote: >>>>>> Hi All, >>>>>> >>>>>> I have a box that I believe was not powered down correctly and after >>>>>> transporting it to a different location it doesn't boot anymore >>>>>> stopping at BIOS check "Verifying DMI Pool Data". >>>>>> >>>>>> The box have 6 drives and after instructing the BIOS to boot from the >>>>>> first drive I managed to boot the OS (Fedora 23) after commenting out >>>>>> 2 /etc/fstab entries , output for "uname -a; cat /etc/fstab" follows: >>>>>> >>>>>> [root@lamachine ~]# uname -a; cat /etc/fstab >>>>>> Linux lamachine 4.3.3-303.fc23.x86_64 #1 SMP Tue Jan 19 18:31:55 UTC >>>>>> 2016 x86_64 x86_64 x86_64 GNU/Linux >>>>>> >>>>>> # >>>>>> # /etc/fstab >>>>>> # Created by anaconda on Tue Mar 24 19:31:21 2015 >>>>>> # >>>>>> # Accessible filesystems, by reference, are maintained under '/dev/disk' >>>>>> # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info >>>>>> # >>>>>> /dev/mapper/vg_bigblackbox-LogVol_root / ext4 >>>>>> defaults 1 1 >>>>>> UUID=4e51f903-37ca-4479-9197-fac7b2280557 /boot ext4 >>>>>> defaults 1 2 >>>>>> /dev/mapper/vg_bigblackbox-LogVol_opt /opt ext4 >>>>>> defaults 1 2 >>>>>> /dev/mapper/vg_bigblackbox-LogVol_tmp /tmp ext4 >>>>>> defaults 1 2 >>>>>> /dev/mapper/vg_bigblackbox-LogVol_var /var ext4 >>>>>> defaults 1 2 >>>>>> UUID=9194f492-881a-4fc3-ac09-ca4e1cc2985a swap swap >>>>>> defaults 0 0 >>>>>> /dev/md2 /home ext4 defaults 1 2 >>>>>> #/dev/vg_media/lv_media /mnt/media ext4 defaults 1 2 >>>>>> #/dev/vg_virt_dir/lv_virt_dir1 /mnt/guest_images/ ext4 defaults 1 2 >>>>>> [root@lamachine ~]# >>>>>> >>>>>> When checking mdstat I can see that 2 of the arrays are showing up as >>>>>> inactive, but not sure how to safely activate these so looking for >>>>>> some knowledgeable advice on how to proceed here. >>>>>> >>>>>> Thanks in advance, >>>>>> >>>>>> Daniel >>>>>> >>>>>> Below some more relevant outputs: >>>>>> >>>>>> [root@lamachine ~]# cat /proc/mdstat >>>>>> Personalities : [raid10] [raid6] [raid5] [raid4] [raid0] >>>>>> md127 : active raid0 sda5[0] sdc5[2] sdb5[1] >>>>>> 94367232 blocks super 1.2 512k chunks >>>>>> >>>>>> md2 : active raid5 sda3[0] sdc2[2] sdb2[1] >>>>>> 511999872 blocks level 5, 64k chunk, algorithm 2 [3/3] [UUU] >>>>>> >>>>>> md128 : inactive sdf1[3](S) >>>>>> 2147352576 blocks super 1.2 >>>>>> >>>>>> md129 : inactive sdf2[2](S) >>>>>> 524156928 blocks super 1.2 >>>>>> >>>>>> md126 : active raid10 sda2[0] sdc1[1] >>>>>> 30719936 blocks 2 near-copies [2/2] [UU] >>>>>> >>>>>> unused devices: <none> >>>>>> [root@lamachine ~]# cat /etc/mdadm.conf >>>>>> # mdadm.conf written out by anaconda >>>>>> MAILADDR root >>>>>> AUTO +imsm +1.x -all >>>>>> ARRAY /dev/md2 level=raid5 num-devices=3 >>>>>> UUID=2cff15d1:e411447b:fd5d4721:03e44022 >>>>>> ARRAY /dev/md126 level=raid10 num-devices=2 >>>>>> UUID=9af006ca:8845bbd3:bfe78010:bc810f04 >>>>>> ARRAY /dev/md127 level=raid0 num-devices=3 >>>>>> UUID=acd5374f:72628c93:6a906c4b:5f675ce5 >>>>>> ARRAY /dev/md128 metadata=1.2 spares=1 name=lamachine:128 >>>>>> UUID=f2372cb9:d3816fd6:ce86d826:882ec82e >>>>>> ARRAY /dev/md129 metadata=1.2 name=lamachine:129 >>>>>> UUID=895dae98:d1a496de:4f590b8b:cb8ac12a >>>>>> [root@lamachine ~]# mdadm --detail /dev/md1* >>>>>> /dev/md126: >>>>>> Version : 0.90 >>>>>> Creation Time : Thu Dec 3 22:12:12 2009 >>>>>> Raid Level : raid10 >>>>>> Array Size : 30719936 (29.30 GiB 31.46 GB) >>>>>> Used Dev Size : 30719936 (29.30 GiB 31.46 GB) >>>>>> Raid Devices : 2 >>>>>> Total Devices : 2 >>>>>> Preferred Minor : 126 >>>>>> Persistence : Superblock is persistent >>>>>> >>>>>> Update Time : Tue Aug 2 07:46:39 2016 >>>>>> State : clean >>>>>> Active Devices : 2 >>>>>> Working Devices : 2 >>>>>> Failed Devices : 0 >>>>>> Spare Devices : 0 >>>>>> >>>>>> Layout : near=2 >>>>>> Chunk Size : 64K >>>>>> >>>>>> UUID : 9af006ca:8845bbd3:bfe78010:bc810f04 >>>>>> Events : 0.264152 >>>>>> >>>>>> Number Major Minor RaidDevice State >>>>>> 0 8 2 0 active sync set-A /dev/sda2 >>>>>> 1 8 33 1 active sync set-B /dev/sdc1 >>>>>> /dev/md127: >>>>>> Version : 1.2 >>>>>> Creation Time : Tue Jul 26 19:00:28 2011 >>>>>> Raid Level : raid0 >>>>>> Array Size : 94367232 (90.00 GiB 96.63 GB) >>>>>> Raid Devices : 3 >>>>>> Total Devices : 3 >>>>>> Persistence : Superblock is persistent >>>>>> >>>>>> Update Time : Tue Jul 26 19:00:28 2011 >>>>>> State : clean >>>>>> Active Devices : 3 >>>>>> Working Devices : 3 >>>>>> Failed Devices : 0 >>>>>> Spare Devices : 0 >>>>>> >>>>>> Chunk Size : 512K >>>>>> >>>>>> Name : reading.homeunix.com:3 >>>>>> UUID : acd5374f:72628c93:6a906c4b:5f675ce5 >>>>>> Events : 0 >>>>>> >>>>>> Number Major Minor RaidDevice State >>>>>> 0 8 5 0 active sync /dev/sda5 >>>>>> 1 8 21 1 active sync /dev/sdb5 >>>>>> 2 8 37 2 active sync /dev/sdc5 >>>>>> /dev/md128: >>>>>> Version : 1.2 >>>>>> Raid Level : raid0 >>>>>> Total Devices : 1 >>>>>> Persistence : Superblock is persistent >>>>>> >>>>>> State : inactive >>>>>> >>>>>> Name : lamachine:128 (local to host lamachine) >>>>>> UUID : f2372cb9:d3816fd6:ce86d826:882ec82e >>>>>> Events : 4154 >>>>>> >>>>>> Number Major Minor RaidDevice >>>>>> >>>>>> - 8 81 - /dev/sdf1 >>>>>> /dev/md129: >>>>>> Version : 1.2 >>>>>> Raid Level : raid0 >>>>>> Total Devices : 1 >>>>>> Persistence : Superblock is persistent >>>>>> >>>>>> State : inactive >>>>>> >>>>>> Name : lamachine:129 (local to host lamachine) >>>>>> UUID : 895dae98:d1a496de:4f590b8b:cb8ac12a >>>>>> Events : 0 >>>>>> >>>>>> Number Major Minor RaidDevice >>>>>> >>>>>> - 8 82 - /dev/sdf2 >>>>>> [root@lamachine ~]# mdadm --detail /dev/md2 >>>>>> /dev/md2: >>>>>> Version : 0.90 >>>>>> Creation Time : Mon Feb 11 07:54:36 2013 >>>>>> Raid Level : raid5 >>>>>> Array Size : 511999872 (488.28 GiB 524.29 GB) >>>>>> Used Dev Size : 255999936 (244.14 GiB 262.14 GB) >>>>>> Raid Devices : 3 >>>>>> Total Devices : 3 >>>>>> Preferred Minor : 2 >>>>>> Persistence : Superblock is persistent >>>>>> >>>>>> Update Time : Mon Aug 1 20:24:23 2016 >>>>>> State : clean >>>>>> Active Devices : 3 >>>>>> Working Devices : 3 >>>>>> Failed Devices : 0 >>>>>> Spare Devices : 0 >>>>>> >>>>>> Layout : left-symmetric >>>>>> Chunk Size : 64K >>>>>> >>>>>> UUID : 2cff15d1:e411447b:fd5d4721:03e44022 (local to host lamachine) >>>>>> Events : 0.611 >>>>>> >>>>>> Number Major Minor RaidDevice State >>>>>> 0 8 3 0 active sync /dev/sda3 >>>>>> 1 8 18 1 active sync /dev/sdb2 >>>>>> 2 8 34 2 active sync /dev/sdc2 >>>>>> [root@lamachine ~]# >>>>>> -- >>>>>> To unsubscribe from this list: send the line "unsubscribe linux-raid" in >>>>>> the body of a message to majordomo@xxxxxxxxxxxxxxx >>>>>> More majordomo info at http://vger.kernel.org/majordomo-info.html >>>>>> >>>>> -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html