Hello, I, personally, just use the device and array lines. You're welcome to keep tracking down why UUID detection doesn't work for you if you so chose. Example: DEVICE /dev/sda1 /dev/sdb1 ... ARRAY /dev/md0 metadata=1.2 spares=1 name=jackie:0 /dev/sda1 /dev/sdb1 ... If you just want the array to work (for now), then: mdadm --assemble /dev/md0 /dev/sd{a,b,e,f,g}1 should do the trick. One question though, why does sdc not have a partition table? I mean, it doesn't really matter if you use a RAID array without one, but it stands out from the rest of the info as erroneous. Sort of like either you goofed (hopefully), or the drive isn't being detected/working properly. Sincerely, David PS: I'm subscribed to the list. No need to CC me. On Tue, 23 Jan 2024 16:16:12 +0000 (UTC) RJ Marquette <rjm1@xxxxxxxxx> wrote: > (Sorry if this came through twice without the mdadm.conf contents, > somehow I accidentally hit send when I was trying to paste in.) > > Thanks. All drives in the system are being detected (/dev/sdd is my > system drive - the rest are all of the array): > > rj@jackie:~$ ls -l /dev/sd* > brw-rw---- 1 root disk 8, 0 Jan 21 19:08 /dev/sda > brw-rw---- 1 root disk 8, 1 Jan 21 19:08 /dev/sda1 > brw-rw---- 1 root disk 8, 16 Jan 21 19:08 /dev/sdb > brw-rw---- 1 root disk 8, 17 Jan 21 19:08 /dev/sdb1 > brw-rw---- 1 root disk 8, 32 Jan 21 19:08 /dev/sdc > brw-rw---- 1 root disk 8, 48 Jan 21 19:08 /dev/sdd > brw-rw---- 1 root disk 8, 49 Jan 21 19:08 /dev/sdd1 > brw-rw---- 1 root disk 8, 50 Jan 21 19:08 /dev/sdd2 > brw-rw---- 1 root disk 8, 53 Jan 21 19:08 /dev/sdd5 > brw-rw---- 1 root disk 8, 54 Jan 21 19:08 /dev/sdd6 > brw-rw---- 1 root disk 8, 55 Jan 21 19:08 /dev/sdd7 > brw-rw---- 1 root disk 8, 56 Jan 21 19:08 /dev/sdd8 > brw-rw---- 1 root disk 8, 64 Jan 21 19:08 /dev/sde > brw-rw---- 1 root disk 8, 65 Jan 21 19:08 /dev/sde1 > brw-rw---- 1 root disk 8, 80 Jan 21 19:08 /dev/sdf > brw-rw---- 1 root disk 8, 81 Jan 21 19:08 /dev/sdf1 > brw-rw---- 1 root disk 8, 96 Jan 21 19:08 /dev/sdg > brw-rw---- 1 root disk 8, 97 Jan 21 19:08 /dev/sdg1 > > > The devices are not listed in the mdadm.conf, nor were they ever. > Here's everything (except the initial header comments about updating > initramfs and all) from that file: > > # by default (built-in), scan all partitions (/proc/partitions) and all > # containers for MD superblocks. alternatively, specify devices to > scan, using # wildcards if desired. > #DEVICE partitions containers > > # automatically tag new arrays as belonging to the local system > HOMEHOST <system> > > # instruct the monitoring daemon where to send mail alerts > MAILADDR rj > > # definitions of existing MD arrays > #ARRAY /dev/md/0 metadata=1.2 UUID=74a11272:9b233a5b:2506f763:27693ccc > name=jackie:0 > > # This configuration was auto-generated on Wed, 27 Nov 2019 15:53:23 > -0500 by mkconf > UUID=74a11272:9b233a5b:2506f763:27693ccc > > > I assume that last line was added when I added the spare drive. Should > I add the drives to the mdadm.conf then run the assemble command you > suggested? > > It's like mdadm was assembling them automatically upon bootup, but that > stopped working with the new motherboard for some reason. > > Thanks. > --RJ > > > > > > > On Tuesday, January 23, 2024 at 11:06:30 AM EST, David Niklas > <simd@xxxxxxxxxxx> wrote: > > > > > > Hello, > > As someone who's a bit more experienced in RAID array failures, I'd like > to suggest the following: > > # Check that all drives are being detected. > ls /dev/sd* > > # Verify what exactly is being scanned. > grep DEVICE /etc/mdadm/mdadm.conf > > Assuming both of these give satisfactory results*, your next step would > be to try assembling them out of order and see what happens. For > example: > > -> mdadm --assemble /dev/md0 /dev/sda /dev/sdb > Mdadm: Error Not part of array /dev/sdb > -> mdadm --assemble /dev/md0 /dev/sda /dev/sdc > Mdadm: Error too few drives to start array /dev/md0 > > Please note that I made up what mdadm is saying there. But it still > tells you what's going on. > * for the ls command you should see all the drives you have. For the > grep command you should get a listing like "/dev/sda /dev/sdb"... > Obviously, all the drives that might have a RAID array on them should > be listed. > > > Sincerely, > David > > > > > > On Tue, 23 Jan 2024 01:52:31 +0000 (UTC) > RJ Marquette <rjm1@xxxxxxxxx> wrote: > > I meant to add that my /proc/mdstat looked much more like yours on the > > old system. But nothing is showing on this one. > > > > I may try swapping back to the old motherboard. Another possibility > > that might be factor - UEFI vs Legacy BIOS. > > > > Thanks. > > --RJ > > > > > > On Monday, January 22, 2024 at 07:45:29 PM EST, RJ Marquette > > <rjm1@xxxxxxxxx> wrote: > > > > > > > > > > > > That's all. > > > > If I run: > > > > root@jackie:~# mdadm --assemble --scan > > mdadm: /dev/md0 assembled from 0 drives and 1 spare - not enough to > > start the array. > > > > root@jackie:~# cat /proc/mdstat > > Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] > > [raid4] [raid10] unused devices: <none> > > > > root@jackie:~# ls -l /dev/md* > > ls: cannot access '/dev/md*': No such file or directory > > > > It seems to be recognizing the spare drive, but not the 5 that > > actually have data, for some reason. > > > > Thanks. > > --RJ > > > > > > > > > > > > > > > > > > On Monday, January 22, 2024 at 06:49:50 PM EST, Reindl Harald > > <h.reindl@xxxxxxxxxxxxx> wrote: > > > > > > > > > > > > > > > > Am 22.01.24 um 23:13 schrieb RJ Marquette: > > > Sorry! > > > > > > rj@jackie:~$ cat /proc/mdstat > > > Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] > > > [raid4] [raid10] unused devices: <none> > > > > that's all and where is the ton of raid-types coming from with no > > single array shown? > > > > [root@srv-rhsoft:~]$ cat /proc/mdstat > > Personalities : [raid1] > > md0 : active raid1 sdb2[2] sda2[0] > > 30740480 blocks super 1.2 [2/2] [UU] > > bitmap: 0/1 pages [0KB], 65536KB chunk > > > > md1 : active raid1 sda3[0] sdb3[2] > > 3875717120 blocks super 1.2 [2/2] [UU] > > bitmap: 5/29 pages [20KB], 65536KB chunk > > > > > > unused devices: <none> > > > > > On Monday, January 22, 2024 at 04:55:50 PM EST, Reindl Harald > > > <h.reindl@xxxxxxxxxxxxx> wrote: > > > > > > a ton of "mdadm --examine" outputs but i can't see a > > > "cat /proc/mdstat" > > > > > > /dev/sdX is completly irrelevant when it comes to raid - you can > > > even connect a random disk via USB adapter without a change from > > > the view of the array > > > > > > Am 22.01.24 um 20:52 schrieb RJ Marquette: > > >> Hi, all. I have a Raid5 array with 5 disks in use and a 6th in > > >> reserve that I built using 3TB drives in 2019. It has been running > > >> fine since, not even a single drive failure. The system also has a > > >> 7th hard drive for OS, home directory, etc. The motherboard had > > >> four SATA ports, so I added an adapter card that has 4 more ports, > > >> with three drives connected to it. The server runs Debian that I > > >> keep relatively current. > > >> > > >> Yesterday, I swapped a newer motherboard into the computer > > >> (upgraded my desktop and moved the guts to my server). I never > > >> disconnected the cables from the adapter card (whew, I think), so > > >> I know which four drives were connected to the motherboard. > > >> Unfortunately I didn't really note how they were hooked to the > > >> motherboard (SATA1-4 ports). Didn't even think it would be an > > >> issue. I'm reasonably confident the array drives on the > > >> motherboard were sda-sdc, but I'm not certain. > > >> > > >> Now I can't get the array to come up. I'm reasonably certain I > > >> haven't done anything to write to the drives - but mdadm will not > > >> assemble the drives (I have not tried to force it). I'm not > > >> entirely sure what's up and would really appreciate any help. > > >> > > >> I've tried various incantations of mdadm --assemble --scan, with no > > >> luck. I've seen the posts about certain motherboards that can mess > > >> up the drives, and I'm hoping I'm not in that boat. The "new" > > >> motherboard is a Asus Z96-K/CSM. > > >> > > >> I assume using --force is in my future...I see various pages that > > >> say use --force then check it, but will that damage it if I'm > > >> wrong? If not, how will I know it's correct? Is the order of > > >> drives important with --force? I see conflicting info on that. > > >> > > >> I'm no expert but it looks like each drive has the mdadm > > >> superblock...so I'm not sure why it won't assemble. Please help! > > >> > > >> Thanks in advance. > > >> --RJ > > >> > > >> root@jackie:~# uname -a > > >> Linux jackie 5.10.0-27-amd64 #1 SMP Debian 5.10.205-2 (2023-12-31) > > >> x86_64 GNU/Linux > > >> > > >> root@jackie:~# mdadm --version > > >> mdadm - v4.1 - 2018-10-01 > > >> > > >> root@jackie:~# mdadm --examine /dev/sda > > >> /dev/sda: MBR Magic : aa55 > > >> Partition[0] : 4294967295 sectors at 1 (type ee) > > >> > > >> root@jackie:~# mdadm --examine /dev/sda1 > > >> mdadm: No md superblock detected on /dev/sda1. > > >> > > >> root@jackie:~# mdadm --examine /dev/sdb > > >> /dev/sdb: MBR Magic : aa55 > > >> Partition[0] : 4294967295 sectors at 1 (type ee) > > >> > > >> root@jackie:~# mdadm --examine /dev/sdb1 > > >> mdadm: No md superblock detected on /dev/sdb1. > > >> > > >> root@jackie:~# mdadm --examine /dev/sdc > > >> /dev/sdc: Magic : a92b4efc Version : 1.2 > > >> Feature Map : 0x0 > > >> Array UUID : 74a11272:9b233a5b:2506f763:27693ccc > > >> Name : jackie:0 (local to host jackie) > > >> Creation Time : Sat Dec 8 19:32:07 2018 > > >> Raid Level : raid5 > > >> Raid Devices : 5 Avail > > >> Dev Size : 5860271024 (2794.39 GiB 3000.46 GB) > > >> Array Size : 11720540160 (11177.58 GiB 12001.83 GB) > > >> Used Dev Size : 5860270080 (2794.39 GiB 3000.46 GB) > > >> Data Offset : 262144 sectors > > >> Super Offset : 8 sectors > > >> Unused Space : before=261864 sectors, after=944 sectors > > >> State : clean > > >> Device UUID : a2b677bb:4004d8fb:a298a923:bab4df8a > > >> Update Time : Fri Jan 19 15:25:37 2024 > > >> Bad Block Log : 512 entries available at offset 264 sectors > > >> Checksum : 2487f053 - correct > > >> Events : 5958 > > >> Layout : left-symmetric > > >> Chunk Size : 512K > > >> Device Role : spare > > >> Array State : AAAAA ('A' == active, '.' == missing, 'R' == > > >> replacing) > > >> > > >> root@jackie:~# mdadm --examine /dev/sdc1 > > >> mdadm: cannot open /dev/sdc1: No such file or directory > > >> > > >> root@jackie:~# mdadm --examine /dev/sde > > >> /dev/sde: MBR Magic : aa55 > > >> Partition[0] : 4294967295 sectors at 1 (type ee) > > >> > > >> root@jackie:~# mdadm --examine /dev/sde1 > > >> mdadm: No md superblock detected on /dev/sde1. > > >> > > >> root@jackie:~# mdadm --examine /dev/sdf > > >> /dev/sdf: MBR Magic : aa55 > > >> Partition[0] : 4294967295 sectors at 1 (type ee) > > >> > > >> root@jackie:~# mdadm --examine /dev/sdf1 > > >> mdadm: No md superblock detected on /dev/sdf1. > > >> > > >> root@jackie:~# mdadm --examine /dev/sdg > > >> /dev/sdg: MBR Magic : aa55 > > >> Partition[0] : 4294967295 sectors at 1 (type ee) > > >> > > >> root@jackie:~# mdadm --examine /dev/sdg1 > > >> mdadm: No md superblock detected on /dev/sdg1. > > >> > > >> root@jackie:~# lsdrv > > >> PCI [ahci] 00:1f.2 SATA controller: Intel Corporation 9 Series > > >> Chipset Family SATA Controller [AHCI Mode] ├scsi 0:0:0:0 ATA > > >> ST3000VN007-2E41 {Z7317D1A} │└sda 2.73t [8:0] Partitioned > > >> (gpt) │ └sda1 2.73t [8:1] Empty/Unknown > > >> ├scsi 1:0:0:0 ATA Hitachi HUS72403 {P8GSA1WR} > > >> │└sdb 2.73t [8:16] Partitioned (gpt) > > >> │ └sdb1 2.73t [8:17] Empty/Unknown > > >> ├scsi 2:0:0:0 ATA Hitachi HUA72303 {MK0371YVGSZ9RA} > > >> │└sdc 2.73t [8:32] MD raid5 (5) inactive > > >> 'jackie:0' {74a11272-9b23-3a5b-2506-f76327693ccc} └scsi 3:0:0:0 ATA > > >> ST32000542AS {5XW110LY} └sdd 1.82t [8:48] Partitioned > > >> (dos) ├sdd1 23.28g [8:49] Partitioned (dos) > > >> {d94cc2c8-037a-49c5-8a1e-01bb47d78624} │└Mounted as /dev/sdd1 @ / > > >> ├sdd2 1.00k [8:50] Partitioned (dos) > > >> ├sdd5 9.31g [8:53] ext4 {6eb3b4d0-8c7f-4b06-a431-4c292d5bda86} > > >> │└Mounted as /dev/sdd5 @ /var > > >> ├sdd6 3.96g [8:54] swap {901cd56d-ef11-4866-824b-d9ec4ae6fe6e} > > >> ├sdd7 1.86g [8:55] ext4 {69ba0889-322b-4fc8-b9d3-a2d133c97e5e} > > >> │└Mounted as /dev/sdd7 @ /tmp > > >> └sdd8 1.78t [8:56] ext4 {4ed408d4-6b22-46e0-baed-2e0589ff41fb} > > >> └Mounted as /dev/sdd8 @ /home PCI [ahci] > > >> > > >> 06:00.0 SATA controller: Marvell Technology Group Ltd. 88SE9215 > > >> PCIe 2.0 x1 4-port SATA 6 Gb/s Controller (rev 11) ├scsi 6:0:0:0 > > >> ATA Hitachi HUS72403 {P8G84LEP} │└sde 2.73t [8:64] Partitioned > > >> (gpt) │ └sde1 2.73t [8:65] Empty/Unknown > > >> ├scsi 7:0:0:0 ATA ST3000VN007-2E41 {Z7317D46} > > >> │└sdf 2.73t [8:80] Partitioned (gpt) > > >> │ └sdf1 2.73t [8:81] Empty/Unknown > > >> └scsi 8:0:0:0 ATA ST3000VN007-2E41 {Z7317JTX} > > >> └sdg 2.73t [8:96] Partitioned (gpt) > > >> └sdg1 2.73t [8:97] Empty/Unknown > > >> > > >> root@jackie:~# cat /etc/mdadm/mdadm.conf > > >> # This configuration was auto-generated on Wed, 27 Nov 2019 > > >>15:53:23 -0500 by mkconf > > >> ARRAY /dev/md0 metadata=1.2 spares=1 name=jackie:0 > > >> UUID=74a11272:9b233a5b:2506f763:27693cccr > > >