Justin, This is a Sun X4500 (Thumper) box, so it's got 48 drives inside. /dev/sd[a-z] are all there as well, just in other RAID sets. Once you get to /dev/sdz, it starts up at /dev/sdaa, sdab, etc. I'd be curious if what I'm experiencing is a bug. What should I try to restore the array? Norman On 2/19/08, Justin Piszcz <jpiszcz@xxxxxxxxxxxxxxx> wrote: > Neil, > > Is this a bug? > > Also, I have a question for Norman-- how come your drives are sda[a-z]1? > Typically it is /dev/sda1 /dev/sdb1 etc? > > Justin. > > On Tue, 19 Feb 2008, Norman Elton wrote: > > > But why do two show up as "removed"?? I would expect /dev/sdal1 to show up > > someplace, either active or failed. > > > > Any ideas? > > > > Thanks, > > > > Norman > > > > > > > > On Feb 19, 2008, at 12:31 PM, Justin Piszcz wrote: > > > >> How many drives actually failed? > >>> Failed Devices : 1 > >> > >> > >> On Tue, 19 Feb 2008, Norman Elton wrote: > >> > >>> So I had my first "failure" today, when I got a report that one drive > >>> (/dev/sdam) failed. I've attached the output of "mdadm --detail". It > >>> appears that two drives are listed as "removed", but the array is > >>> still functioning. What does this mean? How many drives actually > >>> failed? > >>> > >>> This is all a test system, so I can dink around as much as necessary. > >>> Thanks for any advice! > >>> > >>> Norman Elton > >>> > >>> ====== OUTPUT OF MDADM ===== > >>> > >>> Version : 00.90.03 > >>> Creation Time : Fri Jan 18 13:17:33 2008 > >>> Raid Level : raid5 > >>> Array Size : 6837319552 (6520.58 GiB 7001.42 GB) > >>> Device Size : 976759936 (931.51 GiB 1000.20 GB) > >>> Raid Devices : 8 > >>> Total Devices : 7 > >>> Preferred Minor : 4 > >>> Persistence : Superblock is persistent > >>> > >>> Update Time : Mon Feb 18 11:49:13 2008 > >>> State : clean, degraded > >>> Active Devices : 6 > >>> Working Devices : 6 > >>> Failed Devices : 1 > >>> Spare Devices : 0 > >>> > >>> Layout : left-symmetric > >>> Chunk Size : 64K > >>> > >>> UUID : b16bdcaf:a20192fb:39c74cb8:e5e60b20 > >>> Events : 0.110 > >>> > >>> Number Major Minor RaidDevice State > >>> 0 66 1 0 active sync /dev/sdag1 > >>> 1 66 17 1 active sync /dev/sdah1 > >>> 2 66 33 2 active sync /dev/sdai1 > >>> 3 66 49 3 active sync /dev/sdaj1 > >>> 4 66 65 4 active sync /dev/sdak1 > >>> 5 0 0 5 removed > >>> 6 0 0 6 removed > >>> 7 66 113 7 active sync /dev/sdan1 > >>> > >>> 8 66 97 - faulty spare /dev/sdam1 > >>> - > >>> To unsubscribe from this list: send the line "unsubscribe linux-raid" in > >>> the body of a message to majordomo@xxxxxxxxxxxxxxx > >>> More majordomo info at http://vger.kernel.org/majordomo-info.html > > > - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html