Hi all. On a SLES 10 SP 3 x86_64 version I'm encountering the following problem: Two disks sda and sdb are in a RAID 1 configuration. When pulling one disk and replacing it it becomes sdc instead of sda. The RAID is now degraded: kblhbe101:~ # cat /proc/mdstat Personalities : [raid1] [raid0] [raid5] [raid4] [linear] md2 : active raid1 sda8[2](F) sdb8[1] 286125568 blocks [2/1] [_U] md0 : active raid1 sda5[2](F) sdb5[1] 529984 blocks [2/1] [_U] md3 : active raid1 sda7[2](F) sdb7[1] 4200896 blocks [2/1] [_U] md1 : active raid1 sda6[2](F) sdb6[1] 2104384 blocks [2/1] [_U] unused devices: <none> But now I can't remove the sda partitions from the RAID: kblhbe101:~ # mdadm /dev/md0 -r /dev/sda5 mdadm: cannot find /dev/sda5: No such file or directory kblhbe101:~ # mdadm /dev/md0 -r sda5 mdadm: cannot find sda5: No such file or directory What am I doing wrong here? kblhbe101:~ # mdadm --version mdadm - v2.6 - 21 December 2006 Trying mdadm -Iq /dev/md0 results in a segfault... Kind regards, Benjamin
Attachment:
signature.asc
Description: Digital signature