On Sunday July 1, infomails@xxxxxxxxxxxxxxxxxxxxx wrote: > Hello RAID-Experts, > > I have three RAID5 consisting of different partitions on 3 disks (Debian > stable) running the root-filesystem on a md (/boot is a separate non-raid > partition) which is running rather nicely. For convenience I plugged all > drives into the first ide controller making them hda, hdb and hdc. So far, so > good. The partitions are flagged "fd", i.e. Linux raid autodetect. > > As I have another builtin-ide-controller onboard, I'd like to distribute the > disks for performance reasons, moving hdb to hde and hdc to hdg. The arrays > would then consist of drives hda, hde and hdg. > > This should not be a problem, as the arrays should assemble themselves using > the superblocks on the partitions, shouldn't it? > > However, when I switch one drive (hdc), the array starts degraded with two > drives present because it is still looking for hdc, which of course now is > hdg. This shouldn't be happening. > > Well, then I re-added hdg to the degraded array, which went well and the array > rebuilded itself. I now had healthy arrays consisting of hda, hdb and hdg. > But after a reboot the array was degraded again and the system wanted its hdc > drive. > > And yes, I edited /boot/grub/device.map and changed hdc to hdg, so that can't > be the reason. > > I seem to be missing something here, but what is it? Kernel logs from the boot would help here. Maybe /etc/mdadm/mdadm.conf lists "device=...." where it shouldn't. Maybe the other IDE controller uses a module that it loaded late. Logs would help. NeilBrown < - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html