On Fri, Apr 24, 2009 at 12:44:08PM +0200, Jacek Danecki wrote: > Joel Granados wrote: >> On Thu, Apr 23, 2009 at 05:46:02PM -0500, David Lehman wrote: >>> On Thu, 2009-04-23 at 19:36 +0200, Joel Granados Moreno wrote: >>>> --- >>>> storage/devicelibs/mdraid.py | 2 +- >>>> 1 files changed, 1 insertions(+), 1 deletions(-) >>> I am truly amazed that mdadm requires this argument for arrays it is >>> creating, but otherwise it seems fine. >> >> I fail to understand it too. But on all my tests, every time I wanted >> to create a md device (even on tty2) it began to resync the array, >> slowing down the install (it seemed to freeze actually). > > Did you try these commands? Yes this is one of my test cases. But I also did a `dd if=/dev/zero of=/dev/DEVICE bs=1M count=200` to both of the devices that contained the mdraid metadata and this behavior continued. Also I am interested in the --create not the --assemble. So the patch only touches the code that we use to create the md array. The question is. Why, after doing a `mdadm --zero DEVICE` to the devices that compose the mdraid array, is there still metadata in the device? Regards > > # mdadm -C /dev/md0 --run --assume-clean -l 1 -n 2 /dev/sdb /dev/sdc > mdadm: array /dev/md/0 started. > > # cat /proc/mdstat > Personalities : [raid1] [raid10] md0 : active raid1 sdc[1] sdb[0] > 156290816 blocks [2/2] [UU] > unused devices: <none> > > # mdadm -E /dev/sdb > /dev/sdb: > Magic : a92b4efc > Version : 0.90.00 > UUID : c2037f61:6821ab3c:fb7e1339:6a84174c (local to host localhost.localdomain) > Creation Time : Fri Apr 24 12:38:24 2009 > Raid Level : raid1 > Used Dev Size : 156290816 (149.05 GiB 160.04 GB) > Array Size : 156290816 (149.05 GiB 160.04 GB) > Raid Devices : 2 > Total Devices : 2 > Preferred Minor : 0 > > Update Time : Fri Apr 24 12:38:24 2009 > State : clean > Active Devices : 2 > Working Devices : 2 > Failed Devices : 0 > Spare Devices : 0 > Checksum : d68749f2 - correct > Events : 1 > > > Number Major Minor RaidDevice State > this 0 8 16 0 active sync /dev/sdb > > 0 0 8 16 0 active sync /dev/sdb > 1 1 8 32 1 active sync /dev/sdc > > # mdadm -Ss > mdadm: stopped /dev/md0 > # mdadm --zero /dev/sdb /dev/sdc > > > # mdadm -C /dev/md0 --run -l 1 -n 2 /dev/sdb /dev/sdc > mdadm: array /dev/md/0 started. > > # cat /proc/mdstat Personalities : [raid1] [raid10] md0 : active raid1 > sdc[1] sdb[0] > 156290816 blocks [2/2] [UU] > [>....................] resync = 0.2% (317056/156290816) finish=40.9min speed=63411K/sec > unused devices: <none> > > # mdadm -E /dev/sdb > /dev/sdb: > Magic : a92b4efc > Version : 0.90.00 > UUID : d7873388:7a8e6745:fb7e1339:6a84174c (local to host localhost.localdomain) > Creation Time : Fri Apr 24 12:41:41 2009 > Raid Level : raid1 > Used Dev Size : 156290816 (149.05 GiB 160.04 GB) > Array Size : 156290816 (149.05 GiB 160.04 GB) > Raid Devices : 2 > Total Devices : 2 > Preferred Minor : 0 > > Update Time : Fri Apr 24 12:41:41 2009 > State : active > Active Devices : 2 > Working Devices : 2 > Failed Devices : 0 > Spare Devices : 0 > Checksum : fe77bbab - correct > Events : 1 > > > Number Major Minor RaidDevice State > this 0 8 16 0 active sync /dev/sdb > > 0 0 8 16 0 active sync /dev/sdb > 1 1 8 32 1 active sync /dev/sdc > > # mdadm --version > mdadm - v3.0-devel3 - 10th March 2009 > # uname -r > 2.6.29.1-70.fc11.i686.PAE > > _______________________________________________ > Anaconda-devel-list mailing list > Anaconda-devel-list@xxxxxxxxxx > https://www.redhat.com/mailman/listinfo/anaconda-devel-list -- Joel Andres Granados Brno, Czech Republic, Red Hat. _______________________________________________ Anaconda-devel-list mailing list Anaconda-devel-list@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/anaconda-devel-list