Locking bugs in 2.4 md.c

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Neil,

I can pretty reliably Oops the 2.4 md raid1 in md_update_sb with a

while true
	mdadm /dev/md0 -f /dev/foo -r /dev/foo
	mdadm /dev/md0 -a /dev/foo
done &
while true
	mdadm /dev/md1 -f /dev/bar -r /dev/bar
	mdadm /dev/md1 -a /dev/bar
done &

in parallel on several md devices. It will finally die in md_update_sb,
and appears to be related to some locking bugs.

(The above is a stress test and makes it occur faster; customers report
it happens in the field too, that's why I went looking.)

In March (see http://www.spinics.net/lists/raid/msg02335.html) you wrote
you had a patch which made the locking in 2.4 "better", even though it
was rather ugly. Well, ugly it might be, but needed still ;-)

I assume you want to use the reconfig_sem for it? If you don't have a
recent patch, could you sketch out what you wanted to do so I could
start from there?

Thanks!


Sincerely,
    Lars Marowsky-Brée <lmb@suse.de>

(Or maybe you want to backport the 2.6 md to 2.4... ;-)

-- 
High Availability & Clustering		ever tried. ever failed. no matter.
SuSE Labs				try again. fail again. fail better.
Research & Development, SuSE Linux AG		-- Samuel Beckett

-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux