On Thu, Nov 03, 2011 at 06:24:58PM +0400, Dmitry Monakhov wrote: > During stress testing we want to cover as much code paths as possible > fsstress is very good for this purpose. But it has expandable nature > (disk usage almost continually grow). So once it goes in no ENOSPC > condition it will be where till the end. But by running 'dd' writers > in parallel we can regularly trigger ENOSPC but only for a limited > periods of time because each time it opens the same file with O_TRUNC. ..... So you have a 512MB filesystem, and you do: > +# Disable all sync operations to get higher load > +FSSTRESS_AVOID="$FSSTRESS_AVOID -ffsync=0 -fsync=0 -ffdatasync=0" > +_workout() > +{ > + echo "" > + echo "Run fsstress" > + echo "" > + num_iterations=10 > + enospc_time=2 > + out=$SCRATCH_MNT/fsstress.$$ > + args="-p128 -n999999999 -f setattr=1 $FSSTRESS_AVOID -d $out" > + echo "fsstress $args" >> $here/$seq.full > + $FSSTRESS_PROG $args > /dev/null 2>&1 & run a bunch of fsstress processes > + pid=$! > + echo "Run dd writers in parallel" > + for ((i=0; i < num_iterations; i++)) > + do > + # File will be opened with O_TRUNC each time > + dd if=/dev/zero of=$SCRATCH_MNT/SPACE_CONSUMER bs=1M count=1 \ > + > /dev/null 2>&1 > + sleep $enospc_time > + done Then write the same 1MB file 10 times, 2 seconds apart, giving a total space usage of the dd processes of 1MB over 20s. > + kill $pid > + wait $pid > +} Then kill the fsstress. AFAICT, fsstress won't always fill 511MB in 20s - on my test systems the fill rate is typically around 5s per 100MB, which would result in the filesystem not being filled with this test and hence not exercising ENOSPC. Perhaps this would be better done like test 083, which uses a fixed number of write-only operations per fsstress process that is known to end up at ENOSPC, rather than hoping it gets there in 20s. Cheers, Dave. -- Dave Chinner david@xxxxxxxxxxxxx -- To unsubscribe from this list: send the line "unsubscribe linux-fsdevel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html