On Mon, 6 Dec 2010, Jim Schutt wrote: > Hi Sage, > > On Sat, 2010-12-04 at 21:59 -0700, Sage Weil wrote: > > > > > > Also, a possibly related behavior I've noticed is that > > > an 'ls' on a directory where I'm writing files > > > does not return until all the writers are finished. > > > > > > I realize it's likely related to caps, but > > > I'm hoping that can be fixed up somehow? > > > > It depends. If the clients "wrote" that data into the buffer cache and > > it's just taking a long time to flush it out, then things are working as > > intended (given current locking state machine). That can be improved, but > > hasn't been a priority (see #541). If the dd's are still writing and they > > don't stop, something is wrong, either on the mds or kclient. > > > > So here's the results from a couple trials. In the > below results, "do_pdd" is a simple wrapper around > "pdsh -w <clients> dd" that computes aggregate > results. > > In one window I launch the parallel dd commands; > in another window, on a client of my filesystem, > I do the ls. Can you retry it with 'ls -f' and 'ls -al'? What I think is going on is the default ls options also stat files to get a file mode and colorize, and that's going slow. It should be happening on the client, file-by-file. Either there is something broken that's blocking entirely, or it's just slow enough per-file to do that that you don't make it through before it's done. 'ls -f' avoids statting files. If that's fast it means there is no problem with the readdir part of things. 'ls -al' output will tell us if the stats are making slow progress or no progress. Thanks! sage > > The date stamps are my attempt to show the ls > doesn't generate any output until the dd commands > have finished. > > trial one: > > ---- window 1 ---- > $ date; ./do_pdd write 16; date > Mon Dec 6 16:05:18 MST 2010 > > On 64 clients: > dd conv=fdatasync if=/dev/zero of=/mnt/ceph/zero.`hostname -s` bs=4k count=16k > > Elapsed time: 39.39 seconds > Total data: 4294.967 MB (4096 MiB) > Aggregate rate: 109.037 MB/s > Mon Dec 6 16:05:57 MST 2010 > ---- window 2 ---- > $ date;ls /mnt/ceph;date > Mon Dec 6 16:06:22 MST 2010 > zero.an1000 zero.an1010 zero.an1020 zero.an358 zero.an368 zero.an378 zero.an996 > zero.an1001 zero.an1011 zero.an1021 zero.an359 zero.an369 zero.an379 zero.an997 > zero.an1002 zero.an1012 zero.an1022 zero.an360 zero.an370 zero.an380 zero.an998 > zero.an1003 zero.an1013 zero.an1023 zero.an361 zero.an371 zero.an381 zero.an999 > zero.an1004 zero.an1014 zero.an1024 zero.an362 zero.an372 zero.an382 > zero.an1005 zero.an1015 zero.an353 zero.an363 zero.an373 zero.an383 > zero.an1006 zero.an1016 zero.an354 zero.an364 zero.an374 zero.an384 > zero.an1007 zero.an1017 zero.an355 zero.an365 zero.an375 zero.an993 > zero.an1008 zero.an1018 zero.an356 zero.an366 zero.an376 zero.an994 > zero.an1009 zero.an1019 zero.an357 zero.an367 zero.an377 zero.an995 > Mon Dec 6 16:06:46 MST 2010 > > > trial two: > > ---- window 1 ---- > $ date; ./do_pdd write 16; date > Mon Dec 6 16:07:01 MST 2010 > > On 64 clients: > dd conv=fdatasync if=/dev/zero of=/mnt/ceph/zero.`hostname -s` bs=4k count=16k > > Elapsed time: 35.31 seconds > Total data: 4294.967 MB (4096 MiB) > Aggregate rate: 121.636 MB/s > Mon Dec 6 16:07:36 MST 2010 > ---- window 2 ---- > $ date;ls /mnt/ceph;date > Mon Dec 6 16:07:12 MST 2010 > zero.an1000 zero.an1010 zero.an1020 zero.an358 zero.an368 zero.an378 zero.an996 > zero.an1001 zero.an1011 zero.an1021 zero.an359 zero.an369 zero.an379 zero.an997 > zero.an1002 zero.an1012 zero.an1022 zero.an360 zero.an370 zero.an380 zero.an998 > zero.an1003 zero.an1013 zero.an1023 zero.an361 zero.an371 zero.an381 zero.an999 > zero.an1004 zero.an1014 zero.an1024 zero.an362 zero.an372 zero.an382 > zero.an1005 zero.an1015 zero.an353 zero.an363 zero.an373 zero.an383 > zero.an1006 zero.an1016 zero.an354 zero.an364 zero.an374 zero.an384 > zero.an1007 zero.an1017 zero.an355 zero.an365 zero.an375 zero.an993 > zero.an1008 zero.an1018 zero.an356 zero.an366 zero.an376 zero.an994 > zero.an1009 zero.an1019 zero.an357 zero.an367 zero.an377 zero.an995 > Mon Dec 6 16:07:36 MST 2010 > > > Thanks -- Jim > > > > -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html