Hello, On 08/18/2010 09:30 PM, Vladislav Bolkhovitin wrote: > Basically, I measured how iSCSI link utilization depends from amount > of queued commands and queued data size. This is why I made it as a > table. From it you can see which improvement you will have removing > queue draining after 1, 2, 4, etc. commands depending of commands > sizes. > > For instance, on my previous XFS rm example, where rm of 4 files > took 3.5 minutes with nobarrier option, I could see that XFS was > sending 1-3 32K commands in a row. From my table you can see that if > it sent all them at once without draining, it would have about > 150-200% speed increase. You compared barrier off/on. Of course, it will make a big difference. I think good part of that gain should be realized by the currently proposed patchset which removes draining. What's needed to be demonstrated is the difference between ordered-by-waiting and ordered-by-tag. We've never had code to do that properly. The original ordered-by-tag we had only applied tag ordering to two or three command sequences inside a barrier, which doesn't amount to much (and could even be harmful as it imposes draining of all simple commands inside the device only to reduce issue latencies for a few commands). You'll need to hook into filesystem and somehow export the ordering information down to the driver so that whatever needs ordering is sent out as ordered commands. As I've wrote multiple times, I'm pretty skeptical it will bring much. Ordered tag mandates draining inside the device just like the original barrier implementation. Sure, it's done at a lower layer and command issue latencies will be reduced thanks to that but ordered-by-waiting doesn't require _any_ draining at all. The whole pipeline can be kept full all the time. I'm often wrong tho, so please feel free to go ahead and prove me wrong. :-) Thanks. -- tejun -- dm-devel mailing list dm-devel@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/dm-devel