On Tuesday February 19, dgc@xxxxxxx wrote: > On Mon, Feb 18, 2008 at 04:24:27PM +0300, Michael Tokarev wrote: > > First, I still don't understand why in God's sake barriers are "working" > > while regular cache flushes are not. Almost no consumer-grade hard drive > > supports write barriers, but they all support regular cache flushes, and > > the latter should be enough (while not the most speed-optimal) to ensure > > data safety. Why to require write cache disable (like in XFS FAQ) instead > > of going the flush-cache-when-appropriate (as opposed to write-barrier- > > when-appropriate) way? > > Devil's advocate: > > Why should we need to support multiple different block layer APIs > to do the same thing? Surely any hardware that doesn't support barrier > operations can emulate them with cache flushes when they receive a > barrier I/O from the filesystem.... The simple answer to "why multiple APIs" is "different performance trade-offs". If barriers are implemented in at the end of the pipeline, they can presumably be reasonably cheap. If they have to be implemented at the top of the pipeline, thus stalling the whole pipeline, they are likely to be more expensive. A filesystem may be able to mitigate the expense if it knows something about the purpose of the data. e.g. ext3 in data=writeback mode could wait only for journal writes to complete before submitting the (would-be) barrier write of the commit block, and would not bother to wait for data writes. However, consistent APIs are also a good thing. I would easily accept an argument that a BIO_RW_BARRER request must *always* be correctly ordered around all other requests to the same device. If a layered device cannot get the service it requires from lower level devices, it must do that flush/write/wait itself. That should be paired with a way for the upper levels to find out how efficient barriers are. I guess the three levels of barrier efficiency are: 1/ handled above the elevator - least efficient 2/ handled between elevator and device (by 'flush request'), medium 3/ handled inside device (e.g. ordered SCSI request) most efficient. NeilBrown -- dm-devel mailing list dm-devel@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/dm-devel