Nick Piggin <npiggin@xxxxxxx> wrote: > Yeah ... the name isn't ideal. _mb is a nice one, but I don't want to use it > unless it is guaranteed to be a full barrier (rather than lock/unlock > barriers). barrier similarly has an existing meaning. How about _iobarrier? Or how about calling them begin_io_section() and end_io_section()? > > Does having 'CACHEABLE' imply that 'FULL' ones are or that they aren't? > > Are or aren't what? Hmmm... Nevermind. I'm not sure what I was thinking of applies anyway. Perhaps you should state that a FULL barrier implies a CACHEABLE barrier and it implies an I/O barrier and it also orders ordinary memory accesses with respect to I/O accesses. Also, I object to 'CACHEABLE' because memory barriers may still be required even if there aren't any caches. > The problem with just calling them mandatory or SMP conditional is that it > doesn't suggest the important *what* is being ordered. Well, it is called 'memory-barriers.txt':-) But I know what you mean, you've expanded the whole scope of the document in a way. > Pretty verbose, isn't it ;) Not only does a CPU appear self consistent, it > is. And not only will overlapping accesses be ordered correctly, so will > completely out of order ones. Sometimes it's a good idea to explicitly state your assumptions, just so that people know. > > > +[!] Note that CACHEABLE or FULL memory barriers must be used to control the > > > +ordering of references to shared memory on SMP systems, though the use of > > > +locking instead is sufficient. > > > > So a spin_lock() or a down() will do the trick? > > I hope so. I didn't write this (just changed slightly). I think this paragraph implies that use of a spin lock render I/O barriers redundant for that particular situation. However your io_lock() implies otherwise. > > > Do not get smart. > > > > I'd very much recommend against saying that. > > Really? What if you grep drivers/*, do you still recommend against? ;( Bad practice elsewhere doesn't condone it here. In fact, you're even more constrained in formal documentation. > I just figure it is vague, unneccesary, hardware specific. Can take it out. You still have to say why. You should check this with Paul McKenney. I think he added the statement you've removed. > > > +XXX: get rid of this, and just prefer io_lock()/io_unlock() ? > > > > What do you mean? > > I mean that although the readl can obviate the need for mmiowb on PCI, > will it be a problem just to encourage the use of the IO locks instead. > Just for simplicity and clarity. So if I have a hardware device with a register that I want to make three reads of, and I expect each read to have side effects, I can just do: io_lock() readl() readl() readl() io_unlock() One thing I'm not entirely clear on. Do you expect two I/O accesses from one CPU remain ordered wrt to each other? Or is the following required: io_lock() readl() io_rmb() readl() io_rmb() readl() io_unlock() when talking to a device? David - To unsubscribe from this list: send the line "unsubscribe linux-arch" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html