> On Apr 28, 2019, at 2:08 PM, Song Liu <liu.song.a23@xxxxxxxxx> wrote: > > We will bring this proposal up in THP discussions. Would you like to share more > thoughts on pros and cons of the two solutions? Or in other words, do you have > strong reasons to dislike either of them? I think it's a performance issue that needs to be hashed out. The obvious thing to do is read the whole large page and then map it, but depending on the architecture or I/O speed, mapping one PAGESIZE page to satisfy the single fault while the large page is being read in could potentially be faster. However, as with all swags without actual data who can say. You can also bring up the question of whether with SSDs and NVME storage if it makes sense to worry anymore about how long it would take to read a 2M or even 1G page in from storage. I like the idea of simply reading the entire large page purely for neatness reasons - recovering from an error during redhead of a large page seems like it could become rather complex. One other issue is how this will interact with filesystems and how and how to tell filesystems I want a large page's worth of data. Matthew mentioned that compound_order() can be used to detect the page size, so that's one answer, but obviously no such code exists as of yet and it would need to be propagated across all file systems. I really hope the discussions at LSFMM are productive. -- Bill