On 3/6/23 2:03 PM, Muhammad Usama Anjum wrote: > On 3/3/23 9:47 PM, Peter Xu wrote: >> On Fri, Mar 03, 2023 at 11:42:46AM +0500, Muhammad Usama Anjum wrote: >>> On 3/3/23 3:21 AM, Peter Xu wrote: >>>> On Thu, Mar 02, 2023 at 06:38:20PM +0100, David Hildenbrand wrote: >>>>> On 02.03.23 18:19, Muhammad Usama Anjum wrote: >>>>>> On 2/28/23 5:36 AM, Peter Xu wrote: >>>>>>> On Mon, Feb 27, 2023 at 06:00:44PM -0500, Peter Xu wrote: >>>>>>>> This is a new feature that controls how uffd-wp handles none ptes. When >>>>>>>> it's set, the kernel will handle anonymous memory the same way as file >>>>>>>> memory, by allowing the user to wr-protect unpopulated ptes. >>>>>>>> >>>>>>>> File memories handles none ptes consistently by allowing wr-protecting of >>>>>>>> none ptes because of the unawareness of page cache being exist or not. For >>>>>>>> anonymous it was not as persistent because we used to assume that we don't >>>>>>>> need protections on none ptes or known zero pages. >>>>>>>> >>>>>>>> One use case of such a feature bit was VM live snapshot, where if without >>>>>>>> wr-protecting empty ptes the snapshot can contain random rubbish in the >>>>>>>> holes of the anonymous memory, which can cause misbehave of the guest when >>>>>>>> the guest OS assumes the pages should be all zeros. >>>>>>>> >>>>>>>> QEMU worked it around by pre-populate the section with reads to fill in >>>>>>>> zero page entries before starting the whole snapshot process [1]. >>>>>>>> >>>>>>>> Recently there's another need raised on using userfaultfd wr-protect for >>>>>>>> detecting dirty pages (to replace soft-dirty in some cases) [2]. In that >>>>>>>> case if without being able to wr-protect none ptes by default, the dirty >>>>>>>> info can get lost, since we cannot treat every none pte to be dirty (the >>>>>>>> current design is identify a page dirty based on uffd-wp bit being cleared). >>>>>>>> >>>>>>>> In general, we want to be able to wr-protect empty ptes too even for >>>>>>>> anonymous. >>>>>>>> >>>>>>>> This patch implements UFFD_FEATURE_WP_UNPOPULATED so that it'll make >>>>>>>> uffd-wp handling on none ptes being consistent no matter what the memory >>>>>>>> type is underneath. It doesn't have any impact on file memories so far >>>>>>>> because we already have pte markers taking care of that. So it only >>>>>>>> affects anonymous. >>>>>>>> >>>>>>>> The feature bit is by default off, so the old behavior will be maintained. >>>>>>>> Sometimes it may be wanted because the wr-protect of none ptes will contain >>>>>>>> overheads not only during UFFDIO_WRITEPROTECT (by applying pte markers to >>>>>>>> anonymous), but also on creating the pgtables to store the pte markers. So >>>>>>>> there's potentially less chance of using thp on the first fault for a none >>>>>>>> pmd or larger than a pmd. >>>>>>>> >>>>>>>> The major implementation part is teaching the whole kernel to understand >>>>>>>> pte markers even for anonymously mapped ranges, meanwhile allowing the >>>>>>>> UFFDIO_WRITEPROTECT ioctl to apply pte markers for anonymous too when the >>>>>>>> new feature bit is set. >>>>>>>> >>>>>>>> Note that even if the patch subject starts with mm/uffd, there're a few >>>>>>>> small refactors to major mm path of handling anonymous page faults. But >>>>>>>> they should be straightforward. >>>>>>>> >>>>>>>> So far, add a very light smoke test within the userfaultfd kselftest >>>>>>>> pagemap unit test to make sure anon pte markers work. >>>>>>>> >>>>>>>> [1] https://lore.kernel.org/all/20210401092226.102804-4-andrey.gruzdev@xxxxxxxxxxxxx/ >>>>>>>> [1] https://lore.kernel.org/all/Y+v2HJ8+3i%2FKzDBu@x1n/ >>>>>>>> >>>>>>>> Signed-off-by: Peter Xu <peterx@xxxxxxxxxx> >>>>>>>> --- >>>>>>>> v1->v2: >>>>>>>> - Use pte markers rather than populate zero pages when protect [David] >>>>>>>> - Rename WP_ZEROPAGE to WP_UNPOPULATED [David] >>>>>>> >>>>>>> Some very initial performance numbers (I only ran in a VM but it should be >>>>>>> similar, unit is "us") below as requested. The measurement is about time >>>>>>> spent when wr-protecting 10G range of empty but mapped memory. It's done >>>>>>> in a VM, assuming we'll get similar results on bare metal. >>>>>>> >>>>>>> Four test cases: >>>>>>> >>>>>>> - default UFFDIO_WP >>>>>>> - pre-read the memory, then UFFDIO_WP (what QEMU does right now) >>>>>>> - pre-fault using MADV_POPULATE_READ, then default UFFDIO_WP >>>>>>> - UFFDIO_WP with WP_UNPOPULATED >>>>>>> >>>>>>> Results: >>>>>>> >>>>>>> Test DEFAULT: 2 >>>>>>> Test PRE-READ: 3277099 (pre-fault 3253826) >>>>>>> Test MADVISE: 2250361 (pre-fault 2226310) >>>>>>> Test WP-UNPOPULATE: 20850 >>>>>> In your case: >>>>>> Default < WP-UNPOPULATE < MADVISE < PRE-READ >>>>>> >>>>>> >>>>>> In my testing on next-20230228 with this patch and my uffd async patch: >>>>>> >>>>>> Test DEFAULT: 6 >>>>>> Test PRE-READ: 37157 (pre-fault 37006) >>>>>> Test MADVISE: 4884 (pre-fault 4465) >>>>>> Test WP-UNPOPULATE: 17794 >>>>>> >>>>>> DEFAULT < MADVISE < WP-UNPOPULATE < PRE-READ >>>>>> >>>>>> On my setup, MADVISE is performing better than WP-UNPOPULATE consistently. >>>>>> I'm not sure why I'm getting this discrepancy here. I've liked your results >>>>>> to be honest where we perform better with WP-UNPOPULATE than MADVISE. What >>>>>> can be done to get consistent benchmarks over your and my side? >>>>> >>>>> Probably because the current approach from Peter uses uffd-wp markers, and >>>>> these markers can currently only reside on the PTE level, not on the PMD >>>>> level yet. >>>>> >>>>> With MADVISE you get a huge zeropage and avoid dealing with PTEs. >>>> >>>> Yes, probably. But then when write happens it'll be done there when split, >>>> so the overhead was delayed. >>>> >>>> Meanwhile I'll retest again (probably tomorrow..) with bare metals with THP >>>> on/off to double check. >>> Turning on/off THP has effect. >>> >>> (1) With huge page disabled >>> echo madvise > /sys/kernel/mm/transparent_hugepage/enabled >>> ./uffd_wp_perf >>> Test DEFAULT: 4 >>> Test PRE-READ: 1111453 (pre-fault 1101011) >>> Test MADVISE: 278276 (pre-fault 266378) >>> Test WP-UNPOPULATE: 11712 > Hi Paul, > > We were reading, touching or faulting memory before engaging wp. This > reading is called pre-read in this benchmark. We can fault the memory > through madvise as well, it is called madvise here. WP-Unpopulated is being > added in this patch which will remove the need to pre-fault the memory and > it is way faster as shown by the numbers. > >>> >>> (2) With Huge page enabled >>> echo always > /sys/kernel/mm/transparent_hugepage/enabled >>> ./uffd_wp_perf >>> Test DEFAULT: 4 >>> Test PRE-READ: 22521 (pre-fault 22348) >>> Test MADVISE: 4909 (pre-fault 4743) >>> Test WP-UNPOPULATE: 14448 >> >> Thanks. Yes then this is explaine-able to me. As mentioned previously, thp >> will drastically speedup everything during prefault+ioctl(UFFDIO_WP), >> though that overhead will be applied when the 1st write triggers later, >> afaik. >> >>> >>>> >>>> Muhammad, do you think the current performance will work for you? >>> Yes, the current performance is good enough for now. We'll be using normal >>> pages for memory on which PAGEMAP_IOCTL will be used for emulating Windows >>> syscall. When normal size pages are used, (1) show that WP-UNPOPULATE has >>> really good performance (278276 --> 11712). Lets proceed with it. >> >> OK. >> >>> >>>> >>>> Especially I want to double check with you again on whether >>>> XFS/EXT4/... will be needed for the tracking purpose so you can reply here >>>> together. We shouldn't merge anything that doesn't have at least one >>>> existing good use case, and we may need to rethink if it's not. >>> We have been trying to find soft dirty pages correctly on anon memory. >>> Right now with UN_POPULATED, ASYNC and PAGEMAP_IOCTL, we'll be able to >>> achieve this correctly. >>> >>> We want to track the file mapped memory the same way for another Windows >>> api translation in which file memory will be tracked to find out soft-dirty >>> pages. I believe once it is supported in the mm, filesystem doesn't matter >>> as we'll be tracking memory mapped files. >> >> Some more elaboration would be great here. > There is something called "writecopy" in Windows. Paul has given me some > details. > > @Paul can you share more details that: > 1) What is writecopy? > 2) If there is no documentation or documented behavior, Can you please > share the current reference implementation in Wine even though it is buggy > at the time? (I can implement correct implementation with this patch and my > wip v11 pagemap_ioctl patches for testing purpose.) Hi Peter, Just had some offline discussion with Paul, lets leave this second use case for future. We don't want to make discussions more entangled. There isn't much clarity on it right now. Lets only complete the first use case. The current patch seems good. Please can you post the next version by including the missing part. Then I'll also be able to send the v11 of pagemap_ioctl by basing on it. > >> >> For example, what is the other windows API? What will you do with the >> dirty information? What is the impact of supporting only shmem/hugetlbfs >> for "!anon"? >> >>> >>> So we have use cases for both soft-dirty tracking on anon and file backed >>> memories. >>> >>>> >>>> For performance, one approach is probably making uffd-wp async separate >>>> from other features, where we can revert the meaning of uffd-wp bit to >>>> mimic what soft-dirty does (I think this will look closer to what David >>>> mentioned in the other thread), by defining uffd-wp=1 as "written" and >>>> uffd-wp=0 as clean. >>> I'm happy with current state of things. WP Async is just a feature where >>> faults are resolved automatically. The page is WP or not perfectly >>> translates if page is written (dirty) or not with just not (!) operation. >>> If we go and change the meanings of uffd_wp=1 as written, it would cause >>> confusion and probably we'll need a lot more changes to add async feature >>> which is right now quite small changeset. >> >> Yes, that's also why I wanted to keep it like this unless necessary. It >> makes all uffd features still more aligned with everything so more maintainable. >> >>> >>>> >>>> IIUC that'll make it one bit and work as fast as soft-dirty, meanwhile all >>>> uffd-wp marker things can hopefully still be maintained. However I really >>>> don't like it to violate a lot of things, e.g., when UFFDIO_WRITEPROTECT >>>> another round we'll need to DROP uffd-wp if async, but APPLY uffd-wp if >>>> sync... So in general it'll need more thoughts and slower to do. >>> Sorry, I'm unable to understand your example here. Can you please elaborate? >> >> I think it's mostly what you stated above, the changeset will just get more >> involved because of either the redefinition of uffd-wp depending on feature >> enabled, or using yet another bit (e.g., if you still remember the very >> original reply I was planning to reuse soft-dirty bit but used together >> with VM_UFFD_WP; I think that resolves the degraded perf on unpopulated >> memory too because it'll work more like soft-dirty). >> >> This is some detail which, IMHO, we can leave that for later. To me I'm >> more curious on knowing more details of your use case, and how uffd could >> suite that the best way it could, assuming that (or part of it like the >> unpopulate work here) can also be leveraged by others. Your answers to >> above questions may help. >> >> Thanks, >> > -- BR, Muhammad Usama Anjum