Re: [RFC] arm64: mm: update max_pfn after memory hotplug

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



+Anshuman

Thanks for the responses. Do we see any issues with having this patch go for merge? I agree with David that 'max_pfn' is just a hint about maximum possible PFN that can go. So it doesn't need updating when memory is hotunplugged out.

Also regarding comment about possible reserved memory region with "no-map" attribute - I'm not sure if we such reserved memory allocations are allowed post early init when memory blocks are hot added later on after arm64_memblock_init is done? These are all done during setup_arch right?

We found this issue/bug where kernel inits with limited boot memory using 'mem=' and a driver hot-adds memory blocks during module init using add_memory_driver_managed and owns it. This makes 'max_pfn' out of sync and breaks /proc/kpageXXX kpageflags_read() functionality where any processes using pages beyond this outdated 'max_pfn', stable_page_flags() would simply return KPF_NOPAGE.

I think this is generic bug that exists with arm64 memory hotplug where any consumers of 'max_pfn' would break its functionality after memory blocks are hotplugged in after init. If this patch looks OK and be ACKed, we can have it in GKI. Other suggestions on proper fix or feedback are also welcome.

On 9/24/2021 1:17 AM, David Hildenbrand wrote:
On 24.09.21 04:47, Florian Fainelli wrote:


On 9/23/2021 3:54 PM, Chris Goldsworthy wrote:
From: Sudarshan Rajagopalan <quic_sudaraja@xxxxxxxxxxx>

After new memory blocks have been hotplugged, max_pfn and max_low_pfn
needs updating to reflect on new PFNs being hot added to system.

Signed-off-by: Sudarshan Rajagopalan <quic_sudaraja@xxxxxxxxxxx>
Signed-off-by: Chris Goldsworthy <quic_cgoldswo@xxxxxxxxxxx>
---
   arch/arm64/mm/mmu.c | 5 +++++
   1 file changed, 5 insertions(+)

diff --git a/arch/arm64/mm/mmu.c b/arch/arm64/mm/mmu.c
index cfd9deb..fd85b51 100644
--- a/arch/arm64/mm/mmu.c
+++ b/arch/arm64/mm/mmu.c
@@ -1499,6 +1499,11 @@ int arch_add_memory(int nid, u64 start, u64 size,
       if (ret)
           __remove_pgd_mapping(swapper_pg_dir,
                        __phys_to_virt(start), size);
+    else {
+        max_pfn = PFN_UP(start + size);
+        max_low_pfn = max_pfn;
+    }

This is a drive by review, but it got me thinking about your changes a bit:

- if you raise max_pfn when you hotplug memory, don't you need to lower
it when you hot unplug memory as well?

The issue with lowering is that you actually have to do some search to figure out the actual value -- and it's not really worth the trouble. Raising the limit is easy.

With memory hotunplug, anybody wanting to take a look at a "struct page" via a pfn has to do a pfn_to_online_page() either way. That will fail if there isn't actually a memmap anymore because the memory has been unplugged. So "max_pfn" is actually rather a hint what maximum pfn to look at, and it can be bigger than it actually is.

The a look at the example usage in fs/proc/page.c:kpageflags_read()

pfn_to_online_page() will simply fail and stable_page_flags() will indicate a KPF_NOPAGE.

Just like we would have a big memory hole now at the end of memory.


- suppose that you have a platform which maps physical memory into the
CPU's address space at 0x00_4000_0000 (1GB offset) and the kernel boots
with 2GB of DRAM plugged by default. At that point we have not
registered a swiotlb because we have less than 4GB of addressable
physical memory, there is no IOMMU in that system, it's a happy world.
Now assume that we plug an additional 2GB of DRAM into that system
adjacent to the previous 2GB, from 0x00_C0000_0000 through
0x14_0000_0000, now we have physical addresses above 4GB, but we still
don't have a swiotlb, some of our DMA_BIT_MASK(32) peripherals are going
to be unable to DMA from that hot plugged memory, but they could if we
had a swiotlb.

That's why platforms that hotplug memory should indicate the maximum possible PFN via some mechanism during boot. On x86-64 (and IIRC also arm64 now), this is done via the ACPI SRAT.

And that's where "max_possible_pfn" and "max_pfn" differ. See drivers/acpi/numa/srat.c:acpi_numa_memory_affinity_init():

    max_possible_pfn = max(max_possible_pfn, PFN_UP(end - 1));$


Using max_possible_pfn, the OS can properly setup the swiotlb, even thought it wouldn't currently be required when just looking at max_pfn.

I documented that for virtio-mem in
    https://virtio-mem.gitlab.io/user-guide/user-guide-linux.html
"swiotlb and DMA memory".


- now let's go even further but this is very contrived. Assume that the
firmware has somewhat created a reserved memory region with a 'no-map'
attribute thus indicating it does not want a struct page to be created
for a specific PFN range, is it valid to "blindly" raise max_pfn if that
region were to be at the end of the just hot-plugged memory?

no-map means that no direct mapping is to be created, right? We would still have a memmap IIRC, and the pages are PG_reserved.

Again, I think this is very similar to just having no-map regions like random memory holes within the existing memory layout.


What Chris proposes here is very similar to arch/x86/mm/init_64.c:update_end_of_memory_vars() called during arch_add_memory()->add_pages() on x86-64.





[Index of Archives]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux