On 12/23/2017 03:43 AM, Ross Zwisler wrote: > On Fri, Dec 22, 2017 at 08:39:41AM +0530, Anshuman Khandual wrote: >> On 12/14/2017 07:40 AM, Ross Zwisler wrote: >>> ==== Quick Summary ==== >>> >>> Platforms exist today which have multiple types of memory attached to a >>> single CPU. These disparate memory ranges have some characteristics in >>> common, such as CPU cache coherence, but they can have wide ranges of >>> performance both in terms of latency and bandwidth. >> >> Right. >> >>> >>> For example, consider a system that contains persistent memory, standard >>> DDR memory and High Bandwidth Memory (HBM), all attached to the same CPU. >>> There could potentially be an order of magnitude or more difference in >>> performance between the slowest and fastest memory attached to that CPU. >> >> Right. >> >>> >>> With the current Linux code NUMA nodes are CPU-centric, so all the memory >>> attached to a given CPU will be lumped into the same NUMA node. This makes >>> it very difficult for userspace applications to understand the performance >>> of different memory ranges on a given CPU. >> >> Right but that might require fundamental changes to the NUMA representation. >> Plugging those memory as separate NUMA nodes, identify them through sysfs >> and try allocating from it through mbind() seems like a short term solution. >> >> Though if we decide to go in this direction, sysfs interface or something >> similar is required to enumerate memory properties. > > Yep, and this patch series is trying to be the sysfs interface that is > required to the memory properties. :) It's a certainty that we will have > memory-only NUMA nodes, at least on platforms that support ACPI. Supporting > memory-only proximity domains (which Linux turns in to memory-only NUMA nodes) > is explicitly supported with the introduction of the HMAT in ACPI 6.2. Yeah, even on POWER platforms can have memory only NUMA nodes. > > It also turns out that the existing memory management code already deals with > them just fine - you see this with my hmat_examples setup: > > https://github.com/rzwisler/hmat_examples > > Both configurations created by this repo create memory-only NUMA nodes, even > with upstream kernels. My patches don't change that, they just provide a > sysfs representation of the HMAT so users can discover the memory that exists > in the system. Once its a NUMA node everything will work as is from MM interface point of view. But the point is how we export these properties to user space. My only concern is lets not do it in a way which will be locked without first going through NUMA redesign for these new attribute based memory, thats all. > >>> We solve this issue by providing userspace with performance information on >>> individual memory ranges. This performance information is exposed via >>> sysfs: >>> >>> # grep . mem_tgt2/* mem_tgt2/local_init/* 2>/dev/null >>> mem_tgt2/firmware_id:1 >>> mem_tgt2/is_cached:0 >>> mem_tgt2/local_init/read_bw_MBps:40960 >>> mem_tgt2/local_init/read_lat_nsec:50 >>> mem_tgt2/local_init/write_bw_MBps:40960 >>> mem_tgt2/local_init/write_lat_nsec:50 >> >> I might have missed discussions from earlier versions, why we have this >> kind of a "source --> target" model ? We will enlist properties for all >> possible "source --> target" on the system ? Right now it shows only >> bandwidth and latency properties, can it accommodate other properties >> as well in future ? > > The initiator/target model is useful in preventing us from needing a > MAX_NUMA_NODES x MAX_NUMA_NODES sized table for each performance attribute. I > talked about it a little more here: That makes it even more complex. Not only we have a memory attribute like bandwidth specific to the range, we are also exporting it's relative values as seen from different CPU nodes. Its again kind of a NUMA distance table being exported in the generic sysfs path like /sys/devices/. The problem is possible future memory attributes like 'reliability', 'density', 'power consumption' might not have a need for a "source --> destination" kind of model as they dont change based on which CPU node is accessing it. > > https://lists.01.org/pipermail/linux-nvdimm/2017-December/013654.html > >>> This allows applications to easily find the memory that they want to use. >>> We expect that the existing NUMA APIs will be enhanced to use this new >>> information so that applications can continue to use them to select their >>> desired memory. >> >> I had presented a proposal for NUMA redesign in the Plumbers Conference this >> year where various memory devices with different kind of memory attributes >> can be represented in the kernel and be used explicitly from the user space. >> Here is the link to the proposal if you feel interested. The proposal is >> very intrusive and also I dont have a RFC for it yet for discussion here. >> >> https://linuxplumbersconf.org/2017/ocw//system/presentations/4656/original/Hierarchical_NUMA_Design_Plumbers_2017.pdf > > I'll take a look, but my first reaction is that I agree with Dave that it > seems hard to re-teach systems a new NUMA scheme. This patch series doesn't > attempt to do that - it is very unintrusive and only informs users about the > memory-only NUMA nodes that will already exist in their ACPI-based systems. But while not trying to address how NUMA should treat these heterogeneous memory attribute nodes, the patch series might be trying to lock down sysfs interfaces which might not be as appropriate or extensible when redesigning NUMA in future. Sure, pass on the information to user space but not through generic interfaces like /sys/device. -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>