On Mon, Nov 26, 2018 at 10:24:26AM +0100, David Hildenbrand wrote: >On 26.11.18 02:44, Wei Yang wrote: >> On Fri, Nov 23, 2018 at 09:46:52AM +0100, David Hildenbrand wrote: >>> On 23.11.18 09:42, Michal Hocko wrote: >>>> On Thu 22-11-18 16:26:40, David Hildenbrand wrote: >>>>> On 22.11.18 11:12, Wei Yang wrote: >>>>>> During online_pages phase, pgdat->nr_zones will be updated in case this >>>>>> zone is empty. >>>>>> >>>>>> Currently the online_pages phase is protected by the global lock >>>>>> mem_hotplug_begin(), which ensures there is no contention during the >>>>>> update of nr_zones. But this global lock introduces scalability issues. >>>>>> >>>>>> This patch is a preparation for removing the global lock during >>>>>> online_pages phase. Also this patch changes the documentation of >>>>>> node_size_lock to include the protectioin of nr_zones. >>>>> >>>>> I looked into locking recently, and there is more to it. >>>>> >>>>> Please read: >>>>> >>>>> commit dee6da22efac451d361f5224a60be2796d847b51 >>>>> Author: David Hildenbrand <david@xxxxxxxxxx> >>>>> Date: Tue Oct 30 15:10:44 2018 -0700 >>>>> >>>>> memory-hotplug.rst: add some details about locking internals >>>>> >>>>> Let's document the magic a bit, especially why device_hotplug_lock is >>>>> required when adding/removing memory and how it all play together with >>>>> requests to online/offline memory from user space. >>>>> >>>>> Short summary: Onlining/offlining of memory requires the device_hotplug_lock >>>>> as of now. >>>> >>>> Well, I would tend to disagree here. You might be describing the current >>>> state of art but the device_hotplug_lock doesn't make much sense for the >>>> memory hotplug in principle. There is absolutely nothing in the core MM >>> >>> There are collisions with CPU hotplug that require this lock (when nodes >>> come and go as far as I remember). And there is the problematic lock >>> inversion that can happen when adding/remving memory. This all has to be >>> sorted out, we'll have to see if we really need it for >>> onlining/offlining, though, however ... >>> >> >> Seems I get a little understanding on this part. >> >> There are two hotplug: >> * CPU hotplug >> * Memory hotplug. >> >> There are two phase for Memory hotplug: >> * physical add/remove >> * logical online/offline >> >> All of them are protected by device_hotplug_lock now, so we need to be >> careful to release this in any case. Is my understanding correct? > >Yes, e.g. the acpi driver always held the device_hotplug_lock (due to >possible problems with concurrent cpu/memory hot(un)plug). Onlining >offlining of devices (including cpu/memory) from user space always held >the device_hotplug_lock. So this part was executed sequentially for a >long time. > >I recently made sure that any adding/removing/onlining/offlining >correctly grabs the device_hotplug_lock AND the mem_hotplug_lock in any >case (because it was inconsistent and broken), so it is all executed >sequentially. > >So when getting rid of mem_hotplug_lock we only have to care about all >users that don't take the device_hotplug_lock. > Thanks for your sharing. So there are two global locks to protect the hotplug procedure. * device_hotplug_lock * mem_hotplug_lock Seems even removing mem_hotplug_lock doesn't help much on scalability? Maybe we need to move one by one. -- Wei Yang Help you, Help me