On Tue, Nov 27, 2018 at 1:44 PM Logan Gunthorpe <logang@xxxxxxxxxxxx> wrote: > > Hey Dan, > > On 2018-11-20 4:13 p.m., Dan Williams wrote: > > The last step before devm_memremap_pages() returns success is to > > allocate a release action, devm_memremap_pages_release(), to tear the > > entire setup down. However, the result from devm_add_action() is not > > checked. > > > > Checking the error from devm_add_action() is not enough. The api > > currently relies on the fact that the percpu_ref it is using is killed > > by the time the devm_memremap_pages_release() is run. Rather than > > continue this awkward situation, offload the responsibility of killing > > the percpu_ref to devm_memremap_pages_release() directly. This allows > > devm_memremap_pages() to do the right thing relative to init failures > > and shutdown. > > > > Without this change we could fail to register the teardown of > > devm_memremap_pages(). The likelihood of hitting this failure is tiny as > > small memory allocations almost always succeed. However, the impact of > > the failure is large given any future reconfiguration, or > > disable/enable, of an nvdimm namespace will fail forever as subsequent > > calls to devm_memremap_pages() will fail to setup the pgmap_radix since > > there will be stale entries for the physical address range. > > > > An argument could be made to require that the ->kill() operation be set > > in the @pgmap arg rather than passed in separately. However, it helps > > code readability, tracking the lifetime of a given instance, to be able > > to grep the kill routine directly at the devm_memremap_pages() call > > site. > > > > Cc: <stable@xxxxxxxxxxxxxxx> > > Fixes: e8d513483300 ("memremap: change devm_memremap_pages interface...") > > Reviewed-by: "Jérôme Glisse" <jglisse@xxxxxxxxxx> > > Reported-by: Logan Gunthorpe <logang@xxxxxxxxxxxx> > > Reviewed-by: Logan Gunthorpe <logang@xxxxxxxxxxxx> > > Reviewed-by: Christoph Hellwig <hch@xxxxxx> > > Signed-off-by: Dan Williams <dan.j.williams@xxxxxxxxx> > > I recently realized this patch, which was recently added to the mm tree, > will break p2pdma. This is largely because the patch was written and > reviewed before p2pdma was merged (in 4.20). Originally, I think we both > expected this patch would be merged before p2pdma but that's not what > happened. Indeed, sorry I missed this. > > Also, while testing this, I found the teardown is still not quite > correct. In p2pdma, the struct pages will be removed before all of the > percpu references have released and if the device is unbound while pages > are in use, there will be a kernel panic. This is because we wait on the > completion that indicates all references have been free'd after > devm_memremap_pages_release() is called and the pages are removed. This > is fairly easily fixed by waiting for the completion in the kill > function and moving the call after the last put_page(). I suspect device > DAX also has this problem but I'm not entirely certain if something else > might be preventing us from hitting this bug. > > Ideally, as part of this patch we need to update the p2pdma call site > for devm_memremap_pages() and fix the completion issue. The diff for all > this is below, but if you'd like I can send a proper patch. Yes, please send a proper patch. Although, I'm still not sure I see the problem with the order of the percpu-ref kill. It's likely more efficient to put the kill after the put_page() loop because the percpu-ref will still be in "fast" per-cpu mode, but the kernel panic should not be possible as long as their is a wait_for_completion() before the exit, unless something else is wrong. Certainly you can't move the wait_for_completion() into your ->kill() callback without switching the ordering, but I'm not on board with that change until I understand a bit more about why you think device-dax might be broken? I took a look at the p2pdma shutdown path and the: if (percpu_ref_is_dying(ref)) return; ...looks fishy. If multiple agents can overlap their requests for the same range why not track that simply as additional refs? Could it be the crash that you are seeing is a result of mis-accounting when it is safe to assume the page allocation can be freed?