Minchan Kim <minchan@xxxxxxxxxx> writes: > On Thu, Dec 07, 2017 at 04:29:37PM -0800, Andrew Morton wrote: >> On Thu, 7 Dec 2017 09:14:26 +0800 "Huang, Ying" <ying.huang@xxxxxxxxx> wrote: >> >> > When the swapin is performed, after getting the swap entry information >> > from the page table, the PTL (page table lock) will be released, then >> > system will go to swap in the swap entry, without any lock held to >> > prevent the swap device from being swapoff. This may cause the race >> > like below, >> > >> > CPU 1 CPU 2 >> > ----- ----- >> > do_swap_page >> > swapin_readahead >> > __read_swap_cache_async >> > swapoff swapcache_prepare >> > p->swap_map = NULL __swap_duplicate >> > p->swap_map[?] /* !!! NULL pointer access */ >> > >> > Because swap off is usually done when system shutdown only, the race >> > may not hit many people in practice. But it is still a race need to >> > be fixed. >> >> swapoff is so rare that it's hard to get motivated about any fix which >> adds overhead to the regular codepaths. > > That was my concern, too when I see this patch. > >> >> Is there something we can do to ensure that all the overhead of this >> fix is placed into the swapoff side? stop_machine() may be a bit >> brutal, but a surprising amount of code uses it. Any other ideas? > > How about this? > > I think It's same approach with old where we uses si->lock everywhere > instead of more fine-grained cluster lock. > > The reason I repeated to reset p->max to zero in the loop is to avoid > using lockdep annotation(maybe, spin_lock_nested(something) to prevent > false positive. > > diff --git a/mm/swapfile.c b/mm/swapfile.c > index 42fe5653814a..9ce007a42bbc 100644 > --- a/mm/swapfile.c > +++ b/mm/swapfile.c > @@ -2644,6 +2644,19 @@ SYSCALL_DEFINE1(swapoff, const char __user *, specialfile) > swap_file = p->swap_file; > old_block_size = p->old_block_size; > p->swap_file = NULL; > + > + if (p->flags & SWP_SOLIDSTATE) { > + unsigned long ci, nr_cluster; > + > + nr_cluster = DIV_ROUND_UP(p->max, SWAPFILE_CLUSTER); > + for (ci = 0; ci < nr_cluster; ci++) { > + struct swap_cluster_info *sci; > + > + sci = lock_cluster(p, ci * SWAPFILE_CLUSTER); > + p->max = 0; > + unlock_cluster(sci); > + } > + } > p->max = 0; > swap_map = p->swap_map; > p->swap_map = NULL; > @@ -3369,10 +3382,10 @@ static int __swap_duplicate(swp_entry_t entry, unsigned char usage) > goto bad_file; > p = swap_info[type]; > offset = swp_offset(entry); > - if (unlikely(offset >= p->max)) > - goto out; > > ci = lock_cluster_or_swap_info(p, offset); > + if (unlikely(offset >= p->max)) > + goto unlock_out; > > count = p->swap_map[offset]; > Sorry, this doesn't work, because lock_cluster_or_swap_info() Need to read p->cluster_info, which may be freed during swapoff too. To reduce the added overhead in regular code path, Maybe we can use SRCU to implement get_swap_device() and put_swap_device()? There is only increment/decrement on CPU local variable in srcu_read_lock/unlock(). Should be acceptable in not so hot swap path? This needs to select CONFIG_SRCU if CONFIG_SWAP is enabled. But I guess that should be acceptable too? Best Regards, Huang, Ying -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>