Re: [linux-next:master 5756/5946] mm/userfaultfd.c:212:6: warning: variable 'vm_alloc_shared' set but not used

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Jun 1, 2021 at 5:40 PM kernel test robot <lkp@xxxxxxxxx> wrote:
>
> tree:   https://git.kernel.org/pub/scm/linux/kernel/git/next/linux-next.git master
> head:   392d24c0d06bc89a762ba66977db41e53c21bfb5
> commit: 1786d001262006df52cdcda4cbc0c8087a0200ec [5756/5946] mm, hugetlb: fix racy resv_huge_pages underflow on UFFDIO_COPY
> config: powerpc64-randconfig-r022-20210601 (attached as .config)
> compiler: powerpc-linux-gcc (GCC) 9.3.0
> reproduce (this is a W=1 build):
>         wget https://raw.githubusercontent.com/intel/lkp-tests/master/sbin/make.cross -O ~/bin/make.cross
>         chmod +x ~/bin/make.cross
>         # https://git.kernel.org/pub/scm/linux/kernel/git/next/linux-next.git/commit/?id=1786d001262006df52cdcda4cbc0c8087a0200ec
>         git remote add linux-next https://git.kernel.org/pub/scm/linux/kernel/git/next/linux-next.git
>         git fetch --no-tags linux-next master
>         git checkout 1786d001262006df52cdcda4cbc0c8087a0200ec
>         # save the attached .config to linux build tree
>         COMPILER_INSTALL_PATH=$HOME/0day COMPILER=gcc-9.3.0 make.cross ARCH=powerpc64
>
> If you fix the issue, kindly add following tag as appropriate
> Reported-by: kernel test robot <lkp@xxxxxxxxx>
>
> All warnings (new ones prefixed by >>):
>
>    mm/userfaultfd.c: In function '__mcopy_atomic_hugetlb':
> >> mm/userfaultfd.c:212:6: warning: variable 'vm_alloc_shared' set but not used [-Wunused-but-set-variable]
>      212 |  int vm_alloc_shared = dst_vma->vm_flags & VM_SHARED;
>          |      ^~~~~~~~~~~~~~~
>

Looks like a false warning. vm_alloc_shared is set here within the
same function.

mutex_unlock(&hugetlb_fault_mutex_table[hash]);
i_mmap_unlock_read(mapping);
vm_alloc_shared = vm_shared;

> vim +/vm_alloc_shared +212 mm/userfaultfd.c
>
> c1a4de99fada21 Andrea Arcangeli  2015-09-04  199
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  200  #ifdef CONFIG_HUGETLB_PAGE
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  201  /*
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  202   * __mcopy_atomic processing for HUGETLB vmas.  Note that this routine is
> c1e8d7c6a7a682 Michel Lespinasse 2020-06-08  203   * called with mmap_lock held, it will release mmap_lock before returning.
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  204   */
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  205  static __always_inline ssize_t __mcopy_atomic_hugetlb(struct mm_struct *dst_mm,
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  206                                              struct vm_area_struct *dst_vma,
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  207                                              unsigned long dst_start,
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  208                                              unsigned long src_start,
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  209                                              unsigned long len,
> f619147104c8ea Axel Rasmussen    2021-05-04  210                                              enum mcopy_atomic_mode mode)
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  211  {
> 1c9e8def43a345 Mike Kravetz      2017-02-22 @212        int vm_alloc_shared = dst_vma->vm_flags & VM_SHARED;
> 1c9e8def43a345 Mike Kravetz      2017-02-22  213        int vm_shared = dst_vma->vm_flags & VM_SHARED;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  214        ssize_t err;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  215        pte_t *dst_pte;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  216        unsigned long src_addr, dst_addr;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  217        long copied;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  218        struct page *page;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  219        unsigned long vma_hpagesize;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  220        pgoff_t idx;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  221        u32 hash;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  222        struct address_space *mapping;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  223
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  224        /*
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  225         * There is no default zero huge page for all huge page sizes as
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  226         * supported by hugetlb.  A PMD_SIZE huge pages may exist as used
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  227         * by THP.  Since we can not reliably insert a zero page, this
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  228         * feature is not supported.
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  229         */
> f619147104c8ea Axel Rasmussen    2021-05-04  230        if (mode == MCOPY_ATOMIC_ZEROPAGE) {
> d8ed45c5dcd455 Michel Lespinasse 2020-06-08  231                mmap_read_unlock(dst_mm);
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  232                return -EINVAL;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  233        }
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  234
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  235        src_addr = src_start;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  236        dst_addr = dst_start;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  237        copied = 0;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  238        page = NULL;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  239        vma_hpagesize = vma_kernel_pagesize(dst_vma);
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  240
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  241        /*
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  242         * Validate alignment based on huge page size
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  243         */
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  244        err = -EINVAL;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  245        if (dst_start & (vma_hpagesize - 1) || len & (vma_hpagesize - 1))
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  246                goto out_unlock;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  247
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  248  retry:
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  249        /*
> c1e8d7c6a7a682 Michel Lespinasse 2020-06-08  250         * On routine entry dst_vma is set.  If we had to drop mmap_lock and
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  251         * retry, dst_vma will be set to NULL and we must lookup again.
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  252         */
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  253        if (!dst_vma) {
> 27d02568f529e9 Mike Rapoport     2017-02-24  254                err = -ENOENT;
> 643aa36eadebdc Wei Yang          2019-11-30  255                dst_vma = find_dst_vma(dst_mm, dst_start, len);
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  256                if (!dst_vma || !is_vm_hugetlb_page(dst_vma))
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  257                        goto out_unlock;
> 1c9e8def43a345 Mike Kravetz      2017-02-22  258
> 27d02568f529e9 Mike Rapoport     2017-02-24  259                err = -EINVAL;
> 27d02568f529e9 Mike Rapoport     2017-02-24  260                if (vma_hpagesize != vma_kernel_pagesize(dst_vma))
> 27d02568f529e9 Mike Rapoport     2017-02-24  261                        goto out_unlock;
> 27d02568f529e9 Mike Rapoport     2017-02-24  262
> 1c9e8def43a345 Mike Kravetz      2017-02-22  263                vm_shared = dst_vma->vm_flags & VM_SHARED;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  264        }
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  265
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  266        /*
> 1c9e8def43a345 Mike Kravetz      2017-02-22  267         * If not shared, ensure the dst_vma has a anon_vma.
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  268         */
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  269        err = -ENOMEM;
> 1c9e8def43a345 Mike Kravetz      2017-02-22  270        if (!vm_shared) {
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  271                if (unlikely(anon_vma_prepare(dst_vma)))
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  272                        goto out_unlock;
> 1c9e8def43a345 Mike Kravetz      2017-02-22  273        }
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  274
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  275        while (src_addr < src_start + len) {
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  276                BUG_ON(dst_addr >= dst_start + len);
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  277
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  278                /*
> c0d0381ade7988 Mike Kravetz      2020-04-01  279                 * Serialize via i_mmap_rwsem and hugetlb_fault_mutex.
> c0d0381ade7988 Mike Kravetz      2020-04-01  280                 * i_mmap_rwsem ensures the dst_pte remains valid even
> c0d0381ade7988 Mike Kravetz      2020-04-01  281                 * in the case of shared pmds.  fault mutex prevents
> c0d0381ade7988 Mike Kravetz      2020-04-01  282                 * races with other faulting threads.
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  283                 */
> ddeaab32a89f04 Mike Kravetz      2019-01-08  284                mapping = dst_vma->vm_file->f_mapping;
> c0d0381ade7988 Mike Kravetz      2020-04-01  285                i_mmap_lock_read(mapping);
> c0d0381ade7988 Mike Kravetz      2020-04-01  286                idx = linear_page_index(dst_vma, dst_addr);
> 188b04a7d93860 Wei Yang          2019-11-30  287                hash = hugetlb_fault_mutex_hash(mapping, idx);
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  288                mutex_lock(&hugetlb_fault_mutex_table[hash]);
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  289
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  290                err = -ENOMEM;
> aec44e0f0213e3 Peter Xu          2021-05-04  291                dst_pte = huge_pte_alloc(dst_mm, dst_vma, dst_addr, vma_hpagesize);
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  292                if (!dst_pte) {
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  293                        mutex_unlock(&hugetlb_fault_mutex_table[hash]);
> c0d0381ade7988 Mike Kravetz      2020-04-01  294                        i_mmap_unlock_read(mapping);
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  295                        goto out_unlock;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  296                }
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  297
> f619147104c8ea Axel Rasmussen    2021-05-04  298                if (mode != MCOPY_ATOMIC_CONTINUE &&
> f619147104c8ea Axel Rasmussen    2021-05-04  299                    !huge_pte_none(huge_ptep_get(dst_pte))) {
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  300                        err = -EEXIST;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  301                        mutex_unlock(&hugetlb_fault_mutex_table[hash]);
> c0d0381ade7988 Mike Kravetz      2020-04-01  302                        i_mmap_unlock_read(mapping);
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  303                        goto out_unlock;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  304                }
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  305
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  306                err = hugetlb_mcopy_atomic_pte(dst_mm, dst_pte, dst_vma,
> f619147104c8ea Axel Rasmussen    2021-05-04  307                                               dst_addr, src_addr, mode, &page);
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  308
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  309                mutex_unlock(&hugetlb_fault_mutex_table[hash]);
> c0d0381ade7988 Mike Kravetz      2020-04-01  310                i_mmap_unlock_read(mapping);
> 1c9e8def43a345 Mike Kravetz      2017-02-22  311                vm_alloc_shared = vm_shared;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  312
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  313                cond_resched();
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  314
> 9e368259ad9883 Andrea Arcangeli  2018-11-30  315                if (unlikely(err == -ENOENT)) {
> d8ed45c5dcd455 Michel Lespinasse 2020-06-08  316                        mmap_read_unlock(dst_mm);
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  317                        BUG_ON(!page);
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  318
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  319                        err = copy_huge_page_from_user(page,
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  320                                                (const void __user *)src_addr,
> 4fb07ee6510280 Wei Yang          2019-11-30  321                                                vma_hpagesize / PAGE_SIZE,
> 4fb07ee6510280 Wei Yang          2019-11-30  322                                                true);
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  323                        if (unlikely(err)) {
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  324                                err = -EFAULT;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  325                                goto out;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  326                        }
> d8ed45c5dcd455 Michel Lespinasse 2020-06-08  327                        mmap_read_lock(dst_mm);
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  328
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  329                        dst_vma = NULL;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  330                        goto retry;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  331                } else
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  332                        BUG_ON(page);
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  333
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  334                if (!err) {
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  335                        dst_addr += vma_hpagesize;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  336                        src_addr += vma_hpagesize;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  337                        copied += vma_hpagesize;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  338
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  339                        if (fatal_signal_pending(current))
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  340                                err = -EINTR;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  341                }
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  342                if (err)
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  343                        break;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  344        }
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  345
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  346  out_unlock:
> d8ed45c5dcd455 Michel Lespinasse 2020-06-08  347        mmap_read_unlock(dst_mm);
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  348  out:
> 1786d001262006 Mina Almasry      2021-06-01  349        if (page)
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  350                put_page(page);
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  351        BUG_ON(copied < 0);
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  352        BUG_ON(err > 0);
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  353        BUG_ON(!copied && !err);
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  354        return copied ? copied : err;
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  355  }
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  356  #else /* !CONFIG_HUGETLB_PAGE */
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  357  /* fail at build time if gcc attempts to use this */
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  358  extern ssize_t __mcopy_atomic_hugetlb(struct mm_struct *dst_mm,
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  359                                      struct vm_area_struct *dst_vma,
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  360                                      unsigned long dst_start,
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  361                                      unsigned long src_start,
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  362                                      unsigned long len,
> f619147104c8ea Axel Rasmussen    2021-05-04  363                                      enum mcopy_atomic_mode mode);
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  364  #endif /* CONFIG_HUGETLB_PAGE */
> 60d4d2d2b40e44 Mike Kravetz      2017-02-22  365
>
> :::::: The code at line 212 was first introduced by commit
> :::::: 1c9e8def43a3452e7af658b340f5f4f4ecde5c38 userfaultfd: hugetlbfs: add UFFDIO_COPY support for shared mappings
>
> :::::: TO: Mike Kravetz <mike.kravetz@xxxxxxxxxx>
> :::::: CC: Linus Torvalds <torvalds@xxxxxxxxxxxxxxxxxxxx>
>
> ---
> 0-DAY CI Kernel Test Service, Intel Corporation
> https://lists.01.org/hyperkitty/list/kbuild-all@xxxxxxxxxxxx




[Index of Archives]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux