On Tue, Sep 27, 2016 at 02:48:00PM -0600, Ross Zwisler wrote: > DAX PMDs have been disabled since Jan Kara introduced DAX radix tree based > locking. This patch allows DAX PMDs to participate in the DAX radix tree > based locking scheme so that they can be re-enabled using the new struct > iomap based fault handlers. > > There are currently three types of DAX 4k entries: 4k zero pages, 4k DAX > mappings that have an associated block allocation, and 4k DAX empty > entries. The empty entries exist to provide locking for the duration of a > given page fault. > > This patch adds three equivalent 2MiB DAX entries: Huge Zero Page (HZP) > entries, PMD DAX entries that have associated block allocations, and 2 MiB > DAX empty entries. > > Unlike the 4k case where we insert a struct page* into the radix tree for > 4k zero pages, for HZP we insert a DAX exceptional entry with the new > RADIX_DAX_HZP flag set. This is because we use a single 2 MiB zero page in > every 2MiB hole mapping, and it doesn't make sense to have that same struct > page* with multiple entries in multiple trees. This would cause contention > on the single page lock for the one Huge Zero Page, and it would break the > page->index and page->mapping associations that are assumed to be valid in > many other places in the kernel. > > One difficult use case is when one thread is trying to use 4k entries in > radix tree for a given offset, and another thread is using 2 MiB entries > for that same offset. The current code handles this by making the 2 MiB > user fall back to 4k entries for most cases. This was done because it is > the simplest solution, and because the use of 2MiB pages is already > opportunistic. > > If we were to try to upgrade from 4k pages to 2MiB pages for a given range, > we run into the problem of how we lock out 4k page faults for the entire > 2MiB range while we clean out the radix tree so we can insert the 2MiB > entry. We can solve this problem if we need to, but I think that the cases > where both 2MiB entries and 4K entries are being used for the same range > will be rare enough and the gain small enough that it probably won't be > worth the complexity. > > Signed-off-by: Ross Zwisler <ross.zwisler@xxxxxxxxxxxxxxx> .... > +#if defined(CONFIG_TRANSPARENT_HUGEPAGE) > +/* > + * The 'colour' (ie low bits) within a PMD of a page offset. This comes up > + * more often than one might expect in the below functions. > + */ > +#define PG_PMD_COLOUR ((PMD_SIZE >> PAGE_SHIFT) - 1) > + > +static void __dax_pmd_dbg(struct iomap *iomap, unsigned long address, > + const char *reason, const char *fn) > +{ > + if (iomap) { > + char bname[BDEVNAME_SIZE]; > + > + bdevname(iomap->bdev, bname); > + pr_debug("%s: %s addr %lx dev %s type %#x blkno %ld " > + "offset %lld length %lld fallback: %s\n", fn, > + current->comm, address, bname, iomap->type, > + iomap->blkno, iomap->offset, iomap->length, reason); > + } else { > + pr_debug("%s: %s addr: %lx fallback: %s\n", fn, > + current->comm, address, reason); > + } > +} Yuck! Tracepoints for debugging information like this, please, not printk awfulness. > + > +#define dax_pmd_dbg(bh, address, reason) \ > + __dax_pmd_dbg(bh, address, reason, __func__) > + > +static int iomap_pmd_insert_mapping(struct vm_area_struct *vma, pmd_t *pmd, > + struct vm_fault *vmf, unsigned long address, > + struct iomap *iomap, loff_t pos, bool write, void **entryp) Please put a "dax" in the function name. grepping, cscope, etc are much easier when static function names are namespaced properly. > +{ > + struct address_space *mapping = vma->vm_file->f_mapping; > + struct block_device *bdev = iomap->bdev; > + struct blk_dax_ctl dax = { > + .sector = iomap_dax_sector(iomap, pos), > + .size = PMD_SIZE, > + }; > + long length = dax_map_atomic(bdev, &dax); > + void *ret; > + > + if (length < 0) { > + dax_pmd_dbg(iomap, address, "dax-error fallback"); > + return VM_FAULT_FALLBACK; > + } Fails to unmap. Please use an goto based error stack. And tracepoints make this much neater: trace_dax_pmd_insert_mapping(iomap, address, &dax, length); if (length < 0) goto unmap_fallback; if (length < PMD_SIZE) goto unmap_fallback; ..... trace_dax_pmd_insert_mapping_done(iomap, address, &dax, length); return vmf_insert_pfn_pmd(vma, address, pmd, dax.pfn, write); unmap_fallback: dax_unmap_atomic(bdev, &dax); fallback: trace_dax_pmd_insert_fallback(iomap, address, &dax, length); return VM_FAULT_FALLBACK; } i.e. we don't need need all those debug printks to tell us what failed - the first tracepoint tells use everything about the context we are about to check, and the last tracepoint tells us whether we are falling back or about to try mapping a PMD. If you really need custom printk output for debugging, then use trace_printk() so that it shows up in the trace output along with all the trace points.... Same goes for all the other pr_debug() cals in this code - they need to go and be replaced with tracepoints. > +int iomap_dax_pmd_fault(struct vm_area_struct *vma, unsigned long address, > + pmd_t *pmd, unsigned int flags, struct iomap_ops *ops) dax_iomap_pmd_fault() - dax_ is the namespace prefix for the code in fs/dax.c, not iomap_... > +{ > + struct address_space *mapping = vma->vm_file->f_mapping; > + unsigned long pmd_addr = address & PMD_MASK; > + bool write = flags & FAULT_FLAG_WRITE; > + struct inode *inode = mapping->host; > + struct iomap iomap = { 0 }; > + int error, result = 0; > + pgoff_t size, pgoff; > + struct vm_fault vmf; > + void *entry; > + loff_t pos; > + > + /* dax pmd mappings require pfn_t_devmap() */ > + if (!IS_ENABLED(CONFIG_FS_DAX_PMD)) > + return VM_FAULT_FALLBACK; So we build all this stuff in, even if CONFIG_FS_DAX_PMD=n? Shouldn't we just have a simple function that returns VM_FAULT_FALLBACK when CONFIG_FS_DAX_PMD=n? Cheers, Dave. -- Dave Chinner david@xxxxxxxxxxxxx -- To unsubscribe from this list: send the line "unsubscribe linux-ext4" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html