On Wed, Feb 27, 2019 at 05:55:38PM +0100, Jan Kara wrote: > On Wed 27-02-19 04:24:51, Matthew Wilcox wrote: > Looking more into perf traces, I didn't notice any other obvious low > hanging fruit. There is one suboptimality in the fact that > __clear_shadow_entry() does xas_load() and the first thing xas_store() does > is xas_load() again. Removing this double tree traversal does bring > something back but since the traversals are so close together, everything > is cache hot and so the overall gain is small (but still): Calling xas_load() twice isn't too bad; the iterator stays at the base of the tree, so it's just one pointer reload. Still, I think we can avoid it. > COMMIT AVG STDDEV > singleiter 1467763.900000 1078.261049 > > So still 34 ms to go to the original time. > > What profiles do show is there's slightly more time spent here and there > adding to overall larger xas_store() time (compared to > __radix_tree_replace()) mostly due to what I'd blame to cache misses > (xas_store() is responsible for ~3.4% of cache misses after the patch while > xas_store() + __radix_tree_replace() caused only 1.5% together before). > > Some of the expensive loads seem to be from 'xas' structure (kind > of matches with what Nick said), other expensive loads seem to be loads from > xa_node. And I don't have a great explanation why e.g. a load of > xa_node->count is expensive when we looked at xa_node->shift before - > either the cache line fell out of cache or the byte accesses somehow > confuse the CPU. Also xas_store() has some new accesses compared to > __radix_tree_replace() - e.g. it did not previously touch node->shift. > > So overall I don't see easy way how to speed up xarray code further so > maybe just batching truncate to make up for some of the losses and live > with them where we cannot batch will be as good as it gets... Here's what I'm currently looking at. xas_store() becomes a wrapper around xas_replace() and xas_replace() avoids the xas_init_marks() and xas_load() calls: diff --git a/include/linux/xarray.h b/include/linux/xarray.h index 0e01e6129145..26fdba17ce67 100644 --- a/include/linux/xarray.h +++ b/include/linux/xarray.h @@ -1455,6 +1455,7 @@ static inline bool xas_retry(struct xa_state *xas, const void *entry) void *xas_load(struct xa_state *); void *xas_store(struct xa_state *, void *entry); +void xas_replace(struct xa_state *, void *curr, void *entry); void *xas_find(struct xa_state *, unsigned long max); void *xas_find_conflict(struct xa_state *); diff --git a/lib/test_xarray.c b/lib/test_xarray.c index 5d4bad8bd96a..b2e2cdf4eb74 100644 --- a/lib/test_xarray.c +++ b/lib/test_xarray.c @@ -38,6 +38,12 @@ static void *xa_store_index(struct xarray *xa, unsigned long index, gfp_t gfp) return xa_store(xa, index, xa_mk_index(index), gfp); } +static void xa_insert_index(struct xarray *xa, unsigned long index) +{ + XA_BUG_ON(xa, xa_insert(xa, index, xa_mk_index(index), + GFP_KERNEL) != 0); +} + static void xa_alloc_index(struct xarray *xa, unsigned long index, gfp_t gfp) { u32 id; @@ -338,6 +344,20 @@ static noinline void check_xa_shrink(struct xarray *xa) } } +static noinline void check_insert(struct xarray *xa) +{ + unsigned long i; + + for (i = 0; i < 1024; i++) { + xa_insert_index(xa, i); + XA_BUG_ON(xa, xa_load(xa, i - 1) != NULL); + XA_BUG_ON(xa, xa_load(xa, i + 1) != NULL); + xa_erase_index(xa, i); + } + + XA_BUG_ON(xa, !xa_empty(xa)); +} + static noinline void check_cmpxchg(struct xarray *xa) { void *FIVE = xa_mk_value(5); @@ -1527,6 +1547,7 @@ static int xarray_checks(void) check_xa_mark(&array); check_xa_shrink(&array); check_xas_erase(&array); + check_insert(&array); check_cmpxchg(&array); check_reserve(&array); check_reserve(&xa0); diff --git a/lib/xarray.c b/lib/xarray.c index 6be3acbb861f..8ff605bd0fee 100644 --- a/lib/xarray.c +++ b/lib/xarray.c @@ -613,7 +613,7 @@ static int xas_expand(struct xa_state *xas, void *head) /* * xas_create() - Create a slot to store an entry in. * @xas: XArray operation state. - * @allow_root: %true if we can store the entry in the root directly + * @entry: Entry which will be stored in the slot. * * Most users will not need to call this function directly, as it is called * by xas_store(). It is useful for doing conditional store operations @@ -623,14 +623,14 @@ static int xas_expand(struct xa_state *xas, void *head) * If the slot was newly created, returns %NULL. If it failed to create the * slot, returns %NULL and indicates the error in @xas. */ -static void *xas_create(struct xa_state *xas, bool allow_root) +static void *xas_create(struct xa_state *xas, void *entry) { struct xarray *xa = xas->xa; - void *entry; void __rcu **slot; struct xa_node *node = xas->xa_node; int shift; unsigned int order = xas->xa_shift; + bool allow_root = !xa_is_node(entry) && !xa_is_zero(entry); if (xas_top(node)) { entry = xa_head_locked(xa); @@ -701,7 +701,7 @@ void xas_create_range(struct xa_state *xas) xas->xa_sibs = 0; for (;;) { - xas_create(xas, true); + xas_create(xas, XA_ZERO_ENTRY); if (xas_error(xas)) goto restore; if (xas->xa_index <= (index | XA_CHUNK_MASK)) @@ -745,44 +745,36 @@ static void update_node(struct xa_state *xas, struct xa_node *node, } /** - * xas_store() - Store this entry in the XArray. + * xas_replace() - Replace one XArray entry with another. * @xas: XArray operation state. + * @curr: Current entry. * @entry: New entry. * - * If @xas is operating on a multi-index entry, the entry returned by this - * function is essentially meaningless (it may be an internal entry or it - * may be %NULL, even if there are non-NULL entries at some of the indices - * covered by the range). This is not a problem for any current users, - * and can be changed if needed. - * - * Return: The old entry at this index. + * This is not a cmpxchg operation. The caller asserts that @curr is the + * current entry at the index referred to by @xas and wishes to replace it + * with @entry. The slot must have already been created by xas_create() + * or by virtue of @curr being non-NULL. The marks are not changed by + * this operation. */ -void *xas_store(struct xa_state *xas, void *entry) +void xas_replace(struct xa_state *xas, void *curr, void *entry) { struct xa_node *node; void __rcu **slot = &xas->xa->xa_head; unsigned int offset, max; int count = 0; int values = 0; - void *first, *next; + void *next; bool value = xa_is_value(entry); - if (entry) { - bool allow_root = !xa_is_node(entry) && !xa_is_zero(entry); - first = xas_create(xas, allow_root); - } else { - first = xas_load(xas); - } - if (xas_invalid(xas)) - return first; + return; node = xas->xa_node; if (node && (xas->xa_shift < node->shift)) xas->xa_sibs = 0; - if ((first == entry) && !xas->xa_sibs) - return first; + if ((curr == entry) && !xas->xa_sibs) + return; - next = first; + next = curr; offset = xas->xa_offset; max = xas->xa_offset + xas->xa_sibs; if (node) { @@ -790,8 +782,6 @@ void *xas_store(struct xa_state *xas, void *entry) if (xas->xa_sibs) xas_squash_marks(xas); } - if (!entry) - xas_init_marks(xas); for (;;) { /* @@ -807,7 +797,7 @@ void *xas_store(struct xa_state *xas, void *entry) if (!node) break; count += !next - !entry; - values += !xa_is_value(first) - !value; + values += !xa_is_value(curr) - !value; if (entry) { if (offset == max) break; @@ -821,13 +811,41 @@ void *xas_store(struct xa_state *xas, void *entry) if (!xa_is_sibling(next)) { if (!entry && (offset > max)) break; - first = next; + curr = next; } slot++; } update_node(xas, node, count, values); - return first; +} + +/** + * xas_store() - Store this entry in the XArray. + * @xas: XArray operation state. + * @entry: New entry. + * + * If @xas is operating on a multi-index entry, the entry returned by this + * function is essentially meaningless (it may be an internal entry or it + * may be %NULL, even if there are non-NULL entries at some of the indices + * covered by the range). This is not a problem for any current users, + * and can be changed if needed. + * + * Return: The old entry at this index. + */ +void *xas_store(struct xa_state *xas, void *entry) +{ + void *curr; + + if (entry) { + curr = xas_create(xas, entry); + } else { + curr = xas_load(xas); + if (curr) + xas_init_marks(xas); + } + + xas_replace(xas, curr, entry); + return curr; } EXPORT_SYMBOL_GPL(xas_store); @@ -1472,9 +1490,9 @@ int __xa_insert(struct xarray *xa, unsigned long index, void *entry, gfp_t gfp) entry = XA_ZERO_ENTRY; do { - curr = xas_load(&xas); + curr = xas_create(&xas, entry); if (!curr) { - xas_store(&xas, entry); + xas_replace(&xas, curr, entry); if (xa_track_free(xa)) xas_clear_mark(&xas, XA_FREE_MARK); } else { @@ -1553,7 +1571,7 @@ void *xa_store_range(struct xarray *xa, unsigned long first, if (last + 1) order = __ffs(last + 1); xas_set_order(&xas, last, order); - xas_create(&xas, true); + xas_create(&xas, entry); if (xas_error(&xas)) goto unlock; } @@ -1606,11 +1624,13 @@ int __xa_alloc(struct xarray *xa, u32 *id, void *entry, do { xas.xa_index = limit.min; xas_find_marked(&xas, limit.max, XA_FREE_MARK); - if (xas.xa_node == XAS_RESTART) + if (xas.xa_node == XAS_RESTART) { xas_set_err(&xas, -EBUSY); - else + } else { + xas_create(&xas, entry); *id = xas.xa_index; - xas_store(&xas, entry); + } + xas_replace(&xas, NULL, entry); xas_clear_mark(&xas, XA_FREE_MARK); } while (__xas_nomem(&xas, gfp)); diff --git a/mm/filemap.c b/mm/filemap.c index 9f5e323e883e..56a7ef579879 100644 --- a/mm/filemap.c +++ b/mm/filemap.c @@ -131,8 +131,8 @@ static void page_cache_delete(struct address_space *mapping, VM_BUG_ON_PAGE(PageTail(page), page); VM_BUG_ON_PAGE(nr != 1 && shadow, page); - xas_store(&xas, shadow); xas_init_marks(&xas); + xas_replace(&xas, page, shadow); page->mapping = NULL; /* Leave page->index set: truncation lookup relies upon it */ @@ -326,7 +326,7 @@ static void page_cache_delete_batch(struct address_space *mapping, != pvec->pages[i]->index, page); tail_pages--; } - xas_store(&xas, NULL); + xas_replace(&xas, page, NULL); total_pages++; } mapping->nrpages -= total_pages; @@ -771,7 +771,7 @@ int replace_page_cache_page(struct page *old, struct page *new, gfp_t gfp_mask) new->index = offset; xas_lock_irqsave(&xas, flags); - xas_store(&xas, new); + xas_replace(&xas, old, new); old->mapping = NULL; /* hugetlb pages do not participate in page cache accounting. */ diff --git a/mm/migrate.c b/mm/migrate.c index d4fd680be3b0..083f52797d11 100644 --- a/mm/migrate.c +++ b/mm/migrate.c @@ -459,13 +459,13 @@ int migrate_page_move_mapping(struct address_space *mapping, SetPageDirty(newpage); } - xas_store(&xas, newpage); + xas_replace(&xas, page, newpage); if (PageTransHuge(page)) { int i; for (i = 1; i < HPAGE_PMD_NR; i++) { xas_next(&xas); - xas_store(&xas, newpage + i); + xas_replace(&xas, page + i, newpage + i); } } @@ -536,7 +536,7 @@ int migrate_huge_page_move_mapping(struct address_space *mapping, get_page(newpage); - xas_store(&xas, newpage); + xas_replace(&xas, page, newpage); page_ref_unfreeze(page, expected_count - 1); diff --git a/mm/shmem.c b/mm/shmem.c index 6ece1e2fe76e..83925601089d 100644 --- a/mm/shmem.c +++ b/mm/shmem.c @@ -341,7 +341,7 @@ static int shmem_replace_entry(struct address_space *mapping, item = xas_load(&xas); if (item != expected) return -ENOENT; - xas_store(&xas, replacement); + xas_replace(&xas, item, replacement); return 0; } diff --git a/mm/truncate.c b/mm/truncate.c index 798e7ccfb030..0682b2f9ac0e 100644 --- a/mm/truncate.c +++ b/mm/truncate.c @@ -38,7 +38,7 @@ static inline void __clear_shadow_entry(struct address_space *mapping, xas_set_update(&xas, workingset_update_node); if (xas_load(&xas) != entry) return; - xas_store(&xas, NULL); + xas_replace(&xas, entry, NULL); mapping->nrexceptional--; }