The patch titled Subject: slab: implement bulk alloc in SLAB allocator has been added to the -mm tree. Its filename is slab-implement-bulk-alloc-in-slab-allocator.patch This patch should soon appear at http://ozlabs.org/~akpm/mmots/broken-out/slab-implement-bulk-alloc-in-slab-allocator.patch and later at http://ozlabs.org/~akpm/mmotm/broken-out/slab-implement-bulk-alloc-in-slab-allocator.patch Before you just go and hit "reply", please: a) Consider who else should be cc'ed b) Prefer to cc a suitable mailing list as well c) Ideally: find the original patch on the mailing list and do a reply-to-all to that, adding suitable additional cc's *** Remember to use Documentation/SubmitChecklist when testing your code *** The -mm tree is included into linux-next and is updated there every 3-4 working days ------------------------------------------------------ From: Jesper Dangaard Brouer <brouer@xxxxxxxxxx> Subject: slab: implement bulk alloc in SLAB allocator This patch implements the alloc side of bulk API for the SLAB allocator. Further optimization are still possible by changing the call to __do_cache_alloc() into something that can return multiple objects. This optimization is left for later, given end results already show in the area of 80% speedup. Signed-off-by: Jesper Dangaard Brouer <brouer@xxxxxxxxxx> Cc: Christoph Lameter <cl@xxxxxxxxx> Cc: Pekka Enberg <penberg@xxxxxxxxxx> Cc: David Rientjes <rientjes@xxxxxxxxxx> Cc: Joonsoo Kim <iamjoonsoo.kim@xxxxxxx> Cc: Vladimir Davydov <vdavydov@xxxxxxxxxxxxx> Signed-off-by: Andrew Morton <akpm@xxxxxxxxxxxxxxxxxxxx> --- mm/slab.c | 37 +++++++++++++++++++++++++++++++++++-- 1 file changed, 35 insertions(+), 2 deletions(-) diff -puN mm/slab.c~slab-implement-bulk-alloc-in-slab-allocator mm/slab.c --- a/mm/slab.c~slab-implement-bulk-alloc-in-slab-allocator +++ a/mm/slab.c @@ -3390,9 +3390,42 @@ void kmem_cache_free_bulk(struct kmem_ca EXPORT_SYMBOL(kmem_cache_free_bulk); int kmem_cache_alloc_bulk(struct kmem_cache *s, gfp_t flags, size_t size, - void **p) + void **p) { - return __kmem_cache_alloc_bulk(s, flags, size, p); + size_t i; + + s = slab_pre_alloc_hook(s, flags); + if (!s) + return 0; + + cache_alloc_debugcheck_before(s, flags); + + local_irq_disable(); + for (i = 0; i < size; i++) { + void *objp = __do_cache_alloc(s, flags); + + /* this call could be done outside IRQ disabled section */ + objp = cache_alloc_debugcheck_after(s, flags, objp, _RET_IP_); + + if (unlikely(!objp)) + goto error; + p[i] = objp; + } + local_irq_enable(); + + /* Clear memory outside IRQ disabled section */ + if (unlikely(flags & __GFP_ZERO)) + for (i = 0; i < size; i++) + memset(p[i], 0, s->object_size); + + slab_post_alloc_hook(s, flags, size, p); + /* FIXME: Trace call missing. Christoph would like a bulk variant */ + return size; +error: + local_irq_enable(); + slab_post_alloc_hook(s, flags, i, p); + __kmem_cache_free_bulk(s, i, p); + return 0; } EXPORT_SYMBOL(kmem_cache_alloc_bulk); _ Patches currently in -mm which might be from brouer@xxxxxxxxxx are slub-cleanup-code-for-kmem-cgroup-support-to-kmem_cache_free_bulk.patch mm-slab-move-slub-alloc-hooks-to-common-mm-slabh.patch mm-fault-inject-take-over-bootstrap-kmem_cache-check.patch slab-use-slab_pre_alloc_hook-in-slab-allocator-shared-with-slub.patch mm-kmemcheck-skip-object-if-slab-allocation-failed.patch slab-use-slab_post_alloc_hook-in-slab-allocator-shared-with-slub.patch slab-implement-bulk-alloc-in-slab-allocator.patch slab-avoid-running-debug-slab-code-with-irqs-disabled-for-alloc_bulk.patch slab-implement-bulk-free-in-slab-allocator.patch mm-new-api-kfree_bulk-for-slabslub-allocators.patch mm-fix-some-spelling.patch -- To unsubscribe from this list: send the line "unsubscribe mm-commits" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html