From: Ganesh Mahendran <opensource.ganesh@xxxxxxxxx> Subject: mm/zsmalloc: take obj index back from find_alloced_obj the obj index value should be updated after return from find_alloced_obj() to avoid CPU burning caused by unnecessary object scanning. Link: http://lkml.kernel.org/r/1467882338-4300-2-git-send-email-opensource.ganesh@xxxxxxxxx Signed-off-by: Ganesh Mahendran <opensource.ganesh@xxxxxxxxx> Reviewed-by: Sergey Senozhatsky <sergey.senozhatsky@xxxxxxxxx> Acked-by: Minchan Kim <minchan@xxxxxxxxxx> Signed-off-by: Andrew Morton <akpm@xxxxxxxxxxxxxxxxxxxx> --- mm/zsmalloc.c | 8 ++++++-- 1 file changed, 6 insertions(+), 2 deletions(-) diff -puN mm/zsmalloc.c~mm-zsmalloc-take-obj-index-back-from-find_alloced_obj mm/zsmalloc.c --- a/mm/zsmalloc.c~mm-zsmalloc-take-obj-index-back-from-find_alloced_obj +++ a/mm/zsmalloc.c @@ -1741,10 +1741,11 @@ static void zs_object_copy(struct size_c * return handle. */ static unsigned long find_alloced_obj(struct size_class *class, - struct page *page, int index) + struct page *page, int *obj_idx) { unsigned long head; int offset = 0; + int index = *obj_idx; unsigned long handle = 0; void *addr = kmap_atomic(page); @@ -1765,6 +1766,9 @@ static unsigned long find_alloced_obj(st } kunmap_atomic(addr); + + *obj_idx = index; + return handle; } @@ -1790,7 +1794,7 @@ static int migrate_zspage(struct zs_pool int ret = 0; while (1) { - handle = find_alloced_obj(class, s_page, obj_idx); + handle = find_alloced_obj(class, s_page, &obj_idx); if (!handle) { s_page = get_next_page(s_page); if (!s_page) _ -- To unsubscribe from this list: send the line "unsubscribe mm-commits" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html