From: Andrea Arcangeli <aarcange@xxxxxxxxxx> Page migration requires rmap to be able to find all migration ptes created by migration. If the second rmap_walk clearing migration PTEs misses an entry, it is left dangling causing a BUG_ON to trigger during fault. For example; [ 511.201534] kernel BUG at include/linux/swapops.h:105! [ 511.201534] invalid opcode: 0000 [#1] PREEMPT SMP [ 511.201534] last sysfs file: /sys/block/sde/size [ 511.201534] CPU 0 [ 511.201534] Modules linked in: kvm_amd kvm dm_crypt loop i2c_piix4 serio_raw tpm_tis shpchp evdev tpm i2c_core pci_hotplug tpm_bios wmi processor button ext3 jbd mbcache dm_mirror dm_region_hash dm_log dm_snapshot dm_mod raid10 raid456 async_raid6_recov async_pq raid6_pq async_xor xor async_memcpy async_tx raid1 raid0 multipath linear md_mod sg sr_mod cdrom sd_mod ata_generic ahci libahci libata ide_pci_generic ehci_hcd ide_core r8169 mii ohci_hcd scsi_mod floppy thermal fan thermal_sys [ 511.888526] [ 511.888526] Pid: 20431, comm: date Not tainted 2.6.34-rc4-mm1-fix-swapops #6 GA-MA790GP-UD4H/GA-MA790GP-UD4H [ 511.888526] RIP: 0010:[<ffffffff811094ff>] [<ffffffff811094ff>] migration_entry_wait+0xc1/0x129 [ 512.173545] RSP: 0018:ffff880037b979d8 EFLAGS: 00010246 [ 512.198503] RAX: ffffea0000000000 RBX: ffffea0001a2ba10 RCX: 0000000000029830 [ 512.329617] RDX: 0000000001a2ba10 RSI: ffffffff818264b8 RDI: 000000000ef45c3e [ 512.380001] RBP: ffff880037b97a08 R08: ffff880078003f00 R09: ffff880037b979e8 [ 512.380001] R10: ffffffff8114ddaa R11: 0000000000000246 R12: 0000000037304000 [ 512.380001] R13: ffff88007a9ed5c8 R14: f800000000077a2e R15: 000000000ef45c3e [ 512.380001] FS: 00007f3d346866e0(0000) GS:ffff880002200000(0000) knlGS:0000000000000000 [ 512.380001] CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b [ 512.380001] CR2: 00007fff6abec9c1 CR3: 0000000037a15000 CR4: 00000000000006f0 [ 512.380001] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000 [ 513.004775] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400 [ 513.068415] Process date (pid: 20431, threadinfo ffff880037b96000, task ffff880078003f00) [ 513.068415] Stack: [ 513.068415] ffff880037b97b98 ffff880037b97a18 ffff880037b97be8 0000000000000c00 [ 513.228068] <0> ffff880037304f60 00007fff6abec9c1 ffff880037b97aa8 ffffffff810e951a [ 513.228068] <0> ffff880037b97a88 0000000000000246 0000000000000000 ffffffff8130c5c2 [ 513.228068] Call Trace: [ 513.228068] [<ffffffff810e951a>] handle_mm_fault+0x3f8/0x76a [ 513.228068] [<ffffffff8130c5c2>] ? do_page_fault+0x26a/0x46e [ 513.228068] [<ffffffff8130c7a2>] do_page_fault+0x44a/0x46e [ 513.720755] [<ffffffff8130875d>] ? trace_hardirqs_off_thunk+0x3a/0x3c [ 513.789278] [<ffffffff8114ddaa>] ? load_elf_binary+0x14a1/0x192b [ 513.851506] [<ffffffff813099b5>] page_fault+0x25/0x30 [ 513.851506] [<ffffffff8114ddaa>] ? load_elf_binary+0x14a1/0x192b [ 513.851506] [<ffffffff811c1e27>] ? strnlen_user+0x3f/0x57 [ 513.851506] [<ffffffff8114de33>] load_elf_binary+0x152a/0x192b [ 513.851506] [<ffffffff8111329b>] search_binary_handler+0x173/0x313 [ 513.851506] [<ffffffff8114c909>] ? load_elf_binary+0x0/0x192b [ 513.851506] [<ffffffff81114896>] do_execve+0x219/0x30a [ 513.851506] [<ffffffff8111887f>] ? getname+0x14d/0x1b3 [ 513.851506] [<ffffffff8100a5c6>] sys_execve+0x43/0x5e [ 514.483501] [<ffffffff8100320a>] stub_execve+0x6a/0xc0 [ 514.548357] Code: 74 05 83 f8 1f 75 68 48 b8 ff ff ff ff ff ff ff 07 48 21 c2 48 b8 00 00 00 00 00 ea ff ff 48 6b d2 38 48 8d 1c 02 f6 03 01 75 04 <0f> 0b eb fe 8b 4b 08 48 8d 73 08 85 c9 74 35 8d 41 01 89 4d e0 [ 514.704292] RIP [<ffffffff811094ff>] migration_entry_wait+0xc1/0x129 [ 514.808221] RSP <ffff880037b979d8> [ 514.906179] ---[ end trace 4f88495edc224d6b ]--- This particular BUG_ON is caused by a race between shift_arg_pages and migration. During exec, a temporary stack is created and later moved to its final location. If migration selects a page within the temporary stack, the page tables and migration PTE can be copied to the new location before rmap_walk is able to find the copy. This leaves a dangling migration PTE behind that later triggers the bug. This patch fixes the problem by using two VMAs - one which covers the temporary stack and the other which covers the new location. This guarantees that rmap can always find the migration PTE even if it is copied while rmap_walk is taking place. [mel@xxxxxxxxx: Tested and rewrote changelog] Signed-off-by: Andrea Arcangeli <aarcange@xxxxxxxxxx> Signed-off-by: Mel Gorman <mel@xxxxxxxxx> --- fs/exec.c | 37 +++++++++++++++++++++++++++++++++---- 1 files changed, 33 insertions(+), 4 deletions(-) diff --git a/fs/exec.c b/fs/exec.c index 725d7ef..fd0abff 100644 --- a/fs/exec.c +++ b/fs/exec.c @@ -55,6 +55,7 @@ #include <linux/fsnotify.h> #include <linux/fs_struct.h> #include <linux/pipe_fs_i.h> +#include <linux/rmap.h> #include <asm/uaccess.h> #include <asm/mmu_context.h> @@ -503,7 +504,9 @@ static int shift_arg_pages(struct vm_area_struct *vma, unsigned long shift) unsigned long length = old_end - old_start; unsigned long new_start = old_start - shift; unsigned long new_end = old_end - shift; + unsigned long moved_length; struct mmu_gather *tlb; + struct vm_area_struct *tmp_vma; BUG_ON(new_start > new_end); @@ -515,17 +518,43 @@ static int shift_arg_pages(struct vm_area_struct *vma, unsigned long shift) return -EFAULT; /* + * We need to create a fake temporary vma and index it in the + * anon_vma list in order to allow the pages to be reachable + * at all times by the rmap walk for migrate, while + * move_page_tables() is running. + */ + tmp_vma = kmem_cache_alloc(vm_area_cachep, GFP_KERNEL); + if (!tmp_vma) + return -ENOMEM; + *tmp_vma = *vma; + INIT_LIST_HEAD(&tmp_vma->anon_vma_chain); + /* * cover the whole range: [new_start, old_end) */ - if (vma_adjust(vma, new_start, old_end, vma->vm_pgoff, NULL)) + tmp_vma->vm_start = new_start; + /* + * The tmp_vma destination of the copy (with the new vm_start) + * has to be at the end of the anon_vma list for the rmap_walk + * to find the moved pages at all times. + */ + if (unlikely(anon_vma_clone(tmp_vma, vma))) { + kmem_cache_free(vm_area_cachep, tmp_vma); return -ENOMEM; + } /* * move the page tables downwards, on failure we rely on * process cleanup to remove whatever mess we made. */ - if (length != move_page_tables(vma, old_start, - vma, new_start, length)) + moved_length = move_page_tables(vma, old_start, + vma, new_start, length); + + vma->vm_start = new_start; + /* rmap walk will already find all pages using the new_start */ + unlink_anon_vmas(tmp_vma); + kmem_cache_free(vm_area_cachep, tmp_vma); + + if (length != moved_length) return -ENOMEM; lru_add_drain(); @@ -551,7 +580,7 @@ static int shift_arg_pages(struct vm_area_struct *vma, unsigned long shift) /* * Shrink the vma to just the new range. Always succeeds. */ - vma_adjust(vma, new_start, new_end, vma->vm_pgoff, NULL); + vma->vm_end = new_end; return 0; } -- 1.6.5 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxxx For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>