(re-cc'ing ceph-devel list) Not right offhand, but I do have some comments about this code for the maintainers. See below: On Fri, 2017-10-13 at 23:11 +0000, Anish Gupta wrote: > Hello Jeff, > > > I did not see the issue with 4.4.14-rc4. > > > Would you or Zheng know which set of changes from 4.4 onwards might have helped fix this? > PS: I worked with Zheng on chage-id: 62a65f36d016fff32179acdbfcb8b2d8d9e54757 > > thanks, > Anish > > > On Monday, October 9, 2017, 9:12:54 AM PDT, Jeff Layton <jlayton@xxxxxxxxxx> wrote: > > > On Mon, 2017-10-09 at 16:09 +0000, Anish Gupta wrote: > > > > > > > Hello Ceph gurus, > > > > I was running a script in a loop that was doing the following: > > > > > Loop: > > > mount 2 different ceph "clients" using their own shares and secret keys and FSCache > > > write data into a sub-directory of client1 > > > let i=0; while (($i<10)); do sudo dd if=/dev/urandom of=data_file_$i bs=10M count=1; let i=$i+1; done > > > write data into client 2 using same as previous step. > > > read data from client 1 > > > cp -r /client1/data/ /tmp/scratch1_dir > > > read data from client 2 > > > cp -r /client2/data/ /tmp/scratch2_dir > > > unmount the 2 clients > > > > > > > > > > > > > Linux Kernel: 4.4.0-81-generic > > $ ceph --version > > ceph version 10.2.7 (50e863e0f4bc8f4b9e31156de690d765af245185) > > > > Only one script of above loop was running and after 4th iteration ran into this bug: > > > > Back trace looks like this: > > [2183922.263463] ------------[ cut here ]------------ > > [2183922.264771] kernel BUG at /root/ubuntu-xenial/fs/ceph/mds_client.c:1258! > > [2183922.264771] invalid opcode: 0000 [#1] SMP > > [2183922.264771] Modules linked in: nfs lockd grace xt_nat ceph(OE) libceph libcrc32c veth ip6t_REJECT nf_reject_ipv6 nf_log_ipv6 xt_recent ipt_MASQUERADE nf_nat_masquerade_ipv4 xfrm_user xfrm_algo iptable_nat nf_nat_ipv4 br_netfilter bridge stp llc overlay binfmt_misc cachefiles(OE) fscache(OE) xt_multiport xenfs xen_privcmd ppdev nls_utf8 isofs intel_rapl crct10dif_pclmul crc32_pclmul ghash_clmulni_intel nvidia_uvm(POE) aesni_intel aes_x86_64 lrw gf128mul glue_helper ablk_helper cryptd input_leds joydev serio_raw i2c_piix4 8250_fintek parport_pc parport mac_hid xt_hl ip6t_rt nf_conntrack_ipv6 nf_defrag_ipv6 ipt_REJECT nf_reject_ipv4 nf_log_ipv4 nf_log_common xt_LOG xt_limit xt_tcpudp xt_addrtype nf_conntrack_ipv4 nf_defrag_ipv4 xt_conntrack ip6table_filter ip6_tables nf_conntrack_netbios_ns nf_conntrack_broadcast > > [2183922.264771] nf_nat_ftp nf_nat nf_conntrack_ftp nf_conntrack iptable_filter ip_tables x_tables sunrpc autofs4 hid_generic usbhid hid nvidia_drm(POE) nvidia_modeset(POE) nvidia(POE) drm_kms_helper syscopyarea sysfillrect sysimgblt psmouse fb_sys_fops drm pata_acpi floppy fjes > > [2183922.264771] CPU: 0 PID: 19728 Comm: umount Tainted: P OE 4.4.0-81-generic #104-Ubuntu > > [2183922.264771] Hardware name: Xen HVM domU, BIOS 4.6.1-xs137595 03/23/2017 > > [2183922.264771] task: ffff880d3aa12a00 ti: ffff880f321a0000 task.ti: ffff880f321a0000 > > [2183922.264771] RIP: 0010:[<ffffffffc12218ac>] [<ffffffffc12218ac>] remove_session_caps+0x13c/0x150 [ceph] > > [2183922.264771] RSP: 0018:ffff880f321a3d38 EFLAGS: 00010202 > > [2183922.264771] RAX: 0000000000000001 RBX: ffff8809976b7000 RCX: ffff8807c1ee5028 > > [2183922.264771] RDX: 0000000000000039 RSI: ffff880f321a3d00 RDI: ffff880f3978a608 > > [2183922.264771] RBP: ffff880f321a3d78 R08: ffff880f321a0000 R09: 0000000000000000 > > [2183922.264771] R10: 0000000000032800 R11: 0000000000000000 R12: ffff8809976b7540 > > [2183922.264771] R13: ffff8807c1ee5168 R14: ffff8809976b7548 R15: ffff8809976b2800 > > [2183922.264771] FS: 00007f4fb7e28840(0000) GS:ffff880f42a00000(0000) knlGS:0000000000000000 > > [2183922.264771] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 > > [2183922.264771] CR2: 00007f4fb7656170 CR3: 0000000f34af2000 CR4: 00000000003406f0 > > [2183922.264771] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000 > > [2183922.264771] DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400 > > [2183922.264771] Stack: > > [2183922.264771] 0000010007dee895 fffffffffffffffe 0000000039d663f2 ffff880f3978a400 > > [2183922.264771] 0000000000000000 ffff8809976b7000 ffff880f3978a408 ffff8809976b7020 > > [2183922.264771] ffff880f321a3df0 ffffffffc12268f8 ffff88007bb9d200 00000001976b2e08 > > [2183922.264771] Call Trace: > > [2183922.264771] [<ffffffffc12268f8>] ceph_mdsc_close_sessions+0x1d8/0x2e0 [ceph] > > [2183922.264771] [<ffffffff8122bf27>] ? evict_inodes+0x177/0x1a0 > > [2183922.264771] [<ffffffffc12031b3>] ceph_put_super+0x23/0x50 [ceph] > > [2183922.264771] [<ffffffff812115df>] generic_shutdown_super+0x6f/0x100 > > [2183922.264771] [<ffffffffc1203136>] ceph_kill_sb+0x36/0x90 [ceph] > > [2183922.264771] [<ffffffff81211ac3>] deactivate_locked_super+0x43/0x70 > > [2183922.264771] [<ffffffff81211f9c>] deactivate_super+0x5c/0x60 > > [2183922.264771] [<ffffffff8123000f>] cleanup_mnt+0x3f/0x90 > > [2183922.264771] [<ffffffff812300a2>] __cleanup_mnt+0x12/0x20 > > [2183922.264771] [<ffffffff8109f031>] task_work_run+0x81/0xa0 > > [2183922.264771] [<ffffffff81003242>] exit_to_usermode_loop+0xc2/0xd0 > > [2183922.264771] [<ffffffff81003c6e>] syscall_return_slowpath+0x4e/0x60 > > [2183922.264771] [<ffffffff81840bd0>] int_ret_from_sys_call+0x25/0x8f > > [2183922.264771] Code: 2e 48 83 c4 18 5b 41 5c 41 5d 41 5e 41 5f 5d c3 48 89 fa 48 c7 c6 80 79 23 c1 48 c7 c7 40 4d 24 c1 e8 59 2c 20 c0 e9 f9 fe ff ff <0f> 0b 0f 0b e8 4b f8 e5 bf 90 66 2e 0f 1f 84 00 00 00 00 00 0f > > [2183922.264771] RIP [<ffffffffc12218ac>] remove_session_caps+0x13c/0x150 [ceph] > > [2183922.264771] RSP <ffff880f321a3d38> > > [2183922.307469] ---[ end trace e6fc86f4c2c986fd ]--- > > [2183922.263463] ------------[ cut here ]------------ > > [2183922.264771] kernel BUG at /root/ubuntu-xenial/fs/ceph/mds_client.c:1258! > > [2183922.264771] invalid opcode: 0000 [#1] SMP > > [2183922.264771] Modules linked in: nfs lockd grace xt_nat ceph(OE) libceph libcrc32c veth ip6t_REJECT nf_reject_ipv6 nf_log_ipv6 xt_recent ipt_MASQUERADE nf_nat_masquerade_ipv4 xfrm_user xfrm_algo iptable_nat nf_nat_ipv4 br_netfilter bridge stp llc overlay binfmt_misc cachefiles(OE) fscache(OE) xt_multiport xenfs xen_privcmd ppdev nls_utf8 isofs intel_rapl crct10dif_pclmul crc32_pclmul ghash_clmulni_intel nvidia_uvm(POE) aesni_intel aes_x86_64 lrw gf128mul glue_helper ablk_helper cryptd input_leds joydev serio_raw i2c_piix4 8250_fintek parport_pc parport mac_hid xt_hl ip6t_rt nf_conntrack_ipv6 nf_defrag_ipv6 ipt_REJECT nf_reject_ipv4 nf_log_ipv4 nf_log_common xt_LOG xt_limit xt_tcpudp xt_addrtype nf_conntrack_ipv4 nf_defrag_ipv4 xt_conntrack ip6table_filter ip6_tables nf_conntrack_netbios_ns nf_conntrack_broadcast > > [2183922.264771] nf_nat_ftp nf_nat nf_conntrack_ftp nf_conntrack iptable_filter ip_tables x_tables sunrpc autofs4 hid_generic usbhid hid nvidia_drm(POE) nvidia_modeset(POE) nvidia(POE) drm_kms_helper syscopyarea sysfillrect sysimgblt psmouse fb_sys_fops drm pata_acpi floppy fjes > > [2183922.264771] CPU: 0 PID: 19728 Comm: umount Tainted: P OE 4.4.0-81-generic #104-Ubuntu > > [2183922.264771] Hardware name: Xen HVM domU, BIOS 4.6.1-xs137595 03/23/2017 > > [2183922.264771] task: ffff880d3aa12a00 ti: ffff880f321a0000 task.ti: ffff880f321a0000 > > [2183922.264771] RIP: 0010:[<ffffffffc12218ac>] [<ffffffffc12218ac>] remove_session_caps+0x13c/0x150 [ceph] > > [2183922.264771] RSP: 0018:ffff880f321a3d38 EFLAGS: 00010202 > > [2183922.264771] RAX: 0000000000000001 RBX: ffff8809976b7000 RCX: ffff8807c1ee5028 > > [2183922.264771] RDX: 0000000000000039 RSI: ffff880f321a3d00 RDI: ffff880f3978a608 > > [2183922.264771] RBP: ffff880f321a3d78 R08: ffff880f321a0000 R09: 0000000000000000 > > [2183922.264771] R10: 0000000000032800 R11: 0000000000000000 R12: ffff8809976b7540 > > [2183922.264771] R13: ffff8807c1ee5168 R14: ffff8809976b7548 R15: ffff8809976b2800 > > [2183922.264771] FS: 00007f4fb7e28840(0000) GS:ffff880f42a00000(0000) knlGS:0000000000000000 > > [2183922.264771] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 > > [2183922.264771] CR2: 00007f4fb7656170 CR3: 0000000f34af2000 CR4: 00000000003406f0 > > [2183922.264771] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000 > > [2183922.264771] DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400 > > [2183922.264771] Stack: > > [2183922.264771] 0000010007dee895 fffffffffffffffe 0000000039d663f2 ffff880f3978a400 > > [2183922.264771] 0000000000000000 ffff8809976b7000 ffff880f3978a408 ffff8809976b7020 > > [2183922.264771] ffff880f321a3df0 ffffffffc12268f8 ffff88007bb9d200 00000001976b2e08 > > [2183922.264771] Call Trace: > > [2183922.264771] [<ffffffffc12268f8>] ceph_mdsc_close_sessions+0x1d8/0x2e0 [ceph] > > [2183922.264771] [<ffffffff8122bf27>] ? evict_inodes+0x177/0x1a0 > > [2183922.264771] [<ffffffffc12031b3>] ceph_put_super+0x23/0x50 [ceph] > > [2183922.264771] [<ffffffff812115df>] generic_shutdown_super+0x6f/0x100 > > [2183922.264771] [<ffffffffc1203136>] ceph_kill_sb+0x36/0x90 [ceph] > > [2183922.264771] [<ffffffff81211ac3>] deactivate_locked_super+0x43/0x70 > > [2183922.264771] [<ffffffff81211f9c>] deactivate_super+0x5c/0x60 > > [2183922.264771] [<ffffffff8123000f>] cleanup_mnt+0x3f/0x90 > > [2183922.264771] [<ffffffff812300a2>] __cleanup_mnt+0x12/0x20 > > [2183922.264771] [<ffffffff8109f031>] task_work_run+0x81/0xa0 > > [2183922.264771] [<ffffffff81003242>] exit_to_usermode_loop+0xc2/0xd0 > > [2183922.264771] [<ffffffff81003c6e>] syscall_return_slowpath+0x4e/0x60 > > [2183922.264771] [<ffffffff81840bd0>] int_ret_from_sys_call+0x25/0x8f > > [2183922.264771] Code: 2e 48 83 c4 18 5b 41 5c 41 5d 41 5e 41 5f 5d c3 48 89 fa 48 c7 c6 80 79 23 c1 48 c7 c7 40 4d 24 c1 e8 59 2c 20 c0 e9 f9 fe ff ff <0f> 0b 0f 0b e8 4b f8 e5 bf 90 66 2e 0f 1f 84 00 00 00 00 00 0f > > [2183922.264771] RIP [<ffffffffc12218ac>] remove_session_caps+0x13c/0x150 [ceph] > > [2183922.264771] RSP <ffff880f321a3d38> > > [2183922.307469] ---[ end trace e6fc86f4c2c986fd ]--- > > > > This is different from a previous one that Zheng had looked into few months back as it seems to > > point that s_nr_caps still remains > 0 ? > > > > 1 > > 1218 /* > > 1219 * caller must hold session s_mutex > > 1220 */ > > 1221 static void remove_session_caps(struct ceph_mds_session *session) > > 1222 { > > 1223 dout("remove_session_caps on %p\n", session); > > 1224 iterate_session_caps(session, remove_session_caps_cb, NULL); > > 1225 > > 1226 spin_lock(&session->s_cap_lock); > > 1227 if (session->s_nr_caps > 0) { > > 1228 struct super_block *sb = session->s_mdsc->fsc->sb; > > 1229 struct inode *inode; > > 1230 struct ceph_cap *cap, *prev = NULL; > > 1231 struct ceph_vino vino; > > 1232 /* > > 1233 * iterate_session_caps() skips inodes that are being > > 1234 * deleted, we need to wait until deletions are complete. > > 1235 * __wait_on_freeing_inode() is designed for the job, > > 1236 * but it is not exported, so use lookup inode function > > 1237 * to access it. > > 1238 */ > > 1239 while (!list_empty(&session->s_caps)) { > > 1240 cap = list_entry(session->s_caps.next, > > 1241 struct ceph_cap, session_caps); > > 1242 if (cap == prev) > > 1243 break; > > 1244 prev = cap; > > 1245 vino = cap->ci->i_vino; > > 1246 spin_unlock(&session->s_cap_lock); > > 1247 > > 1248 inode = ceph_find_inode(sb, vino); > > 1249 iput(inode); > > 1250 > > 1251 spin_lock(&session->s_cap_lock); > > 1252 } > > 1253 } > > 1254 > > 1255 // drop cap expires and unlock s_cap_lock > > 1256 cleanup_cap_releases(session->s_mdsc, session); > > 1257 > > 1258 BUG_ON(session->s_nr_caps > 0); > > 1259 BUG_ON(!list_empty(&session->s_cap_flushing)); > > 1260 } > > 1261 > > > > Easily reproducible multiple times. > FWIW, the locking in this function looks really wrong. Both s_caps list and s_nr_caps are protected by the s_cap_lock. That function drops that lock in the middle of walking the list to call ceph_find_inode. What prevents concurrent alterations to that list? Maybe some extra serialization was added that papered over this problem recently? > > > > > > 4.4.0 is pretty ancient. Is it reproducible vs. a more recent kernel? Say, something v4.14-rc ish? > -- > Jeff Layton <jlayton@xxxxxxxxxx > > -- Jeff Layton <jlayton@xxxxxxxxxx> -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html