On Mon, 2016-10-24 at 21:00 -0400, Chuck Lever wrote: > > > > On Oct 24, 2016, at 8:57 PM, Jeff Layton <jlayton@xxxxxxxxxx> wrote: > > > > On Mon, 2016-10-24 at 17:38 -0400, Chuck Lever wrote: > > > > > > > > > > > > > > > On Oct 24, 2016, at 4:40 PM, J. Bruce Fields <bfields@xxxxxxxxxxxx> wrote: > > > > > > > > On Mon, Oct 24, 2016 at 03:17:34PM -0400, Jeff Layton wrote: > > > > > > > > > > > > > > > On Mon, 2016-10-24 at 14:08 -0400, J. Bruce Fields wrote: > > > > > > > > > > > > > > > > > > On Mon, Oct 24, 2016 at 11:24:40AM -0400, Jeff Layton wrote: > > > > > > > > > > > > > > > > > > > > > > > > > > > > On Mon, 2016-10-24 at 11:19 -0400, Jeff Layton wrote: > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > On Mon, 2016-10-24 at 09:51 -0400, Chuck Lever wrote: > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > On Oct 24, 2016, at 9:31 AM, Jeff Layton <jlayton@xxxxxxxxxx> wrote: > > > > > > > > > > > > > > > > > > > > On Mon, 2016-10-24 at 11:15 +0800, Eryu Guan wrote: > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > On Sun, Oct 23, 2016 at 02:21:15PM -0400, J. Bruce Fields wrote: > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > I'm getting an intermittent crash in the nfs server as of > > > > > > > > > > > > 68778945e46f143ed7974b427a8065f69a4ce944 "SUNRPC: Separate buffer > > > > > > > > > > > > pointers for RPC Call and Reply messages". > > > > > > > > > > > > > > > > > > > > > > > > I haven't tried to understand that commit or why it would be a problem yet, I > > > > > > > > > > > > don't see an obvious connection--I can take a closer look Monday. > > > > > > > > > > > > > > > > > > > > > > > > Could even be that I just landed on this commit by chance, the problem is a > > > > > > > > > > > > little hard to reproduce so I don't completely trust my testing. > > > > > > > > > > > > > > > > > > > > > > I've hit the same crash on 4.9-rc1 kernel, and it's reproduced for me > > > > > > > > > > > reliably by running xfstests generic/013 case, on a loopback mounted > > > > > > > > > > > NFSv4.1 (or NFSv4.2), XFS is the underlying exported fs. More details > > > > > > > > > > > please see > > > > > > > > > > > > > > > > > > > > > > http://marc.info/?l=linux-nfs&m=147714320129362&w=2 > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > Looks like you landed at the same commit as Bruce, so that's probably > > > > > > > > > > legit. That commit is very small though. The only real change that > > > > > > > > > > doesn't affect the new field is this: > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > @@ -1766,7 +1766,7 @@ rpc_xdr_encode(struct rpc_task *task) > > > > > > > > > > req->rq_buffer, > > > > > > > > > > req->rq_callsize); > > > > > > > > > > xdr_buf_init(&req->rq_rcv_buf, > > > > > > > > > > - (char *)req->rq_buffer + req->rq_callsize, > > > > > > > > > > + req->rq_rbuffer, > > > > > > > > > > req->rq_rcvsize); > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > So I'm guessing this is breaking the callback channel somehow? > > > > > > > > > > > > > > > > > > Could be the TCP backchannel code is using rq_buffer in a different > > > > > > > > > way than RDMA backchannel or the forward channel code. > > > > > > > > > > > > > > > > > > > > > > > > > Well, it basically allocates a page per rpc_rqst and then maps that. > > > > > > > > > > > > > > > > One thing I notice is that this patch ensures that rq_rbuffer gets set > > > > > > > > up in rpc_malloc and xprt_rdma_allocate, but it looks like > > > > > > > > xprt_alloc_bc_req didn't get the same treatment. > > > > > > > > > > > > > > > > I suspect that that may be the problem... > > > > > > > > > > > > > > > In fact, maybe we just need this here? (untested and probably > > > > > > > whitespace damaged): > > > > > > > > > > > > No change in results for me. > > > > > > > > > > > > --b. > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > diff --git a/net/sunrpc/backchannel_rqst.c b/net/sunrpc/backchannel_rqst.c > > > > > > > index ac701c28f44f..c561aa8ce05b 100644 > > > > > > > --- a/net/sunrpc/backchannel_rqst.c > > > > > > > +++ b/net/sunrpc/backchannel_rqst.c > > > > > > > @@ -100,6 +100,7 @@ struct rpc_rqst *xprt_alloc_bc_req(struct rpc_xprt *xprt, gfp_t gfp_flags) > > > > > > > goto out_free; > > > > > > > } > > > > > > > req->rq_rcv_buf.len = PAGE_SIZE; > > > > > > > + req->rq_rbuffer = req->rq_rcv_buf.head[0].iov_base; > > > > > > > > > > > > > > /* Preallocate one XDR send buffer */ > > > > > > > if (xprt_alloc_xdr_buf(&req->rq_snd_buf, gfp_flags) < 0) { > > > > > > > > > > Ahh ok, I think I see. > > > > > > > > > > We probably also need to set rq_rbuffer in bc_malloc and and > > > > > xprt_rdma_bc_allocate. > > > > > > > > > > My guess is that we're ending up in rpc_xdr_encode with a NULL > > > > > rq_rbuffer pointer, so the right fix would seem to be to ensure that it > > > > > is properly set whenever rq_buffer is set. > > > > > > > > > > So I think this may be what we want, actually. I'll plan to test it out > > > > > but may not get to it before tomorrow. > > > > > > > > It passes here. > > > > > > Without Jeff's patch, my server locks up during generic/013 with NFS/RDMA > > > and NFSv4.1. With it, I get all the way to generic/089, and then encounter > > > this: > > > > > > > Thanks for testing. I just sent an "official" patch submission with the > > same patch. > > > > > > > > > > Oct 24 17:31:11 klimt kernel: general protection fault: 0000 [#1] SMP > > > Oct 24 17:31:11 klimt kernel: Modules linked in: cts rpcsec_gss_krb5 sb_edac edac_core x86_pkg_temp_thermal intel_powerclamp coretemp kvm_intel kvm btrfs irqbypass crct10dif_pclmul crc32_pclmul ghash_clmulni_intel iTCO_wdt iTCO_vendor_support aesni_intel lrw gf128mul glue_helper ablk_helper cryptd xor raid6_pq rpcrdma pcspkr i2c_i801 lpc_ich ib_ipoib i2c_smbus mfd_core mei_me sg mei rdma_ucm shpchp ioatdma wmi ib_ucm ipmi_si ipmi_msghandler ib_uverbs ib_umad rdma_cm ib_cm iw_cm acpi_pad acpi_power_meter nfsd auth_rpcgss nfs_acl lockd grace sunrpc ip_tables xfs libcrc32c mlx4_ib ib_core mlx4_en sd_mod sr_mod cdrom ast drm_kms_helper syscopyarea sysfillrect sysimgblt fb_sys_fops ttm drm mlx4_core ahci libahci igb crc32c_intel ptp libata pps_core dca i2c_algo_bit i2c_core dm_mirror dm_region_hash dm_log dm_mod > > > Oct 24 17:31:11 klimt kernel: CPU: 3 PID: 1681 Comm: nfsd Not tainted 4.9.0-rc2-00006-g0e8f985 #2 > > > Oct 24 17:31:11 klimt kernel: Hardware name: Supermicro Super Server/X10SRL-F, BIOS 1.0c 09/09/2015 > > > Oct 24 17:31:11 klimt kernel: task: ffff88083feb15c0 task.stack: ffffc900058a0000 > > > Oct 24 17:31:11 klimt kernel: RIP: 0010:[<ffffffff811ed1f0>] [<ffffffff811ed1f0>] __kmalloc_track_caller+0x1b0/0x220 > > > Oct 24 17:31:11 klimt kernel: RSP: 0018:ffffc900058a3b88 EFLAGS: 00010286 > > > Oct 24 17:31:11 klimt kernel: RAX: 0000000000000000 RBX: 00000000024000c0 RCX: 0000000000293b6b > > > Oct 24 17:31:11 klimt kernel: RDX: 0000000000293b6a RSI: 0000000000000000 RDI: 0000000000000002 > > > Oct 24 17:31:11 klimt kernel: RBP: ffffc900058a3bc8 R08: 000000000001c560 R09: ffff88085f003a40 > > > Oct 24 17:31:11 klimt kernel: R10: ffff88085f003a40 R11: ffff88083fc9b3c0 R12: 00000000024000c0 > > > Oct 24 17:31:11 klimt kernel: R13: 0000000000000018 R14: ffff88081993d600 R15: c0d435750e5f8620 > > > Oct 24 17:31:11 klimt kernel: FS: 0000000000000000(0000) GS:ffff88087fcc0000(0000) knlGS:0000000000000000 > > > Oct 24 17:31:11 klimt kernel: CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 > > > Oct 24 17:31:11 klimt kernel: CR2: 00007fca9725a298 CR3: 0000000001c06000 CR4: 00000000001406e0 > > > Oct 24 17:31:11 klimt kernel: Stack: > > > Oct 24 17:31:11 klimt kernel: ffff88085f003a40 ffff88085f003a40 ffffffff812c227b 0000000000000018 > > > Oct 24 17:31:11 klimt kernel: ffff880852ee6b00 ffff88085583a000 ffff88081993d600 ffff880839a95840 > > > Oct 24 17:31:11 klimt kernel: ffffc900058a3be8 ffffffff811a65d0 ffff88081993d600 00000000024000c0 > > > Oct 24 17:31:11 klimt kernel: Call Trace: > > > Oct 24 17:31:11 klimt kernel: [<ffffffff812c227b>] ? selinux_cred_prepare+0x1b/0x30 > > > Oct 24 17:31:11 klimt kernel: [<ffffffff811a65d0>] kmemdup+0x20/0x50 > > > Oct 24 17:31:11 klimt kernel: [<ffffffff812c227b>] selinux_cred_prepare+0x1b/0x30 > > > Oct 24 17:31:11 klimt kernel: [<ffffffff812bd3f9>] security_prepare_creds+0x39/0x60 > > > Oct 24 17:31:11 klimt kernel: [<ffffffff810a486f>] prepare_creds+0x12f/0x150 > > > Oct 24 17:31:11 klimt kernel: [<ffffffffa043ee6c>] nfsd_setuser+0x8c/0x250 [nfsd] > > > Oct 24 17:31:11 klimt kernel: [<ffffffffa0438077>] nfsd_setuser_and_check_port+0x77/0xa0 [nfsd] > > > Oct 24 17:31:11 klimt kernel: [<ffffffff812bd3f9>] ? security_prepare_creds+0x39/0x60 > > > Oct 24 17:31:11 klimt kernel: [<ffffffffa04a33f6>] ? write_bytes_to_xdr_buf+0xa6/0xd0 [sunrpc] > > > Oct 24 17:31:11 klimt kernel: [<ffffffffa0438dcc>] fh_verify+0x5cc/0x610 [nfsd] > > > Oct 24 17:31:11 klimt kernel: [<ffffffffa044659b>] nfsd4_getattr+0x1b/0x80 [nfsd] > > > Oct 24 17:31:11 klimt kernel: [<ffffffffa04481fd>] nfsd4_proc_compound+0x40d/0x690 [nfsd] > > > Oct 24 17:31:11 klimt kernel: [<ffffffffa04352b4>] nfsd_dispatch+0xd4/0x1d0 [nfsd] > > > Oct 24 17:31:11 klimt kernel: [<ffffffffa049b249>] svc_process_common+0x3d9/0x700 [sunrpc] > > > Oct 24 17:31:11 klimt kernel: [<ffffffffa049c0c1>] svc_process+0xf1/0x1d0 [sunrpc] > > > Oct 24 17:31:11 klimt kernel: [<ffffffffa0434d4f>] nfsd+0xff/0x160 [nfsd] > > > Oct 24 17:31:11 klimt kernel: [<ffffffffa0434c50>] ? nfsd_destroy+0x60/0x60 [nfsd] > > > Oct 24 17:31:11 klimt kernel: [<ffffffff810a2235>] kthread+0xe5/0xf0 > > > Oct 24 17:31:11 klimt kernel: [<ffffffff810a2150>] ? kthread_stop+0x120/0x120 > > > Oct 24 17:31:11 klimt kernel: [<ffffffff816aa795>] ret_from_fork+0x25/0x30 > > > Oct 24 17:31:11 klimt kernel: Code: 85 d2 75 dd eb d3 48 89 c3 66 90 eb 6e 81 e3 00 00 10 00 0f 84 e4 fe ff ff 66 90 e9 e7 fe ff ff 49 63 42 20 48 8d 4a 01 4d 8b 02 <49> 8b 1c 07 4c 89 f8 65 49 0f c7 08 0f 94 c0 84 c0 0f 85 3b ff > > > Oct 24 17:31:11 klimt kernel: RIP [<ffffffff811ed1f0>] __kmalloc_track_caller+0x1b0/0x220 > > > Oct 24 17:31:11 klimt kernel: RSP <ffffc900058a3b88> > > > Oct 24 17:31:11 klimt kernel: ---[ end trace af2b44f9db7f5a36 ]--- > > > > > > This appears to be unrelated. Maybe. > > > > > > Bruce, do you have CONFIG_VMAP_STACK set? I took the default, just > > > wondering if that was an especially safe thing to try. > > > > > > > > > > Yeah, looks unrelated to me too. It died down in the kmalloc code, so > > I'd have to wonder if there is some slab corruption or something? > > > > Chuck, can you do something like this on the kernel where you saw this > > panic? > > > > $ gdb vmlinux > > gdb> list *(__kmalloc_track_caller+0x1b0) > > > > It'd be good to see where this crashed. > > [cel@manet 4.9.0-rc2-00012-gf3e49aa]$ gdb vmlinux > GNU gdb (GDB) Red Hat Enterprise Linux 7.6.1-80.el7 > Copyright (C) 2013 Free Software Foundation, Inc. > License GPLv3+: GNU GPL version 3 or later <http://gnu.org/licenses/gpl.html> > This is free software: you are free to change and redistribute it. > There is NO WARRANTY, to the extent permitted by law. Type "show copying" > and "show warranty" for details. > This GDB was configured as "x86_64-redhat-linux-gnu". > For bug reporting instructions, please see: > <http://www.gnu.org/software/gdb/bugs/>... > Reading symbols from /usr/lib/modules/4.9.0-rc2-00012-gf3e49aa/vmlinux...done. > (gdb) list *(__kmalloc_track_caller+0x1b0) > 0xffffffff811ecee0 is in __kmalloc_track_caller (/home/cel/src/linux/linux-2.6/mm/slub.c:241). > 236 * Core slab cache functions > 237 *******************************************************************/ > 238 > 239 static inline void *get_freepointer(struct kmem_cache *s, void *object) > 240 { > 241 return *(void **)(object + s->offset); > 242 } > 243 > 244 static void prefetch_freepointer(const struct kmem_cache *s, void *object) > 245 { > (gdb) > > -- > Chuck Lever > > > Thanks -- I'm guessing that means that "s" is bogus there, so yeah...slab corruption. I hit some panics in the same function about a year or so ago when I was developing some fsnotify patches, and it turned out to be double- free. Wonder what's sharing that slab? -- Jeff Layton <jlayton@xxxxxxxxxx> -- To unsubscribe from this list: send the line "unsubscribe linux-nfs" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html