On Fri, 24 Oct 2014 18:59:47 +0300 Trond Myklebust <trond.myklebust@xxxxxxxxxxxxxxx> wrote: > On Fri, Oct 24, 2014 at 5:57 PM, Jeff Layton > <jeff.layton@xxxxxxxxxxxxxxx> wrote: > >> @@ -1228,6 +1231,8 @@ svc_process_common(struct svc_rqst *rqstp, struct kvec *argv, struct kvec *resv) > >> dropit: > >> svc_authorise(rqstp); /* doesn't hurt to call this twice */ > >> dprintk("svc: svc_process dropit\n"); > > > > I don't think this will fix it either. I turned the above dprintk into > > a normal printk and it never fired during the test. As best I can tell, > > svc_process_common is not returning 0 when this occurs. > > OK. Is perhaps the "revisit canceled" triggering in svc_revisit()? I'm > having trouble understanding the call chain for that stuff, but it too > looks as if it can trigger some strange behaviour. > I don't think that's it either. I turned the dprintks in svc_revisit into a printks just to be sure, and they didn't fire either. Basically, I don't think we ever do anything in svc_defer for v4.1 requests, due to this at the top of it: if (rqstp->rq_arg.page_len || !rqstp->rq_usedeferral) return NULL; /* if more than a page, give up FIXME */ ...basically rq_usedeferral should be set in most cases for v4.1 requests. It gets set when processing the compound and then unset afterward. That said, I suppose you could end up deferring the request if it occurs before the pc_func gets called, but I haven't seen any evidence of that happening so far with this test. I do concur with Christoph that I've only been able to reproduce this while running on the loopback interface. If I have server and client in different VMs, then this test runs just fine. Could this be related to the changes that Neil sent in recently to make loopback mounts work better? One idea might be reasonable to backport 2aca5b869ace67 to something v3.17-ish and see whether it's still reproducible? -- Jeff Layton <jlayton@xxxxxxxxxxxxxxx> -- To unsubscribe from this list: send the line "unsubscribe linux-nfs" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html