What is the pattern other NFS client is writing to the file? Can't it be a legitimate NUL by any chance? On Thu, Aug 16, 2012 at 1:22 PM, Ranjan Sinha <rnjn.sinha@xxxxxxxxx> wrote: > On Thu, Aug 16, 2012 at 1:00 PM, Rajat Sharma <fs.rajat@xxxxxxxxx> wrote: >> Correct me if I am reading something wrong, in your program listing, >> while printing the buffer you are passing a total_count variable, >> while vfs_read returned value is collected in count variable. >> >> debug_dump("Read buffer", buf, total_count); > > My apologies. Please read that as count only. A typo in the listing. > >> >> One suggestion, please fill up buf with some fixed known pattern >> before vfs_read. > > I tried that as well. It still comes out as ASCII NUL. > >> >>> We have also noticed that the expected increase (inc) and the size >> returned in (vfs_read()) is different. >> >> There is nothing which is blocking updates to file size between >> vfs_getattr() and vfs_read(), right? no locking? > > No locking. On second thoughts I think this is ok since more data could be > available between the calls to vfs_getattr and vfs_read as the other NFS client > is continuously writing to that file. > > -- > Ranjan > > >> >> -Rajat >> >> On Thu, Aug 16, 2012 at 12:01 PM, Ranjan Sinha <rnjn.sinha@xxxxxxxxx> wrote: >>> Hi, >>> >>> On Tue, Aug 14, 2012 at 4:19 PM, Rajat Sharma <fs.rajat@xxxxxxxxx> wrote: >>>> Try mounting with noac nfs mount option to disable attribute caching. >>>> >>>> ac / noac >>>> >>>> "Selects whether the client may cache file attributes. If neither >>>> option is specified (or if ac is specified), the client caches file >>>> attributes." >>> >>> i don't think this is because of attribute caching. The size does change and >>> that is why we go to the read call (think of this is a simplified case of >>> tail -f). The only problem is that sometimes when we read we get ASCII NUL bytes >>> at the end. If we read the same block again, we get the correct data. >>> >>> In addition, we cannot force specific mount options in actual deployment >>> scenarios. >>> >>> >>> <edit> >>> >>>>> On Tue, Aug 14, 2012 at 5:10 PM, Ranjan Sinha <rnjn.sinha@xxxxxxxxx> wrote: >>>>> > For now, /etc/export file has the following setting >>>>> > *(rw,sync,no_root_squash) >>>>> >>>>> hm, AFAIK that means synchronous method is selected. So, >>>>> theoritically, if there is no further data, the other end of NFS >>>>> should just wait. >>>>> >>>>> Are you using blocking or non blocking read, btw? Sorry, i am not >>>>> really that good reading VFS code... >>>>> >>> >>> This is a blocking read call. I think this is not because there is no data, >>> rather somehow the updated data is not present in the VM buffers but the >>> inode size has changed. As I just said, if we read the file again from the >>> exact same location, we get the actual contents. Though after going through the >>> code I don't understand how is this possible. >>> >>>>> > On client side we have not specified any options explicitly. This is >>>>> > from /proc/mounts entry >>>>> > >rw,vers=3,rsize=32768,wsize=32768,hard,proto=tcp,timeo=600,retrans=2,sec=sys >>>>> >>>>> hm, not sure, maybe in your case, read and write buffer should be >>>>> reduced so any new data should be transmitted ASAP. I was inspired by >>>>> bufferbloat handling, but maybe I am wrong here somewhere.... >>>>> >>> >>> -- >>> Regards, >>> Ranjan _______________________________________________ Kernelnewbies mailing list Kernelnewbies@xxxxxxxxxxxxxxxxx http://lists.kernelnewbies.org/mailman/listinfo/kernelnewbies