Re: Cannot run VMware Virtual Machines on GlusterFS

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Please let me know if this patch fixes your problem:

http://review.gluster.com/3617

Thanks for your help and patience so far!

Avati

On Mon, Jun 25, 2012 at 7:50 PM, Anand Avati <anand.avati@xxxxxxxxx> wrote:
Tomaoki, excellent debugging! Please add yourself to CC - https://bugzilla.redhat.com/show_bug.cgi?id=835336

Avati


On Sun, Jun 24, 2012 at 10:55 PM, Tomoaki Sato <tsato@xxxxxxxxxxxxx> wrote:
Avati,

Are these intended ?:
- hashcount value of 'bar'(0) is not same as 'foo/..'(2) and,
- hashcount value of 'foo'(1) is not same as 'foo/../foo'(3).


# tshark -i 1 -R nfs
Running as user "root" and group "root". This could be dangerous.
Capturing on eth0
 2.386732 192.168.1.23 -> 192.168.1.132 NFS V3 GETATTR Call, FH:0x43976ad5
 2.387772 192.168.1.132 -> 192.168.1.23 NFS V3 GETATTR Reply (Call In 7)  Directory mode:0755 uid:0 gid:0
 3.666252 192.168.1.23 -> 192.168.1.132 NFS V3 GETATTR Call, FH:0x43976ad5
 3.667112 192.168.1.132 -> 192.168.1.23 NFS V3 GETATTR Reply (Call In 17)  Directory mode:0755 uid:0 gid:0
 3.667260 192.168.1.23 -> 192.168.1.132 NFS V3 LOOKUP Call, DH:0x43976ad5/foo                  /* bar/foo */
 3.668321 192.168.1.132 -> 192.168.1.23 NFS V3 LOOKUP Reply (Call In 19), FH:0x3f9fd887
 11.386638 192.168.1.23 -> 192.168.1.132 NFS V3 GETATTR Call, FH:0x43976ad5
 11.387664 192.168.1.132 -> 192.168.1.23 NFS V3 GETATTR Reply (Call In 52)  Directory mode:0755 uid:0 gid:0
 20.386438 192.168.1.23 -> 192.168.1.132 NFS V3 GETATTR Call, FH:0x43976ad5
 20.387436 192.168.1.132 -> 192.168.1.23 NFS V3 GETATTR Reply (Call In 95)  Directory mode:0755 uid:0 gid:0
 29.382531 192.168.1.23 -> 192.168.1.132 NFS V3 GETATTR Call, FH:0x43976ad5
 29.383796 192.168.1.132 -> 192.168.1.23 NFS V3 GETATTR Reply (Call In 126)  Directory mode:0755 uid:0 gid:0
 33.666658 192.168.1.23 -> 192.168.1.132 NFS V3 LOOKUP Call, DH:0x3f9fd887/..                   /* foo/.. */
 33.668097 192.168.1.132 -> 192.168.1.23 NFS V3 LOOKUP Reply (Call In 144), FH:0x42966b36
 33.668310 192.168.1.23 -> 192.168.1.132 NFS V3 READDIRPLUS Call, FH:0x42966b36
 33.669996 192.168.1.132 -> 192.168.1.23 NFS V3 READDIRPLUS Reply (Call In 146) .. foo .
 33.670188 192.168.1.23 -> 192.168.1.132 NFS V3 LOOKUP Call, DH:0x42966b36/..                   /* bar/.. */
 33.671279 192.168.1.132 -> 192.168.1.23 NFS V3 LOOKUP Reply (Call In 148), FH:0xbc1b2900
 33.671425 192.168.1.23 -> 192.168.1.132 NFS V3 LOOKUP Call, DH:0x42966b36/foo                  /* bar/foo */
 33.672421 192.168.1.132 -> 192.168.1.23 NFS V3 LOOKUP Reply (Call In 150), FH:0x3e9ed964
20 packets captured

# egrep "nfs3_log_fh_entry_call|nfs3_log_newfh_res" /var/log/glusterfs/nfs.log | tail -8
[2012-06-25 14:28:40.090333] D [nfs3-helpers.c:1645:nfs3_log_fh_entry_call] 0-nfs-nfsv3: XID: 3d78d872, LOOKUP: args: FH: hashcount 0, exportid b2d75589-8370-4528-ab4e-b543b3abdc3b, gfid 00000000-0000-0000-0000-000000000001, name: foo                      /* bar/foo */
[2012-06-25 14:28:40.091108] D [nfs3-helpers.c:3462:nfs3_log_newfh_res] 0-nfs-nfsv3: XID: 3d78d872, LOOKUP: NFS: 0(Call completed successfully.), POSIX: 0(Success), FH: hashcount 1, exportid b2d75589-8370-4528-ab4e-b543b3abdc3b, gfid 7c4b5a51-0108-4ac9-8fd2-4b843dcb2715
[2012-06-25 14:29:10.089791] D [nfs3-helpers.c:1645:nfs3_log_fh_entry_call] 0-nfs-nfsv3: XID: 3d78d879, LOOKUP: args: FH: hashcount 1, exportid b2d75589-8370-4528-ab4e-b543b3abdc3b, gfid 7c4b5a51-0108-4ac9-8fd2-4b843dcb2715, name: ..                       /* foo/.. */
[2012-06-25 14:29:10.090872] D [nfs3-helpers.c:3462:nfs3_log_newfh_res] 0-nfs-nfsv3: XID: 3d78d879, LOOKUP: NFS: 0(Call completed successfully.), POSIX: 0(Success), FH: hashcount 2, exportid b2d75589-8370-4528-ab4e-b543b3abdc3b, gfid 00000000-0000-0000-0000-000000000001
[2012-06-25 14:29:10.093266] D [nfs3-helpers.c:1645:nfs3_log_fh_entry_call] 0-nfs-nfsv3: XID: 3d78d87b, LOOKUP: args: FH: hashcount 2, exportid b2d75589-8370-4528-ab4e-b543b3abdc3b, gfid 00000000-0000-0000-0000-000000000001, name: ..                       /* bar/.. */
[2012-06-25 14:29:10.094056] D [nfs3-helpers.c:3462:nfs3_log_newfh_res] 0-nfs-nfsv3: XID: 3d78d87b, LOOKUP: NFS: 0(Call completed successfully.), POSIX: 0(Success), FH: hashcount 3, exportid b2d75589-8370-4528-ab4e-b543b3abdc3b, gfid 6edd430d-bc57-470e-8e98-eacfe1a91040
[2012-06-25 14:29:10.094498] D [nfs3-helpers.c:1645:nfs3_log_fh_entry_call] 0-nfs-nfsv3: XID: 3d78d87c, LOOKUP: args: FH: hashcount 2, exportid b2d75589-8370-4528-ab4e-b543b3abdc3b, gfid 00000000-0000-0000-0000-000000000001, name: foo                      /* bar/foo */
[2012-06-25 14:29:10.095198] D [nfs3-helpers.c:3462:nfs3_log_newfh_res] 0-nfs-nfsv3: XID: 3d78d87c, LOOKUP: NFS: 0(Call completed successfully.), POSIX: 0(Success), FH: hashcount 3, exportid b2d75589-8370-4528-ab4e-b543b3abdc3b, gfid 7c4b5a51-0108-4ac9-8fd2-4b843dcb2715

Regards,

Tomo

Anand Avati wrote:
Tomoaki, this is very useful. I will look deeper soon.
Thanks!

Avati

On Thu, Jun 21, 2012 at 9:21 PM, Tomoaki Sato <tsato@xxxxxxxxxxxxx <mailto:tsato@xxxxxxxxxxxxx>> wrote:

   Avati,

   tshark says ...
   FH values that the linux kernel NFS server returns stays constant for every LOOKUP 'foo' but,
   FH values that the GlusterFS(NFS) returns are non-constant.

   operaions at the ESXi host:

   ~ # ./getcwd /vmfs/volumes/94925201-78f190e0/foo
   ========= sleep 30 ================
   /vmfs/volumes/94925201-78f190e0/foo
   ~ #

   tshark's output at the linux kernel NFS server:

   # tshark -i 2 -R nfs
   Running as user "root" and group "root". This could be dangerous.
   Capturing on br0
   /* chdir */
     2.056680 192.168.1.23 -> 192.168.1.254 NFS V3 GETATTR Call, FH:0x1ffd38ff
     2.056990 192.168.1.254 -> 192.168.1.23 NFS V3 GETATTR Reply (Call In 13)  Directory mode:0755 uid:0 gid:0
     9.848666 192.168.1.23 -> 192.168.1.254 NFS V3 GETATTR Call, FH:0x1ffd38ff
     9.848767 192.168.1.254 -> 192.168.1.23 NFS V3 GETATTR Reply (Call In 60)  Directory mode:0755 uid:0 gid:0
     9.848966 192.168.1.23 -> 192.168.1.254 NFS V3 LOOKUP Call, DH:0x1ffd38ff/foo
     9.849049 192.168.1.254 -> 192.168.1.23 NFS V3 LOOKUP Reply (Call In 62), FH:0xdb05b90a         <=====
     20.055508 192.168.1.23 -> 192.168.1.254 NFS V3 GETATTR Call, FH:0x1ffd38ff
     20.055702 192.168.1.254 -> 192.168.1.23 NFS V3 GETATTR Reply (Call In 103)  Directory mode:0755 uid:0 gid:0
     29.054939 192.168.1.23 -> 192.168.1.254 NFS V3 GETATTR Call, FH:0x1ffd38ff
     29.055180 192.168.1.254 -> 192.168.1.23 NFS V3 GETATTR Reply (Call In 132)  Directory mode:0755 uid:0 gid:0
     38.054338 192.168.1.23 -> 192.168.1.254 NFS V3 GETATTR Call, FH:0x1ffd38ff
     38.054583 192.168.1.254 -> 192.168.1.23 NFS V3 GETATTR Reply (Call In 151)  Directory mode:0755 uid:0 gid:0
   /* getcwd */
     39.849107 192.168.1.23 -> 192.168.1.254 NFS V3 LOOKUP Call, DH:0xdb05b90a/..
     39.849449 192.168.1.254 -> 192.168.1.23 NFS V3 LOOKUP Reply (Call In 170), FH:0x1ffd38ff
     39.849676 192.168.1.23 -> 192.168.1.254 NFS V3 READDIRPLUS Call, FH:0x1ffd38ff
     39.849833 192.168.1.254 -> 192.168.1.23 NFS V3 READDIRPLUS Reply (Call In 172) . .. foo
     39.850071 192.168.1.23 -> 192.168.1.254 NFS V3 LOOKUP Call, DH:0x1ffd38ff/foo
     39.850149 192.168.1.254 -> 192.168.1.23 NFS V3 LOOKUP Reply (Call In 174), FH:0xdb05b90a
     39.850746 192.168.1.23 -> 192.168.1.254 NFS V3 LOOKUP Call, DH:0xdb05b90a/..
     39.850814 192.168.1.254 -> 192.168.1.23 NFS V3 LOOKUP Reply (Call In 176), FH:0x1ffd38ff
     39.851014 192.168.1.23 -> 192.168.1.254 NFS V3 READDIRPLUS Call, FH:0x1ffd38ff
     39.851095 192.168.1.254 -> 192.168.1.23 NFS V3 READDIRPLUS Reply (Call In 178) . .. foo
     39.851329 192.168.1.23 -> 192.168.1.254 NFS V3 LOOKUP Call, DH:0x1ffd38ff/foo
     39.851438 192.168.1.254 -> 192.168.1.23 NFS V3 LOOKUP Reply (Call In 180), FH:0xdb05b90a            <=====

   operations at the ESXi host:

   ~ # ./getcwd /vmfs/volumes/ef172a87-e5ae817f/foo
   ========= sleep 30 ================
   getcwd: No such file or directory
   ~ #

   tshark's output at the GlusterFS(NFS) server:

   # tshark -i 1 -R nfs
   Running as user "root" and group "root". This could be dangerous.
   Capturing on eth0
   /* chdir */
     1.228396 192.168.1.23 -> 192.168.1.136 NFS V3 GETATTR Call, FH:0x43976ad5
     1.229406 192.168.1.136 -> 192.168.1.23 NFS V3 GETATTR Reply (Call In 6)  Directory mode:0755 uid:0 gid:0
     4.445894 192.168.1.23 -> 192.168.1.136 NFS V3 GETATTR Call, FH:0x43976ad5
     4.446916 192.168.1.136 -> 192.168.1.23 NFS V3 GETATTR Reply (Call In 16)  Directory mode:0755 uid:0 gid:0
     4.447099 192.168.1.23 -> 192.168.1.136 NFS V3 LOOKUP Call, DH:0x43976ad5/foo
     4.448147 192.168.1.136 -> 192.168.1.23 NFS V3 LOOKUP Reply (Call In 18), FH:0x3f9fd887     <=====
     10.228438 192.168.1.23 -> 192.168.1.136 NFS V3 GETATTR Call, FH:0x43976ad5
     10.229432 192.168.1.136 -> 192.168.1.23 NFS V3 GETATTR Reply (Call In 31)  Directory mode:0755 uid:0 gid:0
     19.228321 192.168.1.23 -> 192.168.1.136 NFS V3 GETATTR Call, FH:0x43976ad5
     19.229309 192.168.1.136 -> 192.168.1.23 NFS V3 GETATTR Reply (Call In 47)  Directory mode:0755 uid:0 gid:0
     28.228139 192.168.1.23 -> 192.168.1.136 NFS V3 GETATTR Call, FH:0x43976ad5
     28.229112 192.168.1.136 -> 192.168.1.23 NFS V3 GETATTR Reply (Call In 70)  Directory mode:0755 uid:0 gid:0
   /* getcwd */
     34.448796 192.168.1.23 -> 192.168.1.136 NFS V3 LOOKUP Call, DH:0x3f9fd887/..
     34.450119 192.168.1.136 -> 192.168.1.23 NFS V3 LOOKUP Reply (Call In 81), FH:0x42966b36
     34.450343 192.168.1.23 -> 192.168.1.136 NFS V3 READDIRPLUS Call, FH:0x42966b36
     34.452105 192.168.1.136 -> 192.168.1.23 NFS V3 READDIRPLUS Reply (Call In 83) .. foo .
     34.452311 192.168.1.23 -> 192.168.1.136 NFS V3 LOOKUP Call, DH:0x42966b36/..
     34.453464 192.168.1.136 -> 192.168.1.23 NFS V3 LOOKUP Reply (Call In 85), FH:0xbc1b2900
     34.453648 192.168.1.23 -> 192.168.1.136 NFS V3 LOOKUP Call, DH:0x42966b36/foo
     34.454677 192.168.1.136 -> 192.168.1.23 NFS V3 LOOKUP Reply (Call In 87), FH:0x3e9ed964     <======

   Regards,

   Tomo

   (2012年06月20日 16:28), Tomoaki Sato wrote:
    > Avati,
    >
    > I've tried following:
    > 1) 'esxcfg-nas -d gluster_nfs' at the ESXi host.
    > 2) 'volume set bar nfs.enable-ino32 on' at the 192.168.1.136 host.
    > 3) 'volume stop bar' and 'volume start bar' at the 192.168.1.136 host.
    > 4) 'esxcfg-nas -a -o 192.168.1.136 -s /bar gluster_nfs' at the ESXi host.
    >
    > on the ESXi host:
    >
    > ~ # uname -m
    > x86_64
    > ~ # mkdir /vmfs/volumes/ef172a87-e5ae817f/after-enable-ino32-on
    > ~ # ls -liR /vmfs/volumes/ef172a87-e5ae817f
    > /vmfs/volumes/ef172a87-e5ae817f:
    > -2118204814 drwxr-xr-x 1 root root 4096 Jun 20 07:13 after-enable-ino32-on
    > 1205893126 drwxr-xr-x 1 root root 4096 Jun 20 07:08 baz
    > -1291907235 drwx------ 1 root root 16384 Jun 6 23:41 lost+found
    >
    > /vmfs/volumes/ef172a87-e5ae817f/after-enable-ino32-on:
    >
    > /vmfs/volumes/ef172a87-e5ae817f/baz:
    > -1374929331 drwxr-xr-x 1 root root 4096 Jun 19 06:41 foo
    >
    > /vmfs/volumes/ef172a87-e5ae817f/baz/foo:
    >
    > /vmfs/volumes/ef172a87-e5ae817f/lost+found:
    > ~ # ./getcwd /vmfs/volumes/ef172a87-e5ae817f/after-enable-ino32-on
    > getcwd: No such file or directory
    > ~ #
    >
    > on the 192.168.1.136 host:
    >
    > # gluster volume info bar
    >
    > Volume Name: bar
    > Type: Distribute
    > Volume ID: b2d75589-8370-4528-ab4e-b543b3abdc3b
    > Status: Started
    > Number of Bricks: 1
    > Transport-type: tcp
    > Bricks:
    > Brick1: bar-1-private:/mnt/brick
    > Options Reconfigured:
    > diagnostics.brick-log-level: TRACE
    > diagnostics.client-log-level: TRACE
    > nfs.enable-ino32: on
    >
    > please fine attached nfs.log5.
    >
    > Regards,
    >
    > Tomo
    >
    > (2012/06/20 16:11), Anand Avati wrote:
    >> -1374929331 drwxr-xr-x 1 root root 4096 Jun 19 06:41 foo
    >>
    >> ...
    >>
    >> 2920037965 drwxr-xr-x 2 root root 4096 Jun 19 15:41 foo
    >>
    >>
    >> Ouch!
    >>
    >> -1374929331 == (int32_t) 2920037965
    >>
    >> 'uname -m' from the ESXi host please! Is it a 32bit OS? Can you try 'gluster volume set bar nfs.enable-ino32 on' and retry?
    >>
    >> Avati
    >







[Index of Archives]     [Gluster Users]     [Ceph Users]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Security]     [Bugtraq]     [Linux]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux