(no subject)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Jeroen,
Sorry for the delay.We recommend glusterfs-volgen generated configuration/volume files for gluster-3.0.x 
and gluster cli for 3.1.x


-- 
----
Cheers,
Lakshmipathi.G
FOSS Programmer.
----- Original Message -----
From: "Jeroen Koekkoek" <j.koekkoek at perrit.nl>
To: "Lakshmipathi (lakshmipathi at gluster.com)" <lakshmipathi at gluster.com>
Cc: "gluster-users at gluster.org" <gluster-users at gluster.org>
Sent: Wednesday, December 15, 2010 10:10:18 PM
Subject: RE: (no subject)

Hi Lakshmipathi,

I decoupled the client and server and tested again. The problem did not occur. That leads me to the following question: Is my original configuration a supported one? Meaning is my original configuration supposed to work? Because that configuration was really, really fast compared to the traditional client/server model.

I tested this with GlusterFS 3.0.7. I'll repeat the steps with 3.1.1 tomorrow.

Regards,
Jeroen

> -----Original Message-----
> From: Jeroen Koekkoek
> Sent: Wednesday, December 15, 2010 2:00 PM
> To: 'Lakshmipathi'
> Subject: RE: (no subject)
> 
> Hi Lakshmipathi,
> 
> I forgot to mention that I use a single volfile for both the server and
> the client, so that the client is actually the server and vice-versa.
> The same process is connected to the mount point and serving the brick
> over tcp. Below is my configuration for a single host.
> 
> ----- cut -----
> ### file: glusterfs.vol
> 
> ################################################
> ###  GlusterFS Server and Client Volume File  ##
> ################################################
> 
> volume posix
>   type storage/posix
>   option directory /var/vmail_local
> end-volume
> 
> volume local_brick_mta1
>   type features/locks
>   subvolumes posix
> end-volume
> 
> volume server
>   type protocol/server
>   option transport-type tcp
>   option transport.socket.bind-address 172.16.104.21
>   option auth.addr.local_brick_mta1.allow 172.16.104.22
>   subvolumes local_brick_mta1
> end-volume
> 
> volume remote_brick
>   type protocol/client
>   option transport-type tcp
>   option remote-host 172.16.104.22
>   option remote-subvolume local_brick_mta2 end-volume
> 
> volume afr
>   type cluster/replicate
> #  option read-subvolume local_brick
>   subvolumes remote_brick local_brick_mta1 end-volume
> 
> volume writebehind
>   type performance/write-behind
>   option cache-size 1MB
>   subvolumes afr
> end-volume
> 
> volume quickread
>   type performance/quick-read
>   option cache-timeout 1
>   option max-file-size 1MB
>   subvolumes writebehind
> end-volume
> ----- /cut -----
> 
> Regards,
> Jeroen
> 
> > -----Original Message-----
> > From: Lakshmipathi [mailto:lakshmipathi at gluster.com]
> > Sent: Wednesday, December 15, 2010 10:34 AM
> > To: Jeroen Koekkoek
> > Subject: Re: (no subject)
> >
> > Hi Jeroen Koekkoek,
> > We are unable to  reproduce your issue with 3.1.1.
> >
> > Steps -
> > setup 2 afr server and mount it on 2 clients.
> >
> > client1-mntpt#touch file.txt
> >
> > this file avail. on both client mounts - verified with ls command.
> > client1-mntpt#ls -l file.txt
> > client2-mntpt#ls -l file.txt
> >
> > Now unmounted client2.
> > umount <mntpt-client>
> >
> > now removed the file from client1.
> > client1-mntpt#rm file.txt
> >
> > then mount client2 again and did a ls - client2-mntpt#ls -l
> >
> > file.txt is not available on both clients now,as expected.
> >
> >
> > If you are still facing this issue,sent us the server and client logs
> > along with exact steps to reproduce this issue.
> > Thanks.
> >
> >
> >
> > --
> > ----
> > Cheers,
> > Lakshmipathi.G
> > FOSS Programmer.
> >
> > ----- Original Message -----
> > From: "Jeroen Koekkoek" <j.koekkoek at perrit.nl>
> > To: gluster-users at gluster.org
> > Sent: Wednesday, December 15, 2010 1:05:50 PM
> > Subject: (no subject)
> >
> > Hi,
> >
> > I have a question regarding glusterfs and replicate. I have a two node
> > setup. The following problem arises: if I create a file on the mount
> > point, then unmount gfs on the 2nd machine, remove the file from the
> > 1st (through the mount point) and brind the mount point on the 2nd
> > machine back up. The file is removed (from the 2nd) if I `ls` the
> > mount point on the 1st machine, the file is created (on the 1st) if I
> > `ls` the mount point on the 2nd.
> >
> > If I update the file instead of removing it, everything goes fine. The
> > file is up-to-date on both machines.
> >
> > I looked at the .landfill directory, but that is only used in a self-
> > heal situation. Is there a way I can work around this? Maybe using the
> > trash translator?
> >
> > Best regards,
> > Jeroen Koekkoek
> > _______________________________________________
> > Gluster-users mailing list
> > Gluster-users at gluster.org
> > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users




[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux