Re: GlusterFS Example

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Mon, 30 Jul 2007, Matt Paine wrote:

> Hi Nathan,
>
> > http://www.gluster.org/docs/index.php/GlusterFS_High_Availability_Storage_with_GlusterFS
> >
> > If anyone hasthe time for a few questions:
>
> I will do my best :)

Thanks!

> > 1) In the past, my server configuration had only the local volume as in
> > http://www.gluster.org/docs/index.php/GlusterFS_Configuration_Example_for_Four_Bricks
> > Your config looks like each server has its volume and all the others.
>
> This setup (i'm assuming you are talking about the high availability
> storage with glusterfs article) is specifically for high availability,

I am referencing the first link:
http://www.gluster.org/docs/index.php/GlusterFS_High_Availability_Storage_with_GlusterFS

I understand that the 2nd link is talking about doing everything on one
server, my question is why in the first link that looks like it is talking
about 3 physical boxes and each config file lists all the volumes
(mailspool-santa2-ds, santa3-ds) where in the 2nd link a server.vol file
only has the local volume information.

> if a brick goes down then there needs to be redundancy built in to
> provide a fallback option. If it was a simple unify solution, and a
> brick goes down, then you will be missing files (hence, not highly
> available at all unless you can guarantee 100% uptime of all bricks
> under all circumstances - which as far as I know noone offers :)
>
> In that specific example, it has been chosen to provide every brick with
>   a copy of the files. In this case two bricks could fail and the
> cluster would still be up and running :). Extending this to 10 bricks,
> then you could potentially have 9 bricks fail and it would still be
> working.
>
> Point being, GlusterFS is highly configurable, you can set it up any way
> you like, more redundancy (afr), or more space (unify) or both. Your choice.
>
>
> > 2) Why do you use loopback IP vs the public IP in the first example?
>
> The first example introduces itself with this paragraph....
>
> --------8<---------------------------------
> In this article, I will show you how to configure GlusterFS to simulate
> a cluster with:
>
>      * one single mount point
>      * 4 storage nodes serving the files
>
> I said simulate since we will deploy such a cluster on only one
> computer, but you can as well set it up over physical different
> computers by changing the IP addresses.
> ------------------------->8----------------
>
> Its only a simulation, run on one computer. If you wish to use the
> example over an actual cluster of computers, you will need to change the
> IP numbers accordingly.

Understood, however in the fist link the example has the first brick as a
127 address and the other 2 bricks as 192.168.252 address. My guess is
that it was a typo since when I looked at the page this morning the 172
address was corrected to be 192.168.242.2, consistent with the rest of the
example.

> > 3) Can I change option replicate *:3 to *:2 to just have two copies? I
> > like having an extra copy, but in my application having 3 or 3 note or 4
> > for 4 node is a bit overkill.
>
> Yes. As I said its highly configurable. You can even choose to replicate
> only certain files. E.g.
>
> option replicate *.tmp:1,*.html:4,*.php:4,*.db:1,*:2
>
> In this case any .tmp or .db files will not get replicated (will sit on
> the first afr brick only), all html and php files will be mirrored
> across the first four bricks of afr, and all other files will be
> mirrorred across the first two bricks. (note all on one line, note comma
> delimited, note also no spaces before or after the comma).

Right, but is it possible to just tell the config you want to store 2
copies of everything without telling each system where the redundant
copies should be stored? I.E. I am looking for something like RAID 6,
where I have say 8 bricks with two copies of every file. If I just enter
*:2 then the spare copies will just be on the first two bricks and since
each 8 brick has the same amount of storage I will run out of space on the
first two before the others vs the redundant copies distributed over the 8
bricks.

> > 4) In my application the server is also the client, in my current config
> > when one out of 3 servers is done I can no longer write, even tho I am
> > telling my config to write to the local brick. Is there any way to fix
> > this? If a note is down I understand how I would not have access to the
> > files, but I don't want to take down the full cluster.
>
> I'm note sure sorry. Maybe one of the devs can clarify that. I know
> there was an issue with an earlier relase to do with AFR. When the first
> brick was down then afr was no longer available.

Ah... That could be it.

> > P.S. Our setup consists of 3 servers each with a 4 port video card. They
> > continually capture video from the 4 inputs and save them in /share. Each
> > box also acts as a video server accepting connections and reading any file
> > in /share. I don't care much about redundancy, but I do care that /share
> > has files from all servers, or all servers that are up anyway.
> >
> > My server.vol file, only change is the bind IP address.
> >
> > volume brick
> >         type storage/posix
> >         option directory /md0
> > end-volume
> >
> > volume server
> >         type protocol/server
> >         option transport-type tcp/server
> >         option listen-port 6996
> >         option bind-address 192.168.0.60
> >         subvolumes brick
> >         option auth.ip.brick.allow 192.168.0.*
> > end-volume
> >
> > My client.vol file, the only change is option nufa.local-volume-name
> > client0 changes to client1 or client2 for each box. The above works as
> > long as all 3 servers are up, if one is down I can no longer read or write
> > to /share. I found that I could continue to write to /md0 on each server
> > and it still was replicated correctly, however if any server is down reads
> > on /share would break. : (
>
> Could you post your client spec file? That would make it clearer to me.
> I have a feeling you are using separate client files for each of the
> clients, perhaps it would be easier for you to use a single spec file
> for your entire cluster, but we could clarify that if you post your
> client file :)
>
>
> >
> > Any help would rock!
>
> I hope that rocked at least a little :p
>
>
> Matt.
>
>
>
>
>
> _______________________________________________
> Gluster-devel mailing list
> Gluster-devel@xxxxxxxxxx
> http://lists.nongnu.org/mailman/listinfo/gluster-devel
>




[Index of Archives]     [Gluster Users]     [Ceph Users]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Security]     [Bugtraq]     [Linux]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux