On Wed, 17 Oct 2007, Chris Johnson wrote:
I think a light just came on. I don't know how 'self healing'
would work on the client saide. That sounds like a server side deal.
Is it possible to set up two AFR servers between two RAIDs and access
them in client side AFR mode? Would that provide my failover and
'self-healing' when a failed node came back up? Somehow I think the
servers would need to talk to each other to pull this off. Or is AFR
something that can only run on one node between two file systems?
That wouldn't be as fault tolerant.
On Wed, 17 Oct 2007, Daniel van Ham Colchete wrote:
Ok, clearly missing something here. I had no idea you could do
this AFR just from the client side. That's cool. And that would
seem to allow for failover too, yes? You still need two servers.
They self heal if one comes back here? How does that work? Does
there need to be server support for this?
Well aware of the short comings. When you live off grants and
shoes strings you do what you can. Neat box, yes. Weighs a ton.
Unless you strip it you need a fork lift to move it.
Any other thoughts on how to get as much availabilty out of this
as possible with GlusterFS would be helpfull.
Chris,
so both servers are accessing the same SATABeast to export the same
filesystem? If so, AFR is not what you are looking for. AFR will try to
replicate the files to both servers and they already to this in the
back-end. If you have two different iSCSI virtual disks to each server,
them
AFR is what you are looking for.
Right now AFR will always read from the first available server. Not
splitting the read traffic. This will change with GlusterFS 1.4 (with the
HA
translator, not sure though). But you can have two AFRs, each with one
server as it's first subvolume. It would be like (doing AFR on the client
side):
=== BEGIN CLIENT SPEC FILE ===
volume s1-b1
type protocol/client
option transport-type tcp/client
option remote-host 172.16.0.1
option remote-subvolume b1
option transport-timeout 5
end-volume
volume s2-b1
type protocol/client
option transport-type tcp/client
option remote-host 172.16.0.2
option remote-subvolume b1
option transport-timeout 5
end-volume
volume s1-b2
type protocol/client
option transport-type tcp/client
option remote-host 172.16.0.1
option remote-subvolume b2
option transport-timeout 5
end-volume
volume s2-b2
type protocol/client
option transport-type tcp/client
option remote-host 172.16.0.2
option remote-subvolume b2
option transport-timeout 5
end-volume
volume s1-bn
type protocol/client
option transport-type tcp/client
option remote-host 172.16.0.1
option remote-subvolume bn
option transport-timeout 5
end-volume
volume afr1
type cluster/afr
subvolumes s1-b1 s2-b1
option replicate *:2
end-volume
volume afr2
type cluster/afr
subvolumes s2-b2 s1-b2
option replicate *:2
end-volume
volume unify
type cluster/unify
subvolumes afr1 afr2
option namespace s1-bn
option scheduler rr
option rr.limits.min-free-disk 5
end-volume
=== END CLIENT SPEC FILE ===
With this you have to replication you need, plus sharing the read traffic
between the front-end storage servers. The write performance is always
limited to the minimum write performance of each server.
Please pay attention to the fact that you have a serious
single-point-of-failure. A fire, electrical problems, human error and many
other things can happen with that single SATABeast. I would have two. I
always pair everything. But I really liked that SATABeast and having 42
disks with only 4U.
What do you think?
Best regards,
Daniel
_______________________________________________
Gluster-devel mailing list
Gluster-devel@xxxxxxxxxx
http://lists.nongnu.org/mailman/listinfo/gluster-devel
-------------------------------------------------------------------------------
Chris Johnson |Internet: johnson@xxxxxxxxxxxxxxxxxxx
Systems Administrator |Web:
http://www.nmr.mgh.harvard.edu/~johnson
NMR Center |Voice: 617.726.0949
Mass. General Hospital |FAX: 617.726.7422
149 (2301) 13th Street |Man's a kind of missing link
Charlestown, MA., 02129 USA |fondly thinking he can think. Piet Hein
-------------------------------------------------------------------------------
-------------------------------------------------------------------------------
Chris Johnson |Internet: johnson@xxxxxxxxxxxxxxxxxxx
Systems Administrator |Web: http://www.nmr.mgh.harvard.edu/~johnson
NMR Center |Voice: 617.726.0949
Mass. General Hospital |FAX: 617.726.7422
149 (2301) 13th Street |Life stinks. If you're very lucky, sometimes
Charlestown, MA., 02129 USA |it stinks a little less. Me
-------------------------------------------------------------------------------