Migration questions

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I'm looking for some help on a few technical questions about setting up
Gluster. I have setup a few tests using Gluster v3.1.2 on CentOS 5 with 2
bricks using distributed volumes and then again with distributed striped
volumes. The install and initial configuration is very straight forward and
simple. I'm now looking to start setting up a Gluster production environment
and have a few questions.

We currently use Lustre and would like to move to Gluster.

Current Setup:
  Our environment consists of about 200 grid nodes reading and writing files
to a ~210TB Lustre volume. In addition we have 30-40 clients mounting the
Lustre volume via 6 Lustre/NFS reshare servers. The Lustre servers and grid
node clients are CentOS 5 and the desktop clients are a mix of mostly Mac /
Linux desktops with a few Windows desktops. The desktop clients usually only
read from the volume. The grid node clients read and write to the volume
heavily, although typically each file on the volume will only be read by one
grid node at a time. The files on the volume are currently striped across
about 5 OSTs (Lustre "bricks") to help speed up the writing and reading of
the files. Our file stripe size is 1MB. Our file size ranges from ~5MB to
100GB, but average size is a 2-3 GBs. In the Lustre tradition we run RAIDed
volumes behind the scenes and worry about data integrity outside of the FS.

- I've not seen anyone running a volume of this size with this many clients
on Gluster. It seems as though it shouldn't have issues at this scale. Am I
missing anything?

- With the latest version of Gluster is there still a concept of a
"translators"? I could find little help explaining it in the 3.1 docs, but
some article alluded to it.

- In the documentation it says "For best results, you should use distributed
striped volumes only in high concurrency environments accessing very large
files." Even though in our current environment each file is typically
accessed by only 1 client at a time, wouldn't you still get some speed-ups
by striping? Besides the risk of data is there another reason why striping
is not encouraged? What do you consider "very large files"? Do our 100GB
files meet that criteria?

- How do I know if the grid nodes should be native clients? What metrics are
us to determine that choice?

- In my tests it seems that on a client I can mount the volume via NFS using
any of the "brick" machines with the same result. When mounting the volume
via NFS, which host should the client mount? If I want to mount 60 desktops
to the volume via NFS would I point them all to the same gluster server
hostname? or setup something like round-robin dns so the clients mount all
the gluster servers equally? I'm just worried that if all 60 desktops mount
the volume using the same server hostname / ip address it could cause that
server to slow to a crawl.

Thanks
Scott Barber
Senior Sysadmin
iMemories.com


[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux