Re: SSI, Virtual Servers, ShareRoot, Etc

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



What I badly need right now is a shared root style system. Perhaps where all
nodes boot from the FC SAN using their HBA's and all have access to GFS
storage all around the network.

There are various reasons I would like to do this but one of them also
includes trying to save on power. Say I took 32 machines and was able to get
them all booting off the network without drives, then I could use a 12 drive
FC chassis as the boot server.

You could do that, but see the comment further down about needing local disks for at least scratch space. It ultimately depends on your application.

What I had worked on last year was partitioning one of these chassis into 32
partitions, one for each system but I think there is a better way and, maybe
even gaining some benefits.

Indeed. That sounds rather like you were using a SAN just for the sake of using a SAN, and taking all the disadvantages without any of the advantages.

The problem with that was that partitions were
fixed and inaccessible as individual partitions once formatted on the storage
chassis. A shared root system would be better because then I don't have to
have fixed partitions, just files. Then, each node would have it's storage
over other storage chassis on the network. This is what I would like to
achieve, so far, without success.

Not sure what you mean there, the last two sentences didn't quite parse. Can you please elaborate?

maybe before going right into SSI which seems to be quite an effort to
get running, you might want to think about "cloning" system images by
using any kind of volume manager and copy on write volumes. lvm2 for
instance does support this.

That would make things more complicated to set up AND more complicated to maintain and keep in sync afterwards. SSI isn't hard to set up. Follow the OSR howto and you'll have it up and running in no time.

The only thing I'd do differently than the howto is that I wouldn't unshare the whole of /var, but only /var/log, /var/lock and /var/run (off the top of my head. It's useful to keep things like /var/cache shared. But that's all fairly minor stuff. The howto will get you up and running.

you basically create one master image, do all the basic configuration,
such as routes, connection to your ldap server, whatever, and then just
make snapshots of that image for every node / vm you want to create.
you can make those snapshots writeable, which basically creates a
copy-on-write datafile.

Or you could just create a shared root volume after you've set up the first one, and get everything talking to that. The only minor downside is that you still need a local disk for the initrd base root (otherwise you waste about 120MB of RAM), swap and /tmp, plus any other major file systems you want unshared for performance (e.g. replicated DB copies that are replicated outside of the clustering framework).

The major upshot of SSI is that you only need to manage one file system, which means you can both use smaller disks in the nodes, and save yourself the hassle of keeping all the packages/libraries/configs in sync.

Gordan

--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

[Index of Archives]     [Corosync Cluster Engine]     [GFS]     [Linux Virtualization]     [Centos Virtualization]     [Centos]     [Linux RAID]     [Fedora Users]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite Camping]

  Powered by Linux