Jason, the docs should run through the creation of the pool devices. They can be a bit of a labrynth though, so here is an example called "pool_cca.cfg". <----cut here----> poolname pool_cca #name of the pool/volume to create subpools 1 #how many subpools make up this pool/volume (always starts as 1) subpool 0 128 1 gfs_data #first subpool, zero indexed, 128k stripe, 1 devices pooldevice 0 0 /dev/sdb1 #physical device for pool 0, device 0 (again, zero indexed) <-end cut here --> Additional pools just need a different "poolname" and "pooldevice". NOTE, the cluster nodes need to be "seeing" the devices listed as pooldevices the same way. node1 sees the second physical disk as /dev/sdb, then third as /dev/sdc and so on. Now, if you make /dev/sdb1 about 10MB, you'll have enough space to create a cluster config pool. Then to actually use it, you need to do the following... pool_tool -c pool_cca.cfg then you can issue ... service pool start on all nodes. Just make sure all nodes have a clean view of the partition table (reboot, or issue partprobe). Once you have the cca pool created and activated, you can apply the cluster config to it... ccs_tool create /path/to/configs/ /dev/pool/pool_cca Then do a "service ccsd start" on all nodes followed by "service lock_gulmd start" on all nodes.. To check to see if things are working...do... gulm_tool nodelist nameofalockserver and you should see a list of your nodes and some info about each one. That's should be enough to get you started. to add storage for actual gfs filesystems, simply create more pools. you can also expand pools by adding subpools after creation. It's sort of a poor mans volume management if you will. It can be done to a running system and the filesystem on top of it can be expaned live as well. Anyway, hope this helps... Corey -----Original Message----- From: linux-cluster-bounces@xxxxxxxxxx [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Jason Sent: Wednesday, May 10, 2006 8:53 PM To: linux clustering Subject: Re: question about creating partitions and gfs ummm I was thinking that was the answer too, but I have no idea what the "pool" device is.. how can I tell? Jason On Wed, May 10, 2006 at 08:33:04AM -0400, Kovacs, Corey J. wrote: > Jason, I just realized what the problem is. You need to apply the > config to a "pool" > not a normal device. What do your pooll definitions look like? The > one you created for the config is where you need to point ccs_tool at > to activate the config... > > > Corey > > -----Original Message----- > From: linux-cluster-bounces@xxxxxxxxxx > [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Kovacs, Corey J. > Sent: Wednesday, May 10, 2006 8:31 AM > To: linux clustering > Subject: RE: question about creating partitions and > gfs > > Jason, couple of questions.... (And I assume you are working with > RHEL3+GFS6.0x) > > > 1. Are you actually using raw devices? if so, why? > 2. Does the device /dev/raw/raw64 actually exist on tf2? > > > GFS does not use raw devices for anything. The standard Redhat Cluster > suite does, but not GFS. GFS uses "storage pools". Also, if memory > servs me right, later versions of GFS for RHEL3 need to be told what > pools to use in the "/etc/sysconfig/gfs" config file. Used to be that > GFS just did a scan and "found" the pools, but no longer I believe. > > Hope this helps. If not, can you give more details about your config? > > > > Corey > > > -----Original Message----- > From: linux-cluster-bounces@xxxxxxxxxx > [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Jason > Sent: Tuesday, May 09, 2006 8:23 PM > To: linux clustering > Subject: Re: question about creating partitions and > gfs > > yes, both boxes are connected to the storage, its a dell powervault > 220S configured for cluster mode. > > [root@tf1 cluster]# fdisk -l /dev/sdb > > Disk /dev/sdb: 146.5 GB, 146548981760 bytes > 255 heads, 63 sectors/track, 17816 cylinders Units = cylinders of > 16065 * 512 = 8225280 bytes > > Device Boot Start End Blocks Id System > /dev/sdb1 1 2433 19543041 83 Linux > [root@tf1 cluster]# > > [root@tf2 cluster]# fdisk -l /dev/sdb > > Disk /dev/sdb: 146.5 GB, 146548981760 bytes > 255 heads, 63 sectors/track, 17816 cylinders Units = cylinders of > 16065 * 512 = 8225280 bytes > > Device Boot Start End Blocks Id System > /dev/sdb1 1 2433 19543041 83 Linux > [root@tf2 cluster]# > > > so both sides see the storage. > > on tf1, I can start ccsd fine, but on tf2, I cant, and I see May 8 > 22:00:21 > tf2 ccsd: Unable to open /dev/sdb1 (/dev/raw/raw64): No such device or > address May 8 22:00:21 tf2 ccsd: startup failed May 9 20:17:21 tf2 ccsd: > Unable to open /dev/sdb1 (/dev/raw/raw64): No such device or address > May 9 > 20:17:21 tf2 ccsd: startup failed May 9 20:17:30 tf2 ccsd: Unable to > open > /dev/sdb1 (/dev/raw/raw64): No such device or address May 9 20:17:30 > tf2 > ccsd: startup failed > [root@tf2 cluster]# > > in the logs > > Jason > > > > > On Tue, May 09, 2006 at 08:16:07AM -0400, Kovacs, Corey J. wrote: > > Jason, if IIRC, the dells internal disks show up as /dev/sd* devices. > > Do you have a shared storage device? If /dev/sdb1 is not a shared > > device, then I think you might need to take a step back and get a > > hold of a SAN of some type. If you are just playing around, there > > are ways to get some firewire drives to accept > > > > two hosts and act like a cheap shared devices. There are docs on the > > Oracle site documenting the process of setting up the drive and the > > kernel. Note, that you'll only be able to use two nodes using the > > firewire idea. > > > > Also, you should specify a partition for the command below. That > > partition can be very small. Something on the order of 10MB sounds > > right. Even that is probably way too big. Then use the rest for GFS > > storage pools. > > > > > > Corey > > > > -----Original Message----- > > From: linux-cluster-bounces@xxxxxxxxxx > > [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Jason > > Sent: Monday, May 08, 2006 9:32 PM > > To: linux-cluster@xxxxxxxxxx > > Subject: question about creating partitions and gfs > > > > so still following instructions at > > http://www.gyrate.org/archives/9 > > im at the part that says > > > > "# ccs_tool create /root/cluster /dev/iscsi/bus0/target0/lun0/part1" > > > > in my config, I have the dell PERC 4/DC cards, and I believe the > > logical drive showed up as /dev/sdb > > > > so do I need to create a partition on this logical drive with fdisk > > first before I run > > > > ccs_tool create /root/cluster /dev/sdb1 > > > > or am I totally off track here? > > > > i did ccs_tool create /root/cluster /dev/sdb and it seemed to work > > fine, but doesnt seem right.. > > > > Jason > > > > -- > > > > Linux-cluster@xxxxxxxxxx > > https://www.redhat.com/mailman/listinfo/linux-cluster > > > > -- > > > > Linux-cluster@xxxxxxxxxx > > https://www.redhat.com/mailman/listinfo/linux-cluster > > -- > ================================================ > | Jason Welsh jason@xxxxxxxxxxxxxx | > | http://monsterjam.org DSS PGP: 0x5E30CC98 | > | gpg key: http://monsterjam.org/gpg/ | > ================================================ > > -- > > Linux-cluster@xxxxxxxxxx > https://www.redhat.com/mailman/listinfo/linux-cluster > > -- > > Linux-cluster@xxxxxxxxxx > https://www.redhat.com/mailman/listinfo/linux-cluster > > -- > > Linux-cluster@xxxxxxxxxx > https://www.redhat.com/mailman/listinfo/linux-cluster -- ================================================ | Jason Welsh jason@xxxxxxxxxxxxxx | | http://monsterjam.org DSS PGP: 0x5E30CC98 | | gpg key: http://monsterjam.org/gpg/ | ================================================ -- Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster -- Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster