RE: question about creating partitions and gfs

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Yes, you need to create pool devices for all things gfs, the first of which 
is the cluster configuration archive. You'll need to make more for actual GFS
filesystems you want to create. You can think of pools as cluster aware
volumes.
Just as in LVM, pools relate to volumes which relate to "presented devices".

Make sense?  Good luck!


Corey

 

-----Original Message-----
From: linux-cluster-bounces@xxxxxxxxxx
[mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Jason
Sent: Thursday, May 11, 2006 9:52 PM
To: linux clustering
Subject: Re:  question about creating partitions and gfs

ok, so reading the docs and your example, they reference /dev/sdb1 this is
still the 10 meg partition that i create with fdisk.. right?
then what about the rest of the disk? do I need to reference it as a
pooldevice as well?
i.e. 
/dev/sdb1   <-10 meg partition
/dev/sdb2   <--- rest of logical disk ??

Jason

On Thu, May 11, 2006 at 07:16:14AM -0400, Kovacs, Corey J. wrote:
> Jason, the docs should run through the creation of the pool devices. 
> They can be a bit of a labrynth though, so here is an example called 
> "pool_cca.cfg".
> 
> 
> <----cut here---->
> poolname pool_cca			#name of the pool/volume to create
> subpools 1				#how many subpools make up this
> pool/volume (always starts as 1)
> subpool 0 128 1 gfs_data	#first subpool, zero indexed, 128k stripe, 1
> devices
> pooldevice 0 0 /dev/sdb1	#physical device for pool 0, device 0 (again,
> zero indexed)
> <-end cut here -->
> 
> Additional pools just need a different "poolname" and "pooldevice". 
> 
> NOTE, the cluster nodes need to be "seeing" the devices listed as 
> pooldevices the same way. node1 sees the second physical disk as 
> /dev/sdb, then third as /dev/sdc and so on.
> 
> 
> Now, if you make /dev/sdb1 about 10MB, you'll have enough space to 
> create a cluster config pool. Then to actually use it, you need to do 
> the following...
> 
> pool_tool -c pool_cca.cfg
> 
> then you can issue ...
> 
> service pool start
> 
> on all nodes. Just make sure all nodes have a clean view of the 
> partition table (reboot, or issue partprobe).
> 
> Once you have the cca pool created and activated, you can apply the 
> cluster config to it...
> 
> ccs_tool create /path/to/configs/   /dev/pool/pool_cca
> 
> Then do a "service ccsd start" on all nodes followed by "service 
> lock_gulmd start"
> on all nodes..
> 
> To check to see if things are working...do...
> 
> gulm_tool nodelist nameofalockserver
> 
> and you should see a list of your nodes and some info about each one.
> 
> That's should be enough to get you started. to add storage for actual 
> gfs filesystems, simply create more pools. you can also expand pools 
> by adding subpools after creation. It's sort of a poor mans volume 
> management if you will. It can be done to a running system and the 
> filesystem on top of it can be expaned live as well.
> 
> 
> Anyway, hope this helps...
> 
> 
> Corey
> 
> 
> -----Original Message-----
> From: linux-cluster-bounces@xxxxxxxxxx 
> [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Jason
> Sent: Wednesday, May 10, 2006 8:53 PM
> To: linux clustering
> Subject: Re:  question about creating partitions and 
> gfs
> 
> ummm I was thinking that was the answer too, but I have no idea what 
> the "pool" device is..
> how can I tell?
> 
> Jason
> 
> 
> On Wed, May 10, 2006 at 08:33:04AM -0400, Kovacs, Corey J. wrote:
> > Jason, I just realized what the problem is. You need to apply the 
> > config to a "pool"
> > not a normal device.  What do your pooll definitions look like? The 
> > one you created for the config is where you need to point ccs_tool 
> > at to activate the config...
> > 
> > 
> > Corey
> > 
> > -----Original Message-----
> > From: linux-cluster-bounces@xxxxxxxxxx 
> > [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Kovacs, Corey J.
> > Sent: Wednesday, May 10, 2006 8:31 AM
> > To: linux clustering
> > Subject: RE:  question about creating partitions and 
> > gfs
> > 
> > Jason, couple of questions.... (And I assume you are working with
> > RHEL3+GFS6.0x)
> > 
> > 
> > 1. Are you actually using raw devices? if so, why? 
> > 2. Does the device /dev/raw/raw64 actually exist on tf2?
> > 
> > 
> > GFS does not use raw devices for anything. The standard Redhat 
> > Cluster suite does, but not GFS. GFS uses "storage pools".  Also, if 
> > memory servs me right, later versions of GFS for RHEL3 need to be 
> > told what pools to use in the "/etc/sysconfig/gfs" config file. Used 
> > to be that GFS just did a scan and "found" the pools, but no longer I
believe.
> > 
> > Hope this helps. If not, can you give more details about your config? 
> > 
> > 
> > 
> > Corey
> > 
> > 
> > -----Original Message-----
> > From: linux-cluster-bounces@xxxxxxxxxx 
> > [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Jason
> > Sent: Tuesday, May 09, 2006 8:23 PM
> > To: linux clustering
> > Subject: Re:  question about creating partitions and 
> > gfs
> > 
> > yes, both boxes are connected to the storage, its a dell powervault 
> > 220S configured for cluster mode.
> > 
> > [root@tf1 cluster]#  fdisk -l /dev/sdb
> > 
> > Disk /dev/sdb: 146.5 GB, 146548981760 bytes
> > 255 heads, 63 sectors/track, 17816 cylinders Units = cylinders of
> > 16065 * 512 = 8225280 bytes
> > 
> >    Device Boot    Start       End    Blocks   Id  System
> > /dev/sdb1             1      2433  19543041   83  Linux
> > [root@tf1 cluster]#
> > 
> > [root@tf2 cluster]# fdisk -l /dev/sdb
> > 
> > Disk /dev/sdb: 146.5 GB, 146548981760 bytes
> > 255 heads, 63 sectors/track, 17816 cylinders Units = cylinders of
> > 16065 * 512 = 8225280 bytes
> > 
> >    Device Boot    Start       End    Blocks   Id  System
> > /dev/sdb1             1      2433  19543041   83  Linux
> > [root@tf2 cluster]#
> > 
> > 
> > so both sides see the storage.  
> > 
> > on tf1, I can start ccsd fine, but on tf2, I cant, and I see May  8
> > 22:00:21
> > tf2 ccsd: Unable to open /dev/sdb1 (/dev/raw/raw64): No such device 
> > or address May  8 22:00:21 tf2 ccsd: startup failed May  9 20:17:21 tf2
ccsd:
> > Unable to open /dev/sdb1 (/dev/raw/raw64): No such device or address 
> > May  9
> > 20:17:21 tf2 ccsd: startup failed May  9 20:17:30 tf2 ccsd: Unable 
> > to open
> > /dev/sdb1 (/dev/raw/raw64): No such device or address May  9 
> > 20:17:30
> > tf2
> > ccsd: startup failed
> > [root@tf2 cluster]#
> > 
> > in the logs
> > 
> > Jason
> > 
> > 
> > 
> > 
> > On Tue, May 09, 2006 at 08:16:07AM -0400, Kovacs, Corey J. wrote:
> > > Jason, if IIRC, the dells internal disks show up as /dev/sd* devices. 
> > > Do you have a shared storage device? If /dev/sdb1 is not a shared 
> > > device, then I think you might need to take a step back and get a 
> > > hold of a SAN of some type. If you are just playing around, there 
> > > are ways to get some firewire drives to accept
> > > 
> > > two hosts and act like a cheap shared devices. There are docs on 
> > > the Oracle site documenting the process of setting up the drive 
> > > and the kernel. Note, that you'll only be able to use two nodes 
> > > using the firewire idea.
> > > 
> > > Also, you should specify a partition for the command below. That 
> > > partition can be very small. Something on the order of 10MB sounds 
> > > right. Even that is probably way too big. Then use the rest for 
> > > GFS storage pools.
> > > 
> > > 
> > > Corey
> > > 
> > > -----Original Message-----
> > > From: linux-cluster-bounces@xxxxxxxxxx 
> > > [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Jason
> > > Sent: Monday, May 08, 2006 9:32 PM
> > > To: linux-cluster@xxxxxxxxxx
> > > Subject:  question about creating partitions and 
> > > gfs
> > > 
> > > so still following instructions at
> > > http://www.gyrate.org/archives/9
> > > im at the part that says
> > > 
> > > "# ccs_tool create /root/cluster /dev/iscsi/bus0/target0/lun0/part1"
> > > 
> > > in my config, I have the dell PERC 4/DC cards, and I believe the 
> > > logical drive showed up as /dev/sdb
> > > 
> > > so do I need to create a partition on this logical drive with 
> > > fdisk first before I run
> > > 
> > >  ccs_tool create /root/cluster  /dev/sdb1
> > > 
> > > or am I totally off track here?
> > > 
> > > i did ccs_tool create /root/cluster /dev/sdb and it seemed to work 
> > > fine, but doesnt seem right..
> > > 
> > > Jason
> > > 
> > > --
> > > 
> > > Linux-cluster@xxxxxxxxxx
> > > https://www.redhat.com/mailman/listinfo/linux-cluster
> > > 
> > > --
> > > 
> > > Linux-cluster@xxxxxxxxxx
> > > https://www.redhat.com/mailman/listinfo/linux-cluster
> > 
> > --
> > ================================================
> > |    Jason Welsh   jason@xxxxxxxxxxxxxx        |
> > | http://monsterjam.org    DSS PGP: 0x5E30CC98 |
> > |    gpg key: http://monsterjam.org/gpg/       |
> > ================================================
> > 
> > --
> > 
> > Linux-cluster@xxxxxxxxxx
> > https://www.redhat.com/mailman/listinfo/linux-cluster
> > 
> > --
> > 
> > Linux-cluster@xxxxxxxxxx
> > https://www.redhat.com/mailman/listinfo/linux-cluster
> > 
> > --
> > 
> > Linux-cluster@xxxxxxxxxx
> > https://www.redhat.com/mailman/listinfo/linux-cluster
> 
> --
> ================================================
> |    Jason Welsh   jason@xxxxxxxxxxxxxx        |
> | http://monsterjam.org    DSS PGP: 0x5E30CC98 |
> |    gpg key: http://monsterjam.org/gpg/       |
> ================================================
> 
> --
> 
> Linux-cluster@xxxxxxxxxx
> https://www.redhat.com/mailman/listinfo/linux-cluster
> 
> --
> 
> Linux-cluster@xxxxxxxxxx
> https://www.redhat.com/mailman/listinfo/linux-cluster

--
================================================
|    Jason Welsh   jason@xxxxxxxxxxxxxx        |
| http://monsterjam.org    DSS PGP: 0x5E30CC98 |
|    gpg key: http://monsterjam.org/gpg/       |
================================================

--

Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

--

Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

[Index of Archives]     [Corosync Cluster Engine]     [GFS]     [Linux Virtualization]     [Centos Virtualization]     [Centos]     [Linux RAID]     [Fedora Users]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite Camping]

  Powered by Linux