Hello,
Here is the case and what done till now :
i have installed RHEL_ES_V3_update 4 on the two
nodes and the shared can be seen from the two nodes through
/dev/cciss/c0d0 i had configured each node and done with the packages
needed by oracle, then
i had made two partitions on the shared each with 100M for using as Raw1 and Raw2 for the cluster suite quorum then i had add them on the /etc/sysconfig/rawdevices file then i run service rawdevices restart command and everything run ok, then i had installed RedHat cluster suite update 4 on the two nodes and configured the raws needed there . THEN and after that i had mailed the list about the need of using the cluster suite to have a cluster running oracle RAC 10g Then the list has answered my question that i dont need the cluster suite. So , i have continued and installed the RedHat Global File System (GFS 6 ) update 4 on the two nodes and i have used the documentation in configuring it . i have made another two partition on the shared for use for Oracle ,
and i have build the configuration files for GFS and after that and before
starting Oracle installation i tried to test shutting down one of the
nodes after mounting the two gfs partitions on the shared , then an error
message appeared about the Lock_gulm server the it is still running
and it still apears for a long time without shutting down even i power
them down from the power botton on the server .
i dont understand the lock_gulm or the files of ccs sonfiguration files cause i maybe bulid an error in any of them (i think the fence file cause i didnot understand it ) . i will give you the files that i have build till now , and im using
two proliant ML250 servers and a shared of raid 5 disks remaing a 208.5 G of its
space ...... here is the files :
the pool00.cfg for the first partition
on the shared (for using oracle) :
poolname
pool00
minor 0 subpools 1 subpool 0 128 1 gfs_data pooldevice 0 0 /dev/cciss/c0d0p5 the pool01.cfg for the second partition on the shared (for using oracle) :
poolname
pool01
minor 1 subpools 1 subpool 0 128 1 gfs_data pooldevice 0 0 /dev/cciss/c0d0p6 then i had done with the pool_tool command for them and everything is ok. then i had create a directory for the ccs files on the home of the root : /root/clu
and i have put in the /root/clu the following ccs files
:
cluster.ccs :
cluster {
name = "oracluster" lock_gulm { servers = ["orat1", "orat2"] heartbeat_rate = 0.3 allowed_misses = 1 } } fence.ccs : fence_devices
{
admin { agent = "fence_manual" } } ------------------------------------------------------------
nodes.ccs :
nodes {
orat1 { ip_interfaces { eth1 = "10.0.0.2" } fence { human { admin { ipaddr = "10.0.0.2" } } } } orat2 {
ip_interfaces { eth1 = "10.0.0.3" } fence { human { admin { ipaddr = "10.0.0.2" } } } } }
--------------------------------------------------------------------------
here is the files ... i had done with them . but i dont under stand the fence file use and is it true for my case i had choosed to be a manual fence , but do i need another hardware to use for it or what ??????? Sorry for the long E-mail
but just to let the reader have info about my case correctlly.
any answer for my case please . ????????????????/ any answer about the fence and it is not
recommended to be manual , do i have to have another hardware in my case
??? power switch ___ what do it mean by power switch ....... ??
any answer for my questions ??
Regards ------------------------------------------------- Yazan
|
-- Linux-cluster@xxxxxxxxxx http://www.redhat.com/mailman/listinfo/linux-cluster