RE: share experience migrating cluster suite from centos 5.3 to centos 5.4

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



One problem with the below workflow.
 
7. Your going to need to copy this over manually otherwise it will fail, I've fallen victim of this before. All cluster nodes need to start on the current revision of the file before you update it. I think this is a chicken and egg problem.
 
One of this things I have configured on my clusters is that all clustered services start on it's own runlevel, in my case I have cluster services running on runlevel 3 but default boot to runlelvel 2.  This allows a node to boot up and get network before racing into the cluster (ideal for wanting to find out why it got fenced and solving the problem).
 
Everything else will work as I've just done this myself (except 5 nodes). Your downtime should be quite minimal.
 
Regards,

Peter Tiggerdine
HPC & eResearch Specialist
High Performance Computing Group
Information Technology Services
University of Queensland
Phone: +61 7 3346 6634
  Fax: +61 7 3346 6630
Email: peter.tiggerdine@xxxxxxxxx

 


From: linux-cluster-bounces@xxxxxxxxxx [mailto:linux-cluster-bounces@xxxxxxxxxx] On Behalf Of Gianluca Cecchi
Sent: Tuesday, 3 November 2009 7:29 AM
To: David Teigland
Cc: linux-cluster@xxxxxxxxxx
Subject: Re: share experience migrating cluster suite from centos 5.3 to centos 5.4


On Mon, Nov 2, 2009 at 6:25 PM, David Teigland <teigland@xxxxxxxxxx> wrote:

The out-of-memory should be fixed in 5.4:

https://bugzilla.redhat.com/show_bug.cgi?id=508829

The fix for dlm_send spinning is not released yet:

https://bugzilla.redhat.com/show_bug.cgi?id=521093

Dave


Thank you so much for the feedback.
So I have to expect this freeze and possible downtime...... also if my real nmodes  a safer method could be this one below for my two nodes + quorum disk cluster?
1) shutdown and restart in single user mode of the passive node
So now the cluster is composed of only one node in 5.3 without loss of service, at the moment
2) start network and update the passive node (as in steps of the first mail)
3) reboot in single user mode of the just updated node, and test correct funcionality (without cluster)
4) shutdown again of the just updated node
5) shutdown of the active node --- NOW we have downtime (planned)
6) startup of the updated node, now in 5.4 (and with 508829 bug corrected)
This node should form the cluster with 2 votes, itself and the quorum, correct?

7) IDEA: make a dummy update to the config on this new running node, only incrementing version number by one, so that after, when the other node comes up, it gets the config....
Does it make sense or no need/no problems for this when the second node will join?

8) power on in single user mode of the node still in 5.3
9) start network on it and update system as in steps 2)
10) reboot the just updated node and let it start in single user mode to test its functionality (without cluster enabled)
11) reboot again and let it normally join the cluster

Expected result: correct join of the cluster, correct?

12) Test a relocation of the service  ----- NOW another little downtime, but to be sure that in case of need we get relocation without problems

I'm going to test this tomorrow (here half past ten pm now) after restore of initial situation with both in 5.3, so if there are any comments, they are welcome..

Thanks
Gianluca

--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster

[Index of Archives]     [Corosync Cluster Engine]     [GFS]     [Linux Virtualization]     [Centos Virtualization]     [Centos]     [Linux RAID]     [Fedora Users]     [Fedora SELinux]     [Big List of Linux Books]     [Yosemite Camping]

  Powered by Linux