We'd need to know how your storage is configured.
Everything from how storage is connected to the system (if it's
external) or if you're using an internal RAID controller. What sort of
mirroring are you using, how many locally connected HDD drives do you
have, how are the clustered file systems connected to this system
(iSCSI, Fibre Channel, something else), the contents of /etc/fstab,
/etc/cluster/cluster.conf, output from the lvm commands (IIRC, #pvs,
#lvs and #vgs will work on RHEL to display any LVM configuration
hightlights). Also, what version of the cluster stack are you using (and
what components - things like chkconfig --list, rpm -qa | grep cman,
etc, etc). There are a couple of documents on the web that will provide
you with examples/suggestions for the types of data you can/should
collect before reporting a problem with the Red Hat Cluster Stack.
Also, what caused you to change the drive, how did it get changed and
was anything done as part of that change in order to get the system back
online again.
If you don't know what you need to look at, I'm not sure you're the
optimal person to try and fix this problem. Data recovery can be a
fairly low-level thing in the right (wrong) set of circumstances.
// Thomas
On Thu, 06 Jan 2011 19:51:18 +0000, Luis Cebamanos <luiceur@xxxxxxxxx>
wrote:
Well, there are not backups of that valuable lost data but the
cluster was using disk mirroring but no one has a clue of how to take
advantage of that. Nobody here is a cluster expert and that has been
the problem I guess.
We haven't really "touch" any important system file, after physically
installed the hard disk, we realized that the cluster wasn't properly
working. We rebooted the cluster without the old disks and that has
been the result.
Worst scenario, we will need to call an expert, but we think it can
not be a big deal as I said, we haven't modified the previous
configuration...
On 01/06/2011 07:44 PM, Digimer wrote:
On 01/06/2011 02:36 PM, Luis Cebamanos wrote:
Is a cluster with 16 nodes and I suspect the problem is in the head
node:
$cat /proc/version
Linux version 2.6.11.4-21.11-smp (geeko@buildhost) (gcc version
3.3.5
20050117 (prerelease) (SUSE Linux)) #1 SMP Thu Feb 2 20:54:26 GMT
2006
df -T
Filesystem Type 1K-blocks Used Available Use% Mounted on
/dev/sda2 ext3 33032228 19971720 11382520 64% /
tmpfs tmpfs 1027632 0 1027632 0% /dev/shm
/dev/sda1 ext3 124427 9412 108591 8% /boot
/dev/sda6 ext3 2063504 33820 1924864 2% /tmp
/dev/sda9 ext3 166698068 119970708 38259504 76% /users
/dev/sda8 ext3 32250392 966320 29645848 4% /usr/local
/dev/sda7 ext3 2063504 901804 1056880 47% /var
We were trying to install new hard drives to the system but
something
that we don't know went wrong and it ended up in almost 4 years of
work
lost!!!
Please, let me know what else can I do to be able to get the data
back!
Best
It's a bit late now, but I suppose you don't have backups?
As for how to help, what you provided was only marginally helpful.
We
need a much more extensive overview of your setup and configuration,
versions, etc. before we can have any idea if we can help.
In the short term, don't try anything yourself without careful
thought.
If the data is very valuable, consider hiring a data recovery firm
near
you who can come and look at your setup.
--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster
--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster