Hi
I have 2-nodes cluster without quorum disks.ı noticed a problem at below:
when I want to move resources to other node it is failed to relocate services to other node and again services run the orginal node.
but when I want to restart node it is ok
could you have any ideas?
From: Fabio M. Di Nitto <fdinitto@xxxxxxxxxx>
To: linux-cluster@xxxxxxxxxx
Sent: Tuesday, July 3, 2012 7:04 AM
Subject: Re: CLVM in a 3-node cluster
To: linux-cluster@xxxxxxxxxx
Sent: Tuesday, July 3, 2012 7:04 AM
Subject: Re: CLVM in a 3-node cluster
On 07/02/2012 11:39 PM, urgrue wrote:
> On 2/7/12 19:14, Digimer wrote:
>> On 07/02/2012 01:08 PM, urgrue wrote:
>>> I'm trying to set up a 3-node cluster with clvm. Problem is, one node
>>> can't access the storage, and I'm getting:
>>> Error locking on node node3: Volume group for uuid not found: <snip>
>>> whenever I try to activate the LVs on one of the working nodes.
>>>
>>> This can't be "by design", can it?
>>
>> Does pvscan show the right device? Are all nodes in the cluster? What
>> does 'cman_tool status' and 'dlm_tool ls' show?
>>
>
> Sorry, I realize now I was misleading, let me clarify:
> The third node cannot access the storage, this is by design. I have
> three datacenters but only two have access to the active storage. The
> third datacenter only has an async copy, and will only activate
> (manually) in case of a massive disaster (failure of both the other
> datacenters).
> So I deliberately have a failover domain with only node1 and node2.
> node3's function is to provide quorum, but also be able to be activated
> (manually is fine) in case of a massive disaster.
> In other words node3 is part of the cluster, but it can't see the
> storage during normal operation.
> Looking at it another way, it's kind of as if we had a 3-node cluster
> where one node had an HBA failure but is otherwise working. Surely node1
> and node2 should be able to continue running the services?
> So my question is, do I have an error somehwere, or is clvm really
> actually not able to function without all nodes being active and able to
> access storage?
CLVM requires a consistent view of the storage from all nodes in the
cluster. This is by design.
A storage failure during operations (aka you start with all nodes able
to access the storage and then downgrade) is handle correctly.
Fabio
--
Linux-cluster mailing list
Linux-cluster@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-cluster
-- Linux-cluster mailing list Linux-cluster@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/linux-cluster