[Pacemaker] Different Corosync Rings for Different Nodes in Same Cluster?
Andrew Martin
amartin at xes-inc.com
Mon Jul 2 15:19:11 UTC 2012
Hi Steve,
Thanks for the clarification. Am I correct in understanding that in a complete network, corosync will automatically re-add nodes that drop out and reappear for any reason (e.g. maintenance, network connectivity loss, STONITH, etc)?
Thanks,
Andrew
----- Original Message -----
From: "Steven Dake" <sdake at redhat.com>
To: "The Pacemaker cluster resource manager" <pacemaker at oss.clusterlabs.org>
Cc: discuss at corosync.org
Sent: Friday, June 29, 2012 9:40:43 AM
Subject: Re: [Pacemaker] Different Corosync Rings for Different Nodes in Same Cluster?
On 06/29/2012 01:42 AM, Dan Frincu wrote:
> Hi,
>
> On Thu, Jun 28, 2012 at 6:13 PM, Andrew Martin <amartin at xes-inc.com> wrote:
>> Hi Dan,
>>
>> Thanks for the help. If I configure the network as I described - ring 0 as
>> the network all 3 nodes are on, ring 1 as the network only 2 of the nodes
>> are on, and using "passive" - and the ring 0 network goes down, corosync
>> will start using ring 1. Does this mean that the quorum node will appear to
>> be offline to the cluster? Will the cluster attempt to STONITH it? Once the
>> ring 0 network is available again, will corosync transition back to using it
>> as the communication ring, or will it continue to use ring 1 until it fails?
>>
>> The ideal behavior would be when ring 0 fails it then communicates over ring
>> 1, but keeps periodically checking to see if ring 0 is working again. Once
>> it is, it returns to using ring 0. Is this possible?
>
> Added corosync ML in CC as I think this is better asked here as well.
>
> Regards,
> Dan
>
>>
>> Thanks,
>>
>> Andrew
>>
>> ________________________________
>> From: "Dan Frincu" <df.cluster at gmail.com>
>> To: "The Pacemaker cluster resource manager" <pacemaker at oss.clusterlabs.org>
>> Sent: Wednesday, June 27, 2012 3:42:42 AM
>> Subject: Re: [Pacemaker] Different Corosync Rings for Different Nodes
>> in Same Cluster?
>>
>>
>> Hi,
>>
>> On Tue, Jun 26, 2012 at 9:53 PM, Andrew Martin <amartin at xes-inc.com> wrote:
>>> Hello,
>>>
>>> I am setting up a 3 node cluster with Corosync + Pacemaker on Ubuntu 12.04
>>> server. Two of the nodes are "real" nodes, while the 3rd is in standby
>>> mode
>>> as a quorum node. The two "real" nodes each have two NICs, one that is
>>> connected to a shared LAN and the other that is directly connected between
>>> the two nodes (for DRBD replication). The quorum node is only connected to
>>> the shared LAN. I would like to have multiple Corosync rings for
>>> redundancy,
>>> however I do not know if this would cause problems for the quorum node. Is
>>> it possible for me to configure the shared LAN as ring 0 (which all 3
>>> nodes
>>> are connected to) and set the rrp_mode to passive so that it will use ring
>>> 0
>>> unless there is a failure, but to also configure the direct link between
>>> the
>>> two "real" nodes as ring 1?
>>
In general I think you cannot do what you describe. Let me repeat it so
its clear:
A B C - NET #1
A B - Net #2
Where A, B are your cluster nodes, and C is your quorum node.
You want Net #1 and Net #2 to serve as redundant rings. Since C is
missing, Net #2 will automatically be detected as faulty.
The part about corosync automatically repairing nodes is correct, that
would work (If you had a complete network).
Regards
-steve
>> Short answer, yes.
>>
>> Longer answer. I have a setup with two nodes with two interfaces, one
>> is connected via a switch to the other node and one is a back-to-back
>> link for DRBD replication. In Corosync I have two rings, one that goes
>> via the switch and one via the back-to-back link (rrp_mode: active).
>> With rrp_mode: passive it should work the way you mentioned.
>>
>> HTH,
>> Dan
>>
>>>
>>> Thanks,
>>>
>>> Andrew
>>>
>>> _______________________________________________
>>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>
>>> Project Home: http://www.clusterlabs.org
>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>> Bugs: http://bugs.clusterlabs.org
>>>
>>
>>
>>
>> --
>> Dan Frincu
>> CCNA, RHCE
>>
>> _______________________________________________
>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>
>> Project Home: http://www.clusterlabs.org
>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> Bugs: http://bugs.clusterlabs.org
>>
>>
>> _______________________________________________
>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>
>> Project Home: http://www.clusterlabs.org
>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> Bugs: http://bugs.clusterlabs.org
>>
>
>
>
_______________________________________________
Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker
Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20120702/a5efc615/attachment-0003.html>
More information about the Pacemaker
mailing list