[Pacemaker] No communication between nodes (setup problem)

Dan Frincu df.cluster at gmail.com
Wed Jan 30 15:01:38 CET 2013


On Wed, Jan 30, 2013 at 3:27 PM, Keith Ouellette
<Keith.Ouellette at airgas.com> wrote:
> Hans,
>
>
>
>    Is the multicast port 5405 "opened" in the firewall? That has bitten me
> before.

5405 and 5404.

>
>
>
> Thanks,
>
> Keith
>
>
>
> ________________________________
> From: Hans Bert [dadeda2002 at yahoo.de]
> Sent: Wednesday, January 30, 2013 8:22 AM
> To: and k; The Pacemaker cluster resource manager
> Subject: Re: [Pacemaker] No communication between nodes (setup problem)
>
> Hi,
>
> in the meantime I modified the configuration to check if it works with
> multicast
>
> totem {
>   version: 2
>   secauth: off
>   cluster_name: mcscluster
>   interface {
>     ringnumber: 0
>     bindnetaddr: 192.168.100.0
>     mcastaddr: 239.255.1.12
>     mcastport: 5405
>     ttl: 1
>   }
> }
>
> but unfortunately it is still not working.
> I started wireshark and I can see on both hosts MC packages from both hosts.
>
> YES selinux is disabled on both nodes
>
> [root at server1 corosync]# selinuxenabled
> [root at server1 corosync]# echo $?
> 1
>
>
> Something else I found out is:
>
> [root at server1 corosync]# pcs status nodes both
> Error mapping 192.168.100.111
> Error mapping 192.168.100.112
> Corosync Nodes:
>  Online:
>  Offline: 192.168.100.111 192.168.100.112
> Pacemaker Nodes:
>  Online: server1
>  Standby:
>  Offline:
>
>
> [root at server2 corosync]# pcs status nodes both
> Error mapping 192.168.100.111
> Error mapping 192.168.100.112
> Corosync Nodes:
>  Online:
>  Offline: 192.168.100.111 192.168.100.112
> Pacemaker Nodes:
>  Online: server2
>  Standby:
>  Offline:
>
>
>
> any further hints?
>
>
> Best regards,
> Hans
>
>
>
>
> ________________________________
>
>
> Hi,
>
> It seem to be problem with network traffic.
>
> Have you tried to sniff network traffic to be sure that udp traffic reaches
> from one node to another ??
>
> Try on server1:
>
> tcpdump -i interface -p udp -s 192.168.100.112
>
> on server2:
>
> tcpdump -i interface -p udp -s 192.168.100.111
>
>
> if there will be no packet traffic, that means you have some network issue.
>
> BTW: Is SELinux enabled on nodes ??
>
> --
> Regards
> Andrew
>
>
>
> 2013/1/30 Hans Bert
>
> Hello,
>
> we had to move from Fedora 16 to Fedora 18 and wanted to set up Corosync
> with Pacemaker and PCS as management tool.
> With F16 our cluster was running pretty good, but with F18 after 5 days we
> are reaching the point were we don't have
> got ideas what might be the problem(s).
>
>
> The cluster is build of two servers (server1=192.168.100.111;
> server2=192.168.100.112)
>
> Based on the Howto for F18 with pcs we created the following corosync.conf:
>
> totem {
>   version: 2
>   secauth: off
>   cluster_name: mcscluster
>   transport: udpu
> }
>
> nodelist {
>   node {
>     ring0_addr: 192.168.100.111
>   }
>   node {
>     ring0_addr: 192.168.100.112
>   }
> }
>
> quorum {
>   provider: corosync_votequorum
> }
>
> logging {
>   fileline: off
>   to_stderr: no
>   to_logfile: yes
>   to_syslog: yes
>   logfile: /var/log/cluster/corosync.log
>   debug: on
>   timestamp: on
> }
>
>
>
> After we started the server a status check shows us:
>
>
> [root at server1 corosync]#pcs status corosync
>
> Membership information
> ----------------------
>     Nodeid      Votes Name
> 1868867776          1 server1 (local)
>
> [root at server1 ~]# pcs status
> Last updated: Wed Jan 30 10:45:17 2013
> Last change: Wed Jan 30 10:18:56 2013 via cibadmin on server1
> Stack: corosync
> Current DC: server1 (1868867776) - partition WITHOUT quorum
> Version: 1.1.8-3.fc18-394e906
> 1 Nodes configured, unknown expected votes
> 0 Resources configured.
>
>
> Online: [ server1 ]
>
> Full list of resources:
>
>
>
> And on the other server:
>
>
> [root at server2 corosync]# pcs status corosync
>
> Membership information
> ----------------------
>     Nodeid      Votes Name
> 1885644992          1 server2 (local)
>
> [root at server2 corosync]# pcs status
> Last updated: Wed Jan 30 10:44:40 2013
> Last change: Wed Jan 30 10:19:36 2013 via cibadmin on server2
> Stack: corosync
> Current DC: server2 (1885644992) - partition WITHOUT quorum
> Version: 1.1.8-3.fc18-394e906
> 1 Nodes configured, unknown expected votes
> 0 Resources configured.
>
>
> Online: [ server2 ]
>
>
>
>
>
>
> The only warnings and errors in the logfile are:
>
> [root at server1 ~]# cat /var/log/cluster/corosync.log | egrep "warning|error"
> Jan 30 10:25:59 [1608] server1       crmd:  warning: do_log:    FSA: Input
> I_DC_TIMEOUT from crm_timer_popped() received in state S_PENDING
> Jan 30 10:25:59 [1607] server1    pengine:  warning: cluster_status:    We
> do not have quorum - fencing and resource management disabled
> Jan 30 10:28:25 [1525] server1 corosync debug   [QUORUM] getinfo response
> error: 1
> Jan 30 10:40:59 [1607] server1    pengine:  warning: cluster_status:    We
> do not have quorum - fencing and resource management disabled
>
>
> root at server2 corosync]# cat /var/log/cluster/corosync.log | egrep
> "warning|error"
> Jan 30 10:27:18 [1458] server2       crmd:  warning: do_log:    FSA: Input
> I_DC_TIMEOUT from crm_timer_popped() received in state S_PENDING
> Jan 30 10:27:18 [1457] server2    pengine:  warning: cluster_status:    We
> do not have quorum - fencing and resource management disabled
> Jan 30 10:29:19 [1349] server2 corosync debug   [QUORUM] getinfo response
> error: 1
> Jan 30 10:42:18 [1457] server2    pengine:  warning: cluster_status:    We
> do not have quorum - fencing and resource management disabled
> Jan 30 10:44:36 [1349] server2 corosync debug   [QUORUM] getinfo response
> error: 1
>
>
>
>
> We have installed the following packages:
>
> corosync-2.2.0-1.fc18.i686
> corosynclib-2.2.0-1.fc18.i686
> drbd-bash-completion-8.3.13-1.fc18.i686
> drbd-pacemaker-8.3.13-1.fc18.i686
> drbd-utils-8.3.13-1.fc18.i686
> pacemaker-1.1.8-3.fc18.i686
> pacemaker-cli-1.1.8-3.fc18.i686
> pacemaker-cluster-libs-1.1.8-3.fc18.i686
> pacemaker-libs-1.1.8-3.fc18.i686
> pcs-0.9.27-3.fc18.i686
>
>
>
> Firewalls are disabled, Pinging and SSH communication is working without any
> problems.
>
> With best regards
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>
>
>
>
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>



-- 
Dan Frincu
CCNA, RHCE



More information about the Pacemaker mailing list