[Pacemaker] Error: cluster is not currently running on this node
Miha
miha at softnet.si
Tue Aug 19 08:05:14 CEST 2014
sorry, here is it:
<cluster config_version="9" name="sipproxy">
<fence_daemon/>
<clusternodes>
<clusternode name="sip1" nodeid="1">
<fence>
<method name="pcmk-method">
<device name="pcmk-redirect" port="sip1"/>
</method>
</fence>
</clusternode>
<clusternode name="sip2" nodeid="2">
<fence>
<method name="pcmk-method">
<device name="pcmk-redirect" port="sip2"/>
</method>
</fence>
</clusternode>
</clusternodes>
<cman expected_votes="1" two_node="1"/>
<fencedevices>
<fencedevice agent="fence_pcmk" name="pcmk-redirect"/>
</fencedevices>
<rm>
<failoverdomains/>
<resources/>
</rm>
</cluster>
br
miha
Dne 8/18/2014 11:33 AM, piše emmanuel segura:
> your cman /etc/cluster/cluster.conf ?
>
> 2014-08-18 7:08 GMT+02:00 Miha <miha at softnet.si>:
>> Hi Emmanuel,
>>
>> this is my config:
>>
>>
>> Pacemaker Nodes:
>> sip1 sip2
>>
>> Resources:
>> Master: ms_drbd_mysql
>> Meta Attrs: master-max=1 master-node-max=1 clone-max=2 clone-node-max=1
>> notify=true
>> Resource: p_drbd_mysql (class=ocf provider=linbit type=drbd)
>> Attributes: drbd_resource=clusterdb_res
>> Operations: monitor interval=29s role=Master (p_drbd_mysql-monitor-29s)
>> monitor interval=31s role=Slave (p_drbd_mysql-monitor-31s)
>> Group: g_mysql
>> Resource: p_fs_mysql (class=ocf provider=heartbeat type=Filesystem)
>> Attributes: device=/dev/drbd0 directory=/var/lib/mysql_drbd fstype=ext4
>> Meta Attrs: target-role=Started
>> Resource: p_ip_mysql (class=ocf provider=heartbeat type=IPaddr2)
>> Attributes: ip=XXX.XXX.XXX.XXX cidr_netmask=24 nic=eth2
>> Resource: p_mysql (class=ocf provider=heartbeat type=mysql)
>> Attributes: datadir=/var/lib/mysql_drbd/data/ user=root group=root
>> config=/var/lib/mysql_drbd/my.cnf pid=/var/run/mysqld/mysqld.pid
>> socket=/var/lib/mysql/mysql.sock binary=/usr/bin/mysqld_safe
>> additional_parameters="--bind-address=212.13.249.55 --user=root"
>> Meta Attrs: target-role=Started
>> Operations: start interval=0 timeout=120s (p_mysql-start-0)
>> stop interval=0 timeout=120s (p_mysql-stop-0)
>> monitor interval=20s timeout=30s (p_mysql-monitor-20s)
>> Clone: cl_ping
>> Meta Attrs: interleave=true
>> Resource: p_ping (class=ocf provider=pacemaker type=ping)
>> Attributes: name=ping multiplier=1000 host_list=XXX.XXX.XXX.XXXX
>> Operations: monitor interval=15s timeout=60s (p_ping-monitor-15s)
>> start interval=0s timeout=60s (p_ping-start-0s)
>> stop interval=0s (p_ping-stop-0s)
>> Resource: opensips (class=lsb type=opensips)
>> Meta Attrs: target-role=Started
>> Operations: start interval=0 timeout=120 (opensips-start-0)
>> stop interval=0 timeout=120 (opensips-stop-0)
>>
>> Stonith Devices:
>> Resource: fence_sip1 (class=stonith type=fence_bladecenter_snmp)
>> Attributes: action=off ipaddr=172.30.0.2 port=8 community=test login=snmp8
>> passwd=soft1234
>> Meta Attrs: target-role=Started
>> Resource: fence_sip2 (class=stonith type=fence_bladecenter_snmp)
>> Attributes: action=off ipaddr=172.30.0.2 port=9 community=test1
>> login=snmp8 passwd=soft1234
>> Meta Attrs: target-role=Started
>> Fencing Levels:
>>
>> Location Constraints:
>> Resource: ms_drbd_mysql
>> Constraint: l_drbd_master_on_ping
>> Rule: score=-INFINITY role=Master boolean-op=or
>> (id:l_drbd_master_on_ping-rule)
>> Expression: not_defined ping (id:l_drbd_master_on_ping-expression)
>> Expression: ping lte 0 type=number
>> (id:l_drbd_master_on_ping-expression-0)
>> Ordering Constraints:
>> promote ms_drbd_mysql then start g_mysql (INFINITY)
>> (id:o_drbd_before_mysql)
>> g_mysql then start opensips (INFINITY) (id:opensips_after_mysql)
>> Colocation Constraints:
>> g_mysql with ms_drbd_mysql (INFINITY) (with-rsc-role:Master)
>> (id:c_mysql_on_drbd)
>> opensips with g_mysql (INFINITY) (id:c_opensips_on_mysql)
>>
>> Cluster Properties:
>> cluster-infrastructure: cman
>> dc-version: 1.1.10-14.el6-368c726
>> no-quorum-policy: ignore
>> stonith-enabled: true
>> Node Attributes:
>> sip1: standby=off
>> sip2: standby=off
>>
>>
>> br
>> miha
>>
>> Dne 8/14/2014 3:05 PM, piše emmanuel segura:
>>
>>> ncomplete=10, Source=/var/lib/pacemaker/pengine/pe-warn-7.bz2): Stopped
>>> Jul 03 14:10:51 [2701] sip2 crmd: notice:
>>> too_many_st_failures: No devices found in cluster to fence
>>> sip1, giving up
>>>
>>> Jul 03 14:10:54 [2697] sip2 stonith-ng: info: stonith_command:
>>> Processed st_query reply from sip2: OK (0)
>>> Jul 03 14:10:54 [2697] sip2 stonith-ng: error: remote_op_done:
>>> Operation reboot of sip1 by sip2 for
>>> stonith_admin.cman.28299 at sip2.94474607: No such device
>>>
>>> Jul 03 14:10:54 [2697] sip2 stonith-ng: info: stonith_command:
>>> Processed st_notify reply from sip2: OK (0)
>>> Jul 03 14:10:54 [2701] sip2 crmd: notice:
>>> tengine_stonith_notify: Peer sip1 was not terminated (reboot) by
>>> sip2 for sip2: No such device
>>> (ref=94474607-8cd2-410d-bbf7-5bc7df614a50) by client
>>> stonith_admin.cman.28299
>>>
>>>
>>> :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::
>>>
>>> Sorry for the short answer, have you tested your cluster fencing ? can
>>> you show your cluster.conf xml?
>>>
>>> 2014-08-14 14:44 GMT+02:00 Miha <miha at softnet.si>:
>>>> emmanuel,
>>>>
>>>> tnx. But how to know why fancing stop working?
>>>>
>>>> br
>>>> miha
>>>>
>>>> Dne 8/14/2014 2:35 PM, piše emmanuel segura:
>>>>
>>>>> Node sip2: UNCLEAN (offline) is unclean because the cluster fencing
>>>>> failed to complete the operation
>>>>>
>>>>> 2014-08-14 14:13 GMT+02:00 Miha <miha at softnet.si>:
>>>>>> hi.
>>>>>>
>>>>>> another thing.
>>>>>>
>>>>>> On node I pcs is running:
>>>>>> [root at sip1 ~]# pcs status
>>>>>> Cluster name: sipproxy
>>>>>> Last updated: Thu Aug 14 14:13:37 2014
>>>>>> Last change: Sat Feb 1 20:10:48 2014 via crm_attribute on sip1
>>>>>> Stack: cman
>>>>>> Current DC: sip1 - partition with quorum
>>>>>> Version: 1.1.10-14.el6-368c726
>>>>>> 2 Nodes configured
>>>>>> 10 Resources configured
>>>>>>
>>>>>>
>>>>>> Node sip2: UNCLEAN (offline)
>>>>>> Online: [ sip1 ]
>>>>>>
>>>>>> Full list of resources:
>>>>>>
>>>>>> Master/Slave Set: ms_drbd_mysql [p_drbd_mysql]
>>>>>> Masters: [ sip2 ]
>>>>>> Slaves: [ sip1 ]
>>>>>> Resource Group: g_mysql
>>>>>> p_fs_mysql (ocf::heartbeat:Filesystem): Started sip2
>>>>>> p_ip_mysql (ocf::heartbeat:IPaddr2): Started sip2
>>>>>> p_mysql (ocf::heartbeat:mysql): Started sip2
>>>>>> Clone Set: cl_ping [p_ping]
>>>>>> Started: [ sip1 sip2 ]
>>>>>> opensips (lsb:opensips): Stopped
>>>>>> fence_sip1 (stonith:fence_bladecenter_snmp): Started sip2
>>>>>> fence_sip2 (stonith:fence_bladecenter_snmp): Started sip2
>>>>>>
>>>>>>
>>>>>> [root at sip1 ~]#
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> Dne 8/14/2014 2:12 PM, piše Miha:
>>>>>>
>>>>>>> Hi emmanuel,
>>>>>>>
>>>>>>> i think so, what is the best way to check?
>>>>>>>
>>>>>>> Sorry for my noob question, I have confiured this 6 mouths ago and
>>>>>>> everything was working fine till now. Now I need to find out what
>>>>>>> realy
>>>>>>> heppend beffor I do something stupid.
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> tnx
>>>>>>>
>>>>>>> Dne 8/14/2014 1:58 PM, piše emmanuel segura:
>>>>>>>> are you sure your cluster fencing is working?
>>>>>>>>
>>>>>>>> 2014-08-14 13:40 GMT+02:00 Miha <miha at softnet.si>:
>>>>>>>>> Hi,
>>>>>>>>>
>>>>>>>>> I noticed today that I am having some problem with cluster. I
>>>>>>>>> noticed
>>>>>>>>> the
>>>>>>>>> master server is offilne but still virutal ip is assigned to it and
>>>>>>>>> all
>>>>>>>>> services are running properly (for production).
>>>>>>>>>
>>>>>>>>> If I do this I am getting this notifications:
>>>>>>>>>
>>>>>>>>> [root at sip2 cluster]# pcs status
>>>>>>>>> Error: cluster is not currently running on this node
>>>>>>>>> [root at sip2 cluster]# /etc/init.d/corosync status
>>>>>>>>> corosync dead but pid file exists
>>>>>>>>> [root at sip2 cluster]# pcs status
>>>>>>>>> Error: cluster is not currently running on this node
>>>>>>>>> [root at sip2 cluster]#
>>>>>>>>> [root at sip2 cluster]#
>>>>>>>>> [root at sip2 cluster]# tailf fenced.log
>>>>>>>>> Aug 14 13:34:25 fenced cman_get_cluster error -1 112
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> The main thing is what to do now? Do "pcs start" and hope for the
>>>>>>>>> best
>>>>>>>>> or
>>>>>>>>> what?
>>>>>>>>>
>>>>>>>>> I have pasted log in pastebin: http://pastebin.com/SUp2GcmN
>>>>>>>>>
>>>>>>>>> tnx!
>>>>>>>>>
>>>>>>>>> miha
>>>>>>>>>
>>>>>>>>> _______________________________________________
>>>>>>>>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>>>>>
>>>>>>>>> Project Home: http://www.clusterlabs.org
>>>>>>>>> Getting started:
>>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>>>>> Bugs: http://bugs.clusterlabs.org
>>>>>>>>
>>>>>>>>
>>>>>> _______________________________________________
>>>>>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>>
>>>>>> Project Home: http://www.clusterlabs.org
>>>>>> Getting started:
>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>> Bugs: http://bugs.clusterlabs.org
>>>>>
>>>>>
>>>> _______________________________________________
>>>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>
>>>> Project Home: http://www.clusterlabs.org
>>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>> Bugs: http://bugs.clusterlabs.org
>>>
>>>
>>
>> _______________________________________________
>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>
>> Project Home: http://www.clusterlabs.org
>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> Bugs: http://bugs.clusterlabs.org
>
>
More information about the Pacemaker
mailing list