<div dir="ltr">Hi,<div><br></div><div>Sorry, I have managed to fix this now. I noticed in the logline:</div><div><br></div><div>A<span style="font-family:arial,sans-serif;font-size:13px">ug  6 13:26:23 ldb03 cibadmin[2140]:   notice: crm_log_args: Invoked:</span></div>
<span style="font-family:arial,sans-serif;font-size:13px">cibadmin -M -c -o status --xml-text &lt;node_state id=&quot;ldb03&quot; uname=&quot;ldb03&quot;</span><br style="font-family:arial,sans-serif;font-size:13px"><span style="font-family:arial,sans-serif;font-size:13px">ha=&quot;active&quot; in_ccm=&quot;false&quot; crmd=&quot;offline&quot; join=&quot;member&quot; expected=&quot;down&quot; crm-</span><br style="font-family:arial,sans-serif;font-size:13px">
<span style="font-family:arial,sans-serif;font-size:13px">debug-origin=&quot;manual_clear&quot; shutdown=&quot;0&quot;/&gt;</span><div><span style="font-family:arial,sans-serif;font-size:13px"><br></span></div><div><span style="font-family:arial,sans-serif;font-size:13px">the id is ldb03, not the ID of the node, 12303.</span></div>
<div><span style="font-family:arial,sans-serif;font-size:13px"><br></span></div><div><span style="font-family:arial,sans-serif;font-size:13px">I removed using: crm_node -R &quot;ldb03&quot; --force</span></div><div>and rebooted.</div>
<div><br></div><div>Nodes are now in sync.</div><div><br></div><div>Thanks,</div><div><br></div><div>Jamie.</div></div><div class="gmail_extra"><br><br><div class="gmail_quote">On Wed, Aug 6, 2014 at 2:43 PM, Jamie <span dir="ltr">&lt;<a href="mailto:thisbodydrop@gmail.com" target="_blank">thisbodydrop@gmail.com</a>&gt;</span> wrote:<br>
<blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">Hi,<br>
<br>
I have setup a 2 node cluster, using the following packages:<br>
<br>
pacemaker                           1.1.10+git20130802-1ubuntu2<br>
corosync                            2.3.3-1ubuntu1<br>
<br>
My cluster config is as so:<br>
<br>
node $id=&quot;12303&quot; ldb03<br>
node $id=&quot;12304&quot; ldb04<br>
primitive p_fence_ldb03 stonith:external/vcenter \<br>
        params VI_SERVER=&quot;10.17.248.10&quot;<br>
VI_CREDSTORE=&quot;/root/.vmware/credstore/vicredentials.xml&quot;<br>
HOSTLIST=&quot;ldb03=ldb03&quot; RESETPOWERON=&quot;0&quot; pcmk_host_check=&quot;static-list&quot;<br>
pcmk_host_list=&quot;ldb03&quot; \<br>
        op start interval=&quot;0&quot; timeout=&quot;500s&quot;<br>
primitive p_fence_ldb04 stonith:external/vcenter \<br>
        params VI_SERVER=&quot;10.17.248.10&quot;<br>
VI_CREDSTORE=&quot;/root/.vmware/credstore/vicredentials.xml&quot;<br>
HOSTLIST=&quot;ldb04=ldb04&quot; RESETPOWERON=&quot;0&quot; pcmk_host_check=&quot;static-list&quot;<br>
pcmk_host_list=&quot;ldb04&quot; \<br>
        op start interval=&quot;0&quot; timeout=&quot;500s&quot;<br>
primitive p_fs_mysql ocf:heartbeat:Filesystem \<br>
        params device=&quot;nfsserver:/LDB_Cluster1&quot; directory=&quot;/var/lib/mysql&quot;<br>
fstype=&quot;nfs&quot;<br>
options=&quot;relatime,rw,hard,nointr,rsize=32768,wsize=32768,bg,vers=3,proto=tcp<br>
&quot; \<br>
        op start interval=&quot;0&quot; timeout=&quot;60s&quot; \<br>
        op stop interval=&quot;0&quot; timeout=&quot;120s&quot; \<br>
        op monitor interval=&quot;60s&quot; timeout=&quot;60s&quot; \<br>
        meta is-managed=&quot;true&quot;<br>
primitive p_ip_1 ocf:heartbeat:IPaddr2 \<br>
        params ip=&quot;10.10.10.11&quot; cidr_netmask=&quot;25&quot; \<br>
        op monitor interval=&quot;30s&quot; \<br>
        meta target-role=&quot;Started&quot; is-managed=&quot;true&quot;<br>
primitive p_ip_2 ocf:heartbeat:IPaddr2 \<br>
        params ip=&quot;10.10.10.12&quot; cidr_netmask=&quot;25&quot; \<br>
        op monitor interval=&quot;30s&quot; \<br>
        meta target-role=&quot;Started&quot; is-managed=&quot;true&quot;<br>
primitive p_ip_3 ocf:heartbeat:IPaddr2 \<br>
        params ip=&quot;10.10.10.13&quot; cidr_netmask=&quot;25&quot; \<br>
        op monitor interval=&quot;30s&quot; \<br>
        meta target-role=&quot;Started&quot; is-managed=&quot;true&quot;<br>
primitive p_mysql ocf:heartbeat:mysql \<br>
        params datadir=&quot;/var/lib/mysql&quot; binary=&quot;/usr/bin/mysqld_safe&quot;<br>
socket=&quot;/var/run/mysqld/mysqld.sock&quot; \<br>
        op start interval=&quot;0&quot; timeout=&quot;120&quot; \<br>
        op stop interval=&quot;0&quot; timeout=&quot;120&quot; \<br>
        op monitor interval=&quot;20&quot; timeout=&quot;30&quot; \<br>
        meta target-role=&quot;Started&quot; is-managed=&quot;true&quot;<br>
group g_mysql p_fs_mysql p_mysql p_ip_1 p_ip_2 p_ip_3 \<br>
location l_fence_ldb03 p_fence_ldb03 -inf: ldb03<br>
location l_fence_ldb04 p_fence_ldb04 -inf: ldb04<br>
property $id=&quot;cib-bootstrap-options&quot; \<br>
        dc-version=&quot;1.1.10-42f2063&quot; \<br>
        cluster-infrastructure=&quot;corosync&quot; \<br>
        no-quorum-policy=&quot;ignore&quot; \<br>
        stonith-enabled=&quot;true&quot; \<br>
        stop-all-resources=&quot;false&quot; \<br>
        expected-quorum-votes=&quot;2&quot; \<br>
        last-lrm-refresh=&quot;1407325251&quot;<br>
<br>
<br>
This exact configuration has worked during the setup, but I have encountered<br>
a problem with my inactive node ldb03. Corosync shows this node as up:<br>
<br>
root@ldb03:~# corosync-cmapctl | grep members<br>
runtime.totem.pg.mrp.srp.members.12303.config_version (u64) = 0<br>
runtime.totem.pg.mrp.srp.members.12303.ip (str) = r(0) ip(10.10.10.8)<br>
runtime.totem.pg.mrp.srp.members.12303.join_count (u32) = 1<br>
runtime.totem.pg.mrp.srp.members.12303.status (str) = joined<br>
runtime.totem.pg.mrp.srp.members.12304.config_version (u64) = 0<br>
runtime.totem.pg.mrp.srp.members.12304.ip (str) = r(0) ip(10.10.10.9)<br>
runtime.totem.pg.mrp.srp.members.12304.join_count (u32) = 1<br>
runtime.totem.pg.mrp.srp.members.12304.status (str) = joined<br>
<br>
and crm status and crm node status show it as online:<br>
<br>
Last updated: Wed Aug  6 14:16:24 2014<br>
Last change: Wed Aug  6 14:02:00 2014 via crm_resource on ldb04<br>
Stack: corosync<br>
Current DC: ldb04 (12304) - partition with quorum<br>
Version: 1.1.10-42f2063<br>
2 Nodes configured<br>
7 Resources configured<br>
Online: [ ldb03 ldb04 ]<br>
<br>
root@ldb03:~# crm node status<br>
&lt;nodes&gt;<br>
  &lt;node id=&quot;12304&quot; uname=&quot;ldb04&quot;/&gt;<br>
  &lt;node id=&quot;12303&quot; uname=&quot;ldb03&quot;/&gt;<br>
&lt;/nodes&gt;<br>
<br>
<br>
but....after seeing this entry in my logs:<br>
Aug  6 13:26:23 ldb03 cibadmin[2140]:   notice: crm_log_args: Invoked:<br>
cibadmin -M -c -o status --xml-text &lt;node_state id=&quot;ldb03&quot; uname=&quot;ldb03&quot;<br>
ha=&quot;active&quot; in_ccm=&quot;false&quot; crmd=&quot;offline&quot; join=&quot;member&quot; expected=&quot;down&quot; crm-<br>
debug-origin=&quot;manual_clear&quot; shutdown=&quot;0&quot;/&gt;<br>
<br>
I noticed that cibadmin shows it as normal(offline)<br>
root@ldb03:~# crm node show<br>
ldb04(12304): normal<br>
ldb03(12303): normal(offline)<br>
<br>
The offline state is not present in anything but cibadmin. Not the cib.xml,<br>
not corosync-quorumtool and a tcpdump shows multicast traffic from both<br>
hosts.<br>
<br>
I tried (hesitantly) to delete the line using cibadmin, but I couldn&#39;t quite<br>
get the syntax right. Any tips on how to get this node to show as online and<br>
subsequently be able to run resources? Currently, when I run crm resource<br>
move, this has no effect, no errors and nothing noticeable in the logfiles<br>
either.<br>
<br>
Sorry for long thread....I can attach more logs/config if necessary.<br>
<br>
Thanks,<br>
<br>
Jamie.<br>
<br>
<br>
_______________________________________________<br>
Pacemaker mailing list: <a href="mailto:Pacemaker@oss.clusterlabs.org">Pacemaker@oss.clusterlabs.org</a><br>
<a href="http://oss.clusterlabs.org/mailman/listinfo/pacemaker" target="_blank">http://oss.clusterlabs.org/mailman/listinfo/pacemaker</a><br>
<br>
Project Home: <a href="http://www.clusterlabs.org" target="_blank">http://www.clusterlabs.org</a><br>
Getting started: <a href="http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf" target="_blank">http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf</a><br>
Bugs: <a href="http://bugs.clusterlabs.org" target="_blank">http://bugs.clusterlabs.org</a><br>
</blockquote></div><br></div>