[Pacemaker] vip can not back to active/active status after failover
WenWen
deutschland.gray at gmail.com
Tue Aug 20 11:48:19 UTC 2013
Hi,
I am doing active/active cluster with drbd pacemaker and apache.
when i start cman and pacemaker service on both nodes at the first time
everything is ok.
then I set node1 to standby then online again.
The apache and VIP can not back to active/active again.
vip and apache service only run on one node.
Becoz we set the colocation to the vip and apache resource.
is it the design?
hope somebody can help me.
Here is my configuration file.
----on node1-----
Online: [ node1.test.com node2.test.com ]
Master/Slave Set: WebDataClone [WebData]
Masters: [ node1.test.com node2.test.com ]
Clone Set: WebFSClone [WebFS]
Started: [ node1.test.com node2.test.com ]
Clone Set: WebSiteClone [WebSite]
Started: [ node2.test.com ]
Stopped: [ WebSite:1 ]
Clone Set: WebIP [ClusterIP] (unique)
ClusterIP:0 (ocf::heartbeat:IPaddr2): Started node2.test.com
ClusterIP:1 (ocf::heartbeat:IPaddr2): Started node2.test.com
stonith_fence_virsh_node1 (stonith:fence_virsh): Started node2.test.com
stonith_fence_virsh_node2 (stonith:fence_virsh): Started node1.test.com
------------------------------------
-----------on node2 -----------------------
Online: [ node1.test.com node2.test.com ]
Master/Slave Set: WebDataClone [WebData]
Masters: [ node1.test.com node2.test.com ]
Clone Set: WebFSClone [WebFS]
Started: [ node1.test.com node2.test.com ]
Clone Set: WebSiteClone [WebSite]
Started: [ node2.test.com ]
Stopped: [ WebSite:1 ]
Clone Set: WebIP [ClusterIP] (unique)
ClusterIP:0 (ocf::heartbeat:IPaddr2): Started node2.test.com
ClusterIP:1 (ocf::heartbeat:IPaddr2): Started node2.test.com
stonith_fence_virsh_node1 (stonith:fence_virsh): Started node2.test.com
stonith_fence_virsh_node2 (stonith:fence_virsh): Started node1.test.com
[root at node2 htdocs]#
-----------on node2 -----------------------
------- configuration ----
[root at node1 ~]# crm configure show
node node1.test.com \
attributes standby="off"
node node2.test.com \
attributes standby="off"
primitive ClusterIP ocf:heartbeat:IPaddr2 \
params ip="192.168.119.143" cidr_netmask="24"
clusterip_hash="sourceip" \
op monitor interval="30s"
primitive WebData ocf:linbit:drbd \
params drbd_resource="wwwdata" \
op start interval="0" timeout="240" \
op stop interval="0" timeout="100" \
op monitor interval="29s" role="Master" \
op monitor interval="31s" role="Slave"
primitive WebFS ocf:heartbeat:Filesystem \
params device="/dev/drbd0" directory="/web" fstype="gfs2" \
op start interval="0" timeout="60" \
op stop interval="0" timeout="60" \
op monitor interval="60" timeout="40"
primitive WebSite lsb:httpd \
op start interval="0" timeout="30" \
op stop interval="0" timeout="30" \
op monitor interval="30" timeout="20"
primitive stonith_fence_virsh_node1 stonith:fence_virsh \
params action="reboot" ipaddr="192.168.119.141" login="root"
identity_file="/root/.ssh/id_rsa" port="node1.test.com"
primitive stonith_fence_virsh_node2 stonith:fence_virsh \
params action="reboot" ipaddr="192.168.119.142" login="root"
identity_file="/root/.ssh/id_rsa" port="node2.test.com"
ms WebDataClone WebData \
meta master-max="2" master-node-max="1" clone-max="2"
clone-node-max="1" notify="true"
clone WebFSClone WebFS \
meta interleave="true"
clone WebIP ClusterIP \
meta globally-unique="true" clone-max="2" clone-node-max="2"
clone WebSiteClone WebSite
location l_stonith_fence_virsh_node1_noton_node1 stonith_fence_virsh_node1
-inf: node1.test.com
location l_stonith_fence_virsh_node2_noton_node2 stonith_fence_virsh_node2
-inf: node2.test.com
colocation WebSite-with-WebFS inf: WebSiteClone WebFSClone
colocation fs_on_drbd inf: WebFSClone WebDataClone:Master
colocation website-with-ip inf: WebSiteClone WebIP
order WebFS-after-WebData inf: WebDataClone:promote WebFSClone:start
order WebSite-after-WebFS inf: WebFSClone WebSiteClone
order apache-after-vip inf: WebIP WebSiteClone
property $id="cib-bootstrap-options" \
dc-version="1.1.8-7.el6-394e906" \
------------------------------------------------
----/etc/drbd.d/global_common.conf------------
global {
usage-count yes;
}
common {
handlers {
}
startup {
become-primary-on both;
}
options {
}
disk {
fencing resource-and-stonith;
}
net {
allow-two-primaries;
after-sb-0pri discard-zero-changes;
after-sb-1pri discard-secondary;
after-sb-2pri disconnect;
protocol C;
}
}
-----------------------------
----drbd resource----
resource wwwdata {
device /dev/drbd0;
disk /dev/vg02/webdata;
meta-disk internal;
on node1.test.com {
address 192.168.119.141:7789;
}
on node2.test.com {
address 192.168.119.142:7789;
}
}
---------
More information about the Pacemaker
mailing list