[ClusterLabs] glusterfs 2 active primitives after reboot

Vladislav Bogdanov bubble at hoster-ok.com
Sun May 3 17:56:05 CEST 2015


03.05.2015 15:06:40 GMT+03:00, lukas <lukas.kostyan at gmail.com> wrote:
>Hi,
>
>i am testing a Debian-Wheezy cluster with corosync 1.4.2 and pacamaker 
>1.1.7. On the cluster glusterfs is used
>as storage (replicated) which is mounted to the webservers root 
>directory. To test the down time ping is used.
>Lets assume node1 fails. There is no problem and virtually no 
>interruption during change of one cl_IP instance
>on node2 (both run on node2). If node1 fails node2 restarts node1 with 
>stonith. But as soon as node1 is back
>online again all apache and IP instances stop and start again. This is 
>where the interruption happens which is
>around 5seconds. Is this behaviour to be expected? I thought the 
>instances running on node2 shouldnt stop at
>all when node1 gets back online again.
>
>I guess it has something to do with the primitive active on both nodes 
>(see below)? Why is the
>primitive p_apache started on the failed node?
>
>The CIB looks as follows:
>node vm-1 \
>         attributes standby="off"
>node vm-2 \
>         attributes standby="off"
>primitive IP ocf:heartbeat:IPaddr2 \
>         params ip="192.168.122.200" nic="eth0" 
>clusterip_hash="sourceip-sourceport" \
>         op monitor interval="10s"
>primitive p_apache ocf:heartbeat:apache \
>         params configfile="/etc/apache2/apache2.conf" 
>statusurl="http://localhost/server-status" \
>         op monitor interval="60" timeout="20" \
>         op start interval="0" timeout="40s" start-delay="0" \
>         meta is-managed="true"
>primitive p_fence_N1 stonith:external/libvirt \
>         params hostlist="vm-1:N1" 
>hypervisor_uri="qemu+tcp://192.168.122.1/system" 
>pcmk_reboot_action="reboot" \
>         op monitor interval="60"
>primitive p_fence_N2 stonith:external/libvirt \
>         params hostlist="vm-2:N2" 
>hypervisor_uri="qemu+tcp://192.168.122.1/system" 
>pcmk_reboot_action="reboot" \
>         op monitor interval="60"
>primitive p_gluster_mnt ocf:heartbeat:Filesystem \
>         params device="localhost:/gvolrep" directory="/var/www/html" 
>fstype="glusterfs" \
>         op monitor interval="10"
>primitive p_glusterd ocf:glusterfs:glusterd \
>         op monitor interval="30"
>primitive p_volume_gluster ocf:glusterfs:volume \
>         params volname="gvolrep" \
>         op monitor interval="10"
>clone cl_IP IP \
>         meta globally-unique="true" clone-max="2" clone-node-max="2" 
>interleave="true" \
>         params resource-stickiness="0"
>clone cl_apache p_apache \
>         meta globally-unique="true" target-role="Started"
>clone cl_gluster_mnt p_gluster_mnt \
>         meta globally-unique="true" interleave="true" 
>target-role="Started"
>clone cl_glusterd p_glusterd \
>         meta interleave="true"
>clone cl_glustervol p_volume_gluster \
>         meta interleave="true"
>location l_fence_N1 p_fence_N1 -inf: vm-1
>location l_fence_N2 p_fence_N2 -inf: vm-2
>colocation apache-with-ip inf: cl_apache cl_IP
>colocation c_apache_gluster inf: cl_apache cl_gluster_mnt
>colocation c_gluster_mnt_with_apache inf: cl_gluster_mnt cl_apache
>colocation c_glusterd_with_glustervol inf: cl_glusterd cl_glustervol
>colocation c_glustervol_with_gluster_mnt inf: cl_glustervol
>cl_gluster_mnt
>order o_apacher_after_mnt inf: cl_gluster_mnt cl_apache
>order o_gluster inf: cl_glusterd cl_glustervol cl_gluster_mnt cl_apache
>
>cl_IP
>property $id="cib-bootstrap-options" \
>         dc-version="1.1.7-ee0730e13d124c3d58f00016c3376a1de5323cff" \
>         cluster-infrastructure="openais" \
>         expected-quorum-votes="2" \
>         stonith-enabled="true" \
>         no-quorum-policy="ignore" \
>         last-lrm-refresh="1430571170" \
>         default-resource-stickiness="100"
>op_defaults $id="op-options" \
>         timeout="240s"
>################################################
>root at vm-2:~# tail -f /var/log/corosync.log | grep ERROR
>May 03 13:38:14 vm-2 pengine: [3285]: ERROR: native_create_actions: 
>Resource p_apache:1 (ocf::apache) is active on 2 nodes attempting
>recovery
>May 03 13:38:14 vm-2 pengine: [3285]: ERROR: process_pe_message: 
>Transition 30: ERRORs found during PE processing. PEngine Input stored 
>in: /var/lib/pengine/pe-error-0.bz2
>
>_______________________________________________
>Users mailing list: Users at clusterlabs.org
>http://clusterlabs.org/mailman/listinfo/users
>
>Project Home: http://www.clusterlabs.org
>Getting started:
>http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>Bugs: http://bugs.clusterlabs.org

Do you have apache started by init scripts?



More information about the Users mailing list