[Pacemaker] Filesystem primitive does not start when one of nodes is switched off
Andrew Beekhof
andrew at beekhof.net
Mon Feb 20 12:02:13 CET 2012
2012/2/19 Viacheslav Biriukov <v.v.biriukov at gmail.com>:
> Hi.
> You don't need cman in your configuration.The best way is use gfs-pcmk and
> dlm-pcmk.
Debatable.
>
>
> 16 февраля 2012 г. 12:30 пользователь emmanuel segura <emi2fast at gmail.com>
> написал:
>
>> I don't think so
>>
>> I remember gfs2 && ocfs2 on pacemaker & redhat cluster use clvm
>>
>> In redhat cluster dlm it's started with cman
>>
>> On pacemaker you should do a clone resource
>>
>>
>> 2012/2/16 diego fanesi <diego.fanesi at gmail.com>
>>>
>>> Try to set expected-quorum-votes to 1. I'm not sure but in my opinion
>>> that is the problem.
>>>
>>> Il giorno 14/feb/2012 06:43, "Богомолов Дмитрий Викторович"
>>> <beatseed at mail.ru> ha scritto:
>>>
>>>> Hi,
>>>> I have a trouble with my test configuration.
>>>> I build an Actice/Active cluster
>>>> Ubuntu(11.10)+DRBD+Cman+Pacemaker+gfs2+Xen for test purpose.
>>>> Now i am doing some tests with availability. I am try to start cluster
>>>> on one node.
>>>>
>>>> Trouble is - the Filesystem primitive ClusterFS (fs type=gfs2) does not
>>>> start when one of two nodes is switched off.
>>>>
>>>> Here my configuration:
>>>>
>>>> node blaster \
>>>> attributes standby="off"
>>>> node turrel \
>>>> attributes standby="off"
>>>> primitive ClusterData ocf:linbit:drbd \
>>>> params drbd_resource="clusterdata" \
>>>> op monitor interval="60s"
>>>> primitive ClusterFS ocf:heartbeat:Filesystem \
>>>> params device="/dev/drbd/by-res/clusterdata"
>>>> directory="/mnt/cluster" fstype="gfs2" \
>>>> op start interval="0" timeout="60s" \
>>>> op stop interval="0" timeout="60s" \
>>>> op monitor interval="60s" timeout="60s"
>>>> primitive ClusterIP ocf:heartbeat:IPaddr2 \
>>>> params ip="192.168.122.252" cidr_netmask="32"
>>>> clusterip_hash="sourceip" \
>>>> op monitor interval="30s"
>>>> primitive SSH-stonith stonith:ssh \
>>>> params hostlist="turrel blaster" \
>>>> op monitor interval="60s"
>>>> primitive XenDom ocf:heartbeat:Xen \
>>>> params xmfile="/etc/xen/xen1.example.com.cfg" \
>>>> meta allow-migrate="true" is-managed="true" target-role="Stopped"
>>>> \
>>>> utilization cores="1" mem="512" \
>>>> op monitor interval="30s" timeout="30s" \
>>>> op start interval="0" timeout="90s" \
>>>> op stop interval="0" timeout="300s"
>>>> ms ClusterDataClone ClusterData \
>>>> meta master-max="2" master-node-max="1" clone-max="2"
>>>> clone-node-max="1" notify="true"
>>>> clone ClusterFSClone ClusterFS \
>>>> meta target-role="Started" is-managed="true"
>>>> clone IP ClusterIP \
>>>> meta globally-unique="true" clone-max="2" clone-node-max="2"
>>>> clone SSH-stonithClone SSH-stonith
>>>> location prefere-blaster XenDom 50: blaster
>>>> colocation XenDom-with-ClusterFS inf: XenDom ClusterFSClone
>>>> colocation fs_on_drbd inf: ClusterFSClone ClusterDataClone:Master
>>>> order ClusterFS-after-ClusterData inf: ClusterDataClone:promote
>>>> ClusterFSClone:start
>>>> order XenDom-after-ClusterFS inf: ClusterFSClone XenDom
>>>> property $id="cib-bootstrap-options" \
>>>> dc-version="1.1.5-01e86afaaa6d4a8c4836f68df80ababd6ca3902f" \
>>>> cluster-infrastructure="cman" \
>>>> expected-quorum-votes="2" \
>>>> stonith-enabled="true" \
>>>> no-quorum-policy="ignore" \
>>>> last-lrm-refresh="1329194925"
>>>> rsc_defaults $id="rsc-options" \
>>>> resource-stickiness="100"
>>>>
>>>> Here is an $crm resource show:
>>>>
>>>> Master/Slave Set: ClusterDataClone [ClusterData]
>>>> Masters: [ turrel ]
>>>> Stopped: [ ClusterData:1 ]
>>>> Clone Set: IP [ClusterIP] (unique)
>>>> ClusterIP:0 (ocf::heartbeat:IPaddr2) Started
>>>> ClusterIP:1 (ocf::heartbeat:IPaddr2) Started
>>>> Clone Set: ClusterFSClone [ClusterFS]
>>>> Stopped: [ ClusterFS:0 ClusterFS:1 ]
>>>> Clone Set: SSH-stonithClone [SSH-stonith]
>>>> Started: [ turrel ]
>>>> Stopped: [ SSH-stonith:1 ]
>>>> XenDom (ocf::heartbeat:Xen) Stopped
>>>>
>>>> I tryed:
>>>> crm(live)resource# cleanup ClusterFSClone
>>>> Cleaning up ClusterFS:0 on turrel
>>>> Cleaning up ClusterFS:1 on turrel
>>>> Waiting for 3 replies from the CRMd... OK
>>>>
>>>> I can see only warn message in /var/log/cluster/corosync.log
>>>> Feb 14 16:25:56 turrel pengine: [1640]: WARN: unpack_rsc_op: Processing
>>>> failed op ClusterFS:0_start_0 on turrel: unknown exec error (-2)
>>>> and
>>>> Feb 14 16:25:56 turrel pengine: [1640]: WARN: common_apply_stickiness:
>>>> Forcing ClusterFSClone away from turrel after 1000000 failures (max=1000000)
>>>> Feb 14 16:25:56 turrel pengine: [1640]: WARN: common_apply_stickiness:
>>>> Forcing ClusterFSClone away from turrel after 1000000 failures (max=1000000)
>>>>
>>>> Direct me, please, what i need to check or else?
>>>>
>>>> Best regards,
>>>> Dmitriy Bogomolov
>>>>
>>>> _______________________________________________
>>>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>
>>>> Project Home: http://www.clusterlabs.org
>>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>> Bugs: http://bugs.clusterlabs.org
>>>
>>>
>>> _______________________________________________
>>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>
>>> Project Home: http://www.clusterlabs.org
>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>> Bugs: http://bugs.clusterlabs.org
>>>
>>
>>
>>
>> --
>> esta es mi vida e me la vivo hasta que dios quiera
>>
>> _______________________________________________
>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>
>> Project Home: http://www.clusterlabs.org
>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> Bugs: http://bugs.clusterlabs.org
>>
>
>
>
> --
> Viacheslav Biriukov
> BR
> http://biriukov.com
>
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>
More information about the Pacemaker
mailing list