[Pacemaker] dependent resource reach max fail count

Michael Fung mike at 3open.org
Fri Jul 2 13:16:31 UTC 2010


I deleted:
location ms_drbd_r0-no-conn ms_drbd_r0 \
  rule $id="ms_drbd_r0-no-conn-rule" $role="Master" \
  -inf: not_defined pingd or pingd number:lte 0

and replaced with:
location ve1011-no-conn ve1011 \
  rule $id="ve1011-no-conn-rule" \
  -inf: not_defined pingd or pingd number:lte 0

Things seems to work now.

It seems to indicate that pacemaker do not shutdown dependent resources
in an orderly manner when the multi-state depended resource change
state. The dependent resources just "crashed". Am I right?


Rgds,
Michael


On 2010/7/2 下午 06:18, Michael Fung wrote:
> Hi All,
> 
> It is a 2-node Active/Passive configuration.
> 
> First, everything is normal.
> 
> Then, all network cable disconneted. On node1 and node2, ms_drbd_r0 auto
> go to Slave role. vs_fs stopped as expected, because:
> 
>   order ms_drbd_r0-b4-vz_fs inf: ms_drbd_r0:promote group_vz:start
>   group group_vz vz_fs vz_svc ve1011
> 
> Then, all cable re-connected. On node1, ms_drbd_r0 auto back to Master role.
> 
> However, vz_fs reached max fail count and pacemaker refuse to start it!
> I need to use the cleanup command to fix it **manually**.
> 
> Is this normal behavior? Can I make vz_fs not reach max fail count?
> 
> 
> Thanks in advance!
> Michael
> 
> Attached simplified configure show below:
> 
> primitive drbd_r0 ocf:linbit:drbd \
>         params drbd_resource="r0" \
>         op start interval="0" timeout="240s" \
>         op stop interval="0" timeout="100s" \
>         op monitor interval="11s" role="Master" \
>         op monitor interval="12s" role="Slave"
> 
> ms ms_drbd_r0 drbd_r0 \
>         meta master-max="1" master-node-max="1" clone-max="2"
> clone-node-max="1" notify="true" globally-unique="false"
> target-role="Started"
> 
> primitive gw ocf:pacemaker:ping \
>         params pidfile="/var/run/ping.pid" host_list="alix" \
>         op start interval="0" timeout="60s" \
>         op stop interval="0" timeout="20s" \
>         op monitor interval="10s" timeout="60s"
> 
> primitive vz_fs ocf:heartbeat:Filesystem \
>         params device="/dev/drbd0" directory="/vz" fstype="ext3" \
>         op start interval="0" timeout="60s" \
>         op stop interval="0" timeout="60s" \
>         op monitor interval="10s" timeout="40s"
> 
> group group_vz vz_fs vz_svc ve1011
> 
> clone clone_gw gw \
>         meta clone-node-max="1" clone-max="2"
> 
> location ms_drbd_r0-master-on-node1 ms_drbd_r0 \
>         rule $id="ms_drbd_r0-master-on-node1-rule" $role="Master" 50:
> #uname eq node1
> 
> location ms_drbd_r0-no-conn ms_drbd_r0 \
>         rule $id="ms_drbd_r0-no-conn-rule" $role="Master" -inf:
> not_defined pingd or pingd number:lte 0
> 
> colocation coloc_vz inf: ms_drbd_r0:Master group_vz
> 
> order ms_drbd_r0-b4-vz_fs inf: ms_drbd_r0:promote group_vz:start
> 
> property $id="cib-bootstrap-options" \
>         dc-version="1.0.8-042548a451fce8400660f6031f4da6f0223dd5dd" \
>         cluster-infrastructure="openais" \
>         expected-quorum-votes="2" \
>         stonith-enabled="false" \
>         no-quorum-policy="ignore" \
>         last-lrm-refresh="1278064262"
> rsc_defaults $id="rsc-options" \
>         resource-stickiness="100"
> 
> 
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker



More information about the Pacemaker mailing list