[Pacemaker] clone ip definition and location stops my resources...
Dejan Muhamedagic
dejanmm at fastmail.fm
Tue May 11 09:58:48 UTC 2010
Hi,
On Tue, May 11, 2010 at 11:15:49AM +0200, Gianluca Cecchi wrote:
> On Mon, May 10, 2010 at 4:39 PM, Vadym Chepkov <vchepkov at gmail.com> wrote:
>
> > # crm ra meta ping
> >
> > name (string, [undef]): Attribute name
> > The name of the attributes to set. This is the name to be used in the
> > constraints.
> >
> > By default is "pingd", but you are checking against pinggw
> >
> > I suggest you do not change name though, but adjust your location
> > constraint to use pingd instead.
> > crm_mon only notices "pingd" at the moment whenn you pass -f argument: it's
> > hardcoded
> >
> >
> > On Mon, May 10, 2010 at 9:34 AM, Gianluca Cecchi <
> > gianluca.cecchi at gmail.com> wrote:
> >
> >> Hello,
> >> using pacemaker 1.0.8 on rh el 5 I have some problems understanding the
> >> way ping clone works to setup monitoring of gw... even after reading docs...
> >>
> >> As soon as I run:
> >> crm configure location nfs-group-with-pinggw nfs-group rule -inf:
> >> not_defined pinggw or pinggw lte 0
> >>
> >> the resources go stopped and don't re-start....
> >>
> >> [snip]
> >>
> >
> hem...
> I changed the location line so that now I have:
> primitive pinggw ocf:pacemaker:ping \
> params host_list="192.168.101.1" multiplier="100" \
> op start interval="0" timeout="90" \
> op stop interval="0" timeout="100"
>
> clone cl-pinggw pinggw \
> meta globally-unique="false"
>
> location nfs-group-with-pinggw nfs-group \
> rule $id="nfs-group-with-pinggw-rule" -inf: not_defined pingd or pingd lte 0
>
> But now nothing happens if I run for example
> iptables -A OUTPUT -p icmp -d 192.168.101.1 -j REJECT (or DROP)
> in the node where nfs-group is running.....
>
> Do I have to name the primitive itself to pingd????
I guess not.
> It seems that the binary /bin/ping is not accessed at all (with ls -lu ...)
Then the ping resource is not running...
> Or do I have to change the general property I previously define to avoide
> failback:
> rsc_defaults $id="rsc-options" \
> resource-stickiness="100"
>
> crm_mon -f -r gives:
> Online: [ ha1 ha2 ]
>
> Full list of resources:
>
> SitoWeb (ocf::heartbeat:apache): Started ha1
> Master/Slave Set: NfsData
> Masters: [ ha1 ]
> Slaves: [ ha2 ]
> Resource Group: nfs-group
> ClusterIP (ocf::heartbeat:IPaddr2): Started ha1
> lv_drbd0 (ocf::heartbeat:LVM): Started ha1
> NfsFS (ocf::heartbeat:Filesystem): Started ha1
> nfssrv (ocf::heartbeat:nfsserver): Started ha1
> nfsclient (ocf::heartbeat:Filesystem): Started ha2
> Clone Set: cl-pinggw
> Started: [ ha2 ha1 ]
... but according to this it is.
> Migration summary:
> * Node ha1: pingd=100
> * Node ha2: pingd=100
Do you see the attribute set in the status section (cibadmin -Ql
| grep -w pingd)? If not, then the problem is with the resource.
> Probably I didn't understand correctly what described at the link:
> http://www.clusterlabs.org/wiki/Pingd_with_resources_on_different_networks
> or it is outdated now... and instead of defining two clones it is better
> (aka works) to populate the host_list parameter as described here in case of
> more networks connected:
>
> http://www.clusterlabs.org/doc/en-US/Pacemaker/1.1/html/Pacemaker_Explained/ch09s03s03.html
The former is when you need to test connectivity on different
networks. I don't know if you need that.
Thanks,
Dejan
> Probably I'm missing something very simple but I don't get a clue to it...
> Gianluca
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
More information about the Pacemaker
mailing list