[Pacemaker] clone ip definition and location stops my resources...

Dejan Muhamedagic dejanmm at fastmail.fm
Tue May 11 05:58:48 EDT 2010


Hi,

On Tue, May 11, 2010 at 11:15:49AM +0200, Gianluca Cecchi wrote:
> On Mon, May 10, 2010 at 4:39 PM, Vadym Chepkov <vchepkov at gmail.com> wrote:
> 
> > # crm ra meta ping
> >
> > name (string, [undef]): Attribute name
> >     The name of the attributes to set.  This is the name to be used in the
> > constraints.
> >
> > By default is "pingd", but you are checking against pinggw
> >
> > I suggest you do not change name though, but adjust your location
> > constraint to use pingd instead.
> > crm_mon only notices "pingd" at the moment whenn you pass -f argument: it's
> > hardcoded
> >
> >
> > On Mon, May 10, 2010 at 9:34 AM, Gianluca Cecchi <
> > gianluca.cecchi at gmail.com> wrote:
> >
> >> Hello,
> >> using pacemaker 1.0.8 on rh el 5 I have some problems understanding the
> >> way ping clone works to setup monitoring of gw... even after reading docs...
> >>
> >> As soon as I run:
> >> crm configure location nfs-group-with-pinggw nfs-group rule -inf:
> >> not_defined pinggw or pinggw lte 0
> >>
> >> the resources go stopped and don't re-start....
> >>
> >> [snip]
> >>
> >
> hem...
> I changed the location line so that now I have:
> primitive pinggw ocf:pacemaker:ping \
> params host_list="192.168.101.1" multiplier="100" \
> op start interval="0" timeout="90" \
> op stop interval="0" timeout="100"
> 
> clone cl-pinggw pinggw \
> meta globally-unique="false"
> 
> location nfs-group-with-pinggw nfs-group \
> rule $id="nfs-group-with-pinggw-rule" -inf: not_defined pingd or pingd lte 0
> 
> But now nothing happens  if I run for example
>  iptables -A OUTPUT -p icmp -d 192.168.101.1 -j REJECT (or DROP)
> in the node where nfs-group is running.....
> 
> Do I have to name the primitive itself to pingd????

I guess not.

> It seems that the binary /bin/ping is not accessed at all (with ls -lu ...)

Then the ping resource is not running...

> Or do I have to change the general property I previously define to avoide
> failback:
> rsc_defaults $id="rsc-options" \
> resource-stickiness="100"
> 
> crm_mon -f -r gives:
> Online: [ ha1 ha2 ]
> 
> Full list of resources:
> 
> SitoWeb (ocf::heartbeat:apache):        Started ha1
>  Master/Slave Set: NfsData
>      Masters: [ ha1 ]
>      Slaves: [ ha2 ]
>  Resource Group: nfs-group
>      ClusterIP  (ocf::heartbeat:IPaddr2): Started ha1
>      lv_drbd0   (ocf::heartbeat:LVM):   Started ha1
>      NfsFS (ocf::heartbeat:Filesystem):    Started ha1
>      nfssrv     (ocf::heartbeat:nfsserver):     Started ha1
> nfsclient (ocf::heartbeat:Filesystem):    Started ha2
>  Clone Set: cl-pinggw
>      Started: [ ha2 ha1 ]

... but according to this it is.

> Migration summary:
> * Node ha1:  pingd=100
> * Node ha2:  pingd=100

Do you see the attribute set in the status section (cibadmin -Ql
| grep -w pingd)? If not, then the problem is with the resource.

> Probably I didn't understand correctly what described at the link:
> http://www.clusterlabs.org/wiki/Pingd_with_resources_on_different_networks
> or it is outdated now... and instead of defining two clones it is better
> (aka works) to populate the host_list parameter as described here in case of
> more networks connected:
> 
> http://www.clusterlabs.org/doc/en-US/Pacemaker/1.1/html/Pacemaker_Explained/ch09s03s03.html

The former is when you need to test connectivity on different
networks. I don't know if you need that.

Thanks,

Dejan

> Probably I'm missing something very simple but I don't get a clue to it...
> Gianluca

> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf





More information about the Pacemaker mailing list