[Pacemaker] Best way to recover from failed STONITH?
Andrew Martin
amartin at xes-inc.com
Fri Dec 21 18:47:39 UTC 2012
Andreas,
Thanks for the help. Please see my replies inline below.
----- Original Message -----
> From: "Andreas Kurz" <andreas at hastexo.com>
> To: pacemaker at oss.clusterlabs.org
> Sent: Friday, December 21, 2012 10:11:08 AM
> Subject: Re: [Pacemaker] Best way to recover from failed STONITH?
>
> On 12/21/2012 04:18 PM, Andrew Martin wrote:
> > Hello,
> >
> > Yesterday a power failure took out one of the nodes and its STONITH
> > device (they share an upstream power source) in a 3-node
> > active/passive cluster (Corosync 2.1.0, Pacemaker 1.1.8). After
> > logging into the cluster, I saw that the STONITH operation had
> > given up in failure and that none of the resources were running on
> > the other nodes:
> > Dec 20 17:59:14 [18909] quorumnode crmd: notice:
> > too_many_st_failures: Too many failures to fence node0 (11),
> > giving up
> >
> > I brought the failed node back online and it rejoined the cluster,
> > but no more STONITH attempts were made and the resources remained
> > stopped. Eventually I set stonith-enabled="false" ran killall on
> > all pacemaker-related processes on the other (remaining) nodes,
> > then restarted pacemaker, and the resources successfully migrated
> > to one of the other nodes. This seems like a rather invasive
> > technique. My questions about this type of situation are:
> > - is there a better way to tell the cluster "I have manually
> > confirmed this node is dead/safe"? I see there is the meatclient
> > command, but can that only be used with the meatware STONITH
> > plugin?
>
> crm node cleanup quorumnode
I'm using the latest version of crmsh (1.2.1) but it doesn't seem to support this command:
root at node0:~# crm --version
1.2.1 (Build unknown)
root at node0:~# crm node
crm(live)node# help
Node management and status commands.
Available commands:
status show nodes' status as XML
show show node
standby put node into standby
online set node online
fence fence node
clearstate Clear node state
delete delete node
attribute manage attributes
utilization manage utilization attributes
status-attr manage status attributes
help show help (help topics for list of topics)
end go back one level
quit exit the program
Also, do I run cleanup on just the node that failed, or all of them?
>
> > - in general, is there a way to force the cluster to start
> > resources, if you just need to get them back online and as a
> > human have confirmed that things are okay? Something like crm
> > resource start rsc --force?
>
> ... see above ;-)
On a related note, is there a way to way to get better information
about why the cluster is in its current state? For example, in this
situation it would be nice to be able to run a command and have the
cluster print "resources stopped until node XXX can be fenced" to
be able to quickly assess the problem with the cluster.
>
> > - how can I completely clear out saved data for the cluster and
> > start over from scratch (last-resort option)? Stopping pacemaker
> > and removing everything from /var/lib/pacemaker/cib and
> > /var/lib/pacemaker/pengine cleans the CIB, but the nodes end up
> > sitting in the "pending" state for a very long time (30 minutes
> > or more). Am I missing another directory that needs to be
> > cleared?
>
> you started with an completely empty cib and the two (or three?)
> nodes
> needed 30min to form a cluster?
Yes, in fact I cleared out both /var/lib/pacemaker/cib and /var/lib/pacemaker/pengine
several times and most of the times after starting pacemaker again
one node would become "online" pretty quickly (less than 5 minutes), but the other two
would remain "pending" for quite some time. I left it going overnight
and this morning all of the nodes
>
> >
> > I am going to look into making the power source for the STONITH
> > device independent of the power source for the node itself,
> > however even with that setup there's still a chance that something
> > could take out both power sources at the same time, in which case
> > manual intervention and confirmation that the node is dead would
> > be required.
>
> Pacemaker 1.1.8 supports (again) stonith topologies ... so more than
> one
> fencing device and they can be "logically" combined.
Where can I find documentation on STONITH topologies and configuring
more than one fencing device for a single node? I don't see it mentioned
in the Cluster Labs documentation (Clusters from Scratch or Pacemaker Explained).
Thanks,
Andrew
>
> Regards,
> Andreas
>
> --
> Need help with Pacemaker?
> http://www.hastexo.com/now
>
> >
> > Thanks,
> >
> > Andrew
> >
> > _______________________________________________
> > Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> >
> > Project Home: http://www.clusterlabs.org
> > Getting started:
> > http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> > Bugs: http://bugs.clusterlabs.org
> >
>
>
>
>
>
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started:
> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>
More information about the Pacemaker
mailing list