[Pacemaker] Breaking pacemaker
jimbob palmer
jimbobpalmer at gmail.com
Tue Feb 16 12:59:17 UTC 2010
Hello,
I have a cluster that is all working perfectly. Time to break it.
This is a two node master/slave cluster with drbd. Failover between
the nodes works backwards and forwards. Everything is happier than a
well fed cat.
I wanted to see what would happen if the drbd device couldn't be
mounted, so on the slave node I deleted the mountpoint, then failed
over.
Oh dear. I broke things so badly that I had to fail back, shutdown
corosync on the slave, delete the config files, and start it again.
Since that's not the right way to do it, I thought I should ask the
list for the right way.
Here are the errors I get on the slave when the fs_BLAH resource tries
to start with a missing mountpoint:
Failed actions:
fs_BLAH_start_0 (node=X, call=X, status=complete): not installed
The logs tell me that the mount point doesn't exist, so I create it,
and try to tell pacemaker.
crm(live)resource# start gr_GROUPNAME
Multiple attributes match name=target-role
(group members listed here)
Okay so starting a group doesn't work. I try to start the filesystem member:
crm(live)resource# start fs_BLAH
That didn't work. I expected it to, but it didn't.
Failover failback doesn't work either.
What am I doing wrong here? This seems fairly sensible..
Thanks
J
More information about the Pacemaker
mailing list