[Pacemaker] Seems to be working but fails to transition to other node.
Steven Silk
steven.silk at noaa.gov
Thu May 31 00:16:56 UTC 2012
All Concerned;
I have been getting slapped around all day with this problem - I can't
solve it.
The system is only half done - I have not yet implemented the nfs portion -
but drbd part is not yet cooperating with corosync.
It appears to be working OK - but when I stop corosync on the DC - the
other node does not start drbd?
Here is how I am setting things up....
Configure quorum<http://docs.homelinux.org/doku.php?id=create_high-available_drbd_device_with_pacemaker#fn__1>and
stonith<http://docs.homelinux.org/doku.php?id=create_high-available_drbd_device_with_pacemaker#fn__2>
property no-quorum-policy="ignore"
property stonith-enabled="false"
On wms1 onfigure DRBD resource
primitive drbd_drbd0 ocf:linbit:drbd \
params drbd_resource="drbd0" \
op monitor interval="30s"
Configure DRBD Master/Slave
ms ms_drbd_drbd0 drbd_drbd0 \
meta master-max="1" master-node-max="1" \
clone-max="2" clone-node-max="1" \
notify="true"
Configure filesystem mountpoint
primitive fs_ftpdata ocf:heartbeat:Filesystem \
params device="/dev/drbd0" \
directory="/mnt/drbd0" fstype="ext3"
When I check the status on the DC....
[root at wms2 ~]# crm
crm(live)# status
============
Last updated: Wed May 30 23:58:43 2012
Last change: Wed May 30 23:52:42 2012 via cibadmin on wms1
Stack: openais
Current DC: wms2 - partition with quorum
Version: 1.1.6-3.el6-a02c0f19a00c1eb2527ad38f146ebc0834814558
2 Nodes configured, 2 expected votes
3 Resources configured.
============
Online: [ wms1 wms2 ]
Master/Slave Set: ms_drbd_drbd0 [drbd_drbd0]
Masters: [ wms2 ]
Slaves: [ wms1 ]
fs_ftpdata (ocf::heartbeat:Filesystem): Started wms2
[root at wms2 ~]# mount -l | grep drbd
/dev/drbd0 on /mnt/drbd0 type ext3 (rw)
So I stop corosync - but the other node...
[root at wms1 ~]# crm
crm(live)# status
============
Last updated: Thu May 31 00:12:17 2012
Last change: Wed May 30 23:52:42 2012 via cibadmin on wms1
Stack: openais
Current DC: wms1 - partition WITHOUT quorum
Version: 1.1.6-3.el6-a02c0f19a00c1eb2527ad38f146ebc0834814558
2 Nodes configured, 2 expected votes
3 Resources configured.
============
Online: [ wms1 ]
OFFLINE: [ wms2 ]
Master/Slave Set: ms_drbd_drbd0 [drbd_drbd0]
Masters: [ wms1 ]
Stopped: [ drbd_drbd0:1 ]
Fails to mount /dev/drbd0?
Any ideas?
I tailed /var/log/cluster/corosync.log and get this....
May 31 00:02:36 wms1 attrd: [1266]: WARN: attrd_cib_callback: Update 22 for
master-drbd_drbd0:0=5 failed: Remote node did not respond
May 31 00:03:06 wms1 attrd: [1266]: WARN: attrd_cib_callback: Update 25 for
master-drbd_drbd0:0=5 failed: Remote node did not respond
May 31 00:03:10 wms1 crmd: [1268]: WARN: cib_rsc_callback: Resource update
15 failed: (rc=-41) Remote node did not respond
May 31 00:03:36 wms1 attrd: [1266]: WARN: attrd_cib_callback: Update 28 for
master-drbd_drbd0:0=5 failed: Remote node did not respond
May 31 00:04:06 wms1 attrd: [1266]: WARN: attrd_cib_callback: Update 31 for
master-drbd_drbd0:0=5 failed: Remote node did not respond
May 31 00:04:10 wms1 attrd: [1266]: WARN: attrd_cib_callback: Update 34 for
master-drbd_drbd0:0=5 failed: Remote node did not respond
May 31 00:04:10 wms1 attrd: [1266]: WARN: attrd_cib_callback: Update 37 for
master-drbd_drbd0:0=5 failed: Remote node did not respond
May 31 00:04:10 wms1 attrd: [1266]: WARN: attrd_cib_callback: Update 40 for
master-drbd_drbd0:0=5 failed: Remote node did not respond
May 31 00:08:02 wms1 cib: [1257]: info: cib_stats: Processed 58 operations
(0.00us average, 0% utilization) in the last 10min
May 31 00:08:02 wms1 cib: [1264]: info: cib_stats: Processed 117 operations
(256.00us average, 0% utilization) in the last 10min
[root at wms2 ~]# tail /var/log/cluster/corosync.log
May 31 00:02:16 corosync [pcmk ] info: update_member: Node wms2 now has
process list: 00000000000000000000000000000002 (2)
May 31 00:02:16 corosync [pcmk ] notice: pcmk_shutdown: Shutdown complete
May 31 00:02:16 corosync [SERV ] Service engine unloaded: Pacemaker
Cluster Manager 1.1.6
May 31 00:02:16 corosync [SERV ] Service engine unloaded: corosync
extended virtual synchrony service
May 31 00:02:16 corosync [SERV ] Service engine unloaded: corosync
configuration service
May 31 00:02:16 corosync [SERV ] Service engine unloaded: corosync cluster
closed process group service v1.01
May 31 00:02:16 corosync [SERV ] Service engine unloaded: corosync cluster
config database access v1.01
May 31 00:02:16 corosync [SERV ] Service engine unloaded: corosync profile
loading service
May 31 00:02:16 corosync [SERV ] Service engine unloaded: corosync cluster
quorum service v0.1
May 31 00:02:16 corosync [MAIN ] Corosync Cluster Engine exiting with
status 0 at main.c:1858.
the example that I am working from talks about doing the following....
group services fs_drbd0
But this fails miserable... services being undefined?
--
Steven Silk
CSC
303 497 3112
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20120530/31cc9bd6/attachment-0003.html>
More information about the Pacemaker
mailing list