[Pacemaker] Corosync and Pacemaker Hangs
Norbert Kiam Maclang
norbert.kiam.maclang at gmail.com
Mon Sep 15 01:24:59 UTC 2014
Hi Vladislav and Andrew,
After adding fencing/stonith (resource level) and fencing handlers on drbd,
I am not getting monitor timeouts on drbd but I am experiencing a different
problem now. As per my understanding, logs on node01 showed that it detects
node02 to be disconnected (and moved the resources to itself) but crm_mon
shows that the resources are still started on node02 which is not.
Node01:
node01 crmd[952]: notice: do_state_transition: State transition S_IDLE ->
S_POLICY_ENGINE [ input=I_PE_CALC cause=C_TIMER_POPPED
origin=crm_timer_popped ]
node01 pengine[951]: notice: unpack_config: On loss of CCM Quorum: Ignore
node01 crmd[952]: notice: run_graph: Transition 260 (Complete=0,
Pending=0, Fired=0, Skipped=0, Incomplete=0,
Source=/var/lib/pacemaker/pengine/pe-input-78.bz2): Complete
node01 crmd[952]: notice: do_state_transition: State transition
S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_FSA_INTERNAL
origin=notify_crmd ]
node01 pengine[951]: notice: process_pe_message: Calculated Transition
260: /var/lib/pacemaker/pengine/pe-input-78.bz2
node01 corosync[917]: [TOTEM ] A processor failed, forming new
configuration.
node01 corosync[917]: [TOTEM ] A new membership (10.2.131.20:352) was
formed. Members left: 167936789
node01 crmd[952]: warning: match_down_event: No match for shutdown action
on 167936789
node01 crmd[952]: notice: peer_update_callback: Stonith/shutdown of
node02 not matched
node01 crmd[952]: notice: do_state_transition: State transition S_IDLE ->
S_POLICY_ENGINE [ input=I_PE_CALC cause=C_FSA_INTERNAL
origin=abort_transition_graph ]
node01 pengine[951]: notice: unpack_config: On loss of CCM Quorum: Ignore
node01 pengine[951]: warning: pe_fence_node: Node node02 will be fenced
because our peer process is no longer available
node01 pengine[951]: warning: determine_online_status: Node node02 is
unclean
node01 pengine[951]: warning: stage6: Scheduling Node node02 for STONITH
node01 pengine[951]: notice: LogActions: Move fs_pg#011(Started node02
-> node01)
node01 pengine[951]: notice: LogActions: Move ip_pg#011(Started node02
-> node01)
node01 pengine[951]: notice: LogActions: Move lsb_pg#011(Started
node02 -> node01)
node01 pengine[951]: notice: LogActions: Demote drbd_pg:0#011(Master ->
Stopped node02)
node01 pengine[951]: notice: LogActions: Promote drbd_pg:1#011(Slave ->
Master node01)
node01 pengine[951]: notice: LogActions: Stop p_fence:0#011(node02)
node01 crmd[952]: notice: te_rsc_command: Initiating action 2: cancel
drbd_pg_cancel_31000 on node01 (local)
node01 crmd[952]: notice: te_fence_node: Executing reboot fencing
operation (54) on node02 (timeout=60000)
node01 stonith-ng[948]: notice: handle_request: Client crmd.952.6d7ac808
wants to fence (reboot) 'node02' with device '(any)'
node01 stonith-ng[948]: notice: initiate_remote_stonith_op: Initiating
remote operation reboot for node02: 96530c7b-1c80-42c4-82cf-840bf3d5bb5f (0)
node01 crmd[952]: notice: te_rsc_command: Initiating action 68: notify
drbd_pg_pre_notify_demote_0 on node02
node01 crmd[952]: notice: te_rsc_command: Initiating action 70: notify
drbd_pg_pre_notify_demote_0 on node01 (local)
node01 pengine[951]: warning: process_pe_message: Calculated Transition
261: /var/lib/pacemaker/pengine/pe-warn-0.bz2
node01 crmd[952]: notice: process_lrm_event: LRM operation
drbd_pg_notify_0 (call=63, rc=0, cib-update=0, confirmed=true) ok
node01 kernel: [230495.836024] d-con pg: PingAck did not arrive in time.
node01 kernel: [230495.836176] d-con pg: peer( Primary -> Unknown ) conn(
Connected -> NetworkFailure ) pdsk( UpToDate -> DUnknown )
node01 kernel: [230495.837204] d-con pg: asender terminated
node01 kernel: [230495.837216] d-con pg: Terminating drbd_a_pg
node01 kernel: [230495.837286] d-con pg: Connection closed
node01 kernel: [230495.837298] d-con pg: conn( NetworkFailure ->
Unconnected )
node01 kernel: [230495.837299] d-con pg: receiver terminated
node01 kernel: [230495.837300] d-con pg: Restarting receiver thread
node01 kernel: [230495.837304] d-con pg: receiver (re)started
node01 kernel: [230495.837314] d-con pg: conn( Unconnected -> WFConnection
)
node01 crmd[952]: warning: action_timer_callback: Timer popped
(timeout=20000, abort_level=1000000, complete=false)
node01 crmd[952]: error: print_synapse: [Action 2]: Completed rsc op
drbd_pg_cancel_31000 on node01 (priority: 0, waiting: none)
node01 crmd[952]: warning: action_timer_callback: Timer popped
(timeout=20000, abort_level=1000000, complete=false)
node01 crmd[952]: error: print_synapse: [Action 68]: In-flight rsc op
drbd_pg_pre_notify_demote_0 on node02 (priority: 0, waiting: none)
node01 crmd[952]: warning: cib_action_update: rsc_op 68:
drbd_pg_pre_notify_demote_0 on node02 timed out
node01 crmd[952]: error: cib_action_updated: Update 297 FAILED: Timer
expired
node01 crmd[952]: error: stonith_async_timeout_handler: Async call 2
timed out after 120000ms
node01 crmd[952]: notice: tengine_stonith_callback: Stonith operation
2/54:261:0:6978227d-ce2d-4dc6-955a-eb9313f112a5: Timer expired (-62)
node01 crmd[952]: notice: tengine_stonith_callback: Stonith operation 2
for node02 failed (Timer expired): aborting transition.
node01 crmd[952]: notice: run_graph: Transition 261 (Complete=6,
Pending=0, Fired=0, Skipped=29, Incomplete=15,
Source=/var/lib/pacemaker/pengine/pe-warn-0.bz2): Stopped
node01 pengine[951]: notice: unpack_config: On loss of CCM Quorum: Ignore
node01 pengine[951]: warning: pe_fence_node: Node node02 will be fenced
because our peer process is no longer available
node01 pengine[951]: warning: determine_online_status: Node node02 is
unclean
node01 pengine[951]: warning: stage6: Scheduling Node node02 for STONITH
node01 pengine[951]: notice: LogActions: Move fs_pg#011(Started node02
-> node01)
node01 pengine[951]: notice: LogActions: Move ip_pg#011(Started node02
-> node01)
node01 pengine[951]: notice: LogActions: Move lsb_pg#011(Started
node02 -> node01)
node01 pengine[951]: notice: LogActions: Demote drbd_pg:0#011(Master ->
Stopped node02)
node01 pengine[951]: notice: LogActions: Promote drbd_pg:1#011(Slave ->
Master node01)
node01 pengine[951]: notice: LogActions: Stop p_fence:0#011(node02)
node01 crmd[952]: notice: te_fence_node: Executing reboot fencing
operation (53) on node02 (timeout=60000)
node01 stonith-ng[948]: notice: handle_request: Client crmd.952.6d7ac808
wants to fence (reboot) 'node02' with device '(any)'
node01 stonith-ng[948]: notice: initiate_remote_stonith_op: Initiating
remote operation reboot for node02: a4fae8ce-3a6c-4fe5-a934-b5b83ae123cb (0)
node01 crmd[952]: notice: te_rsc_command: Initiating action 67: notify
drbd_pg_pre_notify_demote_0 on node02
node01 crmd[952]: notice: te_rsc_command: Initiating action 69: notify
drbd_pg_pre_notify_demote_0 on node01 (local)
node01 pengine[951]: warning: process_pe_message: Calculated Transition
262: /var/lib/pacemaker/pengine/pe-warn-1.bz2
node01 crmd[952]: notice: process_lrm_event: LRM operation
drbd_pg_notify_0 (call=66, rc=0, cib-update=0, confirmed=true) ok
Last updated: Mon Sep 15 01:15:59 2014
Last change: Sat Sep 13 15:23:45 2014 via cibadmin on node01
Stack: corosync
Current DC: node01 (167936788) - partition with quorum
Version: 1.1.10-42f2063
2 Nodes configured
7 Resources configured
Node node02 (167936789): UNCLEAN (online)
Online: [ node01 ]
Resource Group: PGServer
fs_pg (ocf::heartbeat:Filesystem): Started node02
ip_pg (ocf::heartbeat:IPaddr2): Started node02
lsb_pg (lsb:postgresql): Started node02
Master/Slave Set: ms_drbd_pg [drbd_pg]
Masters: [ node02 ]
Slaves: [ node01 ]
Clone Set: cln_p_fence [p_fence]
Started: [ node01 node02 ]
Thank you,
Norbert
On Fri, Sep 12, 2014 at 12:06 PM, Vladislav Bogdanov <bubble at hoster-ok.com>
wrote:
> 12.09.2014 05:00, Norbert Kiam Maclang wrote:
> > Hi,
> >
> > After adding resource level fencing on drbd, I still ended up having
> > problems with timeouts on drbd. Is there a recommended settings for
> > this? I followed what is written in the drbd documentation -
> >
> http://www.drbd.org/users-guide-emb/s-pacemaker-crm-drbd-backed-service.html
> > , Another thing I can't understand is why during initial tests, even I
> > reboot the vms several times, failover works. But after I soak it for a
> > couple of hours (say for example 8 hours or more) and continue with the
> > tests, it will not failover and experience split brain. I confirmed it
> > though that everything is healthy before performing a reboot. Disk
> > health and network is good, drbd is synced, time beetween servers is
> good.
>
> I recall I've seen something similar a year ago (near the time your
> pacemaker version is dated). I do not remember what was the exact
> problem cause, but I saw that drbd RA timeouts because it waits for
> something (fencing) in the kernel space to be done. drbd calls userspace
> scripts from within kernelspace, and you'll see them in the process list
> with the drbd kernel thread as a parent.
>
> I'd also upgrade your corosync configuration from "member" to "nodelist"
> syntax, specifying "name" parameter together with ring0_addr for nodes
> (that parameter is not referenced in corosync docs but should be
> somewhere in the Pacemaker Explained - it is used only by the pacemaker).
>
> Also there is trace_ra functionality support in both pacemaker and crmsh
> (cannot say if that is supported in versions you have though, probably
> yes) so you may want to play with that to get the exact picture from the
> resource agent.
>
> Anyways, upgrading to 1.1.12 and more recent crmsh is nice to have for
> you because you may be just hitting a long-ago solved and forgotten
> bug/issue.
>
> Concerning your
> > expected-quorum-votes="1"
>
> You need to configure votequorum in corosync with two_node: 1 instead of
> that line.
>
> >
> > # Logs:
> > node01 lrmd[1036]: warning: child_timeout_callback:
> > drbd_pg_monitor_29000 process (PID 27744) timed out
> > node01 lrmd[1036]: warning: operation_finished:
> > drbd_pg_monitor_29000:27744 - timed out after 20000ms
> > node01 crmd[1039]: error: process_lrm_event: LRM operation
> > drbd_pg_monitor_29000 (69) Timed Out (timeout=20000ms)
> > node01 crmd[1039]: warning: update_failcount: Updating failcount for
> > drbd_pg on tyo1mqdb01p after failed monitor: rc=1 (update=value++,
> > time=1410486352)
> >
> > Thanks,
> > Kiam
> >
> > On Thu, Sep 11, 2014 at 6:58 PM, Norbert Kiam Maclang
> > <norbert.kiam.maclang at gmail.com <mailto:norbert.kiam.maclang at gmail.com>>
> > wrote:
> >
> > Thank you Vladislav.
> >
> > I have configured resource level fencing on drbd and removed
> > wfc-timeout and defr-wfc-timeout (is this required?). My drbd
> > configuration is now:
> >
> > resource pg {
> > device /dev/drbd0;
> > disk /dev/vdb;
> > meta-disk internal;
> > disk {
> > fencing resource-only;
> > on-io-error detach;
> > resync-rate 40M;
> > }
> > handlers {
> > fence-peer "/usr/lib/drbd/crm-fence-peer.sh";
> > after-resync-target "/usr/lib/drbd/crm-unfence-peer.sh";
> > split-brain "/usr/lib/drbd/notify-split-brain.sh nkbm";
> > }
> > on node01 {
> > address 10.2.136.52:7789 <http://10.2.136.52:7789>;
> > }
> > on node02 {
> > address 10.2.136.55:7789 <http://10.2.136.55:7789>;
> > }
> > net {
> > verify-alg md5;
> > after-sb-0pri discard-zero-changes;
> > after-sb-1pri discard-secondary;
> > after-sb-2pri disconnect;
> > }
> > }
> >
> > Failover works on my initial test (restarting both nodes alternately
> > - this always works). Will wait for a couple of hours after doing a
> > failover test again (Which always fail on my previous setup).
> >
> > Thank you!
> > Kiam
> >
> > On Thu, Sep 11, 2014 at 2:14 PM, Vladislav Bogdanov
> > <bubble at hoster-ok.com <mailto:bubble at hoster-ok.com>> wrote:
> >
> > 11.09.2014 05:57, Norbert Kiam Maclang wrote:
> > > Is this something to do with quorum? But I already set
> >
> > You'd need to configure fencing at the drbd resources level.
> >
> >
> http://www.drbd.org/users-guide-emb/s-pacemaker-fencing.html#s-pacemaker-fencing-cib
> >
> >
> > >
> > > property no-quorum-policy="ignore" \
> > > expected-quorum-votes="1"
> > >
> > > Thanks in advance,
> > > Kiam
> > >
> > > On Thu, Sep 11, 2014 at 10:09 AM, Norbert Kiam Maclang
> > > <norbert.kiam.maclang at gmail.com
> > <mailto:norbert.kiam.maclang at gmail.com>
> > <mailto:norbert.kiam.maclang at gmail.com
> > <mailto:norbert.kiam.maclang at gmail.com>>>
> > > wrote:
> > >
> > > Hi,
> > >
> > > Please help me understand what is causing the problem. I
> > have a 2
> > > node cluster running on vms using KVM. Each vm (I am using
> > Ubuntu
> > > 14.04) runs on a separate hypervisor on separate machines.
> > All are
> > > working well during testing (I restarted the vms
> > alternately), but
> > > after a day when I kill the other node, I always end up
> > corosync and
> > > pacemaker hangs on the surviving node. Date and time on
> > the vms are
> > > in sync, I use unicast, tcpdump shows both nodes exchanges,
> > > confirmed that DRBD is healthy and crm_mon show good
> > status before I
> > > kill the other node. Below are my configurations and
> > versions I used:
> > >
> > > corosync 2.3.3-1ubuntu1
> > > crmsh 1.2.5+hg1034-1ubuntu3
> > > drbd8-utils 2:8.4.4-1ubuntu1
> > > libcorosync-common4 2.3.3-1ubuntu1
> > > libcrmcluster4 1.1.10+git20130802-1ubuntu2
> > > libcrmcommon3 1.1.10+git20130802-1ubuntu2
> > > libcrmservice1 1.1.10+git20130802-1ubuntu2
> > > pacemaker 1.1.10+git20130802-1ubuntu2
> > > pacemaker-cli-utils 1.1.10+git20130802-1ubuntu2
> > > postgresql-9.3 9.3.5-0ubuntu0.14.04.1
> > >
> > > # /etc/corosync/corosync:
> > > totem {
> > > version: 2
> > > token: 3000
> > > token_retransmits_before_loss_const: 10
> > > join: 60
> > > consensus: 3600
> > > vsftype: none
> > > max_messages: 20
> > > clear_node_high_bit: yes
> > > secauth: off
> > > threads: 0
> > > rrp_mode: none
> > > interface {
> > > member {
> > > memberaddr: 10.2.136.56
> > > }
> > > member {
> > > memberaddr: 10.2.136.57
> > > }
> > > ringnumber: 0
> > > bindnetaddr: 10.2.136.0
> > > mcastport: 5405
> > > }
> > > transport: udpu
> > > }
> > > amf {
> > > mode: disabled
> > > }
> > > quorum {
> > > provider: corosync_votequorum
> > > expected_votes: 1
> > > }
> > > aisexec {
> > > user: root
> > > group: root
> > > }
> > > logging {
> > > fileline: off
> > > to_stderr: yes
> > > to_logfile: no
> > > to_syslog: yes
> > > syslog_facility: daemon
> > > debug: off
> > > timestamp: on
> > > logger_subsys {
> > > subsys: AMF
> > > debug: off
> > > tags:
> > enter|leave|trace1|trace2|trace3|trace4|trace6
> > > }
> > > }
> > >
> > > # /etc/corosync/service.d/pcmk:
> > > service {
> > > name: pacemaker
> > > ver: 1
> > > }
> > >
> > > /etc/drbd.d/global_common.conf:
> > > global {
> > > usage-count no;
> > > }
> > >
> > > common {
> > > net {
> > > protocol C;
> > > }
> > > }
> > >
> > > # /etc/drbd.d/pg.res:
> > > resource pg {
> > > device /dev/drbd0;
> > > disk /dev/vdb;
> > > meta-disk internal;
> > > startup {
> > > wfc-timeout 15;
> > > degr-wfc-timeout 60;
> > > }
> > > disk {
> > > on-io-error detach;
> > > resync-rate 40M;
> > > }
> > > on node01 {
> > > address 10.2.136.56:7789 <http://10.2.136.56:7789>
> > <http://10.2.136.56:7789>;
> > > }
> > > on node02 {
> > > address 10.2.136.57:7789 <http://10.2.136.57:7789>
> > <http://10.2.136.57:7789>;
> > > }
> > > net {
> > > verify-alg md5;
> > > after-sb-0pri discard-zero-changes;
> > > after-sb-1pri discard-secondary;
> > > after-sb-2pri disconnect;
> > > }
> > > }
> > >
> > > # Pacemaker configuration:
> > > node $id="167938104" node01
> > > node $id="167938105" node02
> > > primitive drbd_pg ocf:linbit:drbd \
> > > params drbd_resource="pg" \
> > > op monitor interval="29s" role="Master" \
> > > op monitor interval="31s" role="Slave"
> > > primitive fs_pg ocf:heartbeat:Filesystem \
> > > params device="/dev/drbd0"
> > directory="/var/lib/postgresql/9.3/main"
> > > fstype="ext4"
> > > primitive ip_pg ocf:heartbeat:IPaddr2 \
> > > params ip="10.2.136.59" cidr_netmask="24" nic="eth0"
> > > primitive lsb_pg lsb:postgresql
> > > group PGServer fs_pg lsb_pg ip_pg
> > > ms ms_drbd_pg drbd_pg \
> > > meta master-max="1" master-node-max="1" clone-max="2"
> > > clone-node-max="1" notify="true"
> > > colocation pg_on_drbd inf: PGServer ms_drbd_pg:Master
> > > order pg_after_drbd inf: ms_drbd_pg:promote PGServer:start
> > > property $id="cib-bootstrap-options" \
> > > dc-version="1.1.10-42f2063" \
> > > cluster-infrastructure="corosync" \
> > > stonith-enabled="false" \
> > > no-quorum-policy="ignore"
> > > rsc_defaults $id="rsc-options" \
> > > resource-stickiness="100"
> > >
> > > # Logs on node01
> > > Sep 10 10:25:33 node01 crmd[1019]: notice:
> > peer_update_callback:
> > > Our peer on the DC is dead
> > > Sep 10 10:25:33 node01 crmd[1019]: notice:
> > do_state_transition:
> > > State transition S_NOT_DC -> S_ELECTION [ input=I_ELECTION
> > > cause=C_CRMD_STATUS_CALLBACK origin=peer_update_callback ]
> > > Sep 10 10:25:33 node01 crmd[1019]: notice:
> > do_state_transition:
> > > State transition S_ELECTION -> S_INTEGRATION [
> > input=I_ELECTION_DC
> > > cause=C_FSA_INTERNAL origin=do_election_check ]
> > > Sep 10 10:25:33 node01 corosync[940]: [TOTEM ] A new
> > membership
> > > (10.2.136.56:52 <http://10.2.136.56:52>
> > <http://10.2.136.56:52>) was formed. Members left:
> > > 167938105
> > > Sep 10 10:25:45 node01 kernel: [74452.740024] d-con pg:
> > PingAck did
> > > not arrive in time.
> > > Sep 10 10:25:45 node01 kernel: [74452.740169] d-con pg:
> peer(
> > > Primary -> Unknown ) conn( Connected -> NetworkFailure )
> pdsk(
> > > UpToDate -> DUnknown )
> > > Sep 10 10:25:45 node01 kernel: [74452.740987] d-con pg:
> > asender
> > > terminated
> > > Sep 10 10:25:45 node01 kernel: [74452.740999] d-con pg:
> > Terminating
> > > drbd_a_pg
> > > Sep 10 10:25:45 node01 kernel: [74452.741235] d-con pg:
> > Connection
> > > closed
> > > Sep 10 10:25:45 node01 kernel: [74452.741259] d-con pg:
> conn(
> > > NetworkFailure -> Unconnected )
> > > Sep 10 10:25:45 node01 kernel: [74452.741260] d-con pg:
> > receiver
> > > terminated
> > > Sep 10 10:25:45 node01 kernel: [74452.741261] d-con pg:
> > Restarting
> > > receiver thread
> > > Sep 10 10:25:45 node01 kernel: [74452.741262] d-con pg:
> > receiver
> > > (re)started
> > > Sep 10 10:25:45 node01 kernel: [74452.741269] d-con pg:
> conn(
> > > Unconnected -> WFConnection )
> > > Sep 10 10:26:12 node01 lrmd[1016]: warning:
> > child_timeout_callback:
> > > drbd_pg_monitor_31000 process (PID 8445) timed out
> > > Sep 10 10:26:12 node01 lrmd[1016]: warning:
> > operation_finished:
> > > drbd_pg_monitor_31000:8445 - timed out after 20000ms
> > > Sep 10 10:26:12 node01 crmd[1019]: error:
> > process_lrm_event: LRM
> > > operation drbd_pg_monitor_31000 (30) Timed Out
> > (timeout=20000ms)
> > > Sep 10 10:26:32 node01 crmd[1019]: warning:
> cib_rsc_callback:
> > > Resource update 23 failed: (rc=-62) Timer expired
> > > Sep 10 10:27:03 node01 lrmd[1016]: warning:
> > child_timeout_callback:
> > > drbd_pg_monitor_31000 process (PID 8693) timed out
> > > Sep 10 10:27:03 node01 lrmd[1016]: warning:
> > operation_finished:
> > > drbd_pg_monitor_31000:8693 - timed out after 20000ms
> > > Sep 10 10:27:54 node01 lrmd[1016]: warning:
> > child_timeout_callback:
> > > drbd_pg_monitor_31000 process (PID 8938) timed out
> > > Sep 10 10:27:54 node01 lrmd[1016]: warning:
> > operation_finished:
> > > drbd_pg_monitor_31000:8938 - timed out after 20000ms
> > > Sep 10 10:28:33 node01 crmd[1019]: error:
> crm_timer_popped:
> > > Integration Timer (I_INTEGRATED) just popped in state
> > S_INTEGRATION!
> > > (180000ms)
> > > Sep 10 10:28:33 node01 crmd[1019]: warning:
> > do_state_transition:
> > > Progressed to state S_FINALIZE_JOIN after C_TIMER_POPPED
> > > Sep 10 10:28:33 node01 crmd[1019]: warning:
> > do_state_transition: 1
> > > cluster nodes failed to respond to the join offer.
> > > Sep 10 10:28:33 node01 crmd[1019]: notice:
> > crmd_join_phase_log:
> > > join-1: node02=none
> > > Sep 10 10:28:33 node01 crmd[1019]: notice:
> > crmd_join_phase_log:
> > > join-1: node01=welcomed
> > > Sep 10 10:28:45 node01 lrmd[1016]: warning:
> > child_timeout_callback:
> > > drbd_pg_monitor_31000 process (PID 9185) timed out
> > > Sep 10 10:28:45 node01 lrmd[1016]: warning:
> > operation_finished:
> > > drbd_pg_monitor_31000:9185 - timed out after 20000ms
> > > Sep 10 10:29:36 node01 lrmd[1016]: warning:
> > child_timeout_callback:
> > > drbd_pg_monitor_31000 process (PID 9432) timed out
> > > Sep 10 10:29:36 node01 lrmd[1016]: warning:
> > operation_finished:
> > > drbd_pg_monitor_31000:9432 - timed out after 20000ms
> > > Sep 10 10:30:27 node01 lrmd[1016]: warning:
> > child_timeout_callback:
> > > drbd_pg_monitor_31000 process (PID 9680) timed out
> > > Sep 10 10:30:27 node01 lrmd[1016]: warning:
> > operation_finished:
> > > drbd_pg_monitor_31000:9680 - timed out after 20000ms
> > > Sep 10 10:31:18 node01 lrmd[1016]: warning:
> > child_timeout_callback:
> > > drbd_pg_monitor_31000 process (PID 9927) timed out
> > > Sep 10 10:31:18 node01 lrmd[1016]: warning:
> > operation_finished:
> > > drbd_pg_monitor_31000:9927 - timed out after 20000ms
> > > Sep 10 10:32:09 node01 lrmd[1016]: warning:
> > child_timeout_callback:
> > > drbd_pg_monitor_31000 process (PID 10174) timed out
> > > Sep 10 10:32:09 node01 lrmd[1016]: warning:
> > operation_finished:
> > > drbd_pg_monitor_31000:10174 - timed out after 20000ms
> > >
> > > #crm_mon on node01 before I kill the other vm:
> > > Stack: corosync
> > > Current DC: node02 (167938104) - partition with quorum
> > > Version: 1.1.10-42f2063
> > > 2 Nodes configured
> > > 5 Resources configured
> > >
> > > Online: [ node01 node02 ]
> > >
> > > Resource Group: PGServer
> > > fs_pg (ocf::heartbeat:Filesystem): Started
> node02
> > > lsb_pg (lsb:postgresql): Started node02
> > > ip_pg (ocf::heartbeat:IPaddr2): Started
> node02
> > > Master/Slave Set: ms_drbd_pg [drbd_pg]
> > > Masters: [ node02 ]
> > > Slaves: [ node01 ]
> > >
> > > Thank you,
> > > Kiam
> > >
> > >
> > >
> > >
> > > _______________________________________________
> > > Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> > <mailto:Pacemaker at oss.clusterlabs.org>
> > > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> > >
> > > Project Home: http://www.clusterlabs.org
> > > Getting started:
> > http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> > > Bugs: http://bugs.clusterlabs.org
> > >
> >
> >
> > _______________________________________________
> > Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> > <mailto:Pacemaker at oss.clusterlabs.org>
> > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> >
> > Project Home: http://www.clusterlabs.org
> > Getting started:
> > http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> > Bugs: http://bugs.clusterlabs.org
> >
> >
> >
> >
> >
> > _______________________________________________
> > Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> >
> > Project Home: http://www.clusterlabs.org
> > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> > Bugs: http://bugs.clusterlabs.org
> >
>
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20140915/1076ae17/attachment.htm>
More information about the Pacemaker
mailing list