[Pacemaker] Trying to figure out a constraint
Digimer
lists at alteeve.ca
Wed Jun 18 06:03:32 CEST 2014
Hi all,
I am trying to setup a basic pacemaker 1.1.10 on RHEL 6.5 with DRBD
8.3.16.
I've setup DRBD and configured one clustered LVM volume group using
that drbd resource as the PV. With DRBD configured alone, I can
stop/start pacemaker repeatedly without issue. However, when I add the
LVM VG using ocf:heartbeat:LVM and setup a constraint, subsequent
restarts of pacemaker almost always end up with a fence. I have to think
then that I am messing up my constraints...
Config:
====
Cluster Name: an-anvil-04
Corosync Nodes:
Pacemaker Nodes:
an-a04n01.alteeve.ca an-a04n02.alteeve.ca
Resources:
Master: drbd_r0_Clone
Meta Attrs: master-max=2 master-node-max=1 clone-max=2
clone-node-max=1 notify=true
Resource: drbd_r0 (class=ocf provider=linbit type=drbd)
Attributes: drbd_resource=r0
Operations: monitor interval=30s (drbd_r0-monitor-interval-30s)
Master: lvm_n01_vg0_Clone
Meta Attrs: master-max=2 master-node-max=1 clone-max=2
clone-node-max=1 notify=true
Resource: lvm_n01_vg0 (class=ocf provider=heartbeat type=LVM)
Attributes: volgrpname=an-a04n01_vg0
Operations: monitor interval=30s (lvm_n01_vg0-monitor-interval-30s)
Stonith Devices:
Resource: fence_n01_ipmi (class=stonith type=fence_ipmilan)
Attributes: pcmk_host_list=an-a04n01.alteeve.ca ipaddr=an-a04n01.ipmi
action=reboot login=admin passwd=Initial1 delay=15
Operations: monitor interval=60s (fence_n01_ipmi-monitor-interval-60s)
Resource: fence_n02_ipmi (class=stonith type=fence_ipmilan)
Attributes: pcmk_host_list=an-a04n02.alteeve.ca ipaddr=an-a04n02.ipmi
action=reboot login=admin passwd=Initial1
Operations: monitor interval=60s (fence_n02_ipmi-monitor-interval-60s)
Fencing Levels:
Location Constraints:
Ordering Constraints:
promote drbd_r0_Clone then start lvm_n01_vg0_Clone (Mandatory)
(id:order-drbd_r0_Clone-lvm_n01_vg0_Clone-mandatory)
Colocation Constraints:
Cluster Properties:
cluster-infrastructure: cman
dc-version: 1.1.10-14.el6_5.3-368c726
last-lrm-refresh: 1403062921
no-quorum-policy: ignore
stonith-enabled: true
====
Constraint:
====
Location Constraints:
Ordering Constraints:
promote drbd_r0_Clone then start lvm_n01_vg0_Clone (Mandatory)
(id:order-drbd_r0_Clone-lvm_n01_vg0_Clone-mandatory)
Colocation Constraints:
====
Logs from 'an-a04n01', starting with '/etc/init.d/pacemaker start'
(always survives and fences 'an-a04n02'):
====
Jun 17 23:55:32 an-a04n01 corosync[28088]: [MAIN ] Corosync Cluster
Engine ('1.4.1'): started and ready to provide service.
Jun 17 23:55:32 an-a04n01 corosync[28088]: [MAIN ] Corosync built-in
features: nss dbus rdma snmp
Jun 17 23:55:32 an-a04n01 corosync[28088]: [MAIN ] Successfully read
config from /etc/cluster/cluster.conf
Jun 17 23:55:32 an-a04n01 corosync[28088]: [MAIN ] Successfully
parsed cman config
Jun 17 23:55:32 an-a04n01 corosync[28088]: [TOTEM ] Initializing
transport (UDP/IP Multicast).
Jun 17 23:55:32 an-a04n01 corosync[28088]: [TOTEM ] Initializing
transmit/receive security: libtomcrypt SOBER128/SHA1HMAC (mode 0).
Jun 17 23:55:32 an-a04n01 corosync[28088]: [TOTEM ] The network
interface [10.20.40.1] is now up.
Jun 17 23:55:32 an-a04n01 corosync[28088]: [QUORUM] Using quorum
provider quorum_cman
Jun 17 23:55:32 an-a04n01 corosync[28088]: [SERV ] Service engine
loaded: corosync cluster quorum service v0.1
Jun 17 23:55:32 an-a04n01 corosync[28088]: [CMAN ] CMAN 3.0.12.1
(built Apr 3 2014 05:12:26) started
Jun 17 23:55:32 an-a04n01 corosync[28088]: [SERV ] Service engine
loaded: corosync CMAN membership service 2.90
Jun 17 23:55:32 an-a04n01 corosync[28088]: [SERV ] Service engine
loaded: openais checkpoint service B.01.01
Jun 17 23:55:32 an-a04n01 corosync[28088]: [SERV ] Service engine
loaded: corosync extended virtual synchrony service
Jun 17 23:55:32 an-a04n01 corosync[28088]: [SERV ] Service engine
loaded: corosync configuration service
Jun 17 23:55:32 an-a04n01 corosync[28088]: [SERV ] Service engine
loaded: corosync cluster closed process group service v1.01
Jun 17 23:55:32 an-a04n01 corosync[28088]: [SERV ] Service engine
loaded: corosync cluster config database access v1.01
Jun 17 23:55:32 an-a04n01 corosync[28088]: [SERV ] Service engine
loaded: corosync profile loading service
Jun 17 23:55:32 an-a04n01 corosync[28088]: [QUORUM] Using quorum
provider quorum_cman
Jun 17 23:55:32 an-a04n01 corosync[28088]: [SERV ] Service engine
loaded: corosync cluster quorum service v0.1
Jun 17 23:55:32 an-a04n01 corosync[28088]: [MAIN ] Compatibility mode
set to whitetank. Using V1 and V2 of the synchronization engine.
Jun 17 23:55:32 an-a04n01 corosync[28088]: [TOTEM ] A processor joined
or left the membership and a new membership was formed.
Jun 17 23:55:32 an-a04n01 corosync[28088]: [CMAN ] quorum regained,
resuming activity
Jun 17 23:55:32 an-a04n01 corosync[28088]: [QUORUM] This node is
within the primary component and will provide service.
Jun 17 23:55:32 an-a04n01 corosync[28088]: [QUORUM] Members[1]: 1
Jun 17 23:55:32 an-a04n01 corosync[28088]: [QUORUM] Members[1]: 1
Jun 17 23:55:32 an-a04n01 corosync[28088]: [CPG ] chosen downlist:
sender r(0) ip(10.20.40.1) ; members(old:0 left:0)
Jun 17 23:55:32 an-a04n01 corosync[28088]: [MAIN ] Completed service
synchronization, ready to provide service.
Jun 17 23:55:33 an-a04n01 corosync[28088]: [TOTEM ] A processor joined
or left the membership and a new membership was formed.
Jun 17 23:55:33 an-a04n01 corosync[28088]: [QUORUM] Members[2]: 1 2
Jun 17 23:55:33 an-a04n01 corosync[28088]: [QUORUM] Members[2]: 1 2
Jun 17 23:55:33 an-a04n01 corosync[28088]: [CPG ] chosen downlist:
sender r(0) ip(10.20.40.1) ; members(old:1 left:0)
Jun 17 23:55:33 an-a04n01 corosync[28088]: [MAIN ] Completed service
synchronization, ready to provide service.
Jun 17 23:55:36 an-a04n01 fenced[28143]: fenced 3.0.12.1 started
Jun 17 23:55:36 an-a04n01 dlm_controld[28169]: dlm_controld 3.0.12.1 started
Jun 17 23:55:37 an-a04n01 gfs_controld[28218]: gfs_controld 3.0.12.1 started
Jun 17 23:55:38 an-a04n01 pacemaker: Attempting to start clvmd
Jun 17 23:55:39 an-a04n01 kernel: dlm: Using TCP for communications
Jun 17 23:55:40 an-a04n01 kernel: dlm: connecting to 2
Jun 17 23:55:40 an-a04n01 clvmd: Cluster LVM daemon started - connected
to CMAN
Jun 17 23:55:41 an-a04n01 pacemaker: Starting Pacemaker Cluster Manager
Jun 17 23:55:42 an-a04n01 pacemakerd[28349]: notice: crm_add_logfile:
Additional logging available in /var/log/cluster/corosync.log
Jun 17 23:55:42 an-a04n01 pacemakerd[28349]: notice: main: Starting
Pacemaker 1.1.10-14.el6_5.3 (Build: 368c726): generated-manpages
agent-manpages ascii-docs publican-docs ncurses libqb-logging libqb-ipc
nagios corosync-plugin cman
Jun 17 23:55:42 an-a04n01 cib[28355]: notice: crm_add_logfile:
Additional logging available in /var/log/cluster/corosync.log
Jun 17 23:55:42 an-a04n01 lrmd[28357]: notice: crm_add_logfile:
Additional logging available in /var/log/cluster/corosync.log
Jun 17 23:55:42 an-a04n01 attrd[28358]: notice: crm_add_logfile:
Additional logging available in /var/log/cluster/corosync.log
Jun 17 23:55:42 an-a04n01 pengine[28359]: notice: crm_add_logfile:
Additional logging available in /var/log/cluster/corosync.log
Jun 17 23:55:42 an-a04n01 attrd[28358]: notice: crm_cluster_connect:
Connecting to cluster infrastructure: cman
Jun 17 23:55:42 an-a04n01 crmd[28360]: notice: crm_add_logfile:
Additional logging available in /var/log/cluster/corosync.log
Jun 17 23:55:42 an-a04n01 crmd[28360]: notice: main: CRM Git Version:
368c726
Jun 17 23:55:42 an-a04n01 stonith-ng[28356]: notice: crm_add_logfile:
Additional logging available in /var/log/cluster/corosync.log
Jun 17 23:55:42 an-a04n01 stonith-ng[28356]: notice:
crm_cluster_connect: Connecting to cluster infrastructure: cman
Jun 17 23:55:42 an-a04n01 attrd[28358]: notice: main: Starting mainloop...
Jun 17 23:55:42 an-a04n01 cib[28355]: notice: crm_cluster_connect:
Connecting to cluster infrastructure: cman
Jun 17 23:55:43 an-a04n01 crmd[28360]: notice: crm_cluster_connect:
Connecting to cluster infrastructure: cman
Jun 17 23:55:43 an-a04n01 crmd[28360]: notice: cman_event_callback:
Membership 276: quorum acquired
Jun 17 23:55:43 an-a04n01 crmd[28360]: notice: crm_update_peer_state:
cman_event_callback: Node an-a04n01.alteeve.ca[1] - state is now member
(was (null))
Jun 17 23:55:43 an-a04n01 crmd[28360]: notice: crm_update_peer_state:
cman_event_callback: Node an-a04n02.alteeve.ca[2] - state is now member
(was (null))
Jun 17 23:55:43 an-a04n01 stonith-ng[28356]: notice: setup_cib:
Watching for stonith topology changes
Jun 17 23:55:43 an-a04n01 crmd[28360]: notice: do_started: The local
CRM is operational
Jun 17 23:55:43 an-a04n01 crmd[28360]: notice: do_state_transition:
State transition S_STARTING -> S_PENDING [ input=I_PENDING
cause=C_FSA_INTERNAL origin=do_started ]
Jun 17 23:55:43 an-a04n01 stonith-ng[28356]: notice: unpack_config: On
loss of CCM Quorum: Ignore
Jun 17 23:55:44 an-a04n01 stonith-ng[28356]: notice:
stonith_device_register: Added 'fence_n01_ipmi' to the device list (1
active devices)
Jun 17 23:55:45 an-a04n01 stonith-ng[28356]: notice:
stonith_device_register: Added 'fence_n02_ipmi' to the device list (2
active devices)
Jun 17 23:56:04 an-a04n01 crmd[28360]: notice: do_state_transition:
State transition S_ELECTION -> S_INTEGRATION [ input=I_ELECTION_DC
cause=C_FSA_INTERNAL origin=do_election_check ]
Jun 17 23:56:04 an-a04n01 attrd[28358]: notice: attrd_local_callback:
Sending full refresh (origin=crmd)
Jun 17 23:56:04 an-a04n01 pengine[28359]: notice: unpack_config: On
loss of CCM Quorum: Ignore
Jun 17 23:56:04 an-a04n01 pengine[28359]: notice: LogActions: Start
fence_n01_ipmi#011(an-a04n01.alteeve.ca)
Jun 17 23:56:04 an-a04n01 pengine[28359]: notice: LogActions: Start
fence_n02_ipmi#011(an-a04n02.alteeve.ca)
Jun 17 23:56:04 an-a04n01 pengine[28359]: notice: LogActions: Start
drbd_r0:0#011(an-a04n01.alteeve.ca)
Jun 17 23:56:04 an-a04n01 pengine[28359]: notice: LogActions: Start
drbd_r0:1#011(an-a04n02.alteeve.ca)
Jun 17 23:56:04 an-a04n01 pengine[28359]: notice: LogActions: Start
lvm_n01_vg0:0#011(an-a04n01.alteeve.ca - blocked)
Jun 17 23:56:04 an-a04n01 pengine[28359]: notice: LogActions: Start
lvm_n01_vg0:1#011(an-a04n02.alteeve.ca - blocked)
Jun 17 23:56:04 an-a04n01 pengine[28359]: notice: process_pe_message:
Calculated Transition 0: /var/lib/pacemaker/pengine/pe-input-152.bz2
Jun 17 23:56:04 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 9: monitor fence_n01_ipmi_monitor_0 on
an-a04n02.alteeve.ca
Jun 17 23:56:04 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 4: monitor fence_n01_ipmi_monitor_0 on
an-a04n01.alteeve.ca (local)
Jun 17 23:56:04 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 10: monitor fence_n02_ipmi_monitor_0 on
an-a04n02.alteeve.ca
Jun 17 23:56:04 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 5: monitor fence_n02_ipmi_monitor_0 on
an-a04n01.alteeve.ca (local)
Jun 17 23:56:04 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 6: monitor drbd_r0:0_monitor_0 on an-a04n01.alteeve.ca
(local)
Jun 17 23:56:04 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 11: monitor drbd_r0:1_monitor_0 on an-a04n02.alteeve.ca
Jun 17 23:56:04 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 7: monitor lvm_n01_vg0:0_monitor_0 on
an-a04n01.alteeve.ca (local)
Jun 17 23:56:04 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 12: monitor lvm_n01_vg0:1_monitor_0 on
an-a04n02.alteeve.ca
Jun 17 23:56:04 an-a04n01 LVM(lvm_n01_vg0)[28419]: WARNING: LVM Volume
an-a04n01_vg0 is not available (stopped)
Jun 17 23:56:05 an-a04n01 crmd[28360]: notice: process_lrm_event: LRM
operation lvm_n01_vg0_monitor_0 (call=19, rc=7, cib-update=28,
confirmed=true) not running
Jun 17 23:56:05 an-a04n01 crmd[28360]: notice: process_lrm_event: LRM
operation drbd_r0_monitor_0 (call=14, rc=7, cib-update=29,
confirmed=true) not running
Jun 17 23:56:05 an-a04n01 crmd[28360]: notice: process_lrm_event:
an-a04n01.alteeve.ca-drbd_r0_monitor_0:14 [ \n ]
Jun 17 23:56:05 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 3: probe_complete probe_complete on
an-a04n01.alteeve.ca (local) - no waiting
Jun 17 23:56:05 an-a04n01 attrd[28358]: notice: attrd_trigger_update:
Sending flush op to all hosts for: probe_complete (true)
Jun 17 23:56:05 an-a04n01 attrd[28358]: notice: attrd_perform_update:
Sent update 4: probe_complete=true
Jun 17 23:56:05 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 8: probe_complete probe_complete on
an-a04n02.alteeve.ca - no waiting
Jun 17 23:56:05 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 13: start fence_n01_ipmi_start_0 on
an-a04n01.alteeve.ca (local)
Jun 17 23:56:05 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 15: start fence_n02_ipmi_start_0 on an-a04n02.alteeve.ca
Jun 17 23:56:05 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 17: start drbd_r0:0_start_0 on an-a04n01.alteeve.ca
(local)
Jun 17 23:56:06 an-a04n01 stonith-ng[28356]: notice:
stonith_device_register: Device 'fence_n01_ipmi' already existed in
device list (2 active devices)
Jun 17 23:56:06 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 19: start drbd_r0:1_start_0 on an-a04n02.alteeve.ca
Jun 17 23:56:06 an-a04n01 crmd[28360]: notice: process_lrm_event: LRM
operation fence_n01_ipmi_start_0 (call=25, rc=0, cib-update=30,
confirmed=true) ok
Jun 17 23:56:06 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 14: monitor fence_n01_ipmi_monitor_60000 on
an-a04n01.alteeve.ca (local)
Jun 17 23:56:06 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 16: monitor fence_n02_ipmi_monitor_60000 on
an-a04n02.alteeve.ca
Jun 17 23:56:06 an-a04n01 crmd[28360]: notice: process_lrm_event: LRM
operation fence_n01_ipmi_monitor_60000 (call=30, rc=0, cib-update=31,
confirmed=false) ok
Jun 17 23:56:06 an-a04n01 kernel: block drbd0: Starting worker thread
(from cqueue [3274])
Jun 17 23:56:06 an-a04n01 kernel: block drbd0: disk( Diskless ->
Attaching )
Jun 17 23:56:06 an-a04n01 kernel: block drbd0: Found 4 transactions (126
active extents) in activity log.
Jun 17 23:56:06 an-a04n01 kernel: block drbd0: Method to ensure write
ordering: flush
Jun 17 23:56:06 an-a04n01 kernel: block drbd0: drbd_bm_resize called
with capacity == 909525832
Jun 17 23:56:06 an-a04n01 kernel: block drbd0: resync bitmap:
bits=113690729 words=1776418 pages=3470
Jun 17 23:56:06 an-a04n01 kernel: block drbd0: size = 434 GB (454762916 KB)
Jun 17 23:56:06 an-a04n01 kernel: block drbd0: bitmap READ of 3470 pages
took 9 jiffies
Jun 17 23:56:06 an-a04n01 kernel: block drbd0: recounting of set bits
took additional 16 jiffies
Jun 17 23:56:06 an-a04n01 kernel: block drbd0: 0 KB (0 bits) marked
out-of-sync by on disk bit-map.
Jun 17 23:56:06 an-a04n01 kernel: block drbd0: disk( Attaching ->
Consistent )
Jun 17 23:56:06 an-a04n01 kernel: block drbd0: attached to UUIDs
C71081B1CBAFC620:0000000000000000:F9F9DA52F6D93990:F9F8DA52F6D93991
Jun 17 23:56:06 an-a04n01 kernel: block drbd0: conn( StandAlone ->
Unconnected )
Jun 17 23:56:06 an-a04n01 kernel: block drbd0: Starting receiver thread
(from drbd0_worker [28524])
Jun 17 23:56:06 an-a04n01 kernel: block drbd0: receiver (re)started
Jun 17 23:56:06 an-a04n01 kernel: block drbd0: conn( Unconnected ->
WFConnection )
Jun 17 23:56:06 an-a04n01 attrd[28358]: notice: attrd_trigger_update:
Sending flush op to all hosts for: master-drbd_r0 (5)
Jun 17 23:56:06 an-a04n01 attrd[28358]: notice: attrd_perform_update:
Sent update 9: master-drbd_r0=5
Jun 17 23:56:06 an-a04n01 crmd[28360]: notice: process_lrm_event: LRM
operation drbd_r0_start_0 (call=27, rc=0, cib-update=32, confirmed=true) ok
Jun 17 23:56:06 an-a04n01 attrd[28358]: notice: attrd_perform_update:
Sent update 11: master-drbd_r0=5
Jun 17 23:56:06 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 82: notify drbd_r0:0_post_notify_start_0 on
an-a04n01.alteeve.ca (local)
Jun 17 23:56:06 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 83: notify drbd_r0:1_post_notify_start_0 on
an-a04n02.alteeve.ca
Jun 17 23:56:06 an-a04n01 crmd[28360]: notice: process_lrm_event: LRM
operation drbd_r0_notify_0 (call=34, rc=0, cib-update=0, confirmed=true) ok
Jun 17 23:56:06 an-a04n01 crmd[28360]: notice: run_graph: Transition 0
(Complete=25, Pending=0, Fired=0, Skipped=2, Incomplete=0,
Source=/var/lib/pacemaker/pengine/pe-input-152.bz2): Stopped
Jun 17 23:56:06 an-a04n01 pengine[28359]: notice: unpack_config: On
loss of CCM Quorum: Ignore
Jun 17 23:56:06 an-a04n01 pengine[28359]: notice: LogActions: Promote
drbd_r0:0#011(Slave -> Master an-a04n01.alteeve.ca)
Jun 17 23:56:06 an-a04n01 pengine[28359]: notice: LogActions: Promote
drbd_r0:1#011(Slave -> Master an-a04n02.alteeve.ca)
Jun 17 23:56:06 an-a04n01 pengine[28359]: notice: LogActions: Start
lvm_n01_vg0:0#011(an-a04n01.alteeve.ca)
Jun 17 23:56:06 an-a04n01 pengine[28359]: notice: LogActions: Start
lvm_n01_vg0:1#011(an-a04n02.alteeve.ca)
Jun 17 23:56:06 an-a04n01 pengine[28359]: notice: process_pe_message:
Calculated Transition 1: /var/lib/pacemaker/pengine/pe-input-153.bz2
Jun 17 23:56:06 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 84: notify drbd_r0_pre_notify_promote_0 on
an-a04n01.alteeve.ca (local)
Jun 17 23:56:06 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 86: notify drbd_r0_pre_notify_promote_0 on
an-a04n02.alteeve.ca
Jun 17 23:56:06 an-a04n01 crmd[28360]: notice: process_lrm_event: LRM
operation drbd_r0_notify_0 (call=37, rc=0, cib-update=0, confirmed=true) ok
Jun 17 23:56:06 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 13: promote drbd_r0_promote_0 on an-a04n01.alteeve.ca
(local)
Jun 17 23:56:06 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 16: promote drbd_r0_promote_0 on an-a04n02.alteeve.ca
Jun 17 23:56:06 an-a04n01 kernel: block drbd0: helper command:
/sbin/drbdadm fence-peer minor-0
Jun 17 23:56:07 an-a04n01 kernel: block drbd0: Handshake successful:
Agreed network protocol version 97
Jun 17 23:56:07 an-a04n01 stonith_admin[28637]: notice: crm_log_args:
Invoked: stonith_admin --fence an-a04n02.alteeve.ca
Jun 17 23:56:07 an-a04n01 stonith-ng[28356]: notice: handle_request:
Client stonith_admin.28637.6ed13ba6 wants to fence (off)
'an-a04n02.alteeve.ca' with device '(any)'
Jun 17 23:56:07 an-a04n01 stonith-ng[28356]: notice:
initiate_remote_stonith_op: Initiating remote operation off for
an-a04n02.alteeve.ca: 382bfa3d-55da-4eed-ad8a-a1a883022a35 (0)
Jun 17 23:56:07 an-a04n01 stonith-ng[28356]: notice:
can_fence_host_with_device: fence_n02_ipmi can fence
an-a04n02.alteeve.ca: static-list
Jun 17 23:56:07 an-a04n01 stonith-ng[28356]: notice:
can_fence_host_with_device: fence_n01_ipmi can not fence
an-a04n02.alteeve.ca: static-list
Jun 17 23:56:07 an-a04n01 stonith-ng[28356]: notice:
can_fence_host_with_device: fence_n02_ipmi can fence
an-a04n02.alteeve.ca: static-list
Jun 17 23:56:07 an-a04n01 stonith-ng[28356]: notice:
can_fence_host_with_device: fence_n01_ipmi can not fence
an-a04n02.alteeve.ca: static-list
Jun 17 23:56:07 an-a04n01 stonith-ng[28356]: notice:
can_fence_host_with_device: fence_n02_ipmi can not fence
an-a04n01.alteeve.ca: static-list
Jun 17 23:56:07 an-a04n01 stonith-ng[28356]: notice:
can_fence_host_with_device: fence_n01_ipmi can fence
an-a04n01.alteeve.ca: static-list
Jun 17 23:56:23 an-a04n01 stonith-ng[28356]: notice: log_operation:
Operation 'off' [28638] (call 2 from stonith_admin.28637) for host
'an-a04n02.alteeve.ca' with device 'fence_n02_ipmi' returned: 0 (OK)
Jun 17 23:56:25 an-a04n01 corosync[28088]: [TOTEM ] A processor
failed, forming new configuration.
Jun 17 23:56:26 an-a04n01 lrmd[28357]: warning: child_timeout_callback:
drbd_r0_promote_0 process (PID 28604) timed out
Jun 17 23:56:26 an-a04n01 lrmd[28357]: warning: operation_finished:
drbd_r0_promote_0:28604 - timed out after 20000ms
Jun 17 23:56:26 an-a04n01 crmd[28360]: error: process_lrm_event: LRM
operation drbd_r0_promote_0 (40) Timed Out (timeout=20000ms)
Jun 17 23:56:26 an-a04n01 crmd[28360]: notice: process_lrm_event:
an-a04n01.alteeve.ca-drbd_r0_promote_0:40 [ allow-two-primaries;\n ]
Jun 17 23:56:26 an-a04n01 crmd[28360]: warning: status_from_rc: Action
13 (drbd_r0_promote_0) on an-a04n01.alteeve.ca failed (target: 0 vs. rc:
1): Error
Jun 17 23:56:26 an-a04n01 crmd[28360]: warning: update_failcount:
Updating failcount for drbd_r0 on an-a04n01.alteeve.ca after failed
promote: rc=1 (update=value++, time=1403063786)
Jun 17 23:56:26 an-a04n01 crmd[28360]: warning: update_failcount:
Updating failcount for drbd_r0 on an-a04n01.alteeve.ca after failed
promote: rc=1 (update=value++, time=1403063786)
Jun 17 23:56:26 an-a04n01 attrd[28358]: notice: attrd_trigger_update:
Sending flush op to all hosts for: fail-count-drbd_r0 (1)
Jun 17 23:56:26 an-a04n01 attrd[28358]: notice: attrd_perform_update:
Sent update 14: fail-count-drbd_r0=1
Jun 17 23:56:26 an-a04n01 attrd[28358]: notice: attrd_trigger_update:
Sending flush op to all hosts for: last-failure-drbd_r0 (1403063786)
Jun 17 23:56:26 an-a04n01 attrd[28358]: notice: attrd_perform_update:
Sent update 17: last-failure-drbd_r0=1403063786
Jun 17 23:56:26 an-a04n01 attrd[28358]: notice: attrd_trigger_update:
Sending flush op to all hosts for: fail-count-drbd_r0 (2)
Jun 17 23:56:26 an-a04n01 attrd[28358]: notice: attrd_perform_update:
Sent update 19: fail-count-drbd_r0=2
Jun 17 23:56:26 an-a04n01 attrd[28358]: notice: attrd_trigger_update:
Sending flush op to all hosts for: last-failure-drbd_r0 (1403063786)
Jun 17 23:56:26 an-a04n01 attrd[28358]: notice: attrd_perform_update:
Sent update 21: last-failure-drbd_r0=1403063786
Jun 17 23:56:27 an-a04n01 corosync[28088]: [QUORUM] Members[1]: 1
Jun 17 23:56:27 an-a04n01 corosync[28088]: [TOTEM ] A processor joined
or left the membership and a new membership was formed.
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice: crm_update_peer_state:
cman_event_callback: Node an-a04n02.alteeve.ca[2] - state is now lost
(was member)
Jun 17 23:56:27 an-a04n01 crmd[28360]: warning: match_down_event: No
match for shutdown action on an-a04n02.alteeve.ca
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice: peer_update_callback:
Stonith/shutdown of an-a04n02.alteeve.ca not matched
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice:
fail_incompletable_actions: Action 87 (87) is scheduled for
an-a04n02.alteeve.ca (offline)
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice:
fail_incompletable_actions: Action 16 (16) was pending on
an-a04n02.alteeve.ca (offline)
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice:
fail_incompletable_actions: Action 89 (89) is scheduled for
an-a04n02.alteeve.ca (offline)
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice:
fail_incompletable_actions: Action 44 (44) is scheduled for
an-a04n02.alteeve.ca (offline)
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice:
fail_incompletable_actions: Action 43 (43) is scheduled for
an-a04n02.alteeve.ca (offline)
Jun 17 23:56:27 an-a04n01 crmd[28360]: warning:
fail_incompletable_actions: Node an-a04n02.alteeve.ca shutdown resulted
in un-runnable actions
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice:
fail_incompletable_actions: Action 87 (87) is scheduled for
an-a04n02.alteeve.ca (offline)
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice:
fail_incompletable_actions: Action 16 (16) was pending on
an-a04n02.alteeve.ca (offline)
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice:
fail_incompletable_actions: Action 89 (89) is scheduled for
an-a04n02.alteeve.ca (offline)
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice:
fail_incompletable_actions: Action 44 (44) is scheduled for
an-a04n02.alteeve.ca (offline)
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice:
fail_incompletable_actions: Action 43 (43) is scheduled for
an-a04n02.alteeve.ca (offline)
Jun 17 23:56:27 an-a04n01 kernel: dlm: closing connection to node 2
Jun 17 23:56:27 an-a04n01 crmd[28360]: warning:
fail_incompletable_actions: Node an-a04n02.alteeve.ca shutdown resulted
in un-runnable actions
Jun 17 23:56:27 an-a04n01 corosync[28088]: [CPG ] chosen downlist:
sender r(0) ip(10.20.40.1) ; members(old:2 left:1)
Jun 17 23:56:27 an-a04n01 corosync[28088]: [MAIN ] Completed service
synchronization, ready to provide service.
Jun 17 23:56:27 an-a04n01 stonith-ng[28356]: notice: remote_op_done:
Operation off of an-a04n02.alteeve.ca by an-a04n01.alteeve.ca for
stonith_admin.28637 at an-a04n01.alteeve.ca.382bfa3d: OK
Jun 17 23:56:27 an-a04n01 attrd[28358]: notice: attrd_local_callback:
Sending full refresh (origin=crmd)
Jun 17 23:56:27 an-a04n01 attrd[28358]: notice: attrd_trigger_update:
Sending flush op to all hosts for: master-drbd_r0 (5)
Jun 17 23:56:27 an-a04n01 crmd[28360]: warning: match_down_event: No
match for shutdown action on an-a04n02.alteeve.ca
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice: peer_update_callback:
Stonith/shutdown of an-a04n02.alteeve.ca not matched
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice:
fail_incompletable_actions: Action 87 (87) is scheduled for
an-a04n02.alteeve.ca (offline)
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice:
fail_incompletable_actions: Action 16 (16) was pending on
an-a04n02.alteeve.ca (offline)
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice:
fail_incompletable_actions: Action 89 (89) is scheduled for
an-a04n02.alteeve.ca (offline)
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice:
fail_incompletable_actions: Action 44 (44) is scheduled for
an-a04n02.alteeve.ca (offline)
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice:
fail_incompletable_actions: Action 43 (43) is scheduled for
an-a04n02.alteeve.ca (offline)
Jun 17 23:56:27 an-a04n01 crmd[28360]: warning:
fail_incompletable_actions: Node an-a04n02.alteeve.ca shutdown resulted
in un-runnable actions
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice: tengine_stonith_notify:
Peer an-a04n02.alteeve.ca was terminated (off) by an-a04n01.alteeve.ca
for an-a04n01.alteeve.ca: OK (ref=382bfa3d-55da-4eed-ad8a-a1a883022a35)
by client stonith_admin.28637
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice: tengine_stonith_notify:
Notified CMAN that 'an-a04n02.alteeve.ca' is now fenced
Jun 17 23:56:27 an-a04n01 fenced[28143]: fencing node an-a04n02.alteeve.ca
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 85: notify drbd_r0_post_notify_promote_0 on
an-a04n01.alteeve.ca (local)
Jun 17 23:56:27 an-a04n01 stonith_admin-fence-peer.sh[28708]:
stonith_admin successfully fenced peer an-a04n02.alteeve.ca.
Jun 17 23:56:27 an-a04n01 kernel: block drbd0: helper command:
/sbin/drbdadm fence-peer minor-0 exit code 7 (0x700)
Jun 17 23:56:27 an-a04n01 kernel: block drbd0: fence-peer helper
returned 7 (peer was stonithed)
Jun 17 23:56:27 an-a04n01 kernel: block drbd0: role( Secondary ->
Primary ) disk( Consistent -> UpToDate ) pdsk( DUnknown -> Outdated )
Jun 17 23:56:27 an-a04n01 kernel: block drbd0: new current UUID
B704B7175D09E91D:C71081B1CBAFC620:F9F9DA52F6D93990:F9F8DA52F6D93991
Jun 17 23:56:27 an-a04n01 kernel: block drbd0: conn( WFConnection ->
WFReportParams )
Jun 17 23:56:27 an-a04n01 kernel: block drbd0: Starting asender thread
(from drbd0_receiver [28542])
Jun 17 23:56:27 an-a04n01 kernel: block drbd0: data-integrity-alg:
<not-used>
Jun 17 23:56:27 an-a04n01 attrd[28358]: notice: attrd_trigger_update:
Sending flush op to all hosts for: last-failure-drbd_r0 (1403063786)
Jun 17 23:56:27 an-a04n01 attrd[28358]: notice: attrd_trigger_update:
Sending flush op to all hosts for: fail-count-drbd_r0 (2)
Jun 17 23:56:27 an-a04n01 attrd[28358]: notice: attrd_perform_update:
Sent update 27: fail-count-drbd_r0=2
Jun 17 23:56:27 an-a04n01 attrd[28358]: notice: attrd_trigger_update:
Sending flush op to all hosts for: probe_complete (true)
Jun 17 23:56:27 an-a04n01 attrd[28358]: notice: attrd_trigger_update:
Sending flush op to all hosts for: master-drbd_r0 (10000)
Jun 17 23:56:27 an-a04n01 attrd[28358]: notice: attrd_perform_update:
Sent update 31: master-drbd_r0=10000
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice: process_lrm_event: LRM
operation drbd_r0_notify_0 (call=43, rc=0, cib-update=0, confirmed=true) ok
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice: run_graph: Transition 1
(Complete=12, Pending=0, Fired=0, Skipped=8, Incomplete=4,
Source=/var/lib/pacemaker/pengine/pe-input-153.bz2): Stopped
Jun 17 23:56:27 an-a04n01 pengine[28359]: notice: unpack_config: On
loss of CCM Quorum: Ignore
Jun 17 23:56:27 an-a04n01 pengine[28359]: warning: unpack_rsc_op:
Processing failed op promote for drbd_r0:0 on an-a04n01.alteeve.ca:
unknown error (1)
Jun 17 23:56:27 an-a04n01 pengine[28359]: notice: LogActions: Start
fence_n02_ipmi#011(an-a04n01.alteeve.ca)
Jun 17 23:56:27 an-a04n01 pengine[28359]: notice: LogActions: Demote
drbd_r0:0#011(Master -> Slave an-a04n01.alteeve.ca)
Jun 17 23:56:27 an-a04n01 pengine[28359]: notice: LogActions: Recover
drbd_r0:0#011(Master an-a04n01.alteeve.ca)
Jun 17 23:56:27 an-a04n01 pengine[28359]: notice: LogActions: Start
lvm_n01_vg0:0#011(an-a04n01.alteeve.ca - blocked)
Jun 17 23:56:27 an-a04n01 pengine[28359]: notice: process_pe_message:
Calculated Transition 2: /var/lib/pacemaker/pengine/pe-input-154.bz2
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 8: start fence_n02_ipmi_start_0 on
an-a04n01.alteeve.ca (local)
Jun 17 23:56:27 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 75: notify drbd_r0_pre_notify_demote_0 on
an-a04n01.alteeve.ca (local)
Jun 17 23:56:27 an-a04n01 fence_pcmk[28761]: Requesting Pacemaker fence
an-a04n02.alteeve.ca (reset)
Jun 17 23:56:27 an-a04n01 stonith-ng[28356]: notice:
stonith_device_register: Device 'fence_n02_ipmi' already existed in
device list (2 active devices)
Jun 17 23:56:27 an-a04n01 stonith_admin[28763]: notice: crm_log_args:
Invoked: stonith_admin --reboot an-a04n02.alteeve.ca --tolerance 5s
--tag cman
Jun 17 23:56:27 an-a04n01 stonith-ng[28356]: notice: handle_request:
Client stonith_admin.cman.28763.4e2c3020 wants to fence (reboot)
'an-a04n02.alteeve.ca' with device '(any)'
Jun 17 23:56:27 an-a04n01 stonith-ng[28356]: notice:
initiate_remote_stonith_op: Initiating remote operation reboot for
an-a04n02.alteeve.ca: bbb6c5c4-d1a7-4df7-a8b0-e33f4ad74860 (0)
Jun 17 23:56:27 an-a04n01 stonith-ng[28356]: notice:
can_fence_host_with_device: fence_n02_ipmi can fence
an-a04n02.alteeve.ca: static-list
Jun 17 23:56:27 an-a04n01 stonith-ng[28356]: notice:
can_fence_host_with_device: fence_n01_ipmi can not fence
an-a04n02.alteeve.ca: static-list
Jun 17 23:56:27 an-a04n01 stonith-ng[28356]: notice:
can_fence_host_with_device: fence_n02_ipmi can fence
an-a04n02.alteeve.ca: static-list
Jun 17 23:56:27 an-a04n01 stonith-ng[28356]: notice:
can_fence_host_with_device: fence_n01_ipmi can not fence
an-a04n02.alteeve.ca: static-list
Jun 17 23:56:27 an-a04n01 kernel: block drbd0: PingAck did not arrive in
time.
Jun 17 23:56:27 an-a04n01 kernel: block drbd0: conn( WFReportParams ->
NetworkFailure )
Jun 17 23:56:27 an-a04n01 kernel: block drbd0: asender terminated
Jun 17 23:56:27 an-a04n01 kernel: block drbd0: Terminating drbd0_asender
Jun 17 23:56:27 an-a04n01 kernel: block drbd0: Connection closed
Jun 17 23:56:27 an-a04n01 kernel: block drbd0: conn( NetworkFailure ->
Unconnected )
Jun 17 23:56:27 an-a04n01 kernel: block drbd0: receiver terminated
Jun 17 23:56:27 an-a04n01 kernel: block drbd0: Restarting drbd0_receiver
Jun 17 23:56:27 an-a04n01 kernel: block drbd0: receiver (re)started
Jun 17 23:56:27 an-a04n01 kernel: block drbd0: conn( Unconnected ->
WFConnection )
Jun 17 23:56:28 an-a04n01 crmd[28360]: notice: process_lrm_event: LRM
operation fence_n02_ipmi_start_0 (call=46, rc=0, cib-update=48,
confirmed=true) ok
Jun 17 23:56:28 an-a04n01 crmd[28360]: notice: process_lrm_event: LRM
operation drbd_r0_notify_0 (call=48, rc=0, cib-update=0, confirmed=true) ok
Jun 17 23:56:28 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 10: demote drbd_r0_demote_0 on an-a04n01.alteeve.ca
(local)
Jun 17 23:56:28 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 9: monitor fence_n02_ipmi_monitor_60000 on
an-a04n01.alteeve.ca (local)
Jun 17 23:56:28 an-a04n01 kernel: block drbd0: role( Primary -> Secondary )
Jun 17 23:56:28 an-a04n01 kernel: block drbd0: bitmap WRITE of 0 pages
took 0 jiffies
Jun 17 23:56:28 an-a04n01 kernel: block drbd0: 0 KB (0 bits) marked
out-of-sync by on disk bit-map.
Jun 17 23:56:28 an-a04n01 crmd[28360]: notice: process_lrm_event: LRM
operation drbd_r0_demote_0 (call=52, rc=0, cib-update=49, confirmed=true) ok
Jun 17 23:56:28 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 76: notify drbd_r0_post_notify_demote_0 on
an-a04n01.alteeve.ca (local)
Jun 17 23:56:28 an-a04n01 crmd[28360]: notice: process_lrm_event: LRM
operation drbd_r0_notify_0 (call=57, rc=0, cib-update=0, confirmed=true) ok
Jun 17 23:56:28 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 74: notify drbd_r0_pre_notify_stop_0 on
an-a04n01.alteeve.ca (local)
Jun 17 23:56:28 an-a04n01 crmd[28360]: notice: process_lrm_event: LRM
operation drbd_r0_notify_0 (call=60, rc=0, cib-update=0, confirmed=true) ok
Jun 17 23:56:28 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 2: stop drbd_r0_stop_0 on an-a04n01.alteeve.ca (local)
Jun 17 23:56:28 an-a04n01 kernel: block drbd0: conn( WFConnection ->
Disconnecting )
Jun 17 23:56:28 an-a04n01 kernel: block drbd0: Discarding network
configuration.
Jun 17 23:56:28 an-a04n01 kernel: block drbd0: Connection closed
Jun 17 23:56:28 an-a04n01 kernel: block drbd0: conn( Disconnecting ->
StandAlone )
Jun 17 23:56:28 an-a04n01 kernel: block drbd0: receiver terminated
Jun 17 23:56:28 an-a04n01 kernel: block drbd0: Terminating drbd0_receiver
Jun 17 23:56:28 an-a04n01 kernel: block drbd0: disk( UpToDate -> Failed )
Jun 17 23:56:28 an-a04n01 kernel: block drbd0: bitmap WRITE of 0 pages
took 0 jiffies
Jun 17 23:56:28 an-a04n01 kernel: block drbd0: 0 KB (0 bits) marked
out-of-sync by on disk bit-map.
Jun 17 23:56:28 an-a04n01 kernel: block drbd0: disk( Failed -> Diskless )
Jun 17 23:56:28 an-a04n01 kernel: block drbd0: drbd_bm_resize called
with capacity == 0
Jun 17 23:56:28 an-a04n01 kernel: block drbd0: worker terminated
Jun 17 23:56:28 an-a04n01 kernel: block drbd0: Terminating drbd0_worker
Jun 17 23:56:28 an-a04n01 attrd[28358]: notice: attrd_trigger_update:
Sending flush op to all hosts for: master-drbd_r0 (<null>)
Jun 17 23:56:28 an-a04n01 attrd[28358]: notice: attrd_perform_update:
Sent delete 33: node=an-a04n01.alteeve.ca, attr=master-drbd_r0,
id=<n/a>, set=(null), section=status
Jun 17 23:56:28 an-a04n01 crmd[28360]: notice: process_lrm_event: LRM
operation drbd_r0_stop_0 (call=63, rc=0, cib-update=50, confirmed=true) ok
Jun 17 23:56:43 an-a04n01 stonith-ng[28356]: notice: log_operation:
Operation 'reboot' [28771] (call 2 from stonith_admin.cman.28763) for
host 'an-a04n02.alteeve.ca' with device 'fence_n02_ipmi' returned: 0 (OK)
Jun 17 23:56:43 an-a04n01 stonith-ng[28356]: notice: remote_op_done:
Operation reboot of an-a04n02.alteeve.ca by an-a04n01.alteeve.ca for
stonith_admin.cman.28763 at an-a04n01.alteeve.ca.bbb6c5c4: OK
Jun 17 23:56:43 an-a04n01 crmd[28360]: notice: tengine_stonith_notify:
Peer an-a04n02.alteeve.ca was terminated (reboot) by
an-a04n01.alteeve.ca for an-a04n01.alteeve.ca: OK
(ref=bbb6c5c4-d1a7-4df7-a8b0-e33f4ad74860) by client
stonith_admin.cman.28763
Jun 17 23:56:43 an-a04n01 crmd[28360]: notice: tengine_stonith_notify:
Notified CMAN that 'an-a04n02.alteeve.ca' is now fenced
Jun 17 23:56:43 an-a04n01 fenced[28143]: fence an-a04n02.alteeve.ca success
Jun 17 23:56:43 an-a04n01 crmd[28360]: notice: process_lrm_event: LRM
operation fence_n02_ipmi_monitor_60000 (call=54, rc=0, cib-update=54,
confirmed=false) ok
Jun 17 23:56:43 an-a04n01 crmd[28360]: notice: run_graph: Transition 2
(Complete=19, Pending=0, Fired=0, Skipped=6, Incomplete=4,
Source=/var/lib/pacemaker/pengine/pe-input-154.bz2): Stopped
Jun 17 23:56:43 an-a04n01 pengine[28359]: notice: unpack_config: On
loss of CCM Quorum: Ignore
Jun 17 23:56:43 an-a04n01 pengine[28359]: warning: unpack_rsc_op:
Processing failed op promote for drbd_r0:0 on an-a04n01.alteeve.ca:
unknown error (1)
Jun 17 23:56:43 an-a04n01 pengine[28359]: notice: LogActions: Start
drbd_r0:0#011(an-a04n01.alteeve.ca)
Jun 17 23:56:43 an-a04n01 pengine[28359]: notice: LogActions: Start
lvm_n01_vg0:0#011(an-a04n01.alteeve.ca - blocked)
Jun 17 23:56:43 an-a04n01 pengine[28359]: notice: process_pe_message:
Calculated Transition 3: /var/lib/pacemaker/pengine/pe-input-155.bz2
Jun 17 23:56:43 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 10: start drbd_r0_start_0 on an-a04n01.alteeve.ca (local)
Jun 17 23:56:43 an-a04n01 kernel: block drbd0: Starting worker thread
(from cqueue [3274])
Jun 17 23:56:43 an-a04n01 kernel: block drbd0: disk( Diskless ->
Attaching )
Jun 17 23:56:43 an-a04n01 kernel: block drbd0: Found 4 transactions (126
active extents) in activity log.
Jun 17 23:56:43 an-a04n01 kernel: block drbd0: Method to ensure write
ordering: flush
Jun 17 23:56:43 an-a04n01 kernel: block drbd0: drbd_bm_resize called
with capacity == 909525832
Jun 17 23:56:43 an-a04n01 kernel: block drbd0: resync bitmap:
bits=113690729 words=1776418 pages=3470
Jun 17 23:56:43 an-a04n01 kernel: block drbd0: size = 434 GB (454762916 KB)
Jun 17 23:56:43 an-a04n01 kernel: block drbd0: bitmap READ of 3470 pages
took 9 jiffies
Jun 17 23:56:43 an-a04n01 kernel: block drbd0: recounting of set bits
took additional 16 jiffies
Jun 17 23:56:43 an-a04n01 kernel: block drbd0: 0 KB (0 bits) marked
out-of-sync by on disk bit-map.
Jun 17 23:56:43 an-a04n01 kernel: block drbd0: disk( Attaching ->
UpToDate ) pdsk( DUnknown -> Outdated )
Jun 17 23:56:43 an-a04n01 kernel: block drbd0: attached to UUIDs
B704B7175D09E91D:C71081B1CBAFC620:F9F9DA52F6D93990:F9F8DA52F6D93991
Jun 17 23:56:43 an-a04n01 kernel: block drbd0: conn( StandAlone ->
Unconnected )
Jun 17 23:56:43 an-a04n01 kernel: block drbd0: Starting receiver thread
(from drbd0_worker [29023])
Jun 17 23:56:43 an-a04n01 kernel: block drbd0: receiver (re)started
Jun 17 23:56:43 an-a04n01 kernel: block drbd0: conn( Unconnected ->
WFConnection )
Jun 17 23:56:43 an-a04n01 attrd[28358]: notice: attrd_trigger_update:
Sending flush op to all hosts for: master-drbd_r0 (10000)
Jun 17 23:56:43 an-a04n01 attrd[28358]: notice: attrd_perform_update:
Sent update 37: master-drbd_r0=10000
Jun 17 23:56:43 an-a04n01 crmd[28360]: notice: process_lrm_event: LRM
operation drbd_r0_start_0 (call=67, rc=0, cib-update=56, confirmed=true) ok
Jun 17 23:56:43 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 73: notify drbd_r0_post_notify_start_0 on
an-a04n01.alteeve.ca (local)
Jun 17 23:56:43 an-a04n01 crmd[28360]: notice: process_lrm_event: LRM
operation drbd_r0_notify_0 (call=70, rc=0, cib-update=0, confirmed=true) ok
Jun 17 23:56:43 an-a04n01 crmd[28360]: notice: run_graph: Transition 3
(Complete=8, Pending=0, Fired=0, Skipped=1, Incomplete=0,
Source=/var/lib/pacemaker/pengine/pe-input-155.bz2): Stopped
Jun 17 23:56:43 an-a04n01 pengine[28359]: notice: unpack_config: On
loss of CCM Quorum: Ignore
Jun 17 23:56:43 an-a04n01 pengine[28359]: warning: unpack_rsc_op:
Processing failed op promote for drbd_r0:0 on an-a04n01.alteeve.ca:
unknown error (1)
Jun 17 23:56:43 an-a04n01 pengine[28359]: notice: LogActions: Promote
drbd_r0:0#011(Slave -> Master an-a04n01.alteeve.ca)
Jun 17 23:56:43 an-a04n01 pengine[28359]: notice: LogActions: Start
lvm_n01_vg0:0#011(an-a04n01.alteeve.ca)
Jun 17 23:56:43 an-a04n01 pengine[28359]: notice: process_pe_message:
Calculated Transition 4: /var/lib/pacemaker/pengine/pe-input-156.bz2
Jun 17 23:56:43 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 77: notify drbd_r0_pre_notify_promote_0 on
an-a04n01.alteeve.ca (local)
Jun 17 23:56:43 an-a04n01 crmd[28360]: notice: process_lrm_event: LRM
operation drbd_r0_notify_0 (call=73, rc=0, cib-update=0, confirmed=true) ok
Jun 17 23:56:43 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 12: promote drbd_r0_promote_0 on an-a04n01.alteeve.ca
(local)
Jun 17 23:56:43 an-a04n01 kernel: block drbd0: role( Secondary -> Primary )
Jun 17 23:56:43 an-a04n01 crmd[28360]: notice: process_lrm_event: LRM
operation drbd_r0_promote_0 (call=76, rc=0, cib-update=58,
confirmed=true) ok
Jun 17 23:56:43 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 78: notify drbd_r0_post_notify_promote_0 on
an-a04n01.alteeve.ca (local)
Jun 17 23:56:43 an-a04n01 crmd[28360]: notice: process_lrm_event: LRM
operation drbd_r0_notify_0 (call=79, rc=0, cib-update=0, confirmed=true) ok
Jun 17 23:56:43 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 37: start lvm_n01_vg0_start_0 on an-a04n01.alteeve.ca
(local)
Jun 17 23:56:44 an-a04n01 LVM(lvm_n01_vg0)[29173]: INFO: Activating
volume group an-a04n01_vg0
Jun 17 23:56:44 an-a04n01 LVM(lvm_n01_vg0)[29173]: INFO: Reading all
physical volumes. This may take a while... Found volume group
"an-a04n01_vg0" using metadata type lvm2
Jun 17 23:56:44 an-a04n01 LVM(lvm_n01_vg0)[29173]: INFO: 1 logical
volume(s) in volume group "an-a04n01_vg0" now active
Jun 17 23:56:44 an-a04n01 crmd[28360]: notice: process_lrm_event: LRM
operation lvm_n01_vg0_start_0 (call=82, rc=0, cib-update=59,
confirmed=true) ok
Jun 17 23:56:44 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 79: notify lvm_n01_vg0_post_notify_start_0 on
an-a04n01.alteeve.ca (local)
Jun 17 23:56:44 an-a04n01 crmd[28360]: notice: process_lrm_event: LRM
operation lvm_n01_vg0_notify_0 (call=85, rc=0, cib-update=0,
confirmed=true) ok
Jun 17 23:56:44 an-a04n01 crmd[28360]: notice: te_rsc_command:
Initiating action 38: monitor lvm_n01_vg0_monitor_30000 on
an-a04n01.alteeve.ca (local)
Jun 17 23:56:44 an-a04n01 crmd[28360]: notice: process_lrm_event: LRM
operation lvm_n01_vg0_monitor_30000 (call=88, rc=0, cib-update=60,
confirmed=false) ok
Jun 17 23:56:44 an-a04n01 crmd[28360]: notice: run_graph: Transition 4
(Complete=18, Pending=0, Fired=0, Skipped=0, Incomplete=0,
Source=/var/lib/pacemaker/pengine/pe-input-156.bz2): Complete
Jun 17 23:56:44 an-a04n01 crmd[28360]: notice: do_state_transition:
State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS
cause=C_FSA_INTERNAL origin=notify_crmd ]
====
Logs from the always-fenced 'an-a04n02', starting with
'/etc/init.d/pacemaker start':
====
Jun 17 23:55:32 an-a04n02 kernel: DLM (built Apr 11 2014 17:28:07) installed
Jun 17 23:55:33 an-a04n02 corosync[7176]: [MAIN ] Corosync Cluster
Engine ('1.4.1'): started and ready to provide service.
Jun 17 23:55:33 an-a04n02 corosync[7176]: [MAIN ] Corosync built-in
features: nss dbus rdma snmp
Jun 17 23:55:33 an-a04n02 corosync[7176]: [MAIN ] Successfully read
config from /etc/cluster/cluster.conf
Jun 17 23:55:33 an-a04n02 corosync[7176]: [MAIN ] Successfully parsed
cman config
Jun 17 23:55:33 an-a04n02 corosync[7176]: [TOTEM ] Initializing
transport (UDP/IP Multicast).
Jun 17 23:55:33 an-a04n02 corosync[7176]: [TOTEM ] Initializing
transmit/receive security: libtomcrypt SOBER128/SHA1HMAC (mode 0).
Jun 17 23:55:33 an-a04n02 corosync[7176]: [TOTEM ] The network
interface [10.20.40.2] is now up.
Jun 17 23:55:33 an-a04n02 corosync[7176]: [QUORUM] Using quorum
provider quorum_cman
Jun 17 23:55:33 an-a04n02 corosync[7176]: [SERV ] Service engine
loaded: corosync cluster quorum service v0.1
Jun 17 23:55:33 an-a04n02 corosync[7176]: [CMAN ] CMAN 3.0.12.1
(built Apr 3 2014 05:12:26) started
Jun 17 23:55:33 an-a04n02 corosync[7176]: [SERV ] Service engine
loaded: corosync CMAN membership service 2.90
Jun 17 23:55:33 an-a04n02 corosync[7176]: [SERV ] Service engine
loaded: openais checkpoint service B.01.01
Jun 17 23:55:33 an-a04n02 corosync[7176]: [SERV ] Service engine
loaded: corosync extended virtual synchrony service
Jun 17 23:55:33 an-a04n02 corosync[7176]: [SERV ] Service engine
loaded: corosync configuration service
Jun 17 23:55:33 an-a04n02 corosync[7176]: [SERV ] Service engine
loaded: corosync cluster closed process group service v1.01
Jun 17 23:55:33 an-a04n02 corosync[7176]: [SERV ] Service engine
loaded: corosync cluster config database access v1.01
Jun 17 23:55:33 an-a04n02 corosync[7176]: [SERV ] Service engine
loaded: corosync profile loading service
Jun 17 23:55:33 an-a04n02 corosync[7176]: [QUORUM] Using quorum
provider quorum_cman
Jun 17 23:55:33 an-a04n02 corosync[7176]: [SERV ] Service engine
loaded: corosync cluster quorum service v0.1
Jun 17 23:55:33 an-a04n02 corosync[7176]: [MAIN ] Compatibility mode
set to whitetank. Using V1 and V2 of the synchronization engine.
Jun 17 23:55:33 an-a04n02 corosync[7176]: [TOTEM ] A processor joined
or left the membership and a new membership was formed.
Jun 17 23:55:33 an-a04n02 corosync[7176]: [TOTEM ] A processor joined
or left the membership and a new membership was formed.
Jun 17 23:55:33 an-a04n02 corosync[7176]: [CMAN ] quorum regained,
resuming activity
Jun 17 23:55:33 an-a04n02 corosync[7176]: [QUORUM] This node is within
the primary component and will provide service.
Jun 17 23:55:33 an-a04n02 corosync[7176]: [QUORUM] Members[1]: 2
Jun 17 23:55:33 an-a04n02 corosync[7176]: [QUORUM] Members[1]: 2
Jun 17 23:55:33 an-a04n02 corosync[7176]: [QUORUM] Members[2]: 1 2
Jun 17 23:55:33 an-a04n02 corosync[7176]: [QUORUM] Members[2]: 1 2
Jun 17 23:55:33 an-a04n02 corosync[7176]: [CPG ] chosen downlist:
sender r(0) ip(10.20.40.1) ; members(old:1 left:0)
Jun 17 23:55:33 an-a04n02 corosync[7176]: [MAIN ] Completed service
synchronization, ready to provide service.
Jun 17 23:55:37 an-a04n02 fenced[7231]: fenced 3.0.12.1 started
Jun 17 23:55:37 an-a04n02 dlm_controld[7254]: dlm_controld 3.0.12.1 started
Jun 17 23:55:38 an-a04n02 gfs_controld[7306]: gfs_controld 3.0.12.1 started
Jun 17 23:55:39 an-a04n02 pacemaker: Attempting to start clvmd
Jun 17 23:55:40 an-a04n02 kernel: dlm: Using TCP for communications
Jun 17 23:55:40 an-a04n02 kernel: dlm: got connection from 1
Jun 17 23:55:41 an-a04n02 clvmd: Cluster LVM daemon started - connected
to CMAN
Jun 17 23:55:41 an-a04n02 pacemaker: Starting Pacemaker Cluster Manager
Jun 17 23:55:42 an-a04n02 pacemakerd[7437]: notice: crm_add_logfile:
Additional logging available in /var/log/cluster/corosync.log
Jun 17 23:55:42 an-a04n02 pacemakerd[7437]: notice: main: Starting
Pacemaker 1.1.10-14.el6_5.3 (Build: 368c726): generated-manpages
agent-manpages ascii-docs publican-docs ncurses libqb-logging libqb-ipc
nagios corosync-plugin cman
Jun 17 23:55:42 an-a04n02 lrmd[7445]: notice: crm_add_logfile:
Additional logging available in /var/log/cluster/corosync.log
Jun 17 23:55:42 an-a04n02 stonith-ng[7444]: notice: crm_add_logfile:
Additional logging available in /var/log/cluster/corosync.log
Jun 17 23:55:42 an-a04n02 cib[7443]: notice: crm_add_logfile:
Additional logging available in /var/log/cluster/corosync.log
Jun 17 23:55:42 an-a04n02 crmd[7448]: notice: crm_add_logfile:
Additional logging available in /var/log/cluster/corosync.log
Jun 17 23:55:42 an-a04n02 pengine[7447]: notice: crm_add_logfile:
Additional logging available in /var/log/cluster/corosync.log
Jun 17 23:55:42 an-a04n02 attrd[7446]: notice: crm_add_logfile:
Additional logging available in /var/log/cluster/corosync.log
Jun 17 23:55:42 an-a04n02 stonith-ng[7444]: notice:
crm_cluster_connect: Connecting to cluster infrastructure: cman
Jun 17 23:55:42 an-a04n02 crmd[7448]: notice: main: CRM Git Version:
368c726
Jun 17 23:55:42 an-a04n02 attrd[7446]: notice: crm_cluster_connect:
Connecting to cluster infrastructure: cman
Jun 17 23:55:42 an-a04n02 attrd[7446]: notice: main: Starting mainloop...
Jun 17 23:55:42 an-a04n02 cib[7443]: notice: crm_cluster_connect:
Connecting to cluster infrastructure: cman
Jun 17 23:55:43 an-a04n02 crmd[7448]: notice: crm_cluster_connect:
Connecting to cluster infrastructure: cman
Jun 17 23:55:43 an-a04n02 crmd[7448]: notice: cman_event_callback:
Membership 276: quorum acquired
Jun 17 23:55:43 an-a04n02 crmd[7448]: notice: crm_update_peer_state:
cman_event_callback: Node an-a04n01.alteeve.ca[1] - state is now member
(was (null))
Jun 17 23:55:43 an-a04n02 crmd[7448]: notice: crm_update_peer_state:
cman_event_callback: Node an-a04n02.alteeve.ca[2] - state is now member
(was (null))
Jun 17 23:55:43 an-a04n02 stonith-ng[7444]: notice: setup_cib:
Watching for stonith topology changes
Jun 17 23:55:43 an-a04n02 crmd[7448]: notice: do_started: The local
CRM is operational
Jun 17 23:55:43 an-a04n02 crmd[7448]: notice: do_state_transition:
State transition S_STARTING -> S_PENDING [ input=I_PENDING
cause=C_FSA_INTERNAL origin=do_started ]
Jun 17 23:55:43 an-a04n02 stonith-ng[7444]: notice: unpack_config: On
loss of CCM Quorum: Ignore
Jun 17 23:55:44 an-a04n02 stonith-ng[7444]: notice:
stonith_device_register: Added 'fence_n01_ipmi' to the device list (1
active devices)
Jun 17 23:55:45 an-a04n02 stonith-ng[7444]: notice:
stonith_device_register: Added 'fence_n02_ipmi' to the device list (2
active devices)
Jun 17 23:56:04 an-a04n02 crmd[7448]: warning: do_log: FSA: Input
I_DC_TIMEOUT from crm_timer_popped() received in state S_PENDING
Jun 17 23:56:04 an-a04n02 crmd[7448]: notice: do_state_transition:
State transition S_ELECTION -> S_PENDING [ input=I_PENDING
cause=C_FSA_INTERNAL origin=do_election_count_vote ]
Jun 17 23:56:04 an-a04n02 attrd[7446]: notice: attrd_local_callback:
Sending full refresh (origin=crmd)
Jun 17 23:56:04 an-a04n02 crmd[7448]: notice: do_state_transition:
State transition S_PENDING -> S_NOT_DC [ input=I_NOT_DC
cause=C_HA_MESSAGE origin=do_cl_join_finalize_respond ]
Jun 17 23:56:05 an-a04n02 LVM(lvm_n01_vg0)[7509]: WARNING: LVM Volume
an-a04n01_vg0 is not available (stopped)
Jun 17 23:56:05 an-a04n02 crmd[7448]: notice: process_lrm_event: LRM
operation lvm_n01_vg0_monitor_0 (call=20, rc=7, cib-update=11,
confirmed=true) not running
Jun 17 23:56:05 an-a04n02 crmd[7448]: notice: process_lrm_event: LRM
operation drbd_r0_monitor_0 (call=15, rc=7, cib-update=12,
confirmed=true) not running
Jun 17 23:56:05 an-a04n02 crmd[7448]: notice: process_lrm_event:
an-a04n02.alteeve.ca-drbd_r0_monitor_0:15 [ \n ]
Jun 17 23:56:05 an-a04n02 attrd[7446]: notice: attrd_trigger_update:
Sending flush op to all hosts for: probe_complete (true)
Jun 17 23:56:05 an-a04n02 attrd[7446]: notice: attrd_perform_update:
Sent update 5: probe_complete=true
Jun 17 23:56:06 an-a04n02 stonith-ng[7444]: notice:
stonith_device_register: Device 'fence_n02_ipmi' already existed in
device list (2 active devices)
Jun 17 23:56:06 an-a04n02 crmd[7448]: notice: process_lrm_event: LRM
operation fence_n02_ipmi_start_0 (call=25, rc=0, cib-update=13,
confirmed=true) ok
Jun 17 23:56:06 an-a04n02 crmd[7448]: notice: process_lrm_event: LRM
operation fence_n02_ipmi_monitor_60000 (call=30, rc=0, cib-update=14,
confirmed=false) ok
Jun 17 23:56:06 an-a04n02 kernel: block drbd0: Starting worker thread
(from cqueue [3220])
Jun 17 23:56:06 an-a04n02 kernel: block drbd0: disk( Diskless ->
Attaching )
Jun 17 23:56:06 an-a04n02 kernel: block drbd0: Found 3 transactions (3
active extents) in activity log.
Jun 17 23:56:06 an-a04n02 kernel: block drbd0: Method to ensure write
ordering: flush
Jun 17 23:56:06 an-a04n02 kernel: block drbd0: drbd_bm_resize called
with capacity == 909525832
Jun 17 23:56:06 an-a04n02 kernel: block drbd0: resync bitmap:
bits=113690729 words=1776418 pages=3470
Jun 17 23:56:06 an-a04n02 kernel: block drbd0: size = 434 GB (454762916 KB)
Jun 17 23:56:06 an-a04n02 kernel: block drbd0: bitmap READ of 3470 pages
took 8 jiffies
Jun 17 23:56:06 an-a04n02 kernel: block drbd0: recounting of set bits
took additional 17 jiffies
Jun 17 23:56:06 an-a04n02 kernel: block drbd0: 0 KB (0 bits) marked
out-of-sync by on disk bit-map.
Jun 17 23:56:06 an-a04n02 kernel: block drbd0: disk( Attaching ->
Consistent )
Jun 17 23:56:06 an-a04n02 kernel: block drbd0: attached to UUIDs
C71081B1CBAFC620:0000000000000000:F9F9DA52F6D93991:F9F8DA52F6D93991
Jun 17 23:56:06 an-a04n02 kernel: block drbd0: conn( StandAlone ->
Unconnected )
Jun 17 23:56:06 an-a04n02 kernel: block drbd0: Starting receiver thread
(from drbd0_worker [7613])
Jun 17 23:56:06 an-a04n02 kernel: block drbd0: receiver (re)started
Jun 17 23:56:06 an-a04n02 kernel: block drbd0: conn( Unconnected ->
WFConnection )
Jun 17 23:56:06 an-a04n02 attrd[7446]: notice: attrd_trigger_update:
Sending flush op to all hosts for: master-drbd_r0 (5)
Jun 17 23:56:06 an-a04n02 attrd[7446]: notice: attrd_perform_update:
Sent update 8: master-drbd_r0=5
Jun 17 23:56:06 an-a04n02 crmd[7448]: notice: process_lrm_event: LRM
operation drbd_r0_start_0 (call=27, rc=0, cib-update=15, confirmed=true) ok
Jun 17 23:56:06 an-a04n02 attrd[7446]: notice: attrd_perform_update:
Sent update 10: master-drbd_r0=5
Jun 17 23:56:06 an-a04n02 crmd[7448]: notice: process_lrm_event: LRM
operation drbd_r0_notify_0 (call=34, rc=0, cib-update=0, confirmed=true) ok
Jun 17 23:56:06 an-a04n02 crmd[7448]: notice: process_lrm_event: LRM
operation drbd_r0_notify_0 (call=37, rc=0, cib-update=0, confirmed=true) ok
Jun 17 23:56:06 an-a04n02 kernel: block drbd0: helper command:
/sbin/drbdadm fence-peer minor-0
Jun 17 23:56:07 an-a04n02 kernel: block drbd0: Handshake successful:
Agreed network protocol version 97
Jun 17 23:56:07 an-a04n02 stonith-ng[7444]: notice:
can_fence_host_with_device: fence_n02_ipmi can fence
an-a04n02.alteeve.ca: static-list
Jun 17 23:56:07 an-a04n02 stonith-ng[7444]: notice:
can_fence_host_with_device: fence_n01_ipmi can not fence
an-a04n02.alteeve.ca: static-list
Jun 17 23:56:07 an-a04n02 stonith_admin[7726]: notice: crm_log_args:
Invoked: stonith_admin --fence an-a04n01.alteeve.ca
Jun 17 23:56:07 an-a04n02 stonith-ng[7444]: notice: handle_request:
Client stonith_admin.7726.0f660392 wants to fence (off)
'an-a04n01.alteeve.ca' with device '(any)'
Jun 17 23:56:07 an-a04n02 stonith-ng[7444]: notice:
initiate_remote_stonith_op: Initiating remote operation off for
an-a04n01.alteeve.ca: fd2fafff-174a-4744-b83c-e762c88ed12b (0)
Jun 17 23:56:07 an-a04n02 stonith-ng[7444]: notice:
can_fence_host_with_device: fence_n02_ipmi can not fence
an-a04n01.alteeve.ca: static-list
Jun 17 23:56:07 an-a04n02 stonith-ng[7444]: notice:
can_fence_host_with_device: fence_n01_ipmi can fence
an-a04n01.alteeve.ca: static-list
Jun 17 23:56:07 an-a04n02 stonith-ng[7444]: notice:
can_fence_host_with_device: fence_n02_ipmi can not fence
an-a04n01.alteeve.ca: static-list
Jun 17 23:56:07 an-a04n02 stonith-ng[7444]: notice:
can_fence_host_with_device: fence_n01_ipmi can fence
an-a04n01.alteeve.ca: static-list
Jun 17 23:56:08 an-a04n02 ntpd[2540]: 0.0.0.0 c612 02 freq_set kernel
16.841 PPM
Jun 17 23:56:08 an-a04n02 ntpd[2540]: 0.0.0.0 c615 05 clock_sync
====
Cluestick beatins welcomed...
--
Digimer
Papers and Projects: https://alteeve.ca/w/
What if the cure for cancer is trapped in the mind of a person without
access to education?
More information about the Pacemaker
mailing list