[Pacemaker] Pacemaker Digest, Vol 53, Issue 40
emmanuel segura
emi2fast at gmail.com
Thu Apr 19 10:12:37 CEST 2012
vgchange -cy XenVG
your vg it's not clustered and check the clvmd process
ps -ef | grep clvmd
Il giorno 19 aprile 2012 09:45, Frank Meier <frank.meier at hr-group.de> ha
scritto:
> xencluster1:~ # vgs
> VG #PV #LV #SN Attr VSize VFree
> XenVG 8 32 0 wz--n- 938.16g 530.98g
>
>
> Mit freundlichen Grüßen
>
> Frank Meier
> UNIX-Basis
>
> Hamm Reno Group GmbH
> Industriegebiet West | D-66987 Thaleischweiler-Fröschen
> T.+49(0)6334 444-8322 | F.+49(0)6334 444-8190
> frank.meier at hr-group.de | www.reno.de
> ___________________________________________________________________
>
> Sitz: Am Tie 7 | D-49086 Osnabrück
> Handelsregister Osnabrück HRB 19587
> Geschäftsführer: Hans-Jürgen de Fries,
> Jens Gransee, Manfred Klumpp,
> Robert Reisch
>
> Am 19.04.2012 09:32, schrieb pacemaker-request at oss.clusterlabs.org:
> > Send Pacemaker mailing list submissions to
> > pacemaker at oss.clusterlabs.org
> >
> > To subscribe or unsubscribe via the World Wide Web, visit
> > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> > or, via email, send a message with subject or body 'help' to
> > pacemaker-request at oss.clusterlabs.org
> >
> > You can reach the person managing the list at
> > pacemaker-owner at oss.clusterlabs.org
> >
> > When replying, please edit your Subject line so it is more specific
> > than "Re: Contents of Pacemaker digest..."
> >
> >
> > Today's Topics:
> >
> > 1. Re: Pacemaker Digest, Vol 53, Issue 38 (emmanuel segura)
> >
> >
> > ----------------------------------------------------------------------
> >
> > Message: 1
> > Date: Thu, 19 Apr 2012 09:37:31 +0200
> > From: emmanuel segura <emi2fast at gmail.com>
> > To: The Pacemaker cluster resource manager
> > <pacemaker at oss.clusterlabs.org>
> > Subject: Re: [Pacemaker] Pacemaker Digest, Vol 53, Issue 38
> > Message-ID:
> > <CAE7pJ3CovBAsjf+vbbJGEKuqTTvFiWWpn0S=UvOTSoyQpeTa=
> w at mail.gmail.com>
> > Content-Type: text/plain; charset="iso-8859-1"
> >
> > give this command
> >
> > vgs
> >
> > Il giorno 19 aprile 2012 09:32, Frank Meier <frank.meier at hr-group.de> ha
> > scritto:
> >
> >> Hi,
> >>
> >> here my lvm.conf:
> >>
> >> devices {
> >> dir = "/dev"
> >> scan = [ "/dev" ]
> >> filter = [ "a|/dev/mapper|", "r/.*/" ]
> >> cache = "/etc/lvm/.cache"
> >> write_cache_state = 1
> >> types = [ "device-mapper", 16 ]
> >> sysfs_scan = 1
> >> md_component_detection = 1
> >> }
> >> log {
> >> verbose = 0
> >> syslog = 1
> >> overwrite = 0
> >> level = 0
> >> indent = 1
> >> command_names = 0
> >> prefix = " "
> >> }
> >> backup {
> >> backup = 1
> >> backup_dir = "/etc/lvm/backup"
> >> archive = 1
> >> archive_dir = "/etc/lvm/archive"
> >> retain_min = 10
> >> retain_days = 30
> >> }
> >> shell {
> >> history_size = 100
> >> }
> >> global {
> >> umask = 077
> >> test = 0
> >> activation = 1
> >> proc = "/proc"
> >> locking_type = 3
> >> locking_dir = "/var/lock/lvm"
> >> }
> >> activation {
> >> missing_stripe_filler = "error"
> >> mirror_region_size = 512
> >> reserved_stack = 256
> >> reserved_memory = 8192
> >> process_priority = -18
> >> }
> >> dmeventd {
> >> mirror_library = "libdevmapper-event-lvm2mirror.so.2.02"
> >> snapshot_library = "libdevmapper-event-lvm2snapshot.so.2.02"
> >> }
> >>
> >>
> >> Mit freundlichen Gr??en
> >>
> >> Frank Meier
> >> UNIX-Basis
> >>
> >> Hamm Reno Group GmbH
> >> Industriegebiet West | D-66987 Thaleischweiler-Fr?schen
> >> T.+49(0)6334 444-8322 | F.+49(0)6334 444-8190
> >> frank.meier at hr-group.de | www.reno.de
> >> ___________________________________________________________________
> >>
> >> Sitz: Am Tie 7 | D-49086 Osnabr?ck
> >> Handelsregister Osnabr?ck HRB 19587
> >> Gesch?ftsf?hrer: Hans-J?rgen de Fries,
> >> Jens Gransee, Manfred Klumpp,
> >> Robert Reisch
> >>
> >> Am 19.04.2012 09:13, schrieb pacemaker-request at oss.clusterlabs.org:
> >>> Send Pacemaker mailing list submissions to
> >>> pacemaker at oss.clusterlabs.org
> >>>
> >>> To subscribe or unsubscribe via the World Wide Web, visit
> >>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> >>> or, via email, send a message with subject or body 'help' to
> >>> pacemaker-request at oss.clusterlabs.org
> >>>
> >>> You can reach the person managing the list at
> >>> pacemaker-owner at oss.clusterlabs.org
> >>>
> >>> When replying, please edit your Subject line so it is more specific
> >>> than "Re: Contents of Pacemaker digest..."
> >>>
> >>>
> >>> Today's Topics:
> >>>
> >>> 1. Re: LVM restarts after SLES upgrade (emmanuel segura)
> >>>
> >>>
> >>> ----------------------------------------------------------------------
> >>>
> >>> Message: 1
> >>> Date: Thu, 19 Apr 2012 09:18:37 +0200
> >>> From: emmanuel segura <emi2fast at gmail.com>
> >>> To: The Pacemaker cluster resource manager
> >>> <pacemaker at oss.clusterlabs.org>
> >>> Subject: Re: [Pacemaker] LVM restarts after SLES upgrade
> >>> Message-ID:
> >>> <CAE7pJ3Ab1ZzSpou0=vvrO=
> >> mb3ur2+K1XhALLJEpj-sOOUEfwTA at mail.gmail.com>
> >>> Content-Type: text/plain; charset="iso-8859-1"
> >>>
> >>> maybe you have problem with clvm
> >>>
> >>> can you show me your vm configuration?
> >>>
> >>> Il giorno 19 aprile 2012 08:29, Frank Meier <frank.meier at hr-group.de>
> ha
> >>> scritto:
> >>>
> >>>> Hi,
> >>>>
> >>>> I've installed a 2-Node Xen-Cluster with SLES 11 SP1.
> >>>>
> >>>> After an upgrade to SLES11 SP2 the cluster won't work as the old one.
> >>>>
> >>>> After some time, all LVs reorganize and so the VM where restarted.
> >>>>
> >>>> Any clue?
> >>>>
> >>>> CIB:
> >>>>
> >>>> node xencluster1 \
> >>>> attributes standby="off"
> >>>> node xencluster2 \
> >>>> attributes standby="off"
> >>>> primitive STONITH--Node1 stonith:meatware \
> >>>> meta is-managed="true" target-role="Started" \
> >>>> operations $id="STONITH--Node1-operations" \
> >>>> op monitor interval="15" timeout="15" start-delay="15" \
> >>>> params hostlist="xencluster1"
> >>>> primitive STONITH-Node2 stonith:meatware \
> >>>> meta target-role="started" is-managed="true" \
> >>>> operations $id="STONITH-Node2-operations" \
> >>>> op monitor interval="15" timeout="15" start-delay="15" \
> >>>> params hostlist="xencluster2"
> >>>> primitive clvm ocf:lvm2:clvmd \
> >>>> operations $id="clvm-operations" \
> >>>> op monitor interval="10" timeout="20" \
> >>>> op start interval="0" timeout="90" \
> >>>> op stop interval="0" timeout="100" \
> >>>> params daemon_timeout="30"
> >>>> primitive clvm-xenvg ocf:heartbeat:LVM \
> >>>> operations $id="clvm-xenvg-operations" \
> >>>> op monitor interval="10" timeout="30" \
> >>>> params volgrpname="XenVG"
> >>>> primitive cmirror ocf:lvm2:cmirrord \
> >>>> operations $id="cmirror-operations" \
> >>>> op monitor interval="10" timeout="20" \
> >>>> op start interval="0" timeout="90" \
> >>>> op stop interval="0" timeout="100" \
> >>>> params daemon_timeout="30"
> >>>> primitive dlm ocf:pacemaker:controld \
> >>>> operations $id="dlm-operations" \
> >>>> op monitor interval="10" timeout="20" start-delay="0" \
> >>>> op start interval="0" timeout="90" \
> >>>> op stop interval="0" timeout="100"
> >>>> primitive fs-config-xen ocf:heartbeat:Filesystem \
> >>>> meta is-managed="true" target-role="Started" \
> >>>> operations $id="fs-config-xen-operations" \
> >>>> op monitor interval="20" timeout="40" \
> >>>> params device="/dev/mapper/XenVG-xenconfig"
> >> directory="/etc/xen/vm"
> >>>> fstype="ocfs2"
> >>>> primitive o2cb ocf:ocfs2:o2cb \
> >>>> operations $id="o2cb-operations" \
> >>>> op monitor interval="10" timeout="20" \
> >>>> op start interval="0" timeout="90" \
> >>>> op stop interval="0" timeout="100"
> >>>> primitive vm-cim ocf:heartbeat:Xen \
> >>>> meta target-role="Started" is-managed="true"
> >> allow-migrate="true" \
> >>>> operations $id="vm-cim-operations" \
> >>>> op monitor interval="10" timeout="30" \
> >>>> op start interval="0" timeout="60" \
> >>>> op stop interval="0" timeout="60" \
> >>>> op migrate_to interval="0" timeout="360" \
> >>>> params xmfile="/etc/xen/vm/cim"
> >>>> primitive vm-deprepo ocf:heartbeat:Xen \
> >>>> meta target-role="started" is-managed="true"
> >> allow-migrate="true" \
> >>>> operations $id="vm-deprepo-operations" \
> >>>> op monitor interval="10" timeout="30" \
> >>>> op start interval="0" timeout="60" \
> >>>> op stop interval="0" timeout="60" \
> >>>> op migrate_to interval="0" timeout="360" \
> >>>> params xmfile="/etc/xen/vm/debrepo"
> >>>>
> >>>>
> >>>> group clvm-glue dlm clvm o2cb cmirror \
> >>>> meta target-role="Started"
> >>>> group xen-vg-fs clvm-xenvg fs-config-xen
> >>>> clone c-clvm-glue clvm-glue \
> >>>> meta interleave="true" ordered="true" target-role="Started"
> >>>> clone c-xen-vg-fs xen-vg-fs \
> >>>> meta interleave="true" ordered="true" is-managed="true"
> >>>> target-role="Started"
> >>>> location STONITHnode1Constraint STONITH--Node1 -inf: xencluster1
> >>>> colocation colo-clvmglue-xenvgsfs inf: c-xen-vg-fs c-clvm-glue
> >>>> order o-cim inf: c-xen-vg-fs vm-cim
> >>>> order o-clvmglue-xenvgfs inf: c-clvm-glue c-xen-vg-fs
> >>>> order o-deprepo inf: c-xen-vg-fs vm-deprepo
> >>>> order o-ilmt inf: c-xen-vg-fs vm-ilmt
> >>>> order o-intranet inf: c-xen-vg-fs vm-intranet
> >>>> order o-nagois inf: c-xen-vg-fs vm-nagios
> >>>> order o-oneiroi inf: c-xen-vg-fs vm-oneiroi
> >>>> order o-otbridge inf: c-xen-vg-fs vm-otbridge
> >>>> order o-soa inf: c-xen-vg-fs vm-soa
> >>>> order o-topdesk-entw inf: c-xen-vg-fs vm-topdesk-entw
> >>>> order o-traumschiff inf: c-xen-vg-fs vm-traumschiff
> >>>> order o-virenscanner inf: c-xen-vg-fs vm-virenscanner
> >>>> property $id="cib-bootstrap-options" \
> >>>> dc-version="1.1.6-b988976485d15cb702c9307df55512d323831a5e" \
> >>>> no-quorum-policy="ignore" \
> >>>> default-resource-stickiness="100000" \
> >>>> last-lrm-refresh="1334518148" \
> >>>> cluster-infrastructure="openais" \
> >>>> expected-quorum-votes="2" \
> >>>> default-action-timeout="120s" \
> >>>> maintenance-mode="true"
> >>>> op_defaults $id="op_defaults-options" \
> >>>> record-pending="false"
> >>>>
> >>>> Node 1Logs:
> >>>>
> >>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: Sending message to all
> cluster
> >>>> nodes
> >>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: process_work_item: local
> >>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: process_local_command:
> unknown
> >>>> (0x2d) msg=0x7f5c40010e80, msglen =32, client=0x7f5c400578d0
> >>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: Syncing device names
> >>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: Reply from node 4914420a: 0
> >> bytes
> >>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: Got 1 replies, expecting: 2
> >>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: LVM thread waiting for work
> >>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: 1226064394 got message from
> >>>> nodeid 1226064394 for 0. len 32
> >>>> Apr 15 22:01:40 xencluster1 lrmd: [8093]: WARN: clvm-xenvg:0:monitor
> >>>> process (PID 31785) timed out (try 1). Killing with signal SIGTERM
> >> (15).
> >>>> Apr 15 22:01:40 xencluster1 lrmd: [8093]: WARN: operation monitor[92]
> on
> >>>> clvm-xenvg:0 for client 8096: pid 31785 timed out
> >>>> Apr 15 22:01:40 xencluster1 crmd: [8096]: ERROR: process_lrm_event:
> LRM
> >>>> operation clvm-xenvg:0_monitor_10000 (92) Timed Out (timeout=30000ms)
> >>>> Apr 15 22:01:42 xencluster1 corosync[8071]: [TOTEM ] Retransmit
> List:
> >>>> 21ac
> >>>> Apr 15 22:01:42 xencluster1 corosync[8071]: [TOTEM ] Retransmit
> List:
> >>>> 21ab
> >>>> Apr 15 22:01:42 xencluster1 corosync[8071]: [TOTEM ] Retransmit
> List:
> >>>> 21ac
> >>>> Apr 15 22:01:42 xencluster1 corosync[8071]: [TOTEM ] Retransmit
> List:
> >>>> 21ab
> >>>>
> >>>>
> >>>> pr 15 22:01:50 xencluster1 clvmd[8763]: Send local reply
> >>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Read on local socket 14, len
> =
> >> 32
> >>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: check_all_clvmds_running
> >>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: down_callback. node
> 1226064394,
> >>>> state = 3
> >>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: down_callback. node
> 1309950474,
> >>>> state = 3
> >>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Got pre command condition...
> >>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Writing status 0 down pipe 16
> >>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Waiting to do post command -
> >>>> state = 0
> >>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: read on PIPE 15: 4 bytes:
> >> status:
> >>>> 0
> >>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: background routine status was
> >>>> 0, sock_client=0x7f5c40057c10
> >>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: distribute command: XID =
> 15473
> >>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: num_nodes = 2
> >>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: add_to_lvmqueue:
> >>>> cmd=0x7f5c4000dbd0. client=0x7f5c40057c10, msg=0x7f5c40057880, len=32,
> >>>> csid=(nil), xid=15473
> >>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Sending message to all
> cluster
> >>>> nodes
> >>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: process_work_item: local
> >>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: process_local_command:
> unknown
> >>>> (0x2d) msg=0x7f5c40010e80, msglen =32, client=0x7f5c40057c10
> >>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Syncing device names
> >>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Reply from node 4914420a: 0
> >> bytes
> >>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Got 1 replies, expecting: 2
> >>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: LVM thread waiting for work
> >>>> Apr 15 22:01:50 xencluster1 corosync[8071]: [TOTEM ] Retransmit
> List:
> >>>> 21ab
> >>>> Apr 15 22:01:50 xencluster1 corosync[8071]: [TOTEM ] Retransmit
> List:
> >>>> 21ac
> >>>> Apr 15 22:01:50 xencluster1 corosync[8071]: [TOTEM ] Retransmit
> List:
> >>>> 21ab
> >>>> ...
> >>>> Apr 15 22:02:19 xencluster1 corosync[8071]: [TOTEM ] Retransmit
> List:
> >>>> 21ac
> >>>> Apr 15 22:02:20 xencluster1 corosync[8071]: [TOTEM ] Retransmit
> List:
> >>>> 21ab
> >>>> Apr 15 22:02:20 xencluster1 lrmd: [8093]: WARN: clvm-xenvg:0:monitor
> >>>> process (PID 771) timed out (try 1). Killing with signal SIGTERM
> (15).
> >>>> Apr 15 22:02:20 xencluster1 lrmd: [8093]: WARN: operation monitor[92]
> on
> >>>> clvm-xenvg:0 for client 8096: pid 771 timed out
> >>>> Apr 15 22:02:20 xencluster1 crmd: [8096]: ERROR: process_lrm_event:
> LRM
> >>>> operation clvm-xenvg:0_monitor_10000 (92) Timed Out (timeout=30000ms)
> >>>> Apr 15 22:02:20 xencluster1 corosync[8071]: [TOTEM ] Retransmit
> List:
> >>>> 21ac
> >>>> Apr 15 22:02:20 xencluster1 corosync[8071]: [TOTEM ] Retransmit
> List:
> >>>> 21ab
> >>>> ...
> >>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: Got new connection on fd 17
> >>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: Read on local socket 17, len
> =
> >> 28
> >>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: creating pipe, [18, 19]
> >>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: Creating pre&post thread
> >>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: Created pre&post thread,
> state
> >> = 0
> >>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: in sub thread: client =
> >>>> 0x7f5c40058080
> >>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: Sub thread ready for work.
> >>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: doing PRE command LOCK_VG
> >>>> 'V_XenVG' at 1 (client=0x7f5c40058080)
> >>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: lock_resource 'V_XenVG',
> >>>> flags=0, mode=3
> >>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: lock_resource returning 0,
> >>>> lock_id=3a0001
> >>>>
> >>>>
> >>>>
> >>>> Apr 15 22:02:48 xencluster1 corosync[8071]: [TOTEM ] Retransmit
> List:
> >>>> 21ab
> >>>> Apr 15 22:02:48 xencluster1 cib: [8092]: info: cib_stats: Processed 4
> >>>> operations (2500.00us average, 0% utilization) in the last 10min
> >>>> Apr 15 22:02:48 xencluster1 corosync[8071]: [TOTEM ] Retransmit
> List:
> >>>> 21ac
> >>>>
> >>>>
> >>>>
> >>>> Apr 15 22:03:00 xencluster1 corosync[8071]: [TOTEM ] Retransmit
> List:
> >>>> 21ac
> >>>> Apr 15 22:03:00 xencluster1 lrmd: [8093]: WARN: clvm-xenvg:0:monitor
> >>>> process (PID 2256) timed out (try 1). Killing with signal SIGTERM
> (15).
> >>>> Apr 15 22:03:00 xencluster1 lrmd: [8093]: WARN: operation monitor[92]
> on
> >>>> clvm-xenvg:0 for client 8096: pid 2256 timed out
> >>>> Apr 15 22:03:00 xencluster1 crmd: [8096]: ERROR: process_lrm_event:
> LRM
> >>>> operation clvm-xenvg:0_monitor_10000 (92) Timed Out (timeout=30000ms)
> >>>> Apr 15 22:03:00 xencluster1 corosync[8071]: [TOTEM ] Retransmit
> List:
> >>>> 21ab
> >>>>
> >>>>
> >>>> Node2 Log
> >>>>
> >>>> Apr 15 22:00:51 xencluster2 clvmd[12666]: Joined child thread
> >>>> Apr 15 22:00:51 xencluster2 clvmd[12666]: ret == 0, errno = 0.
> removing
> >>>> client
> >>>> Apr 15 22:00:51 xencluster2 clvmd[12666]: add_to_lvmqueue:
> >>>> cmd=0x7fa3d0028710. client=0x7fa3d0028760, msg=(nil), len=0,
> csid=(nil),
> >>>> xid=14778
> >>>> Apr 15 22:00:51 xencluster2 clvmd[12666]: process_work_item: free fd
> -1
> >>>> Apr 15 22:00:51 xencluster2 clvmd[12666]: LVM thread waiting for work
> >>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: crm_timer_popped:
> >>>> PEngine Recheck Timer (I_PE_CALC) just popped (900000ms)
> >>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_state_transition:
> >>>> State transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC
> >>>> cause=C_TIMER_POPPED origin=crm_timer_popped ]
> >>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_state_transition:
> >>>> Progressed to state S_POLICY_ENGINE after C_TIMER_POPPED
> >>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_state_transition:
> All
> >>>> 2 cluster nodes are eligible to run resources.
> >>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_pe_invoke: Query
> 983:
> >>>> Requesting the current CIB: S_POLICY_ENGINE
> >>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_pe_invoke_callback:
> >>>> Invoking the PE: query=983, ref=pe_calc-dc-1334520052-604, seq=2212,
> >>>> quorate=1
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_config: On
> >>>> loss of CCM Quorum: Ignore
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
> >>>> Operation clvm:1_last_failure_0 found resource clvm:1 active on
> >> xencluster2
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
> >>>> Operation STONITH-Node2_last_failure_0 found resource STONITH-Node2
> >>>> active on xencluster2
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
> >>>> Operation STONITH--Node1_last_failure_0 found resource STONITH--Node1
> >>>> active on xencluster2
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
> >>>> Operation cmirror:1_last_failure_0 found resource cmirror:1 active on
> >>>> xencluster2
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
> >>>> Operation dlm:1_last_failure_0 found resource dlm:1 active on
> >> xencluster2
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
> >>>> Operation o2cb:1_last_failure_0 found resource o2cb:1 active on
> >> xencluster2
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
> >>>> Operation dlm:0_last_failure_0 found resource dlm:0 active on
> >> xencluster1
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
> >>>> Operation clvm:0_last_failure_0 found resource clvm:0 active on
> >> xencluster1
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
> >>>> Operation o2cb:0_last_failure_0 found resource o2cb:0 active on
> >> xencluster1
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
> >>>> Operation clvm-xenvg:0_last_failure_0 found resource clvm-xenvg:0
> active
> >>>> on xencluster1
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: WARN: unpack_rsc_op:
> >>>> Processing failed op vm-deprepo_last_failure_0 on xencluster1: unknown
> >>>> error (1)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> STONITH-Node2 (Started xencluster2)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> STONITH--Node1 (Started xencluster2)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> dlm:0 (Started xencluster1)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> clvm:0 (Started xencluster1)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> o2cb:0 (Started xencluster1)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> cmirror:0 (Started xencluster1)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> dlm:1 (Started xencluster2)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> clvm:1 (Started xencluster2)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> o2cb:1 (Started xencluster2)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> cmirror:1 (Started xencluster2)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> clvm-xenvg:0 (Started xencluster1)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> fs-config-xen:0 (Started xencluster1)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> clvm-xenvg:1 (Started xencluster2)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> fs-config-xen:1 (Started xencluster2)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> vm-ilmt (Started xencluster1)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> vm-topdesk-entw (Started xencluster1)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> vm-otbridge (Started xencluster2)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> vm-virenscanner (Started xencluster1)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> vm-deprepo (Started xencluster1)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> vm-traumschiff (Started xencluster2)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> vm-cim (Started xencluster2)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> vm-oneiroi (Started xencluster1)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> vm-intranet (Started xencluster1)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> vm-soa (Started xencluster1)
> >>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice:
> process_pe_message:
> >>>> Transition 80: PEngine Input stored in:
> >> /var/lib/pengine/pe-input-271.bz2
> >>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_state_transition:
> >>>> State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [
> >>>> input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
> >>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: unpack_graph: Unpacked
> >>>> transition 80: 0 actions in 0 synapses
> >>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_te_invoke:
> Processing
> >>>> graph 80 (ref=pe_calc-dc-1334520052-604) derived from
> >>>> /var/lib/pengine/pe-input-271.bz2
> >>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: run_graph:
> >>>> ====================================================
> >>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: notice: run_graph:
> Transition
> >>>> 80 (Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0,
> >>>> Source=/var/lib/pengine/pe-input-271.bz2): Complete
> >>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: te_graph_trigger:
> >>>> Transition 80 is now complete
> >>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: notify_crmd:
> Transition
> >>>> 80 status: done - <null>
> >>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_state_transition:
> >>>> State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS
> >>>> cause=C_FSA_INTERNAL origin=notify_crmd ]
> >>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_state_transition:
> >>>> Starting PEngine Recheck Timer
> >>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: 1309950474 got message from
> >>>> nodeid 1226064394 for 0. len 32
> >>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: add_to_lvmqueue:
> >>>> cmd=0x7fa3d0028780. client=0x6934a0, msg=0x7fa3d910063c, len=32,
> >>>> csid=0x7fffb4e5d944, xid=0
> >>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: process_work_item: remote
> >>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: process_remote_command
> unknown
> >>>> (0x2d) for clientid 0x5000000 XID 15337 on node 4914420a
> >>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: Syncing device names
> >>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: LVM thread waiting for work
> >>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: 1309950474 got message from
> >>>> nodeid 1309950474 for 1226064394. len 18
> >>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: 1309950474 got message from
> >>>> nodeid 1226064394 for 0. len 32
> >>>>
> >>>>
> >>>>
> >>>>
> >>>> Apr 15 22:01:02 xencluster2 clvmd[12666]: Joined child thread
> >>>> Apr 15 22:01:02 xencluster2 clvmd[12666]: ret == 0, errno = 0.
> removing
> >>>> client
> >>>> Apr 15 22:01:02 xencluster2 clvmd[12666]: add_to_lvmqueue:
> >>>> cmd=0x7fa3d004d7a0. client=0x7fa3d004d7f0, msg=(nil), len=0,
> csid=(nil),
> >>>> xid=14848
> >>>> Apr 15 22:01:02 xencluster2 clvmd[12666]: process_work_item: free fd
> -1
> >>>> Apr 15 22:01:02 xencluster2 clvmd[12666]: LVM thread waiting for work
> >>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info: stonith_command:
> >>>> Processed st_execute from lrmd: rc=-1
> >>>> Apr 15 22:01:06 xencluster2 stonith: parse config info
> info=xencluster1
> >>>> Apr 15 22:01:06 xencluster2 stonith: meatware device OK.
> >>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info: log_operation:
> >>>> STONITH--Node1: Performing: stonith -t meatware -S
> >>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info: log_operation:
> >>>> STONITH--Node1: success: 0
> >>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info:
> >>>> stonith_device_execute: Nothing to do for STONITH--Node1
> >>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info: stonith_command:
> >>>> Processed st_execute from lrmd: rc=-1
> >>>> Apr 15 22:01:06 xencluster2 stonith: parse config info
> info=xencluster2
> >>>> Apr 15 22:01:06 xencluster2 stonith: meatware device OK.
> >>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info: log_operation:
> >>>> STONITH-Node2: Performing: stonith -t meatware -S
> >>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info: log_operation:
> >>>> STONITH-Node2: success: 0
> >>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info:
> >>>> stonith_device_execute: Nothing to do for STONITH-Node2
> >>>> Apr 15 22:01:10 xencluster2 clvmd[12666]: 1309950474 got message from
> >>>> nodeid 1226064394 for 0. len 32
> >>>> Apr 15 22:01:10 xencluster2 clvmd[12666]: add_to_lvmqueue:
> >>>> cmd=0x7fa3d004d810. client=0x6934a0, msg=0x7fa3d9100cfc, len=32,
> >>>> csid=0x7fffb4e5d944, xid=0
> >>>> Apr 15 22:01:10 xencluster2 clvmd[12666]: process_work_item: remote
> >>>> Apr 15 22:01:10 xencluster2 clvmd[12666]: process_remote_command
> unknown
> >>>> (0x2d) for clientid 0x5000000 XID 15407 on node 4914420a
> >>>> Apr 15 22:01:10 xencluster2 clvmd[12666]: Syncing device names
> >>>> Apr 15 22:01:10 xencluster2 clvmd[12666]: LVM thread waiting for work
> >>>> Apr 15 22:01:10 xencluster2 clvmd[12666]: 1309950474 got message from
> >>>> nodeid 1309950474 for 1226064394. len 18
> >>>> Apr 15 22:01:10 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196
> >>>> Apr 15 22:01:10 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196
> >>>> Apr 15 22:01:10 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196
> >>>> ....
> >>>> Apr 15 22:01:11 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196
> >>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: Got new connection on fd 5
> >>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: Read on local socket 5, len
> =
> >> 28
> >>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: creating pipe, [12, 13]
> >>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: Creating pre&post thread
> >>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: Created pre&post thread,
> >> state =
> >>>> 0
> >>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: in sub thread: client =
> >>>> 0x7fa3d004d810
> >>>>
> >>>>
> >>>>
> >>>>
> >>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: Syncing device names
> >>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: Reply from node 4e14420a: 0
> >> bytes
> >>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: Got 1 replies, expecting: 2
> >>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: LVM thread waiting for work
> >>>> Apr 15 22:01:12 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196
> >>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: 1309950474 got message from
> >>>> nodeid 1309950474 for 0. len 32
> >>>> Apr 15 22:01:12 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196
> >>>> Apr 15 22:01:12 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196
> >>>> ....
> >>>> Apr 15 22:01:21 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197
> >>>> Apr 15 22:01:21 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197
> >>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info: stonith_command:
> >>>> Processed st_execute from lrmd: rc=-1
> >>>> Apr 15 22:01:21 xencluster2 stonith: parse config info
> info=xencluster1
> >>>> Apr 15 22:01:21 xencluster2 stonith: meatware device OK.
> >>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info: log_operation:
> >>>> STONITH--Node1: Performing: stonith -t meatware -S
> >>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info: log_operation:
> >>>> STONITH--Node1: success: 0
> >>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info:
> >>>> stonith_device_execute: Nothing to do for STONITH--Node1
> >>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info: stonith_command:
> >>>> Processed st_execute from lrmd: rc=-1
> >>>> Apr 15 22:01:21 xencluster2 stonith: parse config info
> info=xencluster2
> >>>> Apr 15 22:01:21 xencluster2 stonith: meatware device OK.
> >>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info: log_operation:
> >>>> STONITH-Node2: Performing: stonith -t meatware -S
> >>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info: log_operation:
> >>>> STONITH-Node2: success: 0
> >>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info:
> >>>> stonith_device_execute: Nothing to do for STONITH-Node2
> >>>> Apr 15 22:01:21 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197
> >>>> Apr 15 22:01:22 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197
> >>>> Apr 15 22:01:22 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197
> >>>> ....
> >>>> Apr 15 22:01:36 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197
> >>>> Apr 15 22:01:36 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197
> >>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info: stonith_command:
> >>>> Processed st_execute from lrmd: rc=-1
> >>>> Apr 15 22:01:36 xencluster2 stonith: parse config info
> info=xencluster1
> >>>> Apr 15 22:01:36 xencluster2 stonith: meatware device OK.
> >>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info: log_operation:
> >>>> STONITH--Node1: Performing: stonith -t meatware -S
> >>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info: log_operation:
> >>>> STONITH--Node1: success: 0
> >>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info:
> >>>> stonith_device_execute: Nothing to do for STONITH--Node1
> >>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info: stonith_command:
> >>>> Processed st_execute from lrmd: rc=-1
> >>>> Apr 15 22:01:36 xencluster2 stonith: parse config info
> info=xencluster2
> >>>> Apr 15 22:01:36 xencluster2 stonith: meatware device OK.
> >>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info: log_operation:
> >>>> STONITH-Node2: Performing: stonith -t meatware -S
> >>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info: log_operation:
> >>>> STONITH-Node2: success: 0
> >>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info:
> >>>> stonith_device_execute: Nothing to do for STONITH-Node2
> >>>> Apr 15 22:01:36 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197
> >>>> Apr 15 22:01:37 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197
> >>>> ....
> >>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197
> >>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197
> >>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: process_graph_event:
> >>>> Detected action clvm-xenvg:0_monitor_10000 from a different
> transition:
> >>>> 62 vs. 80
> >>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info:
> abort_transition_graph:
> >>>> process_graph_event:476 - Triggered transition abort (complete=1,
> >>>> tag=lrm_rsc_op, id=clvm-xenvg:0_monitor_10000,
> >>>> magic=2:-2;61:62:0:72497771-fdb2-4ca7-a723-2cc360861bb0,
> cib=0.5816.4) :
> >>>> Old event
> >>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: WARN: update_failcount:
> >>>> Updating failcount for clvm-xenvg:0 on xencluster1 after failed
> monitor:
> >>>> rc=-2 (update=value++, time=1334520100)
> >>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: do_state_transition:
> >>>> State transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC
> >>>> cause=C_FSA_INTERNAL origin=abort_transition_graph ]
> >>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: do_state_transition:
> All
> >>>> 2 cluster nodes are eligible to run resources.
> >>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: do_pe_invoke: Query
> 984:
> >>>> Requesting the current CIB: S_POLICY_ENGINE
> >>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197
> >>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197
> >>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197
> >>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197
> >>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197
> >>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197
> >>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197 219a 219b 219c 219d
> >>>> ...
> >>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197 219a 219b 219c 219d
> >>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: do_pe_invoke_callback:
> >>>> Invoking the PE: query=984, ref=pe_calc-dc-1334520100-605, seq=2212,
> >>>> quorate=1
> >>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197 219a 219b 219c 219d
> >>>> ...
> >>>>
> >>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197 219a 219b 219c 219d
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_config: On
> >>>> loss of CCM Quorum: Ignore
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
> >>>> Operation clvm:1_last_failure_0 found resource clvm:1 active on
> >> xencluster2
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
> >>>> Operation STONITH-Node2_last_failure_0 found resource STONITH-Node2
> >>>> active on xencluster2
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
> >>>> Operation STONITH--Node1_last_failure_0 found resource STONITH--Node1
> >>>> active on xencluster2
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
> >>>> Operation cmirror:1_last_failure_0 found resource cmirror:1 active on
> >>>> xencluster2
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
> >>>> Operation dlm:1_last_failure_0 found resource dlm:1 active on
> >> xencluster2
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
> >>>> Operation o2cb:1_last_failure_0 found resource o2cb:1 active on
> >> xencluster2
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
> >>>> Operation cmirror:0_last_failure_0 found resource cmirror:0 active on
> >>>> xencluster1
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
> >>>> Operation dlm:0_last_failure_0 found resource dlm:0 active on
> >> xencluster1
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
> >>>> Operation clvm:0_last_failure_0 found resource clvm:0 active on
> >> xencluster1
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
> >>>> Operation o2cb:0_last_failure_0 found resource o2cb:0 active on
> >> xencluster1
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
> >>>> Operation clvm-xenvg:0_last_failure_0 found resource clvm-xenvg:0
> active
> >>>> on xencluster1
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: WARN: unpack_rsc_op:
> >>>> Processing failed op clvm-xenvg:0_monitor_10000 on xencluster1:
> unknown
> >>>> exec error (-2)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: WARN: unpack_rsc_op:
> >>>> Processing failed op vm-deprepo_last_failure_0 on xencluster1: unknown
> >>>> error (1)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: RecurringOp:
> Start
> >>>> recurring monitor (10s) for clvm-xenvg:0 on xencluster1
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> STONITH-Node2 (Started xencluster2)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> STONITH--Node1 (Started xencluster2)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> dlm:0 (Started xencluster1)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> clvm:0 (Started xencluster1)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> o2cb:0 (Started xencluster1)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> cmirror:0 (Started xencluster1)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> dlm:1 (Started xencluster2)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> clvm:1 (Started xencluster2)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> o2cb:1 (Started xencluster2)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> cmirror:1 (Started xencluster2)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions:
> Recover
> >>>> clvm-xenvg:0 (Started xencluster1)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions:
> Restart
> >>>> fs-config-xen:0 (Started xencluster1)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> clvm-xenvg:1 (Started xencluster2)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> fs-config-xen:1 (Started xencluster2)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions:
> Restart
> >>>> vm-ilmt (Started xencluster1)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions:
> Restart
> >>>> vm-topdesk-entw (Started xencluster1)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> vm-otbridge (Started xencluster2)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions:
> Restart
> >>>> vm-virenscanner (Started xencluster1)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions:
> Restart
> >>>> vm-deprepo (Started xencluster1)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> vm-traumschiff (Started xencluster2)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
> >>>> vm-cim (Started xencluster2)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions:
> Restart
> >>>> vm-oneiroi (Started xencluster1)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions:
> Restart
> >>>> vm-intranet (Started xencluster1)
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions:
> Restart
> >>>> vm-soa (Started xencluster1)
> >>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: do_state_transition:
> >>>> State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [
> >>>> input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
> >>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: unpack_graph: Unpacked
> >>>> transition 81: 39 actions in 39 synapses
> >>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: do_te_invoke:
> Processing
> >>>> graph 81 (ref=pe_calc-dc-1334520100-605) derived from
> >>>> /var/lib/pengine/pe-input-272.bz2
> >>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: te_rsc_command:
> >>>> Initiating action 82: stop vm-ilmt_stop_0 on xencluster1
> >>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: te_rsc_command:
> >>>> Initiating action 86: stop vm-topdesk-entw_stop_0 on xencluster1
> >>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: te_rsc_command:
> >>>> Initiating action 90: stop vm-virenscanner_stop_0 on xencluster1
> >>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: te_rsc_command:
> >>>> Initiating action 92: stop vm-deprepo_stop_0 on xencluster1
> >>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: te_rsc_command:
> >>>> Initiating action 98: stop vm-oneiroi_stop_0 on xencluster1
> >>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: te_rsc_command:
> >>>> Initiating action 100: stop vm-intranet_stop_0 on xencluster1
> >>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: te_rsc_command:
> >>>> Initiating action 102: stop vm-soa_stop_0 on xencluster1
> >>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice:
> process_pe_message:
> >>>> Transition 81: PEngine Input stored in:
> >> /var/lib/pengine/pe-input-272.bz2
> >>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197 219a 219b 219c 219d
> >>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197 219a 219b 219c 219d
> >>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197 219a 219b 219c 219d
> >>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197 219a 219b 219c 219d
> >>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197 219a 219b 219c 219d
> >>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197 219a 219b 219c 219d 219e 219f 21a0 21a1 21a2 21a3 21a4 21a5
> >>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197 219a 219b 219c 219d 219e 219f 21a0 21a1 21a2 21a3 21a4 21a5
> >>>> ...
> >>>> Apr 15 22:01:41 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197 219a 219b 219c 219d 219e 219f 21a0 21a1 21a2 21a3 21a4 21a5
> >>>> Apr 15 22:01:42 xencluster2 lrmd: [7675]: WARN: clvm-xenvg:1:monitor
> >>>> process (PID 29785) timed out (try 1). Killing with signal SIGTERM
> >> (15).
> >>>> Apr 15 22:01:42 xencluster2 lrmd: [7675]: WARN: operation monitor[125]
> >>>> on clvm-xenvg:1 for client 7678: pid 29785 timed out
> >>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: ERROR: process_lrm_event:
> LRM
> >>>> operation clvm-xenvg:1_monitor_10000 (125) Timed Out (timeout=30000ms)
> >>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: info: process_graph_event:
> >>>> Detected action clvm-xenvg:1_monitor_10000 from a different
> transition:
> >>>> 68 vs. 81
> >>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: info:
> abort_transition_graph:
> >>>> process_graph_event:476 - Triggered transition abort (complete=0,
> >>>> tag=lrm_rsc_op, id=clvm-xenvg:1_monitor_10000,
> >>>> magic=2:-2;61:68:0:72497771-fdb2-4ca7-a723-2cc360861bb0,
> cib=0.5816.5) :
> >>>> Old event
> >>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: info: update_abort_priority:
> >>>> Abort priority upgraded from 0 to 1000000
> >>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: info: update_abort_priority:
> >>>> Abort action done superceeded by restart
> >>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: WARN: update_failcount:
> >>>> Updating failcount for clvm-xenvg:1 on xencluster2 after failed
> monitor:
> >>>> rc=-2 (update=value++, time=1334520102)
> >>>> Apr 15 22:01:42 xencluster2 attrd: [7676]: notice:
> attrd_trigger_update:
> >>>> Sending flush op to all hosts for: fail-count-clvm-xenvg:1 (1)
> >>>> Apr 15 22:01:42 xencluster2 attrd: [7676]: notice:
> attrd_perform_update:
> >>>> Sent update 448: fail-count-clvm-xenvg:1=1
> >>>> Apr 15 22:01:42 xencluster2 attrd: [7676]: notice:
> attrd_trigger_update:
> >>>> Sending flush op to all hosts for: last-failure-clvm-xenvg:1
> >> (1334520102)
> >>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: info:
> abort_transition_graph:
> >>>> te_update_diff:176 - Triggered transition abort (complete=0,
> tag=nvpair,
> >>>> id=status-xencluster2-fail-count-clvm-xenvg.1,
> >>>> name=fail-count-clvm-xenvg:1, value=1, magic=NA, cib=0.5816.6) :
> >>>> Transient attribute: update
> >>>> Apr 15 22:01:42 xencluster2 attrd: [7676]: notice:
> attrd_perform_update:
> >>>> Sent update 451: last-failure-clvm-xenvg:1=1334520102
> >>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: info:
> abort_transition_graph:
> >>>> te_update_diff:176 - Triggered transition abort (complete=0,
> tag=nvpair,
> >>>> id=status-xencluster2-last-failure-clvm-xenvg.1,
> >>>> name=last-failure-clvm-xenvg:1, value=1334520102, magic=NA,
> >>>> cib=0.5816.7) : Transient attribute: update
> >>>> Apr 15 22:01:42 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197 219a 219b 219c 219d 219e 219f 21a0 21a1 21a2 21a3 21a4 21a5
> >>>> Apr 15 22:01:42 xencluster2 corosync[7666]: [TOTEM ] Retransmit
> List:
> >>>> 2196 2197 219a 219b 219c 219d 219e 219f 21a0 21a1 21a2 21a3 21a4 21a5
> >>>>
> >>>> --
> >>>> Mit freundlichen Gr??en
> >>>>
> >>>> Frank Meier
> >>>> UNIX-Basis
> >>>>
> >>>> Hamm Reno Group GmbH
> >>>> Industriegebiet West | D-66987 Thaleischweiler-Fr?schen
> >>>> T.+49(0)6334 444-8322 | F.+49(0)6334 444-8190
> >>>> frank.meier at hr-group.de | www.reno.de
> >>>> ___________________________________________________________________
> >>>>
> >>>> Sitz: Am Tie 7 | D-49086 Osnabr?ck
> >>>> Handelsregister Osnabr?ck HRB 19587
> >>>> Gesch?ftsf?hrer: Hans-J?rgen de Fries,
> >>>> Jens Gransee, Manfred Klumpp,
> >>>> Robert Reisch
> >>>>
> >>>> _______________________________________________
> >>>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> >>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> >>>>
> >>>> Project Home: http://www.clusterlabs.org
> >>>> Getting started:
> >> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> >>>> Bugs: http://bugs.clusterlabs.org
> >>>>
> >>>
> >>>
> >>>
> >>> --
> >>> esta es mi vida e me la vivo hasta que dios quiera
> >>> -------------- next part --------------
> >>> An HTML attachment was scrubbed...
> >>> URL: <
> >>
> http://oss.clusterlabs.org/pipermail/pacemaker/attachments/20120419/00a36dbe/attachment.html
> >>>
> >>>
> >>> ------------------------------
> >>>
> >>> _______________________________________________
> >>> Pacemaker mailing list
> >>> Pacemaker at oss.clusterlabs.org
> >>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> >>>
> >>>
> >>> End of Pacemaker Digest, Vol 53, Issue 38
> >>> *****************************************
> >>
> >> _______________________________________________
> >> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> >> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> >>
> >> Project Home: http://www.clusterlabs.org
> >> Getting started:
> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> >> Bugs: http://bugs.clusterlabs.org
> >>
> >
> >
> >
> > --
> > esta es mi vida e me la vivo hasta que dios quiera
> > -------------- next part --------------
> > An HTML attachment was scrubbed...
> > URL: <
> http://oss.clusterlabs.org/pipermail/pacemaker/attachments/20120419/63b888a3/attachment.html
> >
> >
> > ------------------------------
> >
> > _______________________________________________
> > Pacemaker mailing list
> > Pacemaker at oss.clusterlabs.org
> > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> >
> >
> > End of Pacemaker Digest, Vol 53, Issue 40
> > *****************************************
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>
--
esta es mi vida e me la vivo hasta que dios quiera
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://oss.clusterlabs.org/pipermail/pacemaker/attachments/20120419/f21bf7c7/attachment-0001.html>
More information about the Pacemaker
mailing list