[Pacemaker] Announce: Pacemaker 1.1.10 now available

Takatoshi MATSUO matsuo.tak at gmail.com
Wed Jul 31 20:18:42 EDT 2013


Hi Andrew

I'm about to collect logs of crm_report,
but Rainer already provides it.

Could you see his reports ?

Thanks,
Takatoshi MATSUO


2013/8/1 Rainer Brestan <rainer.brestan at gmx.net>:
> Base situation for the logs:
> Pacemaker stop on int2node1 and int2node2
> Master/slave resource msABC already configured.
> Included in the crm_report is also per node a file "a", this is the one,
> which the modified Stateful RA writes to log each action performed.
>
> 1.) 19:22:25 start Pacemaker on int2node1
> https://www.dropbox.com/s/ftbdl71ol2iyi42/step1.log.tar.bz2
> monitor on master is called
>
> 2.) 19:32:14 start Pacemaker on int2node2
> https://www.dropbox.com/s/s3jnxqvod9mlyz1/step2.log.tar.bz2
> monitor on master is not called any more
>
> 3.) 19:37:14 stop Pacemaker on int2node2
> https://www.dropbox.com/s/w75myab6fxh7mak/step3.log.tar.bz2
> monitor on master is still not called any more
>
> 4.) 19:42:14 start Pacemaker on in2node2
> https://www.dropbox.com/s/p00wl9kx4vwhilh/step4.log.tar.bz2
> monitor on master is called normally
>
> Hope this gives a clearer picture which component has forgotten the monitor
> action.
>
> Rainer
> Gesendet: Mittwoch, 31. Juli 2013 um 14:19 Uhr
>
> Von: "Andrew Beekhof" <andrew at beekhof.net>
> An: "The Pacemaker cluster resource manager" <pacemaker at oss.clusterlabs.org>
> Betreff: Re: [Pacemaker] Announce: Pacemaker 1.1.10 now available
>
> On 31/07/2013, at 5:17 PM, Rainer Brestan <rainer.brestan at gmx.net> wrote:
>
>> Modified the RA to log each action call performed and from this log there
>> is no call of monitor action.
>>
>> From the logs i do not think it is the policy engine, it might be the LRM
>> part of crmd (the is the only relevant change be seen after git diff between
>> 1.1.10-rc7 and 1.1.10).
>
> Ok. Can you still send me a crm_report though?
> Even if the PE isn't at fault, it shows me what the cib looked like at the
> time which can be surprisingly helpful.
> And it would have all the logs...
>
>>
>> Explanation of the below log:
>> primitive resABC ocf:heartbeat:Stateful \
>> op start interval="0s" timeout="60s" on-fail="restart" \
>> op monitor interval="30s" timeout="60s" on-fail="restart" \
>> op promote interval="0s" timeout="60s" on-fail="restart" \
>> op demote interval="0" timeout="60s" on-fail="restart" \
>> op stop interval="0" timeout="60s" on-fail="restart" \
>> op monitor interval="20" role="Master" timeout="60"
>> ms msABC resABC \
>> meta master-max="1" master-node-max="1" clone-max="2" clone-node-max="1"
>> notify="true"
>> crm_mon at begin of log:
>> Last updated: Wed Jul 31 08:30:57 2013
>> Last change: Tue Jul 30 13:01:36 2013 via crmd on int2node1
>> Stack: corosync
>> Current DC: int2node1 (1743917066) - partition with quorum
>> Version: 1.1.10-1.el6-368c726
>> 2 Nodes configured
>> 5 Resources configured
>> Online: [ int2node1 int2node2 ]
>> Master/Slave Set: msABC [resABC]
>> Masters: [ int2node1 ]
>> Slaves: [ int2node2 ]
>> crm_mon at end of log:
>> Last updated: Wed Jul 31 08:55:29 2013
>> Last change: Tue Jul 30 13:01:36 2013 via crmd on int2node1
>> Stack: corosync
>> Current DC: int2node1 (1743917066) - partition with quorum
>> Version: 1.1.10-1.el6-368c726
>> 2 Nodes configured
>> 5 Resources configured
>> Online: [ int2node1 ]
>> OFFLINE: [ int2node2 ]
>> Master/Slave Set: msABC [resABC]
>> Masters: [ int2node1 ]
>>
>> int2node1 is running, int2node2 is started
>> 2013-07-31T08:30:52.631+02:00 int2node1 pengine[16443] notice: notice:
>> LogActions: Start resABC:1 (int2node2)
>> 2013-07-31T08:30:52.638+02:00 int2node1 crmd[16444] notice: notice:
>> te_rsc_command: Initiating action 9: monitor resABC:1_monitor_0 on int2node2
>> 2013-07-31T08:30:52.638+02:00 int2node1 crmd[16444] notice: notice:
>> te_rsc_command: Initiating action 54: notify resABC_pre_notify_start_0 on
>> int2node1 (local)
>> 2013-07-31T08:30:52.681+02:00 int2node1 crmd[16444] notice: notice:
>> process_lrm_event: LRM operation resABC_notify_0 (call=64, rc=0,
>> cib-update=0, confirmed=true) ok
>> 2013-07-31T08:30:52.780+02:00 int2node1 crmd[16444] notice: notice:
>> te_rsc_command: Initiating action 25: start resABC:1_start_0 on int2node2
>> 2013-07-31T08:30:52.940+02:00 int2node1 crmd[16444] notice: notice:
>> te_rsc_command: Initiating action 55: notify resABC_post_notify_start_0 on
>> int2node1 (local)
>> 2013-07-31T08:30:52.943+02:00 int2node1 crmd[16444] notice: notice:
>> te_rsc_command: Initiating action 56: notify resABC:1_post_notify_start_0 on
>> int2node2
>> 2013-07-31T08:30:52.982+02:00 int2node1 crmd[16444] notice: notice:
>> process_lrm_event: LRM operation resABC_notify_0 (call=67, rc=0,
>> cib-update=0, confirmed=true) ok
>> 2013-07-31T08:30:52.992+02:00 int2node1 crmd[16444] notice: notice:
>> te_rsc_command: Initiating action 24: monitor resABC_monitor_20000 on
>> int2node1 (local)
>> 2013-07-31T08:30:52.996+02:00 int2node1 crmd[16444] notice: notice:
>> te_rsc_command: Initiating action 26: monitor resABC:1_monitor_30000 on
>> int2node2
>> 2013-07-31T08:30:53.035+02:00 int2node1 crmd[16444] notice: notice:
>> process_lrm_event: LRM operation resABC_monitor_20000 (call=70, rc=8,
>> cib-update=149, confirmed=false) master
>>
>> At this point int2node2 is stopped.
>> 2013-07-31T08:37:51.457+02:00 int2node1 crmd[16444] notice: notice:
>> do_state_transition: State transition S_IDLE -> S_POLICY_ENGINE [
>> input=I_PE_CALC cause=C_FSA_INTERNAL origin=abort_transition_graph ]
>> 2013-07-31T08:37:51.462+02:00 int2node1 pengine[16443] notice: notice:
>> unpack_config: On loss of CCM Quorum: Ignore
>> 2013-07-31T08:37:51.465+02:00 int2node1 pengine[16443] notice: notice:
>> stage6: Scheduling Node int2node2 for shutdown
>> 2013-07-31T08:37:51.466+02:00 int2node1 pengine[16443] notice: notice:
>> LogActions: Stop resABC:1 (int2node2)
>> 2013-07-31T08:37:51.469+02:00 int2node1 pengine[16443] notice: notice:
>> process_pe_message: Calculated Transition 86:
>> /var/lib/pacemaker/pengine/pe-input-125.bz2
>> 2013-07-31T08:37:51.471+02:00 int2node1 crmd[16444] notice: notice:
>> te_rsc_command: Initiating action 56: notify resABC_pre_notify_stop_0 on
>> int2node1 (local)
>> 2013-07-31T08:37:51.474+02:00 int2node1 crmd[16444] notice: notice:
>> te_rsc_command: Initiating action 58: notify resABC_pre_notify_stop_0 on
>> int2node2
>> 2013-07-31T08:37:51.512+02:00 int2node1 crmd[16444] notice: notice:
>> process_lrm_event: LRM operation resABC_notify_0 (call=74, rc=0,
>> cib-update=0, confirmed=true) ok
>> 2013-07-31T08:37:51.514+02:00 int2node1 crmd[16444] notice: notice:
>> te_rsc_command: Initiating action 23: stop resABC_stop_0 on int2node2
>> 2013-07-31T08:37:51.654+02:00 int2node1 crmd[16444] notice: notice:
>> te_rsc_command: Initiating action 57: notify resABC_post_notify_stop_0 on
>> int2node1 (local)
>> 2013-07-31T08:37:51.699+02:00 int2node1 crmd[16444] notice: notice:
>> process_lrm_event: LRM operation resABC_notify_0 (call=78, rc=0,
>> cib-update=0, confirmed=true) ok
>> 2013-07-31T08:37:51.699+02:00 int2node1 crmd[16444] notice: notice:
>> run_graph: Transition 86 (Complete=13, Pending=0, Fired=0, Skipped=2,
>> Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-125.bz2): Stopped
>> 2013-07-31T08:37:51.705+02:00 int2node1 pengine[16443] notice: notice:
>> unpack_config: On loss of CCM Quorum: Ignore
>> 2013-07-31T08:37:51.705+02:00 int2node1 pengine[16443] notice: notice:
>> stage6: Scheduling Node int2node2 for shutdown
>> 2013-07-31T08:37:51.706+02:00 int2node1 pengine[16443] notice: notice:
>> process_pe_message: Calculated Transition 87:
>> /var/lib/pacemaker/pengine/pe-input-126.bz2
>> 2013-07-31T08:37:51.707+02:00 int2node1 crmd[16444] notice: notice:
>> run_graph: Transition 87 (Complete=1, Pending=0, Fired=0, Skipped=0,
>> Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-126.bz2): Complete
>> 2013-07-31T08:37:51.707+02:00 int2node1 crmd[16444] notice: notice:
>> do_state_transition: State transition S_TRANSITION_ENGINE -> S_IDLE [
>> input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd ]
>> 2013-07-31T08:37:51.720+02:00 int2node1 crmd[16444] notice: notice:
>> peer_update_callback: do_shutdown of int2node2 (op 45) is complete
>>
>> Output from RA on int2node1:
>> Wed Jul 31 08:30:52 CEST 2013 resABC: operation notify, type pre,
>> operation start
>> Wed Jul 31 08:30:52 CEST 2013 resABC: operation notify, type post,
>> operation start
>> Wed Jul 31 08:30:53 CEST 2013 resABC: operation monitor, type , operation
>> Wed Jul 31 08:31:13 CEST 2013 resABC: operation monitor, type , operation
>> Wed Jul 31 08:31:33 CEST 2013 resABC: operation monitor, type , operation
>> Wed Jul 31 08:31:53 CEST 2013 resABC: operation monitor, type , operation
>> Wed Jul 31 08:32:13 CEST 2013 resABC: operation monitor, type , operation
>> Wed Jul 31 08:32:33 CEST 2013 resABC: operation monitor, type , operation
>> Wed Jul 31 08:32:53 CEST 2013 resABC: operation monitor, type , operation
>> Wed Jul 31 08:33:13 CEST 2013 resABC: operation monitor, type , operation
>> Wed Jul 31 08:33:33 CEST 2013 resABC: operation monitor, type , operation
>> Wed Jul 31 08:33:53 CEST 2013 resABC: operation monitor, type , operation
>> Wed Jul 31 08:34:13 CEST 2013 resABC: operation monitor, type , operation
>> Wed Jul 31 08:34:33 CEST 2013 resABC: operation monitor, type , operation
>> Wed Jul 31 08:34:53 CEST 2013 resABC: operation monitor, type , operation
>> Wed Jul 31 08:35:13 CEST 2013 resABC: operation monitor, type , operation
>> Wed Jul 31 08:35:33 CEST 2013 resABC: operation monitor, type , operation
>> Wed Jul 31 08:35:53 CEST 2013 resABC: operation monitor, type , operation
>> Wed Jul 31 08:36:13 CEST 2013 resABC: operation monitor, type , operation
>> Wed Jul 31 08:36:33 CEST 2013 resABC: operation monitor, type , operation
>> Wed Jul 31 08:36:53 CEST 2013 resABC: operation monitor, type , operation
>> Wed Jul 31 08:37:13 CEST 2013 resABC: operation monitor, type , operation
>> Wed Jul 31 08:37:33 CEST 2013 resABC: operation monitor, type , operation
>> Wed Jul 31 08:37:51 CEST 2013 resABC: operation notify, type pre,
>> operation stop
>> Wed Jul 31 08:37:51 CEST 2013 resABC: operation notify, type post,
>> operation stop
>>
>> After 08:37:51 no log output from Pacemaker for resABC, nor any output
>> from RA on int2node1.
>>
>> Gesendet: Mittwoch, 31. Juli 2013 um 02:10 Uhr
>> Von: "Andrew Beekhof" <andrew at beekhof.net>
>> An: "The Pacemaker cluster resource manager"
>> <pacemaker at oss.clusterlabs.org>
>> Betreff: Re: [Pacemaker] Announce: Pacemaker 1.1.10 now available
>>
>> On 30/07/2013, at 9:13 PM, Rainer Brestan <rainer.brestan at gmx.net> wrote:
>>
>> > I can agree, Master monitor operation is broken in 1.1.10 release.
>> > When the slave monitor action is started, the master monitor action is
>> > not called any more.
>>
>> Based on?
>>
>> >
>> > I have created a setup with Stateful resource with two nodes.
>> > Then the Pacemaker installation is changed to different versions without
>> > changing the configuration part of the CIB.
>> >
>> > Result:
>> > 1.1.10-rc5, 1.1.10-rc6 and 1.1.10-rc7 does not have this error
>> > 1.1.10-1 release has the error
>> >
>> > Installation order (just that anybody know how it was done):
>> > 1.1.10-1 -> error
>> > 1.1.10-rc5 -> no error
>> > 1.1.10-rc6 -> no error
>> > 1.1.10-rc7 -> no error
>> > 1.1.10-1 -> error
>> >
>> > Rainer
>> > Gesendet: Freitag, 26. Juli 2013 um 09:32 Uhr
>> > Von: "Takatoshi MATSUO" <matsuo.tak at gmail.com>
>> > An: "The Pacemaker cluster resource manager"
>> > <pacemaker at oss.clusterlabs.org>
>> > Betreff: Re: [Pacemaker] Announce: Pacemaker 1.1.10 now available
>> > Hi
>> >
>> > I used Stateful RA and caught a same issue.
>> >
>> > 1. before starting slave
>> >
>> > # crm_simulate -VVV -S -x /var/lib/pacemaker/pengine/pe-input-1543.bz2
>> > | grep "Resource action"
>> > * Resource action: stateful monitor=2000 on 16-sl6
>> >
>> > 2. starting slave
>> > # crm_simulate -VVV -S -x /var/lib/pacemaker/pengine/pe-input-1544.bz2
>> > | grep "Resource action"
>> > * Resource action: stateful monitor on 17-sl6
>> > * Resource action: stateful notify on 16-sl6
>> > * Resource action: stateful start on 17-sl6
>> > * Resource action: stateful notify on 16-sl6
>> > * Resource action: stateful notify on 17-sl6
>> > * Resource action: stateful monitor=3000 on 17-sl6
>> >
>> > 3. after
>> > # crm_simulate -VVV -S -x /var/lib/pacemaker/pengine/pe-input-1545.bz2
>> > | grep "Resource action"
>> > * Resource action: stateful monitor=3000 on 17-sl6
>> >
>> > Monitor=2000 is deleted.
>> > Is this correct ?
>> >
>> >
>> > My setting
>> > --------
>> > property \
>> > no-quorum-policy="ignore" \
>> > stonith-enabled="false"
>> >
>> > rsc_defaults \
>> > resource-stickiness="INFINITY" \
>> > migration-threshold="1"
>> >
>> > ms msStateful stateful \
>> > meta \
>> > master-max="1" \
>> > master-node-max="1" \
>> > clone-max="2" \
>> > clone-node-max="1" \
>> > notify="true"
>> >
>> > primitive stateful ocf:heartbeat:Stateful \
>> > op start timeout="60s" interval="0s" on-fail="restart" \
>> > op monitor timeout="60s" interval="3s" on-fail="restart" \
>> > op monitor timeout="60s" interval="2s" on-fail="restart" role="Master" \
>> > op promote timeout="60s" interval="0s" on-fail="restart" \
>> > op demote timeout="60s" interval="0s" on-fail="stop" \
>> > op stop timeout="60s" interval="0s" on-fail="block"
>> > --------
>> >
>> > Regards,
>> > Takatoshi MATSUO
>> >
>> > 2013/7/26 Takatoshi MATSUO <matsuo.tak at gmail.com>:
>> > > Hi
>> > >
>> > > My report is late for 1.1.10 :(
>> > >
>> > > I am using pacemaker 1.1.10-0.1.ab2e209.git.
>> > > It seems that master's monitor is stopped when slave is started.
>> > >
>> > > Does someone encounter same problem ?
>> > > I attach a log and settings.
>> > >
>> > >
>> > > Thanks,
>> > > Takatoshi MATSUO
>> > >
>> > > 2013/7/26 Digimer <lists at alteeve.ca>:
>> > >> Congrats!! I know this was a long time in the making.
>> > >>
>> > >> digimer
>> > >>
>> > >>
>> > >> On 25/07/13 20:43, Andrew Beekhof wrote:
>> > >>>
>> > >>> Announcing the release of Pacemaker 1.1.10
>> > >>>
>> > >>> https://github.com/ClusterLabs/pacemaker/releases/Pacemaker-1.1.10
>> > >>>
>> > >>> There were three changes of note since rc7:
>> > >>>
>> > >>> + Bug cl#5161 - crmd: Prevent memory leak in operation cache
>> > >>> + cib: Correctly read back archived configurations if the primary is
>> > >>> corrupted
>> > >>> + cman: Do not pretend we know the state of nodes we've never seen
>> > >>>
>> > >>> Along with assorted bug fixes, the major topics for this release
>> > >>> were:
>> > >>>
>> > >>> - stonithd fixes
>> > >>> - fixing memory leaks, often caused by incorrect use of glib
>> > >>> reference
>> > >>> counting
>> > >>> - supportability improvements (code cleanup and deduplication,
>> > >>> standardized error codes)
>> > >>>
>> > >>> Release candidates for the next Pacemaker release (1.1.11) can be
>> > >>> expected some time around Novemeber.
>> > >>>
>> > >>> A big thankyou to everyone that spent time testing the release
>> > >>> candidates and/or contributed patches. However now that Pacemaker is
>> > >>> perfect, anyone reporting bugs will be shot :-)
>> > >>>
>> > >>> To build `rpm` packages:
>> > >>>
>> > >>> 1. Clone the current sources:
>> > >>>
>> > >>> # git clone --depth 0 git://github.com/ClusterLabs/pacemaker.git
>> > >>> # cd pacemaker
>> > >>>
>> > >>> 1. Install dependancies (if you haven't already)
>> > >>>
>> > >>> [Fedora] # sudo yum install -y yum-utils
>> > >>> [ALL] # make rpm-dep
>> > >>>
>> > >>> 1. Build Pacemaker
>> > >>>
>> > >>> # make release
>> > >>>
>> > >>> 1. Copy and deploy as needed
>> > >>>
>> > >>> ## Details - 1.1.10 - final
>> > >>>
>> > >>> Changesets: 602
>> > >>> Diff: 143 files changed, 8162 insertions(+), 5159 deletions(-)
>> > >>>
>> > >>> ## Highlights
>> > >>>
>> > >>> ### Features added since Pacemaker-1.1.9
>> > >>>
>> > >>> + Core: Convert all exit codes to positive errno values
>> > >>> + crm_error: Add the ability to list and print error symbols
>> > >>> + crm_resource: Allow individual resources to be reprobed
>> > >>> + crm_resource: Allow options to be set recursively
>> > >>> + crm_resource: Implement --ban for moving resources away from nodes
>> > >>> and --clear (replaces --unmove)
>> > >>> + crm_resource: Support OCF tracing when using
>> > >>> --force-(check|start|stop)
>> > >>> + PE: Allow active nodes in our current membership to be fenced
>> > >>> without
>> > >>> quorum
>> > >>> + PE: Suppress meaningless IDs when displaying anonymous clone
>> > >>> status
>> > >>> + Turn off auto-respawning of systemd services when the cluster
>> > >>> starts
>> > >>> them
>> > >>> + Bug cl#5128 - pengine: Support maintenance mode for a single node
>> > >>>
>> > >>> ### Changes since Pacemaker-1.1.9
>> > >>>
>> > >>> + crmd: cib: stonithd: Memory leaks resolved and improved use of
>> > >>> glib
>> > >>> reference counting
>> > >>> + attrd: Fixes deleted attributes during dc election
>> > >>> + Bug cf#5153 - Correctly display clone failcounts in crm_mon
>> > >>> + Bug cl#5133 - pengine: Correctly observe on-fail=block for failed
>> > >>> demote operation
>> > >>> + Bug cl#5148 - legacy: Correctly remove a node that used to have a
>> > >>> different nodeid
>> > >>> + Bug cl#5151 - Ensure node names are consistently compared without
>> > >>> case
>> > >>> + Bug cl#5152 - crmd: Correctly clean up fenced nodes during
>> > >>> membership
>> > >>> changes
>> > >>> + Bug cl#5154 - Do not expire failures when on-fail=block is present
>> > >>> + Bug cl#5155 - pengine: Block the stop of resources if any
>> > >>> depending
>> > >>> resource is unmanaged
>> > >>> + Bug cl#5157 - Allow migration in the absence of some colocation
>> > >>> constraints
>> > >>> + Bug cl#5161 - crmd: Prevent memory leak in operation cache
>> > >>> + Bug cl#5164 - crmd: Fixes crash when using pacemaker-remote
>> > >>> + Bug cl#5164 - pengine: Fixes segfault when calculating transition
>> > >>> with remote-nodes.
>> > >>> + Bug cl#5167 - crm_mon: Only print "stopped" node list for
>> > >>> incomplete
>> > >>> clone sets
>> > >>> + Bug cl#5168 - Prevent clones from being bounced around the cluster
>> > >>> due to location constraints
>> > >>> + Bug cl#5170 - Correctly support on-fail=block for clones
>> > >>> + cib: Correctly read back archived configurations if the primary is
>> > >>> corrupted
>> > >>> + cib: The result is not valid when diffs fail to apply cleanly for
>> > >>> CLI
>> > >>> tools
>> > >>> + cib: Restore the ability to embed comments in the configuration
>> > >>> + cluster: Detect and warn about node names with capitals
>> > >>> + cman: Do not pretend we know the state of nodes we've never seen
>> > >>> + cman: Do not unconditionally start cman if it is already running
>> > >>> + cman: Support non-blocking CPG calls
>> > >>> + Core: Ensure the blackbox is saved on abnormal program termination
>> > >>> + corosync: Detect the loss of members for which we only know the
>> > >>> nodeid
>> > >>> + corosync: Do not pretend we know the state of nodes we've never
>> > >>> seen
>> > >>> + corosync: Ensure removed peers are erased from all caches
>> > >>> + corosync: Nodes that can persist in sending CPG messages must be
>> > >>> alive afterall
>> > >>> + crmd: Do not get stuck in S_POLICY_ENGINE if a node we couldn't
>> > >>> fence
>> > >>> returns
>> > >>> + crmd: Do not update fail-count and last-failure for old failures
>> > >>> + crmd: Ensure all membership operations can complete while trying
>> > >>> to
>> > >>> cancel a transition
>> > >>> + crmd: Ensure operations for cleaned up resources don't block
>> > >>> recovery
>> > >>> + crmd: Ensure we return to a stable state if there have been too
>> > >>> many
>> > >>> fencing failures
>> > >>> + crmd: Initiate node shutdown if another node claims to have
>> > >>> successfully fenced us
>> > >>> + crmd: Prevent messages for remote crmd clients from being relayed
>> > >>> to
>> > >>> wrong daemons
>> > >>> + crmd: Properly handle recurring monitor operations for remote-node
>> > >>> agent
>> > >>> + crmd: Store last-run and last-rc-change for all operations
>> > >>> + crm_mon: Ensure stale pid files are updated when a new process is
>> > >>> started
>> > >>> + crm_report: Correctly collect logs when 'uname -n' reports fully
>> > >>> qualified names
>> > >>> + fencing: Fail the operation once all peers have been exhausted
>> > >>> + fencing: Restore the ability to manually confirm that fencing
>> > >>> completed
>> > >>> + ipc: Allow unpriviliged clients to clean up after server failures
>> > >>> + ipc: Restore the ability for members of the haclient group to
>> > >>> connect
>> > >>> to the cluster
>> > >>> + legacy: Support "crm_node --remove" with a node name for corosync
>> > >>> plugin (bnc#805278)
>> > >>> + lrmd: Default to the upstream location for resource agent scratch
>> > >>> directory
>> > >>> + lrmd: Pass errors from lsb metadata generation back to the caller
>> > >>> + pengine: Correctly handle resources that recover before we operate
>> > >>> on
>> > >>> them
>> > >>> + pengine: Delete the old resource state on every node whenever the
>> > >>> resource type is changed
>> > >>> + pengine: Detect constraints with inappropriate actions (ie.
>> > >>> promote
>> > >>> for a clone)
>> > >>> + pengine: Ensure per-node resource parameters are used during
>> > >>> probes
>> > >>> + pengine: If fencing is unavailable or disabled, block further
>> > >>> recovery for resources that fail to stop
>> > >>> + pengine: Implement the rest of get_timet_now() and rename to
>> > >>> get_effective_time
>> > >>> + pengine: Re-initiate _active_ recurring monitors that previously
>> > >>> failed but have timed out
>> > >>> + remote: Workaround for inconsistent tls handshake behavior between
>> > >>> gnutls versions
>> > >>> + systemd: Ensure we get shut down correctly by systemd
>> > >>> + systemd: Reload systemd after adding/removing override files for
>> > >>> cluster services
>> > >>> + xml: Check for and replace non-printing characters with their
>> > >>> octal
>> > >>> equivalent while exporting xml text
>> > >>> + xml: Prevent lockups by setting a more reliable buffer allocation
>> > >>> strategy
>> > >>>
>> > >>>
>> > >>> _______________________________________________
>> > >>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>> > >>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>> > >>>
>> > >>> Project Home: http://www.clusterlabs.org
>> > >>> Getting started:
>> > >>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> > >>> Bugs: http://bugs.clusterlabs.org
>> > >>>
>> > >>
>> > >>
>> > >> --
>> > >> Digimer
>> > >> Papers and Projects: https://alteeve.ca/w/
>> > >> What if the cure for cancer is trapped in the mind of a person
>> > >> without
>> > >> access to education?
>> > >>
>> > >>
>> > >> _______________________________________________
>> > >> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>> > >> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>> > >>
>> > >> Project Home: http://www.clusterlabs.org
>> > >> Getting started:
>> > >> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> > >> Bugs: http://bugs.clusterlabs.org
>> >
>> > _______________________________________________
>> > Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>> > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>> >
>> > Project Home: http://www.clusterlabs.org
>> > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> > Bugs: http://bugs.clusterlabs.org
>> > _______________________________________________
>> > Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>> > http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>> >
>> > Project Home: http://www.clusterlabs.org
>> > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> > Bugs: http://bugs.clusterlabs.org
>>
>>
>> _______________________________________________
>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>
>> Project Home: http://www.clusterlabs.org
>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> Bugs: http://bugs.clusterlabs.org
>> _______________________________________________
>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>
>> Project Home: http://www.clusterlabs.org
>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> Bugs: http://bugs.clusterlabs.org
>
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>




More information about the Pacemaker mailing list