[Pacemaker] Pacemaker and LDAP (389 Directory Service)
veghead
sean at studyblue.com
Mon Jun 27 21:33:12 UTC 2011
Sorry for the questions. Some days my brain is just slow. :)
Serge Dubrouski <sergeyfd at ...> writes:
> If you want to make your LDAP independent from IP just remove your
> collocation:colocation ldap-with-eip inf: elastic_ip ldap-clone
Is that really what I want to do? I mean, I need the elastic ip assigned to
~one~ of the machines... And if LDAP fails on that machine, I need Pacemaker to
start the Elastic IP on the other machine.
If I remove the co-location, won't the elastic_ip resource just stay where it
is? Regardless of what happens to LDAP?
> But I'd rather try to find out why monitoring for IP fails. May bet
> it just needs an increased timeout on monitor operation, though it
> looks like you've already increased it. What's in your log files
> when that monitor fails?
Originally, I had the monitor on the elastic_ip resource set to 10 seconds. The
error in the logs was:
---snip---
pengine: [16980]: notice: unpack_rsc_op: Operation elastic_ip_monitor_0 found
resource elastic_ip active on ldap1.example.ec2
pengine: [16980]: WARN: unpack_rsc_op: Processing failed op
elastic_ip_monitor_10000 on ldap1.example.ec2: unknown exec error (-2)
pengine: [16980]: WARN: unpack_rsc_op: Processing failed op elastic_ip_stop_0 on
ldap1.example.ec2: unknown exec error (-2)
pengine: [16980]: info: native_add_running: resource elastic_ip isnt managed
pengine: [16980]: notice: unpack_rsc_op: Operation ldap:1_monitor_0 found
resource ldap:1 active on ldap2.example.ec2
pengine: [16980]: WARN: unpack_rsc_op: Processing failed op elastic_ip_start_0
on ldap2.example.ec2: unknown exec error (-2)
pengine: [16980]: notice: native_print: elastic_ip (lsb:elastic-ip):
Started ldap1.example.ec2 (unmanaged) FAILED
pengine: [16980]: notice: clone_print: Clone Set: ldap-clone
pengine: [16980]: notice: short_print: Stopped: [ ldap:0 ldap:1 ]
pengine: [16980]: info: get_failcount: elastic_ip has failed INFINITY times on
ldap1.example.ec2
pengine: [16980]: WARN: common_apply_stickiness: Forcing elastic_ip away from
ldap1.example.ec2 after 1000000 failures (max=1000000)
pengine: [16980]: info: get_failcount: elastic_ip has failed INFINITY times on
ldap2.example.ec2
pengine: [16980]: WARN: common_apply_stickiness: Forcing elastic_ip away from
ldap2.example.ec2 after 1000000 failures (max=1000000)
pengine: [16980]: info: native_color: Unmanaged resource elastic_ip allocated to
'nowhere': failed
pengine: [16980]: notice: RecurringOp: Start recurring monitor (15s) for ldap:0
on ldap1.example.ec2
pengine: [16980]: notice: RecurringOp: Start recurring monitor (15s) for ldap:1
on ldap2.example.ec2
pengine: [16980]: notice: LogActions: Leave resource elastic_ip
(Started unmanaged)
pengine: [16980]: notice: LogActions: Start ldap:0 (ldap1.example.ec2)
pengine: [16980]: notice: LogActions: Start ldap:1 (ldap2.example.ec2)
---snip---
Now that I have set the monitor interval for the elastic_ip resource to "0", it
keeps thinking everything is either stopped or should be stopped:
---snip---
pengine: [7287]: notice: unpack_rsc_op: Operation elastic_ip_monitor_0 found
resource elastic_ip active on ldap1.example.ec2
pengine: [7287]: notice: unpack_rsc_op: Operation ldap:0_monitor_0 found
resource ldap:0 active on ldap2.example.ec2
pengine: [7287]: notice: native_print: elastic_ip (lsb:elastic-ip):
Stopped
pengine: [7287]: notice: clone_print: Clone Set: ldap-clone
pengine: [7287]: notice: short_print: Stopped: [ ldap:0 ldap:1 ]
pengine: [7287]: notice: LogActions: Leave resource elastic_ip (Stopped)
pengine: [7287]: notice: LogActions: Leave resource ldap:0 (Stopped)
pengine: [7287]: notice: LogActions: Leave resource ldap:1 (Stopped)
---snip---
Very strange.
More information about the Pacemaker
mailing list