[Pacemaker] 'crm configure edit' failed with "Timer expired"

Lonni J Friedman netllama at gmail.com
Thu Oct 18 14:05:45 EDT 2012


Nevermind, I figured it out.  What I had neglected to mention
originally was that I had just upgraded both nodes from Fedora16 to
17, and with that came the jump from corosync-1.x to 2.x.  Apparently
the corosync.conf that I was using with 1.x doesn't work with 2.x.  I
blew it away, and created a new corosync.conf, and now the nodes are
talking again.

Sorry for the noise.

On Thu, Oct 18, 2012 at 10:25 AM, Lonni J Friedman <netllama at gmail.com> wrote:
> Both nodes can ssh to each other, selinux is disabled, and there are
> currently no iptables rules in force.  So I'm not sure why the systems
> can't communicate.
>
> I looked at /var/log/cluster/corosync.log but didn't see any obvious
> problems.  Would you like to see the logs from both nodes?
>
> What else can I do to investigate & debug this?
>
> thanks
>
>
> On Wed, Oct 17, 2012 at 8:24 PM, Andrew Beekhof <andrew at beekhof.net> wrote:
>> This is your resource:
>>    Current DC: NONE
>>
>> All the "crm configure" commands try to talk to the DC, and you don't have one.
>> Normally one would be elected quite quickly, you may have a
>> network/filewall issue.
>>
>> On Thu, Oct 18, 2012 at 10:37 AM, Lonni J Friedman <netllama at gmail.com> wrote:
>>> I'm running Fedora17, with pacemaker-1.18.  I just tried to make a
>>> configuration change with crmsh, and it failed as follows:
>>> ##########
>>> # crm configure edit
>>> Call cib_replace failed (-62): Timer expired
>>> <null>
>>> ERROR: could not replace cib
>>> INFO: offending xml: <configuration>
>>>         <crm_config>
>>>                 <cluster_property_set id="cib-bootstrap-options">
>>>                         <nvpair id="cib-bootstrap-options-dc-version" name="dc-version"
>>> value="1.1.8-2.fc16-394e906"/>
>>>                         <nvpair id="cib-bootstrap-options-cluster-infrastructure"
>>> name="cluster-infrastructure" value="openais"/>
>>>                         <nvpair id="cib-bootstrap-options-expected-quorum-votes"
>>> name="expected-quorum-votes" value="2"/>
>>>                         <nvpair id="cib-bootstrap-options-stonith-enabled"
>>> name="stonith-enabled" value="false"/>
>>>                         <nvpair id="cib-bootstrap-options-no-quorum-policy"
>>> name="no-quorum-policy" value="ignore"/>
>>>                 </cluster_property_set>
>>>         </crm_config>
>>>         <nodes>
>>>                 <node id="farm-ljf0" uname="farm-ljf0">
>>>                         <instance_attributes id="nodes-farm-ljf0">
>>>                                 <nvpair id="nodes-farm-ljf0-standby" name="standby" value="off"/>
>>>                         </instance_attributes>
>>>                 </node>
>>>                 <node id="farm-ljf1" uname="farm-ljf1"/>
>>>         </nodes>
>>>         <resources>
>>>                 <master id="FS0_Clone">
>>>                         <meta_attributes id="FS0_Clone-meta_attributes">
>>>                                 <nvpair id="FS0_Clone-meta_attributes-master-max"
>>> name="master-max" value="1"/>
>>>                                 <nvpair id="FS0_Clone-meta_attributes-master-node-max"
>>> name="master-node-max" value="1"/>
>>>                                 <nvpair id="FS0_Clone-meta_attributes-clone-max" name="clone-max"
>>> value="2"/>
>>>                                 <nvpair id="FS0_Clone-meta_attributes-clone-node-max"
>>> name="clone-node-max" value="1"/>
>>>                                 <nvpair id="FS0_Clone-meta_attributes-notify" name="notify" value="true"/>
>>>                         </meta_attributes>
>>>                         <primitive class="ocf" id="FS0" provider="linbit" type="drbd">
>>>                                 <instance_attributes id="FS0-instance_attributes">
>>>                                         <nvpair id="FS0-instance_attributes-drbd_resource"
>>> name="drbd_resource" value="r0"/>
>>>                                 </instance_attributes>
>>>                                 <operations>
>>>                                         <op id="FS0-monitor-10s" interval="10s" name="monitor" role="Master"/>
>>>                                         <op id="FS0-monitor-30s" interval="30s" name="monitor" role="Slave"/>
>>>                                 </operations>
>>>                         </primitive>
>>>                 </master>
>>>                 <group id="g_services">
>>>                         <primitive class="ocf" id="ClusterIP" provider="heartbeat" type="IPaddr2">
>>>                                 <instance_attributes id="ClusterIP-instance_attributes">
>>>                                         <nvpair id="ClusterIP-instance_attributes-ip" name="ip"
>>> value="10.31.99.8"/>
>>>                                         <nvpair id="ClusterIP-instance_attributes-cidr_netmask"
>>> name="cidr_netmask" value="22"/>
>>>                                         <nvpair id="ClusterIP-instance_attributes-nic" name="nic" value="eth1"/>
>>>                                 </instance_attributes>
>>>                                 <operations>
>>>                                         <op id="ClusterIP-monitor-10s" interval="10s" name="monitor"/>
>>>                                 </operations>
>>>                                 <meta_attributes id="ClusterIP-meta_attributes">
>>>                                         <nvpair id="ClusterIP-meta_attributes-target-role"
>>> name="target-role" value="Started"/>
>>>                                 </meta_attributes>
>>>                         </primitive>
>>>                         <primitive class="ocf" id="FS0_drbd" provider="heartbeat" type="Filesystem">
>>>                                 <instance_attributes id="FS0_drbd-instance_attributes">
>>>                                         <nvpair id="FS0_drbd-instance_attributes-device" name="device"
>>> value="/dev/drbd0"/>
>>>                                         <nvpair id="FS0_drbd-instance_attributes-directory"
>>> name="directory" value="/mnt/sdb1"/>
>>>                                         <nvpair id="FS0_drbd-instance_attributes-fstype" name="fstype"
>>> value="xfs"/>
>>>                                 </instance_attributes>
>>>                                 <meta_attributes id="FS0_drbd-meta_attributes">
>>>                                         <nvpair id="FS0_drbd-meta_attributes-target-role"
>>> name="target-role" value="Started"/>
>>>                                 </meta_attributes>
>>>                         </primitive>
>>>                         <primitive class="systemd" id="FS0_nfs" type="nfs-server">
>>>                                 <operations>
>>>                                         <op id="FS0_nfs-monitor-10s" interval="10s" name="monitor"/>
>>>                                 </operations>
>>>                                 <meta_attributes id="FS0_nfs-meta_attributes">
>>>                                         <nvpair id="FS0_nfs-meta_attributes-target-role"
>>> name="target-role" value="Started"/>
>>>                                 </meta_attributes>
>>>                         </primitive>
>>>                 </group>
>>>         </resources>
>>>         <constraints>
>>>                 <rsc_colocation id="fs0_on_drbd" rsc="g_services" score="INFINITY"
>>> with-rsc="FS0_Clone" with-rsc-role="Master"/>
>>>                 <rsc_order first="FS0_Clone" first-action="promote"
>>> id="FS0_drbd-after-FS0" score="INFINITY" then="g_services"
>>> then-action="start"/>
>>>         </constraints>
>>> </configuration>
>>> ##########
>>>
>>> I'm confused why this failed.  In fact, no matter what I try to
>>> change, it always fails in the same fashion when I attempt to save the
>>> changes.
>>>
>>> "crm status" currently shows the following:
>>> #######
>>> Last updated: Wed Oct 17 16:36:45 2012
>>> Last change: Tue Oct 16 14:23:18 2012 via cibadmin on farm-ljf1
>>> Stack: openais
>>> Current DC: NONE
>>> 2 Nodes configured, 2 expected votes
>>> 5 Resources configured.
>>>
>>>
>>> OFFLINE: [ farm-ljf0 farm-ljf1 ]
>>> ######
>>>
>>> help?




More information about the Pacemaker mailing list