[Pacemaker] 'crm configure edit' failed with "Timer expired"
Lonni J Friedman
netllama at gmail.com
Thu Oct 18 18:05:45 UTC 2012
Nevermind, I figured it out. What I had neglected to mention
originally was that I had just upgraded both nodes from Fedora16 to
17, and with that came the jump from corosync-1.x to 2.x. Apparently
the corosync.conf that I was using with 1.x doesn't work with 2.x. I
blew it away, and created a new corosync.conf, and now the nodes are
talking again.
Sorry for the noise.
On Thu, Oct 18, 2012 at 10:25 AM, Lonni J Friedman <netllama at gmail.com> wrote:
> Both nodes can ssh to each other, selinux is disabled, and there are
> currently no iptables rules in force. So I'm not sure why the systems
> can't communicate.
>
> I looked at /var/log/cluster/corosync.log but didn't see any obvious
> problems. Would you like to see the logs from both nodes?
>
> What else can I do to investigate & debug this?
>
> thanks
>
>
> On Wed, Oct 17, 2012 at 8:24 PM, Andrew Beekhof <andrew at beekhof.net> wrote:
>> This is your resource:
>> Current DC: NONE
>>
>> All the "crm configure" commands try to talk to the DC, and you don't have one.
>> Normally one would be elected quite quickly, you may have a
>> network/filewall issue.
>>
>> On Thu, Oct 18, 2012 at 10:37 AM, Lonni J Friedman <netllama at gmail.com> wrote:
>>> I'm running Fedora17, with pacemaker-1.18. I just tried to make a
>>> configuration change with crmsh, and it failed as follows:
>>> ##########
>>> # crm configure edit
>>> Call cib_replace failed (-62): Timer expired
>>> <null>
>>> ERROR: could not replace cib
>>> INFO: offending xml: <configuration>
>>> <crm_config>
>>> <cluster_property_set id="cib-bootstrap-options">
>>> <nvpair id="cib-bootstrap-options-dc-version" name="dc-version"
>>> value="1.1.8-2.fc16-394e906"/>
>>> <nvpair id="cib-bootstrap-options-cluster-infrastructure"
>>> name="cluster-infrastructure" value="openais"/>
>>> <nvpair id="cib-bootstrap-options-expected-quorum-votes"
>>> name="expected-quorum-votes" value="2"/>
>>> <nvpair id="cib-bootstrap-options-stonith-enabled"
>>> name="stonith-enabled" value="false"/>
>>> <nvpair id="cib-bootstrap-options-no-quorum-policy"
>>> name="no-quorum-policy" value="ignore"/>
>>> </cluster_property_set>
>>> </crm_config>
>>> <nodes>
>>> <node id="farm-ljf0" uname="farm-ljf0">
>>> <instance_attributes id="nodes-farm-ljf0">
>>> <nvpair id="nodes-farm-ljf0-standby" name="standby" value="off"/>
>>> </instance_attributes>
>>> </node>
>>> <node id="farm-ljf1" uname="farm-ljf1"/>
>>> </nodes>
>>> <resources>
>>> <master id="FS0_Clone">
>>> <meta_attributes id="FS0_Clone-meta_attributes">
>>> <nvpair id="FS0_Clone-meta_attributes-master-max"
>>> name="master-max" value="1"/>
>>> <nvpair id="FS0_Clone-meta_attributes-master-node-max"
>>> name="master-node-max" value="1"/>
>>> <nvpair id="FS0_Clone-meta_attributes-clone-max" name="clone-max"
>>> value="2"/>
>>> <nvpair id="FS0_Clone-meta_attributes-clone-node-max"
>>> name="clone-node-max" value="1"/>
>>> <nvpair id="FS0_Clone-meta_attributes-notify" name="notify" value="true"/>
>>> </meta_attributes>
>>> <primitive class="ocf" id="FS0" provider="linbit" type="drbd">
>>> <instance_attributes id="FS0-instance_attributes">
>>> <nvpair id="FS0-instance_attributes-drbd_resource"
>>> name="drbd_resource" value="r0"/>
>>> </instance_attributes>
>>> <operations>
>>> <op id="FS0-monitor-10s" interval="10s" name="monitor" role="Master"/>
>>> <op id="FS0-monitor-30s" interval="30s" name="monitor" role="Slave"/>
>>> </operations>
>>> </primitive>
>>> </master>
>>> <group id="g_services">
>>> <primitive class="ocf" id="ClusterIP" provider="heartbeat" type="IPaddr2">
>>> <instance_attributes id="ClusterIP-instance_attributes">
>>> <nvpair id="ClusterIP-instance_attributes-ip" name="ip"
>>> value="10.31.99.8"/>
>>> <nvpair id="ClusterIP-instance_attributes-cidr_netmask"
>>> name="cidr_netmask" value="22"/>
>>> <nvpair id="ClusterIP-instance_attributes-nic" name="nic" value="eth1"/>
>>> </instance_attributes>
>>> <operations>
>>> <op id="ClusterIP-monitor-10s" interval="10s" name="monitor"/>
>>> </operations>
>>> <meta_attributes id="ClusterIP-meta_attributes">
>>> <nvpair id="ClusterIP-meta_attributes-target-role"
>>> name="target-role" value="Started"/>
>>> </meta_attributes>
>>> </primitive>
>>> <primitive class="ocf" id="FS0_drbd" provider="heartbeat" type="Filesystem">
>>> <instance_attributes id="FS0_drbd-instance_attributes">
>>> <nvpair id="FS0_drbd-instance_attributes-device" name="device"
>>> value="/dev/drbd0"/>
>>> <nvpair id="FS0_drbd-instance_attributes-directory"
>>> name="directory" value="/mnt/sdb1"/>
>>> <nvpair id="FS0_drbd-instance_attributes-fstype" name="fstype"
>>> value="xfs"/>
>>> </instance_attributes>
>>> <meta_attributes id="FS0_drbd-meta_attributes">
>>> <nvpair id="FS0_drbd-meta_attributes-target-role"
>>> name="target-role" value="Started"/>
>>> </meta_attributes>
>>> </primitive>
>>> <primitive class="systemd" id="FS0_nfs" type="nfs-server">
>>> <operations>
>>> <op id="FS0_nfs-monitor-10s" interval="10s" name="monitor"/>
>>> </operations>
>>> <meta_attributes id="FS0_nfs-meta_attributes">
>>> <nvpair id="FS0_nfs-meta_attributes-target-role"
>>> name="target-role" value="Started"/>
>>> </meta_attributes>
>>> </primitive>
>>> </group>
>>> </resources>
>>> <constraints>
>>> <rsc_colocation id="fs0_on_drbd" rsc="g_services" score="INFINITY"
>>> with-rsc="FS0_Clone" with-rsc-role="Master"/>
>>> <rsc_order first="FS0_Clone" first-action="promote"
>>> id="FS0_drbd-after-FS0" score="INFINITY" then="g_services"
>>> then-action="start"/>
>>> </constraints>
>>> </configuration>
>>> ##########
>>>
>>> I'm confused why this failed. In fact, no matter what I try to
>>> change, it always fails in the same fashion when I attempt to save the
>>> changes.
>>>
>>> "crm status" currently shows the following:
>>> #######
>>> Last updated: Wed Oct 17 16:36:45 2012
>>> Last change: Tue Oct 16 14:23:18 2012 via cibadmin on farm-ljf1
>>> Stack: openais
>>> Current DC: NONE
>>> 2 Nodes configured, 2 expected votes
>>> 5 Resources configured.
>>>
>>>
>>> OFFLINE: [ farm-ljf0 farm-ljf1 ]
>>> ######
>>>
>>> help?
More information about the Pacemaker
mailing list