[Pacemaker] 'crm configure edit' failed with "Timer expired"
Lonni J Friedman
netllama at gmail.com
Thu Oct 18 17:25:25 UTC 2012
Both nodes can ssh to each other, selinux is disabled, and there are
currently no iptables rules in force. So I'm not sure why the systems
can't communicate.
I looked at /var/log/cluster/corosync.log but didn't see any obvious
problems. Would you like to see the logs from both nodes?
What else can I do to investigate & debug this?
thanks
On Wed, Oct 17, 2012 at 8:24 PM, Andrew Beekhof <andrew at beekhof.net> wrote:
> This is your resource:
> Current DC: NONE
>
> All the "crm configure" commands try to talk to the DC, and you don't have one.
> Normally one would be elected quite quickly, you may have a
> network/filewall issue.
>
> On Thu, Oct 18, 2012 at 10:37 AM, Lonni J Friedman <netllama at gmail.com> wrote:
>> I'm running Fedora17, with pacemaker-1.18. I just tried to make a
>> configuration change with crmsh, and it failed as follows:
>> ##########
>> # crm configure edit
>> Call cib_replace failed (-62): Timer expired
>> <null>
>> ERROR: could not replace cib
>> INFO: offending xml: <configuration>
>> <crm_config>
>> <cluster_property_set id="cib-bootstrap-options">
>> <nvpair id="cib-bootstrap-options-dc-version" name="dc-version"
>> value="1.1.8-2.fc16-394e906"/>
>> <nvpair id="cib-bootstrap-options-cluster-infrastructure"
>> name="cluster-infrastructure" value="openais"/>
>> <nvpair id="cib-bootstrap-options-expected-quorum-votes"
>> name="expected-quorum-votes" value="2"/>
>> <nvpair id="cib-bootstrap-options-stonith-enabled"
>> name="stonith-enabled" value="false"/>
>> <nvpair id="cib-bootstrap-options-no-quorum-policy"
>> name="no-quorum-policy" value="ignore"/>
>> </cluster_property_set>
>> </crm_config>
>> <nodes>
>> <node id="farm-ljf0" uname="farm-ljf0">
>> <instance_attributes id="nodes-farm-ljf0">
>> <nvpair id="nodes-farm-ljf0-standby" name="standby" value="off"/>
>> </instance_attributes>
>> </node>
>> <node id="farm-ljf1" uname="farm-ljf1"/>
>> </nodes>
>> <resources>
>> <master id="FS0_Clone">
>> <meta_attributes id="FS0_Clone-meta_attributes">
>> <nvpair id="FS0_Clone-meta_attributes-master-max"
>> name="master-max" value="1"/>
>> <nvpair id="FS0_Clone-meta_attributes-master-node-max"
>> name="master-node-max" value="1"/>
>> <nvpair id="FS0_Clone-meta_attributes-clone-max" name="clone-max"
>> value="2"/>
>> <nvpair id="FS0_Clone-meta_attributes-clone-node-max"
>> name="clone-node-max" value="1"/>
>> <nvpair id="FS0_Clone-meta_attributes-notify" name="notify" value="true"/>
>> </meta_attributes>
>> <primitive class="ocf" id="FS0" provider="linbit" type="drbd">
>> <instance_attributes id="FS0-instance_attributes">
>> <nvpair id="FS0-instance_attributes-drbd_resource"
>> name="drbd_resource" value="r0"/>
>> </instance_attributes>
>> <operations>
>> <op id="FS0-monitor-10s" interval="10s" name="monitor" role="Master"/>
>> <op id="FS0-monitor-30s" interval="30s" name="monitor" role="Slave"/>
>> </operations>
>> </primitive>
>> </master>
>> <group id="g_services">
>> <primitive class="ocf" id="ClusterIP" provider="heartbeat" type="IPaddr2">
>> <instance_attributes id="ClusterIP-instance_attributes">
>> <nvpair id="ClusterIP-instance_attributes-ip" name="ip"
>> value="10.31.99.8"/>
>> <nvpair id="ClusterIP-instance_attributes-cidr_netmask"
>> name="cidr_netmask" value="22"/>
>> <nvpair id="ClusterIP-instance_attributes-nic" name="nic" value="eth1"/>
>> </instance_attributes>
>> <operations>
>> <op id="ClusterIP-monitor-10s" interval="10s" name="monitor"/>
>> </operations>
>> <meta_attributes id="ClusterIP-meta_attributes">
>> <nvpair id="ClusterIP-meta_attributes-target-role"
>> name="target-role" value="Started"/>
>> </meta_attributes>
>> </primitive>
>> <primitive class="ocf" id="FS0_drbd" provider="heartbeat" type="Filesystem">
>> <instance_attributes id="FS0_drbd-instance_attributes">
>> <nvpair id="FS0_drbd-instance_attributes-device" name="device"
>> value="/dev/drbd0"/>
>> <nvpair id="FS0_drbd-instance_attributes-directory"
>> name="directory" value="/mnt/sdb1"/>
>> <nvpair id="FS0_drbd-instance_attributes-fstype" name="fstype"
>> value="xfs"/>
>> </instance_attributes>
>> <meta_attributes id="FS0_drbd-meta_attributes">
>> <nvpair id="FS0_drbd-meta_attributes-target-role"
>> name="target-role" value="Started"/>
>> </meta_attributes>
>> </primitive>
>> <primitive class="systemd" id="FS0_nfs" type="nfs-server">
>> <operations>
>> <op id="FS0_nfs-monitor-10s" interval="10s" name="monitor"/>
>> </operations>
>> <meta_attributes id="FS0_nfs-meta_attributes">
>> <nvpair id="FS0_nfs-meta_attributes-target-role"
>> name="target-role" value="Started"/>
>> </meta_attributes>
>> </primitive>
>> </group>
>> </resources>
>> <constraints>
>> <rsc_colocation id="fs0_on_drbd" rsc="g_services" score="INFINITY"
>> with-rsc="FS0_Clone" with-rsc-role="Master"/>
>> <rsc_order first="FS0_Clone" first-action="promote"
>> id="FS0_drbd-after-FS0" score="INFINITY" then="g_services"
>> then-action="start"/>
>> </constraints>
>> </configuration>
>> ##########
>>
>> I'm confused why this failed. In fact, no matter what I try to
>> change, it always fails in the same fashion when I attempt to save the
>> changes.
>>
>> "crm status" currently shows the following:
>> #######
>> Last updated: Wed Oct 17 16:36:45 2012
>> Last change: Tue Oct 16 14:23:18 2012 via cibadmin on farm-ljf1
>> Stack: openais
>> Current DC: NONE
>> 2 Nodes configured, 2 expected votes
>> 5 Resources configured.
>>
>>
>> OFFLINE: [ farm-ljf0 farm-ljf1 ]
>> ######
>>
>> help?
More information about the Pacemaker
mailing list