[Pacemaker] Infinite fail-count and migration-threshold after node fail-back

Dan Frincu dfrincu at streamwide.ro
Sat Oct 9 18:49:33 UTC 2010


Hi,

Since it was brought to my attention that I should upgrade from 
openais-0.80 to a more recent version of corosync, I've done just that, 
however I'm experiencing a strange behavior on the cluster.

The same setup was used with the below packages:

# rpm -qa | grep -i "(openais|cluster|heartbeat|pacemaker|resource)"
openais-0.80.5-15.2
cluster-glue-1.0-12.2
pacemaker-1.0.5-4.2
cluster-glue-libs-1.0-12.2
resource-agents-1.0-31.5
pacemaker-libs-1.0.5-4.2
pacemaker-mgmt-1.99.2-7.2
libopenais2-0.80.5-15.2
heartbeat-3.0.0-33.3
pacemaker-mgmt-client-1.99.2-7.2

Now I've migrated to the most recent stable packages I could find (on 
the clusterlabs.org website) for RHEL5:

# rpm -qa | grep -i "(openais|cluster|heartbeat|pacemaker|resource)"
cluster-glue-1.0.6-1.6.el5
pacemaker-libs-1.0.9.1-1.el5
pacemaker-1.0.9.1-1.el5
heartbeat-libs-3.0.3-2.el5
heartbeat-3.0.3-2.el5
openaislib-1.1.3-1.6.el5
resource-agents-1.0.3-2.el5
cluster-glue-libs-1.0.6-1.6.el5
openais-1.1.3-1.6.el5

Expected behavior:
- all the resources the in group should go (based on location 
preference) to bench1
- if bench1 goes down, resources migrate to bench2
- if bench1 comes back up, resources stay on bench2, unless manually 
told otherwise.

On the previous incantation, this worked, by using the new packages, not 
so much. Now if bench1 goes down (crm node standby `uname -n`), failover 
occurs, but when bench1 comes backup up, resources migrate back, even if 
default-resource-stickiness is set, and more than that, 2 drbd block 
devices reach infinite metrics, most notably because they try to promote 
the resources to a Master state on bench1, but fail to do so due to the 
resource being held open (by some process, I could not identify it).

Strangely enough, the resources (drbd) fail to be promoted to a Master 
status on bench1, so they fail back to bench2, where they are mounted 
(functional), but crm_mon shows:

Migration summary:
* Node bench2.streamwide.ro:
   drbd_mysql:1: migration-threshold=1000000 fail-count=1000000
   drbd_home:1: migration-threshold=1000000 fail-count=1000000
* Node bench1.streamwide.ro:

.... infinite metrics on bench2, while the drbd resources are available

version: 8.3.2 (api:88/proto:86-90)
GIT-hash: dd7985327f146f33b86d4bff5ca8c94234ce840e build by 
mockbuild at v20z-x86-64.home.local, 2009-08-29 14:07:55
 0: cs:Connected ro:Primary/Secondary ds:UpToDate/UpToDate C r----
    ns:1632 nr:1864 dw:3512 dr:3933 al:11 bm:19 lo:0 pe:0 ua:0 ap:0 ep:1 
wo:b oos:0
 1: cs:Connected ro:Primary/Secondary ds:UpToDate/UpToDate C r----
    ns:4 nr:24 dw:28 dr:25 al:1 bm:1 lo:0 pe:0 ua:0 ap:0 ep:1 wo:b oos:0
 2: cs:Connected ro:Primary/Secondary ds:UpToDate/UpToDate C r----
    ns:4 nr:24 dw:28 dr:85 al:1 bm:1 lo:0 pe:0 ua:0 ap:0 ep:1 wo:b oos:0

and mounted

/dev/drbd1 on /home type ext3 (rw,noatime,nodiratime)
/dev/drbd0 on /mysql type ext3 (rw,noatime,nodiratime)
/dev/drbd2 on /storage type ext3 (rw,noatime,nodiratime)

Attached is the hb_report.

Thank you in advance.

Best regards

-- 
Dan FRINCU
Systems Engineer
CCNA, RHCE
Streamwide Romania

-------------- next part --------------
A non-text attachment was scrubbed...
Name: hb_report-Sat-09-Oct-2010.tar.bz2
Type: application/octet-stream
Size: 122777 bytes
Desc: not available
URL: <http://oss.clusterlabs.org/pipermail/pacemaker/attachments/20101009/bba3f22e/attachment-0001.obj>


More information about the Pacemaker mailing list