[Pacemaker] "ERROR: Wrong stack o2cb" when trying to start o2cb service in Pacemaker cluster
David Guyot
david.guyot at europecamions-interactive.com
Wed Jun 20 13:49:27 UTC 2012
Actually, yes, I start DRBD manually, because this is currently a test
configuration which relies on OpenVPN for the communications between
these 2 nodes. I have no order and collocation constraints because I'm
discovering these software and trying to configure them step by step and
make resources work before ordering them (nevertheless, I just tried to
configure DLM/O2CB constraints, but they fail, apparently because they
are relying on O2CB, which causes the problem I wrote you about.) And I
have no OCFS2 mounts because I was on the assumption that OCFS2 wouldn't
mount partitions without O2CB and DLM, which seems to be right :
root at Malastare:/home/david# crm_mon --one-shot -VroA
============
Last updated: Wed Jun 20 15:32:50 2012
Last change: Wed Jun 20 15:28:34 2012 via crm_shadow on Malastare
Stack: openais
Current DC: Vindemiatrix - partition with quorum
Version: 1.1.7-ee0730e13d124c3d58f00016c3376a1de5323cff
2 Nodes configured, 2 expected votes
14 Resources configured.
============
Online: [ Vindemiatrix Malastare ]
Full list of resources:
soapi-fencing-malastare (stonith:external/ovh): Started Vindemiatrix
soapi-fencing-vindemiatrix (stonith:external/ovh): Started Malastare
Master/Slave Set: ms_drbd_ocfs2_pgsql [p_drbd_ocfs2_pgsql]
Masters: [ Malastare Vindemiatrix ]
Master/Slave Set: ms_drbd_ocfs2_backupvi [p_drbd_ocfs2_backupvi]
Masters: [ Malastare Vindemiatrix ]
Master/Slave Set: ms_drbd_ocfs2_svn [p_drbd_ocfs2_svn]
Masters: [ Malastare Vindemiatrix ]
Master/Slave Set: ms_drbd_ocfs2_www [p_drbd_ocfs2_www]
Masters: [ Malastare Vindemiatrix ]
Clone Set: cl_ocfs2mgmt [g_ocfs2mgmt]
Stopped: [ g_ocfs2mgmt:0 g_ocfs2mgmt:1 ]
Node Attributes:
* Node Vindemiatrix:
+ master-p_drbd_ocfs2_backupvi:1 : 10000
+ master-p_drbd_ocfs2_pgsql:1 : 10000
+ master-p_drbd_ocfs2_svn:1 : 10000
+ master-p_drbd_ocfs2_www:1 : 10000
* Node Malastare:
+ master-p_drbd_ocfs2_backupvi:0 : 10000
+ master-p_drbd_ocfs2_pgsql:0 : 10000
+ master-p_drbd_ocfs2_svn:0 : 10000
+ master-p_drbd_ocfs2_www:0 : 10000
Operations:
* Node Vindemiatrix:
p_drbd_ocfs2_pgsql:1: migration-threshold=1000000
+ (4) probe: rc=8 (master)
p_drbd_ocfs2_backupvi:1: migration-threshold=1000000
+ (5) probe: rc=8 (master)
p_drbd_ocfs2_svn:1: migration-threshold=1000000
+ (6) probe: rc=8 (master)
p_drbd_ocfs2_www:1: migration-threshold=1000000
+ (7) probe: rc=8 (master)
soapi-fencing-malastare: migration-threshold=1000000
+ (10) start: rc=0 (ok)
p_o2cb:1: migration-threshold=1000000
+ (9) probe: rc=5 (not installed)
* Node Malastare:
p_drbd_ocfs2_pgsql:0: migration-threshold=1000000
+ (4) probe: rc=8 (master)
p_drbd_ocfs2_backupvi:0: migration-threshold=1000000
+ (5) probe: rc=8 (master)
p_drbd_ocfs2_svn:0: migration-threshold=1000000
+ (6) probe: rc=8 (master)
soapi-fencing-vindemiatrix: migration-threshold=1000000
+ (10) start: rc=0 (ok)
p_drbd_ocfs2_www:0: migration-threshold=1000000
+ (7) probe: rc=8 (master)
p_o2cb:0: migration-threshold=1000000
+ (9) probe: rc=5 (not installed)
Failed actions:
p_o2cb:1_monitor_0 (node=Vindemiatrix, call=9, rc=5,
status=complete): not installed
p_o2cb:0_monitor_0 (node=Malastare, call=9, rc=5, status=complete):
not installed
root at Malastare:/home/david# mount -t ocfs2 /dev/drbd1 /media/ocfs/
mount.ocfs2: Cluster stack specified does not match the one currently
running while trying to join the group
Concerning the notify meta-attribute, I didn't configured it because it
wasn't even referred to in the DRBD official guide (
http://www.drbd.org/users-guide-8.3/s-ocfs2-pacemaker.html), and I don't
know what it does, so, by default, I stupidly followed the official
guide. What does this meta-attribute sets? If you know a better guide,
could you please tell me about, so I can check my config based on this
other guide?
And, last but not least, I run Debian Squeeze 3.2.13-grsec-xxxx-grs-ipv6-64.
Thank you in advance.
Kind regards.
PS: if you find me a bit rude, please accept my apologies; I'm working
on it for weeks following the official DRBD guide and it's frustrating
to ask help as a last resort and to be answered with something which
sounds like "What's this bloody mess !?!" to my tired nerve cells. Once
again, please accept my apologies.
Le 20/06/2012 15:09, Andreas Kurz a écrit :
> On 06/20/2012 02:22 PM, David Guyot wrote:
>> Hello.
>>
>> Oops, an omission.
>>
>> Here comes my Pacemaker config :
>> root at Malastare:/home/david# crm configure show
>> node Malastare
>> node Vindemiatrix
>> primitive p_controld ocf:pacemaker:controld
>> primitive p_drbd_ocfs2_backupvi ocf:linbit:drbd \
>> params drbd_resource="backupvi"
>> primitive p_drbd_ocfs2_pgsql ocf:linbit:drbd \
>> params drbd_resource="postgresql"
>> primitive p_drbd_ocfs2_svn ocf:linbit:drbd \
>> params drbd_resource="svn"
>> primitive p_drbd_ocfs2_www ocf:linbit:drbd \
>> params drbd_resource="www"
>> primitive p_o2cb ocf:pacemaker:o2cb \
>> meta target-role="Started"
>> primitive soapi-fencing-malastare stonith:external/ovh \
>> params reversedns="ns208812.ovh.net"
>> primitive soapi-fencing-vindemiatrix stonith:external/ovh \
>> params reversedns="ns235795.ovh.net"
>> ms ms_drbd_ocfs2_backupvi p_drbd_ocfs2_backupvi \
>> meta master-max="2" clone-max="2"
>> ms ms_drbd_ocfs2_pgsql p_drbd_ocfs2_pgsql \
>> meta master-max="2" clone-max="2"
>> ms ms_drbd_ocfs2_svn p_drbd_ocfs2_svn \
>> meta master-max="2" clone-max="2"
>> ms ms_drbd_ocfs2_www p_drbd_ocfs2_www \
>> meta master-max="2" clone-max="2"
>> location stonith-malastare soapi-fencing-malastare -inf: Malastare
>> location stonith-vindemiatrix soapi-fencing-vindemiatrix -inf: Vindemiatrix
>> property $id="cib-bootstrap-options" \
>> dc-version="1.1.7-ee0730e13d124c3d58f00016c3376a1de5323cff" \
>> cluster-infrastructure="openais" \
>> expected-quorum-votes="2"
>>
> I have absolutely no idea why your configuration can run at all without
> more errors ... do you start the drbd resources manually before the cluster?
>
> You are missing the notify meta-attribute for all your DRBD ms
> resources, you have no order and colocation constraints or groups at all
> and you don't clone controld and o2cb ... and there are no ocfs2 mounts?
>
> Also quite important: what distribution are you using?
>
>> The STONITH resources are custom ones which use my provider SOAP API to
>> electrically reboot fenced nodes.
>>
>> Concerning the web page you talked me about, I tried to insert the
>> referred environment variable, but it did not solved the problem :
> Really have a look at the crm configuration snippet on that page and
> read manuals about setting up DRBD in Pacemaker.
>
> Regards,
> Andreas
>
>> root at Malastare:/home/david# crm_mon --one-shot -VroA
>> ============
>> Last updated: Wed Jun 20 14:14:41 2012
>> Last change: Wed Jun 20 09:22:39 2012 via cibadmin on Malastare
>> Stack: openais
>> Current DC: Vindemiatrix - partition with quorum
>> Version: 1.1.7-ee0730e13d124c3d58f00016c3376a1de5323cff
>> 2 Nodes configured, 2 expected votes
>> 12 Resources configured.
>> ============
>>
>> Online: [ Vindemiatrix Malastare ]
>>
>> Full list of resources:
>>
>> soapi-fencing-malastare (stonith:external/ovh): Stopped
>> p_controld (ocf::pacemaker:controld): Started Malastare
>> p_o2cb (ocf::pacemaker:o2cb): Started Vindemiatrix FAILED
>> soapi-fencing-vindemiatrix (stonith:external/ovh): Stopped
>> Master/Slave Set: ms_drbd_ocfs2_pgsql [p_drbd_ocfs2_pgsql]
>> Masters: [ Vindemiatrix Malastare ]
>> Master/Slave Set: ms_drbd_ocfs2_backupvi [p_drbd_ocfs2_backupvi]
>> Masters: [ Vindemiatrix Malastare ]
>> Master/Slave Set: ms_drbd_ocfs2_svn [p_drbd_ocfs2_svn]
>> Masters: [ Vindemiatrix Malastare ]
>> Master/Slave Set: ms_drbd_ocfs2_www [p_drbd_ocfs2_www]
>> Masters: [ Vindemiatrix Malastare ]
>>
>> Node Attributes:
>> * Node Vindemiatrix:
>> + master-p_drbd_ocfs2_backupvi:0 : 10000
>> + master-p_drbd_ocfs2_pgsql:0 : 10000
>> + master-p_drbd_ocfs2_svn:0 : 10000
>> + master-p_drbd_ocfs2_www:0 : 10000
>> * Node Malastare:
>> + master-p_drbd_ocfs2_backupvi:1 : 10000
>> + master-p_drbd_ocfs2_pgsql:1 : 10000
>> + master-p_drbd_ocfs2_svn:1 : 10000
>> + master-p_drbd_ocfs2_www:1 : 10000
>>
>> Operations:
>> * Node Vindemiatrix:
>> p_o2cb: migration-threshold=1000000 fail-count=1000000
>> + (11) start: rc=5 (not installed)
>> p_drbd_ocfs2_pgsql:0: migration-threshold=1000000
>> + (6) probe: rc=8 (master)
>> p_drbd_ocfs2_backupvi:0: migration-threshold=1000000
>> + (7) probe: rc=8 (master)
>> p_drbd_ocfs2_svn:0: migration-threshold=1000000
>> + (8) probe: rc=8 (master)
>> p_drbd_ocfs2_www:0: migration-threshold=1000000
>> + (9) probe: rc=8 (master)
>> * Node Malastare:
>> p_controld: migration-threshold=1000000
>> + (10) start: rc=0 (ok)
>> p_o2cb: migration-threshold=1000000
>> + (4) probe: rc=5 (not installed)
>> p_drbd_ocfs2_pgsql:1: migration-threshold=1000000
>> + (6) probe: rc=8 (master)
>> p_drbd_ocfs2_backupvi:1: migration-threshold=1000000
>> + (7) probe: rc=8 (master)
>> p_drbd_ocfs2_svn:1: migration-threshold=1000000
>> + (8) probe: rc=8 (master)
>> p_drbd_ocfs2_www:1: migration-threshold=1000000
>> + (9) probe: rc=8 (master)
>>
>> Failed actions:
>> p_o2cb_start_0 (node=Vindemiatrix, call=11, rc=5, status=complete):
>> not installed
>> p_o2cb_monitor_0 (node=Malastare, call=4, rc=5, status=complete):
>> not installed
>>
>> Thank you in advance for your help!
>>
>> Kind regards.
>>
>> Le 20/06/2012 14:02, Andreas Kurz a écrit :
>>> On 06/20/2012 01:43 PM, David Guyot wrote:
>>>> Hello, everybody.
>>>>
>>>> I'm trying to configure Pacemaker for using DRBD + OCFS2 storage, but
>>>> I'm stuck with DRBD and controld up and o2cb doggedly displaying "not
>>>> installed" errors. To do this, I followed the DRBD guide (
>>>> http://www.drbd.org/users-guide-8.3/ch-ocfs2.html), with the difference
>>>> that I was forced to disable DRBD fencing because it was interfering
>>>> with Pacemaker fencing and stopping each nodes as often as it could.
>>> Unfortunately you didn't share your Pacemaker configuration but you
>>> definitely must not start any ocfs2 init script but let all be managed
>>> by the cluster-manager.
>>>
>>> Here is a brief setup description, also mentioning the tune.ocfs2 when
>>> the Pacemaker stack is running:
>>>
>>> http://www.hastexo.com/resources/hints-and-kinks/ocfs2-pacemaker-debianubuntu
>>>
>>> And once this is running as expected you really want to reactivate the
>>> DRBD fencing configuration.
>>>
>>> Regards,
>>> Andreas
>>>
>>>
>>>
>>> _______________________________________________
>>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>
>>> Project Home: http://www.clusterlabs.org
>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>> Bugs: http://bugs.clusterlabs.org
>>
>>
>>
>> _______________________________________________
>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>
>> Project Home: http://www.clusterlabs.org
>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> Bugs: http://bugs.clusterlabs.org
>>
>
>
>
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
-------------- next part --------------
A non-text attachment was scrubbed...
Name: signature.asc
Type: application/pgp-signature
Size: 554 bytes
Desc: OpenPGP digital signature
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20120620/75393b4e/attachment-0004.sig>
More information about the Pacemaker
mailing list