[Pacemaker] corosync/openais fails to start

Gianluca Cecchi gianluca.cecchi at gmail.com
Thu May 27 13:20:38 EDT 2010


On Thu, May 27, 2010 at 5:50 PM, Steven Dake <sdake at redhat.com> wrote:

> On 05/27/2010 08:40 AM, Diego Remolina wrote:
>
>> Is there any workaround for this? Perhaps a slightly older version of
>> the rpms? If so where do I find those?
>>
>>
> Corosync 1.2.1 doesn't have this issue apparently.  With corosync 1.2.1,
> please don't use "debug: on" keyword in your config options.  I am not sure
> where Andrew has corosync 1.2.1 rpms available.
>
> The corosync project itself doesn't release rpms.  See our policy on this
> topic:
>
> http://www.corosync.org/doku.php?id=faq:release_binaries
>
> Regards
> -steve
>
>
>
In my case, using pacemaker/corosync from clusterlabs repo on rh el 5.5 32
bit I had:
- both nodes ha1 and ha2 with
[root at ha1 ~]# rpm -qa corosync\* pacemaker\*
pacemaker-1.0.8-6.el5
corosynclib-1.2.1-1.el5
corosync-1.2.1-1.el5
pacemaker-libs-1.0.8-6.el5

- stop of corosync on node ha1
- update (using clusterlabs repo proposed and applied packages for pacemaker
with same version... donna if same bits..)
This takes corosync to 1.2.2
- start of corosync on ha1 and successfull join with the still corosync
1.2.1 one
 May 27 18:59:23 ha1 corosync[5136]:   [MAIN  ] Corosync Cluster Engine
exiting with status -1 at main.c:160.
May 27 19:06:19 ha1 yum: Updated: corosynclib-1.2.2-1.1.el5.i386
May 27 19:06:19 ha1 yum: Updated: pacemaker-libs-1.0.8-6.1.el5.i386
May 27 19:06:19 ha1 yum: Updated: corosync-1.2.2-1.1.el5.i386
May 27 19:06:20 ha1 yum: Updated: pacemaker-1.0.8-6.1.el5.i386
May 27 19:06:20 ha1 yum: Updated: corosynclib-devel-1.2.2-1.1.el5.i386
May 27 19:06:22 ha1 yum: Updated: pacemaker-libs-devel-1.0.8-6.1.el5.i386
May 27 19:06:59 ha1 corosync[7442]:   [MAIN  ] Corosync Cluster Engine
('1.2.2'): started and ready to provide service.
May 27 19:06:59 ha1 corosync[7442]:   [MAIN  ] Corosync built-in features:
nss rdma
May 27 19:06:59 ha1 corosync[7442]:   [MAIN  ] Successfully read main
configuration file '/etc/corosync/corosync.conf'.
May 27 19:06:59 ha1 corosync[7442]:   [TOTEM ] Initializing transport
(UDP/IP).
May 27 19:06:59 ha1 corosync[7442]:   [TOTEM ] Initializing transmit/receive
security: libtomcrypt SOBER128/SHA1HMAC (mode 0).

this implies also start of resources on it (nfsclient and apache in my case)

- move (and unmove to be able to take them again) of resources from ha2 to
the updated node ha1 (nfs-group in my case)
 Resource Group: nfs-group
     lv_drbd0   (ocf::heartbeat:LVM):   Started ha1
     ClusterIP  (ocf::heartbeat:IPaddr2):       Started ha1
     NfsFS      (ocf::heartbeat:Filesystem):    Started ha1
     nfssrv     (ocf::heartbeat:nfsserver):     Started ha1

- stop of corosync 1.2.1 on ha2
- update of pacemaker and corosync on ha2
- startup of corosync on ha2 and correct join to cluster with start of its
resources (nfsclient and apache in my case)
May 27 19:14:42 ha2 corosync[30954]:   [pcmk  ] notice: pcmk_shutdown: cib
confirmed stopped
May 27 19:14:42 ha2 corosync[30954]:   [pcmk  ] notice: stop_child: Sent -15
to stonithd: [30961]
May 27 19:14:42 ha2 stonithd: [30961]: notice: /usr/lib/heartbeat/stonithd
normally quit.
May 27 19:14:42 ha2 corosync[30954]:   [pcmk  ] info: pcmk_ipc_exit: Client
stonithd (conn=0x82aee48, async-conn=0x82aee48) left
May 27 19:14:43 ha2 corosync[30954]:   [pcmk  ] notice: pcmk_shutdown:
stonithd confirmed stopped
May 27 19:14:43 ha2 corosync[30954]:   [pcmk  ] info: update_member: Node
ha2 now has process list: 00000000000000000000000000000002 (2)
May 27 19:14:43 ha2 corosync[30954]:   [pcmk  ] notice: pcmk_shutdown:
Shutdown complete
May 27 19:14:43 ha2 corosync[30954]:   [SERV  ] Service engine unloaded:
Pacemaker Cluster Manager 1.0.8
May 27 19:14:43 ha2 corosync[30954]:   [SERV  ] Service engine unloaded:
corosync extended virtual synchrony service
May 27 19:14:43 ha2 corosync[30954]:   [SERV  ] Service engine unloaded:
corosync configuration service
May 27 19:14:43 ha2 corosync[30954]:   [SERV  ] Service engine unloaded:
corosync cluster closed process group service v1.01
May 27 19:14:43 ha2 corosync[30954]:   [SERV  ] Service engine unloaded:
corosync cluster config database access v1.01
May 27 19:14:43 ha2 corosync[30954]:   [SERV  ] Service engine unloaded:
corosync profile loading service
May 27 19:14:43 ha2 corosync[30954]:   [SERV  ] Service engine unloaded:
corosync cluster quorum service v0.1
May 27 19:14:43 ha2 corosync[30954]:   [MAIN  ] Corosync Cluster Engine
exiting with status -1 at main.c:160.
May 27 19:15:51 ha2 yum: Updated: corosynclib-1.2.2-1.1.el5.i386
May 27 19:15:51 ha2 yum: Updated: pacemaker-libs-1.0.8-6.1.el5.i386
May 27 19:15:52 ha2 yum: Updated: corosync-1.2.2-1.1.el5.i386
May 27 19:15:52 ha2 yum: Updated: pacemaker-1.0.8-6.1.el5.i386
May 27 19:17:00 ha2 corosync[3430]:   [MAIN  ] Corosync Cluster Engine
('1.2.2'): started and ready to provide service.
May 27 19:17:00 ha2 corosync[3430]:   [MAIN  ] Corosync built-in features:
nss rdma
May 27 19:17:00 ha2 corosync[3430]:   [MAIN  ] Successfully read main
configuration file '/etc/corosync/corosync.conf'.
May 27 19:17:00 ha2 corosync[3430]:   [TOTEM ] Initializing transport
(UDP/IP).
May 27 19:17:00 ha2 corosync[3430]:   [TOTEM ] Initializing transmit/receive
security: libtomcrypt SOBER128/SHA1HMAC (mode 0).

So in my case the sw upgrade was successfull with no downtime.

Gianluca
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20100527/4041f0bf/attachment-0001.html>


More information about the Pacemaker mailing list