[ClusterLabs] Problem in starting cman
Vijay Partha
vijaysarathy94 at gmail.com
Mon Aug 3 12:37:51 UTC 2015
I have formed a cluster by making use of 2 nodes.
On Mon, Aug 3, 2015 at 6:07 PM, Vijay Partha <vijaysarathy94 at gmail.com>
wrote:
> service cman start
> Starting cluster:
> Checking if cluster has been disabled at boot... [ OK ]
> Checking Network Manager... [ OK ]
> Global setup... [ OK ]
> Loading kernel modules... [ OK ]
> Mounting configfs... [ OK ]
> Starting cman... [ OK ]
> Waiting for quorum... [ OK ]
> Starting fenced... [ OK ]
> Starting dlm_controld... [ OK ]
> Tuning DLM kernel config... [ OK ]
> Starting gfs_controld... [ OK ]
> Unfencing self... [ OK ]
> Joining fence domain...
>
> It doesnt go beyond this.
>
> This is my pacemaker log
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info:
> crm_ipc_connect: Could not establish pacemakerd connection: Connection
> refused (111)
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info:
> config_find_next: Processing additional service options...
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info:
> get_config_opt: Found 'corosync_quorum' for option: name
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info:
> config_find_next: Processing additional service options...
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info:
> get_config_opt: Found 'corosync_cman' for option: name
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info:
> config_find_next: Processing additional service options...
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info:
> get_config_opt: Found 'openais_ckpt' for option: name
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info:
> config_find_next: No additional configuration supplied for: service
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info:
> config_find_next: Processing additional quorum options...
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info:
> get_config_opt: Found 'quorum_cman' for option: provider
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info:
> get_cluster_type: Detected an active 'cman' cluster
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info:
> mcp_read_config: Reading configure for stack: cman
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info:
> config_find_next: Processing additional logging options...
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info:
> get_config_opt: Defaulting to 'off' for option: debug
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info:
> get_config_opt: Found 'yes' for option: to_logfile
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info:
> get_config_opt: Found '/var/log/cluster/corosync.log' for option:
> logfile
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: notice:
> crm_add_logfile: Additional logging available in
> /var/log/cluster/corosync.log
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info:
> get_config_opt: Found 'yes' for option: to_syslog
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info:
> get_config_opt: Found 'local4' for option: syslog_facility
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: notice: main:
> Starting Pacemaker 1.1.11 (Build: 97629de): generated-manpages
> agent-manpages ascii-docs ncurses libqb-logging libqb-ipc nagios
> corosync-plugin cman acls
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info: main:
> Maximum core file size is: 18446744073709551615
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd: info:
> qb_ipcs_us_publish: server name: pacemakerd
> Jul 29 13:46:17 [22563] vmx-occ-004 pacemakerd: error:
> cluster_connect_cpg: Could not join the CPG group 'pacemakerd': 6
> Jul 29 13:46:17 [22563] vmx-occ-004 pacemakerd: error: main:
> Couldn't connect to Corosync's CPG service
> Jul 29 13:46:17 [22563] vmx-occ-004 pacemakerd: info:
> crm_xml_cleanup: Cleaning up memory from libxml2
>
> this is my message log
>
> Aug 3 14:35:31 vmx-occ-004 dlm_controld[2383]: daemon cpg_join error
> retrying
> Aug 3 14:35:33 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() for [my
> cluster] failed to contact node 10.61.40.194
> Aug 3 14:35:33 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() got no
> answer from any [my cluster] datasource
> Aug 3 14:35:34 vmx-occ-004 fenced[2359]: daemon cpg_join error retrying
> Aug 3 14:35:37 vmx-occ-004 /usr/sbin/gmond[4590]: Error creating
> multicast server mcast_join=10.61.40.194 port=8649 mcast_if=NULL
> family='inet4'. Will try again...#012
> Aug 3 14:35:38 vmx-occ-004 gfs_controld[2458]: daemon cpg_join error
> retrying
> Aug 3 14:35:41 vmx-occ-004 dlm_controld[2383]: daemon cpg_join error
> retrying
> Aug 3 14:35:44 vmx-occ-004 fenced[2359]: daemon cpg_join error retrying
> Aug 3 14:35:48 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() for [my
> cluster] failed to contact node 10.61.40.194
> Aug 3 14:35:48 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() got no
> answer from any [my cluster] datasource
> Aug 3 14:35:48 vmx-occ-004 gfs_controld[2458]: daemon cpg_join error
> retrying
> Aug 3 14:35:51 vmx-occ-004 dlm_controld[2383]: daemon cpg_join error
> retrying
> Aug 3 14:35:54 vmx-occ-004 fenced[2359]: daemon cpg_join error retrying
> Aug 3 14:35:58 vmx-occ-004 gfs_controld[2458]: daemon cpg_join error
> retrying
> Aug 3 14:36:01 vmx-occ-004 dlm_controld[2383]: daemon cpg_join error
> retrying
> Aug 3 14:36:03 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() for [my
> cluster] failed to contact node 10.61.40.194
> Aug 3 14:36:03 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() got no
> answer from any [my cluster] datasource
> Aug 3 14:36:04 vmx-occ-004 fenced[2359]: daemon cpg_join error retrying
> Aug 3 14:36:08 vmx-occ-004 gfs_controld[2458]: daemon cpg_join error
> retrying
> Aug 3 14:36:11 vmx-occ-004 dlm_controld[2383]: daemon cpg_join error
> retrying
> Aug 3 14:36:14 vmx-occ-004 fenced[2359]: daemon cpg_join error retrying
> Aug 3 14:36:18 vmx-occ-004 gfs_controld[2458]: daemon cpg_join error
> retrying
> Aug 3 14:36:18 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() for [my
> cluster] failed to contact node 10.61.40.194
> Aug 3 14:36:18 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() got no
> answer from any [my cluster] datasource
> Aug 3 14:36:21 vmx-occ-004 dlm_controld[2383]: daemon cpg_join error
> retrying
> Aug 3 14:36:24 vmx-occ-004 fenced[2359]: daemon cpg_join error retrying
>
>
> On Mon, Aug 3, 2015 at 6:02 PM, emmanuel segura <emi2fast at gmail.com>
> wrote:
>
>> Sorry, but I think is more easy to help you, If you provide more
>> information about your problem.
>>
>> 2015-08-03 14:14 GMT+02:00 Vijay Partha <vijaysarathy94 at gmail.com>:
>> > Hi
>> >
>> > When i start cman it hangs in joining fence domain.
>> >
>> > this is my message log.
>> >
>> >
>> > Aug 3 14:12:16 vmx-occ-005 dlm_controld[2112]: daemon cpg_join error
>> > retrying
>> > Aug 3 14:12:19 vmx-occ-005 gfs_controld[2191]: daemon cpg_join error
>> > retrying
>> > Aug 3 14:12:24 vmx-occ-005 fenced[2098]: daemon cpg_join error retrying
>> > Aug 3 14:12:27 vmx-occ-005 dlm_controld[2112]: daemon cpg_join error
>> > retrying
>> > Aug 3 14:12:29 vmx-occ-005 gfs_controld[2191]: daemon cpg_join error
>> > retrying
>> > Aug 3 14:12:34 vmx-occ-005 fenced[2098]: daemon cpg_join error retrying
>> > Aug 3 14:12:37 vmx-occ-005 dlm_controld[2112]: daemon cpg_join error
>> > retrying
>> > Aug 3 14:12:39 vmx-occ-005 gfs_controld[2191]: daemon cpg_join error
>> > retrying
>> > Aug 3 14:12:44 vmx-occ-005 fenced[2098]: daemon cpg_join error retrying
>> > Aug 3 14:12:47 vmx-occ-005 dlm_controld[2112]: daemon cpg_join error
>> > retrying
>> > Aug 3 14:12:49 vmx-occ-005 gfs_controld[2191]: daemon cpg_join error
>> > retrying
>> > Aug 3 14:12:54 vmx-occ-005 fenced[2098]: daemon cpg_join error retrying
>> >
>> > How to solve this issue?
>> > --
>> > With Regards
>> > P.Vijay
>> >
>> > _______________________________________________
>> > Users mailing list: Users at clusterlabs.org
>> > http://clusterlabs.org/mailman/listinfo/users
>> >
>> > Project Home: http://www.clusterlabs.org
>> > Getting started:
>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> > Bugs: http://bugs.clusterlabs.org
>> >
>>
>>
>>
>> --
>> .~.
>> /V\
>> // \\
>> /( )\
>> ^`~'^
>>
>> _______________________________________________
>> Users mailing list: Users at clusterlabs.org
>> http://clusterlabs.org/mailman/listinfo/users
>>
>> Project Home: http://www.clusterlabs.org
>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> Bugs: http://bugs.clusterlabs.org
>>
>
>
>
> --
> With Regards
> P.Vijay
>
--
With Regards
P.Vijay
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/users/attachments/20150803/9ebad409/attachment.htm>
More information about the Users
mailing list