[ClusterLabs] Problem in starting cman

Vijay Partha vijaysarathy94 at gmail.com
Mon Aug 3 12:37:51 UTC 2015


I have formed a cluster by making use of 2 nodes.

On Mon, Aug 3, 2015 at 6:07 PM, Vijay Partha <vijaysarathy94 at gmail.com>
wrote:

> service cman start
> Starting cluster:
>    Checking if cluster has been disabled at boot...        [  OK  ]
>    Checking Network Manager...                             [  OK  ]
>    Global setup...                                         [  OK  ]
>    Loading kernel modules...                               [  OK  ]
>    Mounting configfs...                                    [  OK  ]
>    Starting cman...                                        [  OK  ]
>    Waiting for quorum...                                   [  OK  ]
>    Starting fenced...                                      [  OK  ]
>    Starting dlm_controld...                                [  OK  ]
>    Tuning DLM kernel config...                             [  OK  ]
>    Starting gfs_controld...                                [  OK  ]
>    Unfencing self...                                       [  OK  ]
>    Joining fence domain...
>
> It doesnt go beyond this.
>
> This is my pacemaker log
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd:     info:
> crm_ipc_connect:      Could not establish pacemakerd connection: Connection
> refused (111)
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd:     info:
> config_find_next:     Processing additional service options...
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd:     info:
> get_config_opt:       Found 'corosync_quorum' for option: name
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd:     info:
> config_find_next:     Processing additional service options...
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd:     info:
> get_config_opt:       Found 'corosync_cman' for option: name
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd:     info:
> config_find_next:     Processing additional service options...
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd:     info:
> get_config_opt:       Found 'openais_ckpt' for option: name
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd:     info:
> config_find_next:     No additional configuration supplied for: service
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd:     info:
> config_find_next:     Processing additional quorum options...
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd:     info:
> get_config_opt:       Found 'quorum_cman' for option: provider
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd:     info:
> get_cluster_type:     Detected an active 'cman' cluster
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd:     info:
> mcp_read_config:      Reading configure for stack: cman
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd:     info:
> config_find_next:     Processing additional logging options...
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd:     info:
> get_config_opt:       Defaulting to 'off' for option: debug
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd:     info:
> get_config_opt:       Found 'yes' for option: to_logfile
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd:     info:
> get_config_opt:       Found '/var/log/cluster/corosync.log' for option:
> logfile
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd:   notice:
> crm_add_logfile:      Additional logging available in
> /var/log/cluster/corosync.log
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd:     info:
> get_config_opt:       Found 'yes' for option: to_syslog
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd:     info:
> get_config_opt:       Found 'local4' for option: syslog_facility
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd:   notice: main:
> Starting Pacemaker 1.1.11 (Build: 97629de):  generated-manpages
> agent-manpages ascii-docs ncurses libqb-logging libqb-ipc nagios
> corosync-plugin cman acls
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd:     info: main:
> Maximum core file size is: 18446744073709551615
> Jul 29 13:38:32 [22563] vmx-occ-004 pacemakerd:     info:
> qb_ipcs_us_publish:   server name: pacemakerd
> Jul 29 13:46:17 [22563] vmx-occ-004 pacemakerd:    error:
> cluster_connect_cpg:  Could not join the CPG group 'pacemakerd': 6
> Jul 29 13:46:17 [22563] vmx-occ-004 pacemakerd:    error: main:
> Couldn't connect to Corosync's CPG service
> Jul 29 13:46:17 [22563] vmx-occ-004 pacemakerd:     info:
> crm_xml_cleanup:      Cleaning up memory from libxml2
>
> this is my message log
>
> Aug  3 14:35:31 vmx-occ-004 dlm_controld[2383]: daemon cpg_join error
> retrying
> Aug  3 14:35:33 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() for [my
> cluster] failed to contact node 10.61.40.194
> Aug  3 14:35:33 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() got no
> answer from any [my cluster] datasource
> Aug  3 14:35:34 vmx-occ-004 fenced[2359]: daemon cpg_join error retrying
> Aug  3 14:35:37 vmx-occ-004 /usr/sbin/gmond[4590]: Error creating
> multicast server mcast_join=10.61.40.194 port=8649 mcast_if=NULL
> family='inet4'.  Will try again...#012
> Aug  3 14:35:38 vmx-occ-004 gfs_controld[2458]: daemon cpg_join error
> retrying
> Aug  3 14:35:41 vmx-occ-004 dlm_controld[2383]: daemon cpg_join error
> retrying
> Aug  3 14:35:44 vmx-occ-004 fenced[2359]: daemon cpg_join error retrying
> Aug  3 14:35:48 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() for [my
> cluster] failed to contact node 10.61.40.194
> Aug  3 14:35:48 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() got no
> answer from any [my cluster] datasource
> Aug  3 14:35:48 vmx-occ-004 gfs_controld[2458]: daemon cpg_join error
> retrying
> Aug  3 14:35:51 vmx-occ-004 dlm_controld[2383]: daemon cpg_join error
> retrying
> Aug  3 14:35:54 vmx-occ-004 fenced[2359]: daemon cpg_join error retrying
> Aug  3 14:35:58 vmx-occ-004 gfs_controld[2458]: daemon cpg_join error
> retrying
> Aug  3 14:36:01 vmx-occ-004 dlm_controld[2383]: daemon cpg_join error
> retrying
> Aug  3 14:36:03 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() for [my
> cluster] failed to contact node 10.61.40.194
> Aug  3 14:36:03 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() got no
> answer from any [my cluster] datasource
> Aug  3 14:36:04 vmx-occ-004 fenced[2359]: daemon cpg_join error retrying
> Aug  3 14:36:08 vmx-occ-004 gfs_controld[2458]: daemon cpg_join error
> retrying
> Aug  3 14:36:11 vmx-occ-004 dlm_controld[2383]: daemon cpg_join error
> retrying
> Aug  3 14:36:14 vmx-occ-004 fenced[2359]: daemon cpg_join error retrying
> Aug  3 14:36:18 vmx-occ-004 gfs_controld[2458]: daemon cpg_join error
> retrying
> Aug  3 14:36:18 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() for [my
> cluster] failed to contact node 10.61.40.194
> Aug  3 14:36:18 vmx-occ-004 /usr/sbin/gmetad[4295]: data_thread() got no
> answer from any [my cluster] datasource
> Aug  3 14:36:21 vmx-occ-004 dlm_controld[2383]: daemon cpg_join error
> retrying
> Aug  3 14:36:24 vmx-occ-004 fenced[2359]: daemon cpg_join error retrying
>
>
> On Mon, Aug 3, 2015 at 6:02 PM, emmanuel segura <emi2fast at gmail.com>
> wrote:
>
>> Sorry, but I think is more easy to help you, If you provide more
>> information about your problem.
>>
>> 2015-08-03 14:14 GMT+02:00 Vijay Partha <vijaysarathy94 at gmail.com>:
>> > Hi
>> >
>> > When i start cman it hangs in joining fence domain.
>> >
>> > this is my message log.
>> >
>> >
>> > Aug  3 14:12:16 vmx-occ-005 dlm_controld[2112]: daemon cpg_join error
>> > retrying
>> > Aug  3 14:12:19 vmx-occ-005 gfs_controld[2191]: daemon cpg_join error
>> > retrying
>> > Aug  3 14:12:24 vmx-occ-005 fenced[2098]: daemon cpg_join error retrying
>> > Aug  3 14:12:27 vmx-occ-005 dlm_controld[2112]: daemon cpg_join error
>> > retrying
>> > Aug  3 14:12:29 vmx-occ-005 gfs_controld[2191]: daemon cpg_join error
>> > retrying
>> > Aug  3 14:12:34 vmx-occ-005 fenced[2098]: daemon cpg_join error retrying
>> > Aug  3 14:12:37 vmx-occ-005 dlm_controld[2112]: daemon cpg_join error
>> > retrying
>> > Aug  3 14:12:39 vmx-occ-005 gfs_controld[2191]: daemon cpg_join error
>> > retrying
>> > Aug  3 14:12:44 vmx-occ-005 fenced[2098]: daemon cpg_join error retrying
>> > Aug  3 14:12:47 vmx-occ-005 dlm_controld[2112]: daemon cpg_join error
>> > retrying
>> > Aug  3 14:12:49 vmx-occ-005 gfs_controld[2191]: daemon cpg_join error
>> > retrying
>> > Aug  3 14:12:54 vmx-occ-005 fenced[2098]: daemon cpg_join error retrying
>> >
>> > How to solve this issue?
>> > --
>> > With Regards
>> > P.Vijay
>> >
>> > _______________________________________________
>> > Users mailing list: Users at clusterlabs.org
>> > http://clusterlabs.org/mailman/listinfo/users
>> >
>> > Project Home: http://www.clusterlabs.org
>> > Getting started:
>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> > Bugs: http://bugs.clusterlabs.org
>> >
>>
>>
>>
>> --
>>   .~.
>>   /V\
>>  //  \\
>> /(   )\
>> ^`~'^
>>
>> _______________________________________________
>> Users mailing list: Users at clusterlabs.org
>> http://clusterlabs.org/mailman/listinfo/users
>>
>> Project Home: http://www.clusterlabs.org
>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> Bugs: http://bugs.clusterlabs.org
>>
>
>
>
> --
> With Regards
> P.Vijay
>



-- 
With Regards
P.Vijay
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/users/attachments/20150803/9ebad409/attachment.htm>


More information about the Users mailing list