[Pacemaker] Problems with jboss on pacemaker

Benjamin Knoth knoth at mpdl.mpg.de
Wed May 4 09:44:02 EDT 2011



Am 04.05.2011 13:18, schrieb Benjamin Knoth:
> Hi,
> 
> Am 04.05.2011 12:18, schrieb Dejan Muhamedagic:
>> Hi,
>>
>> On Wed, May 04, 2011 at 10:37:40AM +0200, Benjamin Knoth wrote:
>>
>>
>> Am 04.05.2011 09:42, schrieb Florian Haas:
>>>>> On 05/04/2011 09:31 AM, Benjamin Knoth wrote:
>>>>>> Hi Florian,
>>>>>> i test  it with ocf, but i couldn't run.
>>>>>
>>>>> Well that's really helpful information. Logs? Error messages? Anything?
>>
>> Logs
>>
>> May  4 09:55:10 vm36 lrmd: [19214]: WARN: p_jboss_ocf:start process (PID
>> 27702) timed out (try 1).  Killing with signal SIGTERM (15).
>>
>>> You need to set/increase the timeout for the start operation to
>>> match the maximum expected start time. Take a look at "crm ra
>>> info jboss" for minimum values.
>>
>> May  4 09:55:10 vm36 attrd: [19215]: info: find_hash_entry: Creating
>> hash entry for fail-count-p_jboss_ocf
>> May  4 09:55:10 vm36 lrmd: [19214]: WARN: operation start[342] on
>> ocf::jboss::p_jboss_ocf for client 19217, its parameters:
>> CRM_meta_name=[start] crm_feature_set=[3.0.1]
>> java_home=[/usr/lib64/jvm/java] CRM_meta_timeout=[240000] jboss_sto
>> p_timeout=[30] jboss_home=[/usr/share/jboss] jboss_pstring=[java
>> -Dprogram.name=run.sh] : pid [27702] timed out
>> May  4 09:55:10 vm36 attrd: [19215]: info: attrd_trigger_update: Sending
>> flush op to all hosts for: fail-count-p_jboss_ocf (INFINITY)
>> May  4 09:55:10 vm36 crmd: [19217]: WARN: status_from_rc: Action 64
>> (p_jboss_ocf_start_0) on vm36 failed (target: 0 vs. rc: -2): Error
>> May  4 09:55:10 vm36 lrmd: [19214]: info: rsc:p_jboss_ocf:346: stop
>> May  4 09:55:10 vm36 attrd: [19215]: info: attrd_perform_update: Sent
>> update 2294: fail-count-p_jboss_ocf=INFINITY
>> May  4 09:55:10 vm36 pengine: [19216]: notice: unpack_rsc_op: Hard error
>> - p_jboss_lsb_monitor_0 failed with rc=5: Preventing p_jboss_lsb from
>> re-starting on vm36
>> May  4 09:55:10 vm36 crmd: [19217]: WARN: update_failcount: Updating
>> failcount for p_jboss_ocf on vm36 after failed start: rc=-2
>> (update=INFINITY, time=1304495710)
>> May  4 09:55:10 vm36 attrd: [19215]: info: find_hash_entry: Creating
>> hash entry for last-failure-p_jboss_ocf
>> May  4 09:55:10 vm36 pengine: [19216]: notice: unpack_rsc_op: Operation
>> p_jboss_cs_monitor_0 found resource p_jboss_cs active on vm36
>> May  4 09:55:10 vm36 crmd: [19217]: info: abort_transition_graph:
>> match_graph_event:272 - Triggered transition abort (complete=0,
>> tag=lrm_rsc_op, id=p_jboss_ocf_start_0,
>> magic=2:-2;64:1375:0:fc16910d-2fe9-4daa-834a-348a4c7645ef, cib=0.53
>> 5.2) : Event failed
>> May  4 09:55:10 vm36 attrd: [19215]: info: attrd_trigger_update: Sending
>> flush op to all hosts for: last-failure-p_jboss_ocf (1304495710)
>> May  4 09:55:10 vm36 pengine: [19216]: notice: unpack_rsc_op: Hard error
>> - p_jboss_init_monitor_0 failed with rc=5: Preventing p_jboss_init from
>> re-starting on vm36
>> May  4 09:55:10 vm36 crmd: [19217]: info: match_graph_event: Action
>> p_jboss_ocf_start_0 (64) confirmed on vm36 (rc=4)
>> May  4 09:55:10 vm36 attrd: [19215]: info: attrd_perform_update: Sent
>> update 2297: last-failure-p_jboss_ocf=1304495710
>> May  4 09:55:10 vm36 pengine: [19216]: WARN: unpack_rsc_op: Processing
>> failed op p_jboss_ocf_start_0 on vm36: unknown exec error (-2)
>> May  4 09:55:10 vm36 crmd: [19217]: info: te_rsc_command: Initiating
>> action 1: stop p_jboss_ocf_stop_0 on vm36 (local)
>> May  4 09:55:10 vm36 pengine: [19216]: notice: unpack_rsc_op: Operation
>> p_jboss_ocf_monitor_0 found resource p_jboss_ocf active on vm37
>> May  4 09:55:10 vm36 crmd: [19217]: info: do_lrm_rsc_op: Performing
>> key=1:1376:0:fc16910d-2fe9-4daa-834a-348a4c7645ef op=p_jboss_ocf_stop_0 )
>> May  4 09:55:10 vm36 pengine: [19216]: notice: native_print: p_jboss_ocf
>>        (ocf::heartbeat:jboss): Stopped
>> May  4 09:55:10 vm36 pengine: [19216]: info: get_failcount: p_jboss_ocf
>> has failed INFINITY times on vm36
>> May  4 09:55:10 vm36 pengine: [19216]: WARN: common_apply_stickiness:
>> Forcing p_jboss_ocf away from vm36 after 1000000 failures (max=1000000)
>> May  4 09:59:10 vm36 pengine: [19216]: info: unpack_config: Node scores:
>> 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
>> May  4 09:59:10 vm36 crmd: [19217]: WARN: status_from_rc: Action 50
>> (p_jboss_ocf_start_0) on vm37 failed (target: 0 vs. rc: -2): Error
>> May  4 09:59:10 vm36 pengine: [19216]: info: determine_online_status:
>> Node vm36 is online
>> May  4 09:59:10 vm36 crmd: [19217]: WARN: update_failcount: Updating
>> failcount for p_jboss_ocf on vm37 after failed start: rc=-2
>> (update=INFINITY, time=1304495950)
>> May  4 09:59:10 vm36 pengine: [19216]: notice: unpack_rsc_op: Hard error
>> - p_jboss_lsb_monitor_0 failed with rc=5: Preventing p_jboss_lsb from
>> re-starting on vm36
>> May  4 09:59:10 vm36 crmd: [19217]: info: abort_transition_graph:
>> match_graph_event:272 - Triggered transition abort (complete=0,
>> tag=lrm_rsc_op, id=p_jboss_ocf_start_0,
>> magic=2:-2;50:1377:0:fc16910d-2fe9-4daa-834a-348a4c7645ef, cib=0.53
>> 5.12) : Event failed
>> May  4 09:59:10 vm36 pengine: [19216]: notice: unpack_rsc_op: Operation
>> p_jboss_cs_monitor_0 found resource p_jboss_cs active on vm36
>> May  4 09:59:10 vm36 crmd: [19217]: info: match_graph_event: Action
>> p_jboss_ocf_start_0 (50) confirmed on vm37 (rc=4)
>> May  4 09:59:10 vm36 pengine: [19216]: notice: native_print: p_jboss_ocf
>>        (ocf::heartbeat:jboss): Stopped
>> May  4 09:59:10 vm36 pengine: [19216]: info: get_failcount: p_jboss_ocf
>> has failed INFINITY times on vm37
>> May  4 09:59:10 vm36 pengine: [19216]: WARN: common_apply_stickiness:
>> Forcing p_jboss_ocf away from vm37 after 1000000 failures (max=1000000)
>> May  4 09:59:10 vm36 pengine: [19216]: info: get_failcount: p_jboss_ocf
>> has failed INFINITY times on vm36
>> May  4 09:59:10 vm36 pengine: [19216]: info: native_color: Resource
>> p_jboss_ocf cannot run anywhere
>> May  4 09:59:10 vm36 pengine: [19216]: notice: LogActions: Leave
>> resource p_jboss_ocf   (Stopped)
>> May  4 09:59:31 vm36 pengine: [19216]: notice: native_print: p_jboss_ocf
>>        (ocf::heartbeat:jboss): Stopped
>> ....
>>
>> Now i don't know how can i reset the resource p_jboss_ocf to test it again.
>>
>>> crm resource cleanup p_jboss_ocf
> 
> That's the now way, but if i start this command on shell or crm shell in
> both i get Cleaning up p_jboss_ocf on vm37
> Cleaning up p_jboss_ocf on vm36
> 
> But if i look on the monitoring with crm_mon -1 i getevery time
> 
> Failed actions:
> p_jboss_ocf_start_0 (node=vm36, call=-1, rc=1, status=Timed Out):
> unknown error
>     p_jboss_monitor_0 (node=vm37, call=205, rc=5, status=complete): not
> installed
>     p_jboss_ocf_start_0 (node=vm37, call=281, rc=-2, status=Timed Out):
> unknown exec error
> 
> p_jboss was deleted in the config yesterday.

For demonstration:

15:34:22 ~ # crm_mon -1

Failed actions:
    p_jboss_ocf_start_0 (node=vm36, call=376, rc=-2, status=Timed Out):
unknown exec error
    p_jboss_monitor_0 (node=vm37, call=205, rc=5, status=complete): not
installed
    p_jboss_ocf_start_0 (node=vm37, call=283, rc=-2, status=Timed Out):
unknown exec error

15:35:02 ~ # crm resource cleanup p_jboss_ocf
INFO: no curses support: you won't see colors
Cleaning up p_jboss_ocf on vm37
Cleaning up p_jboss_ocf on vm36

15:39:12 ~ # crm resource cleanup p_jboss
INFO: no curses support: you won't see colors
Cleaning up p_jboss on vm37
Cleaning up p_jboss on vm36

15:39:19 ~ # crm_mon -1

Failed actions:
    p_jboss_ocf_start_0 (node=vm36, call=376, rc=-2, status=Timed Out):
unknown exec error
    p_jboss_monitor_0 (node=vm37, call=205, rc=5, status=complete): not
installed
    p_jboss_ocf_start_0 (node=vm37, call=283, rc=-2, status=Timed Out):
unknown exec error
> 
>>
>> And after some tests i have some not  more existing resouces in the
>> Failed actions list. How can i delete them?
>>
>>> The same way.
>>
>>> Thanks,
>>
>>> Dejan
>>

Thx

Benjamin
>>
>>
>>
>>>>>
>>>>> Florian
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> _______________________________________________
> 
> 
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker




More information about the Pacemaker mailing list