[Pacemaker] Why "order o inf: VIP A B" starts VIP, A and B simultaneously ?

Sékine Coulibaly scoulibaly at gmail.com
Wed Jun 25 09:56:03 UTC 2014


Additionnaly, stopping VIP73 only stop the ZK resource, not PostgreSQL.
I'm definitely doing something wrong or making bad assumptions. Can anyone
help ?
Thank you !

Jun 25 05:44:11 clustera cib[33828]:     info: cib:diff: - <cib
admin_epoch="0" epoch="498" num_updates="26" />
Jun 25 05:44:11 clustera cib[33828]:     info: cib:diff: + <cib epoch="499"
num_updates="1" admin_epoch="0" validate-with="pacemaker-1.2"
crm_feature_set="3.0.6" update-origin="clustera" update-client="cibadmin"
cib-last-written="Wed Jun 25 04:42:04 2014" have-quorum="0"
dc-uuid="clustera" >
Jun 25 05:44:11 clustera cib[33828]:     info: cib:diff: +   <configuration
>
Jun 25 05:44:11 clustera cib[33828]:     info: cib:diff: +     <resources >
Jun 25 05:44:11 clustera cib[33828]:     info: cib:diff: +       <primitive
class="ocf" id="VIP73" provider="heartbeat" type="IPaddr2" >
Jun 25 05:44:11 clustera cib[33828]:     info: cib:diff: +
<meta_attributes id="VIP73-meta_attributes" __crm_diff_marker__="added:top"
>
Jun 25 05:44:11 clustera cib[33828]:     info: cib:diff: +
<nvpair id="VIP73-meta_attributes-target-role" name="target-role"
value="Stopped" />
Jun 25 05:44:11 clustera cib[33828]:     info: cib:diff: +
</meta_attributes>
Jun 25 05:44:11 clustera cib[33828]:     info: cib:diff: +
</primitive>
Jun 25 05:44:11 clustera cib[33828]:     info: cib:diff: +     </resources>
Jun 25 05:44:11 clustera cib[33828]:     info: cib:diff: +
</configuration>
Jun 25 05:44:11 clustera cib[33828]:     info: cib:diff: + </cib>
Jun 25 05:44:11 clustera cib[33828]:     info: cib_process_request:
Operation complete: op cib_replace for section resources
(origin=local/cibadmin/2, version=0.499.1): ok (rc=0)
Jun 25 05:44:11 clustera crmd[33833]:     info: abort_transition_graph:
te_update_diff:126 - Triggered transition abort (complete=1, tag=diff,
id=(null), magic=NA, cib=0.499.1) : Non-status change
Jun 25 05:44:11 clustera crmd[33833]:   notice: do_state_transition: State
transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_FSA_INTERNAL
origin=abort_transition_graph ]
Jun 25 05:44:11 clustera pengine[33832]:     info: unpack_config: Startup
probes: enabled
Jun 25 05:44:11 clustera pengine[33832]:   notice: unpack_config: On loss
of CCM Quorum: Ignore
Jun 25 05:44:11 clustera pengine[33832]:     info: unpack_config: Node
scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Jun 25 05:44:11 clustera pengine[33832]:     info: unpack_domains:
Unpacking domains
Jun 25 05:44:11 clustera pengine[33832]:     info: determine_online_status:
Node clustera is online
Jun 25 05:44:11 clustera pengine[33832]:     info: native_print:
VIP73#011(ocf::heartbeat:IPaddr2):#011Started clustera
Jun 25 05:44:11 clustera pengine[33832]:     info: clone_print:
 Master/Slave Set: MS_POSTGRESQL [POSTGRESQL]
Jun 25 05:44:11 clustera pengine[33832]:     info: short_print:
 Masters: [ clustera ]
Jun 25 05:44:11 clustera pengine[33832]:     info: short_print:
 Stopped: [ POSTGRESQL:1 ]
Jun 25 05:44:11 clustera pengine[33832]:     info: native_print:
ZK#011(ocf::custom:zookeeper):#011Started clustera
Jun 25 05:44:11 clustera pengine[33832]:     info: native_color: Resource
VIP73 cannot run anywhere
Jun 25 05:44:11 clustera pengine[33832]:     info: native_color: Resource
POSTGRESQL:1 cannot run anywhere
Jun 25 05:44:11 clustera pengine[33832]:     info: master_color: Promoting
POSTGRESQL:0 (Master clustera)
Jun 25 05:44:11 clustera pengine[33832]:     info: master_color:
MS_POSTGRESQL: Promoted 1 instances of a possible 1 to master
Jun 25 05:44:11 clustera pengine[33832]:   notice: LogActions: Stop
 VIP73#011(clustera)
Jun 25 05:44:11 clustera pengine[33832]:     info: LogActions: Leave
POSTGRESQL:0#011(Master clustera)
Jun 25 05:44:11 clustera pengine[33832]:     info: LogActions: Leave
POSTGRESQL:1#011(Stopped)
Jun 25 05:44:11 clustera pengine[33832]:   notice: LogActions: Stop
 ZK#011(Started clustera)
Jun 25 05:44:11 clustera crmd[33833]:   notice: do_state_transition: State
transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS
cause=C_IPC_MESSAGE origin=handle_response ]
Jun 25 05:44:11 clustera crmd[33833]:     info: do_te_invoke: Processing
graph 4 (ref=pe_calc-dc-1403689451-28) derived from
/var/lib/pengine/pe-input-3209.bz2
Jun 25 05:44:11 clustera crmd[33833]:     info: te_rsc_command: Initiating
action 36: stop ZK_stop_0 on clustera (local)
Jun 25 05:44:11 clustera lrmd: [33830]: info: cancel_op: operation
monitor[9] on ocf::zookeeper::ZK for client 33833, its parameters:
CRM_meta_name=[monitor] crm_feature_set=[3.0.6] CRM_meta_timeout=[10000]
CRM_meta_interval=[5000]  cancelled
Jun 25 05:44:11 clustera lrmd: [33830]: info: rsc:ZK:15: stop
Jun 25 05:44:11 clustera crmd[33833]:     info: process_lrm_event: LRM
operation ZK_monitor_5000 (call=9, status=1, cib-update=0, confirmed=true)
Cancelled
Jun 25 05:44:11 clustera pengine[33832]:   notice: process_pe_message:
Transition 4: PEngine Input stored in: /var/lib/pengine/pe-input-3209.bz2
Jun 25 05:44:11 clustera zookeeper(ZK)[45537]: INFO: Killing Zookeeper PID
34282
Jun 25 05:44:11 clustera lrmd: [33830]: info: RA output: (ZK:stop:stdout)
Stopping zookeeper ...
Jun 25 05:44:11 clustera lrmd: [33830]: info: RA output: (ZK:stop:stdout)
STOPPED
Jun 25 05:44:11 clustera lrmd: [33830]: info: Managed ZK:stop process 45537
exited with return code 0.
Jun 25 05:44:11 clustera crmd[33833]:     info: process_lrm_event: LRM
operation ZK_stop_0 (call=15, rc=0, cib-update=40, confirmed=true) ok
Jun 25 05:44:11 clustera crmd[33833]:     info: te_rsc_command: Initiating
action 7: stop VIP73_stop_0 on clustera (local)
Jun 25 05:44:11 clustera lrmd: [33830]: info: cancel_op: operation
monitor[7] on ocf::IPaddr2::VIP73 for client 33833, its parameters:
cidr_netmask=[24] nic=[eth1] broadcast=[192.168.73.255]
crm_feature_set=[3.0.6] ip=[192.168.73.222] iflabel=[VIP73]
CRM_meta_name=[monitor] CRM_meta_timeout=[20000] CRM_meta_interval=[10000]
 cancelled
Jun 25 05:44:11 clustera lrmd: [33830]: info: rsc:VIP73:16: stop
Jun 25 05:44:11 clustera crmd[33833]:     info: process_lrm_event: LRM
operation VIP73_monitor_10000 (call=7, status=1, cib-update=0,
confirmed=true) Cancelled
Jun 25 05:44:11 clustera IPaddr2(VIP73)[45649]: INFO: IP status = ok,
IP_CIP=
Jun 25 05:44:11 clustera avahi-daemon[2065]: Withdrawing address record for
192.168.73.222 on eth1.
Jun 25 05:44:11 clustera lrmd: [33830]: info: Managed VIP73:stop process
45649 exited with return code 0.
Jun 25 05:44:11 clustera crmd[33833]:     info: process_lrm_event: LRM
operation VIP73_stop_0 (call=16, rc=0, cib-update=41, confirmed=true) ok
Jun 25 05:44:11 clustera crmd[33833]:   notice: run_graph: ==== Transition
4 (Complete=3, Pending=0, Fired=0, Skipped=0, Incomplete=0,
Source=/var/lib/pengine/pe-input-3209.bz2): Complete
Jun 25 05:44:11 clustera crmd[33833]:   notice: do_state_transition: State
transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS
cause=C_FSA_INTERNAL origin=notify_crmd ]



2014-06-25 11:36 GMT+02:00 Sékine Coulibaly <scoulibaly at gmail.com>:

> Hi all,
>
> My setup is as follows : RedHat 6.3 (yes, I know,this is quite old) ,
> Pacemaker 1.1.7, Corosync 1.4.1.
>
> I noticed something that is strange because since it doesn't complies with
> what I read (and understood) from the following ressources :
> 1. http://crmsh.nongnu.org/crm.8.html#cmdhelp_configure_order
> 2.
> http://doc.opensuse.org/products/draft/SLE-HA/SLE-ha-guide_sd_draft/cha.ha.manual_config.html
>
>
> The strange behaviour I experience is related to ordering.
> Let's have this very basic case :
>
> One zookeeper process, and one PostgreSQL master. I want them to run
> alltogether on the master node. Since they bind on the VIP IP, I want the
> VIP to be started before ZK and PostgreSQL.
> So I setup :
>
> primitive VIP73 ocf:heartbeat:IPaddr2 \
>         params ip="192.168.73.222" broadcast="192.168.73.255" nic="eth1"
> cidr_netmask="24" iflabel="VIP73" \
>         op monitor interval="10s" timeout="20s"
>
> primitive POSTGRESQL ocf:custom:postgresql \
>         params repmgr_conf="/var/lib/pgsql/repmgr/repmgr.conf"
> pgctl="/usr/pgsql-9.2/bin/pg_ctl" pgdata="/opt/custom/pgdata" \
>         op start interval="0" timeout="90s" \
>         op stop interval="0" timeout="60s" \
>         op promote interval="0" timeout="120s" \
>         op monitor interval="53s" role="Master" \
>         op monitor interval="60s" role="Slave"
>
> primitive ZK ocf:custom:zookeeper \
>         op monitor interval="5s" timeout="10s" \
>         op start interval="0" timeout="10s" \
>         op stop interval="0" timeout="10s"
>
> ms MS_POSTGRESQL POSTGRESQL \
>         meta clone-max="2" target-role="Started" resource-stickiness="100"
> notify="true"
>
> I then add an ordering such as this one :
> order VIP_last inf: VIP73 ZK MS_POSTGRESQL:promote
>
> I expect the VIP to be mounted, and then ZK started, and then PostgreSQL
> master to be promoted. Instead, all the resources seem to be started in
> parallel.
>
> Here comes a /var/log/messages extract, taken just after a corosync
> restart.
> It can be seen at Jun 25 05:19:52 that VIP73 and POSTGRESQL are all
> started simultaneously.
> Do I have something wrong here, or something I didn't understand ?
>
>
>
> primitive VIP73 ocf:heartbeat:IPaddr2 \
>         params ip="192.168.73.222" broadcast="192.168.73.255" nic="eth1"
> cidr_netmask="24" iflabel="VIP73" \
>         op monitor interval="10s" timeout="20s"
>
> primitive POSTGRESQL ocf:custom:postgresql \
>         params repmgr_conf="/var/lib/pgsql/repmgr/repmgr.conf"
> pgctl="/usr/pgsql-9.2/bin/pg_ctl" pgdata="/opt/custom/pgdata" \
>         op start interval="0" timeout="90s" \
>         op stop interval="0" timeout="60s" \
>         op promote interval="0" timeout="120s" \
>         op monitor interval="53s" role="Master" \
>         op monitor interval="60s" role="Slave"
>
> primitive ZK ocf:custom:zookeeper \
>         op monitor interval="5s" timeout="10s" \
>         op start interval="0" timeout="10s" \
>         op stop interval="0" timeout="10s"
> ms MS_POSTGRESQL POSTGRESQL \
>         meta clone-max="2" target-role="Started" resource-stickiness="100"
> notify="true"
>
> I then add an ordering such as this one :
> order VIP_last inf: VIP73 ZK MS_POSTGRESQL:promote
>
> I expect the VIP to be mounted, and then ZK started, and then PostgreSQL
> master to be promoted. Instead, all the resources seem to be started in
> parallel.
>
> Here comes a /var/log/messages extract, taken just after a corosync
> restart.
> It can be seen at Jun 25 05:19:52 that VIP73 and POSTGRESQL are all
> started simultaneously.
> Do I have something wrong here, or something I didn't understand ?
>
> Thank you !
>
> Jun 25 05:19:52 clustera pengine[33832]:     info:
> determine_online_status: Node clustera is online
> Jun 25 05:19:52 clustera pengine[33832]:     info: native_print:
> VIP73#011(ocf::heartbeat:IPaddr2):#011Stopped
> Jun 25 05:19:52 clustera pengine[33832]:     info: clone_print:
>  Master/Slave Set: MS_POSTGRESQL [POSTGRESQL]
> Jun 25 05:19:52 clustera pengine[33832]:     info: short_print:
>  Stopped: [ POSTGRESQL:0 POSTGRESQL:1 ]
> Jun 25 05:19:52 clustera pengine[33832]:     info: native_print:
> ZK#011(ocf::custom:zookeeper):#011Stopped
> Jun 25 05:19:52 clustera pengine[33832]:     info: native_color: Resource
> POSTGRESQL:1 cannot run anywhere
> Jun 25 05:19:52 clustera pengine[33832]:     info: master_color:
> MS_POSTGRESQL: Promoted 0 instances of a possible 1 to master
> Jun 25 05:19:52 clustera pengine[33832]:     info: RecurringOp:  Start
> recurring monitor (10s) for VIP73 on clustera
> Jun 25 05:19:52 clustera pengine[33832]:     info: RecurringOp:  Start
> recurring monitor (60s) for POSTGRESQL:0 on clustera
> Jun 25 05:19:52 clustera pengine[33832]:     info: RecurringOp:  Start
> recurring monitor (60s) for POSTGRESQL:0 on clustera
> Jun 25 05:19:52 clustera pengine[33832]:     info: RecurringOp:  Start
> recurring monitor (5s) for ZK on clustera
> Jun 25 05:19:52 clustera pengine[33832]:   notice: LogActions: Start
> VIP73#011(clustera)
> Jun 25 05:19:52 clustera pengine[33832]:   notice: LogActions: Start
> POSTGRESQL:0#011(clustera)
> Jun 25 05:19:52 clustera pengine[33832]:     info: LogActions: Leave
> POSTGRESQL:1#011(Stopped)
> Jun 25 05:19:52 clustera pengine[33832]:   notice: LogActions: Start
> ZK#011(clustera)
>
> ...
> Jun 25 05:19:52 clustera crmd[33833]:     info: te_rsc_command: Initiating
> action 3: probe_complete probe_complete on clustera (local) - no waiting
> Jun 25 05:19:52 clustera crmd[33833]:     info: te_rsc_command: Initiating
> action 7: start VIP73_start_0 on clustera (local)
> Jun 25 05:19:52 clustera attrd[33831]:     info: find_hash_entry: Creating
> hash entry for probe_complete
> Jun 25 05:19:52 clustera attrd[33831]:   notice: attrd_trigger_update:
> Sending flush op to all hosts for: probe_complete (true)
> Jun 25 05:19:52 clustera lrmd: [33830]: info: rsc:VIP73:5: start
> Jun 25 05:19:52 clustera attrd[33831]:   notice: attrd_perform_update:
> Sent update 4: probe_complete=true
> Jun 25 05:19:52 clustera crmd[33833]:     info: te_rsc_command: Initiating
> action 9: start POSTGRESQL:0_start_0 on clustera (local)
> Jun 25 05:19:52 clustera lrmd: [33830]: info: rsc:POSTGRESQL:0:6: start
> Jun 25 05:19:52 clustera IPaddr2(VIP73)[33953]: INFO: ip -f inet addr add
> 192.168.73.222/24 brd 192.168.73.255 dev eth1 label eth1:VIP73
> Jun 25 05:19:52 clustera avahi-daemon[2065]: Registering new address
> record for 192.168.73.222 on eth1.IPv4.
> Jun 25 05:19:52 clustera IPaddr2(VIP73)[33953]: INFO: ip link set eth1 up
> Jun 25 05:19:52 clustera IPaddr2(VIP73)[33953]: INFO:
> /usr/lib64/heartbeat/send_arp -i 200 -r 5 -p
> /var/run/heartbeat/rsctmp/send_arp-192.168.73.222 eth1 192.168.73.222 auto
> not_used not_used
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> POSTGRESQL:0: Starting
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> ACTION=start
> Jun 25 05:19:52 clustera lrmd: [33830]: info: Managed VIP73:start process
> 33953 exited with return code 0.
> Jun 25 05:19:52 clustera crmd[33833]:     info: process_lrm_event: LRM
> operation VIP73_start_0 (call=5, rc=0, cib-update=29, confirmed=true) ok
> Jun 25 05:19:52 clustera crmd[33833]:     info: te_rsc_command: Initiating
> action 8: monitor VIP73_monitor_10000 on clustera (local)
> Jun 25 05:19:52 clustera lrmd: [33830]: info: rsc:VIP73:7: monitor
> Jun 25 05:19:52 clustera crmd[33833]:     info: te_rsc_command: Initiating
> action 35: start ZK_start_0 on clustera (local)
> Jun 25 05:19:52 clustera lrmd: [33830]: info: rsc:ZK:8: start
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RA_VERSION_MAJOR=1
> Jun 25 05:19:52 clustera zookeeper(ZK)[34058]: INFO: [ZK] No pid file found
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RA_VERSION_MINOR=0
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_clone=0
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_clone_max=2
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_clone_node_max=1
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_globally_unique=false
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_master_max=1
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_master_node_max=1
> Jun 25 05:19:52 clustera lrmd: [33830]: info: Managed VIP73:monitor
> process 34057 exited with return code 0.
> Jun 25 05:19:52 clustera crmd[33833]:     info: process_lrm_event: LRM
> operation VIP73_monitor_10000 (call=7, rc=0, cib-update=30,
> confirmed=false) ok
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_name=start
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_notify=true
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_notify_active_resource=
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_notify_active_uname=
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_notify_demote_resource=
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_notify_demote_uname=
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_notify_inactive_resource=POSTGRESQL:0
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> POSTGRESQL:1
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_notify_master_resource=
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_notify_master_uname=
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_notify_promote_resource=
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_notify_promote_uname=
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_notify_slave_resource=
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_notify_slave_uname=
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_notify_start_resource=POSTGRESQL:0
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_notify_start_uname=clustera
> Jun 25 05:19:52 clustera lrmd: [33830]: info: RA output: (ZK:start:stdout)
> Starting zookeeper ...
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_notify_stop_resource=
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_notify_stop_uname=
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_CRM_meta_timeout=90000
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_crm_feature_set=3.0.6
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_pgctl=/usr/pgsql-9.2/bin/pg_ctl
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_pgdata=/opt/custom/pgdata
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESKEY_repmgr_conf=/var/lib/pgsql/repmgr/repmgr.conf
> Jun 25 05:19:52 clustera avahi-daemon[2065]: Invalid legacy unicast query
> packet.
> Jun 25 05:19:52 clustera avahi-daemon[2065]: Received response from host
> 192.168.72.1 with invalid source port 50070 on interface 'eth0.0'
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESOURCE_INSTANCE=POSTGRESQL:0
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESOURCE_PROVIDER=custom
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_RESOURCE_TYPE=postgresql
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO:
> OCF_ROOT=/usr/lib/ocf
> Jun 25 05:19:52 clustera postgresql(POSTGRESQL:0)[33954]: INFO: Run as
> postgres: /usr/pgsql-9.2/bin/pg_ctl start -w -D /opt/custom/pgdata -l
> /var/log/custom/custom/ha/postgresql/postgres_ha.log -o '-c
> config_file=/opt/custom/pgdata/postgresql.conf' -o '-p 5432'
> Jun 25 05:19:53 clustera avahi-daemon[2065]: Invalid legacy unicast query
> packet.
> Jun 25 05:19:53 clustera avahi-daemon[2065]: Invalid legacy unicast query
> packet.
> Jun 25 05:19:53 clustera avahi-daemon[2065]: Received response from host
> 192.168.72.1 with invalid source port 50070 on interface 'eth0.0'
> Jun 25 05:19:53 clustera avahi-daemon[2065]: Received response from host
> 192.168.72.1 with invalid source port 50070 on interface 'eth0.0'
> Jun 25 05:19:53 clustera lrmd: [33830]: info: RA output: (ZK:start:stdout)
> STARTED
> Jun 25 05:19:53 clustera lrmd: [33830]: info: RA output: (ZK:start:stdout)
> ---------------------------------------------------------#012Starting ZK
> restore#012---------------------------------------------------------
> Jun 25 05:19:53 clustera lrmd: [33830]: info: RA output: (ZK:start:stderr)
> ls: cannot access /var/custom/custom/ha/snapshots/zk_dump_*.xml: No such
> file or directory
> Jun 25 05:19:53 clustera lrmd: [33830]: info: RA output: (ZK:start:stdout)
> [ZK]   Last snapshot :           <NONE>#012[ZK]   WARNING: Nothing to
> restore
> Jun 25 05:19:53 clustera zookeeper(ZK)[34058]: INFO: [ZK] Nothing to
> restore
> Jun 25 05:19:53 clustera lrmd: [33830]: info: RA output: (ZK:start:stdout)
> ---------------------------------------------------------
> Jun 25 05:19:53 clustera lrmd: [33830]: info: Managed ZK:start process
> 34058 exited with return code 0.
> Jun 25 05:19:53 clustera crmd[33833]:     info: process_lrm_event: LRM
> operation ZK_start_0 (call=8, rc=0, cib-update=31, confirmed=true) ok
> Jun 25 05:19:53 clustera crmd[33833]:     info: te_rsc_command: Initiating
> action 36: monitor ZK_monitor_5000 on clustera (local)
> Jun 25 05:19:53 clustera lrmd: [33830]: info: rsc:ZK:9: monitor
> Jun 25 05:19:53 clustera lrmd: [33830]: info: Managed ZK:monitor process
> 34390 exited with return code 0.
> Jun 25 05:19:53 clustera crmd[33833]:     info: process_lrm_event: LRM
> operation ZK_monitor_5000 (call=9, rc=0, cib-update=32, confirmed=false) ok
> Jun 25 05:19:53 clustera postgresql(POSTGRESQL:0)[33954]: INFO: waiting
> for server to start.... done server started
> Jun 25 05:19:54 clustera avahi-daemon[2065]: Received response from host
> 192.168.72.1 with invalid source port 50070 on interface 'eth0.0'
> Jun 25 05:19:56 clustera avahi-daemon[2065]: Received response from host
> 192.168.72.1 with invalid source port 50070 on interface 'eth0.0'
> Jun 25 05:19:56 clustera lrmd: [33830]: info: RA output:
> (VIP73:start:stderr) ARPING 192.168.73.222 from 192.168.73.222 eth1#012Sent
> 5 probes (5 broadcast(s))#
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20140625/ebf6705c/attachment.htm>


More information about the Pacemaker mailing list