May 14 15:06:52 sv07 cib[3153]:   notice: cib:diff: Diff: --- 11.12.7
May 14 15:06:52 sv07 cib[3153]:   notice: cib:diff: Diff: +++ 11.13.1 915614dd85ebe2a23c89ecf05d0834a6
May 14 15:06:52 sv07 cib[3153]:   notice: cib:diff: --           <nvpair id="nodes-sv06-standby" name="standby" value="on"/>
May 14 15:06:52 sv07 cib[3153]:   notice: cib:diff: ++ <cib admin_epoch="11" cib-last-written="Wed May 14 15:06:52 2014" crm_feature_set="3.0.7" epoch="13" num_updates="1" validate-with="pacemaker-1.2" update-origin="sv07" update-client="crm_attribute" have-quorum="1" dc-uuid="sv07"/>
May 14 15:06:52 sv07 crmd[3158]:   notice: do_state_transition: State transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_FSA_INTERNAL origin=abort_transition_graph ]
May 14 15:06:52 sv07 pengine[3157]:   notice: LogActions: Start   res_drbd_vm1:1#011(sv06)
May 14 15:06:52 sv07 pengine[3157]:   notice: LogActions: Start   res_drbd_live:1#011(sv06)
May 14 15:06:52 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 121: notify res_drbd_vm1_pre_notify_start_0 on sv07 (local)
May 14 15:06:52 sv07 pengine[3157]:   notice: process_pe_message: Calculated Transition 410: /var/lib/pacemaker/pengine/pe-input-962.bz2
May 14 15:06:52 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 132: notify res_drbd_live_pre_notify_start_0 on sv07 (local)
May 14 15:06:52 sv07 crmd[3158]:   notice: process_lrm_event: LRM operation res_drbd_vm1_notify_0 (call=2486, rc=0, cib-update=0, confirmed=true) ok
May 14 15:06:52 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 19: start res_drbd_vm1_start_0 on sv06
May 14 15:06:52 sv07 crmd[3158]:   notice: process_lrm_event: LRM operation res_drbd_live_notify_0 (call=2488, rc=0, cib-update=0, confirmed=true) ok
May 14 15:06:52 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 79: start res_drbd_live_start_0 on sv06
May 14 15:06:53 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 122: notify res_drbd_vm1_post_notify_start_0 on sv07 (local)
May 14 15:06:53 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 123: notify res_drbd_vm1_post_notify_start_0 on sv06
May 14 15:06:53 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 133: notify res_drbd_live_post_notify_start_0 on sv07 (local)
May 14 15:06:53 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 134: notify res_drbd_live_post_notify_start_0 on sv06
May 14 15:06:53 sv07 crmd[3158]:   notice: process_lrm_event: LRM operation res_drbd_vm1_notify_0 (call=2492, rc=0, cib-update=0, confirmed=true) ok
May 14 15:06:53 sv07 crmd[3158]:   notice: process_lrm_event: LRM operation res_drbd_live_notify_0 (call=2494, rc=0, cib-update=0, confirmed=true) ok
May 14 15:06:53 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 80: monitor res_drbd_live_monitor_60000 on sv06
May 14 15:06:53 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 20: monitor res_drbd_vm1_monitor_60000 on sv06
May 14 15:06:53 sv07 crmd[3158]:   notice: run_graph: Transition 410 (Complete=26, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-962.bz2): Complete
May 14 15:06:53 sv07 crmd[3158]:   notice: do_state_transition: State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd ]
May 14 15:06:54 sv07 kernel: drbd live: Handshake successful: Agreed network protocol version 101
May 14 15:06:54 sv07 kernel: drbd vm1: Handshake successful: Agreed network protocol version 101
May 14 15:06:54 sv07 kernel: drbd vm1: Agreed to support TRIM on protocol level
May 14 15:06:54 sv07 kernel: drbd vm1: conn( WFConnection -> WFReportParams ) 
May 14 15:06:54 sv07 kernel: drbd vm1: Starting asender thread (from drbd_r_vm1 [26859])
May 14 15:06:54 sv07 kernel: drbd live: Agreed to support TRIM on protocol level
May 14 15:06:54 sv07 kernel: drbd live: conn( WFConnection -> WFReportParams ) 
May 14 15:06:54 sv07 kernel: drbd live: Starting asender thread (from drbd_r_live [26852])
May 14 15:06:54 sv07 kernel: block drbd2: drbd_sync_handshake:
May 14 15:06:54 sv07 kernel: block drbd2: self 3949FB1C48B13D5B:6E9B4E45BA168C37:851DCB4095274155:851CCB4095274155 bits:119 flags:0
May 14 15:06:54 sv07 kernel: block drbd2: peer 6E9B4E45BA168C36:0000000000000000:851DCB4095274154:851CCB4095274155 bits:0 flags:0
May 14 15:06:54 sv07 kernel: block drbd2: uuid_compare()=1 by rule 70
May 14 15:06:54 sv07 kernel: block drbd2: peer( Unknown -> Secondary ) conn( WFReportParams -> WFBitMapS ) pdsk( Outdated -> Consistent ) 
May 14 15:06:55 sv07 kernel: block drbd1: drbd_sync_handshake:
May 14 15:06:55 sv07 kernel: block drbd2: send bitmap stats [Bytes(packets)]: plain 0(0), RLE 74(1), total 74; compression: 100.0%
May 14 15:06:55 sv07 kernel: block drbd2: receive bitmap stats [Bytes(packets)]: plain 0(0), RLE 74(1), total 74; compression: 100.0%
May 14 15:06:55 sv07 kernel: block drbd2: helper command: /sbin/drbdadm before-resync-source minor-2
May 14 15:06:55 sv07 kernel: block drbd2: helper command: /sbin/drbdadm before-resync-source minor-2 exit code 0 (0x0)
May 14 15:06:55 sv07 kernel: block drbd2: conn( WFBitMapS -> SyncSource ) pdsk( Consistent -> Inconsistent ) 
May 14 15:06:55 sv07 kernel: block drbd2: Began resync as SyncSource (will sync 476 KB [119 bits set]).
May 14 15:06:55 sv07 kernel: block drbd2: updated sync UUID 3949FB1C48B13D5B:6E9C4E45BA168C37:6E9B4E45BA168C37:851DCB4095274155
May 14 15:06:55 sv07 kernel: block drbd1: self 4CBC95F3FC5C90F1:0C42AEF8E8A199DB:24720B3E4115A6F9:24710B3E4115A6F9 bits:0 flags:0
May 14 15:06:55 sv07 kernel: block drbd1: peer 0C42AEF8E8A199DA:0000000000000000:24720B3E4115A6F8:24710B3E4115A6F9 bits:0 flags:0
May 14 15:06:55 sv07 kernel: block drbd1: uuid_compare()=1 by rule 70
May 14 15:06:55 sv07 kernel: block drbd1: peer( Unknown -> Secondary ) conn( WFReportParams -> WFBitMapS ) pdsk( Outdated -> Consistent ) 
May 14 15:06:55 sv07 kernel: block drbd1: send bitmap stats [Bytes(packets)]: plain 0(0), RLE 23(1), total 23; compression: 100.0%
May 14 15:06:55 sv07 kernel: block drbd1: receive bitmap stats [Bytes(packets)]: plain 0(0), RLE 23(1), total 23; compression: 100.0%
May 14 15:06:55 sv07 kernel: block drbd1: helper command: /sbin/drbdadm before-resync-source minor-1
May 14 15:06:55 sv07 kernel: block drbd1: helper command: /sbin/drbdadm before-resync-source minor-1 exit code 0 (0x0)
May 14 15:06:55 sv07 kernel: block drbd1: conn( WFBitMapS -> SyncSource ) pdsk( Consistent -> Inconsistent ) 
May 14 15:06:55 sv07 kernel: block drbd2: Resync done (total 1 sec; paused 0 sec; 476 K/sec)
May 14 15:06:55 sv07 kernel: block drbd2: updated UUIDs 3949FB1C48B13D5B:0000000000000000:6E9C4E45BA168C37:6E9B4E45BA168C37
May 14 15:06:55 sv07 kernel: block drbd2: conn( SyncSource -> Connected ) pdsk( Inconsistent -> UpToDate ) 
May 14 15:06:55 sv07 kernel: block drbd1: Began resync as SyncSource (will sync 0 KB [0 bits set]).
May 14 15:06:55 sv07 kernel: block drbd1: updated sync UUID 4CBC95F3FC5C90F1:0C43AEF8E8A199DB:0C42AEF8E8A199DB:24720B3E4115A6F9
May 14 15:06:56 sv07 kernel: block drbd1: Resync done (total 1 sec; paused 0 sec; 0 K/sec)
May 14 15:06:56 sv07 kernel: block drbd1: updated UUIDs 4CBC95F3FC5C90F1:0000000000000000:0C43AEF8E8A199DB:0C42AEF8E8A199DB
May 14 15:06:56 sv07 kernel: block drbd1: conn( SyncSource -> Connected ) pdsk( Inconsistent -> UpToDate ) 
May 14 15:07:53 sv07 crmd[3158]:   notice: do_state_transition: State transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_FSA_INTERNAL origin=abort_transition_graph ]
May 14 15:07:53 sv07 pengine[3157]:   notice: LogActions: Promote res_drbd_vm1:1#011(Slave -> Master sv06)
May 14 15:07:53 sv07 pengine[3157]:   notice: LogActions: Restart res_fs_vm1:0#011(Started sv07)
May 14 15:07:53 sv07 pengine[3157]:   notice: LogActions: Start   res_fs_vm1:1#011(sv06)
May 14 15:07:53 sv07 pengine[3157]:   notice: LogActions: Restart res_vm_nfs_server#011(Started sv07)
May 14 15:07:53 sv07 pengine[3157]:   notice: process_pe_message: Calculated Transition 411: /var/lib/pacemaker/pengine/pe-input-963.bz2
May 14 15:07:53 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 4: cancel res_drbd_vm1_cancel_60000 on sv06
May 14 15:07:53 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 118: stop res_vm_nfs_server_stop_0 on sv07 (local)
May 14 15:07:53 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 132: notify res_drbd_vm1_pre_notify_promote_0 on sv07 (local)
May 14 15:07:53 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 134: notify res_drbd_vm1_pre_notify_promote_0 on sv06
May 14 15:07:53 sv07 crmd[3158]:   notice: process_lrm_event: LRM operation res_drbd_vm1_notify_0 (call=2502, rc=0, cib-update=0, confirmed=true) ok
May 14 15:07:53 sv07 VirtualDomain(res_vm_nfs_server)[29595]: INFO: Issuing graceful shutdown request for domain vm09.
May 14 15:08:00 sv07 kernel: server_lan: port 2(vnet0) entering disabled state
May 14 15:08:00 sv07 kernel: device vnet0 left promiscuous mode
May 14 15:08:00 sv07 kernel: server_lan: port 2(vnet0) entering disabled state
May 14 15:08:00 sv07 crmd[3158]:   notice: process_lrm_event: LRM operation res_vm_nfs_server_stop_0 (call=2499, rc=0, cib-update=1411, confirmed=true) ok
May 14 15:08:00 sv07 crmd[3158]:   notice: run_graph: Transition 411 (Complete=5, Pending=0, Fired=0, Skipped=21, Incomplete=7, Source=/var/lib/pacemaker/pengine/pe-input-963.bz2): Stopped
May 14 15:08:00 sv07 pengine[3157]:   notice: LogActions: Promote res_drbd_vm1:1#011(Slave -> Master sv06)
May 14 15:08:00 sv07 pengine[3157]:   notice: LogActions: Restart res_fs_vm1:0#011(Started sv07)
May 14 15:08:00 sv07 pengine[3157]:   notice: LogActions: Start   res_fs_vm1:1#011(sv06)
May 14 15:08:00 sv07 pengine[3157]:   notice: LogActions: Start   res_vm_nfs_server#011(sv07)
May 14 15:08:00 sv07 pengine[3157]:   notice: process_pe_message: Calculated Transition 412: /var/lib/pacemaker/pengine/pe-input-964.bz2
May 14 15:08:00 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 130: notify res_drbd_vm1_pre_notify_promote_0 on sv07 (local)
May 14 15:08:00 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 132: notify res_drbd_vm1_pre_notify_promote_0 on sv06
May 14 15:08:00 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 54: stop res_fs_vm1_stop_0 on sv07 (local)
May 14 15:08:01 sv07 crmd[3158]:   notice: process_lrm_event: LRM operation res_drbd_vm1_notify_0 (call=2506, rc=0, cib-update=0, confirmed=true) ok
May 14 15:08:01 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 21: promote res_drbd_vm1_promote_0 on sv06
May 14 15:08:01 sv07 Filesystem(res_fs_vm1)[29727]: INFO: Running stop for /dev/drbd/by-res/vm1 on /data/vm1
May 14 15:08:01 sv07 kernel: block drbd2: peer( Secondary -> Primary ) 
May 14 15:08:01 sv07 Filesystem(res_fs_vm1)[29727]: INFO: Trying to unmount /data/vm1
May 14 15:08:01 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 131: notify res_drbd_vm1_post_notify_promote_0 on sv07 (local)
May 14 15:08:01 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 133: notify res_drbd_vm1_post_notify_promote_0 on sv06
May 14 15:08:01 sv07 crmd[3158]:   notice: process_lrm_event: LRM operation res_drbd_vm1_notify_0 (call=2513, rc=0, cib-update=0, confirmed=true) ok
May 14 15:08:01 sv07 Filesystem(res_fs_vm1)[29727]: INFO: unmounted /data/vm1 successfully
May 14 15:08:01 sv07 crmd[3158]:   notice: process_lrm_event: LRM operation res_fs_vm1_stop_0 (call=2509, rc=0, cib-update=1413, confirmed=true) ok
May 14 15:08:01 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 55: start res_fs_vm1_start_0 on sv07 (local)
May 14 15:08:01 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 60: start res_fs_vm1_start_0 on sv06
May 14 15:08:01 sv07 Filesystem(res_fs_vm1)[29860]: INFO: Running start for /dev/drbd/by-res/vm1 on /data/vm1
May 14 15:08:01 sv07 ntpd[2527]: Deleting interface #30 vnet0, fe80::fc54:ff:fe8c:ec85#123, interface stats: received=0, sent=0, dropped=0, active_time=408 secs
May 14 15:08:01 sv07 ntpd[2527]: peers refreshed
May 14 15:08:01 sv07 kernel: GFS2: fsid=: Trying to join cluster "lock_dlm", "jusme:vm1"
May 14 15:08:01 sv07 kernel: dlm: Using TCP for communications
May 14 15:08:01 sv07 kernel: dlm: got connection from 2
May 14 15:08:01 sv07 kernel: dlm: connecting to 2
May 14 15:08:01 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 61: monitor res_fs_vm1_monitor_60000 on sv06
May 14 15:08:01 sv07 kernel: GFS2: fsid=jusme:vm1.1: Joined cluster. Now mounting FS...
May 14 15:08:02 sv07 kernel: GFS2: fsid=jusme:vm1.1: jid=1, already locked for use
May 14 15:08:02 sv07 kernel: GFS2: fsid=jusme:vm1.1: jid=1: Looking at journal...
May 14 15:08:02 sv07 kernel: GFS2: fsid=jusme:vm1.1: jid=1: Done
May 14 15:08:02 sv07 crmd[3158]:   notice: process_lrm_event: LRM operation res_fs_vm1_start_0 (call=2517, rc=0, cib-update=1414, confirmed=true) ok
May 14 15:08:02 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 1: monitor res_fs_vm1_monitor_60000 on sv07 (local)
May 14 15:08:02 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 116: start res_vm_nfs_server_start_0 on sv07 (local)
May 14 15:08:02 sv07 crmd[3158]:   notice: process_lrm_event: LRM operation res_fs_vm1_monitor_60000 (call=2520, rc=0, cib-update=1415, confirmed=false) ok
May 14 15:08:02 sv07 VirtualDomain(res_vm_nfs_server)[29939]: INFO: Domain name "vm09" saved to /var/run/resource-agents/VirtualDomain-res_vm_nfs_server.state.
May 14 15:08:02 sv07 kernel: device vnet0 entered promiscuous mode
May 14 15:08:02 sv07 kernel: server_lan: port 2(vnet0) entering forwarding state
May 14 15:08:03 sv07 crmd[3158]:   notice: process_lrm_event: LRM operation res_vm_nfs_server_start_0 (call=2522, rc=0, cib-update=1416, confirmed=true) ok
May 14 15:08:03 sv07 crmd[3158]:   notice: te_rsc_command: Initiating action 117: monitor res_vm_nfs_server_monitor_60000 on sv07 (local)
May 14 15:08:03 sv07 crmd[3158]:   notice: process_lrm_event: LRM operation res_vm_nfs_server_monitor_60000 (call=2526, rc=0, cib-update=1417, confirmed=false) ok
May 14 15:08:03 sv07 crmd[3158]:   notice: run_graph: Transition 412 (Complete=31, Pending=0, Fired=0, Skipped=0, Incomplete=0, Source=/var/lib/pacemaker/pengine/pe-input-964.bz2): Complete
May 14 15:08:03 sv07 crmd[3158]:   notice: do_state_transition: State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_FSA_INTERNAL origin=notify_crmd ]
May 14 15:08:06 sv07 ntpd[2527]: Listen normally on 31 vnet0 fe80::fc54:ff:fe8c:ec85 UDP 123
May 14 15:08:06 sv07 ntpd[2527]: peers refreshed
May 14 15:08:12 sv07 kernel: kvm: 30038: cpu0 disabled perfctr wrmsr: 0xc1 data 0xabcd
May 14 15:10:11 sv07 root: Mark:pcmk:1400076611
May 14 15:10:18 sv07 crm_verify[32020]:   notice: crm_log_args: Invoked: crm_verify -V -x /root/report-20140514-1/sv07/cib.xml 
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ] ========== Checkpoint Information ===========
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ] global_ckpt_id: 30
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ] Checkpoint dlmplock.vm1 (12):
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    id:       29
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    sec cnt:  0
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    ref cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    unlinked: 0
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ] Checkpoint dlmplock.vm1 (12):
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    id:       28
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    sec cnt:  0
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    ref cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    unlinked: 1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ] Checkpoint dlmplock.vm1 (12):
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    id:       24
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    sec cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    ref cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    unlinked: 1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ] Checkpoint dlmplock.vm1 (12):
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    id:       23
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    sec cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    ref cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    unlinked: 1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ] Checkpoint dlmplock.vm3 (12):
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    id:       21
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    sec cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    ref cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    unlinked: 0
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ] Checkpoint dlmplock.vm2 (12):
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    id:       19
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    sec cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    ref cnt:  0
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    unlinked: 0
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ] Checkpoint dlmplock.vm1 (12):
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    id:       16
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    sec cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    ref cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    unlinked: 1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ] Checkpoint dlmplock.vm3 (12):
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    id:       15
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    sec cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    ref cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    unlinked: 1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ] Checkpoint dlmplock.vm3 (12):
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    id:       13
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    sec cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    ref cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    unlinked: 1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ] Checkpoint dlmplock.vm2 (12):
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    id:       10
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    sec cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    ref cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    unlinked: 1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ] Checkpoint dlmplock.vm1 (12):
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    id:       9
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    sec cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    ref cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    unlinked: 1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ] Checkpoint dlmplock.vm3 (12):
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    id:       7
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    sec cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    ref cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    unlinked: 1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ] Checkpoint dlmplock.vm1 (12):
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    id:       4
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    sec cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    ref cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    unlinked: 1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ] Checkpoint dlmplock.vm2 (12):
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    id:       2
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    sec cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    ref cnt:  1
May 14 15:10:18 sv07 corosync[2908]:   [CKPT  ]    unlinked: 1
