[ClusterLabs] [Linux-HA] file system resource becomes inaccesible when any of the node goes down

Muhammad Sharfuddin M.Sharfuddin at nds.com.pk
Mon Jul 6 20:02:49 UTC 2015


On 07/06/2015 07:04 PM, Dejan Muhamedagic wrote:
> On Mon, Jul 06, 2015 at 03:14:34PM +0500, Muhammad Sharfuddin wrote:
>> On 07/06/2015 02:50 PM, Dejan Muhamedagic wrote:
>>> Hi,
>>>
>>> On Sun, Jul 05, 2015 at 09:13:56PM +0500, Muhammad Sharfuddin wrote:
>>>> SLES 11 SP 3 + online updates(pacemaker-1.1.11-0.8.11.70
>>>> openais-1.1.4-5.22.1.7)
>>>>
>>>> Its a dual primary drbd cluster, which mounts a file system resource
>>>> on both the cluster nodes simultaneously(file system type is ocfs2).
>>>>
>>>> Whenever any of the nodes goes down, the file system(/sharedata)
>>>> become inaccessible for exact 35 seconds on the other
>>>> (surviving/online) node, and then become available again on the
>>>> online node.
>>>>
>>>> Please help me understand why the node which survives or remains
>>>> online unable to access the file system resource(/sharedata) for 35
>>>> seconds ? and how can I fix the cluster so that file system remains
>>>> accessible on the surviving node without any interruption/delay(as
>>>> in my case of about 35 seconds)
>>>>
>>>> By inaccessible, I meant to say that running "ls -l /sharedata" and
>>>> "df /sharedata" does not return any output and does not return the
>>>> prompt back on the online node for exact 35 seconds once the other
>>>> node becomes offline.
>>>>
>>>> e.g "node1" got offline somewhere around  01:37:15, and then
>>>> /sharedata file system was inaccessible during 01:37:35 and 01:38:18
>>>> on the online node i.e "node2".
>>> Before the failing node gets fenced you won't be able to use the
>>> ocfs2 filesystem. In this case, the fencing operation takes 40
>>> seconds:
>> so its expected.
>>>> [...]
>>>> Jul  5 01:37:35 node2 sbd: [6197]: info: Writing reset to node slot node1
>>>> Jul  5 01:37:35 node2 sbd: [6197]: info: Messaging delay: 40
>>>> Jul  5 01:38:15 node2 sbd: [6197]: info: reset successfully
>>>> delivered to node1
>>>> Jul  5 01:38:15 node2 sbd: [6196]: info: Message successfully delivered.
>>>> [...]
>>> You may want to reduce that sbd timeout.
>> Ok, so reducing the sbd timeout(or msgwait) would provide the
>> uninterrupted access to the ocfs2 file system on the
>> surviving/online node ?
>> or would it just minimize the downtime ?
> Only the latter. But note that it is important that once sbd
> reports success, the target node is really down. sbd is
> timeout-based, i.e. it doesn't test whether the node actually
> left. Hence this timeout shouldn't be too short.
>
> Thanks,
>
> Dejan
>
>>> Thanks,
>>>
>>> Dejan
>>> _______________________________________________
>>>
>>>
>> --
>> Regards,
>>
>> Muhammad Sharfuddin
>> _______________________________________________
>>
>>

Now I did two changes
1 - remove the drbd, now there is a shared LUN
2 - reduce the watchdog and msgwait timeout values to 5 and 10 respectively.

Still when either of the node goes down delay of exact 15 seconds 
occurs. My concern/question,  when either of node goes down/offline, is 
there no way to uninterrupted access the ocfs2 file system resource  ?

at 00:43:44, node1 went offline, ocfs2 file system resource remained 
inaccessible exactly in between 00:44:00 and 00:44:14, /var/log/messages 
from node2:
Jul  7 00:43:55 node2 corosync[4361]:  [TOTEM ] A processor failed, 
forming new configuration.
Jul  7 00:44:01 node2 corosync[4361]:  [CLM   ] CLM CONFIGURATION CHANGE
Jul  7 00:44:01 node2 corosync[4361]:  [CLM   ] New Configuration:
Jul  7 00:44:01 node2 corosync[4361]:  [CLM   ]     r(0) ip(172.16.241.132)
Jul  7 00:44:01 node2 corosync[4361]:  [CLM   ] Members Left:
Jul  7 00:44:01 node2 corosync[4361]:  [CLM   ]     r(0) ip(172.16.241.131)
Jul  7 00:44:01 node2 corosync[4361]:  [CLM   ] Members Joined:
Jul  7 00:44:01 node2 corosync[4361]:  [pcmk  ] notice: 
pcmk_peer_update: Transitional membership event on ring 348: memb=1, 
new=0, lost=1
Jul  7 00:44:01 node2 corosync[4361]:  [pcmk  ] info: pcmk_peer_update: 
memb: node2 739307908
Jul  7 00:44:01 node2 corosync[4361]:  [pcmk  ] info: pcmk_peer_update: 
lost: node1 739307907
Jul  7 00:44:01 node2 corosync[4361]:  [CLM   ] CLM CONFIGURATION CHANGE
Jul  7 00:44:01 node2 corosync[4361]:  [CLM   ] New Configuration:
Jul  7 00:44:01 node2 corosync[4361]:  [CLM   ]     r(0) ip(172.16.241.132)
Jul  7 00:44:01 node2 cluster-dlm[4866]:   notice: 
plugin_handle_membership: Membership 348: quorum lost
Jul  7 00:44:01 node2 ocfs2_controld[4917]:   notice: 
plugin_handle_membership: Membership 348: quorum lost
Jul  7 00:44:01 node2 corosync[4361]:  [CLM   ] Members Left:
Jul  7 00:44:01 node2 cluster-dlm[4866]:   notice: 
crm_update_peer_state: plugin_handle_membership: Node node1[739307907] - 
state is now lost (was member)
Jul  7 00:44:01 node2 cluster-dlm[4866]: update_cluster: Processing 
membership 348
Jul  7 00:44:01 node2 cluster-dlm[4866]: del_configfs_node: 
del_configfs_node rmdir "/sys/kernel/config/dlm/cluster/comms/739307907"
Jul  7 00:44:01 node2 cluster-dlm[4866]: dlm_process_node: Removed 
inactive node 739307907: born-on=344, last-seen=344, this-event=348, 
last-event=344
Jul  7 00:44:01 node2 ocfs2_controld[4917]:   notice: 
crm_update_peer_state: plugin_handle_membership: Node node1[739307907] - 
state is now lost (was member)
Jul  7 00:44:01 node2 corosync[4361]:  [CLM   ] Members Joined:
Jul  7 00:44:01 node2 cluster-dlm[4866]: dlm_process_node: Skipped 
active node 739307908: born-on=336, last-seen=348, this-event=348, 
last-event=344
Jul  7 00:44:01 node2 stonith-ng[4367]:   notice: 
plugin_handle_membership: Membership 348: quorum lost
Jul  7 00:44:01 node2 cib[4366]:   notice: plugin_handle_membership: 
Membership 348: quorum lost
Jul  7 00:44:01 node2 crmd[4371]:   notice: plugin_handle_membership: 
Membership 348: quorum lost
Jul  7 00:44:01 node2 ocfs2_controld[4917]: confchg called
Jul  7 00:44:01 node2 corosync[4361]:  [pcmk  ] notice: 
pcmk_peer_update: Stable membership event on ring 348: memb=1, new=0, lost=0
Jul  7 00:44:01 node2 cluster-dlm[4866]: log_config: dlm:controld conf 1 
0 1 memb 739307908 join left 739307907
Jul  7 00:44:01 node2 stonith-ng[4367]:   notice: crm_update_peer_state: 
plugin_handle_membership: Node node1[739307907] - state is now lost (was 
member)
Jul  7 00:44:01 node2 cib[4366]:   notice: crm_update_peer_state: 
plugin_handle_membership: Node node1[739307907] - state is now lost (was 
member)
Jul  7 00:44:01 node2 crmd[4371]:  warning: match_down_event: No match 
for shutdown action on node1
Jul  7 00:44:01 node2 ocfs2_controld[4917]: ocfs2_controld (group 
"ocfs2:controld") confchg: members 1, left 1, joined 0
Jul  7 00:44:01 node2 corosync[4361]:  [pcmk  ] info: pcmk_peer_update: 
MEMB: node2 739307908
Jul  7 00:44:01 node2 cluster-dlm[4866]: log_config: 
dlm:ls:E4C27EC365B44D03B3629077F1B31722 conf 1 0 1 memb 739307908 join 
left 739307907
Jul  7 00:44:01 node2 crmd[4371]:   notice: peer_update_callback: 
Stonith/shutdown of node1 not matched
Jul  7 00:44:01 node2 ocfs2_controld[4917]: node daemon left 739307907
Jul  7 00:44:01 node2 ocfs2_controld[4917]: node down 739307907
Jul  7 00:44:01 node2 corosync[4361]:  [pcmk  ] info: 
ais_mark_unseen_peer_dead: Node node1 was not seen in the previous 
transition
Jul  7 00:44:01 node2 cluster-dlm[4866]: node_history_fail: 
E4C27EC365B44D03B3629077F1B31722 check_fs nodeid 739307907 set
Jul  7 00:44:01 node2 crmd[4371]:   notice: crm_update_peer_state: 
plugin_handle_membership: Node node1[739307907] - state is now lost (was 
member)
Jul  7 00:44:01 node2 ocfs2_controld[4917]: Node 739307907 has left 
mountgroup E4C27EC365B44D03B3629077F1B31722
Jul  7 00:44:01 node2 corosync[4361]:  [pcmk  ] info: update_member: 
Node 739307907/node1 is now: lost
Jul  7 00:44:01 node2 corosync[4361]:  [pcmk  ] info: 
send_member_notification: Sending membership update 348 to 5 children
Jul  7 00:44:01 node2 cluster-dlm[4866]: add_change: 
E4C27EC365B44D03B3629077F1B31722 add_change cg 2 remove nodeid 739307907 
reason 3
Jul  7 00:44:01 node2 crmd[4371]:  warning: match_down_event: No match 
for shutdown action on node1
Jul  7 00:44:01 node2 corosync[4361]:  [TOTEM ] A processor joined or 
left the membership and a new membership was formed.
Jul  7 00:44:01 node2 cluster-dlm[4866]: add_change: 
E4C27EC365B44D03B3629077F1B31722 add_change cg 2 counts member 1 joined 
0 remove 1 failed 1
Jul  7 00:44:01 node2 cluster-dlm[4866]: stop_kernel: 
E4C27EC365B44D03B3629077F1B31722 stop_kernel cg 2
Jul  7 00:44:01 node2 crmd[4371]:   notice: peer_update_callback: 
Stonith/shutdown of node1 not matched
Jul  7 00:44:01 node2 corosync[4361]:  [CPG   ] chosen downlist: sender 
r(0) ip(172.16.241.132) ; members(old:2 left:1)
Jul  7 00:44:01 node2 cluster-dlm[4866]: do_sysfs: write "0" to 
"/sys/kernel/dlm/E4C27EC365B44D03B3629077F1B31722/control"
Jul  7 00:44:01 node2 sbd: [4349]: WARN: CIB: We do NOT have quorum!
Jul  7 00:44:01 node2 sbd: [4347]: WARN: Pacemaker health check: UNHEALTHY
Jul  7 00:44:01 node2 kernel: [ 2362.087877] dlm: closing connection to 
node 739307907
Jul  7 00:44:01 node2 corosync[4361]:  [MAIN  ] Completed service 
synchronization, ready to provide service.
Jul  7 00:44:01 node2 cluster-dlm[4866]: fence_node_time: Node 
739307907/node1 has not been shot yet
Jul  7 00:44:01 node2 ocfs2_controld[4917]: Sending notification of node 
739307907 for "E4C27EC365B44D03B3629077F1B31722"
Jul  7 00:44:01 node2 cluster-dlm[4866]: check_fencing_done: 
E4C27EC365B44D03B3629077F1B31722 check_fencing 739307907 wait add 
1436211736 fail 1436211841 last 0
Jul  7 00:44:01 node2 cluster-dlm[4866]: set_fs_notified: 
E4C27EC365B44D03B3629077F1B31722 set_fs_notified nodeid 739307907
Jul  7 00:44:01 node2 ocfs2_controld[4917]: confchg called
Jul  7 00:44:01 node2 crmd[4371]:   notice: crm_update_quorum: Updating 
quorum status to false (call=645)
Jul  7 00:44:01 node2 ocfs2_controld[4917]: group 
"ocfs2:E4C27EC365B44D03B3629077F1B31722" confchg: members 1, left 1, 
joined 0
Jul  7 00:44:01 node2 crmd[4371]:   notice: do_state_transition: State 
transition S_IDLE -> S_INTEGRATION [ input=I_NODE_JOIN 
cause=C_FSA_INTERNAL origin=check_join_state ]
Jul  7 00:44:01 node2 ocfs2_controld[4917]: message from dlmcontrol
Jul  7 00:44:01 node2 crmd[4371]:   notice: crm_update_quorum: Updating 
quorum status to false (call=650)
Jul  7 00:44:01 node2 attrd[4369]:   notice: attrd_local_callback: 
Sending full refresh (origin=crmd)
Jul  7 00:44:01 node2 ocfs2_controld[4917]: Notified for 
"E4C27EC365B44D03B3629077F1B31722", node 739307907, status 0
Jul  7 00:44:01 node2 pengine[4370]:   notice: unpack_config: On loss of 
CCM Quorum: Ignore
Jul  7 00:44:01 node2 attrd[4369]:   notice: attrd_trigger_update: 
Sending flush op to all hosts for: last-failure-p-sapfs (1436211473)
Jul  7 00:44:01 node2 ocfs2_controld[4917]: Completing notification on 
"E4C27EC365B44D03B3629077F1B31722" for node 739307907
Jul  7 00:44:01 node2 pengine[4370]:  warning: pe_fence_node: Node node1 
will be fenced because the node is no longer part of the cluster
Jul  7 00:44:01 node2 attrd[4369]:   notice: attrd_trigger_update: 
Sending flush op to all hosts for: shutdown (0)
Jul  7 00:44:01 node2 pengine[4370]:  warning: determine_online_status: 
Node node1 is unclean
Jul  7 00:44:01 node2 attrd[4369]:   notice: attrd_trigger_update: 
Sending flush op to all hosts for: probe_complete (true)
Jul  7 00:44:01 node2 pengine[4370]:  warning: custom_action: Action 
p-dlm:0_stop_0 on node1 is unrunnable (offline)
Jul  7 00:44:01 node2 pengine[4370]:  warning: custom_action: Action 
p-dlm:0_stop_0 on node1 is unrunnable (offline)
Jul  7 00:44:01 node2 pengine[4370]:  warning: custom_action: Action 
p-o2cb:0_stop_0 on node1 is unrunnable (offline)
Jul  7 00:44:01 node2 pengine[4370]:  warning: custom_action: Action 
p-o2cb:0_stop_0 on node1 is unrunnable (offline)
Jul  7 00:44:01 node2 pengine[4370]:  warning: custom_action: Action 
p-vgsap:0_stop_0 on node1 is unrunnable (offline)
Jul  7 00:44:01 node2 pengine[4370]:  warning: custom_action: Action 
p-vgsap:0_stop_0 on node1 is unrunnable (offline)
Jul  7 00:44:01 node2 pengine[4370]:  warning: custom_action: Action 
p-sapfs:0_stop_0 on node1 is unrunnable (offline)
Jul  7 00:44:01 node2 pengine[4370]:  warning: custom_action: Action 
p-sapfs:0_stop_0 on node1 is unrunnable (offline)
Jul  7 00:44:01 node2 pengine[4370]:  warning: stage6: Scheduling Node 
node1 for STONITH
Jul  7 00:44:01 node2 pengine[4370]:   notice: LogActions: Stop 
p-dlm:0    (node1)
Jul  7 00:44:01 node2 pengine[4370]:   notice: LogActions: Stop 
p-o2cb:0    (node1)
Jul  7 00:44:01 node2 pengine[4370]:   notice: LogActions: Stop 
p-vgsap:0    (node1)
Jul  7 00:44:01 node2 pengine[4370]:   notice: LogActions: Stop 
p-sapfs:0    (node1)
Jul  7 00:44:01 node2 crmd[4371]:   notice: do_te_invoke: Processing 
graph 37 (ref=pe_calc-dc-1436211841-296) derived from 
/var/lib/pacemaker/pengine/pe-warn-15.bz2
Jul  7 00:44:01 node2 pengine[4370]:  warning: process_pe_message: 
Calculated Transition 37: /var/lib/pacemaker/pengine/pe-warn-15.bz2
Jul  7 00:44:01 node2 crmd[4371]:   notice: te_fence_node: Executing 
reboot fencing operation (48) on node1 (timeout=90000)
Jul  7 00:44:01 node2 stonith-ng[4367]:   notice: handle_request: Client 
crmd.4371.43345a93 wants to fence (reboot) 'node1' with device '(any)'
Jul  7 00:44:01 node2 stonith-ng[4367]:   notice: 
initiate_remote_stonith_op: Initiating remote operation reboot for 
node1: b140f322-d273-4de3-b07f-37b8f5939aff (0)
Jul  7 00:44:01 node2 stonith-ng[4367]:   notice: 
can_fence_host_with_device: sbd_stonith can fence node1: dynamic-list
Jul  7 00:44:01 node2 sbd: [6039]: info: Delivery process handling 
/dev/disk/by-id/scsi-1494554000000000035c71550e83586b70dfc77c3b382dfbd
Jul  7 00:44:01 node2 sbd: [6039]: info: Device UUID: 
5ed06971-5e18-4ab5-9968-620c61a71269
Jul  7 00:44:01 node2 sbd: [6039]: info: Writing reset to node slot node1
Jul  7 00:44:01 node2 sbd: [6039]: info: Messaging delay: 10
Jul  7 00:44:11 node2 sbd: [6039]: info: reset successfully delivered to 
node1
Jul  7 00:44:11 node2 sbd: [6038]: info: Message successfully delivered.
Jul  7 00:44:12 node2 stonith-ng[4367]:   notice: log_operation: 
Operation 'reboot' [6027] (call 2 from crmd.4371) for host 'node1' with 
device 'sbd_stonith' returned: 0 (OK)
Jul  7 00:44:12 node2 stonith-ng[4367]:   notice: remote_op_done: 
Operation reboot of node1 by node2 for crmd.4371 at node2.b140f322: OK
Jul  7 00:44:12 node2 crmd[4371]:   notice: tengine_stonith_callback: 
Stonith operation 2/48:37:0:f3061adb-af45-4c93-9365-bae0cf01a847: OK (0)
Jul  7 00:44:12 node2 crmd[4371]:   notice: tengine_stonith_notify: Peer 
node1 was terminated (reboot) by node2 for node2: OK 
(ref=b140f322-d273-4de3-b07f-37b8f5939aff) by client crmd.4371
Jul  7 00:44:12 node2 crmd[4371]:   notice: run_graph: Transition 37 
(Complete=15, Pending=0, Fired=0, Skipped=0, Incomplete=0, 
Source=/var/lib/pacemaker/pengine/pe-warn-15.bz2): Complete
Jul  7 00:44:12 node2 crmd[4371]:   notice: do_state_transition: State 
transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS 
cause=C_FSA_INTERNAL origin=notify_crmd ]
Jul  7 00:44:12 node2 crmd[4371]:   notice: do_state_transition: State 
transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC 
cause=C_FSA_INTERNAL origin=abort_transition_graph ]
Jul  7 00:44:12 node2 pengine[4370]:   notice: unpack_config: On loss of 
CCM Quorum: Ignore
Jul  7 00:44:12 node2 pengine[4370]:   notice: process_pe_message: 
Calculated Transition 38: /var/lib/pacemaker/pengine/pe-input-154.bz2
Jul  7 00:44:12 node2 crmd[4371]:   notice: do_te_invoke: Processing 
graph 38 (ref=pe_calc-dc-1436211852-297) derived from 
/var/lib/pacemaker/pengine/pe-input-154.bz2
Jul  7 00:44:12 node2 crmd[4371]:   notice: run_graph: Transition 38 
(Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0, 
Source=/var/lib/pacemaker/pengine/pe-input-154.bz2): Complete
Jul  7 00:44:12 node2 crmd[4371]:   notice: do_state_transition: State 
transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS 
cause=C_FSA_INTERNAL origin=notify_crmd ]
Jul  7 00:44:13 node2 cluster-dlm[4866]: fence_node_time: Node 
739307907/node1 was last shot 'now'
Jul  7 00:44:13 node2 cluster-dlm[4866]: check_fencing_done: 
E4C27EC365B44D03B3629077F1B31722 check_fencing 739307907 done add 
1436211736 fail 1436211841 last 1436211853
Jul  7 00:44:13 node2 cluster-dlm[4866]: check_fencing_done: 
E4C27EC365B44D03B3629077F1B31722 check_fencing done
Jul  7 00:44:13 node2 cluster-dlm[4866]: check_quorum_done: 
E4C27EC365B44D03B3629077F1B31722 check_quorum disabled
Jul  7 00:44:13 node2 cluster-dlm[4866]: check_fs_done: 
E4C27EC365B44D03B3629077F1B31722 check_fs nodeid 739307907 clear
Jul  7 00:44:13 node2 cluster-dlm[4866]: check_fs_done: 
E4C27EC365B44D03B3629077F1B31722 check_fs done
Jul  7 00:44:13 node2 cluster-dlm[4866]: send_info: 
E4C27EC365B44D03B3629077F1B31722 send_start cg 2 flags 2 data2 0 counts 
1 1 0 1 1
Jul  7 00:44:13 node2 cluster-dlm[4866]: receive_start: 
E4C27EC365B44D03B3629077F1B31722 receive_start 739307908:2 len 76
Jul  7 00:44:13 node2 cluster-dlm[4866]: match_change: 
E4C27EC365B44D03B3629077F1B31722 match_change 739307908:2 matches cg 2
Jul  7 00:44:13 node2 cluster-dlm[4866]: wait_messages_done: 
E4C27EC365B44D03B3629077F1B31722 wait_messages cg 2 got all 1
Jul  7 00:44:13 node2 cluster-dlm[4866]: start_kernel: 
E4C27EC365B44D03B3629077F1B31722 start_kernel cg 2 member_count 1
Jul  7 00:44:13 node2 cluster-dlm[4866]: update_dir_members: dir_member 
739307908
Jul  7 00:44:13 node2 cluster-dlm[4866]: update_dir_members: dir_member 
739307907
Jul  7 00:44:13 node2 cluster-dlm[4866]: set_configfs_members: 
set_members rmdir 
"/sys/kernel/config/dlm/cluster/spaces/E4C27EC365B44D03B3629077F1B31722/nodes/739307907"
Jul  7 00:44:13 node2 cluster-dlm[4866]: do_sysfs: write "1" to 
"/sys/kernel/dlm/E4C27EC365B44D03B3629077F1B31722/control"
Jul  7 00:44:13 node2 kernel: [ 2374.125296] 
(ocfs2rec,6017,0):ocfs2_replay_journal:1549 Recovering node 739307907 
from slot 0 on device (253,0)
Jul  7 00:44:13 node2 kernel: [ 2374.429695] 
(ocfs2rec,6017,0):ocfs2_begin_quota_recovery:407 Beginning quota 
recovery in slot 0
Jul  7 00:44:13 node2 kernel: [ 2374.444718] 
(kworker/u:28,139,0):ocfs2_finish_quota_recovery:599 Finishing quota 
recovery in slot 0

cib:
primitive p-dlm ocf:pacemaker:controld \
         op monitor interval="120" timeout="30" \
         op start interval="0" timeout="90" \
         op stop interval="0" timeout="100"
primitive p-o2cb ocf:ocfs2:o2cb \
         op monitor interval="120" timeout="30" \
         op start interval="0" timeout="90" \
         op stop interval="0" timeout="100"
primitive p-sapfs ocf:heartbeat:Filesystem \
         params device="/dev/vgsap/lvsap" directory="/sapdata" 
fstype="ocfs2" \
         op monitor interval="60s" timeout="60s" \
         op start interval="0" timeout="90s" \
         op stop interval="0" timeout="90s"
primitive p-vgsap ocf:heartbeat:LVM \
         params volgrpname="vgsap" \
         op monitor interval="60s" timeout="40" \
         op start interval="0" timeout="40" \
         op stop interval="0" timeout="40"
primitive sbd_stonith stonith:external/sbd \
         meta target-role="Started" \
         op monitor interval="3000" timeout="20" \
         op start interval="0" timeout="20" \
         op stop interval="0" timeout="20" \
         params 
sbd_device="/dev/disk/by-id/scsi-1494554000000000035c71550e83586b70dfc77c3b382dfbd"
group g-lock p-dlm p-o2cb
group g-sapdata p-vgsap p-sapfs
clone cl-lock g-lock \
         meta globally-unique="false" interleave="true"
clone cl-sapdata g-sapdata \
         meta globally-unique="false" interleave="true" 
target-role="Started"
colocation col-lock-sapdata inf: cl-sapdata cl-lock
order ord-lock-sapdata inf: cl-lock cl-sapdata


-- 
Regards,

Muhammad Sharfuddin
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/users/attachments/20150707/a2bd0e2d/attachment.htm>


More information about the Users mailing list