[ClusterLabs] crazy setup of fence_sbd, clusterd MD, nvmet
Andrei Borzenkov
arvidjaar at gmail.com
Tue Apr 15 18:04:06 UTC 2025
14.04.2025 19:43, Artem wrote:
> Dear gurus, I need your advice.
>
> We want to build a pacemaker cluster with the following resources.
> Could you please evaluate the idea and give feedback?
>
>
> Pairs of nodes with NVMe disks. Disks are shared from one node to
> another via nvmet. Persistent udev names and partition ids.
> MD raid1 is made on top of pairs of disks from different nodes. I
> suspect it must be clustered MD, and it'll require dlm?
> 2 or 4 clustered VLM volume groups are made on top of MD devices.
> Pacemaker location preference rules for half of VGs to one node and
> another half to another node.
>
> Striped LVs on top of VG with FS for Lustre MDT and OST. 2 main nodes
> in Corosync, other OST nodes are configured as remote resources.
>
> OS network is separate from iBMC, and firewall rules deny this
> traffic, so I decided to use SBD for fencing.
>
SBD requires a shared independent device. Using disks local to each
cluster node for SBD defeats its purpose.
>
> I only found some pieces of such a stack documented, different OS,
> different years ago. Now I'm trying to make it work together. At the
> moment the clustered MD cannot be created as it fails to create a
> lockspace (due to dlm error?). And dlm-clone doesn't want to start
> either on main nodes or (as it should) on remote nodes. OS = RHEL9.
>
> May be such setup is too complicated? I try to avoid split brain
> situations and uncoordinated writes by 2 mdadm processes on different
> nodes in all failure scenarios.
> I know that a common approach is to use JBODs of SAN arrays. But we
> don't have it for this project.
>
> Thanks in advance.
> Kindest regards,
> Artem
> _______________________________________________
> Manage your subscription:
> https://lists.clusterlabs.org/mailman/listinfo/users
>
> ClusterLabs home: https://www.clusterlabs.org/
More information about the Users
mailing list