[ClusterLabs] SBD with shared block storage (and watchdog?)
durwin at mgtsciences.com
durwin at mgtsciences.com
Mon Feb 13 13:34:39 EST 2017
emmanuel segura <emi2fast at gmail.com> wrote on 02/13/2017 10:55:58 AM:
> From: emmanuel segura <emi2fast at gmail.com>
> To: Cluster Labs - All topics related to open-source clustering
> welcomed <users at clusterlabs.org>
> Date: 02/13/2017 10:56 AM
> Subject: Re: [ClusterLabs] SBD with shared block storage (and watchdog?)
>
> modprobe softdog if you don't have an external watchdog
Thank you, that made sbd watch happy.
I now have this running on the 2 nodes.
11:10 AM root at node1 ~
fc25> ps aux|grep sbd
root 24426 0.0 0.6 97888 13848 pts/0 SL 11:00 0:00 sbd:
inquisitor
root 24427 0.0 0.6 97892 13988 pts/0 SL 11:00 0:00 sbd:
watcher: /dev/sdb1 - slot: 0 - uuid: 6094f0f4-2a07-47db-b4f7-6d478464d56a
root 24428 0.0 0.8 102476 18404 pts/0 SL 11:00 0:00 sbd:
watcher: Pacemaker
root 29442 0.0 0.0 118520 1000 pts/0 S+ 11:18 0:00 grep
--color=auto sbd
11:18 AM root at node2 ~
fc25> ps aux|grep sbd
root 22784 0.0 0.6 97884 13844 pts/0 SL 11:18 0:00 sbd:
inquisitor
root 22785 0.0 0.6 97888 13984 pts/0 SL 11:18 0:00 sbd:
watcher: /dev/sdb1 - slot: 1 - uuid: 6094f0f4-2a07-47db-b4f7-6d478464d56a
root 22786 0.0 0.8 102472 18400 pts/0 SL 11:18 0:00 sbd:
watcher: Pacemaker
root 22789 0.0 0.0 118520 952 pts/0 S+ 11:18 0:00 grep
--color=auto sbd
Is the fencing complete?
If so, will 'pcs cluster standby' simulate node failure?
Addressing email that has followed.
The device sbd1 is solely for sbd. Sounds like you're saying it does not
get mounted.
Is that correct? If I just unmount sbd1 will all be ok?
How small can I make the sbd block device?
Thank you,
Durwin
>
> 2017-02-13 18:34 GMT+01:00 <durwin at mgtsciences.com>:
> > I am working to get an active/active cluster running.
> > I have Windows 10 running 2 Fedora 25 Virtualbox VMs.
> > VMs named node1, and node2.
> >
> > I created a vdi disk and set it to shared.
> > I formatted it to gfs2 with this command.
> >
> > mkfs.gfs2 -t msicluster:msigfs2 -j 2 /dev/sdb1
> >
> > After installing 'dlm' and insuring guest additions were
> > installed, I was able to mount the gfs2 parition.
> >
> > I then followed.
> >
> > https://github.com/l-mb/sbd/blob/master/man/sbd.8.pod
> >
> > I used this command.
> >
> > sbd -d /dev/sdb1 create
> >
> > Using sbd to 'list' returns nothing, but 'dump' shows this.
> >
> > fc25> sbd -d /dev/sdb1 dump
> > ==Dumping header on disk /dev/sdb1
> > Header version : 2.1
> > UUID : 6094f0f4-2a07-47db-b4f7-6d478464d56a
> > Number of slots : 255
> > Sector size : 512
> > Timeout (watchdog) : 5
> > Timeout (allocate) : 2
> > Timeout (loop) : 1
> > Timeout (msgwait) : 10
> > ==Header on disk /dev/sdb1 is dumped
> >
> > I then tried the 'watch' command and journalctl shows error listed.
> >
> > sbd -d /dev/sdb1 -W -P watch
> >
> > Feb 13 09:54:09 node1 sbd[6908]: error: watchdog_init: Cannot open
> > watchdog device '/dev/watchdog': No such file or directory (2)
> > Feb 13 09:54:09 node1 sbd[6908]: warning: cleanup_servant_by_pid:
Servant
> > for pcmk (pid: 6910) has terminated
> > Feb 13 09:54:09 node1 sbd[6908]: warning: cleanup_servant_by_pid:
Servant
> > for /dev/sdb1 (pid: 6909) has terminated
> >
> >
> > From
> >
> > http://blog.clusterlabs.org/blog/2015/sbd-fun-and-profit
> >
> > I installed watchdog.
> >
> > my /etc/sysconfig/sbd is.
> >
> > SBD_DELAY_START=no
> > SBD_OPTS=
> > SBD_PACEMAKER=yes
> > SBD_STARTMODE=clean
> > SBD_WATCHDOG_DEV=/dev/watchdog
> > SBD_WATCHDOG_TIMEOUT=5
> >
> > the sbd-fun-and-profit says to use this command.
> >
> > virsh edit vmnode
> >
> > But there is no vmnode and no instructions on how to create it.
> >
> > Is anyone able to piece together the missing steps?
> >
> >
> > Thank you.
> >
> > Durwin F. De La Rue
> > Management Sciences, Inc.
> > 6022 Constitution Ave. NE
> > Albuquerque, NM 87110
> > Phone (505) 255-8611
> >
> >
> > This email message and any attachments are for the sole use of the
intended
> > recipient(s) and may contain proprietary and/or confidential
information
> > which may be privileged or otherwise protected from disclosure. Any
> > unauthorized review, use, disclosure or distribution is prohibited. If
you
> > are not the intended recipient(s), please contact the sender by reply
email
> > and destroy the original message and any copies of the message as well
as
> > any attachments to the original message.
> > _______________________________________________
> > Users mailing list: Users at clusterlabs.org
> > http://lists.clusterlabs.org/mailman/listinfo/users
> >
> > Project Home: http://www.clusterlabs.org
> > Getting started:
http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> > Bugs: http://bugs.clusterlabs.org
> >
>
>
>
> --
> .~.
> /V\
> // \\
> /( )\
> ^`~'^
>
> _______________________________________________
> Users mailing list: Users at clusterlabs.org
> http://lists.clusterlabs.org/mailman/listinfo/users
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
This email message and any attachments are for the sole use of the
intended recipient(s) and may contain proprietary and/or confidential
information which may be privileged or otherwise protected from
disclosure. Any unauthorized review, use, disclosure or distribution is
prohibited. If you are not the intended recipient(s), please contact the
sender by reply email and destroy the original message and any copies of
the message as well as any attachments to the original message.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/users/attachments/20170213/9dbaee5d/attachment-0003.html>
More information about the Users
mailing list