[ClusterLabs] CentOS 8 & drbd 9 second slave is not started
fatcharly at gmx.de
fatcharly at gmx.de
Mon Jan 18 10:52:00 EST 2021
Hi,
I'm trying to install a 2 node pacemaker/drbd cluster on a CentOS 8.3. I'm using this versions:
kmod-drbd90-9.0.25-2.el8_3.elrepo.x86_64
drbd90-utils-9.13.1-1.el8.elrepo.x86_64
pacemaker-cluster-libs-2.0.4-6.el8.x86_64
pacemaker-cli-2.0.4-6.el8.x86_64
pacemaker-schemas-2.0.4-6.el8.noarch
pacemaker-2.0.4-6.el8.x86_64
pacemaker-libs-2.0.4-6.el8.x86_64
clusternode-names are lisbon and susanne
There are two drbd-resources configured which are paired with filesystem-resources. Both are working in simple master/slave configuration.
When I start up the cluster one resource is promoted to a master and a slave, the other is just getting on one the master started, but the slave is not starting up.
Status of the cluster:
Cluster Summary:
* Stack: corosync
* Current DC: lisbon (version 2.0.4-6.el8-2deceaa3ae) - partition with quorum
* Last updated: Mon Jan 18 16:30:21 2021
* Last change: Mon Jan 18 16:30:17 2021 by root via cibadmin on lisbon
* 2 nodes configured
* 7 resource instances configured
Node List:
* Online: [ lisbon susanne ]
Active Resources:
* HA-IP_1 (ocf::heartbeat:IPaddr2): Started susanne
* Clone Set: drbd_database-clone [drbd_database] (promotable):
* Masters: [ susanne ]
* Slaves: [ lisbon ]
* fs_database (ocf::heartbeat:Filesystem): Started susanne
* Clone Set: drbd_logsfiles-clone [drbd_logsfiles] (promotable):
* Masters: [ susanne ]
* fs_logfiles (ocf::heartbeat:Filesystem): Started susanne
drbdadm status
[root at susanne ~]# drbdadm status
drbd1 role:Primary
disk:UpToDate
lisbon role:Secondary
peer-disk:UpToDate
drbd2 role:Primary
disk:UpToDate
lisbon connection:Connecting
[root at lisbon ~]# drbdadm status
drbd1 role:Secondary
disk:UpToDate
susanne role:Primary
peer-disk:UpToDate
cluster-config:
Cluster Name: mysql_cluster
Corosync Nodes:
susanne lisbon
Pacemaker Nodes:
lisbon susanne
Resources:
Resource: HA-IP_1 (class=ocf provider=heartbeat type=IPaddr2)
Attributes: cidr_netmask=24 ip=192.168.18.150
Operations: monitor interval=15s (HA-IP_1-monitor-interval-15s)
start interval=0s timeout=20s (HA-IP_1-start-interval-0s)
stop interval=0s timeout=20s (HA-IP_1-stop-interval-0s)
Clone: drbd_database-clone
Meta Attrs: clone-max=2 clone-node-max=1 notify=true promotable=true promoted-max=1 promoted-node-max=1
Resource: drbd_database (class=ocf provider=linbit type=drbd)
Attributes: drbd_resource=drbd1
Operations: demote interval=0s timeout=90 (drbd_database-demote-interval-0s)
monitor interval=20 role=Slave timeout=20 (drbd_database-monitor-interval-20)
monitor interval=10 role=Master timeout=20 (drbd_database-monitor-interval-10)
notify interval=0s timeout=90 (drbd_database-notify-interval-0s)
promote interval=0s timeout=90 (drbd_database-promote-interval-0s)
reload interval=0s timeout=30 (drbd_database-reload-interval-0s)
start interval=0s timeout=240 (drbd_database-start-interval-0s)
stop interval=0s timeout=100 (drbd_database-stop-interval-0s)
Resource: fs_database (class=ocf provider=heartbeat type=Filesystem)
Attributes: device=/dev/drbd1 directory=/mnt/clusterfs1 fstype=ext4
Operations: monitor interval=20s timeout=40s (fs_database-monitor-interval-20s)
start interval=0s timeout=60s (fs_database-start-interval-0s)
stop interval=0s timeout=60s (fs_database-stop-interval-0s)
Clone: drbd_logsfiles-clone
Meta Attrs: clone-max=2 clone-node-max=1 notify=true promotable=true promoted-max=1 promoted-node-max=1
Resource: drbd_logsfiles (class=ocf provider=linbit type=drbd)
Attributes: drbd_resource=drbd2
Operations: demote interval=0s timeout=90 (drbd_logsfiles-demote-interval-0s)
monitor interval=20 role=Slave timeout=20 (drbd_logsfiles-monitor-interval-20)
monitor interval=10 role=Master timeout=20 (drbd_logsfiles-monitor-interval-10)
notify interval=0s timeout=90 (drbd_logsfiles-notify-interval-0s)
promote interval=0s timeout=90 (drbd_logsfiles-promote-interval-0s)
reload interval=0s timeout=30 (drbd_logsfiles-reload-interval-0s)
start interval=0s timeout=240 (drbd_logsfiles-start-interval-0s)
stop interval=0s timeout=100 (drbd_logsfiles-stop-interval-0s)
Resource: fs_logfiles (class=ocf provider=heartbeat type=Filesystem)
Attributes: device=/dev/drbd2 directory=/mnt/clusterfs2 fstype=ext4
Operations: monitor interval=20s timeout=40s (fs_logfiles-monitor-interval-20s)
start interval=0s timeout=60s (fs_logfiles-start-interval-0s)
stop interval=0s timeout=60s (fs_logfiles-stop-interval-0s)
Stonith Devices:
Fencing Levels:
Location Constraints:
Ordering Constraints:
start drbd_database-clone then start fs_database (kind:Mandatory) (id:order-drbd_database-clone-fs_database-mandatory)
start drbd_logsfiles-clone then start fs_logfiles (kind:Mandatory) (id:order-drbd_logsfiles-clone-fs_logfiles-mandatory)
Colocation Constraints:
fs_database with drbd_database-clone (score:INFINITY) (with-rsc-role:Master) (id:colocation-fs_database-drbd_database-clone-INFINITY)
fs_logfiles with drbd_logsfiles-clone (score:INFINITY) (with-rsc-role:Master) (id:colocation-fs_logfiles-drbd_logsfiles-clone-INFINITY)
drbd_logsfiles-clone with drbd_database-clone (score:INFINITY) (with-rsc-role:Master) (id:colocation-drbd_logsfiles-clone-drbd_database-clone-INFINITY)
Ticket Constraints:
Alerts:
No alerts defined
Resources Defaults:
No defaults set
Operations Defaults:
No defaults set
Cluster Properties:
cluster-infrastructure: corosync
cluster-name: mysql_cluster
dc-version: 2.0.4-6.el8-2deceaa3ae
have-watchdog: false
last-lrm-refresh: 1610382881
stonith-enabled: false
Tags:
No tags defined
Quorum:
Options:
Any suggestions are welcome
stay safe and healty
fatcharly
More information about the Users
mailing list