[ClusterLabs] drbd9 3 node "neverending" sync?
Mevo Govo
govomevo at gmail.com
Mon Oct 8 08:12:13 EDT 2018
Hi,
I tried it with and without drbd quorum options, I resterted all the node,
but same "neverending" refresh. Device is not mounted, so no "outer"
modification. protocol is C (synchron). I recreated the resource on node 3
(xetest3) with tis commands:
drbdadm down r0
dd if=/dev/zero of=/dev/sdb1 bs=100M count=1
drbdadm create-md r0
After this "full resync", the neverending sync stopped (UpToDate on every
nodes). But after a resource down/up, the "Outdated/Inconsistent/UpToDate"
cycle is restarted on xetest3.
Maybe it is normal on 3rd node, even if protocol C. But I do not see it on
our prod environment.
B.r: Laszlo.
Mevo Govo <govomevo at gmail.com> ezt írta (időpont: 2018. okt. 8., H, 11:20):
> Hi,
>
> we are testing drbd9 with 3 node. After a test (network disable, then
> enable on one host), the syncronization is not finishing.
> As a recovery, I switched the pacemaker off, and I tried "primary --force"
> on one node and "--discard-my-data" on the other nodes. It seem solve the
> problem (everything is UpToDate), but after I down/up the resource, the
> "neverending" synchronization is restarted (when the resource is Secondary
> on all the 3 node).
> Why sync is restarted after UpToDate state. Ho can I solve this problem?
> Nodes are xetes1, xetest2, xetest3, here is a checking test ("drbdadm
> status" on every node, then wait 10 secod. This repeated 10 times)
>
> Thanks for answers: Laci.
>
>
> for i in {1..10}
> do
> echo "################## check $i ##################"
> sleep 10
> date
> echo '################## on node xetest1'
> ssh xetest1 sudo drbdadm status
> echo '################## on node xetest2'
> ssh xetest2 sudo drbdadm status
> echo '################## on node xetest3'
> ssh xetest3 sudo drbdadm status
> done
>
> ################## check 1 ##################
>
>
>
> Mon Oct 8 09:11:12 GMT 2018
> ################## on node xetest1
> r0 role:Secondary
> disk:UpToDate
> xetest2 role:Secondary
> peer-disk:UpToDate
> xetest3 role:Secondary
> peer-disk:UpToDate
>
> ################## on node xetest2
> r0 role:Secondary
> disk:UpToDate
> xetest1 role:Secondary
> peer-disk:UpToDate
> xetest3 role:Secondary
> peer-disk:Inconsistent resync-suspended:peer
>
> ################## on node xetest3
> r0 role:Secondary
> disk:Outdated blocked:upper
> xetest1 role:Secondary
> replication:WFBitMapT peer-disk:UpToDate
> xetest2 role:Secondary
> peer-disk:UpToDate
>
> ################## check 2 ##################
> Mon Oct 8 09:11:23 GMT 2018
> ################## on node xetest1
> r0 role:Secondary
> disk:UpToDate
> xetest2 role:Secondary
> peer-disk:UpToDate
> xetest3 role:Secondary
> peer-disk:Inconsistent resync-suspended:peer
>
> ################## on node xetest2
> r0 role:Secondary
> disk:UpToDate
> xetest1 role:Secondary
> peer-disk:UpToDate
> xetest3 role:Secondary
> replication:SyncSource peer-disk:Inconsistent
>
> ################## on node xetest3
> r0 role:Secondary
> disk:Outdated blocked:upper
> xetest1 role:Secondary
> peer-disk:UpToDate
> xetest2 role:Secondary
> replication:WFBitMapT peer-disk:UpToDate
>
> ################## check 3 ##################
> Mon Oct 8 09:11:35 GMT 2018
> ################## on node xetest1
> r0 role:Secondary
> disk:UpToDate
> xetest2 role:Secondary
> peer-disk:UpToDate
> xetest3 role:Secondary
> replication:SyncSource peer-disk:Inconsistent
>
> ################## on node xetest2
> r0 role:Secondary
> disk:UpToDate
> xetest1 role:Secondary
> peer-disk:UpToDate
> xetest3 role:Secondary
> replication:SyncSource peer-disk:Inconsistent
>
> ################## on node xetest3
> r0 role:Secondary
> disk:UpToDate
> xetest1 role:Secondary
> peer-disk:UpToDate
> xetest2 role:Secondary
> peer-disk:UpToDate
>
> ################## check 4 ##################
> Mon Oct 8 09:11:46 GMT 2018
> ################## on node xetest1
> r0 role:Secondary
> disk:UpToDate
> xetest2 role:Secondary
> peer-disk:UpToDate
> xetest3 role:Secondary
> replication:WFBitMapS peer-disk:Outdated
>
> ################## on node xetest2
> r0 role:Secondary
> disk:UpToDate
> xetest1 role:Secondary
> peer-disk:UpToDate
> xetest3 role:Secondary
> replication:SyncSource peer-disk:Inconsistent
>
> ################## on node xetest3
> r0 role:Secondary
> disk:Inconsistent
> xetest1 role:Secondary
> replication:SyncTarget peer-disk:UpToDate
> xetest2 role:Secondary
> peer-disk:UpToDate
>
> ################## check 5 ##################
> Mon Oct 8 09:11:58 GMT 2018
> ################## on node xetest1
> r0 role:Secondary
> disk:UpToDate
> xetest2 role:Secondary
> peer-disk:UpToDate
> xetest3 role:Secondary
> peer-disk:Outdated
>
> ################## on node xetest2
> r0 role:Secondary
> disk:UpToDate
> xetest1 role:Secondary
> peer-disk:UpToDate
> xetest3 role:Secondary
> replication:SyncSource peer-disk:Inconsistent
>
> ################## on node xetest3
> r0 role:Secondary
> disk:Outdated blocked:upper
> xetest1 role:Secondary
> replication:WFBitMapT peer-disk:UpToDate
> xetest2 role:Secondary
> peer-disk:UpToDate
>
> ################## check 6 ##################
> Mon Oct 8 09:12:09 GMT 2018
> ################## on node xetest1
> r0 role:Secondary
> disk:UpToDate
> xetest2 role:Secondary
> peer-disk:UpToDate
> xetest3 role:Secondary
> peer-disk:UpToDate
>
> ################## on node xetest2
> r0 role:Secondary
> disk:UpToDate
> xetest1 role:Secondary
> peer-disk:UpToDate
> xetest3 role:Secondary
> peer-disk:Inconsistent resync-suspended:peer
>
> ################## on node xetest3
> r0 role:Secondary
> disk:Inconsistent
> xetest1 role:Secondary
> replication:SyncTarget peer-disk:UpToDate
> xetest2 role:Secondary
> peer-disk:UpToDate
>
> ################## check 7 ##################
> Mon Oct 8 09:12:21 GMT 2018
> ################## on node xetest1
> r0 role:Secondary
> disk:UpToDate
> xetest2 role:Secondary
> peer-disk:UpToDate
> xetest3 role:Secondary
> peer-disk:UpToDate
>
> ################## on node xetest2
> r0 role:Secondary
> disk:UpToDate
> xetest1 role:Secondary
> peer-disk:UpToDate
> xetest3 role:Secondary
> replication:SyncSource peer-disk:Inconsistent
>
> ################## on node xetest3
> r0 role:Secondary
> disk:Outdated blocked:upper
> xetest1 role:Secondary
> peer-disk:UpToDate
> xetest2 role:Secondary
> replication:WFBitMapT peer-disk:UpToDate
>
> ################## check 8 ##################
> Mon Oct 8 09:12:32 GMT 2018
> ################## on node xetest1
> r0 role:Secondary
> disk:UpToDate
> xetest2 role:Secondary
> peer-disk:UpToDate
> xetest3 role:Secondary
> peer-disk:Outdated
>
> ################## on node xetest2
> r0 role:Secondary
> disk:UpToDate
> xetest1 role:Secondary
> peer-disk:UpToDate
> xetest3 role:Secondary
> replication:SyncSource peer-disk:Inconsistent
>
> ################## on node xetest3
> r0 role:Secondary
> disk:Inconsistent
> xetest1 role:Secondary
> replication:SyncTarget peer-disk:UpToDate
> xetest2 role:Secondary
> peer-disk:UpToDate
>
> ################## check 9 ##################
> Mon Oct 8 09:12:44 GMT 2018
> ################## on node xetest1
> r0 role:Secondary
> disk:UpToDate
> xetest2 role:Secondary
> peer-disk:UpToDate
> xetest3 role:Secondary
> peer-disk:Outdated
>
> ################## on node xetest2
> r0 role:Secondary
> disk:UpToDate
> xetest1 role:Secondary
> peer-disk:UpToDate
> xetest3 role:Secondary
> replication:SyncSource peer-disk:Inconsistent
>
> ################## on node xetest3
> r0 role:Secondary
> disk:Outdated blocked:upper
> xetest1 role:Secondary
> replication:WFBitMapT peer-disk:UpToDate
> xetest2 role:Secondary
> peer-disk:UpToDate
>
> ################## check 10 ##################
> Mon Oct 8 09:12:55 GMT 2018
> ################## on node xetest1
> r0 role:Secondary
> disk:UpToDate
> xetest2 role:Secondary
> peer-disk:UpToDate
> xetest3 role:Secondary
> peer-disk:UpToDate
>
> ################## on node xetest2
> r0 role:Secondary
> disk:UpToDate
> xetest1 role:Secondary
> peer-disk:UpToDate
> xetest3 role:Secondary
> replication:SyncSource peer-disk:Inconsistent
>
> ################## on node xetest3
> r0 role:Secondary
> disk:Outdated blocked:upper
> xetest1 role:Secondary
> replication:WFBitMapT peer-disk:UpToDate
> xetest2 role:Secondary
> peer-disk:UpToDate
>
> [oracle at xetest1 ~]$
>
>
>
>
>
>
> ####### our config file is:
> [oracle at xetest1 ~]$ cat /etc/drbd.conf
> include "drbd.d/global_common.conf";
>
> resource r0 {
> protocol C;
> meta-disk internal;
> device /dev/drbd0;
> options {
> auto-promote no;
> quorum majority;
> on-no-quorum io-error;
> # quorum-minimum-redundancy 2;
> }
> syncer {
> verify-alg sha1;
> }
> net {
> ping-int 5;
> after-sb-0pri discard-zero-changes;
> after-sb-1pri disconnect;
> after-sb-2pri disconnect;
> }
> on xetest1 {
> disk /dev/sdb1;
> address 10.0.1.2:7790;
> node-id 1;
> }
> on xetest2 {
> disk /dev/sdb1;
> address 10.0.2.2:7790;
> node-id 2;
> }
> on xetest3 {
> disk /dev/sdb1;
> address 10.0.0.2:7790;
> node-id 3;
> }
> connection-mesh {
> hosts xetest1 xetest2 xetest3;
> }
> }
> [oracle at xetest1 ~]$
>
>
>
>
>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/users/attachments/20181008/8b30d21d/attachment-0002.html>
More information about the Users
mailing list