[ClusterLabs] drbd9 3 node "neverending" sync?

Mevo Govo govomevo at gmail.com
Mon Oct 8 08:12:13 EDT 2018


Hi,

I tried it with and without drbd quorum options, I resterted all the node,
but same "neverending" refresh. Device is not mounted, so no "outer"
modification. protocol is C (synchron). I recreated the resource on node 3
(xetest3) with tis commands:

drbdadm down r0
dd if=/dev/zero of=/dev/sdb1 bs=100M count=1
drbdadm create-md r0

After this "full resync", the neverending sync stopped (UpToDate on every
nodes). But after a resource down/up, the "Outdated/Inconsistent/UpToDate"
cycle is restarted on xetest3.
Maybe it is normal on 3rd node, even if protocol C. But I do not see it on
our prod environment.

B.r: Laszlo.



Mevo Govo <govomevo at gmail.com> ezt írta (időpont: 2018. okt. 8., H, 11:20):

> Hi,
>
> we are testing drbd9 with 3 node. After a test (network disable, then
> enable on one host), the syncronization is not finishing.
> As a recovery, I switched the pacemaker off, and I tried "primary --force"
> on one node and "--discard-my-data" on the other nodes. It seem solve the
> problem (everything is UpToDate), but after I down/up the resource, the
> "neverending" synchronization is restarted (when the resource is Secondary
> on all the 3 node).
> Why sync is restarted after UpToDate state. Ho can I solve this problem?
> Nodes are xetes1, xetest2, xetest3, here is a checking test ("drbdadm
> status" on every node, then wait 10 secod. This repeated 10 times)
>
> Thanks for answers: Laci.
>
>
> for i in {1..10}
> do
>   echo "################## check $i ##################"
>   sleep 10
>   date
>   echo '################## on node xetest1'
>   ssh xetest1 sudo drbdadm status
>   echo '################## on node xetest2'
>   ssh xetest2 sudo drbdadm status
>   echo '################## on node xetest3'
>   ssh xetest3 sudo drbdadm status
> done
>
> ################## check 1 ##################
>
>
>
> Mon Oct  8 09:11:12 GMT 2018
> ################## on node xetest1
> r0 role:Secondary
>   disk:UpToDate
>   xetest2 role:Secondary
>     peer-disk:UpToDate
>   xetest3 role:Secondary
>     peer-disk:UpToDate
>
> ################## on node xetest2
> r0 role:Secondary
>   disk:UpToDate
>   xetest1 role:Secondary
>     peer-disk:UpToDate
>   xetest3 role:Secondary
>     peer-disk:Inconsistent resync-suspended:peer
>
> ################## on node xetest3
> r0 role:Secondary
>   disk:Outdated blocked:upper
>   xetest1 role:Secondary
>     replication:WFBitMapT peer-disk:UpToDate
>   xetest2 role:Secondary
>     peer-disk:UpToDate
>
> ################## check 2 ##################
> Mon Oct  8 09:11:23 GMT 2018
> ################## on node xetest1
> r0 role:Secondary
>   disk:UpToDate
>   xetest2 role:Secondary
>     peer-disk:UpToDate
>   xetest3 role:Secondary
>     peer-disk:Inconsistent resync-suspended:peer
>
> ################## on node xetest2
> r0 role:Secondary
>   disk:UpToDate
>   xetest1 role:Secondary
>     peer-disk:UpToDate
>   xetest3 role:Secondary
>     replication:SyncSource peer-disk:Inconsistent
>
> ################## on node xetest3
> r0 role:Secondary
>   disk:Outdated blocked:upper
>   xetest1 role:Secondary
>     peer-disk:UpToDate
>   xetest2 role:Secondary
>     replication:WFBitMapT peer-disk:UpToDate
>
> ################## check 3 ##################
> Mon Oct  8 09:11:35 GMT 2018
> ################## on node xetest1
> r0 role:Secondary
>   disk:UpToDate
>   xetest2 role:Secondary
>     peer-disk:UpToDate
>   xetest3 role:Secondary
>     replication:SyncSource peer-disk:Inconsistent
>
> ################## on node xetest2
> r0 role:Secondary
>   disk:UpToDate
>   xetest1 role:Secondary
>     peer-disk:UpToDate
>   xetest3 role:Secondary
>     replication:SyncSource peer-disk:Inconsistent
>
> ################## on node xetest3
> r0 role:Secondary
>   disk:UpToDate
>   xetest1 role:Secondary
>     peer-disk:UpToDate
>   xetest2 role:Secondary
>     peer-disk:UpToDate
>
> ################## check 4 ##################
> Mon Oct  8 09:11:46 GMT 2018
> ################## on node xetest1
> r0 role:Secondary
>   disk:UpToDate
>   xetest2 role:Secondary
>     peer-disk:UpToDate
>   xetest3 role:Secondary
>     replication:WFBitMapS peer-disk:Outdated
>
> ################## on node xetest2
> r0 role:Secondary
>   disk:UpToDate
>   xetest1 role:Secondary
>     peer-disk:UpToDate
>   xetest3 role:Secondary
>     replication:SyncSource peer-disk:Inconsistent
>
> ################## on node xetest3
> r0 role:Secondary
>   disk:Inconsistent
>   xetest1 role:Secondary
>     replication:SyncTarget peer-disk:UpToDate
>   xetest2 role:Secondary
>     peer-disk:UpToDate
>
> ################## check 5 ##################
> Mon Oct  8 09:11:58 GMT 2018
> ################## on node xetest1
> r0 role:Secondary
>   disk:UpToDate
>   xetest2 role:Secondary
>     peer-disk:UpToDate
>   xetest3 role:Secondary
>     peer-disk:Outdated
>
> ################## on node xetest2
> r0 role:Secondary
>   disk:UpToDate
>   xetest1 role:Secondary
>     peer-disk:UpToDate
>   xetest3 role:Secondary
>     replication:SyncSource peer-disk:Inconsistent
>
> ################## on node xetest3
> r0 role:Secondary
>   disk:Outdated blocked:upper
>   xetest1 role:Secondary
>     replication:WFBitMapT peer-disk:UpToDate
>   xetest2 role:Secondary
>     peer-disk:UpToDate
>
> ################## check 6 ##################
> Mon Oct  8 09:12:09 GMT 2018
> ################## on node xetest1
> r0 role:Secondary
>   disk:UpToDate
>   xetest2 role:Secondary
>     peer-disk:UpToDate
>   xetest3 role:Secondary
>     peer-disk:UpToDate
>
> ################## on node xetest2
> r0 role:Secondary
>   disk:UpToDate
>   xetest1 role:Secondary
>     peer-disk:UpToDate
>   xetest3 role:Secondary
>     peer-disk:Inconsistent resync-suspended:peer
>
> ################## on node xetest3
> r0 role:Secondary
>   disk:Inconsistent
>   xetest1 role:Secondary
>     replication:SyncTarget peer-disk:UpToDate
>   xetest2 role:Secondary
>     peer-disk:UpToDate
>
> ################## check 7 ##################
> Mon Oct  8 09:12:21 GMT 2018
> ################## on node xetest1
> r0 role:Secondary
>   disk:UpToDate
>   xetest2 role:Secondary
>     peer-disk:UpToDate
>   xetest3 role:Secondary
>     peer-disk:UpToDate
>
> ################## on node xetest2
> r0 role:Secondary
>   disk:UpToDate
>   xetest1 role:Secondary
>     peer-disk:UpToDate
>   xetest3 role:Secondary
>     replication:SyncSource peer-disk:Inconsistent
>
> ################## on node xetest3
> r0 role:Secondary
>   disk:Outdated blocked:upper
>   xetest1 role:Secondary
>     peer-disk:UpToDate
>   xetest2 role:Secondary
>     replication:WFBitMapT peer-disk:UpToDate
>
> ################## check 8 ##################
> Mon Oct  8 09:12:32 GMT 2018
> ################## on node xetest1
> r0 role:Secondary
>   disk:UpToDate
>   xetest2 role:Secondary
>     peer-disk:UpToDate
>   xetest3 role:Secondary
>     peer-disk:Outdated
>
> ################## on node xetest2
> r0 role:Secondary
>   disk:UpToDate
>   xetest1 role:Secondary
>     peer-disk:UpToDate
>   xetest3 role:Secondary
>     replication:SyncSource peer-disk:Inconsistent
>
> ################## on node xetest3
> r0 role:Secondary
>   disk:Inconsistent
>   xetest1 role:Secondary
>     replication:SyncTarget peer-disk:UpToDate
>   xetest2 role:Secondary
>     peer-disk:UpToDate
>
> ################## check 9 ##################
> Mon Oct  8 09:12:44 GMT 2018
> ################## on node xetest1
> r0 role:Secondary
>   disk:UpToDate
>   xetest2 role:Secondary
>     peer-disk:UpToDate
>   xetest3 role:Secondary
>     peer-disk:Outdated
>
> ################## on node xetest2
> r0 role:Secondary
>   disk:UpToDate
>   xetest1 role:Secondary
>     peer-disk:UpToDate
>   xetest3 role:Secondary
>     replication:SyncSource peer-disk:Inconsistent
>
> ################## on node xetest3
> r0 role:Secondary
>   disk:Outdated blocked:upper
>   xetest1 role:Secondary
>     replication:WFBitMapT peer-disk:UpToDate
>   xetest2 role:Secondary
>     peer-disk:UpToDate
>
> ################## check 10 ##################
> Mon Oct  8 09:12:55 GMT 2018
> ################## on node xetest1
> r0 role:Secondary
>   disk:UpToDate
>   xetest2 role:Secondary
>     peer-disk:UpToDate
>   xetest3 role:Secondary
>     peer-disk:UpToDate
>
> ################## on node xetest2
> r0 role:Secondary
>   disk:UpToDate
>   xetest1 role:Secondary
>     peer-disk:UpToDate
>   xetest3 role:Secondary
>     replication:SyncSource peer-disk:Inconsistent
>
> ################## on node xetest3
> r0 role:Secondary
>   disk:Outdated blocked:upper
>   xetest1 role:Secondary
>     replication:WFBitMapT peer-disk:UpToDate
>   xetest2 role:Secondary
>     peer-disk:UpToDate
>
> [oracle at xetest1 ~]$
>
>
>
>
>
>
> ####### our config file is:
> [oracle at xetest1 ~]$ cat /etc/drbd.conf
> include "drbd.d/global_common.conf";
>
> resource r0 {
>  protocol C;
>  meta-disk internal;
>  device /dev/drbd0;
>  options {
>    auto-promote no;
>    quorum majority;
>    on-no-quorum io-error;
> #   quorum-minimum-redundancy 2;
>  }
>  syncer {
>   verify-alg sha1;
>  }
>  net {
>   ping-int 5;
>   after-sb-0pri discard-zero-changes;
>   after-sb-1pri disconnect;
>   after-sb-2pri disconnect;
>  }
>  on xetest1 {
>   disk   /dev/sdb1;
>   address  10.0.1.2:7790;
>   node-id  1;
>  }
>  on xetest2 {
>   disk   /dev/sdb1;
>   address  10.0.2.2:7790;
>   node-id  2;
>  }
>  on xetest3 {
>   disk   /dev/sdb1;
>   address  10.0.0.2:7790;
>   node-id  3;
>  }
>  connection-mesh {
>         hosts     xetest1 xetest2 xetest3;
>  }
> }
> [oracle at xetest1 ~]$
>
>
>
>
>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/users/attachments/20181008/8b30d21d/attachment-0002.html>


More information about the Users mailing list