<div dir="ltr">Yes. <div><br></div><div>'pcs cluster stop --all' + reboot all nodes</div><div><br></div><div>/Steffen</div></div><br><div class="gmail_quote"><div dir="ltr" class="gmail_attr">On Tue, Jan 12, 2021 at 11:43 AM Klaus Wenninger <<a href="mailto:kwenning@redhat.com">kwenning@redhat.com</a>> wrote:<br></div><blockquote class="gmail_quote" style="margin:0px 0px 0px 0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex">
  
    
  
  <div>
    <div>On 1/12/21 11:23 AM, Steffen Vinther
      Sørensen wrote:<br>
    </div>
    <blockquote type="cite">
      
      <div dir="ltr">Hello Hideo.
        <div><br>
        </div>
        <div>I am overwhelmed by how serious this group is taking good
          care of issues. </div>
        <div><br>
        </div>
        <div>For your information, the 'pending fencing action' status
          disappeared after bringing the nodes offline, and during that
          I found some gfs2 errors that were fixed by fsck.gfs2, and
          since then my cluster has been acting very stable. <br>
        </div>
      </div>
    </blockquote>
    <tt>By bringing offline you mean shutting down pacemaker?</tt><tt><br>
    </tt><tt>That would be expected as fence-history is kept solely in
      RAM.</tt><tt><br>
    </tt><tt>The history-knowledge is synced between the nodes so the</tt><tt><br>
    </tt><tt>history is just lost if all nodes are down at the same
      time.</tt><tt><br>
    </tt><tt>Unfortunately that mechanism keeps unwanted leftovers</tt><tt><br>
    </tt><tt>around as well.</tt><tt><br>
    </tt><tt><br>
    </tt><tt>Regards,</tt><tt><br>
    </tt><tt>Klaus</tt><br>
    <blockquote type="cite">
      <div dir="ltr">
        <div><br>
        </div>
        <div>If I can provide more info let me know. </div>
        <div><br>
        </div>
        <div>/Steffen</div>
      </div>
      <br>
      <div class="gmail_quote">
        <div dir="ltr" class="gmail_attr">On Tue, Jan 12, 2021 at 3:45
          AM <<a href="mailto:renayama19661014@ybb.ne.jp" target="_blank">renayama19661014@ybb.ne.jp</a>>
          wrote:<br>
        </div>
        <blockquote class="gmail_quote" style="margin:0px 0px 0px 0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex">Hi
          Steffen,<br>
          <br>
          I've been experimenting with it since last weekend, but I
          haven't been able to reproduce the same situation.<br>
          It seems that the cause is that the reproduction method cannot
          be limited.<br>
          <br>
          Can I attach a problem log?<br>
          <br>
          Best Regards,<br>
          Hideo Yamauchi.<br>
          <br>
          <br>
          ----- Original Message -----<br>
          > From: Klaus Wenninger <<a href="mailto:kwenning@redhat.com" target="_blank">kwenning@redhat.com</a>><br>
          > To: Steffen Vinther Sørensen <<a href="mailto:svinther@gmail.com" target="_blank">svinther@gmail.com</a>>; Cluster
          Labs - All topics related to open-source clustering welcomed
          <<a href="mailto:users@clusterlabs.org" target="_blank">users@clusterlabs.org</a>><br>
          > Cc: <br>
          > Date: 2021/1/7, Thu 21:42<br>
          > Subject: Re: [ClusterLabs] Pending Fencing Actions shown
          in pcs status<br>
          > <br>
          > On 1/7/21 1:13 PM, Steffen Vinther Sørensen wrote:<br>
          >>  Hi Klaus,<br>
          >> <br>
          >>  Yes then the status does sync to the other nodes.
          Also it looks like<br>
          >>  there are some hostname resolving problems in play
          here, maybe causing<br>
          >>  problems,  here is my notes from restarting
          pacemaker etc.<br>
          > Don't think there are hostname resolving problems.<br>
          > The messages you are seeing, that look as if, are caused<br>
          > by using -EHOSTUNREACH as error-code to fail a pending<br>
          > fence action when a node that is just coming up sees<br>
          > a pending action that is claimed to be handled by
          himself.<br>
          > Back then I chose that error-code as there was none<br>
          > that really matched available right away and it was<br>
          > urgent for some reason so that introduction of something<br>
          > new was to risky at that state.<br>
          > Probably would make sense to introduce something that<br>
          > is more descriptive.<br>
          > Back then the issue was triggered by fenced crashing and<br>
          > being restarted - so not a node-restart but just fenced<br>
          > restarting.<br>
          > And it looks as if building the failed-message failed
          somehow.<br>
          > So that could be the reason why the pending action
          persists.<br>
          > Would be something else then what we solved with Bug
          5401.<br>
          > But what triggers the logs below might as well just be a<br>
          > follow-up issue after the Bug 5401 thing.<br>
          > Will try to find time for a deeper look later today.<br>
          > <br>
          > Klaus<br>
          >> <br>
          >>  pcs cluster standby <a href="http://kvm03-node02.avigol-gcs.dk" rel="noreferrer" target="_blank">kvm03-node02.avigol-gcs.dk</a><br>
          >>  pcs cluster stop <a href="http://kvm03-node02.avigol-gcs.dk" rel="noreferrer" target="_blank">kvm03-node02.avigol-gcs.dk</a><br>
          >>  pcs status<br>
          >> <br>
          >>  Pending Fencing Actions:<br>
          >>  * reboot of <a href="http://kvm03-node02.avigol-gcs.dk" rel="noreferrer" target="_blank">kvm03-node02.avigol-gcs.dk</a>
          pending: client=crmd.37819,<br>
          >>  origin=<a href="http://kvm03-node03.avigol-gcs.dk" rel="noreferrer" target="_blank">kvm03-node03.avigol-gcs.dk</a><br>
          >> <br>
          >>  # From logs on all 3 nodes:<br>
          >>  Jan 07 12:48:18 kvm03-node03 stonith-ng[37815]: 
          warning: received<br>
          >>  pending action we are supposed to be the owner but
          it's not in our<br>
          >>  records -> fail it<br>
          >>  Jan 07 12:48:18 kvm03-node03 stonith-ng[37815]:   
          error: Operation<br>
          >>  'reboot' targeting <a href="http://kvm03-node02.avigol-gcs.dk" rel="noreferrer" target="_blank">kvm03-node02.avigol-gcs.dk</a>
          on <no-one> for<br>
          >>  <a href="mailto:crmd.37819@kvm03-node03.avigol-gcs.dk.56a3018c" target="_blank">crmd.37819@kvm03-node03.avigol-gcs.dk.56a3018c</a>: No
          route to host<br>
          >>  Jan 07 12:48:18 kvm03-node03 stonith-ng[37815]:   
          error:<br>
          >>  stonith_construct_reply: Triggered assert at
          commands.c:2406 : request<br>
          >>  != NULL<br>
          >>  Jan 07 12:48:18 kvm03-node03 stonith-ng[37815]: 
          warning: Can't create<br>
          >>  a sane reply<br>
          >>  Jan 07 12:48:18 kvm03-node03 crmd[37819]:   notice:
          Peer<br>
          >>  <a href="http://kvm03-node02.avigol-gcs.dk" rel="noreferrer" target="_blank">kvm03-node02.avigol-gcs.dk</a>
          was not terminated (reboot) by <anyone> on<br>
          >>  behalf of crmd.37819: No route to host<br>
          >> <br>
          >>  pcs cluster start <a href="http://kvm03-node02.avigol-gcs.dk" rel="noreferrer" target="_blank">kvm03-node02.avigol-gcs.dk</a><br>
          >>  pcs status (now outputs the same on all 3 nodes)<br>
          >> <br>
          >>  Failed Fencing Actions:<br>
          >>  * reboot of <a href="http://kvm03-node02.avigol-gcs.dk" rel="noreferrer" target="_blank">kvm03-node02.avigol-gcs.dk</a>
          failed: delegate=,<br>
          >>  client=crmd.37819, origin=<a href="http://kvm03-node03.avigol-gcs.dk" rel="noreferrer" target="_blank">kvm03-node03.avigol-gcs.dk</a>,<br>
          >>      last-failed='Thu Jan  7 12:48:18 2021'<br>
          >> <br>
          >> <br>
          >>  pcs cluster unstandby <a href="http://kvm03-node02.avigol-gcs.dk" rel="noreferrer" target="_blank">kvm03-node02.avigol-gcs.dk</a><br>
          >> <br>
          >>  # Now libvirtd refuses to start<br>
          >> <br>
          >>  Jan 07 12:51:44 kvm03-node02 dnsmasq[20884]: read
          /etc/hosts - 8 addresses<br>
          >>  Jan 07 12:51:44 kvm03-node02 dnsmasq[20884]: read<br>
          >>  /var/lib/libvirt/dnsmasq/default.addnhosts - 0
          addresses<br>
          >>  Jan 07 12:51:44 kvm03-node02 dnsmasq-dhcp[20884]:
          read<br>
          >>  /var/lib/libvirt/dnsmasq/default.hostsfile<br>
          >>  Jan 07 12:51:44 kvm03-node02 libvirtd[24091]:
          2021-01-07<br>
          >>  11:51:44.729+0000: 24160: info : libvirt version:
          4.5.0, package:<br>
          >>  36.el7_9.3 (CentOS BuildSystem <<a href="http://bugs.centos.org" rel="noreferrer" target="_blank">http://bugs.centos.org</a>
          >,<br>
          >>  2020-11-16-16:25:20, <a href="http://x86-01.bsys.centos.org" rel="noreferrer" target="_blank">x86-01.bsys.centos.org</a>)<br>
          >>  Jan 07 12:51:44 kvm03-node02 libvirtd[24091]:
          2021-01-07<br>
          >>  11:51:44.729+0000: 24160: info : hostname:
          kvm03-node02<br>
          >>  Jan 07 12:51:44 kvm03-node02 libvirtd[24091]:
          2021-01-07<br>
          >>  11:51:44.729+0000: 24160: error :
          qemuMonitorOpenUnix:392 : failed to<br>
          >>  connect to monitor socket: Connection refused<br>
          >>  Jan 07 12:51:44 kvm03-node02 libvirtd[24091]:
          2021-01-07<br>
          >>  11:51:44.729+0000: 24159: error :
          qemuMonitorOpenUnix:392 : failed to<br>
          >>  connect to monitor socket: Connection refused<br>
          >>  Jan 07 12:51:44 kvm03-node02 libvirtd[24091]:
          2021-01-07<br>
          >>  11:51:44.730+0000: 24161: error :
          qemuMonitorOpenUnix:392 : failed to<br>
          >>  connect to monitor socket: Connection refused<br>
          >>  Jan 07 12:51:44 kvm03-node02 libvirtd[24091]:
          2021-01-07<br>
          >>  11:51:44.730+0000: 24162: error :
          qemuMonitorOpenUnix:392 : failed to<br>
          >>  connect to monitor socket: Connection refused<br>
          >> <br>
          >>  pcs status<br>
          >> <br>
          >>  Failed Resource Actions:<br>
          >>  * libvirtd_start_0 on <a href="http://kvm03-node02.avigol-gcs.dk" rel="noreferrer" target="_blank">kvm03-node02.avigol-gcs.dk</a>
          'unknown error' <br>
          > (1):<br>
          >>  call=142, status=complete, exitreason='',<br>
          >>      last-rc-change='Thu Jan  7 12:51:44 2021',
          queued=0ms, <br>
          > exec=2157ms<br>
          >> <br>
          >>  Failed Fencing Actions:<br>
          >>  * reboot of <a href="http://kvm03-node02.avigol-gcs.dk" rel="noreferrer" target="_blank">kvm03-node02.avigol-gcs.dk</a>
          failed: delegate=,<br>
          >>  client=crmd.37819, origin=<a href="http://kvm03-node03.avigol-gcs.dk" rel="noreferrer" target="_blank">kvm03-node03.avigol-gcs.dk</a>,<br>
          >>      last-failed='Thu Jan  7 12:48:18 2021'<br>
          >> <br>
          >> <br>
          >>  # from /etc/hosts on all 3 nodes:<br>
          >> <br>
          >>  172.31.0.31    kvm03-node01 <a href="http://kvm03-node01.avigol-gcs.dk" rel="noreferrer" target="_blank">kvm03-node01.avigol-gcs.dk</a><br>
          >>  172.31.0.32    kvm03-node02 <a href="http://kvm03-node02.avigol-gcs.dk" rel="noreferrer" target="_blank">kvm03-node02.avigol-gcs.dk</a><br>
          >>  172.31.0.33    kvm03-node03 <a href="http://kvm03-node03.avigol-gcs.dk" rel="noreferrer" target="_blank">kvm03-node03.avigol-gcs.dk</a><br>
          >> <br>
          >>  On Thu, Jan 7, 2021 at 11:15 AM Klaus Wenninger <<a href="mailto:kwenning@redhat.com" target="_blank">kwenning@redhat.com</a>> <br>
          > wrote:<br>
          >>>  Hi Steffen,<br>
          >>> <br>
          >>>  If you just see the leftover pending-action on
          one node<br>
          >>>  it would be interesting if restarting of
          pacemaker on<br>
          >>>  one of the other nodes does sync it to all of
          the<br>
          >>>  nodes.<br>
          >>> <br>
          >>>  Regards,<br>
          >>>  Klaus<br>
          >>> <br>
          >>>  On 1/7/21 9:54 AM, <a href="mailto:renayama19661014@ybb.ne.jp" target="_blank">renayama19661014@ybb.ne.jp</a> wrote:<br>
          >>>>  Hi Steffen,<br>
          >>>> <br>
          >>>>>  Unfortunately not sure about the exact
          scenario. But I have <br>
          > been doing<br>
          >>>>>  some recent experiments with node
          standby/unstandby stop/start. <br>
          > This<br>
          >>>>>  to get procedures right for updating
          node rpms etc.<br>
          >>>>> <br>
          >>>>>  Later I noticed the uncomforting
          "pending fencing <br>
          > actions" status msg.<br>
          >>>>  Okay!<br>
          >>>> <br>
          >>>>  Repeat the standby and unstandby steps in
          the same way to check.<br>
          >>>>  We will start checking after tomorrow, so I
          think it will take some <br>
          > time until next week.<br>
          >>>> <br>
          >>>> <br>
          >>>>  Many thanks,<br>
          >>>>  Hideo Yamauchi.<br>
          >>>> <br>
          >>>> <br>
          >>>> <br>
          >>>>  ----- Original Message -----<br>
          >>>>>  From: "<a href="mailto:renayama19661014@ybb.ne.jp" target="_blank">renayama19661014@ybb.ne.jp</a>" <br>
          > <<a href="mailto:renayama19661014@ybb.ne.jp" target="_blank">renayama19661014@ybb.ne.jp</a>><br>
          >>>>>  To: Reid Wahl <<a href="mailto:nwahl@redhat.com" target="_blank">nwahl@redhat.com</a>>; Cluster
          Labs - All <br>
          > topics related to open-source clustering welcomed <<a href="mailto:users@clusterlabs.org" target="_blank">users@clusterlabs.org</a>><br>
          >>>>>  Cc:<br>
          >>>>>  Date: 2021/1/7, Thu 17:51<br>
          >>>>>  Subject: Re: [ClusterLabs] Pending
          Fencing Actions shown in pcs <br>
          > status<br>
          >>>>> <br>
          >>>>>  Hi Steffen,<br>
          >>>>>  Hi Reid,<br>
          >>>>> <br>
          >>>>>  The fencing history is kept inside
          stonith-ng and is not <br>
          > written to cib.<br>
          >>>>>  However, getting the entire cib and
          getting it sent will help <br>
          > you to reproduce<br>
          >>>>>  the problem.<br>
          >>>>> <br>
          >>>>>  Best Regards,<br>
          >>>>>  Hideo Yamauchi.<br>
          >>>>> <br>
          >>>>> <br>
          >>>>>  ----- Original Message -----<br>
          >>>>>>  From: Reid Wahl <<a href="mailto:nwahl@redhat.com" target="_blank">nwahl@redhat.com</a>><br>
          >>>>>>  To: <a href="mailto:renayama19661014@ybb.ne.jp" target="_blank">renayama19661014@ybb.ne.jp</a>;
          Cluster Labs - All topics <br>
          > related to<br>
          >>>>>  open-source clustering welcomed <<a href="mailto:users@clusterlabs.org" target="_blank">users@clusterlabs.org</a>><br>
          >>>>>>  Date: 2021/1/7, Thu 17:39<br>
          >>>>>>  Subject: Re: [ClusterLabs] Pending
          Fencing Actions shown in <br>
          > pcs status<br>
          >>>>>> <br>
          >>>>>> <br>
          >>>>>>  Hi, Steffen. Those attachments don't
          contain the CIB. <br>
          > They contain the<br>
          >>>>>  `pcs config` output. You can get the cib
          with `pcs cluster cib <br>
          >> <br>
          >>>>>  $(hostname).cib.xml`.<br>
          >>>>>>  Granted, it's possible that this
          fence action <br>
          > information wouldn't<br>
          >>>>>  be in the CIB at all. It might be stored
          in fencer memory.<br>
          >>>>>>  On Thu, Jan 7, 2021 at 12:26 AM <br>
          > <<a href="mailto:renayama19661014@ybb.ne.jp" target="_blank">renayama19661014@ybb.ne.jp</a>>
          wrote:<br>
          >>>>>> <br>
          >>>>>>  Hi Steffen,<br>
          >>>>>>>>   Here CIB settings attached
          (pcs config show) for <br>
          > all 3 of my nodes<br>
          >>>>>>>>   (all 3 seems 100%
          identical), node03 is the DC.<br>
          >>>>>>>  Thank you for the attachment.<br>
          >>>>>>> <br>
          >>>>>>>  What is the scenario when this
          situation occurs?<br>
          >>>>>>>  In what steps did the problem
          appear when fencing was <br>
          > performed (or<br>
          >>>>>  failed)?<br>
          >>>>>>>  Best Regards,<br>
          >>>>>>>  Hideo Yamauchi.<br>
          >>>>>>> <br>
          >>>>>>> <br>
          >>>>>>>  ----- Original Message -----<br>
          >>>>>>>>   From: Steffen Vinther
          Sørensen <br>
          > <<a href="mailto:svinther@gmail.com" target="_blank">svinther@gmail.com</a>><br>
          >>>>>>>>   To: <a href="mailto:renayama19661014@ybb.ne.jp" target="_blank">renayama19661014@ybb.ne.jp</a>;
          Cluster Labs - All <br>
          > topics related<br>
          >>>>>  to open-source clustering welcomed <br>
          > <<a href="mailto:users@clusterlabs.org" target="_blank">users@clusterlabs.org</a>><br>
          >>>>>>>>   Cc:<br>
          >>>>>>>>   Date: 2021/1/7, Thu 17:05<br>
          >>>>>>>>   Subject: Re: [ClusterLabs]
          Pending Fencing Actions <br>
          > shown in pcs<br>
          >>>>>  status<br>
          >>>>>>>>   Hi Hideo,<br>
          >>>>>>>> <br>
          >>>>>>>>   Here CIB settings attached
          (pcs config show) for <br>
          > all 3 of my nodes<br>
          >>>>>>>>   (all 3 seems 100%
          identical), node03 is the DC.<br>
          >>>>>>>> <br>
          >>>>>>>>   Regards<br>
          >>>>>>>>   Steffen<br>
          >>>>>>>> <br>
          >>>>>>>>   On Thu, Jan 7, 2021 at 8:06
          AM <br>
          > <<a href="mailto:renayama19661014@ybb.ne.jp" target="_blank">renayama19661014@ybb.ne.jp</a>><br>
          >>>>>  wrote:<br>
          >>>>>>>>>    Hi Steffen,<br>
          >>>>>>>>>    Hi Reid,<br>
          >>>>>>>>> <br>
          >>>>>>>>>    I also checked the
          Centos source rpm and it <br>
          > seems to include a<br>
          >>>>>  fix for the<br>
          >>>>>>>>   problem.<br>
          >>>>>>>>>    As Steffen suggested,
          if you share your CIB <br>
          > settings, I might<br>
          >>>>>  know<br>
          >>>>>>>>   something.<br>
          >>>>>>>>>    If this issue is the
          same as the fix, the <br>
          > display will only be<br>
          >>>>>  displayed on<br>
          >>>>>>>>   the DC node and will not
          affect the operation.<br>
          >>>>>>>>>    The pending actions
          shown will remain for a <br>
          > long time, but<br>
          >>>>>  will not have a<br>
          >>>>>>>>   negative impact on the
          cluster.<br>
          >>>>>>>>>    Best Regards,<br>
          >>>>>>>>>    Hideo Yamauchi.<br>
          >>>>>>>>> <br>
          >>>>>>>>> <br>
          >>>>>>>>>    ----- Original Message
          -----<br>
          >>>>>>>>>    > From: Reid Wahl
          <<a href="mailto:nwahl@redhat.com" target="_blank">nwahl@redhat.com</a>><br>
          >>>>>>>>>    > To: Cluster Labs
          - All topics related to <br>
          > open-source<br>
          >>>>>  clustering<br>
          >>>>>>>>   welcomed <<a href="mailto:users@clusterlabs.org" target="_blank">users@clusterlabs.org</a>><br>
          >>>>>>>>>    > Cc:<br>
          >>>>>>>>>    > Date: 2021/1/7,
          Thu 15:58<br>
          >>>>>>>>>    > Subject: Re:
          [ClusterLabs] Pending <br>
          > Fencing Actions shown<br>
          >>>>>  in pcs status<br>
          >>>>>>>>>    ><br>
          >>>>>>>>>    > It's supposedly
          fixed in that <br>
          > version.<br>
          >>>>>>>>>    >   - <br>
          > <a href="https://bugzilla.redhat.com/show_bug.cgi?id=1787749" rel="noreferrer" target="_blank">https://bugzilla.redhat.com/show_bug.cgi?id=1787749</a>
          <br>
          >>>>>>>>>    >   - <br>
          > <a href="https://access.redhat.com/solutions/4713471" rel="noreferrer" target="_blank">https://access.redhat.com/solutions/4713471</a>
          <br>
          >>>>>>>>>    ><br>
          >>>>>>>>>    > So you may be
          hitting a different issue <br>
          > (unless<br>
          >>>>>  there's a bug in<br>
          >>>>>>>>   the<br>
          >>>>>>>>>    > pcmk 1.1 backport
          of the fix).<br>
          >>>>>>>>>    ><br>
          >>>>>>>>>    > I may be a little
          bit out of my area of <br>
          > knowledge here,<br>
          >>>>>  but can you<br>
          >>>>>>>>>    > share the CIBs
          from nodes 1 and 3? Maybe <br>
          > Hideo, Klaus, or<br>
          >>>>>  Ken has some<br>
          >>>>>>>>>    > insight.<br>
          >>>>>>>>>    ><br>
          >>>>>>>>>    > On Wed, Jan 6,
          2021 at 10:53 PM Steffen <br>
          > Vinther Sørensen<br>
          >>>>>>>>>    > <<a href="mailto:svinther@gmail.com" target="_blank">svinther@gmail.com</a>> wrote:<br>
          >>>>>>>>>    >><br>
          >>>>>>>>>    >>  Hi Hideo,<br>
          >>>>>>>>>    >><br>
          >>>>>>>>>    >>  If the fix
          is not going to make it <br>
          > into the CentOS7<br>
          >>>>>  pacemaker<br>
          >>>>>>>>   version,<br>
          >>>>>>>>>    >>  I guess the
          stable approach to take <br>
          > advantage of it<br>
          >>>>>  is to build<br>
          >>>>>>>>   the<br>
          >>>>>>>>>    >>  cluster on
          another OS than CentOS7 <br>
          > ? A little late<br>
          >>>>>  for that in<br>
          >>>>>>>>   this<br>
          >>>>>>>>>    >>  case though
          :)<br>
          >>>>>>>>>    >><br>
          >>>>>>>>>    >>  Regards<br>
          >>>>>>>>>    >>  Steffen<br>
          >>>>>>>>>    >><br>
          >>>>>>>>>    >><br>
          >>>>>>>>>    >><br>
          >>>>>>>>>    >><br>
          >>>>>>>>>    >>  On Thu, Jan
          7, 2021 at 7:27 AM<br>
          >>>>>  <<a href="mailto:renayama19661014@ybb.ne.jp" target="_blank">renayama19661014@ybb.ne.jp</a>><br>
          >>>>>>>>   wrote:<br>
          >>>>>>>>>    >>  ><br>
          >>>>>>>>>    >>  > Hi
          Steffen,<br>
          >>>>>>>>>    >>  ><br>
          >>>>>>>>>    >>  > The fix
          pointed out by Reid is <br>
          > affecting it.<br>
          >>>>>>>>>    >>  ><br>
          >>>>>>>>>    >>  > Since
          the fencing action <br>
          > requested by the DC<br>
          >>>>>  node exists<br>
          >>>>>>>>   only in the<br>
          >>>>>>>>>    > DC node, such an
          event occurs.<br>
          >>>>>>>>>    >>  > You
          will need to take <br>
          > advantage of the modified<br>
          >>>>>  pacemaker to<br>
          >>>>>>>>   resolve<br>
          >>>>>>>>>    > the issue.<br>
          >>>>>>>>>    >>  ><br>
          >>>>>>>>>    >>  > Best
          Regards,<br>
          >>>>>>>>>    >>  > Hideo
          Yamauchi.<br>
          >>>>>>>>>    >>  ><br>
          >>>>>>>>>    >>  ><br>
          >>>>>>>>>    >>  ><br>
          >>>>>>>>>    >>  > -----
          Original Message -----<br>
          >>>>>>>>>    >>  > >
          From: Reid Wahl <br>
          > <<a href="mailto:nwahl@redhat.com" target="_blank">nwahl@redhat.com</a>><br>
          >>>>>>>>>    >>  > >
          To: Cluster Labs - All <br>
          > topics related to<br>
          >>>>>  open-source<br>
          >>>>>>>>   clustering<br>
          >>>>>>>>>    > welcomed <<a href="mailto:users@clusterlabs.org" target="_blank">users@clusterlabs.org</a>><br>
          >>>>>>>>>    >>  > >
          Cc:<br>
          >>>>>>>>>    >>  > >
          Date: 2021/1/7, Thu 15:07<br>
          >>>>>>>>>    >>  > >
          Subject: Re: <br>
          > [ClusterLabs] Pending Fencing<br>
          >>>>>  Actions<br>
          >>>>>>>>   shown in pcs<br>
          >>>>>>>>>    > status<br>
          >>>>>>>>>    >>  > ><br>
          >>>>>>>>>    >>  > >
          Hi, Steffen. Are your <br>
          > cluster nodes all<br>
          >>>>>  running the<br>
          >>>>>>>>   same<br>
          >>>>>>>>>    > Pacemaker<br>
          >>>>>>>>>    >>  > >
          versions? This looks like <br>
          > Bug 5401[1],<br>
          >>>>>  which is fixed<br>
          >>>>>>>>   by upstream<br>
          >>>>>>>>>    >>  > >
          commit df71a07[2]. <br>
          > I'm a little bit<br>
          >>>>>  confused about<br>
          >>>>>>>>   why it<br>
          >>>>>>>>>    > only shows<br>
          >>>>>>>>>    >>  > > up
          on one out of three <br>
          > nodes though.<br>
          >>>>>>>>>    >>  > ><br>
          >>>>>>>>>    >>  > >
          [1]<br>
          >>>>>  <a href="https://bugs.clusterlabs.org/show_bug.cgi?id=5401" rel="noreferrer" target="_blank">https://bugs.clusterlabs.org/show_bug.cgi?id=5401</a>
          <br>
          >>>>>>>>>    >>  > >
          [2]<br>
          >>>>>>>>   <br>
          > <a href="https://github.com/ClusterLabs/pacemaker/commit/df71a07" rel="noreferrer" target="_blank">https://github.com/ClusterLabs/pacemaker/commit/df71a07</a>
          <br>
          >>>>>>>>>    >>  > ><br>
          >>>>>>>>>    >>  > > On
          Tue, Jan 5, 2021 at <br>
          > 8:31 AM Steffen<br>
          >>>>>  Vinther Sørensen<br>
          >>>>>>>>>    >>  > > <br>
          > <<a href="mailto:svinther@gmail.com" target="_blank">svinther@gmail.com</a>> wrote:<br>
          >>>>>>>>>    >>  >
          >><br>
          >>>>>>>>>    >>  >
          >>  Hello<br>
          >>>>>>>>>    >>  >
          >><br>
          >>>>>>>>>    >>  >
          >>  node 1 is showing <br>
          > this in 'pcs<br>
          >>>>>  status'<br>
          >>>>>>>>>    >>  >
          >><br>
          >>>>>>>>>    >>  >
          >>  Pending Fencing <br>
          > Actions:<br>
          >>>>>>>>>    >>  >
          >>  * reboot of<br>
          >>>>>  <a href="http://kvm03-node02.avigol-gcs.dk" rel="noreferrer" target="_blank">kvm03-node02.avigol-gcs.dk</a>
          pending:<br>
          >>>>>>>>>    >
          client=crmd.37819,<br>
          >>>>>>>>>    >>  >
          >>  <br>
          > origin=<a href="http://kvm03-node03.avigol-gcs.dk" rel="noreferrer" target="_blank">kvm03-node03.avigol-gcs.dk</a><br>
          >>>>>>>>>    >>  >
          >><br>
          >>>>>>>>>    >>  >
          >>  node 2 and node 3 <br>
          > outputs no such<br>
          >>>>>  thing (node 3 is<br>
          >>>>>>>>   DC)<br>
          >>>>>>>>>    >>  >
          >><br>
          >>>>>>>>>    >>  >
          >>  Google is not much <br>
          > help, how to<br>
          >>>>>  investigate this<br>
          >>>>>>>>   further and<br>
          >>>>>>>>>    > get rid<br>
          >>>>>>>>>    >>  >
          >>  of such terrifying <br>
          > status message ?<br>
          >>>>>>>>>    >>  >
          >><br>
          >>>>>>>>>    >>  >
          >>  Regards<br>
          >>>>>>>>>    >>  >
          >>  Steffen<br>
          >>>>>>>>>    >>  >
          >><br>
          >>>>> 
          _______________________________________________<br>
          >>>>>>>>>    >>  >
          >>  Manage your <br>
          > subscription:<br>
          >>>>>>>>>    >>  >
          >><br>
          >>>>>>>>   <br>
          > <a href="https://lists.clusterlabs.org/mailman/listinfo/users" rel="noreferrer" target="_blank">https://lists.clusterlabs.org/mailman/listinfo/users</a>
          <br>
          >>>>>>>>>    >>  >
          >><br>
          >>>>>>>>>    >>  >
          >>  ClusterLabs home:<br>
          >>>>>  <a href="https://www.clusterlabs.org/" rel="noreferrer" target="_blank">https://www.clusterlabs.org/</a>
          <br>
          >>>>>>>>>    >>  >
          >><br>
          >>>>>>>>>    >>  > ><br>
          >>>>>>>>>    >>  > ><br>
          >>>>>>>>>    >>  > > --<br>
          >>>>>>>>>    >>  > >
          Regards,<br>
          >>>>>>>>>    >>  > ><br>
          >>>>>>>>>    >>  > >
          Reid Wahl, RHCA<br>
          >>>>>>>>>    >>  > >
          Senior Software <br>
          > Maintenance Engineer, Red<br>
          >>>>>  Hat<br>
          >>>>>>>>>    >>  > >
          CEE - Platform Support <br>
          > Delivery -<br>
          >>>>>  ClusterHA<br>
          >>>>>>>>>    >>  > ><br>
          >>>>>>>>>    >>  > ><br>
          >>>>> 
          _______________________________________________<br>
          >>>>>>>>>    >>  > >
          Manage your subscription:<br>
          >>>>>>>>>    >>  > ><br>
          >>>>>  <a href="https://lists.clusterlabs.org/mailman/listinfo/users" rel="noreferrer" target="_blank">https://lists.clusterlabs.org/mailman/listinfo/users</a>
          <br>
          >>>>>>>>>    >>  > ><br>
          >>>>>>>>>    >>  > >
          ClusterLabs home:<br>
          >>>>>  <a href="https://www.clusterlabs.org/" rel="noreferrer" target="_blank">https://www.clusterlabs.org/</a>
          <br>
          >>>>>>>>>    >>  > ><br>
          >>>>>>>>>    >>  ><br>
          >>>>>>>>>    >>  > <br>
          > _______________________________________________<br>
          >>>>>>>>>    >>  > Manage
          your subscription:<br>
          >>>>>>>>>    >>  ><br>
          >>>>>  <a href="https://lists.clusterlabs.org/mailman/listinfo/users" rel="noreferrer" target="_blank">https://lists.clusterlabs.org/mailman/listinfo/users</a>
          <br>
          >>>>>>>>>    >>  ><br>
          >>>>>>>>>    >>  >
          ClusterLabs home: <br>
          > <a href="https://www.clusterlabs.org/" rel="noreferrer" target="_blank">https://www.clusterlabs.org/</a>
          <br>
          >>>>>>>>>    >>  <br>
          > _______________________________________________<br>
          >>>>>>>>>    >>  Manage your
          subscription:<br>
          >>>>>>>>>    >>  <br>
          > <a href="https://lists.clusterlabs.org/mailman/listinfo/users" rel="noreferrer" target="_blank">https://lists.clusterlabs.org/mailman/listinfo/users</a>
          <br>
          >>>>>>>>>    >><br>
          >>>>>>>>>    >>  ClusterLabs
          home: <br>
          > <a href="https://www.clusterlabs.org/" rel="noreferrer" target="_blank">https://www.clusterlabs.org/</a>
          <br>
          >>>>>>>>>    ><br>
          >>>>>>>>>    ><br>
          >>>>>>>>>    ><br>
          >>>>>>>>>    > --<br>
          >>>>>>>>>    > Regards,<br>
          >>>>>>>>>    ><br>
          >>>>>>>>>    > Reid Wahl, RHCA<br>
          >>>>>>>>>    > Senior Software
          Maintenance Engineer, <br>
          > Red Hat<br>
          >>>>>>>>>    > CEE - Platform
          Support Delivery - <br>
          > ClusterHA<br>
          >>>>>>>>>    ><br>
          >>>>>>>>>    > <br>
          > _______________________________________________<br>
          >>>>>>>>>    > Manage your
          subscription:<br>
          >>>>>>>>>    > <br>
          > <a href="https://lists.clusterlabs.org/mailman/listinfo/users" rel="noreferrer" target="_blank">https://lists.clusterlabs.org/mailman/listinfo/users</a>
          <br>
          >>>>>>>>>    ><br>
          >>>>>>>>>    > ClusterLabs home:
          <br>
          > <a href="https://www.clusterlabs.org/" rel="noreferrer" target="_blank">https://www.clusterlabs.org/</a>
          <br>
          >>>>>>>>>    ><br>
          >>>>>>>>> <br>
          >>>>>>>>>   <br>
          > _______________________________________________<br>
          >>>>>>>>>    Manage your
          subscription:<br>
          >>>>>>>>>   <br>
          > <a href="https://lists.clusterlabs.org/mailman/listinfo/users" rel="noreferrer" target="_blank">https://lists.clusterlabs.org/mailman/listinfo/users</a>
          <br>
          >>>>>>>>> <br>
          >>>>>>>>>    ClusterLabs home: <br>
          > <a href="https://www.clusterlabs.org/" rel="noreferrer" target="_blank">https://www.clusterlabs.org/</a>
          <br>
          >>>>>>> 
          _______________________________________________<br>
          >>>>>>>  Manage your subscription:<br>
          >>>>>>>  <a href="https://lists.clusterlabs.org/mailman/listinfo/users" rel="noreferrer" target="_blank">https://lists.clusterlabs.org/mailman/listinfo/users</a>
          <br>
          >>>>>>> <br>
          >>>>>>>  ClusterLabs home: <a href="https://www.clusterlabs.org/" rel="noreferrer" target="_blank">https://www.clusterlabs.org/</a>
          <br>
          >>>>>>> <br>
          >>>>>>  --<br>
          >>>>>> <br>
          >>>>>>  Regards,<br>
          >>>>>> <br>
          >>>>>>  Reid Wahl, RHCA<br>
          >>>>>> <br>
          >>>>>>  Senior Software Maintenance
          Engineer, Red Hat<br>
          >>>>>>  CEE - Platform Support Delivery -
          ClusterHA<br>
          >>>>>> <br>
          >>>>>> <br>
          >>>>> 
          _______________________________________________<br>
          >>>>>  Manage your subscription:<br>
          >>>>>  <a href="https://lists.clusterlabs.org/mailman/listinfo/users" rel="noreferrer" target="_blank">https://lists.clusterlabs.org/mailman/listinfo/users</a>
          <br>
          >>>>> <br>
          >>>>>  ClusterLabs home: <a href="https://www.clusterlabs.org/" rel="noreferrer" target="_blank">https://www.clusterlabs.org/</a>
          <br>
          >>>>> <br>
          >>>> 
          _______________________________________________<br>
          >>>>  Manage your subscription:<br>
          >>>>  <a href="https://lists.clusterlabs.org/mailman/listinfo/users" rel="noreferrer" target="_blank">https://lists.clusterlabs.org/mailman/listinfo/users</a>
          <br>
          >>>> <br>
          >>>>  ClusterLabs home: <a href="https://www.clusterlabs.org/" rel="noreferrer" target="_blank">https://www.clusterlabs.org/</a>
          <br>
          >>>  _______________________________________________<br>
          >>>  Manage your subscription:<br>
          >>>  <a href="https://lists.clusterlabs.org/mailman/listinfo/users" rel="noreferrer" target="_blank">https://lists.clusterlabs.org/mailman/listinfo/users</a>
          <br>
          >>> <br>
          >>>  ClusterLabs home: <a href="https://www.clusterlabs.org/" rel="noreferrer" target="_blank">https://www.clusterlabs.org/</a>
          <br>
          > <br>
          > _______________________________________________<br>
          > Manage your subscription:<br>
          > <a href="https://lists.clusterlabs.org/mailman/listinfo/users" rel="noreferrer" target="_blank">https://lists.clusterlabs.org/mailman/listinfo/users</a>
          <br>
          > <br>
          > ClusterLabs home: <a href="https://www.clusterlabs.org/" rel="noreferrer" target="_blank">https://www.clusterlabs.org/</a>
          <br>
          > <br>
          <br>
        </blockquote>
      </div>
    </blockquote>
    <br>
  </div>

</blockquote></div>