[ClusterLabs] Node restart if service is disabled

Enrico Bianchi enrico.bianchi at live.com
Mon Dec 18 07:36:45 EST 2017


Hi,

I have a cluster on CentOS 7 with this configuration:

# pcs property show
Cluster Properties:
  cluster-infrastructure: corosync
  cluster-name: mail_cluster
  dc-version: 1.1.15-11.el7_3.4-e174ec8
  have-watchdog: false
  last-lrm-refresh: 1493425399
  stonith-enabled: true
  symmetric-cluster: true
# pcs cluster status
Cluster Status:
  Stack: corosync
  Current DC: node001.host.local (version 1.1.15-11.el7_3.4-e174ec8) - partition with quorum
  Last updated: Mon Dec 18 13:11:19 2017         Last change: Thu Dec 14 11:59:42 2017 by root via crm_resource on node005.host.local
  5 nodes and 19 resources configured

PCSD Status:
   node005.host.local: Online  node001.host.local: Online

   node003.host.local: Online
   node002.host.local: Online
   node004.host.local: Online
# pcs resource
  mailstore1_fs  (ocf::heartbeat:Filesystem):    Started node001.host.local
  mailstore2_fs  (ocf::heartbeat:Filesystem):    Started node002.host.local
  mailstore3_fs  (ocf::heartbeat:Filesystem):    Started node003.host.local
  mailstore4_fs  (ocf::heartbeat:Filesystem):    Started node004.host.local
  logs_fs        (ocf::heartbeat:Filesystem):    Started node005.host.local
  dsmail_fs      (ocf::heartbeat:Filesystem):    Started node005.host.local
  dovecot1_vip   (ocf::heartbeat:IPaddr2):       Started node001.host.local
  dovecot2_vip   (ocf::heartbeat:IPaddr2):       Started node002.host.local
  dovecot3_vip   (ocf::heartbeat:IPaddr2):       Started node003.host.local
  dovecot4_vip   (ocf::heartbeat:IPaddr2):       Started node004.host.local
  dsmail_vip     (ocf::heartbeat:IPaddr2):       Started node005.host.local
  dovecot1_svr   (systemd:dovecot at 1):    Started node001.host.local
  dovecot2_svr   (systemd:dovecot at 2):    Started node002.host.local
  dovecot3_svr   (systemd:dovecot at 3):    Started node003.host.local
  dovecot4_svr   (systemd:dovecot at 4):    Started node004.host.local
  haproxy_svr    (systemd:haproxy):      Started node005.host.local
  service1_svr      (systemd:mwpec):        Started node005.host.local
  service2_svr       (systemd:mwpecdaemons): Started node005.host.local
# pcs stonith show
  fence_vmware   (stonith:fence_vmware_soap):    Started node005.host.local
# pcs constraint show
Location Constraints:
   Resource: dovecot1_vip
     Enabled on: node001.host.local (score:50)
   Resource: dovecot2_vip
     Enabled on: node002.host.local (score:50)
   Resource: dovecot3_vip
     Enabled on: node003.host.local (score:50)
   Resource: dovecot4_vip
     Enabled on: node004.host.local (score:50)
   Resource: dsmail_fs
     Enabled on: node005.host.local (score:INFINITY) (role: Started)
     Disabled on: node002.host.local (score:-INFINITY) (role: Started)
   Resource: dsmail_vip
     Enabled on: node005.host.local (score:50)
Ordering Constraints:
   start dovecot1_vip then start mailstore1_fs (kind:Mandatory)
   start mailstore1_fs then start dovecot1_svr (kind:Mandatory)
   start dovecot2_vip then start mailstore2_fs (kind:Mandatory)
   start mailstore2_fs then start dovecot2_svr (kind:Mandatory)
   start dovecot3_vip then start mailstore3_fs (kind:Mandatory)
   start mailstore3_fs then start dovecot3_svr (kind:Mandatory)
   start dovecot4_vip then start mailstore4_fs (kind:Mandatory)
   start mailstore4_fs then start dovecot4_svr (kind:Mandatory)
   start dsmail_vip then start dsmail_fs (kind:Mandatory)
   start dsmail_vip then start haproxy_svr (kind:Mandatory)
   start dsmail_fs then start logs_fs (kind:Mandatory)
   start dsmail_vip then start service1_svr (kind:Mandatory)
   start dsmail_vip then start service2_svr (kind:Mandatory)
Colocation Constraints:
   dovecot1_svr with dovecot1_vip (score:INFINITY)
   mailstore1_fs with dovecot1_vip (score:INFINITY)
   dovecot2_svr with dovecot2_vip (score:INFINITY)
   mailstore2_fs with dovecot2_vip (score:INFINITY)
   dovecot3_svr with dovecot3_vip (score:INFINITY)
   mailstore3_fs with dovecot3_vip (score:INFINITY)
   dovecot4_svr with dovecot4_vip (score:INFINITY)
   mailstore4_fs with dovecot4_vip (score:INFINITY)
   dsmail_fs with dsmail_vip (score:INFINITY)
   haproxy_svr with dsmail_vip (score:INFINITY)
   logs_fs with dsmail_fs (score:INFINITY)
   service1_svr with dsmail_vip (score:INFINITY)
   service2_svr with dsmail_vip (score:INFINITY)
Ticket Constraints:
# pcs status
Cluster name: mail_cluster
Stack: corosync
Current DC: node001.host.local (version 1.1.15-11.el7_3.4-e174ec8) - partition with quorum
Last updated: Mon Dec 18 13:28:27 2017          Last change: Thu Dec 14 11:59:42 2017 by root via crm_resource on node005.host.local

5 nodes and 19 resources configured

Online: [ node001.host.local node002.host.local node003.host.local node004.host.local node005.host.local ]

Full list of resources:

  mailstore1_fs  (ocf::heartbeat:Filesystem):    Started node001.host.local
  mailstore2_fs  (ocf::heartbeat:Filesystem):    Started node002.host.local
  mailstore3_fs  (ocf::heartbeat:Filesystem):    Started node003.host.local
  mailstore4_fs  (ocf::heartbeat:Filesystem):    Started node004.host.local
  logs_fs        (ocf::heartbeat:Filesystem):    Started node005.host.local
  dsmail_fs      (ocf::heartbeat:Filesystem):    Started node005.host.local
  dovecot1_vip   (ocf::heartbeat:IPaddr2):       Started node001.host.local
  dovecot2_vip   (ocf::heartbeat:IPaddr2):       Started node002.host.local
  dovecot3_vip   (ocf::heartbeat:IPaddr2):       Started node003.host.local
  dovecot4_vip   (ocf::heartbeat:IPaddr2):       Started node004.host.local
  dsmail_vip     (ocf::heartbeat:IPaddr2):       Started node005.host.local
  dovecot1_svr   (systemd:dovecot at 1):    Started node001.host.local
  dovecot2_svr   (systemd:dovecot at 2):    Started node002.host.local
  dovecot3_svr   (systemd:dovecot at 3):    Started node003.host.local
  dovecot4_svr   (systemd:dovecot at 4):    Started node004.host.local
  haproxy_svr    (systemd:haproxy):      Started node005.host.local
  service1_svr      (systemd:mwpec):        Started node005.host.local
  service2_svr       (systemd:mwpecdaemons): Started node005.host.local
  fence_vmware   (stonith:fence_vmware_soap):    Started node005.host.local

Failed Actions:
* fence_vmware_start_0 on node004.host.local 'unknown error' (1): call=803, status=Timed Out, exitreason='none',
     last-rc-change='Mon Jun 26 18:21:29 2017', queued=0ms, exec=20961ms
* fence_vmware_start_0 on node002.host.local 'unknown error' (1): call=906, status=Error, exitreason='none',
     last-rc-change='Mon Jun 26 18:20:52 2017', queued=3ms, exec=13876ms
* fence_vmware_start_0 on node001.host.local 'unknown error' (1): call=1194, status=Timed Out, exitreason='none',
     last-rc-change='Mon Jun 26 18:20:31 2017', queued=0ms, exec=19676ms
* fence_vmware_start_0 on node003.host.local 'unknown error' (1): call=783, status=Timed Out, exitreason='none',
     last-rc-change='Mon Jun 26 18:21:07 2017', queued=0ms, exec=19954ms


Daemon Status:
   corosync: active/enabled
   pacemaker: active/enabled
   pcsd: active/enabled
#

(please note, "Failed Actions" refers a resolved problem)

Cluster works, but I've noticed that if I disable a resource (e.g. 
service1_svr), stonith restart the node (default action). Is normal? 
Resources service1_svr and service2_svr are java applications started 
with systemd as forking types, is a problem?

Cheers,

Enrico



More information about the Users mailing list