[ClusterLabs] Why did the ticket that was granted become revoked?
yuusuke_iida at newson.co.jp
yuusuke_iida at newson.co.jp
Fri Aug 1 04:40:03 UTC 2025
Hi all,
While testing a multi-site cluster using tickets, I encountered an issue where tickets were revoked automatically when Pacemaker was restarted.
Things to confirm
* Is it intentional that tickets are revoked when Pacemaker is restarted?
What we are trying to do
* We are verifying a multi-site cluster using tickets.
* Since we assume that tickets will be switched manually, we are not using booth.
Expected behavior
* We want tickets to remain granted even after Pacemaker restarts.
What we did
* For maintenance purposes, we stopped Pacemaker on one node with `pcs cluster stop` and then started it with `pcs cluster start`, but the tickets that were in the granted state were revoked.
Best regards,
Cluster status before Pacemaker stop on standby1
```
Cluster name: primary_cluster
Cluster Summary:
* Stack: corosync (Pacemaker is running)
* Current DC: primary1 (1) (version 2.1.8-3.el9-3980678f0) - partition with quorum
* Last updated: Thu Jul 31 23:06:30 2025 on standby1
* Last change: Thu Jul 31 23:06:05 2025 by root via root on primary1
* 2 nodes configured
* 13 resource instances configured
Node List:
* Node primary1 (1): online, feature set 3.19.5
* Node standby1 (2): online, feature set 3.19.5
Full List of Resources:
* Clone Set: pgsql-clone [pgsql] (promotable):
* pgsql (ocf:linuxhajp:pgsql): Promoted primary1
* pgsql (ocf:linuxhajp:pgsql): Unpromoted standby1
* Resource Group: primary-group:
* ipaddr-primary (ocf:heartbeat:IPaddr2): Started primary1
* ipaddr-replication (ocf:heartbeat:IPaddr2): Started primary1
* ipaddr-standby (ocf:heartbeat:IPaddr2): Started standby1
* Resource Group: remote-site-rep-group:
* remote-site-pgsql (ocf:linuxhajp:pgsql): Stopped
* local-site-pgsql (ocf:linuxhajp:pgsql): Stopped
* Clone Set: ping-clone [ping]:
* ping (ocf:pacemaker:ping): Started primary1
* ping (ocf:pacemaker:ping): Started standby1
* Clone Set: storage-mon-clone [storage-mon]:
* storage-mon (ocf:heartbeat:storage-mon): Started primary1
* storage-mon (ocf:heartbeat:storage-mon): Started standby1
* fence1-virsh (stonith:fence_virsh): Started standby1
* fence2-virsh (stonith:fence_virsh): Started primary1
Node Attributes:
* Node: primary1 (1):
* master-pgsql : 1000
* pgsql-data-status : LATEST
* pgsql-master-baseline : 000000002D000168
* pgsql-status : PRI
* ping-status : 1
* Node: standby1 (2):
* master-pgsql : 100
* pgsql-data-status : STREAMING|SYNC
* pgsql-status : HS:sync
* ping-status : 1
Migration Summary:
Fencing History:
* turning off of standby1 successful: delegate=a human, client=stonith_admin.3019770, origin=primary1, completed='2025-07-31 23:02:10.031118 -04:00'
Tickets:
tokyo-dc-dr-site-ticket revoked
tokyo-dc-primary-site-ticket granted (granted=true, last-granted=Thu Jul 31 23:03:47 2025)
PCSD Status:
primary1: Online
standby1: Online
Daemon Status:
corosync: active/disabled
pacemaker: active/disabled
pcsd: active/enabled
```
Cluster status after Pacemaker start on standby1
```
Cluster name: primary_cluster
Cluster Summary:
* Stack: corosync (Pacemaker is running)
* Current DC: primary1 (1) (version 2.1.8-3.el9-3980678f0) - partition with quorum
* Last updated: Thu Jul 31 23:07:08 2025 on standby1
* Last change: Thu Jul 31 23:06:36 2025 by root via root on primary1
* 2 nodes configured
* 13 resource instances configured
Node List:
* Node primary1 (1): online, feature set 3.19.5
* Node standby1 (2): online, feature set 3.19.5
Full List of Resources:
* Clone Set: pgsql-clone [pgsql] (promotable):
* pgsql (ocf:linuxhajp:pgsql): Stopped
* pgsql (ocf:linuxhajp:pgsql): Stopped
* Resource Group: primary-group:
* ipaddr-primary (ocf:heartbeat:IPaddr2): Stopped
* ipaddr-replication (ocf:heartbeat:IPaddr2): Stopped
* ipaddr-standby (ocf:heartbeat:IPaddr2): Stopped
* Resource Group: remote-site-rep-group:
* remote-site-pgsql (ocf:linuxhajp:pgsql): Stopped
* local-site-pgsql (ocf:linuxhajp:pgsql): Stopped
* Clone Set: ping-clone [ping]:
* ping (ocf:pacemaker:ping): Started standby1
* ping (ocf:pacemaker:ping): Started primary1
* Clone Set: storage-mon-clone [storage-mon]:
* storage-mon (ocf:heartbeat:storage-mon): Started standby1
* storage-mon (ocf:heartbeat:storage-mon): Started primary1
* fence1-virsh (stonith:fence_virsh): Started standby1
* fence2-virsh (stonith:fence_virsh): Started primary1
Node Attributes:
* Node: primary1 (1):
* master-pgsql : -INFINITY
* pgsql-data-status : LATEST
* pgsql-status : STOP
* ping-status : 1
* Node: standby1 (2):
* pgsql-data-status : DISCONNECT
* ping-status : 1
Migration Summary:
Fencing History:
* turning off of standby1 successful: delegate=a human, client=stonith_admin.3019770, origin=primary1, completed='2025-07-31 23:02:10.031118 -04:00'
Tickets:
tokyo-dc-dr-site-ticket revoked
tokyo-dc-primary-site-ticket revoked
PCSD Status:
primary1: Online
standby1: Online
Daemon Status:
corosync: active/disabled
pacemaker: active/disabled
pcsd: active/enabled
```
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/users/attachments/20250801/cf680b28/attachment-0001.htm>
More information about the Users
mailing list