[Pacemaker] Does pingd works on openais?

Pinto, Phil (GTI) phil_pinto at ml.com
Tue Mar 11 15:59:01 EDT 2008

Hi - just thought I would jump in here 

>The fact that links get broken completely, but both side can still reach
>the STONITH device, however is statistically very rare.  In fact,
>connectivity to the STONITH device becomes the "quorum token."

This may not always true in a virtual environment if two nodes are running under a VM and the VM in turn runs a virtual Stonith device.  

Phil Pinto 

-----Original Message-----
From: pacemaker-bounces at clusterlabs.org [mailto:pacemaker-bounces at clusterlabs.org] On Behalf Of Lars Marowsky-Bree
Sent: Friday, March 07, 2008 3:29 PM
To: The Pacemaker cluster resource manager
Subject: Re: [Pacemaker] Does pingd works on openais?

On 2008-03-07T12:30:27, Serge Dubrouski <sergeyfd at gmail.com> wrote:

> In fact I'm afraid that CRM lacks one feature that other clustering
> projects have. It's that quorum disk. In RedHat ClusterSuite or in HP
> ServiceGuard quorum disk helps to fight split brain scenario. 

That's not a CRM/Pacemaker-level feature though. That needs to be
provided by the cluster infrastructure below.

You're right that heartbeat does not have a disk quorum plugin. Xinwei
has written a disk-based comm plugin, but that wasn't submitted
upstream. (And, AFAIK, never shipped anywhere.)

> It's not clear how CRM acts when heartbeat link gets broken and nodes
> can't communicate to each other. What I see in my logs both nodes try
> to STONITH each other which isn't the best way to handle this
> problem.

The surviving side would still fence the other, so that's a partially
separate issue. 

The fact that links get broken completely, but both side can still reach
the STONITH device, however is statistically very rare.  In fact,
connectivity to the STONITH device becomes the "quorum token."

I'm not saying it wouldn't be nice to support disk-based comms/quorum
with heartbeat as well, but really it isn't a crucial feature.

> Please do not mistake quorum drive (feature for local cluster) with
> quorumd server which was designed for  geographically spread clusters.

I'm very clear on the distinction ;-)

Though this is not entirely true: it's essentially the same concept. A
3rd party tie-breaker is introduced to decide equal node count splits.
Quorum disk, or quorum server - it's the same idea; one scenario uses
TCP/SSL and the other SCSI reservations, that's the entire difference.
And if, as in this case, the SCSI reservation is in fact handled by the
iSCSI server (over TCP), the distinction becomes pretty much mood.


PS: And please, trim the quotes, if you would be so kind ;-)

Teamlead Kernel, SuSE Labs, Research and Development
SUSE LINUX Products GmbH, GF: Markus Rex, HRB 16746 (AG Nürnberg)
"Experience is the name everyone gives to their mistakes." -- Oscar Wilde

Pacemaker mailing list
Pacemaker at clusterlabs.org

This message w/attachments (message) may be privileged, confidential or proprietary, and if you are not an intended recipient, please notify the sender, do not use or share it and delete it. Unless specifically indicated, this message is not an offer to sell or a solicitation of any investment products or other financial product or service, an official confirmation of any transaction, or an official statement of Merrill Lynch. Subject to applicable law, Merrill Lynch may monitor, review and retain e-communications (EC) traveling through its networks/systems. The laws of the country of each sender/recipient may impact the handling of EC, and EC may be archived, supervised and produced in countries other than the country in which you are located. This message cannot be guaranteed to be secure or error-free. This message is subject to terms available at the following link: http://www.ml.com/e-communications_terms/. By messaging with Merrill Lynch you consent to the foregoing.

More information about the Pacemaker mailing list