[Pacemaker] abrupt power failure problem

Bernd Schubert bs_lists at aakef.fastmail.fm
Tue Jun 15 17:55:49 UTC 2010


Hello Diane,

the problem is that pacemaker is not allowed to take over resources until 
stonith succeeds, as it simply does not know about the state of the other 
server. Lets assume the other node would still be up and running, would have 
mounted a shared storage device an would write to it, but would respond to 
network anymore. If pacemaker would now mount this device again, you would get 
data corruption. To protect you against that, it requires that stonith 
succeeds, or that you manually solve that problem.

The only automatic solution would be a more reliable stonith device, e.g. IPMI 
with an extra power supply for the IPMI card or a PDU.

Cheers,
Bernd

On Tuesday 15 June 2010, Schaefer, Diane E wrote:
> Thanks for the idea. Is there any way to automatically recover resources
>  without manual intervention?
> 
> Diane
> 
> THIS COMMUNICATION MAY CONTAIN CONFIDENTIAL AND/OR OTHERWISE PROPRIETARY
>  MATERIAL and is thus for use only by the intended recipient. If you
>  received this in error, please contact the sender and delete the e-mail
>  and its attachments from all computers.
> 
> 
> -----Original Message-----
> From: Bernd Schubert [mailto:bs_lists at aakef.fastmail.fm]
> Sent: Tuesday, June 15, 2010 1:39 PM
> To: pacemaker at oss.clusterlabs.org
> Cc: Schaefer, Diane E
> Subject: Re: [Pacemaker] abrupt power failure problem
> 
> On Tuesday 15 June 2010, Schaefer, Diane E wrote:
> > Hi,
> >   We are having trouble with our two node cluster after one node
> >  experiences an abrupt power failure.  The resources do not seem to start
> >  on the remaining node (ie DRBD resources do not promote to master).  In
> >  the log we notice:
> >
> > Jan  8 02:12:27 qpr4 stonithd: [6622]: info: external_run_cmd: Calling
> >  '/usr/lib64/stonith/plugins/external/ipmi reset qpr3' returned 256 Jan 
> > 8 02:12:27 qpr4 stonithd: [6622]: CRIT: external_reset_req: 'ipmi reset'
> > for host qpr3 failed with rc 256 Jan  8 02:12:27 qpr4 stonithd: [5854]:
> > info: failed to STONITH node qpr3 with local device stonith0 (exitcode
> > 5), gonna try the next local device Jan  8 02:12:27 qpr4 stonithd:
> > [5854]: info: we can't manage qpr3, broadcast request to other nodes Jan 
> > 8 02:13:27 qpr4 stonithd: [5854]: ERROR: Failed to STONITH the node qpr3:
> > optype=RESET, op_result=TIMEOUT
> >
> > Jan  8 02:13:27 qpr4 stonithd: [6763]: info: external_run_cmd: Calling
> >  '/usr/lib64/stonith/plugins/external/ipmi reset qpr3' returned 256 Jan 
> > 8 02:13:27 qpr4 stonithd: [6763]: CRIT: external_reset_req: 'ipmi reset'
> > for host qpr3 failed with rc 256 Jan  8 02:13:27 qpr4 stonithd: [5854]:
> > info: failed to STONITH node qpr3 with local device stonith0 (exitcode
> > 5), gonna try the next local device Jan  8 02:13:27 qpr4 stonithd:
> > [5854]: info: we can't manage qpr3, broadcast request to other nodes Jan 
> > 8 02:14:27 qpr4 stonithd: [5854]: ERROR: Failed to STONITH the node qpr3:
> > optype=RESET, op_result=TIMEOUT
> 
> Without looking at your hb_report, this already looks pretty clear - this
>  node tries to reset the other node using IPMI and that fails, of course,
>  as the node to be reset is powered off.
> When we had that problem in the past, we simply temporarily removed the
>  failed node from the pacemaker configuration: crm node remove <node-name>
> 
> 
> Cheers,
> Bernd
> 






More information about the Pacemaker mailing list