[Pacemaker] two nodes fenced when drbd link fails

Ivan Coronado icoronado at epcge.com
Fri May 14 15:24:46 UTC 2010


Oh nop, sorry, my mistake, it doesn't works.... :(
 
 
Ivan 

________________________________

De: Ivan Coronado [mailto:icoronado at epcge.com] 
Enviado el: viernes, 14 de mayo de 2010 9:02
Para: The Pacemaker cluster resource manager
Asunto: Re: [Pacemaker] two nodes fenced when drbd link fails


Thanks! 
 
it's works!!!
 
Ivan 

________________________________

De: Vadym Chepkov [mailto:vchepkov at gmail.com] 
Enviado el: viernes, 14 de mayo de 2010 4:03
Para: The Pacemaker cluster resource manager
Asunto: Re: [Pacemaker] two nodes fenced when drbd link fails



On May 13, 2010, at 1:37 PM, Ivan Coronado wrote:


	Hello to everybody,
	 
	I have a problem with the corosync.conf setup. I have a drbd
service runing on eth3, and a general network and the stonith device
(idrac6) in the eth0. If I unplug the eth3 to simulate a network failure
two nodes are fenced (first the slave followed by the master). If I only
leave ringnumber 0 in the coroync.conf file I don't have this problem.
Is this normal operation?
	 
	Here you have the section of corosync.conf where I have the
problem, and thanks for the help.
	 
	        rrp_mode: active
	        interface {
	                # eth0
	                ringnumber: 0
	                bindnetaddr: 200.200.201.0
	                mcastaddr: 226.94.1.1
	                mcastport: 5405
	        }
	        interface {
	                #eth3
	                ringnumber: 1
	                bindnetaddr: 192.168.2.0
	                mcastaddr: 226.94.1.2
	                mcastport: 5406
	        }
	
	-----
	Ivan



I read in the list openais at lists.osdl.org setting ports at least two
apart helps (5405, 5407)

Vadym

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20100514/63768e29/attachment-0002.htm>


More information about the Pacemaker mailing list