<html>
<head>
<meta http-equiv="content-type" content="text/html; charset=ISO-8859-1">
</head>
<body bgcolor="#FFFFFF" text="#000000">
Hello again,<br>
<br>
My PM 1.1.12 cluster is quite large: 22 nodes, ~300 resources.<br>
<br>
When gracefully shutting down the current DC (iow. move resources
elsewhere, node standby, pacemaker stop, corosync stop) the CIB load
increases - on the slowest nodes to close to 100% - until the new DC
gets elected.<br>
What explains this phenomenom ?<br>
(What could I do to limit/circumvent it ?)<br>
<br>
In parallel, when this happens and on those nodes that display the
"throttle_mode: High CIB load detected" message, my "ping" (network
connectivity) RA times out without obvious explanation (the RA
timeout is conservative enough, compared to the ping
timeout/attempts, so that it should never kick in). Looking at the
code of the ".../resource.d/pacemaker/ping", I suspect - though I
may be wrong - the culprit is "attrd_updater".<br>
Hypothesis: "attrd_updater" doesn't return immediately, as it is
supposed to do, because of the high CIB load.<br>
Does this hypothesis make sense ?<br>
(PS: it is very difficult for me to reproduce/debug this issue,
showing up on my production cluster, without risking to wreak havoc
with my services)<br>
<br>
Thank you very much for your response(s)<br>
<br>
Best,<br>
<br>
Cédric<br>
<br>
<div class="moz-signature">-- <br>
<meta http-equiv="Content-Type" content="text/html;
charset=ISO-8859-1">
<style type="text/css">
DIV.signature {FONT:normal 11px sans-serif;COLOR:#000000;}
DIV.signature P {MARGIN:5px 0px;FONT:bold 13px sans-serif;COLOR:#000050;}
</style>
<div class="signature">
<p>Cédric Dufour @ Idiap Research Institute</p>
</div>
</div>
</body>
</html>