[ClusterLabs] pacemaker after upgrade from wheezy to jessie
Toni Tschampke
tt at halle.it
Thu Nov 3 16:39:47 CET 2016
> I'm going to guess you were using the experimental 1.1 schema as the
> "validate-with" at the top of /var/lib/pacemaker/cib/cib.xml. Try
> changing the validate-with to pacemaker-next or pacemaker-1.2 and see if
> you get better results. Don't edit the file directly though; use the
> cibadmin command so it signs the end result properly.
>
> After changing the validate-with, run:
>
> crm_verify -x /var/lib/pacemaker/cib/cib.xml
>
> and fix any errors that show up.
strange, the location of our cib.xml differs from your path, our cib is
located in /var/lib/heartbeat/crm/
running cibadmin --modify --xml-text '<cib validate-with="pacemaker-1.2"/>'
gave no output but was logged to corosync:
cib: info: cib_perform_op: -- <cib num_updates="0"
validate-with="pacemaker-1.1"/>
cib: info: cib_perform_op: ++ <cib admin_epoch="0" epoch="8462"
num_updates="1" validate-with="pacemaker-1.2" crm_feature_set="3.0.6"
have-quorum="1" cib-last-written="Thu Nov 3 10:05:52 2016"
update-origin="nebel1" update-client="cibadmin" update-user="root"/>
I'm guessing this change should be instantly written into the xml file?
If this is the case something is wrong, greping for validate gives the
old string back.
<cib admin_epoch="0" epoch="8462" num_updates="0"
validate-with="pacemaker-1.1" crm_feature_set="3.0.6" have-quorum="1"
cib-last-written="Thu Nov 3 16:19:51 2016" update-origin="nebel1"
update-client="cibadmin" update-user="root">
pacemakerd --features
Pacemaker 1.1.15 (Build: e174ec8)
Supporting v3.0.10:
Should the crm_feature_set be updated this way too? I'm guessing this is
done when "cibadmin --upgrade" succeeds?
We just get an timeout error when trying to upgrade it with cibadmin:
Call cib_upgrade failed (-62): Timer expired
Do have permissions changed from 1.1.7 to 1.1.15? when looking at our
quite big /var/lib/heartbeat/crm/ folder some permissions changed:
-rw------- 1 hacluster root 80K Nov 1 16:56 cib-31.raw
-rw-r--r-- 1 hacluster root 32 Nov 1 16:56 cib-31.raw.sig
-rw------- 1 hacluster haclient 80K Nov 1 18:53 cib-32.raw
-rw------- 1 hacluster haclient 32 Nov 1 18:53 cib-32.raw.sig
cib-31 was before upgrading, cib-32 after starting upgraded pacemaker
--
Mit freundlichen Grüßen
Toni Tschampke | tt at halle.it
bcs kommunikationslösungen
Inh. Dipl. Ing. Carsten Burkhardt
Harz 51 | 06108 Halle (Saale) | Germany
tel +49 345 29849-0 | fax +49 345 29849-22
www.b-c-s.de | www.halle.it | www.wivewa.de
EINFACH ADRESSEN, TELEFONATE UND DOKUMENTE VERWALTEN - MIT WIVEWA -
IHREM WISSENSVERWALTER FUER IHREN BETRIEB!
Weitere Informationen erhalten Sie unter www.wivewa.de
Am 03.11.2016 um 15:39 schrieb Ken Gaillot:
> On 11/03/2016 05:51 AM, Toni Tschampke wrote:
>> Hi,
>>
>> we just upgraded our nodes from wheezy 7.11 (pacemaker 1.1.7) to jessie
>> (pacemaker 1.1.15, corosync 2.3.6).
>> During the upgrade pacemaker was removed (rc) and reinstalled after from
>> jessie-backports, same for crmsh.
>>
>> Now we are encountering multiple problems:
>>
>> First I checked the configuration on a single node running pacemaker &
>> corosync which dropped a strange error, followed by multiple lines
>> stating syntax is wrong. crm configure show then showed up a mixed view
>> of xml and crmsh singleline syntax.
>>
>>> ERROR: Cannot read schema file
>> '/usr/share/pacemaker/pacemaker-1.1.rng': [Errno 2] No such file or
>> directory: '/usr/share/pacemaker/pacemaker-1.1.rng'
>
> pacemaker-1.1.rng was renamed to pacemaker-next.rng in Pacemaker 1.1.12,
> as it was used to hold experimental new features rather than as the
> actual next version of the schema. So, the schema skipped to 1.2.
>
> I'm going to guess you were using the experimental 1.1 schema as the
> "validate-with" at the top of /var/lib/pacemaker/cib/cib.xml. Try
> changing the validate-with to pacemaker-next or pacemaker-1.2 and see if
> you get better results. Don't edit the file directly though; use the
> cibadmin command so it signs the end result properly.
>
> After changing the validate-with, run:
>
> crm_verify -x /var/lib/pacemaker/cib/cib.xml
>
> and fix any errors that show up.
>
>> When we looked into that folder there was pacemaker-1.0.rng, 1.2 and so
>> on. As a quick try we symlinked the 1.2 to 1.1 and the syntax errors
>> were gone. When running crm resource show, all resources showed up, when
>> running crm_mon -1fA the output was unexpected as it showed all nodes
>> offline, with no DC elected:
>>
>>> Stack: corosync
>>> Current DC: NONE
>>> Last updated: Thu Nov 3 11:11:16 2016
>>> Last change: Thu Nov 3 09:54:52 2016 by root via cibadmin on nebel1
>>>
>>> *** Resource management is DISABLED ***
>>> The cluster will not attempt to start, stop or recover services
>>>
>>> 3 nodes and 73 resources configured:
>>> 5 resources DISABLED and 0 BLOCKED from being started due to failures
>>>
>>> OFFLINE: [ nebel1 nebel2 nebel3 ]
>>
>> we tried to manually change dc-version
>>
>> when issuing a simple cleanup command I got the following error:
>>
>>> crm resource cleanup DrbdBackuppcMs
>>> Error signing on to the CRMd service
>>> Error performing operation: Transport endpoint is not connected
>>
>> which looks like crmsh is not able to communicate with crmd and nothing
>> is logged in this case in corosync.log
>>
>> we experimented with multiple config changes (corosync.conf: pacemaker
>> ver 0 > 1)
>> cib-bootstrap-options: cluster-infrastructure from openais to corosync
>>
>>> Package versions:
>>> cman 3.1.8-1.2+b1
>>> corosync 2.3.6-3~bpo8+1
>>> crmsh 2.2.0-1~bpo8+1
>>> csync2 1.34-2.3+b1
>>> dlm-pcmk 3.0.12-3.2+deb7u2
>>> libcman3 3.1.8-1.2+b1
>>> libcorosync-common4:amd64 2.3.6-3~bpo8+1
>>> munin-libvirt-plugins 0.0.6-1
>>> pacemaker 1.1.15-2~bpo8+1
>>> pacemaker-cli-utils 1.1.15-2~bpo8+1
>>> pacemaker-common 1.1.15-2~bpo8+1
>>> pacemaker-resource-agents 1.1.15-2~bpo8+1
>>
>>> Kernel: #1 SMP Debian 3.16.36-1+deb8u2 (2016-10-19) x86_64 GNU/Linux
>>
>> I attached our cib before upgrade and after, as well as the one with the
>> mixed syntax and our corosync.conf.
>>
>> When we tried to connect a second node to the cluster, pacemaker starts
>> it's deamons, starts corosync and dies after 15 tries with following in
>> corosync log:
>>
>>> crmd: info: crm_timer_popped: Wait Timer (I_NULL) just popped (2000ms)
>>> crmd: info: do_cib_control: Could not connect to the CIB service:
>>> Transport endpoint is not connected
>>> crmd: warning: do_cib_control:
>>> Couldn't complete CIB registration 15 times... pause and retry
>>> attrd: error: attrd_cib_connect: Signon to CIB failed:
>>> Transport endpoint is not connected (-107)
>>> attrd: info: main: Shutting down attribute manager
>>> attrd: info: qb_ipcs_us_withdraw: withdrawing server sockets
>>> attrd: info: crm_xml_cleanup: Cleaning up memory from libxml2
>>> crmd: info: crm_timer_popped: Wait Timer (I_NULL) just popped (2000ms)
>>> pacemakerd: warning: pcmk_child_exit:
>>> The attrd process (12761) can no longer be respawned,
>>> shutting the cluster down.
>>> pacemakerd: notice: pcmk_shutdown_worker: Shutting down Pacemaker
>>
>> A third node joins without above error, but crm_mon still shows all
>> nodes as offline.
>>
>> Thanks for any advice how to solve this, I'm out of ideas now.
>>
>> Regards, Toni
>>
>
> _______________________________________________
> Users mailing list: Users at clusterlabs.org
> http://clusterlabs.org/mailman/listinfo/users
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>
More information about the Users
mailing list