[Pacemaker] Pacemaker Digest, Vol 53, Issue 42

Frank Meier frank.meier at hr-group.de
Thu Apr 19 08:41:50 UTC 2012


Oh,

my VMs are in productive enviroment, so I shutdown the cluster and start
Xen on one node whithout cluster.

Mit freundlichen Grüßen

Frank Meier
UNIX-Basis

Hamm Reno Group GmbH
Industriegebiet West | D-66987 Thaleischweiler-Fröschen
T.+49(0)6334 444-8322 | F.+49(0)6334 444-8190
frank.meier at hr-group.de | www.reno.de
___________________________________________________________________

Sitz: Am Tie 7 | D-49086 Osnabrück
Handelsregister Osnabrück HRB 19587
Geschäftsführer: Hans-Jürgen de Fries,
Jens Gransee, Manfred Klumpp,
Robert Reisch

Am 19.04.2012 10:07, schrieb pacemaker-request at oss.clusterlabs.org:
> Send Pacemaker mailing list submissions to
>         pacemaker at oss.clusterlabs.org
> 
> To subscribe or unsubscribe via the World Wide Web, visit
>         http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> or, via email, send a message with subject or body 'help' to
>         pacemaker-request at oss.clusterlabs.org
> 
> You can reach the person managing the list at
>         pacemaker-owner at oss.clusterlabs.org
> 
> When replying, please edit your Subject line so it is more specific
> than "Re: Contents of Pacemaker digest..."
> 
> 
> Today's Topics:
> 
>    1. Re: Pacemaker Digest, Vol 53, Issue 40 (emmanuel segura)
> 
> 
> ----------------------------------------------------------------------
> 
> Message: 1
> Date: Thu, 19 Apr 2012 10:12:37 +0200
> From: emmanuel segura <emi2fast at gmail.com>
> To: The Pacemaker cluster resource manager
>         <pacemaker at oss.clusterlabs.org>
> Subject: Re: [Pacemaker] Pacemaker Digest, Vol 53, Issue 40
> Message-ID:
>         <CAE7pJ3B5L1m2s=OQ=cUDd9=mqy50ORhP+eiDteCRp5_6JLR6jQ at mail.gmail.com>
> Content-Type: text/plain; charset="iso-8859-1"
> 
> vgchange -cy XenVG
> 
> your vg it's not clustered and check the clvmd process
> 
> ps -ef | grep clvmd
> 
> Il giorno 19 aprile 2012 09:45, Frank Meier <frank.meier at hr-group.de> ha
> scritto:
> 
>> xencluster1:~ # vgs
>>  VG    #PV #LV #SN Attr   VSize   VFree
>>  XenVG   8  32   0 wz--n- 938.16g 530.98g
>>
>>
>> Mit freundlichen Gr??en
>>
>> Frank Meier
>> UNIX-Basis
>>
>> Hamm Reno Group GmbH
>> Industriegebiet West | D-66987 Thaleischweiler-Fr?schen
>> T.+49(0)6334 444-8322 | F.+49(0)6334 444-8190
>> frank.meier at hr-group.de | www.reno.de
>> ___________________________________________________________________
>>
>> Sitz: Am Tie 7 | D-49086 Osnabr?ck
>> Handelsregister Osnabr?ck HRB 19587
>> Gesch?ftsf?hrer: Hans-J?rgen de Fries,
>> Jens Gransee, Manfred Klumpp,
>> Robert Reisch
>>
>> Am 19.04.2012 09:32, schrieb pacemaker-request at oss.clusterlabs.org:
>>> Send Pacemaker mailing list submissions to
>>>         pacemaker at oss.clusterlabs.org
>>>
>>> To subscribe or unsubscribe via the World Wide Web, visit
>>>         http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>> or, via email, send a message with subject or body 'help' to
>>>         pacemaker-request at oss.clusterlabs.org
>>>
>>> You can reach the person managing the list at
>>>         pacemaker-owner at oss.clusterlabs.org
>>>
>>> When replying, please edit your Subject line so it is more specific
>>> than "Re: Contents of Pacemaker digest..."
>>>
>>>
>>> Today's Topics:
>>>
>>>    1. Re: Pacemaker Digest, Vol 53, Issue 38 (emmanuel segura)
>>>
>>>
>>> ----------------------------------------------------------------------
>>>
>>> Message: 1
>>> Date: Thu, 19 Apr 2012 09:37:31 +0200
>>> From: emmanuel segura <emi2fast at gmail.com>
>>> To: The Pacemaker cluster resource manager
>>>         <pacemaker at oss.clusterlabs.org>
>>> Subject: Re: [Pacemaker] Pacemaker Digest, Vol 53, Issue 38
>>> Message-ID:
>>>         <CAE7pJ3CovBAsjf+vbbJGEKuqTTvFiWWpn0S=UvOTSoyQpeTa=
>> w at mail.gmail.com>
>>> Content-Type: text/plain; charset="iso-8859-1"
>>>
>>> give this command
>>>
>>> vgs
>>>
>>> Il giorno 19 aprile 2012 09:32, Frank Meier <frank.meier at hr-group.de> ha
>>> scritto:
>>>
>>>> Hi,
>>>>
>>>> here my lvm.conf:
>>>>
>>>> devices {
>>>>    dir = "/dev"
>>>>    scan = [ "/dev" ]
>>>>    filter = [ "a|/dev/mapper|", "r/.*/" ]
>>>>    cache = "/etc/lvm/.cache"
>>>>    write_cache_state = 1
>>>>    types = [ "device-mapper", 16 ]
>>>>    sysfs_scan = 1
>>>>    md_component_detection = 1
>>>> }
>>>> log {
>>>>    verbose = 0
>>>>    syslog = 1
>>>>    overwrite = 0
>>>>    level = 0
>>>>    indent = 1
>>>>    command_names = 0
>>>>    prefix = "  "
>>>> }
>>>> backup {
>>>>    backup = 1
>>>>    backup_dir = "/etc/lvm/backup"
>>>>    archive = 1
>>>>    archive_dir = "/etc/lvm/archive"
>>>>    retain_min = 10
>>>>    retain_days = 30
>>>> }
>>>> shell {
>>>>    history_size = 100
>>>> }
>>>> global {
>>>>    umask = 077
>>>>    test = 0
>>>>    activation = 1
>>>>    proc = "/proc"
>>>>    locking_type = 3
>>>>    locking_dir = "/var/lock/lvm"
>>>> }
>>>> activation {
>>>>    missing_stripe_filler = "error"
>>>>    mirror_region_size = 512
>>>>    reserved_stack = 256
>>>>    reserved_memory = 8192
>>>>    process_priority = -18
>>>> }
>>>> dmeventd {
>>>>        mirror_library = "libdevmapper-event-lvm2mirror.so.2.02"
>>>>        snapshot_library = "libdevmapper-event-lvm2snapshot.so.2.02"
>>>> }
>>>>
>>>>
>>>> Mit freundlichen Gr??en
>>>>
>>>> Frank Meier
>>>> UNIX-Basis
>>>>
>>>> Hamm Reno Group GmbH
>>>> Industriegebiet West | D-66987 Thaleischweiler-Fr?schen
>>>> T.+49(0)6334 444-8322 | F.+49(0)6334 444-8190
>>>> frank.meier at hr-group.de | www.reno.de
>>>> ___________________________________________________________________
>>>>
>>>> Sitz: Am Tie 7 | D-49086 Osnabr?ck
>>>> Handelsregister Osnabr?ck HRB 19587
>>>> Gesch?ftsf?hrer: Hans-J?rgen de Fries,
>>>> Jens Gransee, Manfred Klumpp,
>>>> Robert Reisch
>>>>
>>>> Am 19.04.2012 09:13, schrieb pacemaker-request at oss.clusterlabs.org:
>>>>> Send Pacemaker mailing list submissions to
>>>>>         pacemaker at oss.clusterlabs.org
>>>>>
>>>>> To subscribe or unsubscribe via the World Wide Web, visit
>>>>>         http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>> or, via email, send a message with subject or body 'help' to
>>>>>         pacemaker-request at oss.clusterlabs.org
>>>>>
>>>>> You can reach the person managing the list at
>>>>>         pacemaker-owner at oss.clusterlabs.org
>>>>>
>>>>> When replying, please edit your Subject line so it is more specific
>>>>> than "Re: Contents of Pacemaker digest..."
>>>>>
>>>>>
>>>>> Today's Topics:
>>>>>
>>>>>    1. Re: LVM restarts after SLES upgrade (emmanuel segura)
>>>>>
>>>>>
>>>>> ----------------------------------------------------------------------
>>>>>
>>>>> Message: 1
>>>>> Date: Thu, 19 Apr 2012 09:18:37 +0200
>>>>> From: emmanuel segura <emi2fast at gmail.com>
>>>>> To: The Pacemaker cluster resource manager
>>>>>         <pacemaker at oss.clusterlabs.org>
>>>>> Subject: Re: [Pacemaker] LVM restarts after SLES upgrade
>>>>> Message-ID:
>>>>>         <CAE7pJ3Ab1ZzSpou0=vvrO=
>>>> mb3ur2+K1XhALLJEpj-sOOUEfwTA at mail.gmail.com>
>>>>> Content-Type: text/plain; charset="iso-8859-1"
>>>>>
>>>>> maybe you have problem with clvm
>>>>>
>>>>>  can you show me your vm configuration?
>>>>>
>>>>> Il giorno 19 aprile 2012 08:29, Frank Meier <frank.meier at hr-group.de>
>> ha
>>>>> scritto:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> I've installed a 2-Node Xen-Cluster with SLES 11 SP1.
>>>>>>
>>>>>> After an upgrade to SLES11 SP2 the cluster won't work as the old one.
>>>>>>
>>>>>> After some time, all LVs reorganize and so the VM where restarted.
>>>>>>
>>>>>> Any clue?
>>>>>>
>>>>>> CIB:
>>>>>>
>>>>>> node xencluster1 \
>>>>>>        attributes standby="off"
>>>>>> node xencluster2 \
>>>>>>        attributes standby="off"
>>>>>> primitive STONITH--Node1 stonith:meatware \
>>>>>>        meta is-managed="true" target-role="Started" \
>>>>>>        operations $id="STONITH--Node1-operations" \
>>>>>>        op monitor interval="15" timeout="15" start-delay="15" \
>>>>>>        params hostlist="xencluster1"
>>>>>> primitive STONITH-Node2 stonith:meatware \
>>>>>>        meta target-role="started" is-managed="true" \
>>>>>>        operations $id="STONITH-Node2-operations" \
>>>>>>        op monitor interval="15" timeout="15" start-delay="15" \
>>>>>>        params hostlist="xencluster2"
>>>>>> primitive clvm ocf:lvm2:clvmd \
>>>>>>        operations $id="clvm-operations" \
>>>>>>        op monitor interval="10" timeout="20" \
>>>>>>        op start interval="0" timeout="90" \
>>>>>>        op stop interval="0" timeout="100" \
>>>>>>        params daemon_timeout="30"
>>>>>> primitive clvm-xenvg ocf:heartbeat:LVM \
>>>>>>        operations $id="clvm-xenvg-operations" \
>>>>>>        op monitor interval="10" timeout="30" \
>>>>>>        params volgrpname="XenVG"
>>>>>> primitive cmirror ocf:lvm2:cmirrord \
>>>>>>        operations $id="cmirror-operations" \
>>>>>>        op monitor interval="10" timeout="20" \
>>>>>>        op start interval="0" timeout="90" \
>>>>>>        op stop interval="0" timeout="100" \
>>>>>>        params daemon_timeout="30"
>>>>>> primitive dlm ocf:pacemaker:controld \
>>>>>>        operations $id="dlm-operations" \
>>>>>>        op monitor interval="10" timeout="20" start-delay="0" \
>>>>>>        op start interval="0" timeout="90" \
>>>>>>        op stop interval="0" timeout="100"
>>>>>> primitive fs-config-xen ocf:heartbeat:Filesystem \
>>>>>>        meta is-managed="true" target-role="Started" \
>>>>>>        operations $id="fs-config-xen-operations" \
>>>>>>        op monitor interval="20" timeout="40" \
>>>>>>        params device="/dev/mapper/XenVG-xenconfig"
>>>> directory="/etc/xen/vm"
>>>>>> fstype="ocfs2"
>>>>>> primitive o2cb ocf:ocfs2:o2cb \
>>>>>>        operations $id="o2cb-operations" \
>>>>>>        op monitor interval="10" timeout="20" \
>>>>>>        op start interval="0" timeout="90" \
>>>>>>        op stop interval="0" timeout="100"
>>>>>> primitive vm-cim ocf:heartbeat:Xen \
>>>>>>        meta target-role="Started" is-managed="true"
>>>> allow-migrate="true" \
>>>>>>        operations $id="vm-cim-operations" \
>>>>>>        op monitor interval="10" timeout="30" \
>>>>>>        op start interval="0" timeout="60" \
>>>>>>        op stop interval="0" timeout="60" \
>>>>>>        op migrate_to interval="0" timeout="360" \
>>>>>>        params xmfile="/etc/xen/vm/cim"
>>>>>> primitive vm-deprepo ocf:heartbeat:Xen \
>>>>>>        meta target-role="started" is-managed="true"
>>>> allow-migrate="true" \
>>>>>>        operations $id="vm-deprepo-operations" \
>>>>>>        op monitor interval="10" timeout="30" \
>>>>>>        op start interval="0" timeout="60" \
>>>>>>        op stop interval="0" timeout="60" \
>>>>>>        op migrate_to interval="0" timeout="360" \
>>>>>>        params xmfile="/etc/xen/vm/debrepo"
>>>>>>
>>>>>>
>>>>>> group clvm-glue dlm clvm o2cb cmirror \
>>>>>>        meta target-role="Started"
>>>>>> group xen-vg-fs clvm-xenvg fs-config-xen
>>>>>> clone c-clvm-glue clvm-glue \
>>>>>>        meta interleave="true" ordered="true" target-role="Started"
>>>>>> clone c-xen-vg-fs xen-vg-fs \
>>>>>>        meta interleave="true" ordered="true" is-managed="true"
>>>>>> target-role="Started"
>>>>>> location STONITHnode1Constraint STONITH--Node1 -inf: xencluster1
>>>>>> colocation colo-clvmglue-xenvgsfs inf: c-xen-vg-fs c-clvm-glue
>>>>>> order o-cim inf: c-xen-vg-fs vm-cim
>>>>>> order o-clvmglue-xenvgfs inf: c-clvm-glue c-xen-vg-fs
>>>>>> order o-deprepo inf: c-xen-vg-fs vm-deprepo
>>>>>> order o-ilmt inf: c-xen-vg-fs vm-ilmt
>>>>>> order o-intranet inf: c-xen-vg-fs vm-intranet
>>>>>> order o-nagois inf: c-xen-vg-fs vm-nagios
>>>>>> order o-oneiroi inf: c-xen-vg-fs vm-oneiroi
>>>>>> order o-otbridge inf: c-xen-vg-fs vm-otbridge
>>>>>> order o-soa inf: c-xen-vg-fs vm-soa
>>>>>> order o-topdesk-entw inf: c-xen-vg-fs vm-topdesk-entw
>>>>>> order o-traumschiff inf: c-xen-vg-fs vm-traumschiff
>>>>>> order o-virenscanner inf: c-xen-vg-fs vm-virenscanner
>>>>>> property $id="cib-bootstrap-options" \
>>>>>>        dc-version="1.1.6-b988976485d15cb702c9307df55512d323831a5e" \
>>>>>>        no-quorum-policy="ignore" \
>>>>>>        default-resource-stickiness="100000" \
>>>>>>        last-lrm-refresh="1334518148" \
>>>>>>        cluster-infrastructure="openais" \
>>>>>>        expected-quorum-votes="2" \
>>>>>>        default-action-timeout="120s" \
>>>>>>        maintenance-mode="true"
>>>>>> op_defaults $id="op_defaults-options" \
>>>>>>        record-pending="false"
>>>>>>
>>>>>> Node 1Logs:
>>>>>>
>>>>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: Sending message to all
>> cluster
>>>>>> nodes
>>>>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: process_work_item: local
>>>>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: process_local_command:
>> unknown
>>>>>> (0x2d) msg=0x7f5c40010e80, msglen =32, client=0x7f5c400578d0
>>>>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: Syncing device names
>>>>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: Reply from node 4914420a: 0
>>>> bytes
>>>>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: Got 1 replies, expecting: 2
>>>>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: LVM thread waiting for work
>>>>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: 1226064394 got message from
>>>>>> nodeid 1226064394 for 0. len 32
>>>>>> Apr 15 22:01:40 xencluster1 lrmd: [8093]: WARN: clvm-xenvg:0:monitor
>>>>>> process (PID 31785) timed out (try 1).  Killing with signal SIGTERM
>>>> (15).
>>>>>> Apr 15 22:01:40 xencluster1 lrmd: [8093]: WARN: operation monitor[92]
>> on
>>>>>> clvm-xenvg:0 for client 8096: pid 31785 timed out
>>>>>> Apr 15 22:01:40 xencluster1 crmd: [8096]: ERROR: process_lrm_event:
>> LRM
>>>>>> operation clvm-xenvg:0_monitor_10000 (92) Timed Out (timeout=30000ms)
>>>>>> Apr 15 22:01:42 xencluster1 corosync[8071]:   [TOTEM ] Retransmit
>> List:
>>>>>> 21ac
>>>>>> Apr 15 22:01:42 xencluster1 corosync[8071]:   [TOTEM ] Retransmit
>> List:
>>>>>> 21ab
>>>>>> Apr 15 22:01:42 xencluster1 corosync[8071]:   [TOTEM ] Retransmit
>> List:
>>>>>> 21ac
>>>>>> Apr 15 22:01:42 xencluster1 corosync[8071]:   [TOTEM ] Retransmit
>> List:
>>>>>> 21ab
>>>>>>
>>>>>>
>>>>>> pr 15 22:01:50 xencluster1 clvmd[8763]: Send local reply
>>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Read on local socket 14, len
>> =
>>>> 32
>>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: check_all_clvmds_running
>>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: down_callback. node
>> 1226064394,
>>>>>> state = 3
>>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: down_callback. node
>> 1309950474,
>>>>>> state = 3
>>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Got pre command condition...
>>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Writing status 0 down pipe 16
>>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Waiting to do post command -
>>>>>> state = 0
>>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: read on PIPE 15: 4 bytes:
>>>> status:
>>>>>> 0
>>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: background routine status was
>>>>>> 0, sock_client=0x7f5c40057c10
>>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: distribute command: XID =
>> 15473
>>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: num_nodes = 2
>>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: add_to_lvmqueue:
>>>>>> cmd=0x7f5c4000dbd0. client=0x7f5c40057c10, msg=0x7f5c40057880, len=32,
>>>>>> csid=(nil), xid=15473
>>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Sending message to all
>> cluster
>>>>>> nodes
>>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: process_work_item: local
>>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: process_local_command:
>> unknown
>>>>>> (0x2d) msg=0x7f5c40010e80, msglen =32, client=0x7f5c40057c10
>>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Syncing device names
>>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Reply from node 4914420a: 0
>>>> bytes
>>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Got 1 replies, expecting: 2
>>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: LVM thread waiting for work
>>>>>> Apr 15 22:01:50 xencluster1 corosync[8071]:   [TOTEM ] Retransmit
>> List:
>>>>>> 21ab
>>>>>> Apr 15 22:01:50 xencluster1 corosync[8071]:   [TOTEM ] Retransmit
>> List:
>>>>>> 21ac
>>>>>> Apr 15 22:01:50 xencluster1 corosync[8071]:   [TOTEM ] Retransmit
>> List:
>>>>>> 21ab
>>>>>> ...
>>>>>> Apr 15 22:02:19 xencluster1 corosync[8071]:   [TOTEM ] Retransmit
>> List:
>>>>>> 21ac
>>>>>> Apr 15 22:02:20 xencluster1 corosync[8071]:   [TOTEM ] Retransmit
>> List:
>>>>>> 21ab
>>>>>> Apr 15 22:02:20 xencluster1 lrmd: [8093]: WARN: clvm-xenvg:0:monitor
>>>>>> process (PID 771) timed out (try 1).  Killing with signal SIGTERM
>> (15).
>>>>>> Apr 15 22:02:20 xencluster1 lrmd: [8093]: WARN: operation monitor[92]
>> on
>>>>>> clvm-xenvg:0 for client 8096: pid 771 timed out
>>>>>> Apr 15 22:02:20 xencluster1 crmd: [8096]: ERROR: process_lrm_event:
>> LRM
>>>>>> operation clvm-xenvg:0_monitor_10000 (92) Timed Out (timeout=30000ms)
>>>>>> Apr 15 22:02:20 xencluster1 corosync[8071]:   [TOTEM ] Retransmit
>> List:
>>>>>> 21ac
>>>>>> Apr 15 22:02:20 xencluster1 corosync[8071]:   [TOTEM ] Retransmit
>> List:
>>>>>> 21ab
>>>>>> ...
>>>>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: Got new connection on fd 17
>>>>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: Read on local socket 17, len
>> =
>>>> 28
>>>>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: creating pipe, [18, 19]
>>>>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: Creating pre&post thread
>>>>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: Created pre&post thread,
>> state
>>>> = 0
>>>>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: in sub thread: client =
>>>>>> 0x7f5c40058080
>>>>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: Sub thread ready for work.
>>>>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: doing PRE command LOCK_VG
>>>>>> 'V_XenVG' at 1 (client=0x7f5c40058080)
>>>>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: lock_resource 'V_XenVG',
>>>>>> flags=0, mode=3
>>>>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: lock_resource returning 0,
>>>>>> lock_id=3a0001
>>>>>>
>>>>>>
>>>>>>
>>>>>> Apr 15 22:02:48 xencluster1 corosync[8071]:   [TOTEM ] Retransmit
>> List:
>>>>>> 21ab
>>>>>> Apr 15 22:02:48 xencluster1 cib: [8092]: info: cib_stats: Processed 4
>>>>>> operations (2500.00us average, 0% utilization) in the last 10min
>>>>>> Apr 15 22:02:48 xencluster1 corosync[8071]:   [TOTEM ] Retransmit
>> List:
>>>>>> 21ac
>>>>>>
>>>>>>
>>>>>>
>>>>>> Apr 15 22:03:00 xencluster1 corosync[8071]:   [TOTEM ] Retransmit
>> List:
>>>>>> 21ac
>>>>>> Apr 15 22:03:00 xencluster1 lrmd: [8093]: WARN: clvm-xenvg:0:monitor
>>>>>> process (PID 2256) timed out (try 1).  Killing with signal SIGTERM
>> (15).
>>>>>> Apr 15 22:03:00 xencluster1 lrmd: [8093]: WARN: operation monitor[92]
>> on
>>>>>> clvm-xenvg:0 for client 8096: pid 2256 timed out
>>>>>> Apr 15 22:03:00 xencluster1 crmd: [8096]: ERROR: process_lrm_event:
>> LRM
>>>>>> operation clvm-xenvg:0_monitor_10000 (92) Timed Out (timeout=30000ms)
>>>>>> Apr 15 22:03:00 xencluster1 corosync[8071]:   [TOTEM ] Retransmit
>> List:
>>>>>> 21ab
>>>>>>
>>>>>>
>>>>>> Node2 Log
>>>>>>
>>>>>> Apr 15 22:00:51 xencluster2 clvmd[12666]: Joined child thread
>>>>>> Apr 15 22:00:51 xencluster2 clvmd[12666]: ret == 0, errno = 0.
>> removing
>>>>>> client
>>>>>> Apr 15 22:00:51 xencluster2 clvmd[12666]: add_to_lvmqueue:
>>>>>> cmd=0x7fa3d0028710. client=0x7fa3d0028760, msg=(nil), len=0,
>> csid=(nil),
>>>>>> xid=14778
>>>>>> Apr 15 22:00:51 xencluster2 clvmd[12666]: process_work_item: free fd
>> -1
>>>>>> Apr 15 22:00:51 xencluster2 clvmd[12666]: LVM thread waiting for work
>>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: crm_timer_popped:
>>>>>> PEngine Recheck Timer (I_PE_CALC) just popped (900000ms)
>>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_state_transition:
>>>>>> State transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC
>>>>>> cause=C_TIMER_POPPED origin=crm_timer_popped ]
>>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_state_transition:
>>>>>> Progressed to state S_POLICY_ENGINE after C_TIMER_POPPED
>>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_state_transition:
>> All
>>>>>> 2 cluster nodes are eligible to run resources.
>>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_pe_invoke: Query
>> 983:
>>>>>> Requesting the current CIB: S_POLICY_ENGINE
>>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_pe_invoke_callback:
>>>>>> Invoking the PE: query=983, ref=pe_calc-dc-1334520052-604, seq=2212,
>>>>>> quorate=1
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_config: On
>>>>>> loss of CCM Quorum: Ignore
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
>>>>>> Operation clvm:1_last_failure_0 found resource clvm:1 active on
>>>> xencluster2
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
>>>>>> Operation STONITH-Node2_last_failure_0 found resource STONITH-Node2
>>>>>> active on xencluster2
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
>>>>>> Operation STONITH--Node1_last_failure_0 found resource STONITH--Node1
>>>>>> active on xencluster2
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
>>>>>> Operation cmirror:1_last_failure_0 found resource cmirror:1 active on
>>>>>> xencluster2
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
>>>>>> Operation dlm:1_last_failure_0 found resource dlm:1 active on
>>>> xencluster2
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
>>>>>> Operation o2cb:1_last_failure_0 found resource o2cb:1 active on
>>>> xencluster2
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
>>>>>> Operation dlm:0_last_failure_0 found resource dlm:0 active on
>>>> xencluster1
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
>>>>>> Operation clvm:0_last_failure_0 found resource clvm:0 active on
>>>> xencluster1
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
>>>>>> Operation o2cb:0_last_failure_0 found resource o2cb:0 active on
>>>> xencluster1
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
>>>>>> Operation clvm-xenvg:0_last_failure_0 found resource clvm-xenvg:0
>> active
>>>>>> on xencluster1
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: WARN: unpack_rsc_op:
>>>>>> Processing failed op vm-deprepo_last_failure_0 on xencluster1: unknown
>>>>>> error (1)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  STONITH-Node2  (Started xencluster2)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  STONITH--Node1 (Started xencluster2)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  dlm:0  (Started xencluster1)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  clvm:0 (Started xencluster1)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  o2cb:0 (Started xencluster1)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  cmirror:0      (Started xencluster1)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  dlm:1  (Started xencluster2)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  clvm:1 (Started xencluster2)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  o2cb:1 (Started xencluster2)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  cmirror:1      (Started xencluster2)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  clvm-xenvg:0   (Started xencluster1)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  fs-config-xen:0        (Started xencluster1)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  clvm-xenvg:1   (Started xencluster2)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  fs-config-xen:1        (Started xencluster2)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  vm-ilmt        (Started xencluster1)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  vm-topdesk-entw        (Started xencluster1)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  vm-otbridge    (Started xencluster2)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  vm-virenscanner        (Started xencluster1)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  vm-deprepo     (Started xencluster1)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  vm-traumschiff (Started xencluster2)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  vm-cim (Started xencluster2)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  vm-oneiroi     (Started xencluster1)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  vm-intranet    (Started xencluster1)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  vm-soa (Started xencluster1)
>>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice:
>> process_pe_message:
>>>>>> Transition 80: PEngine Input stored in:
>>>> /var/lib/pengine/pe-input-271.bz2
>>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_state_transition:
>>>>>> State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [
>>>>>> input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
>>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: unpack_graph: Unpacked
>>>>>> transition 80: 0 actions in 0 synapses
>>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_te_invoke:
>> Processing
>>>>>> graph 80 (ref=pe_calc-dc-1334520052-604) derived from
>>>>>> /var/lib/pengine/pe-input-271.bz2
>>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: run_graph:
>>>>>> ====================================================
>>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: notice: run_graph:
>> Transition
>>>>>> 80 (Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0,
>>>>>> Source=/var/lib/pengine/pe-input-271.bz2): Complete
>>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: te_graph_trigger:
>>>>>> Transition 80 is now complete
>>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: notify_crmd:
>> Transition
>>>>>> 80 status: done - <null>
>>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_state_transition:
>>>>>> State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS
>>>>>> cause=C_FSA_INTERNAL origin=notify_crmd ]
>>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_state_transition:
>>>>>> Starting PEngine Recheck Timer
>>>>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: 1309950474 got message from
>>>>>> nodeid 1226064394 for 0. len 32
>>>>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: add_to_lvmqueue:
>>>>>> cmd=0x7fa3d0028780. client=0x6934a0, msg=0x7fa3d910063c, len=32,
>>>>>> csid=0x7fffb4e5d944, xid=0
>>>>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: process_work_item: remote
>>>>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: process_remote_command
>> unknown
>>>>>> (0x2d) for clientid 0x5000000 XID 15337 on node 4914420a
>>>>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: Syncing device names
>>>>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: LVM thread waiting for work
>>>>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: 1309950474 got message from
>>>>>> nodeid 1309950474 for 1226064394. len 18
>>>>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: 1309950474 got message from
>>>>>> nodeid 1226064394 for 0. len 32
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> Apr 15 22:01:02 xencluster2 clvmd[12666]: Joined child thread
>>>>>> Apr 15 22:01:02 xencluster2 clvmd[12666]: ret == 0, errno = 0.
>> removing
>>>>>> client
>>>>>> Apr 15 22:01:02 xencluster2 clvmd[12666]: add_to_lvmqueue:
>>>>>> cmd=0x7fa3d004d7a0. client=0x7fa3d004d7f0, msg=(nil), len=0,
>> csid=(nil),
>>>>>> xid=14848
>>>>>> Apr 15 22:01:02 xencluster2 clvmd[12666]: process_work_item: free fd
>> -1
>>>>>> Apr 15 22:01:02 xencluster2 clvmd[12666]: LVM thread waiting for work
>>>>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info: stonith_command:
>>>>>> Processed st_execute from lrmd: rc=-1
>>>>>> Apr 15 22:01:06 xencluster2 stonith: parse config info
>> info=xencluster1
>>>>>> Apr 15 22:01:06 xencluster2 stonith: meatware device OK.
>>>>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info: log_operation:
>>>>>> STONITH--Node1: Performing: stonith -t meatware -S
>>>>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info: log_operation:
>>>>>> STONITH--Node1: success:  0
>>>>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info:
>>>>>> stonith_device_execute: Nothing to do for STONITH--Node1
>>>>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info: stonith_command:
>>>>>> Processed st_execute from lrmd: rc=-1
>>>>>> Apr 15 22:01:06 xencluster2 stonith: parse config info
>> info=xencluster2
>>>>>> Apr 15 22:01:06 xencluster2 stonith: meatware device OK.
>>>>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info: log_operation:
>>>>>> STONITH-Node2: Performing: stonith -t meatware -S
>>>>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info: log_operation:
>>>>>> STONITH-Node2: success:  0
>>>>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info:
>>>>>> stonith_device_execute: Nothing to do for STONITH-Node2
>>>>>> Apr 15 22:01:10 xencluster2 clvmd[12666]: 1309950474 got message from
>>>>>> nodeid 1226064394 for 0. len 32
>>>>>> Apr 15 22:01:10 xencluster2 clvmd[12666]: add_to_lvmqueue:
>>>>>> cmd=0x7fa3d004d810. client=0x6934a0, msg=0x7fa3d9100cfc, len=32,
>>>>>> csid=0x7fffb4e5d944, xid=0
>>>>>> Apr 15 22:01:10 xencluster2 clvmd[12666]: process_work_item: remote
>>>>>> Apr 15 22:01:10 xencluster2 clvmd[12666]: process_remote_command
>> unknown
>>>>>> (0x2d) for clientid 0x5000000 XID 15407 on node 4914420a
>>>>>> Apr 15 22:01:10 xencluster2 clvmd[12666]: Syncing device names
>>>>>> Apr 15 22:01:10 xencluster2 clvmd[12666]: LVM thread waiting for work
>>>>>> Apr 15 22:01:10 xencluster2 clvmd[12666]: 1309950474 got message from
>>>>>> nodeid 1309950474 for 1226064394. len 18
>>>>>> Apr 15 22:01:10 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196
>>>>>> Apr 15 22:01:10 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196
>>>>>> Apr 15 22:01:10 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196
>>>>>> ....
>>>>>> Apr 15 22:01:11 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196
>>>>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: Got new connection on fd 5
>>>>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: Read on local socket 5, len
>> =
>>>> 28
>>>>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: creating pipe, [12, 13]
>>>>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: Creating pre&post thread
>>>>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: Created pre&post thread,
>>>> state =
>>>>>> 0
>>>>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: in sub thread: client =
>>>>>> 0x7fa3d004d810
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: Syncing device names
>>>>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: Reply from node 4e14420a: 0
>>>> bytes
>>>>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: Got 1 replies, expecting: 2
>>>>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: LVM thread waiting for work
>>>>>> Apr 15 22:01:12 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196
>>>>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: 1309950474 got message from
>>>>>> nodeid 1309950474 for 0. len 32
>>>>>> Apr 15 22:01:12 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196
>>>>>> Apr 15 22:01:12 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196
>>>>>> ....
>>>>>> Apr 15 22:01:21 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197
>>>>>> Apr 15 22:01:21 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197
>>>>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info: stonith_command:
>>>>>> Processed st_execute from lrmd: rc=-1
>>>>>> Apr 15 22:01:21 xencluster2 stonith: parse config info
>> info=xencluster1
>>>>>> Apr 15 22:01:21 xencluster2 stonith: meatware device OK.
>>>>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info: log_operation:
>>>>>> STONITH--Node1: Performing: stonith -t meatware -S
>>>>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info: log_operation:
>>>>>> STONITH--Node1: success:  0
>>>>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info:
>>>>>> stonith_device_execute: Nothing to do for STONITH--Node1
>>>>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info: stonith_command:
>>>>>> Processed st_execute from lrmd: rc=-1
>>>>>> Apr 15 22:01:21 xencluster2 stonith: parse config info
>> info=xencluster2
>>>>>> Apr 15 22:01:21 xencluster2 stonith: meatware device OK.
>>>>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info: log_operation:
>>>>>> STONITH-Node2: Performing: stonith -t meatware -S
>>>>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info: log_operation:
>>>>>> STONITH-Node2: success:  0
>>>>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info:
>>>>>> stonith_device_execute: Nothing to do for STONITH-Node2
>>>>>> Apr 15 22:01:21 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197
>>>>>> Apr 15 22:01:22 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197
>>>>>> Apr 15 22:01:22 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197
>>>>>> ....
>>>>>> Apr 15 22:01:36 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197
>>>>>> Apr 15 22:01:36 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197
>>>>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info: stonith_command:
>>>>>> Processed st_execute from lrmd: rc=-1
>>>>>> Apr 15 22:01:36 xencluster2 stonith: parse config info
>> info=xencluster1
>>>>>> Apr 15 22:01:36 xencluster2 stonith: meatware device OK.
>>>>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info: log_operation:
>>>>>> STONITH--Node1: Performing: stonith -t meatware -S
>>>>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info: log_operation:
>>>>>> STONITH--Node1: success:  0
>>>>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info:
>>>>>> stonith_device_execute: Nothing to do for STONITH--Node1
>>>>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info: stonith_command:
>>>>>> Processed st_execute from lrmd: rc=-1
>>>>>> Apr 15 22:01:36 xencluster2 stonith: parse config info
>> info=xencluster2
>>>>>> Apr 15 22:01:36 xencluster2 stonith: meatware device OK.
>>>>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info: log_operation:
>>>>>> STONITH-Node2: Performing: stonith -t meatware -S
>>>>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info: log_operation:
>>>>>> STONITH-Node2: success:  0
>>>>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info:
>>>>>> stonith_device_execute: Nothing to do for STONITH-Node2
>>>>>> Apr 15 22:01:36 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197
>>>>>> Apr 15 22:01:37 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197
>>>>>> ....
>>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197
>>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197
>>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: process_graph_event:
>>>>>> Detected action clvm-xenvg:0_monitor_10000 from a different
>> transition:
>>>>>> 62 vs. 80
>>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info:
>> abort_transition_graph:
>>>>>> process_graph_event:476 - Triggered transition abort (complete=1,
>>>>>> tag=lrm_rsc_op, id=clvm-xenvg:0_monitor_10000,
>>>>>> magic=2:-2;61:62:0:72497771-fdb2-4ca7-a723-2cc360861bb0,
>> cib=0.5816.4) :
>>>>>> Old event
>>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: WARN: update_failcount:
>>>>>> Updating failcount for clvm-xenvg:0 on xencluster1 after failed
>> monitor:
>>>>>> rc=-2 (update=value++, time=1334520100)
>>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: do_state_transition:
>>>>>> State transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC
>>>>>> cause=C_FSA_INTERNAL origin=abort_transition_graph ]
>>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: do_state_transition:
>> All
>>>>>> 2 cluster nodes are eligible to run resources.
>>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: do_pe_invoke: Query
>> 984:
>>>>>> Requesting the current CIB: S_POLICY_ENGINE
>>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197
>>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197
>>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197
>>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197
>>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197
>>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197
>>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197 219a 219b 219c 219d
>>>>>> ...
>>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197 219a 219b 219c 219d
>>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: do_pe_invoke_callback:
>>>>>> Invoking the PE: query=984, ref=pe_calc-dc-1334520100-605, seq=2212,
>>>>>> quorate=1
>>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197 219a 219b 219c 219d
>>>>>> ...
>>>>>>
>>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197 219a 219b 219c 219d
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_config: On
>>>>>> loss of CCM Quorum: Ignore
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
>>>>>> Operation clvm:1_last_failure_0 found resource clvm:1 active on
>>>> xencluster2
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
>>>>>> Operation STONITH-Node2_last_failure_0 found resource STONITH-Node2
>>>>>> active on xencluster2
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
>>>>>> Operation STONITH--Node1_last_failure_0 found resource STONITH--Node1
>>>>>> active on xencluster2
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
>>>>>> Operation cmirror:1_last_failure_0 found resource cmirror:1 active on
>>>>>> xencluster2
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
>>>>>> Operation dlm:1_last_failure_0 found resource dlm:1 active on
>>>> xencluster2
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
>>>>>> Operation o2cb:1_last_failure_0 found resource o2cb:1 active on
>>>> xencluster2
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
>>>>>> Operation cmirror:0_last_failure_0 found resource cmirror:0 active on
>>>>>> xencluster1
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
>>>>>> Operation dlm:0_last_failure_0 found resource dlm:0 active on
>>>> xencluster1
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
>>>>>> Operation clvm:0_last_failure_0 found resource clvm:0 active on
>>>> xencluster1
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
>>>>>> Operation o2cb:0_last_failure_0 found resource o2cb:0 active on
>>>> xencluster1
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op:
>>>>>> Operation clvm-xenvg:0_last_failure_0 found resource clvm-xenvg:0
>> active
>>>>>> on xencluster1
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: WARN: unpack_rsc_op:
>>>>>> Processing failed op clvm-xenvg:0_monitor_10000 on xencluster1:
>> unknown
>>>>>> exec error (-2)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: WARN: unpack_rsc_op:
>>>>>> Processing failed op vm-deprepo_last_failure_0 on xencluster1: unknown
>>>>>> error (1)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: RecurringOp:
>>  Start
>>>>>> recurring monitor (10s) for clvm-xenvg:0 on xencluster1
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  STONITH-Node2  (Started xencluster2)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  STONITH--Node1 (Started xencluster2)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  dlm:0  (Started xencluster1)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  clvm:0 (Started xencluster1)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  o2cb:0 (Started xencluster1)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  cmirror:0      (Started xencluster1)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  dlm:1  (Started xencluster2)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  clvm:1 (Started xencluster2)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  o2cb:1 (Started xencluster2)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  cmirror:1      (Started xencluster2)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions:
>> Recover
>>>>>> clvm-xenvg:0   (Started xencluster1)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions:
>> Restart
>>>>>> fs-config-xen:0        (Started xencluster1)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  clvm-xenvg:1   (Started xencluster2)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  fs-config-xen:1        (Started xencluster2)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions:
>> Restart
>>>>>> vm-ilmt        (Started xencluster1)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions:
>> Restart
>>>>>> vm-topdesk-entw        (Started xencluster1)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  vm-otbridge    (Started xencluster2)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions:
>> Restart
>>>>>> vm-virenscanner        (Started xencluster1)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions:
>> Restart
>>>>>> vm-deprepo     (Started xencluster1)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  vm-traumschiff (Started xencluster2)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave
>>>>>>  vm-cim (Started xencluster2)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions:
>> Restart
>>>>>> vm-oneiroi     (Started xencluster1)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions:
>> Restart
>>>>>> vm-intranet    (Started xencluster1)
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions:
>> Restart
>>>>>> vm-soa (Started xencluster1)
>>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: do_state_transition:
>>>>>> State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [
>>>>>> input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ]
>>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: unpack_graph: Unpacked
>>>>>> transition 81: 39 actions in 39 synapses
>>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: do_te_invoke:
>> Processing
>>>>>> graph 81 (ref=pe_calc-dc-1334520100-605) derived from
>>>>>> /var/lib/pengine/pe-input-272.bz2
>>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: te_rsc_command:
>>>>>> Initiating action 82: stop vm-ilmt_stop_0 on xencluster1
>>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: te_rsc_command:
>>>>>> Initiating action 86: stop vm-topdesk-entw_stop_0 on xencluster1
>>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: te_rsc_command:
>>>>>> Initiating action 90: stop vm-virenscanner_stop_0 on xencluster1
>>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: te_rsc_command:
>>>>>> Initiating action 92: stop vm-deprepo_stop_0 on xencluster1
>>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: te_rsc_command:
>>>>>> Initiating action 98: stop vm-oneiroi_stop_0 on xencluster1
>>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: te_rsc_command:
>>>>>> Initiating action 100: stop vm-intranet_stop_0 on xencluster1
>>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: te_rsc_command:
>>>>>> Initiating action 102: stop vm-soa_stop_0 on xencluster1
>>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice:
>> process_pe_message:
>>>>>> Transition 81: PEngine Input stored in:
>>>> /var/lib/pengine/pe-input-272.bz2
>>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197 219a 219b 219c 219d
>>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197 219a 219b 219c 219d
>>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197 219a 219b 219c 219d
>>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197 219a 219b 219c 219d
>>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197 219a 219b 219c 219d
>>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197 219a 219b 219c 219d 219e 219f 21a0 21a1 21a2 21a3 21a4 21a5
>>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197 219a 219b 219c 219d 219e 219f 21a0 21a1 21a2 21a3 21a4 21a5
>>>>>> ...
>>>>>> Apr 15 22:01:41 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197 219a 219b 219c 219d 219e 219f 21a0 21a1 21a2 21a3 21a4 21a5
>>>>>> Apr 15 22:01:42 xencluster2 lrmd: [7675]: WARN: clvm-xenvg:1:monitor
>>>>>> process (PID 29785) timed out (try 1).  Killing with signal SIGTERM
>>>> (15).
>>>>>> Apr 15 22:01:42 xencluster2 lrmd: [7675]: WARN: operation monitor[125]
>>>>>> on clvm-xenvg:1 for client 7678: pid 29785 timed out
>>>>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: ERROR: process_lrm_event:
>> LRM
>>>>>> operation clvm-xenvg:1_monitor_10000 (125) Timed Out (timeout=30000ms)
>>>>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: info: process_graph_event:
>>>>>> Detected action clvm-xenvg:1_monitor_10000 from a different
>> transition:
>>>>>> 68 vs. 81
>>>>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: info:
>> abort_transition_graph:
>>>>>> process_graph_event:476 - Triggered transition abort (complete=0,
>>>>>> tag=lrm_rsc_op, id=clvm-xenvg:1_monitor_10000,
>>>>>> magic=2:-2;61:68:0:72497771-fdb2-4ca7-a723-2cc360861bb0,
>> cib=0.5816.5) :
>>>>>> Old event
>>>>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: info: update_abort_priority:
>>>>>> Abort priority upgraded from 0 to 1000000
>>>>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: info: update_abort_priority:
>>>>>> Abort action done superceeded by restart
>>>>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: WARN: update_failcount:
>>>>>> Updating failcount for clvm-xenvg:1 on xencluster2 after failed
>> monitor:
>>>>>> rc=-2 (update=value++, time=1334520102)
>>>>>> Apr 15 22:01:42 xencluster2 attrd: [7676]: notice:
>> attrd_trigger_update:
>>>>>> Sending flush op to all hosts for: fail-count-clvm-xenvg:1 (1)
>>>>>> Apr 15 22:01:42 xencluster2 attrd: [7676]: notice:
>> attrd_perform_update:
>>>>>> Sent update 448: fail-count-clvm-xenvg:1=1
>>>>>> Apr 15 22:01:42 xencluster2 attrd: [7676]: notice:
>> attrd_trigger_update:
>>>>>> Sending flush op to all hosts for: last-failure-clvm-xenvg:1
>>>> (1334520102)
>>>>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: info:
>> abort_transition_graph:
>>>>>> te_update_diff:176 - Triggered transition abort (complete=0,
>> tag=nvpair,
>>>>>> id=status-xencluster2-fail-count-clvm-xenvg.1,
>>>>>> name=fail-count-clvm-xenvg:1, value=1, magic=NA, cib=0.5816.6) :
>>>>>> Transient attribute: update
>>>>>> Apr 15 22:01:42 xencluster2 attrd: [7676]: notice:
>> attrd_perform_update:
>>>>>> Sent update 451: last-failure-clvm-xenvg:1=1334520102
>>>>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: info:
>> abort_transition_graph:
>>>>>> te_update_diff:176 - Triggered transition abort (complete=0,
>> tag=nvpair,
>>>>>> id=status-xencluster2-last-failure-clvm-xenvg.1,
>>>>>> name=last-failure-clvm-xenvg:1, value=1334520102, magic=NA,
>>>>>> cib=0.5816.7) : Transient attribute: update
>>>>>> Apr 15 22:01:42 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197 219a 219b 219c 219d 219e 219f 21a0 21a1 21a2 21a3 21a4 21a5
>>>>>> Apr 15 22:01:42 xencluster2 corosync[7666]:   [TOTEM ] Retransmit
>> List:
>>>>>> 2196 2197 219a 219b 219c 219d 219e 219f 21a0 21a1 21a2 21a3 21a4 21a5
>>>>>>
>>>>>> --
>>>>>> Mit freundlichen Gr??en
>>>>>>
>>>>>> Frank Meier
>>>>>> UNIX-Basis
>>>>>>
>>>>>> Hamm Reno Group GmbH
>>>>>> Industriegebiet West | D-66987 Thaleischweiler-Fr?schen
>>>>>> T.+49(0)6334 444-8322 | F.+49(0)6334 444-8190
>>>>>> frank.meier at hr-group.de | www.reno.de
>>>>>> ___________________________________________________________________
>>>>>>
>>>>>> Sitz: Am Tie 7 | D-49086 Osnabr?ck
>>>>>> Handelsregister Osnabr?ck HRB 19587
>>>>>> Gesch?ftsf?hrer: Hans-J?rgen de Fries,
>>>>>> Jens Gransee, Manfred Klumpp,
>>>>>> Robert Reisch
>>>>>>
>>>>>> _______________________________________________
>>>>>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>>
>>>>>> Project Home: http://www.clusterlabs.org
>>>>>> Getting started:
>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>>>> Bugs: http://bugs.clusterlabs.org
>>>>>>
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> esta es mi vida e me la vivo hasta que dios quiera
>>>>> -------------- next part --------------
>>>>> An HTML attachment was scrubbed...
>>>>> URL: <
>>>>
>> http://oss.clusterlabs.org/pipermail/pacemaker/attachments/20120419/00a36dbe/attachment.html
>>>>>
>>>>>
>>>>> ------------------------------
>>>>>
>>>>> _______________________________________________
>>>>> Pacemaker mailing list
>>>>> Pacemaker at oss.clusterlabs.org
>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>>
>>>>>
>>>>> End of Pacemaker Digest, Vol 53, Issue 38
>>>>> *****************************************
>>>>
>>>> _______________________________________________
>>>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>
>>>> Project Home: http://www.clusterlabs.org
>>>> Getting started:
>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>> Bugs: http://bugs.clusterlabs.org
>>>>
>>>
>>>
>>>
>>> --
>>> esta es mi vida e me la vivo hasta que dios quiera
>>> -------------- next part --------------
>>> An HTML attachment was scrubbed...
>>> URL: <
>> http://oss.clusterlabs.org/pipermail/pacemaker/attachments/20120419/63b888a3/attachment.html
>>>
>>>
>>> ------------------------------
>>>
>>> _______________________________________________
>>> Pacemaker mailing list
>>> Pacemaker at oss.clusterlabs.org
>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>
>>>
>>> End of Pacemaker Digest, Vol 53, Issue 40
>>> *****************************************
>>
>> _______________________________________________
>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>
>> Project Home: http://www.clusterlabs.org
>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> Bugs: http://bugs.clusterlabs.org
>>
> 
> 
> 
> --
> esta es mi vida e me la vivo hasta que dios quiera
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL: <http://oss.clusterlabs.org/pipermail/pacemaker/attachments/20120419/f21bf7c7/attachment.html>
> 
> ------------------------------
> 
> _______________________________________________
> Pacemaker mailing list
> Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> 
> 
> End of Pacemaker Digest, Vol 53, Issue 42
> *****************************************




More information about the Pacemaker mailing list