[Pacemaker] answer:Pacemaker Digest, Vol 33, Issue 65

jiaju liu liujiaju86 at yahoo.com.cn
Mon Aug 30 21:11:49 EDT 2010



--- 10年8月30日,周一, pacemaker-request at oss.clusterlabs.org <pacemaker-request at oss.clusterlabs.org> 写道:


发件人: pacemaker-request at oss.clusterlabs.org <pacemaker-request at oss.clusterlabs.org>
主题: Pacemaker Digest, Vol 33, Issue 65
收件人: pacemaker at oss.clusterlabs.org
日期: 2010年8月30日,周一,下午5:11


Send Pacemaker mailing list submissions to
    pacemaker at oss.clusterlabs.org

To subscribe or unsubscribe via the World Wide Web, visit
    http://oss.clusterlabs.org/mailman/listinfo/pacemaker
or, via email, send a message with subject or body 'help' to
    pacemaker-request at oss.clusterlabs.org

You can reach the person managing the list at
    pacemaker-owner at oss.clusterlabs.org

When replying, please edit your Subject line so it is more specific
than "Re: Contents of Pacemaker digest..."


Today's Topics:

   1. some problem about install pacemaker (jiaju liu)
   2. Re: some problem about install pacemaker (Andrew Beekhof)
   3. Re: drbd diskless -> failover to other node (jimbob palmer)
   4. Re: ocf:pacemaker:o2cb Unable to connect to CKPT
      (Michael Schwartzkopff)
   5. Re: Quorum disk? (Michael Schwartzkopff)
   6. Re: how to keep ftp connection when swap from primary to
      secondary (Michael Schwartzkopff)
   7. Re: Howto upgrade Pacemaker cluster from Version: 1.0.2 to
      the last released on clusterlabs (Roberto Giordani)
   8. Re: cluster-dlm: set_fs_notified: set_fs_notified no nodeid
      1812048064#012 (Roberto Giordani)


----------------------------------------------------------------------

Message: 1
Date: Mon, 30 Aug 2010 08:40:20 +0800 (CST)
From: jiaju liu <liujiaju86 at yahoo.com.cn>
To: pacemaker at oss.clusterlabs.org
Subject: [Pacemaker] some problem about install pacemaker
Message-ID: <897653.16221.qm at web15703.mail.cnb.yahoo.com>
Content-Type: text/plain; charset="iso-8859-1"

hello 
when I use yum install -y pacemaker corosync heartbeat to install I meet a problem
?
libopenais-devel-0.80.5-15.1.x86_64 from installed has depsolving problems
? --> Missing Dependency: openais = 0.80.5-15.1 is needed by package libopenais-devel-0.80.5-15.1.x86_64 (installed)
Error: Missing Dependency: openais = 0.80.5-15.1 is needed by package libopenais-devel-0.80.5-15.1.x86_64 (installed)
?You could try using --skip-broken to work around the problem
?You could try running: package-cleanup --problems
??????????????????????? package-cleanup --dupes
??????????????????????? rpm -Va --nofiles --nodigest
?
Thanks


      
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://oss.clusterlabs.org/pipermail/pacemaker/attachments/20100830/007989f2/attachment-0001.htm>

------------------------------

Message: 2
Date: Mon, 30 Aug 2010 08:17:17 +0200
From: Andrew Beekhof <andrew at beekhof.net>
To: The Pacemaker cluster resource manager
    <pacemaker at oss.clusterlabs.org>
Subject: Re: [Pacemaker] some problem about install pacemaker
Message-ID:
    <AANLkTin5Lw2LqVk6eDUBbv5HaKC_1FhT5fVYfKj8znXA at mail.gmail.com>
Content-Type: text/plain; charset="iso-8859-1"

On Mon, Aug 30, 2010 at 2:40 AM, jiaju liu <liujiaju86 at yahoo.com.cn> wrote:

> hello
> when I use *yum install -y pacemaker corosync heartbeat* to install I meet
> a problem
>

Which version of pacemaker? from where? on which distro?


thank you for your reply
pacemaker vision is
pacemaker-libs-1.0.9.1-1.15.el5
pacemaker-1.0.9.1-1.15.el5
 
and I use cent os 5.4
I have never used pacemaker before so would you please give me some advice and send me some document   about how to install pacemaker corectly?
Thank you a lot
>
> libopenais-devel-0.80.5-15.1.x86_64 from installed has depsolving problems
>   --> Missing Dependency: openais = 0.80.5-15.1 is needed by package
> libopenais-devel-0.80.5-15.1.x86_64 (installed)
> Error: Missing Dependency: openais = 0.80.5-15.1 is needed by package
> libopenais-devel-0.80.5-15.1.x86_64 (installed)
>  You could try using --skip-broken to work around the problem
>  You could try running: package-cleanup --problems
>                         package-cleanup --dupes
>                         rpm -Va --nofiles --nodigest
>
> Thanks
>
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs:
> http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://oss.clusterlabs.org/pipermail/pacemaker/attachments/20100830/257aadb9/attachment-0001.htm>

------------------------------

Message: 3
Date: Mon, 30 Aug 2010 11:06:09 +0200
From: jimbob palmer <jimbobpalmer at gmail.com>
To: The Pacemaker cluster resource manager
    <pacemaker at oss.clusterlabs.org>
Subject: Re: [Pacemaker] drbd diskless -> failover to other node
Message-ID:
    <AANLkTi=qXWsFRzAPa3E3BfEXUfi4YQ9J4rtmUcffihrh at mail.gmail.com>
Content-Type: text/plain; charset=ISO-8859-1

>> Are you saying that if a server loses its disk, it will transparently
>> write to the secondary server without any need to failover at all?
>
> Yes. As long as it still has a network connection to the peer, of course.
>
>> WOW. I never knew DRBD did this. This is a _fantastic_ feature :)
>
> Well, that's what diskless mode is really all about.
> http://www.drbd.org/users-guide/s-handling-disk-errors.html

A final question: does DRBD switch to Protocol C in diskless mode, or
does it stay with the configured Protocol? If it doesn't switch, can
it be configured to?



------------------------------

Message: 4
Date: Wed, 25 Aug 2010 11:05:18 +0200
From: Michael Schwartzkopff <misch at clusterbau.com>
To: The Pacemaker cluster resource manager
    <pacemaker at oss.clusterlabs.org>
Subject: Re: [Pacemaker] ocf:pacemaker:o2cb Unable to connect to CKPT
Message-ID: <1282727118.3864.4.camel at nb001>
Content-Type: text/plain; charset="UTF-8"

Am Mittwoch, den 25.08.2010, 09:43 +0200 schrieb Andrew Beekhof:
> On Fri, Aug 6, 2010 at 3:33 PM, Michael Fung <mike at 3open.org> wrote:
> > Hi All,
> >
> >
> > I am still testing with the Debian Squeeze machine.
> >
> > Unable to start the RA ocf:pacemaker:o2cb
(...)
> 
> No. It just tells corosync to load the extra services like ckpt (part
> of openais) needed by ocfs2


Hi,

how can I tell corosync to load ckpt service?

Thanks.

> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker





------------------------------

Message: 5
Date: Thu, 26 Aug 2010 04:21:39 +0200
From: Michael Schwartzkopff <misch at clusterbau.com>
To: The Pacemaker cluster resource manager
    <pacemaker at oss.clusterlabs.org>
Subject: Re: [Pacemaker] Quorum disk?
Message-ID: <1282789299.2431.0.camel at nb001>
Content-Type: text/plain; charset="UTF-8"

Am Mittwoch, den 25.08.2010, 17:01 -0400 schrieb Ciro Iriarte:
> Hi, I'm planning to use OpanAIS+Pacemaker on SLES11-SP1 and would like
> to know if it's possible to use a quorum disk in a two-node cluster.
> The idea is to avoid adding a third node just for quorum...
> 
> Regards,

Hi,

you could have a look at the sfex resource agent.

Greetings,

Michael Schwartzkopff




------------------------------

Message: 6
Date: Thu, 26 Aug 2010 17:24:47 +0200
From: Michael Schwartzkopff <misch at clusterbau.com>
To: The Pacemaker cluster resource manager
    <pacemaker at oss.clusterlabs.org>
Subject: Re: [Pacemaker] how to keep ftp connection when swap from
    primary to    secondary
Message-ID: <1282836287.5912.6.camel at nb001>
Content-Type: text/plain; charset="UTF-8"

Am Donnerstag, den 26.08.2010, 17:17 +0200 schrieb Raoul Bhatia [IPAX]:
> On 08/26/2010 04:42 PM, Liang.Ma at asc-csa.gc.ca wrote:
> > I have followed the guide in ?Clusters from Scratch? written by Andrew
> > Beekhof and successfully setup an Active/Passive pair of cluster
> > servers. The cluster runs in Fedora 13 and includes services like
> > apache, vsftpd and nfs. Drbd is used to allow data consistence during a
> > failover. Everything works fine except ftp lose its connection when the
> > service swaps from primary to the secondary or vice versa. I know to
> > keep the ftp connection, one may need to keep the connection states for
> > the session across the nodes. But I couldn?t find clue how to do it.
> > Does anyone there have any idea how to keep the ftp connection when
> > swapping nodes, if it is possible?
> 
> hi,
> 
> as of now, we're not syncing our connections between the load
> balancers, but i would suggest
> http://www.linuxvirtualserver.org/docs/sync.html and the like.
> 
> 
> cheers,
> raoul

Even a Load Balancer wouldn't sync the data that the FTP server on the
real servers hold in RAM. You would need a cluster-aware FTP for such
purpose.

On the other hand: How often does a failover happen? Is it really
nescessary to take care for such rare events?

Michael.




------------------------------

Message: 7
Date: Sat, 28 Aug 2010 09:34:35 +0200
From: Roberto Giordani <r.giordani at tiscali.it>
To: The Pacemaker cluster resource manager
    <pacemaker at oss.clusterlabs.org>
Subject: Re: [Pacemaker] Howto upgrade Pacemaker cluster from Version:
    1.0.2 to the last released on clusterlabs
Message-ID: <4C78BC0B.1010109 at tiscali.it>
Content-Type: text/plain; charset=ISO-8859-1

Hello,
but How to migrate the entire cluster configuration (resource, nodes,
stonith)?
Regards,
Roberto.

On 08/26/2010 09:40 AM, Andrew Beekhof wrote:
> On Wed, Aug 18, 2010 at 11:15 PM, Roberto Giordani <r.giordani at libero.it> wrote:
>   
>> Hello,
>> I'd like to know how is it possible to upgrade a running cluster
>> pacemaker on Opensuse 11.2 version 1.02 to the last available on clusterlabs
>> using dlm + ocfs2 too
>>     
> The problem is that the versions of pacemaker on clusterlabs are
> probably incompatible with your existing dlm and ocfs2 packages.
> You'd need to rebuild them against the new pacemaker packages.
>
>   
>> Could someone explain in some steps how to proceed without loose all the
>> cluster configuration up and running?
>>     
> Assuming you have a compatible set of new packages (see above), just
> do a rolling upgrade.
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker
>
>   




------------------------------

Message: 8
Date: Sat, 28 Aug 2010 09:41:41 +0200
From: Roberto Giordani <r.giordani at tiscali.it>
To: The Pacemaker cluster resource manager
    <pacemaker at oss.clusterlabs.org>
Subject: Re: [Pacemaker] cluster-dlm: set_fs_notified: set_fs_notified
    no nodeid 1812048064#012
Message-ID: <4C78BDB5.7000001 at tiscali.it>
Content-Type: text/plain; charset=UTF-8

Thanks,
who should I contact? Which mailing list?
I've discovered that this problem occours when the port of my switch
where the cluster ring is connected became "blocked" due spanning tree.
I've resolved the bug using for the ring a separate switch without
spanning tre enabled and different subnet.
Is there a configuration to avoid that before the spanning tree
recalculate the route due a failure, the cluster nodes doesn't hang?
The hang occurses on SLES11sp1 too where the servers are up running, the
cluster status is ok, but when try to connect to the server with ssh,
after the login hang the session.

Usually the recalculate takes 50 seconds.

Regards,
Roberto.

On 08/26/2010 10:24 AM, Dejan Muhamedagic wrote:
> Hi,
>
> On Thu, Aug 26, 2010 at 09:36:10AM +0200, Andrew Beekhof wrote:
>   
>> On Wed, Aug 18, 2010 at 6:24 PM, Roberto Giordani <r.giordani at libero.it> wrote:
>>     
>>> Hello,
>>> I'll explain what?s happened after a network black-out
>>> I've a cluster with pacemaker on Opensuse 11.2 64bit
>>> ============
>>> Last updated: Wed Aug 18 18:13:33 2010
>>> Current DC: nodo1 (nodo1)
>>> Version: 1.0.2-ec6b0bbee1f3aa72c4c2559997e675db6ab39160
>>> 3 Nodes configured.
>>> 11 Resources configured.
>>> ============
>>>
>>> Node: nodo1 (nodo1): online
>>> Node: nodo3 (nodo3): online
>>> Node: nodo4 (nodo4): online
>>>
>>> Clone Set: dlm-clone
>>>     dlm:0       (ocf::pacemaker:controld):      Started nodo3
>>>     dlm:1       (ocf::pacemaker:controld):      Started nodo1
>>>     dlm:2       (ocf::pacemaker:controld):      Started nodo4
>>> Clone Set: o2cb-clone
>>>     o2cb:0      (ocf::ocfs2:o2cb):      Started nodo3
>>>     o2cb:1      (ocf::ocfs2:o2cb):      Started nodo1
>>>     o2cb:2      (ocf::ocfs2:o2cb):      Started nodo4
>>> Clone Set: XencfgFS-Clone
>>>     XencfgFS:0  (ocf::heartbeat:Filesystem):    Started nodo3
>>>     XencfgFS:1  (ocf::heartbeat:Filesystem):    Started nodo1
>>>     XencfgFS:2  (ocf::heartbeat:Filesystem):    Started nodo4
>>> Clone Set: XenimageFS-Clone
>>>     XenimageFS:0        (ocf::heartbeat:Filesystem):    Started nodo3
>>>     XenimageFS:1        (ocf::heartbeat:Filesystem):    Started nodo1
>>>     XenimageFS:2        (ocf::heartbeat:Filesystem):    Started nodo4
>>> rsa1-fencing    (stonith:external/ibmrsa-telnet):       Started nodo4
>>> rsa2-fencing    (stonith:external/ibmrsa-telnet):       Started nodo3
>>> rsa3-fencing    (stonith:external/ibmrsa-telnet):       Started nodo4
>>> rsa4-fencing    (stonith:external/ibmrsa-telnet):       Started nodo3
>>> mailsrv-rm      (ocf::heartbeat:Xen):   Started nodo3
>>> dbsrv-rm        (ocf::heartbeat:Xen):   Started nodo4
>>> websrv-rm       (ocf::heartbeat:Xen):   Started nodo4
>>>
>>> After a  switch failure all the nodes and the rsa stonith devices was
>>> unreachable.
>>>
>>> On the cluster happen the following error on one node
>>>
>>> Aug 18 13:11:38 nodo1 cluster-dlm: receive_plocks_stored:
>>> receive_plocks_stored 1778493632:2 need_plocks 0#012
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272025] ------------[ cut here
>>> ]------------
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272036] kernel BUG at
>>> /usr/src/packages/BUILD/kernel-xen-2.6.31.12/linux-2.6.31/fs/inode.c:1323!
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272042] invalid opcode: 0000 [#1] SMP
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272046] last sysfs file:
>>> /sys/kernel/dlm/0BB443F896254AD3BA8FB960C425B666/control
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272050] CPU 1
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272053] Modules linked in:
>>> nf_conntrack_ipv4 nf_defrag_ipv4 xt_state nf_conntrack xt_physdev
>>> iptable_filter ip_tables x_tables ocfs2 ocfs2_nodemanager quota_tree
>>> ocfs2_stack_user ocfs2_stackglue dlm configfs netbk coretemp blkbk
>>> blkback_pagemap blktap xenbus_be ipmi_si edd dm_round_robin scsi_dh_rdac
>>> dm_multipath scsi_dh bridge stp llc bonding ipv6 fuse ext4 jbd2 crc16 loop
>>> dm_mod sr_mod ide_pci_generic ide_core iTCO_wdt ata_generic ibmpex i5k_amb
>>> ibmaem iTCO_vendor_support ipmi_msghandler bnx2 i5000_edac 8250_pnp shpchp
>>> ata_piix pcspkr ics932s401 joydev edac_core i2c_i801 ses pci_hotplug 8250
>>> i2c_core serio_raw enclosure serial_core button sg reiserfs usbhid hid
>>> uhci_hcd ehci_hcd xenblk cdrom xennet fan processor pata_acpi lpfc thermal
>>> thermal_sys hwmon aacraid [last unloaded: ocfs2_stackglue]
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272111] Pid: 8889, comm: dlm_send Not
>>> tainted 2.6.31.12-0.2-xen #1 IBM System x3650 -[7979AC1]-
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272113] RIP: e030:[<ffffffff801331c2>]
>>> [<ffffffff801331c2>] iput+0x82/0x90
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272121] RSP: e02b:ffff88014ec03c30
>>> EFLAGS: 00010246
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272122] RAX: 0000000000000000 RBX:
>>> ffff880148a703c8 RCX: 0000000000000000
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272123] RDX: ffffc90000010000 RSI:
>>> ffff880148a70380 RDI: ffff880148a703c8
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272125] RBP: ffff88014ec03c50 R08:
>>> b038000000000000 R09: fe99594c51a57607
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272126] R10: ffff880040410270 R11:
>>> 0000000000000000 R12: ffff8801713e6e08
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272128] R13: ffff88014ec03d20 R14:
>>> 0000000000000000 R15: ffffc9000331d108
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272133] FS: 00007ff4cb11a730(0000)
>>> GS:ffffc90000010000(0000) knlGS:0000000000000000
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272135] CS: e033 DS: 0000 ES: 0000 CR0:
>>> 000000008005003b
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272136] CR2: 00007ff4c5c45000 CR3:
>>> 0000000135b2a000 CR4: 0000000000002660
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272138] DR0: 0000000000000000 DR1:
>>> 0000000000000000 DR2: 0000000000000000
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272140] DR3: 0000000000000000 DR6:
>>> 00000000ffff0ff0 DR7: 0000000000000400
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272142] Process dlm_send (pid: 8889,
>>> threadinfo ffff88014ec02000, task ffff8801381e45c0)
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272143] Stack:
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272144] 0000000000000000
>>> 00000000072f0874 ffff880148a70380 ffff880148a70380
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272146] <0> ffff88014ec03c80
>>> ffffffff803add09 ffff88014ec03c80 00000000072f0874
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272147] <0> ffff8801713e6df8
>>> ffff8801713e6e08 ffff88014ec03de0 ffffffffa05661e1
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272150] Call Trace:
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272164] [<ffffffff803add09>]
>>> sock_release+0x89/0xa0
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272177] [<ffffffffa05661e1>]
>>> tcp_connect_to_sock+0x161/0x2b0 [dlm]
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272206] [<ffffffffa0568764>]
>>> process_send_sockets+0x34/0x60 [dlm]
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272222] [<ffffffff800693f3>]
>>> run_workqueue+0x83/0x230
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272227] [<ffffffff80069654>]
>>> worker_thread+0xb4/0x140
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272231] [<ffffffff8006fac6>]
>>> kthread+0xb6/0xc0
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272236] [<ffffffff8000d38a>]
>>> child_rip+0xa/0x20
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272240] Code: 42 20 48 c7 c2 b0 4c 13
>>> 80 48 85 c0 48 0f 44 c2 48 89 df ff d0 48 8b 45 e8 65 48 33 04 25 28 00 00
>>> 00 75 0b 48 83 c4 18 5b c9 c3 <0f> 0b eb fe e8 35 c6 f1 ff 0f 1f 44 00 00 55
>>> 48 8d 97 10 02 00
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272256] RIP [<ffffffff801331c2>]
>>> iput+0x82/0x90
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272259] RSP <ffff88014ec03c30>
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272264] ---[ end trace 7707d0d92a7f5415
>>> ]---
>>>
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272495] dlm: connect from non cluster
>>> node
>>>
>>> and after few log lines the following line repeated until the node was
>>> killed by me
>>>
>>> Aug 18 13:12:31 nodo1 cluster-dlm: start_kernel: start_kernel cg 3
>>> member_count 1#012
>>>
>>> Aug 18 13:12:31 nodo1 cluster-dlm: update_dir_members: dir_member
>>> 1812048064#012
>>>
>>> Aug 18 13:12:31 nodo1 cluster-dlm: update_dir_members: dir_member
>>> 1778493632#012
>>>
>>> Aug 18 13:12:31 nodo1 cluster-dlm: set_configfs_members: set_members rmdir
>>> "/sys/kernel/config/dlm/cluster/spaces/0BB443F896254AD3BA8FB960C425B666/nodes/1812048064"#012
>>>
>>> Aug 18 13:12:31 nodo1 cluster-dlm: do_sysfs: write "1" to
>>> "/sys/kernel/dlm/0BB443F896254AD3BA8FB960C425B666/control"#012
>>>
>>> Aug 18 13:12:31 nodo1 cluster-dlm: set_fs_notified: set_fs_notified no
>>> nodeid 1812048064#012
>>>
>>> Aug 18 13:12:31 nodo1 cluster-dlm: set_fs_notified: set_fs_notified no
>>> nodeid 1812048064#012
>>>
>>> Aug 18 13:12:31 nodo1 cluster-dlm: set_fs_notified: set_fs_notified no
>>> nodeid 1812048064#012
>>>
>>> Aug 18 13:12:31 nodo1 cluster-dlm: set_fs_notified: set_fs_notified no
>>> nodeid 1812048064#012
>>>
>>> Aug 18 13:12:31 nodo1 cluster-dlm: set_fs_notified: set_fs_notified no
>>> nodeid 1812048064#012
>>>
>>> Aug 18 13:12:31 nodo1 cluster-dlm: set_fs_notified: set_fs_notified no
>>> nodeid 1812048064#012
>>>
>>> Aug 18 13:12:31 nodo1 cluster-dlm: set_fs_notified: set_fs_notified no
>>> nodeid 1812048064#012
>>>
>>> Attached the log file
>>>
>>> Someone can explain what is the reason?
>>>       
>> Perhaps the membership got out of sync...
>>
>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272495] dlm: connect from non cluster node
>>
>> Maybe lmb or dejan can suggest something... I dont have much to do
>> with ocfs2 anymore.
>>     
> Me neither. But this looks like a kernel bug:
>
>   
>>> Aug 18 13:11:38 nodo1 kernel: [ 4154.272036] kernel BUG at
>>> /usr/src/packages/BUILD/kernel-xen-2.6.31.12/linux-2.6.31/fs/inode.c:1323!
>>>       
> Perhaps ask on the kernel ML?
>
> Thanks,
>
> Dejan
>
>
>   
>> _______________________________________________
>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>
>> Project Home: http://www.clusterlabs.org
>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker
>>     
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker
>   





------------------------------

_______________________________________________
Pacemaker mailing list
Pacemaker at oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker


End of Pacemaker Digest, Vol 33, Issue 65
*****************************************



      
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.clusterlabs.org/pipermail/pacemaker/attachments/20100831/62636e9c/attachment.html>


More information about the Pacemaker mailing list