[Pacemaker] [pacemaker] DRBD + corosync + pacemaker + postgresql

Thomaz Luiz Santos thomaz.santos at gmail.com
Fri Oct 11 13:23:19 EDT 2013


I remake my CRM config and reboot the virtual machines, and not have
successful.
the DRBD not start the primary, and not mount the /dev/drbd0 :-(


ha-slave:

version: 8.3.13 (api:88/proto:86-96)
srcversion: 697DE8B1973B1D8914F04DB
 0: cs:Connected ro:Secondary/Secondary ds:UpToDate/UpToDate C r-----
    ns:0 nr:28672 dw:28672 dr:0 al:0 bm:5 lo:0 pe:0 ua:0 ap:0 ep:1 wo:n
oos:0


ha-master:
version: 8.3.13 (api:88/proto:86-96)
srcversion: 697DE8B1973B1D8914F04DB
 0: cs:Connected ro:Secondary/Secondary ds:UpToDate/UpToDate C r-----
    ns:28672 nr:0 dw:0 dr:28672 al:0 bm:5 lo:0 pe:0 ua:0 ap:0 ep:1 wo:n
oos:0





crm(live)# configure
crm(live)configure# show
node ha-master
node ha-slave
primitive drbd_postgresql ocf:heartbeat:drbd \
        params drbd_resource="postgresql"
primitive fs_postgresql ocf:heartbeat:Filesystem \
        params device="/dev/drbd/by-res/postgresql" directory="/mnt"
fstype="ext4"
primitive postgresqld lsb:postgresql
primitive vip_cluster ocf:heartbeat:IPaddr2 \
        params ip="172.70.65.200" nic="eth0:1"
group postgresql fs_postgresql vip_cluster postgresqld \
        meta target-role="Started"
ms ms_drbd_postgresql drbd_postgresql \
        meta master-max="1" master-node-max="1" clone-max="2"
clone-node-max="1" notify="true"
colocation postgresql_on_drbd inf: postgresql ms_drbd_postgresql:Master
order postgresql_after_drbd inf: ms_drbd_postgresql:promote postgresql:start
property $id="cib-bootstrap-options" \
        dc-version="1.1.6-9971ebba4494012a93c03b40a2c58ec0eb60f50c" \
        cluster-infrastructure="openais" \
        expected-quorum-votes="2" \
        stonith-enabled="false" \
        no-quorum-policy="ignore"
rsc_defaults $id="rsc-options" \
        resource-stickiness="100"



crm(live)# resource
crm(live)resource# list
 Master/Slave Set: ms_drbd_postgresql [drbd_postgresql]
     Stopped: [ drbd_postgresql:0 drbd_postgresql:1 ]
 Resource Group: postgresql
     fs_postgresql      (ocf::heartbeat:Filesystem) Stopped
     vip_cluster        (ocf::heartbeat:IPaddr2) Stopped
     postgresqld        (lsb:postgresql) Stopped




============
Last updated: Fri Oct 11 14:22:50 2013
Last change: Fri Oct 11 14:11:06 2013 via cibadmin on ha-slave
Stack: openais
Current DC: ha-slave - partition with quorum
Version: 1.1.6-9971ebba4494012a93c03b40a2c58ec0eb60f50c
2 Nodes configured, 2 expected votes
5 Resources configured.
============

Online: [ ha-slave ha-master ]


Failed actions:
    drbd_postgresql:0_start_0 (node=ha-slave, call=14, rc=1,
status=complete): unknown error
    drbd_postgresql:0_start_0 (node=ha-master, call=18, rc=1,
status=complete): unknown error




On Fri, Oct 11, 2013 at 1:23 PM, emmanuel segura <emi2fast at gmail.com> wrote:

> try with this constrains
>
> colocation col_postgresql inf: postgresql_cluster ms_drbd_postgresql:Master
> order or_postgresql inf: ms_drbd_postgresql:promote
> postgresql_cluster:start
>
>
>
> 2013/10/11 Thomaz Luiz Santos <thomaz.santos at gmail.com>
>
>> Dear all!
>>
>> I'm trying to make a sample cluster, in virtual machine, and after
>> migrate to a physical machine, however i have problems to configure the
>> pacemaker ( crm ),  to startup the resources and failover.
>>
>> I cant mount the device /dev/drbd0 in the primary node and start
>> postgresql manually, but use in crm resource,  dont can mount the device,
>> and start de postgresql.
>>
>>
>>
>> DRBD Version: 8.3.11 (api:88)
>> Corosync Cluster Engine, version '1.4.2'
>> Pacemaker 1.1.6
>>
>>
>>
>> **** that is my configure on crm ****
>>
>> crm(live)# configure
>> crm(live)configure# show
>> node ha-master
>> node ha-slave
>> primitive drbd_postgresql ocf:heartbeat:drbd \
>>         params drbd_resource="postgresql"
>> primitive fs_postgresql ocf:heartbeat:Filesystem \
>>         params device="/dev/drbd0" directory="/mnt" fstype="ext4" \
>>         meta target-role="Started"
>> primitive postgresql lsb:postgresql \
>>         meta target-role="Started"
>> primitive vip_cluster ocf:heartbeat:IPaddr2 \
>>         params ip="172.70.65.200" nic="eth0:1" \
>>         meta target-role="Started"
>> group postgresql_cluster fs_postgresql vip_cluster postgresql
>> ms ms_drbd_postgresql drbd_postgresql \
>>         meta master-max="1" master-node-max="1" clone-max="2"
>> clone-node-max="1" notify="true" target-role="Stopped"
>> colocation col_postgresql inf: postgresql_cluster ms_drbd_postgresql
>> order or_postgresql inf: ms_drbd_postgresql postgresql_cluster
>> property $id="cib-bootstrap-options" \
>>         dc-version="1.1.6-9971ebba4494012a93c03b40a2c58ec0eb60f50c" \
>>         cluster-infrastructure="openais" \
>>         expected-quorum-votes="2" \
>>         stonith-enabled="false" \
>>         no-quorum-policy="ignore"
>> rsc_defaults $id="rsc-options" \
>>         resource-stickiness="100"
>>
>>
>> **** that is my global_common on drbd  ****
>>
>> global {
>>         usage-count yes;
>>         # minor-count dialog-refresh disable-ip-verification
>> }
>>
>> common {
>>         protocol C;
>>
>>         handlers {
>>                 pri-on-incon-degr
>> "/usr/lib/drbd/notify-pri-on-incon-degr.sh; /usr/lib/drbd/not
>>
>>            ify-emergency-reboot.sh; echo b > /proc/sysrq-trigger ; reboot
>> -f";
>>                 pri-lost-after-sb
>> "/usr/lib/drbd/notify-pri-lost-after-sb.sh; /usr/lib/drbd/not
>>
>>            ify-emergency-reboot.sh; echo b > /proc/sysrq-trigger ; reboot
>> -f";
>>                 local-io-error "/usr/lib/drbd/notify-io-error.sh;
>> /usr/lib/drbd/notify-emergenc
>>                                                        y-shutdown.sh; echo
>> o > /proc/sysrq-trigger ; halt -f";
>>                 fence-peer "/usr/lib/drbd/crm-fence-peer.sh";
>>                 after-resync-target "/usr/lib/drbd/crm-unfence-peer.sh";
>>                 # split-brain "/usr/lib/drbd/notify-split-brain.sh root";
>>                 # out-of-sync "/usr/lib/drbd/notify-out-of-sync.sh root";
>>                 # before-resync-target
>> "/usr/lib/drbd/snapshot-resync-target-lvm.sh -p 15 -- -c
>>
>>         16k";
>>                 # after-resync-target
>> /usr/lib/drbd/unsnapshot-resync-target-lvm.sh;
>>         }
>>
>>         startup {
>>                  # wfc-timeout 15;
>>                  # degr-wfc-timeout 60;
>>                  # outdated-wfc-timeout wait-after-sb
>>         }
>>
>>         disk {
>>                 # on-io-error fencing use-bmbv no-disk-barrier
>> no-disk-flushes
>>                 # no-disk-drain no-md-flushes max-bio-bvecs
>>         }
>>
>>         net {
>>                 # cram-hmac-alg sha1;
>>                 # shared-secret "secret";
>>                 # sndbuf-size rcvbuf-size timeout connect-int ping-int
>> ping-timeout max-buffers
>>                 # max-epoch-size ko-count allow-two-primaries
>> cram-hmac-alg shared-secret
>>                 # after-sb-0pri after-sb-1pri after-sb-2pri
>> data-integrity-alg no-tcp-cork
>>         }
>>
>>         syncer {
>>                 # rate 150M;
>>                 # rate after al-extents use-rle cpu-mask verify-alg
>> csums-alg
>>         }
>> }
>>
>>
>> **** that is my postgresql.res ****
>>
>> resource postgresql {
>>   startup {
>>     wfc-timeout 15;
>>     degr-wfc-timeout 60;
>>   }
>>
>>   syncer {
>>     rate 150M;
>>     verify-alg md5;
>>  }
>>
>>  disk {
>>    on-io-error detach;
>>    no-disk-barrier;
>>    no-disk-flushes;
>>    no-disk-drain;
>>    fencing resource-only;
>>  }
>>
>>  on ha-master {
>>     device /dev/drbd0;
>>     disk /dev/sdb1;
>>     address 172.70.65.210:7788;
>>     meta-disk internal;
>>  }
>>
>>  on ha-slave {
>>     device /dev/drbd0;
>>     disk /dev/sdb1;
>>     address 172.70.65.220:7788;
>>     meta-disk internal;
>>  }
>>
>>
>> }
>>
>>
>> **** that is my corosync.conf ****
>>
>>
>> compatibility: whitetank
>>
>> totem {
>>         version: 2
>>         secauth: off
>>         threads: 0
>>         interface {
>>                 ringnumber: 0
>>                 bindnetaddr: 172.70.65.200
>>                 mcastaddr: 226.94.1.1
>>                 mcastport: 5405
>>                 ttl: 1
>>         }
>> }
>>
>> logging {
>>         fileline: off
>>         to_stderr: yes
>>         to_logfile: yes
>>         to_syslog: yes
>>         logfile: /var/log/cluster/corosync.log
>>         debug: on
>>         timestamp: on
>>         logger_subsys {
>>                 subsys: AMF
>>                 debug: off
>>         }
>> }
>>
>> amf {
>>         mode: disabled
>> }
>>
>> aisexec{
>>   user : root
>>   group : root
>> }
>>
>> service{
>>   # Load the Pacemaker Cluster Resource Manager
>>   name : pacemaker
>>   ver : 0
>> }
>>
>>
>>
>> DRBD, postgresql, manually start :
>>
>>
>> version: 8.3.13 (api:88/proto:86-96)
>> srcversion: 697DE8B1973B1D8914F04DB
>>  0: cs:Connected ro:Primary/Secondary ds:UpToDate/UpToDate C r-----
>>     ns:0 nr:0 dw:0 dr:664 al:0 bm:0 lo:0 pe:0 ua:0 ap:0 ep:1 wo:n oos:0
>>
>>
>> version: 8.3.13 (api:88/proto:86-96)
>> srcversion: 697DE8B1973B1D8914F04DB
>>  0: cs:Connected ro:Secondary/Primary ds:UpToDate/UpToDate C r-----
>>     ns:0 nr:0 dw:0 dr:0 al:0 bm:0 lo:0 pe:0 ua:0 ap:0 ep:1 wo:n oos:0
>>
>>
>>
>> root at ha-master:/mnt# df -hT
>> Sist. Arq.     Tipo      Tam. Usado Disp. Uso% Montado em
>> /dev/sda1      ext4      4,0G  1,8G  2,1G  47% /
>> udev           devtmpfs  473M  4,0K  473M   1% /dev
>> tmpfs          tmpfs     193M  264K  193M   1% /run
>> none           tmpfs     5,0M  4,0K  5,0M   1% /run/lock
>> none           tmpfs     482M   17M  466M   4% /run/shm
>> /dev/drbd0     ext4      2,0G   69M  1,9G   4% /mnt
>>
>>
>> root at ha-master:/mnt# service postgresql status
>> Running clusters: 9.1/main
>>
>>
>>
>>
>> --
>> ------------------------------
>> Thomaz Luiz Santos
>> Linux User: #359356
>>
>>
>> _______________________________________________
>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>
>> Project Home: http://www.clusterlabs.org
>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>> Bugs: http://bugs.clusterlabs.org
>>
>>
>
>
> --
> esta es mi vida e me la vivo hasta que dios quiera
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>
>


-- 
------------------------------
Thomaz Luiz Santos
Linux User: #359356
http://thomaz.santos.googlepages.com/
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20131011/98799843/attachment-0003.html>


More information about the Pacemaker mailing list