[Pacemaker] [pacemaker] DRBD + corosync + pacemaker + postgresql

emmanuel segura emi2fast at gmail.com
Tue Oct 15 22:12:20 UTC 2013


Check if your postgres is stopped or not start in the boot time.


2013/10/15 Thomaz Luiz Santos <thomaz.santos at gmail.com>

> dear all :-D
>
> I remake my crm config
>
> node ha-master
> node ha-slave
> primitive drbd_postgresql ocf:linbit:drbd \
>         params drbd_resource="postgresql" \
>         op monitor interval="30" role="Master" \
>         op monitor interval="33" role="Slave"
> primitive fs_postgresql ocf:heartbeat:Filesystem \
>         params device="/dev/drbd0" directory="/mnt" fstype="ext4" \
>         meta target-role="Started"
> primitive postgresqld lsb:postgresql \
>         meta target-role="Started"
> primitive vip_cluster ocf:heartbeat:IPaddr2 \
>         params ip="172.70.65.200" nic="eth0:1" \
>         meta target-role="Started"
> group postgresql fs_postgresql vip_cluster postgresqld \
>         meta is-managed="true"
> ms ms_drbd_postgresql drbd_postgresql \
>         meta master-max="1" master-node-max="1" clone-max="2"
> clone-node-max="1" notify="true" target-role="Started" is-managed="true"
> location cli-prefer-postgresql postgresql \
>         rule $id="cli-prefer-rule-postgresql" inf: #uname eq ha-master
> colocation postgresql_on_drbd inf: postgresql ms_drbd_postgresql:Master
> order postgresql_after_drbd inf: ms_drbd_postgresql:promote
> postgresql:start
>
>
> and fix some problems  but have one error new rsrsrs.
>
> how fix it, its is the main problem ??
>
> on ha-master syslog:
> pengine: [1100]: ERROR: native_create_actions: Resource postgresqld
> (lsb::postgresql) is active on 2 nodes attempting recovery
>
>
>
> On Fri, Oct 11, 2013 at 2:23 PM, Thomaz Luiz Santos <
> thomaz.santos at gmail.com> wrote:
>
>> I remake my CRM config and reboot the virtual machines, and not have
>> successful.
>> the DRBD not start the primary, and not mount the /dev/drbd0 :-(
>>
>>
>> ha-slave:
>>
>> version: 8.3.13 (api:88/proto:86-96)
>> srcversion: 697DE8B1973B1D8914F04DB
>>  0: cs:Connected ro:Secondary/Secondary ds:UpToDate/UpToDate C r-----
>>     ns:0 nr:28672 dw:28672 dr:0 al:0 bm:5 lo:0 pe:0 ua:0 ap:0 ep:1 wo:n
>> oos:0
>>
>>
>> ha-master:
>> version: 8.3.13 (api:88/proto:86-96)
>> srcversion: 697DE8B1973B1D8914F04DB
>>  0: cs:Connected ro:Secondary/Secondary ds:UpToDate/UpToDate C r-----
>>     ns:28672 nr:0 dw:0 dr:28672 al:0 bm:5 lo:0 pe:0 ua:0 ap:0 ep:1 wo:n
>> oos:0
>>
>>
>>
>>
>>
>> crm(live)# configure
>> crm(live)configure# show
>> node ha-master
>> node ha-slave
>> primitive drbd_postgresql ocf:heartbeat:drbd \
>>         params drbd_resource="postgresql"
>> primitive fs_postgresql ocf:heartbeat:Filesystem \
>>         params device="/dev/drbd/by-res/postgresql" directory="/mnt"
>> fstype="ext4"
>> primitive postgresqld lsb:postgresql
>> primitive vip_cluster ocf:heartbeat:IPaddr2 \
>>         params ip="172.70.65.200" nic="eth0:1"
>> group postgresql fs_postgresql vip_cluster postgresqld \
>>         meta target-role="Started"
>> ms ms_drbd_postgresql drbd_postgresql \
>>         meta master-max="1" master-node-max="1" clone-max="2"
>> clone-node-max="1" notify="true"
>> colocation postgresql_on_drbd inf: postgresql ms_drbd_postgresql:Master
>> order postgresql_after_drbd inf: ms_drbd_postgresql:promote
>> postgresql:start
>> property $id="cib-bootstrap-options" \
>>         dc-version="1.1.6-9971ebba4494012a93c03b40a2c58ec0eb60f50c" \
>>         cluster-infrastructure="openais" \
>>         expected-quorum-votes="2" \
>>         stonith-enabled="false" \
>>         no-quorum-policy="ignore"
>> rsc_defaults $id="rsc-options" \
>>         resource-stickiness="100"
>>
>>
>>
>> crm(live)# resource
>> crm(live)resource# list
>>  Master/Slave Set: ms_drbd_postgresql [drbd_postgresql]
>>      Stopped: [ drbd_postgresql:0 drbd_postgresql:1 ]
>>  Resource Group: postgresql
>>      fs_postgresql      (ocf::heartbeat:Filesystem) Stopped
>>      vip_cluster        (ocf::heartbeat:IPaddr2) Stopped
>>      postgresqld        (lsb:postgresql) Stopped
>>
>>
>>
>>
>> ============
>>  Last updated: Fri Oct 11 14:22:50 2013
>> Last change: Fri Oct 11 14:11:06 2013 via cibadmin on ha-slave
>> Stack: openais
>> Current DC: ha-slave - partition with quorum
>> Version: 1.1.6-9971ebba4494012a93c03b40a2c58ec0eb60f50c
>> 2 Nodes configured, 2 expected votes
>> 5 Resources configured.
>> ============
>>
>> Online: [ ha-slave ha-master ]
>>
>>
>> Failed actions:
>>     drbd_postgresql:0_start_0 (node=ha-slave, call=14, rc=1,
>> status=complete): unknown error
>>     drbd_postgresql:0_start_0 (node=ha-master, call=18, rc=1,
>> status=complete): unknown error
>>
>>
>>
>>
>> On Fri, Oct 11, 2013 at 1:23 PM, emmanuel segura <emi2fast at gmail.com>wrote:
>>
>>> try with this constrains
>>>
>>> colocation col_postgresql inf: postgresql_cluster
>>> ms_drbd_postgresql:Master
>>> order or_postgresql inf: ms_drbd_postgresql:promote
>>> postgresql_cluster:start
>>>
>>>
>>>
>>> 2013/10/11 Thomaz Luiz Santos <thomaz.santos at gmail.com>
>>>
>>>>  Dear all!
>>>>
>>>> I'm trying to make a sample cluster, in virtual machine, and after
>>>> migrate to a physical machine, however i have problems to configure the
>>>> pacemaker ( crm ),  to startup the resources and failover.
>>>>
>>>> I cant mount the device /dev/drbd0 in the primary node and start
>>>> postgresql manually, but use in crm resource,  dont can mount the device,
>>>> and start de postgresql.
>>>>
>>>>
>>>>
>>>> DRBD Version: 8.3.11 (api:88)
>>>> Corosync Cluster Engine, version '1.4.2'
>>>> Pacemaker 1.1.6
>>>>
>>>>
>>>>
>>>> **** that is my configure on crm ****
>>>>
>>>> crm(live)# configure
>>>> crm(live)configure# show
>>>> node ha-master
>>>> node ha-slave
>>>> primitive drbd_postgresql ocf:heartbeat:drbd \
>>>>         params drbd_resource="postgresql"
>>>> primitive fs_postgresql ocf:heartbeat:Filesystem \
>>>>         params device="/dev/drbd0" directory="/mnt" fstype="ext4" \
>>>>         meta target-role="Started"
>>>> primitive postgresql lsb:postgresql \
>>>>         meta target-role="Started"
>>>> primitive vip_cluster ocf:heartbeat:IPaddr2 \
>>>>         params ip="172.70.65.200" nic="eth0:1" \
>>>>         meta target-role="Started"
>>>> group postgresql_cluster fs_postgresql vip_cluster postgresql
>>>> ms ms_drbd_postgresql drbd_postgresql \
>>>>         meta master-max="1" master-node-max="1" clone-max="2"
>>>> clone-node-max="1" notify="true" target-role="Stopped"
>>>> colocation col_postgresql inf: postgresql_cluster ms_drbd_postgresql
>>>> order or_postgresql inf: ms_drbd_postgresql postgresql_cluster
>>>> property $id="cib-bootstrap-options" \
>>>>         dc-version="1.1.6-9971ebba4494012a93c03b40a2c58ec0eb60f50c" \
>>>>         cluster-infrastructure="openais" \
>>>>         expected-quorum-votes="2" \
>>>>         stonith-enabled="false" \
>>>>         no-quorum-policy="ignore"
>>>> rsc_defaults $id="rsc-options" \
>>>>         resource-stickiness="100"
>>>>
>>>>
>>>> **** that is my global_common on drbd  ****
>>>>
>>>> global {
>>>>         usage-count yes;
>>>>         # minor-count dialog-refresh disable-ip-verification
>>>> }
>>>>
>>>> common {
>>>>         protocol C;
>>>>
>>>>         handlers {
>>>>                 pri-on-incon-degr
>>>> "/usr/lib/drbd/notify-pri-on-incon-degr.sh; /usr/lib/drbd/not
>>>>
>>>>            ify-emergency-reboot.sh; echo b > /proc/sysrq-trigger ; reboot
>>>> -f";
>>>>                 pri-lost-after-sb
>>>> "/usr/lib/drbd/notify-pri-lost-after-sb.sh; /usr/lib/drbd/not
>>>>
>>>>            ify-emergency-reboot.sh; echo b > /proc/sysrq-trigger ; reboot
>>>> -f";
>>>>                 local-io-error "/usr/lib/drbd/notify-io-error.sh;
>>>> /usr/lib/drbd/notify-emergenc
>>>>                                                        y-shutdown.sh; echo
>>>> o > /proc/sysrq-trigger ; halt -f";
>>>>                 fence-peer "/usr/lib/drbd/crm-fence-peer.sh";
>>>>                 after-resync-target "/usr/lib/drbd/crm-unfence-peer.sh";
>>>>                 # split-brain "/usr/lib/drbd/notify-split-brain.sh
>>>> root";
>>>>                 # out-of-sync "/usr/lib/drbd/notify-out-of-sync.sh
>>>> root";
>>>>                 # before-resync-target
>>>> "/usr/lib/drbd/snapshot-resync-target-lvm.sh -p 15 -- -c
>>>>
>>>>         16k";
>>>>                 # after-resync-target
>>>> /usr/lib/drbd/unsnapshot-resync-target-lvm.sh;
>>>>         }
>>>>
>>>>         startup {
>>>>                  # wfc-timeout 15;
>>>>                  # degr-wfc-timeout 60;
>>>>                  # outdated-wfc-timeout wait-after-sb
>>>>         }
>>>>
>>>>         disk {
>>>>                 # on-io-error fencing use-bmbv no-disk-barrier
>>>> no-disk-flushes
>>>>                 # no-disk-drain no-md-flushes max-bio-bvecs
>>>>         }
>>>>
>>>>         net {
>>>>                 # cram-hmac-alg sha1;
>>>>                 # shared-secret "secret";
>>>>                 # sndbuf-size rcvbuf-size timeout connect-int ping-int
>>>> ping-timeout max-buffers
>>>>                 # max-epoch-size ko-count allow-two-primaries
>>>> cram-hmac-alg shared-secret
>>>>                 # after-sb-0pri after-sb-1pri after-sb-2pri
>>>> data-integrity-alg no-tcp-cork
>>>>         }
>>>>
>>>>         syncer {
>>>>                 # rate 150M;
>>>>                 # rate after al-extents use-rle cpu-mask verify-alg
>>>> csums-alg
>>>>         }
>>>> }
>>>>
>>>>
>>>> **** that is my postgresql.res ****
>>>>
>>>> resource postgresql {
>>>>   startup {
>>>>     wfc-timeout 15;
>>>>     degr-wfc-timeout 60;
>>>>   }
>>>>
>>>>   syncer {
>>>>     rate 150M;
>>>>     verify-alg md5;
>>>>  }
>>>>
>>>>  disk {
>>>>    on-io-error detach;
>>>>    no-disk-barrier;
>>>>    no-disk-flushes;
>>>>    no-disk-drain;
>>>>    fencing resource-only;
>>>>  }
>>>>
>>>>  on ha-master {
>>>>     device /dev/drbd0;
>>>>     disk /dev/sdb1;
>>>>     address 172.70.65.210:7788;
>>>>     meta-disk internal;
>>>>  }
>>>>
>>>>  on ha-slave {
>>>>     device /dev/drbd0;
>>>>     disk /dev/sdb1;
>>>>     address 172.70.65.220:7788;
>>>>     meta-disk internal;
>>>>  }
>>>>
>>>>
>>>> }
>>>>
>>>>
>>>> **** that is my corosync.conf ****
>>>>
>>>>
>>>> compatibility: whitetank
>>>>
>>>> totem {
>>>>         version: 2
>>>>         secauth: off
>>>>         threads: 0
>>>>         interface {
>>>>                 ringnumber: 0
>>>>                 bindnetaddr: 172.70.65.200
>>>>                 mcastaddr: 226.94.1.1
>>>>                 mcastport: 5405
>>>>                 ttl: 1
>>>>         }
>>>> }
>>>>
>>>> logging {
>>>>         fileline: off
>>>>         to_stderr: yes
>>>>         to_logfile: yes
>>>>         to_syslog: yes
>>>>         logfile: /var/log/cluster/corosync.log
>>>>         debug: on
>>>>         timestamp: on
>>>>         logger_subsys {
>>>>                 subsys: AMF
>>>>                 debug: off
>>>>         }
>>>> }
>>>>
>>>> amf {
>>>>         mode: disabled
>>>> }
>>>>
>>>> aisexec{
>>>>   user : root
>>>>   group : root
>>>> }
>>>>
>>>> service{
>>>>   # Load the Pacemaker Cluster Resource Manager
>>>>   name : pacemaker
>>>>   ver : 0
>>>> }
>>>>
>>>>
>>>>
>>>> DRBD, postgresql, manually start :
>>>>
>>>>
>>>> version: 8.3.13 (api:88/proto:86-96)
>>>> srcversion: 697DE8B1973B1D8914F04DB
>>>>  0: cs:Connected ro:Primary/Secondary ds:UpToDate/UpToDate C r-----
>>>>     ns:0 nr:0 dw:0 dr:664 al:0 bm:0 lo:0 pe:0 ua:0 ap:0 ep:1 wo:n oos:0
>>>>
>>>>
>>>> version: 8.3.13 (api:88/proto:86-96)
>>>> srcversion: 697DE8B1973B1D8914F04DB
>>>>  0: cs:Connected ro:Secondary/Primary ds:UpToDate/UpToDate C r-----
>>>>     ns:0 nr:0 dw:0 dr:0 al:0 bm:0 lo:0 pe:0 ua:0 ap:0 ep:1 wo:n oos:0
>>>>
>>>>
>>>>
>>>> root at ha-master:/mnt# df -hT
>>>> Sist. Arq.     Tipo      Tam. Usado Disp. Uso% Montado em
>>>> /dev/sda1      ext4      4,0G  1,8G  2,1G  47% /
>>>> udev           devtmpfs  473M  4,0K  473M   1% /dev
>>>> tmpfs          tmpfs     193M  264K  193M   1% /run
>>>> none           tmpfs     5,0M  4,0K  5,0M   1% /run/lock
>>>> none           tmpfs     482M   17M  466M   4% /run/shm
>>>> /dev/drbd0     ext4      2,0G   69M  1,9G   4% /mnt
>>>>
>>>>
>>>> root at ha-master:/mnt# service postgresql status
>>>> Running clusters: 9.1/main
>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> ------------------------------
>>>> Thomaz Luiz Santos
>>>> Linux User: #359356
>>>>
>>>>
>>>> _______________________________________________
>>>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>>
>>>> Project Home: http://www.clusterlabs.org
>>>> Getting started:
>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>>> Bugs: http://bugs.clusterlabs.org
>>>>
>>>>
>>>
>>>
>>> --
>>> esta es mi vida e me la vivo hasta que dios quiera
>>>
>>> _______________________________________________
>>> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>>>
>>> Project Home: http://www.clusterlabs.org
>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
>>> Bugs: http://bugs.clusterlabs.org
>>>
>>>
>>
>>
>> --
>> ------------------------------
>> Thomaz Luiz Santos
>> Linux User: #359356
>> http://thomaz.santos.googlepages.com/
>>
>
>
>
> --
> ------------------------------
> Thomaz Luiz Santos
> Linux User: #359356
> http://thomaz.santos.googlepages.com/
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker at oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>
>


-- 
esta es mi vida e me la vivo hasta que dios quiera
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.clusterlabs.org/pipermail/pacemaker/attachments/20131016/6d25079b/attachment.htm>


More information about the Pacemaker mailing list