Hello everyone,<br><br>I tried to configure postgres RA and I ran into some problems.<br><br>I configured several resources in my cluster config where pgsql was set to run last, after DRBD, Filesystem, IPAddr2 and nginx.<br>
<br>Here is how it looks like in crm configure:<br><br>crm(live)configure# show<br>node webnode01 \<br> attributes standby="off"<br>node webnode02 \<br> attributes standby="off"<br>primitive ClusterIP ocf:heartbeat:IPaddr2 \<br>
params ip="192.168.10.80" cidr_netmask="32" \<br> op monitor interval="30s"<br>primitive drbd_res ocf:linbit:drbd \<br> params drbd_resource="yorxs" \<br> op monitor interval="60s" \<br>
op start interval="0s" timeout="240s" \<br> op stop interval="0s" timeout="100s"<br>primitive fs_res ocf:heartbeat:Filesystem \<br> params device="/dev/drbd1" directory="/srv" fstype="ext4" \<br>
op start interval="0s" timeout="60s" \<br> op stop interval="0s" timeout="60s" \<br> op monitor interval="60s" timeout="40s"<br>primitive nginx_res ocf:heartbeat:nginx \<br>
params configfile="/etc/nginx/nginx.conf" httpd="/usr/local/sbin/nginx" status10url="http:/<a href="http://127.0.0.1">127.0.0.1</a>" \<br> op monitor interval="10s" timeout="30s" \<br>
op start interval="0" timeout="40s" \<br> op stop interval="0" timeout="60s"<br>primitive postgres_res ocf:heartbeat:pgsql \<br> params psql="/bin/psql" pgdata="/var/lib/postgres/8.4/main" logfile="/var/log/postgres/postgres.log" \<br>
op start interval="0" timeout="120s" \<br> op stop interval="0" timeout="120s" \<br> op monitor interval="30s" timeout="30s"<br>group cluster_1 fs_res ClusterIP nginx_res postgres_res<br>
ms drbd_cluster drbd_res \<br> meta master-max="1" master-node-max="1" clone-max="2" clone-node-max="1" notify="true"<br>location prefer_webnode01 cluster_1 50: webnode01<br>
location prefer_webnode01_drbd drbd_cluster 50: webnode01<br>colocation cluster_1_on_drbd inf: cluster_1 drbd_cluster:Master<br>order cluster_1_after_drbd inf: drbd_cluster:promote cluster_1:start<br>property $id="cib-bootstrap-options" \<br>
dc-version="1.0.9-74392a28b7f31d7ddc86689598bd23114f58978b" \<br> cluster-infrastructure="openais" \<br> expected-quorum-votes="2" \<br> stonith-enabled="false" \<br>
no-quorum-policy="ignore" \<br> last-lrm-refresh="1318326771"<br><br>However, when I run this config, everything except for pgsql starts without problems. For pgsql, I got the following error:<br>
<br>in crm_mon<br>Online: [ webnode02 webnode01 ]<br><br> Master/Slave Set: drbd_cluster<br> Masters: [ webnode01 ]<br> Slaves: [ webnode02 ]<br> Resource Group: cluster_1<br> fs_res (ocf::heartbeat:Filesystem): Started webnode01<br>
ClusterIP (ocf::heartbeat:IPaddr2): Started webnode01<br> nginx_res (ocf::heartbeat:nginx): Started webnode01<br> postgres_res (ocf::heartbeat:pgsql): Stopped<br><br>Failed actions:<br> postgres_res_start_0 (node=webnode01, call=84, rc=5, status=complete): not installed<br>
postgres_res_start_0 (node=webnode02, call=66, rc=5, status=complete): not installed<br><br>in /var/log/syslog<br>webnode01 log # cat syslog |grep postgres_res<br>Oct 11 11:39:34 webnode01 crmd: [921]: info: do_lrm_rsc_op: Performing key=6:93:7:933bf2ab-00d0-435c-a24f-85897e0c9725 op=postgres_res_monitor_0 )<br>
Oct 11 11:39:34 webnode01 lrmd: [914]: info: rsc:postgres_res:27: probe<br>Oct 11 11:39:34 webnode01 crmd: [921]: info: process_lrm_event: LRM operation postgres_res_monitor_0 (call=27, rc=7, cib-update=36, confirmed=true) not running<br>
Oct 11 11:39:50 webnode01 crmd: [921]: info: do_lrm_rsc_op: Performing key=39:96:0:933bf2ab-00d0-435c-a24f-85897e0c9725 op=postgres_res_start_0 )<br>Oct 11 11:39:50 webnode01 lrmd: [914]: info: rsc:postgres_res:39: start<br>
Oct 11 11:39:50 webnode01 crmd: [921]: info: process_lrm_event: LRM operation postgres_res_start_0 (call=39, rc=5, cib-update=47, confirmed=true) not installed<br>Oct 11 11:39:50 webnode01 attrd: [918]: info: find_hash_entry: Creating hash entry for fail-count-postgres_res<br>
Oct 11 11:39:50 webnode01 attrd: [918]: info: attrd_trigger_update: Sending flush op to all hosts for: fail-count-postgres_res (INFINITY)<br>Oct 11 11:39:50 webnode01 attrd: [918]: info: attrd_perform_update: Sent update 63: fail-count-postgres_res=INFINITY<br>
Oct 11 11:39:50 webnode01 attrd: [918]: info: find_hash_entry: Creating hash entry for last-failure-postgres_res<br>Oct 11 11:39:50 webnode01 attrd: [918]: info: attrd_trigger_update: Sending flush op to all hosts for: last-failure-postgres_res (1318325990)<br>
Oct 11 11:39:50 webnode01 attrd: [918]: info: attrd_perform_update: Sent update 66: last-failure-postgres_res=1318325990<br>Oct 11 11:39:50 webnode01 crmd: [921]: info: do_lrm_rsc_op: Performing key=4:97:0:933bf2ab-00d0-435c-a24f-85897e0c9725 op=postgres_res_stop_0 )<br>
Oct 11 11:39:50 webnode01 lrmd: [914]: info: rsc:postgres_res:40: stop<br>Oct 11 11:39:50 webnode01 crmd: [921]: info: process_lrm_event: LRM operation postgres_res_stop_0 (call=40, rc=0, cib-update=49, confirmed=true) ok<br>
<br>Additional info:<br><br>/etc/postgresql, /etc/postgresql-common and /var/lib/postgresql are symlinks on both nodes. Actual directories are on shared DRBD disk.<br>Postgres starts without any problems with init script. On both nodes.<br>
<br>Thanks a lot in advance for any advice.<br><br>-- <br>Amar Prasovic<br>Gaißacher Straße 17<br>D - 81371 München<br>