[pgpool-general: 397] Re: Pacemaker CoroSync + PGPool-II

Videanu Adrian videanuadrian at yahoo.com
Mon Apr 30 14:19:28 JST 2012


Hi list, 
i have manage to get rid of the errors that Steve also encountered here by manually provide to pgpool ocf the paths to my correct system path. Also i have used the original ocf, generated by make, make install of pgpool-ha. So, instead of : 

if [ "x$OCF_RESKEY_pgpoolconf" = 'x' ]; then
        XX=$($PGPOOLCMD --help 2>&1 | grep ' config_file default path: ')
        PGP_CONF=${XX#  config_file default path: }
else
        PGP_CONF=$OCF_RESKEY_pgpoolconf
i have : 
if [ "x$OCF_RESKEY_pgpoolconf" = 'x' ]; then
        PGP_CONF=/usr/local/etc/pgpool.conf
else
        PGP_CONF=$OCF_RESKEY_pgpoolconf

and i did like this all my config directives (i think that this could be overriden if you provide these params when importing ocf script to corosync).
Also do the fact that my pgpool scripts were run as root i have to modify the start and stop commands from : 
su -c "$PGPOOL $PGPOOL_START_ARG" $PGPOOLUSER
to: 
$PGPOOL $PGPOOL_START_ARG

and 
RET=`su -c "$PGPOOL $PGPOOL_FORCE_STOP_ARG | grep ERROR " $PGPOOLUSER`
to
RET=`$PGPOOL $PGPOOL_FORCE_STOP_ARG | grep ERROR`

After all these modifications pgpool seems to work as expected.
Hope this will help someone

--- On Tue, 4/24/12, Steven Bambling <smbambling at arin.net> wrote:

From: Steven Bambling <smbambling at arin.net>
Subject: [pgpool-general: 379] Pacemaker CoroSync + PGPool-II
To: "pgpool-general at pgpool.net" <pgpool-general at pgpool.net>
Date: Tuesday, April 24, 2012, 3:45 PM



 

After doing some searching on setting up "PGPool-HA" to limit pgpool being a single point of failure it looks like development on the heartbeat project has reduced greatly and development has shifted to corosync  (backed by RedHat and Suse) that is recommend
 by pacemaker. 



I've found an article here http://masteinhauser.github.com/blog/2011/09/24/pacemaker-pgpool2/ that explains using pacemaker with pgpool-II.  In the post a resource agent is
 provided.  There was a quick tweak I had to make with the PGPool-II path created by the installed RPMs obtained from http://yum.postgresql.org/9.1/redhat/rhel-$releasever-$basearch.
  I modified the below marked in bold from /var/run/pgpool/ to /var/run/pgpool-II-91




pgpool2_status() {
    if [ ! -r "/var/run/pgpool-II-91/pgpool.pid" ]; then
        return $OCF_NOT_RUNNING
    fi
    ps_info=$(ps ax | grep "pgpool" | grep $(cat /var/run/pgpool-II-91/pgpool.pid))




I used the following parameters to created the resource





crm configure primitive pgPool ocf:heartbeat:pgpool2 
\

params pcp_admin_username=postgres \

params pcp_admin_password=password \

params pcp_admin_port=9898
\

params pcp_admin_host=localhost 
\

params pgpool_bin=/usr/bin/pgpool 
\

params pcp_attach_node_bin=/usr/bin/pcp_attach_node
\

params pcp_detach_node_bin=/usr/bin/pcp_detach_node
\

params pcp_node_count_bin=/usr/bin/pcp_node_count
\

params pcp_node_info_bin=/usr/bin/pcp_node_info
\

params stop_mode=f 
\

params auto_reconnect=t
\



params fail_on_detached=true \


op monitor 
interval=1min





The resource is looks to be created correctly but when I (re)start the corosync service and look at crm_mon I see some failed actions




============
Last updated: Tue Apr 24 08:31:08 2012
Last change: Tue Apr 24 08:02:31 2012 via cibadmin on 
pg1.stage.arin.net
Stack: openais
Current DC: pg2.stage.arin.net - partition with quorum
Version: 1.1.6-3.el6-a02c0f19a00c1eb2527ad38f146ebc0834814558
2 Nodes configured, 2 expected votes
2 Resources configured.
============



Online: [ pg1.stage.net 
pg2.stage.net ]



ClusterIP (ocf::heartbeat:IPaddr2):
Started pg2.stage.net



Failed actions:
    pgPool_monitor_0 (node=pg1.stage.net, call=3, rc=2, status=complete): invalid parameter
    pgPool_monitor_0 (node=pg2.stage.net, call=3, rc=2, status=complete): invalid parameter




When I look in the /var/log/cluster/corosync.log I see this error Apr 24 08:23:48
pg1.stage.net lrmd: [28471]: WARN: Managed pgPool:monitor process 28484 exited with return code 2



Has anyone ran into a similar experience or have suggestions for a cluster solution with pgpool-II.



I know this is not 100% related to PGPool so I've also posted it to the pacemaker mailing list.  



v/r



STEVE








-----Inline Attachment Follows-----

_______________________________________________
pgpool-general mailing list
pgpool-general at pgpool.net
http://www.pgpool.net/mailman/listinfo/pgpool-general
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.sraoss.jp/pipermail/pgpool-general/attachments/20120429/6c6509db/attachment.html>


More information about the pgpool-general mailing list