[pgpool-hackers: 2790] Re: [pgpool-II 0000394]: When multiple failover happens the pgpool pool node status is going into standby -stanby state

Tatsuo Ishii ishii at sraoss.co.jp
Tue Apr 24 18:05:36 JST 2018


Usama,

This seems to be a watchdog issue. Can you please take of this?

Best regards,
--
Tatsuo Ishii
SRA OSS, Inc. Japan
English: http://www.sraoss.co.jp/index_en.php
Japanese:http://www.sraoss.co.jp

From: Pgpool-II Bug Tracker <bugtracker at pgpool.net>
Subject: [pgpool-II 0000394]: When multiple failover happens the pgpool pool node status is going into standby -stanby state
Date: Sat, 21 Apr 2018 20:18:46 +0900
Message-ID: <48b3ffd9f75c3aab03024ca9b2ae3d8e at www.pgpool.net>

> The following issue has been SUBMITTED. 
> ====================================================================== 
> https://www.pgpool.net/mantisbt/view.php?id=394 
> ====================================================================== 
> Reported By:                kavin
> Assigned To:                
> ====================================================================== 
> Project:                    pgpool-II
> Issue ID:                   394
> Category:                   Bug
> Reproducibility:            always
> Severity:                   major
> Priority:                   urgent
> Status:                     new
> ====================================================================== 
> Date Submitted:             2018-04-21 20:18 JST
> Last Modified:              2018-04-21 20:18 JST
> ====================================================================== 
> Summary:                    When multiple failover happens the pgpool pool node
> status is going into standby -stanby state
> Description: 
> I have 2 cluster nodes installed with pgpool-II(pgpool-II-10-3.7.2-1) and
> postgresql(postgresql10-10.2). The two nodes are configured as streaming
> replication mode. Lets name the two clusetr node as vm1(primary) and
> vm2(secondary ).
> 
> Use case:
> 
> 1. Failover from primary(vm1) to secondary(vm2)
> 2. Failover from secondary(vm2) to primary(vm1)
> 
> Thsi scenario leads the pgpool node status to standby-standby state
> 
> Steps to Reproduce: 
> 1. Started postgers DB on primary node (vm1)
> 2. Started the pgpool on primary node (vm1)
> 3. Once the delegate IP is up on the primary node (vm1), started the pgpool on
> the secondary node (vm2).
> 4. pgpool node status is follows
> 
>  node_id |   hostname    | port | status | lb_weight |  role   | select_cnt |
> load_balance_node | replication_delay
> ---------+---------------+------+--------+-----------+---------+------------+-------------------+-------------------
>  0       | vm1 | 5432 | up     | 0.500000  | primary | 0          | true        
>      | 0
>  1       | vm2 | 5432 | down   | 0.500000  | standby | 0          | false       
>      | 0
> (2 rows)
> 
> The secondary node DB is still down.
> 5. Started the pcp recovery for the secondary niode. Output is as follows
> 
> node_id |   hostname    | port | status | lb_weight |  role   | select_cnt |
> load_balance_node | replication_delay
> ---------+---------------+------+--------+-----------+---------+------------+-------------------+-------------------
>  0       | vm1 | 5432 | up     | 0.500000  | primary | 0          | true        
>      | 0
>  1       | vm2 | 5432 | up     | 0.500000  | standby | 0          | false       
>      | 0
> (2 rows)
> 
> The secondary node (vm2) db is stared successfully.
> 6. I want to do the failover from vm1 to vm2. SO i brought down the postgres DB
> on the primary(vm1).Output follows
> 
> node_id |   hostname    | port | status | lb_weight |  role   | select_cnt |
> load_balance_node | replication_delay
> ---------+---------------+------+--------+-----------+---------+------------+-------------------+-------------------
>  0       | vm1 | 5432 | down   | 0.500000  | standby | 0          | false       
>      | 0
>  1       | vm2 | 5432 | up     | 0.500000  | primary | 0          | true        
>      | 0
> (2 rows)
> The secondary(vm2) node became the primary after the failover. 
> 7. Then started the pcp recovery for primary node(vm1). Output follows
> 
> node_id |   hostname    | port | status | lb_weight |  role   | select_cnt |
> load_balance_node | replication_delay
> ---------+---------------+------+--------+-----------+---------+------------+-------------------+-------------------
>  0       | vm1 | 5432 | up     | 0.500000  | standby | 0          | false       
>      | 0
>  1       | vm2| 5432 | up     | 0.500000  | primary | 0          | true         
>     | 0
> (2 rows)
> 
> Prmary(vm1) nod e db is up now.
> 
> 8. Again i want to do the failover from vm2 to vm1. SO i brought down the DB
> from Vm2. . Output as follows
> 
>  node_id |   hostname    | port | status | lb_weight |  role   | select_cnt |
> load_balance_node | replication_delay
> ---------+---------------+------+--------+-----------+---------+------------+-------------------+-------------------
>  0       | vm1 | 5432 | up     | 0.500000  | standby | 0          | true        
>      | 0
>  1       | vm2 | 5432 | down   | 0.500000  | standby | 0          | false       
>      | 0
> (2 rows)
> 
> DOing this leads the pgpool node into standby-standby state.
> 
> 
> ******* Logs and pgpool configirations are posted in additional information
> field***********
> 
> 
> 
> 
> Additional Information: 
> *************************************************************************************
> failover from primary(vm1) to secondary(vm2)
> ****************************************************************************************
> Apr 21 10:16:44 vm1 journal: ealth check process(0)[23041]: [19-1] 2018-04-21
> 10:16:44: pid 23041: ERROR:  failed to make persistent db connection
> Apr 21 10:16:44 vm1 journal: ealth check process(0)[23041]: [19-2] 2018-04-21
> 10:16:44: pid 23041: DETAIL:  connection to host:"vm1:5432" failed
> Apr 21 10:16:44 vm1 journal: ealth check process(0)[23041]: [20-1] 2018-04-21
> 10:16:44: pid 23041: LOG:  health check retrying on DB node: 0 (round:3)
> Apr 21 10:16:49 vm1 journal: orker process[23745]: [1050-1] 2018-04-21 10:16:49:
> pid 23745: ERROR:  Failed to check replication time lag
> Apr 21 10:16:49 vm1 journal: orker process[23745]: [1050-2] 2018-04-21 10:16:49:
> pid 23745: DETAIL:  No persistent db connection for the node 0
> Apr 21 10:16:49 vm1 journal: orker process[23745]: [1050-3] 2018-04-21 10:16:49:
> pid 23745: HINT:  check sr_check_user and sr_check_password
> Apr 21 10:16:49 vm1 journal: orker process[23745]: [1050-4] 2018-04-21 10:16:49:
> pid 23745: CONTEXT:  while checking replication time lag
> Apr 21 10:16:49 vm1 journal: orker process[23745]: [1051-1] 2018-04-21 10:16:49:
> pid 23745: LOG:  failed to connect to PostgreSQL server on "vm1:5432",
> getsockopt() detected error "Connection refused"
> Apr 21 10:16:49 vm1 journal: orker process[23745]: [1052-1] 2018-04-21 10:16:49:
> pid 23745: ERROR:  failed to make persistent db connection
> Apr 21 10:16:49 vm1 journal: orker process[23745]: [1052-2] 2018-04-21 10:16:49:
> pid 23745: DETAIL:  connection to host:"vm1:5432" failed
> Apr 21 10:16:49 vm1 journal: ait for connection request[23482]: [797-1]
> 2018-04-21 10:16:49: pid 23482: LOG:  failback event detected
> Apr 21 10:16:49 vm1 journal: ait for connection request[23482]: [797-2]
> 2018-04-21 10:16:49: pid 23482: DETAIL:  restarting myself
> Apr 21 10:16:49 vm1 journal: ait for connection request[23481]: [795-1]
> 2018-04-21 10:16:49: pid 23481: LOG:  failback event detected
> Apr 21 10:16:49 vm1 journal: ait for connection request[23481]: [795-2]
> 2018-04-21 10:16:49: pid 23481: DETAIL:  restarting myself
> Apr 21 10:16:49 vm1 journal: ccept connection[24131]: [1063-1] 2018-04-21
> 10:16:49: pid 24131: LOG:  failed to connect to PostgreSQL server on "vm1:5432",
> getsockopt() detected error "Connection refused"
> Apr 21 10:16:49 vm1 journal: ccept connection[24131]: [1064-1] 2018-04-21
> 10:16:49: pid 24131: LOG:  received degenerate backend request for node_id: 0
> from pid [24131]
> Apr 21 10:16:49 vm1 atchdog[22511]: [60-1] 2018-04-21 10:16:49: pid 22511: LOG: 
> new IPC connection received
> Apr 21 10:16:49 vm1 atchdog[22511]: [61-1] 2018-04-21 10:16:49: pid 22511: LOG: 
> watchdog received the failover command from local pgpool-II on IPC interface
> Apr 21 10:16:49 vm1 atchdog[22511]: [62-1] 2018-04-21 10:16:49: pid 22511: LOG: 
> watchdog is processing the failover command [DEGENERATE_BACKEND_REQUEST]
> received from local pgpool-II on IPC interface
> Apr 21 10:16:49 vm1 atchdog[22511]: [63-1] 2018-04-21 10:16:49: pid 22511: LOG: 
> we have got the consensus to perform the failover
> Apr 21 10:16:49 vm1 atchdog[22511]: [63-2] 2018-04-21 10:16:49: pid 22511:
> DETAIL:  1 node(s) voted in the favor
> Apr 21 10:16:49 vm1 journal: ccept connection[24131]: [1065-1] 2018-04-21
> 10:16:49: pid 24131: FATAL:  failed to create a backend connection
> Apr 21 10:16:49 vm1 journal: ccept connection[24131]: [1065-2] 2018-04-21
> 10:16:49: pid 24131: DETAIL:  executing failover on backend
> Apr 21 10:16:49 vm1 pgpool[22509]: [1064-1] 2018-04-21 10:16:49: pid 22509: LOG:
>  Pgpool-II parent process has received failover request
> Apr 21 10:16:49 vm1 atchdog[22511]: [64-1] 2018-04-21 10:16:49: pid 22511: LOG: 
> new IPC connection received
> Apr 21 10:16:49 vm1 atchdog[22511]: [65-1] 2018-04-21 10:16:49: pid 22511: LOG: 
> received the failover indication from Pgpool-II on IPC interface
> Apr 21 10:16:49 vm1 atchdog[22511]: [66-1] 2018-04-21 10:16:49: pid 22511: LOG: 
> watchdog is informed of failover end by the main process
> Apr 21 10:16:49 vm1 pgpool[22509]: [1065-1] 2018-04-21 10:16:49: pid 22509: LOG:
>  starting degeneration. shutdown host vm1(5432)
> Apr 21 10:16:49 vm1 pgpool[22509]: [1066-1] 2018-04-21 10:16:49: pid 22509: LOG:
>  Restart all children
> Apr 21 10:16:49 vm1 pgpool[22509]: [1067-1] 2018-04-21 10:16:49: pid 22509: LOG:
>  execute command: failover.sh 0 0 vm2 /usr/local/data
> Apr 21 10:16:50 vm1 pgpool[22509]: [1068-1] 2018-04-21 10:16:50: pid 22509: LOG:
>  find_primary_node_repeatedly: waiting for finding a primary node
> Apr 21 10:16:50 vm1 pgpool[22509]: [1069-1] 2018-04-21 10:16:50: pid 22509: LOG:
>  find_primary_node: checking backend no 0
> Apr 21 10:16:50 vm1 pgpool[22509]: [1070-1] 2018-04-21 10:16:50: pid 22509: LOG:
>  find_primary_node: checking backend no 1
> Apr 21 10:16:50 vm1 pgpool[22509]: [1071-1] 2018-04-21 10:16:50: pid 22509: LOG:
>  find_primary_node: primary node id is 1
> Apr 21 10:16:50 vm1 pgpool[22509]: [1072-1] 2018-04-21 10:16:50: pid 22509: LOG:
>  failover: set new primary node: 1
> Apr 21 10:16:50 vm1 pgpool[22509]: [1073-1] 2018-04-21 10:16:50: pid 22509: LOG:
>  failover: set new master node: 1
> 
> 
> **************************************************************************************
> failover from secondary(vm2) to primary(vm1)
> **************************************************************************************
> Apr 21 10:22:44 vm1 journal: ealth check process(1)[23042]: [75-1] 2018-04-21
> 10:22:44: pid 23042: LOG:  health check retrying on DB node: 1 (round:10)
> Apr 21 10:22:49 vm1 journal: orker process[25487]: [3107-1] 2018-04-21 10:22:49:
> pid 25487: ERROR:  Failed to check replication time lag
> Apr 21 10:22:49 vm1 journal: orker process[25487]: [3107-2] 2018-04-21 10:22:49:
> pid 25487: DETAIL:  No persistent db connection for the node 1
> Apr 21 10:22:49 vm1 journal: orker process[25487]: [3107-3] 2018-04-21 10:22:49:
> pid 25487: HINT:  check sr_check_user and sr_check_password
> Apr 21 10:22:49 vm1 journal: orker process[25487]: [3107-4] 2018-04-21 10:22:49:
> pid 25487: CONTEXT:  while checking replication time lag
> Apr 21 10:22:49 vm1 journal: orker process[25487]: [3108-1] 2018-04-21 10:22:49:
> pid 25487: LOG:  failed to connect to PostgreSQL server on "vm2:5432",
> getsockopt() detected error "Connection refused"
> Apr 21 10:22:49 vm1 journal: orker process[25487]: [3109-1] 2018-04-21 10:22:49:
> pid 25487: ERROR:  failed to make persistent db connection
> Apr 21 10:22:49 vm1 journal: orker process[25487]: [3109-2] 2018-04-21 10:22:49:
> pid 25487: DETAIL:  connection to host:"vm2:5432" failed
> Apr 21 10:22:54 vm1 journal: ealth check process(1)[23042]: [76-1] 2018-04-21
> 10:22:54: pid 23042: LOG:  failed to connect to PostgreSQL server on "vm2:5432",
> getsockopt() detected error "Connection refused"
> Apr 21 10:22:54 vm1 journal: ealth check process(1)[23042]: [77-1] 2018-04-21
> 10:22:54: pid 23042: ERROR:  failed to make persistent db connection
> Apr 21 10:22:54 vm1 journal: ealth check process(1)[23042]: [77-2] 2018-04-21
> 10:22:54: pid 23042: DETAIL:  connection to host:"vm2:5432" failed
> Apr 21 10:22:54 vm1 journal: ealth check process(1)[23042]: [78-1] 2018-04-21
> 10:22:54: pid 23042: LOG:  health check failed on node 1 (timeout:0)
> Apr 21 10:22:54 vm1 journal: ealth check process(1)[23042]: [79-1] 2018-04-21
> 10:22:54: pid 23042: LOG:  received degenerate backend request for node_id: 1
> from pid [23042]
> Apr 21 10:22:54 vm1 atchdog[22511]: [91-1] 2018-04-21 10:22:54: pid 22511: LOG: 
> new IPC connection received
> Apr 21 10:22:54 vm1 atchdog[22511]: [92-1] 2018-04-21 10:22:54: pid 22511: LOG: 
> watchdog received the failover command from local pgpool-II on IPC interface
> Apr 21 10:22:54 vm1 atchdog[22511]: [93-1] 2018-04-21 10:22:54: pid 22511: LOG: 
> watchdog is processing the failover command [DEGENERATE_BACKEND_REQUEST]
> received from local pgpool-II on IPC interface
> Apr 21 10:22:54 vm1 atchdog[22511]: [94-1] 2018-04-21 10:22:54: pid 22511: LOG: 
> we have got the consensus to perform the failover
> Apr 21 10:22:54 vm1 atchdog[22511]: [94-2] 2018-04-21 10:22:54: pid 22511:
> DETAIL:  1 node(s) voted in the favor
> Apr 21 10:22:54 vm1 pgpool[22509]: [4088-1] 2018-04-21 10:22:54: pid 22509: LOG:
>  Pgpool-II parent process has received failover request
> Apr 21 10:22:54 vm1 atchdog[22511]: [95-1] 2018-04-21 10:22:54: pid 22511: LOG: 
> new IPC connection received
> Apr 21 10:22:54 vm1 atchdog[22511]: [96-1] 2018-04-21 10:22:54: pid 22511: LOG: 
> received the failover indication from Pgpool-II on IPC interface
> Apr 21 10:22:54 vm1 atchdog[22511]: [97-1] 2018-04-21 10:22:54: pid 22511: LOG: 
> watchdog is informed of failover end by the main process
> Apr 21 10:22:54 vm1 pgpool[22509]: [4089-1] 2018-04-21 10:22:54: pid 22509: LOG:
>  starting degeneration. shutdown host vm2(5432)
> Apr 21 10:22:54 vm1 pgpool[22509]: [4090-1] 2018-04-21 10:22:54: pid 22509: LOG:
>  Do not restart children because we are switching over node id 1 host: vm2 port:
> 5432 and we are in streaming replication mode
> Apr 21 10:22:54 vm1 pgpool[22509]: [4091-1] 2018-04-21 10:22:54: pid 22509: LOG:
>  execute command: failover.sh 1 0 vm1 /usr/local/data
> Apr 21 10:22:54 vm1 pgpool[22509]: [4092-1] 2018-04-21 10:22:54: pid 22509: LOG:
>  find_primary_node_repeatedly: waiting for finding a primary node
> Apr 21 10:22:54 vm1 pgpool[22509]: [4093-1] 2018-04-21 10:22:54: pid 22509: LOG:
>  find_primary_node: checking backend no 0
> Apr 21 10:22:54 vm1 pgpool[22509]: [4094-1] 2018-04-21 10:22:54: pid 22509: LOG:
>  find_primary_node: checking backend no 1
> Apr 21 10:22:55 vm1 pgpool[22509]: [4095-1] 2018-04-21 10:22:55: pid 22509: LOG:
>  find_primary_node: checking backend no 0
> Apr 21 10:22:55 vm1 pgpool[22509]: [4096-1] 2018-04-21 10:22:55: pid 22509: LOG:
>  find_primary_node: checking backend no 1
> Apr 21 10:22:56 vm1 pgpool[22509]: [4097-1] 2018-04-21 10:22:56: pid 22509: LOG:
>  find_primary_node: checking backend no 0
> Apr 21 10:22:56 vm1 pgpool[22509]: [4098-1] 2018-04-21 10:22:56: pid 22509: LOG:
>  find_primary_node: checking backend no 1
> Apr 21 10:22:57 vm1 pgpool[22509]: [4099-1] 2018-04-21 10:22:57: pid 22509: LOG:
>  find_primary_node: checking backend no 0
> Apr 21 10:22:57 vm1 pgpool[22509]: [4100-1] 2018-04-21 10:22:57: pid 22509: LOG:
>  find_primary_node: checking backend no 1
> Apr 21 10:22:58 vm1 pgpool[22509]: [4101-1] 2018-04-21 10:22:58: pid 22509: LOG:
>  find_primary_node: checking backend no 0
> Apr 21 10:22:58 vm1 pgpool[22509]: [4102-1] 2018-04-21 10:22:58: pid 22509: LOG:
>  find_primary_node: checking backend no 1
> Apr 21 10:22:59 vm1 journal: orker process[25487]: [3110-1] 2018-04-21 10:22:59:
> pid 25487: ERROR:  Failed to check replication time lag
> Apr 21 10:22:59 vm1 journal: orker process[25487]: [3110-2] 2018-04-21 10:22:59:
> pid 25487: DETAIL:  No persistent db connection for the node 1
> Apr 21 10:22:59 vm1 journal: orker process[25487]: [3110-3] 2018-04-21 10:22:59:
> pid 25487: HINT:  check sr_check_user and sr_check_password
> Apr 21 10:22:59 vm1 journal: orker process[25487]: [3110-4] 2018-04-21 10:22:59:
> pid 25487: CONTEXT:  while checking replication time lag
> Apr 21 10:22:59 vm1 journal: orker process[25487]: [3111-1] 2018-04-21 10:22:59:
> pid 25487: LOG:  failed to connect to PostgreSQL server on "vm2:5432",
> getsockopt() detected error "Connection refused"
> Apr 21 10:22:59 vm1 journal: orker process[25487]: [3112-1] 2018-04-21 10:22:59:
> pid 25487: ERROR:  failed to make persistent db connection
> Apr 21 10:22:59 vm1 journal: orker process[25487]: [3112-2] 2018-04-21 10:22:59:
> pid 25487: DETAIL:  connection to host:"vm2:5432" failed
> Apr 21 10:22:59 vm1 atchdog[22511]: [98-1] 2018-04-21 10:22:59: pid 22511: LOG: 
> watchdog received the failover command from remote pgpool-II node "vm2:9999
> Linux vrtmscpub7.in.rdlabs.hpecorp.net"
> Apr 21 10:22:59 vm1 atchdog[22511]: [99-1] 2018-04-21 10:22:59: pid 22511: LOG: 
> watchdog is processing the failover command [DEGENERATE_BACKEND_REQUEST]
> received from vm2:9999 Linux vrtmscpub7.in.rdlabs.hpecorp.net
> Apr 21 10:22:59 vm1 atchdog[22511]: [100-1] 2018-04-21 10:22:59: pid 22511: LOG:
>  we have got the consensus to perform the failover
> Apr 21 10:22:59 vm1 atchdog[22511]: [100-2] 2018-04-21 10:22:59: pid 22511:
> DETAIL:  1 node(s) voted in the favor
> Apr 21 10:22:59 vm1 atchdog[22511]: [101-1] 2018-04-21 10:22:59: pid 22511: LOG:
>  invalid degenerate backend request, node id : 1 status: [3] is not valid for
> failover
> Apr 21 10:22:59 vm1 pgpool[22509]: [4103-1] 2018-04-21 10:22:59: pid 22509: LOG:
>  find_primary_node: checking backend no 0
> Apr 21 10:22:59 vm1 pgpool[22509]: [4104-1] 2018-04-21 10:22:59: pid 22509: LOG:
>  find_primary_node: checking backend no 1
> Apr 21 10:23:00 vm1 pgpool[22509]: [4105-1] 2018-04-21 10:23:00: pid 22509: LOG:
>  find_primary_node: checking backend no 0
> Apr 21 10:23:00 vm1 pgpool[22509]: [4106-1] 2018-04-21 10:23:00: pid 22509: LOG:
>  find_primary_node: checking backend no 1
> Apr 21 10:23:01 vm1 pgpool[22509]: [4107-1] 2018-04-21 10:23:01: pid 22509: LOG:
>  find_primary_node: checking backend no 0
> Apr 21 10:23:01 vm1 pgpool[22509]: [4108-1] 2018-04-21 10:23:01: pid 22509: LOG:
>  find_primary_node: checking backend no 1
> Apr 21 10:23:02 vm1 pgpool[22509]: [4109-1] 2018-04-21 10:23:02: pid 22509: LOG:
>  find_primary_node: checking backend no 0
> Apr 21 10:23:02 vm1 pgpool[22509]: [4110-1] 2018-04-21 10:23:02: pid 22509: LOG:
>  find_primary_node: checking backend no 1
> Apr 21 10:23:03 vm1 pgpool[22509]: [4111-1] 2018-04-21 10:23:03: pid 22509: LOG:
>  find_primary_node: checking backend no 0
> Apr 21 10:23:03 vm1 pgpool[22509]: [4112-1] 2018-04-21 10:23:03: pid 22509: LOG:
>  find_primary_node: checking backend no 1
> Apr 21 10:23:04 vm1 pgpool[22509]: [4113-1] 2018-04-21 10:23:04: pid 22509: LOG:
>  failover: set new primary node: -1
> Apr 21 10:23:04 vm1 pgpool[22509]: [4114-1] 2018-04-21 10:23:04: pid 22509: LOG:
>  failover: set new master node: 0
> Apr 21 10:23:04 vm1 atchdog[22511]: [102-1] 2018-04-21 10:23:04: pid 22511: LOG:
>  new IPC connection received
> Apr 21 10:23:04 vm1 atchdog[22511]: [103-1] 2018-04-21 10:23:04: pid 22511: LOG:
>  received the failover indication from Pgpool-II on IPC interface
> Apr 21 10:23:04 vm1 journal: orker process[25487]: [3113-1] 2018-04-21 10:23:04:
> pid 25487: ERROR:  Failed to check replication time lag
> Apr 21 10:23:04 vm1 journal: orker process[25487]: [3113-2] 2018-04-21 10:23:04:
> pid 25487: DETAIL:  No persistent db connection for the node 1
> 
> ******************************************************************************
> pgpool configurations
> *******************************************************************************
> 
> # ----------------------------
> # pgPool-II configuration file
> # ----------------------------
> #
> # This file consists of lines of the form:
> #
> #   name = value
> #
> # Whitespace may be used.  Comments are introduced with "#" anywhere on a line.
> # The complete list of parameter names and allowed values can be found in the
> # pgPool-II documentation.
> #
> # This file is read on server startup and when the server receives a SIGHUP
> # signal.  If you edit the file on a running system, you have to SIGHUP the
> # server for the changes to take effect, or use "pgpool reload".  Some
> # parameters, which are marked below, require a server shutdown and restart to
> # take effect.
> #
> 
> 
> #------------------------------------------------------------------------------
> # CONNECTIONS
> #------------------------------------------------------------------------------
> 
> # - pgpool Connection Settings -
> 
> listen_addresses = '*'
>                                    # Host name or IP address to listen on:
>                                    # '*' for all, '' for no TCP/IP connections
>                                    # (change requires restart)
> port = 9999
>                                    # Port number
>                                    # (change requires restart)
> socket_dir = '/tmp'
>                                    # Unix domain socket path
>                                    # The Debian package defaults to
>                                    # /var/run/postgresql
>                                    # (change requires restart)
> listen_backlog_multiplier = 2
>                                    # Set the backlog parameter of listen(2) to
> 								   # num_init_children * listen_backlog_multiplier.
>                                    # (change requires restart)
> serialize_accept = off
>                                    # whether to serialize accept() call to avoid
> thundering herd problem
>                                    # (change requires restart)
> 
> # - pgpool Communication Manager Connection Settings -
> 
> pcp_listen_addresses = '*'
>                                    # Host name or IP address for pcp process to
> listen on:
>                                    # '*' for all, '' for no TCP/IP connections
>                                    # (change requires restart)
> pcp_port = 9898
>                                    # Port number for pcp
>                                    # (change requires restart)
> pcp_socket_dir = '/tmp'
>                                    # Unix domain socket path for pcp
>                                    # The Debian package defaults to
>                                    # /var/run/postgresql
>                                    # (change requires restart)
> 
> # - Backend Connection Settings -
> 
> backend_hostname0 = 'vm1'
>                                    # Host name or IP address to connect to for
> backend 0
> backend_port0 = 5432
>                                    # Port number for backend 0
> backend_weight0 = 1
>                                    # Weight for backend 0 (only in load
> balancing mode)
> backend_data_directory0 = '/usr/local/data'
>                                    # Data directory for backend 0
> backend_flag0 = 'ALLOW_TO_FAILOVER'
>                                    # Controls various backend behavior
>                                    # ALLOW_TO_FAILOVER or DISALLOW_TO_FAILOVER
> backend_hostname1 = 'vm2'
> backend_port1 = 5432
> backend_weight1 = 1
> backend_data_directory1 = '/usr/local/data'
> backend_flag1 = 'ALLOW_TO_FAILOVER'
> 
> # - Authentication -
> 
> enable_pool_hba = off
>                                    # Use pool_hba.conf for client authentication
> pool_passwd = 'pool_passwd'
>                                    # File name of pool_passwd for md5
> authentication.
>                                    # "" disables pool_passwd.
>                                    # (change requires restart)
> authentication_timeout = 60
>                                    # Delay in seconds to complete client
> authentication
>                                    # 0 means no timeout.
> 
> # - SSL Connections -
> 
> ssl = off
>                                    # Enable SSL support
>                                    # (change requires restart)
> #ssl_key = './server.key'
>                                    # Path to the SSL private key file
>                                    # (change requires restart)
> #ssl_cert = './server.cert'
>                                    # Path to the SSL public certificate file
>                                    # (change requires restart)
> #ssl_ca_cert = ''
>                                    # Path to a single PEM format file
>                                    # containing CA root certificate(s)
>                                    # (change requires restart)
> #ssl_ca_cert_dir = ''
>                                    # Directory containing CA root certificate(s)
>                                    # (change requires restart)
> 
> 
> #------------------------------------------------------------------------------
> # POOLS
> #------------------------------------------------------------------------------
> 
> # - Concurrent session and pool size -
> 
> num_init_children = 500
>                                    # Number of pools
>                                    # (change requires restart)
> max_pool = 1
>                                    # Number of connections pool caches per
> connection
>                                    # (change requires restart)
> 
> # - Life time -
> 
> child_life_time = 30
>                                    # Pool exits after being idle for this many
> seconds
> child_max_connections = 90
>                                    # Pool exits after receiving that many
> connections
>                                    # 0 means no exit
> connection_life_time = 0
>                                    # Connection to backend closes after being
> idle for this many seconds
>                                    # 0 means no close
> client_idle_limit = 0
>                                    # Client is disconnected after being idle for
> that many seconds
>                                    # (even inside an explicit transactions!)
>                                    # 0 means no disconnection
> 
> 
> #------------------------------------------------------------------------------
> # LOGS
> #------------------------------------------------------------------------------
> 
> # - Where to log -
> 
> log_destination = 'stderr,syslog'
>                                    # Where to log
>                                    # Valid values are combinations of stderr,
>                                    # and syslog. Default to stderr.
> 
> # - What to log -
> 
> log_line_prefix = '%t: pid %p: '   # printf-style string to output at beginning
> of each log line.
> 
> log_connections = off
>                                    # Log connections
> log_hostname = off
>                                    # Hostname will be shown in ps status
>                                    # and in logs if connections are logged
> log_statement = off
>                                    # Log all statements
> log_per_node_statement = off
>                                    # Log all statements
>                                    # with node and backend informations
> log_standby_delay = 'always'
>                                    # Log standby delay
>                                    # Valid values are combinations of always,
>                                    # if_over_threshold, none
> 
> # - Syslog specific -
> 
> syslog_facility = 'LOCAL0'
>                                    # Syslog local facility. Default to LOCAL0
> syslog_ident = 'pgpool'
>                                    # Syslog program identification string
>                                    # Default to 'pgpool'
> 
> # - Debug -
> 
> #log_error_verbosity = default          # terse, default, or verbose messages
> 
> #client_min_messages = notice           # values in order of decreasing detail:
>                                         #   debug5
>                                         #   debug4
>                                         #   debug3
>                                         #   debug2
>                                         #   debug1
>                                         #   log
>                                         #   notice
>                                         #   warning
>                                         #   error
> 
> #log_min_messages = warning             # values in order of decreasing detail:
>                                         #   debug5
>                                         #   debug4
>                                         #   debug3
>                                         #   debug2
>                                         #   debug1
>                                         #   info
>                                         #   notice
>                                         #   warning
>                                         #   error
>                                         #   log
>                                         #   fatal
>                                         #   panic
> 
> #------------------------------------------------------------------------------
> # FILE LOCATIONS
> #------------------------------------------------------------------------------
> 
> pid_file_name = '/usr/local/logs/pgpool.pid'
>                                    # PID file name
>                                    # Can be specified as relative to the"
>                                    # location of pgpool.conf file or
>                                    # as an absolute path
>                                    # (change requires restart)
> logdir = '/usr/local/logs'
>                                    # Directory of pgPool status file
>                                    # (change requires restart)
> 
> 
> #------------------------------------------------------------------------------
> # CONNECTION POOLING
> #------------------------------------------------------------------------------
> 
> connection_cache = on
>                                    # Activate connection pools
>                                    # (change requires restart)
> 
>                                    # Semicolon separated list of queries
>                                    # to be issued at the end of a session
>                                    # The default is for 8.3 and later
> reset_query_list = 'ABORT; DISCARD ALL'
>                                    # The following one is for 8.2 and before
> #reset_query_list = 'ABORT; RESET ALL; SET SESSION AUTHORIZATION DEFAULT'
> 
> 
> #------------------------------------------------------------------------------
> # REPLICATION MODE
> #------------------------------------------------------------------------------
> 
> replication_mode = off
>                                    # Activate replication mode
>                                    # (change requires restart)
> replicate_select = off
>                                    # Replicate SELECT statements
>                                    # when in replication mode
>                                    # replicate_select is higher priority than
>                                    # load_balance_mode.
> 
> insert_lock = on
>                                    # Automatically locks a dummy row or a table
>                                    # with INSERT statements to keep SERIAL data
>                                    # consistency
>                                    # Without SERIAL, no lock will be issued
> lobj_lock_table = ''
>                                    # When rewriting lo_creat command in
>                                    # replication mode, specify table name to
>                                    # lock
> 
> # - Degenerate handling -
> 
> replication_stop_on_mismatch = off
>                                    # On disagreement with the packet kind
>                                    # sent from backend, degenerate the node
>                                    # which is most likely "minority"
>                                    # If off, just force to exit this session
> 
> failover_if_affected_tuples_mismatch = off
>                                    # On disagreement with the number of affected
>                                    # tuples in UPDATE/DELETE queries, then
>                                    # degenerate the node which is most likely
>                                    # "minority".
>                                    # If off, just abort the transaction to
>                                    # keep the consistency
> 
> 
> #------------------------------------------------------------------------------
> # LOAD BALANCING MODE
> #------------------------------------------------------------------------------
> 
> load_balance_mode = off
>                                    # Activate load balancing mode
>                                    # (change requires restart)
> ignore_leading_white_space = on
>                                    # Ignore leading white spaces of each query
> white_function_list = ''
>                                    # Comma separated list of function names
>                                    # that don't write to database
>                                    # Regexp are accepted
> black_function_list = 'nextval,setval'
>                                    # Comma separated list of function names
>                                    # that write to database
>                                    # Regexp are accepted
> 
> database_redirect_preference_list = ''
> 								   # comma separated list of pairs of database and node id.
> 								   # example: postgres:primary,mydb[0-4]:1,mydb[5-9]:2'
> 								   # valid for streaming replicaton mode only.
> 
> app_name_redirect_preference_list = ''
> 								   # comma separated list of pairs of app name and node id.
> 								   # example: 'psql:primary,myapp[0-4]:1,myapp[5-9]:standby'
> 								   # valid for streaming replicaton mode only.
> allow_sql_comments = off
> 								   # if on, ignore SQL comments when judging if load balance or
> 								   # query cache is possible.
> 								   # If off, SQL comments effectively prevent the judgment
> 								   # (pre 3.4 behavior).
> 
> #------------------------------------------------------------------------------
> # MASTER/SLAVE MODE
> #------------------------------------------------------------------------------
> 
> master_slave_mode = on
>                                    # Activate master/slave mode
>                                    # (change requires restart)
> master_slave_sub_mode = 'stream'
>                                    # Master/slave sub mode
>                                    # Valid values are combinations slony or
>                                    # stream. Default is slony.
>                                    # (change requires restart)
> 
> # - Streaming -
> 
> sr_check_period = 10
>                                    # Streaming replication check period
>                                    # Disabled (0) by default
> sr_check_user = 'postgres'
>                                    # Streaming replication check user
>                                    # This is necessary even if you disable
>                                    # streaming replication delay check with
>                                    # sr_check_period = 0
> sr_check_password = 'postgres'
>                                    # Password for streaming replication check
> user
> sr_check_database = 'postgres'
>                                    # Database name for streaming replication
> check
> delay_threshold = 10
>                                    # Threshold before not dispatching query to
> standby node
>                                    # Unit is in bytes
>                                    # Disabled (0) by default
> 
> # - Special commands -
> 
> follow_master_command = ''
>                                    # Executes this command after master failover
>                                    # Special values:
>                                    #   %d = node id
>                                    #   %h = host name
>                                    #   %p = port number
>                                    #   %D = database cluster path
>                                    #   %m = new master node id
>                                    #   %H = hostname of the new master node
>                                    #   %M = old master node id
>                                    #   %P = old primary node id
>                                    #   %r = new master port number
>                                    #   %R = new master database cluster path
>                                    #   %% = '%' character
> 
> 
> #------------------------------------------------------------------------------
> # HEALTH CHECK GLOBAL PARAMETERS
> #------------------------------------------------------------------------------
> 
> health_check_period = 10
>                                    # Health check period
>                                    # Disabled (0) by default
> health_check_timeout = 20
>                                    # Health check timeout
>                                    # 0 means no timeout
> health_check_user = 'postgres'
>                                    # Health check user
> health_check_password = 'postgres'
>                                    # Password for health check user
> health_check_database = 'postgres'
>                                    # Database name for health check. If '',
> tries 'postgres' frist, then 'template1'
> 
> health_check_max_retries = 10
>                                    # Maximum number of times to retry a failed
> health check before giving up.
> health_check_retry_delay = 10
>                                    # Amount of time to wait (in seconds) between
> retries.
> connect_timeout = 10000
>                                    # Timeout value in milliseconds before giving
> up to connect to backend.
> 								   # Default is 10000 ms (10 second). Flaky network user may want to
> increase
> 								   # the value. 0 means no timeout.
> 								   # Note that this value is not only used for health check,
> 								   # but also for ordinary conection to backend.
> 
> #------------------------------------------------------------------------------
> # HEALTH CHECK PER NODE PARAMETERS (OPTIONAL)
> #------------------------------------------------------------------------------
> #health_check_period0 = 0
> #health_check_timeout0 = 20
> #health_check_user0 = 'nobody'
> #health_check_password0 = ''
> #health_check_database0 = ''
> #health_check_max_retries0 = 0
> #health_check_retry_delay0 = 1
> #connect_timeout0 = 10000
> 
> #------------------------------------------------------------------------------
> # FAILOVER AND FAILBACK
> #------------------------------------------------------------------------------
> 
> failover_command = 'failover.sh %d %P %H %R'
>                                    # Executes this command at failover
>                                    # Special values:
>                                    #   %d = node id
>                                    #   %h = host name
>                                    #   %p = port number
>                                    #   %D = database cluster path
>                                    #   %m = new master node id
>                                    #   %H = hostname of the new master node
>                                    #   %M = old master node id
>                                    #   %P = old primary node id
>                                    #   %r = new master port number
>                                    #   %R = new master database cluster path
>                                    #   %% = '%' character
> failback_command = ''
>                                    # Executes this command at failback.
>                                    # Special values:
>                                    #   %d = node id
>                                    #   %h = host name
>                                    #   %p = port number
>                                    #   %D = database cluster path
>                                    #   %m = new master node id
>                                    #   %H = hostname of the new master node
>                                    #   %M = old master node id
>                                    #   %P = old primary node id
>                                    #   %r = new master port number
>                                    #   %R = new master database cluster path
>                                    #   %% = '%' character
> 
> fail_over_on_backend_error = on
>                                    # Initiates failover when reading/writing to
> the
>                                    # backend communication socket fails
>                                    # If set to off, pgpool will report an
>                                    # error and disconnect the session.
> 
> search_primary_node_timeout = 10
>                                    # Timeout in seconds to search for the
>                                    # primary node when a failover occurs.
>                                    # 0 means no timeout, keep searching
>                                    # for a primary node forever.
> 
> #------------------------------------------------------------------------------
> # ONLINE RECOVERY
> #------------------------------------------------------------------------------
> 
> recovery_user = 'postgres'
>                                    # Online recovery user
> recovery_password = 'postgres'
>                                    # Online recovery password
> recovery_1st_stage_command = 'basebackup.sh'
>                                    # Executes a command in first stage
> recovery_2nd_stage_command = ''
>                                    # Executes a command in second stage
> recovery_timeout = 190
>                                    # Timeout in seconds to wait for the
>                                    # recovering node's postmaster to start up
>                                    # 0 means no wait
> client_idle_limit_in_recovery = 0
>                                    # Client is disconnected after being idle
>                                    # for that many seconds in the second stage
>                                    # of online recovery
>                                    # 0 means no disconnection
>                                    # -1 means immediate disconnection
> 
> 
> #------------------------------------------------------------------------------
> # WATCHDOG
> #------------------------------------------------------------------------------
> 
> # - Enabling -
> 
> use_watchdog = on
>                                     # Activates watchdog
>                                     # (change requires restart)
> 
> # -Connection to up stream servers -
> 
> trusted_servers = ''
>                                     # trusted server list which are used
>                                     # to confirm network connection
>                                     # (hostA,hostB,hostC,...)
>                                     # (change requires restart)
> ping_path = '/bin'
>                                     # ping command path
>                                     # (change requires restart)
> 
> # - Watchdog communication Settings -
> 
> wd_hostname = 'vm1'
>                                     # Host name or IP address of this watchdog
>                                     # (change requires restart)
> wd_port = 9000
>                                     # port number for watchdog service
>                                     # (change requires restart)
> wd_priority = 1
> 									# priority of this watchdog in leader election
> 									# (change requires restart)
> 
> wd_authkey = ''
>                                     # Authentication key for watchdog
> communication
>                                     # (change requires restart)
> 
> wd_ipc_socket_dir = '/tmp'
> 									# Unix domain socket path for watchdog IPC socket
> 									# The Debian package defaults to
> 									# /var/run/postgresql
> 									# (change requires restart)
> 
> 
> # - Virtual IP control Setting -
> 
> delegate_IP = '192.168.60.85'
>                                     # delegate IP address
>                                     # If this is empty, virtual IP never bring
> up. 
>                                     # (change requires restart)
> if_cmd_path = '/sbin'
>                                     # path to the directory where if_up/down_cmd
> exists 
>                                     # (change requires restart)
> if_up_cmd = 'ifconfig ens224:0 inet $_IP_$ netmask 255.255.255.0'
>                                     # startup delegate IP command
>                                     # (change requires restart)
> if_down_cmd = 'ifconfig ens224:0 down'
>                                     # shutdown delegate IP command
>                                     # (change requires restart)
> 
> arping_path = '/usr/sbin'           # arping command path
>                                     # (change requires restart)
> 
> arping_cmd = 'arping -U $_IP_$ -w 1'
>                                     # arping command
>                                     # (change requires restart)
> 
> # - Behaivor on escalation Setting -
> 
> clear_memqcache_on_escalation = on
>                                     # Clear all the query cache on shared memory
>                                     # when standby pgpool escalate to active
> pgpool
>                                     # (= virtual IP holder).
>                                     # This should be off if client connects to
> pgpool
>                                     # not using virtual IP.
>                                     # (change requires restart)
> wd_escalation_command = ''
>                                     # Executes this command at escalation on new
> active pgpool.
>                                     # (change requires restart)
> wd_de_escalation_command = ''
> 									# Executes this command when master pgpool resigns from being master.
> 									# (change requires restart)
> 
> # - Watchdog consensus settings for failover -
> 
> failover_when_quorum_exists = on
> 									# Only perform backend node failover
> 									# when the watchdog cluster holds the quorum
> 									# (change requires restart)
> 
> failover_require_consensus = on
> 									# Perform failover when majority of Pgpool-II nodes
> 									# aggrees on the backend node status change
> 									# (change requires restart)
> 
> allow_multiple_failover_requests_from_node = off
> 									# A Pgpool-II node can cast multiple votes
> 									# for building the consensus on failover
> 									# (change requires restart)
> 
> # - Lifecheck Setting - 
> 
> # -- common --
> 
> wd_monitoring_interfaces_list = ''  # Comma separated list of interfaces names
> to monitor.
> 									# if any interface from the list is active the watchdog will
> 									# consider the network is fine
> 									# 'any' to enable monitoring on all interfaces except loopback
> 									# '' to disable monitoring
> 									# (change requires restart)
> 
> 
> wd_lifecheck_method = 'heartbeat'
>                                     # Method of watchdog lifecheck ('heartbeat'
> or 'query')
>                                     # (change requires restart)
> wd_interval = 10
>                                     # lifecheck interval (sec) > 0
>                                     # (change requires restart)
> 
> # -- heartbeat mode --
> 
> wd_heartbeat_port = 9694
>                                     # Port number for receiving heartbeat signal
>                                     # (change requires restart)
> wd_heartbeat_keepalive = 2
>                                     # Interval time of sending heartbeat signal
> (sec)
>                                     # (change requires restart)
> wd_heartbeat_deadtime = 30
>                                     # Deadtime interval for heartbeat signal
> (sec)
>                                     # (change requires restart)
> heartbeat_destination0 = 'vm2'
>                                     # Host name or IP address of destination 0
>                                     # for sending heartbeat signal.
>                                     # (change requires restart)
> heartbeat_destination_port0 = 9694 
>                                     # Port number of destination 0 for sending
>                                     # heartbeat signal. Usually this is the
>                                     # same as wd_heartbeat_port.
>                                     # (change requires restart)
> heartbeat_device0 = 'ens224'
>                                     # Name of NIC device (such like 'eth0')
>                                     # used for sending/receiving heartbeat
>                                     # signal to/from destination 0.
>                                     # This works only when this is not empty
>                                     # and pgpool has root privilege.
>                                     # (change requires restart)
> 
> #heartbeat_destination1 = 'host0_ip2'
> #heartbeat_destination_port1 = 9694
> #heartbeat_device1 = ''
> 
> # -- query mode --
> 
> wd_life_point = 3
>                                     # lifecheck retry times
>                                     # (change requires restart)
> wd_lifecheck_query = 'SELECT 1'
>                                     # lifecheck query to pgpool from watchdog
>                                     # (change requires restart)
> wd_lifecheck_dbname = 'template1'
>                                     # Database name connected for lifecheck
>                                     # (change requires restart)
> wd_lifecheck_user = 'postgres'
>                                     # watchdog user monitoring pgpools in
> lifecheck
>                                     # (change requires restart)
> wd_lifecheck_password = ''
>                                     # Password for watchdog user in lifecheck
>                                     # (change requires restart)
> 
> # - Other pgpool Connection Settings -
> 
> other_pgpool_hostname0 = 'vm2'
>                                     # Host name or IP address to connect to for
> other pgpool 0
>                                     # (change requires restart)
> other_pgpool_port0 = 9999
>                                     # Port number for othet pgpool 0
>                                     # (change requires restart)
> other_wd_port0 = 9000
>                                     # Port number for othet watchdog 0
>                                     # (change requires restart)
> #other_pgpool_hostname1 = 'host1'
> #other_pgpool_port1 = 9999
> #other_wd_port1 = 9000
> 
> 
> #------------------------------------------------------------------------------
> # OTHERS
> #------------------------------------------------------------------------------
> relcache_expire = 0
>                                    # Life time of relation cache in seconds.
>                                    # 0 means no cache expiration(the default).
>                                    # The relation cache is used for cache the
>                                    # query result against PostgreSQL system
>                                    # catalog to obtain various information
>                                    # including table structures or if it's a
>                                    # temporary table or not. The cache is
>                                    # maintained in a pgpool child local memory
>                                    # and being kept as long as it survives.
>                                    # If someone modify the table by using
>                                    # ALTER TABLE or some such, the relcache is
>                                    # not consistent anymore.
>                                    # For this purpose, cache_expiration
>                                    # controls the life time of the cache.
> 
> relcache_size = 256
>                                    # Number of relation cache
>                                    # entry. If you see frequently:
>                                    # "pool_search_relcache: cache replacement
> happend"
>                                    # in the pgpool log, you might want to
> increate this number.
> 
> check_temp_table = on
>                                    # If on, enable temporary table check in
> SELECT statements.
>                                    # This initiates queries against system
> catalog of primary/master
>                                    # thus increases load of master.
>                                    # If you are absolutely sure that your system
> never uses temporary tables
>                                    # and you want to save access to
> primary/master, you could turn this off.
>                                    # Default is on.
> 
> check_unlogged_table = on
>                                    # If on, enable unlogged table check in
> SELECT statements.
>                                    # This initiates queries against system
> catalog of primary/master
>                                    # thus increases load of master.
>                                    # If you are absolutely sure that your system
> never uses unlogged tables
>                                    # and you want to save access to
> primary/master, you could turn this off.
>                                    # Default is on.
> 
> #------------------------------------------------------------------------------
> # IN MEMORY QUERY MEMORY CACHE
> #------------------------------------------------------------------------------
> memory_cache_enabled = off
> 								   # If on, use the memory cache functionality, off by default
> memqcache_method = 'shmem'
> 								   # Cache storage method. either 'shmem'(shared memory) or
> 								   # 'memcached'. 'shmem' by default
>                                    # (change requires restart)
> memqcache_memcached_host = 'localhost'
> 								   # Memcached host name or IP address. Mandatory if
> 								   # memqcache_method = 'memcached'.
> 								   # Defaults to localhost.
>                                    # (change requires restart)
> memqcache_memcached_port = 11211
> 								   # Memcached port number. Mondatory if memqcache_method = 'memcached'.
> 								   # Defaults to 11211.
>                                    # (change requires restart)
> memqcache_total_size = 67108864
> 								   # Total memory size in bytes for storing memory cache.
> 								   # Mandatory if memqcache_method = 'shmem'.
> 								   # Defaults to 64MB.
>                                    # (change requires restart)
> memqcache_max_num_cache = 1000000
> 								   # Total number of cache entries. Mandatory
> 								   # if memqcache_method = 'shmem'.
> 								   # Each cache entry consumes 48 bytes on shared memory.
> 								   # Defaults to 1,000,000(45.8MB).
>                                    # (change requires restart)
> memqcache_expire = 0
> 								   # Memory cache entry life time specified in seconds.
> 								   # 0 means infinite life time. 0 by default.
>                                    # (change requires restart)
> memqcache_auto_cache_invalidation = on
> 								   # If on, invalidation of query cache is triggered by corresponding
> 								   # DDL/DML/DCL(and memqcache_expire).  If off, it is only triggered
> 								   # by memqcache_expire.  on by default.
>                                    # (change requires restart)
> memqcache_maxcache = 409600
> 								   # Maximum SELECT result size in bytes.
> 								   # Must be smaller than memqcache_cache_block_size. Defaults to 400KB.
>                                    # (change requires restart)
> memqcache_cache_block_size = 1048576
> 								   # Cache block size in bytes. Mandatory if memqcache_method = 'shmem'.
> 								   # Defaults to 1MB.
>                                    # (change requires restart)
> memqcache_oiddir = '/var/log/pgpool/oiddir'
> 				   				   # Temporary work directory to record table oids
>                                    # (change requires restart)
> white_memqcache_table_list = ''
>                                    # Comma separated list of table names to
> memcache
>                                    # that don't write to database
>                                    # Regexp are accepted
> black_memqcache_table_list = ''
>                                    # Comma separated list of table names not to
> memcache
>                                    # that don't write to database
>                                    # Regexp are accepted
> 
> ====================================================================== 
> 
> Issue History 
> Date Modified    Username       Field                    Change               
> ====================================================================== 
> 2018-04-21 20:18 kavin          New Issue                                    
> 2018-04-21 20:18 kavin          Issue generated from: 0000393                   
> 
> ======================================================================
> 


More information about the pgpool-hackers mailing list