[pgpool-hackers: 3373] Re: Behavior of resigned master watchdog node

Muhammad Usama m.usama at gmail.com
Thu Aug 8 15:21:24 JST 2019


Hi Ishii-San

Thanks for providing the log files along with the steps to reproduce the
issue. I have pushed the fix for it

https://git.postgresql.org/gitweb/?p=pgpool2.git;a=commitdiff;h=2d7702c6961e8949e97992c23a42c8616af36d84

Thanks
Best Regards
Muhammad Usama


On Fri, Aug 2, 2019 at 4:50 PM Tatsuo Ishii <ishii at sraoss.co.jp> wrote:

> > Hi Ishii-San
> >
> > On Fri, Aug 2, 2019 at 10:25 AM Tatsuo Ishii <ishii at sraoss.co.jp> wrote:
> >
> >> Hi Usama,
> >>
> >> In your commit:
> >>
> >>
> https://git.postgresql.org/gitweb/?p=pgpool2.git;a=commit;h=33df0d33df1ce701f07fecaeef5b87a2707c08f2
> >>
> >> "the master watchdog node should resign from master responsibilities
> >> if the primary backend node gets into quarantine state on that."
> >>
> >> While testing this feature, I noticed that the master watchdog node
> >> resign from master responsibilities as expected in this case. However
> >> none of standby nodes gets promoted to new primary. As a result,
> >> there's no primary node any more after the master watchdog resign.
> >>
> >> This is not very good because users cannot
> >
> >
> > This is not at all expected behavior, Can you please share the logs or
> > steps to reproduce the issue.
> > I will look into this on priority.
>
> Sure. Pgpool.logs attached.
>
> Here are steps to reproduce the issue.
>
> - Pgpool-II master branch head with patch [pgpool-hackers: 3361].
>
> 1) create 3 watchdog/3 PostgreSQL cluster: watchdog_setup -wn 3 -n 3
>
> $ psql -p 50000 -c "show pool_nodes" test
>
>  node_id | hostname | port  | status | lb_weight |  role   | select_cnt |
> load_balance_node | replication_delay | replication_state |
> replication_sync_state | last_status_change
>
> ---------+----------+-------+--------+-----------+---------+------------+-------------------+-------------------+-------------------+------------------------+---------------------
>  0       | /tmp     | 51000 | up     | 0.333333  | primary | 0          |
> true              | 0                 |                   |
>         | 2019-08-02 20:38:25
>  1       | /tmp     | 51001 | up     | 0.333333  | standby | 0          |
> false             | 0                 | streaming         | async
>         | 2019-08-02 20:38:25
>  2       | /tmp     | 51002 | up     | 0.333333  | standby | 0          |
> false             | 0                 | streaming         | async
>         | 2019-08-02 20:38:25
> (3 rows)
>
> $ pcp_watchdog_info -p 50001 -v -w
>
> Watchdog Cluster Information
> Total Nodes          : 3
> Remote Nodes         : 2
> Quorum state         : QUORUM EXIST
> Alive Remote Nodes   : 2
> VIP up on local node : YES
> Master Node Name     : localhost:50000 Linux tishii-CFSV7-1
> Master Host Name     : localhost
>
> Watchdog Node Information
> Node Name      : localhost:50000 Linux tishii-CFSV7-1
> Host Name      : localhost
> Delegate IP    : Not_Set
> Pgpool port    : 50000
> Watchdog port  : 50002
> Node priority  : 3
> Status         : 4
> Status Name    : MASTER
>
> Node Name      : localhost:50004 Linux tishii-CFSV7-1
> Host Name      : localhost
> Delegate IP    : Not_Set
> Pgpool port    : 50004
> Watchdog port  : 50006
> Node priority  : 2
> Status         : 7
> Status Name    : STANDBY
>
> Node Name      : localhost:50008 Linux tishii-CFSV7-1
> Host Name      : localhost
> Delegate IP    : Not_Set
> Pgpool port    : 50008
> Watchdog port  : 50010
> Node priority  : 1
> Status         : 7
> Status Name    : STANDBY
>
> 2) echo "0      down" > pgpoo0/log
>
> 3) check the result.
>
> $psql -p 50000 -c "show pool_nodes" test
>
>  node_id | hostname | port  | status | lb_weight |  role   | select_cnt |
> load_balance_node | replication_delay | replication_state |
> replication_sync_state | last_status_change
>
> ---------+----------+-------+--------+-----------+---------+------------+-------------------+-------------------+-------------------+------------------------+---------------------
>  0       | /tmp     | 51000 | up     | 0.333333  | standby | 0          |
> true              | 0                 |                   |
>         | 2019-08-02 20:39:25
>  1       | /tmp     | 51001 | up     | 0.333333  | standby | 0          |
> false             | 0                 |                   |
>         | 2019-08-02 20:39:25
>  2       | /tmp     | 51002 | up     | 0.333333  | standby | 0          |
> false             | 0                 |                   |
>         | 2019-08-02 20:39:25
> (3 rows)
>
> $ pcp_watchdog_info -p 50001 -v -w
> Watchdog Cluster Information
> Total Nodes          : 3
> Remote Nodes         : 2
> Quorum state         : QUORUM EXIST
> Alive Remote Nodes   : 2
> VIP up on local node : NO
> Master Node Name     : localhost:50004 Linux tishii-CFSV7-1
> Master Host Name     : localhost
>
> Watchdog Node Information
> Node Name      : localhost:50000 Linux tishii-CFSV7-1
> Host Name      : localhost
> Delegate IP    : Not_Set
> Pgpool port    : 50000
> Watchdog port  : 50002
> Node priority  : 3
> Status         : 7
> Status Name    : STANDBY
>
> Node Name      : localhost:50004 Linux tishii-CFSV7-1
> Host Name      : localhost
> Delegate IP    : Not_Set
> Pgpool port    : 50004
> Watchdog port  : 50006
> Node priority  : 2
> Status         : 4
> Status Name    : MASTER
>
> Node Name      : localhost:50008 Linux tishii-CFSV7-1
> Host Name      : localhost
> Delegate IP    : Not_Set
> Pgpool port    : 50008
> Watchdog port  : 50010
> Node priority  : 1
> Status         : 7
> Status Name    : STANDBY
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.sraoss.jp/pipermail/pgpool-hackers/attachments/20190808/1c3523b9/attachment.html>


More information about the pgpool-hackers mailing list