Granthana Biswas granthana.biswas at gmail.com
Thu Oct 30 02:25:37 PDT 2014
Also the nodelock table on node1 and 2 have too many rows for node 3:

select count(*) from "_cluster".sl_nodelock where nl_nodeid =3;
 count
-------
    67


On Thu, Oct 30, 2014 at 2:29 PM, Granthana Biswas <
granthana.biswas at gmail.com> wrote:

> Hi All,
>
> My replication setup is as  db1 -> db2 -> db3.
>
> On adding a new set to the cluster, the merge from node 3 is going on
> waiting state for node 3 to subscribe. Because of this, node 3 is lagging
> behind.
>
> These are the slonik commands that I used to add new set, subscribe and
> merge:
>
> ---------------------------------------------------------------------------------------------------------
> create set ( id = 2, origin = 1, comment = 'replication set for surcharge
> table');
>
> set add table (set id = 2, origin = 1, id = 1744, fully qualified name =
> 'public.t2', comment = 't2 table');
> set add sequence (set id = 2, origin = 1, id = 1756, fully qualified name
> = 'public.s2', comment = 's2 sequence');
>
> subscribe set ( id = 2, provider = 1, receiver = 2, forward = yes );
> subscribe set ( id = 2, provider = 2, receiver = 3, forward = yes );
>
> merge set ( id = 1, add id = 2, origin = 1);
>
> -----------------------------------------------------------------------------------------------------------
>
> Even though it goes it waiting mode, the sl_subscribe table shows the
> following:
>
>  sub_set | sub_provider | sub_receiver | sub_forward | sub_active
> ---------+--------------+--------------+-------------+------------
>        1 |            1 |            2 | t           | t
>        1 |            2 |            3 | t           | t
>        2 |            1 |            2 | t           | t
>        2 |            2 |            3 | t           | t
>
>
> But the slony log on node 3 shows the following repeatedly:
>
> NOTICE:  Slony-I: cleanup stale sl_nodelock entry for pid=29117
> NOTICE:  Slony-I: cleanup stale sl_nodelock entry for pid=30115
> NOTICE:  Slony-I: cleanup stale sl_nodelock entry for pid=30116
> NOTICE:  Slony-I: cleanup stale sl_nodelock entry for pid=30414
> NOTICE:  Slony-I: Logswitch to sl_log_2 initiated
> CONTEXT:  SQL statement "SELECT  "_cluster".logswitch_start()"
> PL/pgSQL function "cleanupevent" line 96 at PERFORM
> NOTICE:  truncate of <NULL> failed - doing delete
> NOTICE:  Slony-I: cleanup stale sl_nodelock entry for pid=31364
> NOTICE:  Slony-I: cleanup stale sl_nodelock entry for pid=31369
> NOTICE:  Slony-I: cleanup stale sl_nodelock entry for pid=31368
> NOTICE:  Slony-I: cleanup stale sl_nodelock entry for pid=32300
> NOTICE:  Slony-I: cleanup stale sl_nodelock entry for pid=1117
> NOTICE:  Slony-I: cleanup stale sl_nodelock entry for pid=1149
> NOTICE:  Slony-I: cleanup stale sl_nodelock entry for pid=1186
> NOTICE:  Slony-I: cleanup stale sl_nodelock entry for pid=1247
> NOTICE:  Slony-I: cleanup stale sl_nodelock entry for pid=1270
> NOTICE:  Slony-I: cleanup stale sl_nodelock entry for pid=1294
>
>
> It is continuously trying to cleanup stale nodelock entry for nl_nodeid=3
> and nl_conncnt=0.
>
> I tried stopping and starting the slon process for node 3 which didn't
> help. I don't see any errors in the other slony log files.
>
> Do I have to stop all slon processes of all nodes and start again?
>
> Regards,
> Granthana
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://lists.slony.info/pipermail/slony1-general/attachments/20141030/a67a2b34/attachment-0001.htm 


More information about the Slony1-general mailing list