Thu Apr 14 22:22:25 PDT 2005
- Previous message: [Slony1-commit] By cbbrowne: Normalized a bunch of stuff to deal with error messages
- Next message: [Slony1-commit] By cbbrowne: Added numerous omitted end tags
- Messages sorted by: [ date ] [ thread ] [ subject ] [ author ]
Log Message:
-----------
Normalize various SGML documentation in order to cut down on
error messages
Modified Files:
--------------
slony1-engine/doc/adminguide:
adminscripts.sgml (r1.22 -> r1.23)
defineset.sgml (r1.14 -> r1.15)
help.sgml (r1.14 -> r1.15)
intro.sgml (r1.13 -> r1.14)
logshipping.sgml (r1.7 -> r1.8)
plainpaths.sgml (r1.7 -> r1.8)
slonik_ref.sgml (r1.20 -> r1.21)
usingslonik.sgml (r1.8 -> r1.9)
-------------- next part --------------
Index: adminscripts.sgml
===================================================================
RCS file: /usr/local/cvsroot/slony1/slony1-engine/doc/adminguide/adminscripts.sgml,v
retrieving revision 1.22
retrieving revision 1.23
diff -Ldoc/adminguide/adminscripts.sgml -Ldoc/adminguide/adminscripts.sgml -u -w -r1.22 -r1.23
--- doc/adminguide/adminscripts.sgml
+++ doc/adminguide/adminscripts.sgml
@@ -17,8 +17,7 @@
couple of occasions, to pretty calamitous actions, so the behavior has
been changed so that the scripts simply submit output to standard
output. An administrator should review the script
-<emphasis>before</emphasis> submitting it to <xref
-linkend="slonik">.</para>
+<emphasis>before</emphasis> submitting it to <xref linkend="slonik">.</para>
<sect2><title>Node/Cluster Configuration - cluster.nodes</title>
Index: slonik_ref.sgml
===================================================================
RCS file: /usr/local/cvsroot/slony1/slony1-engine/doc/adminguide/slonik_ref.sgml,v
retrieving revision 1.20
retrieving revision 1.21
diff -Ldoc/adminguide/slonik_ref.sgml -Ldoc/adminguide/slonik_ref.sgml -u -w -r1.20 -r1.21
--- doc/adminguide/slonik_ref.sgml
+++ doc/adminguide/slonik_ref.sgml
@@ -1,6 +1,5 @@
<article id="slonikref">
<title>Slonik Command Summary</title>
-
<sect1><title>Introduction</title>
<para>
@@ -84,7 +83,7 @@
<para> Those commands are grouped together into one transaction
per participating node. </para>
-<!-- ************************************************************ -->
+<!-- ************************************************************ --></sect3></sect2></sect1></article>
<reference id="metacmds">
<title>Slonik Meta Commands</title>
Index: help.sgml
===================================================================
RCS file: /usr/local/cvsroot/slony1/slony1-engine/doc/adminguide/help.sgml,v
retrieving revision 1.14
retrieving revision 1.15
diff -Ldoc/adminguide/help.sgml -Ldoc/adminguide/help.sgml -u -w -r1.14 -r1.15
--- doc/adminguide/help.sgml
+++ doc/adminguide/help.sgml
@@ -40,11 +40,6 @@
KirovOpenSourceCommunity: Slony</ulink> may be the place to
go.</para></listitem>
-<listitem><para> A <ulink
-url="http://www.kuzin.net/work/sloniki-privet.html"> Russian Setup /
-Example / HOWTO </ulink> is available for Russian readers. </para>
-</listitem>
-
<listitem><para> <ulink url="http://pgpool.projects.postgresql.org/"
id="pgpool"> <application> pgpool </application> </ulink> </para>
@@ -69,8 +64,8 @@
config files in an XML-based format that the tool transforms into a
Slonik script</para></listitem>
-<listitem><Para><ulink url="http://freshmeat.net/projects/slonyi/">
-Freshmeat on &slony1; </ulink>
+<listitem><para><ulink url="http://freshmeat.net/projects/slonyi/">
+Freshmeat on &slony1; </ulink></para></listitem>
</itemizedlist>
</sect2>
Index: logshipping.sgml
===================================================================
RCS file: /usr/local/cvsroot/slony1/slony1-engine/doc/adminguide/logshipping.sgml,v
retrieving revision 1.7
retrieving revision 1.8
diff -Ldoc/adminguide/logshipping.sgml -Ldoc/adminguide/logshipping.sgml -u -w -r1.7 -r1.8
--- doc/adminguide/logshipping.sgml
+++ doc/adminguide/logshipping.sgml
@@ -2,15 +2,15 @@
<sect1 id="logshipping">
<title>Log Shipping - &slony1; with Files</title>
-<para> One of the new features for &slony1; 1.1 is the ability to
-serialize the updates to go out into log files that can be kept in a
-spool directory.
+<para> One of the new features for 1.1 is the ability to serialize the
+updates to go out into log files that can be kept in a spool
+directory.</para>
<para> The spool files could then be transferred via whatever means
was desired to a <quote>slave system,</quote> whether that be via FTP,
rsync, or perhaps even by pushing them onto a 1GB <quote>USB
key</quote> to be sent to the destination by clipping it to the ankle
-of some sort of <quote>avian transport</quote> system.
+of some sort of <quote>avian transport</quote> system.</para>
<para> There are plenty of neat things you can do with a data stream
in this form, including:
@@ -18,48 +18,48 @@
<itemizedlist>
<listitem><para> Replicating to nodes that
- <emphasis>aren't</emphasis> securable
+ <emphasis>aren't</emphasis> securable</para></listitem>
<listitem><para> Replicating to destinations where it is not
- possible to set up bidirection communications
+ possible to set up bidirection communications</para></listitem>
- <listitem><para> Supporting a different form of <acronym/PITR/
+ <listitem><para> Supporting a different form of <acronym>PITR</acronym>
(Point In Time Recovery) that filters out read-only transactions and
- updates to tables that are not of interest.
+ updates to tables that are not of interest.</para></listitem>
<listitem><para> If some disaster strikes, you can look at the logs
- of queries in detail
+ of queries in detail</para>
<para> This makes log shipping potentially useful even though you
- might not intend to actually create a log-shipped node.
+ might not intend to actually create a log-shipped node.</para></listitem>
<listitem><para> This is a really slick scheme for building load for
- doing tests
+ doing tests</para></listitem>
<listitem><para> We have a data <quote>escrow</quote> system that
- would become incredibly cheaper given log shipping
+ would become incredibly cheaper given log shipping</para></listitem>
<listitem><para> You may apply triggers on the <quote>disconnected
- node </quote> to do additional processing on the data
+ node </quote> to do additional processing on the data</para>
<para> For instance, you might take a fairly <quote>stateful</quote>
database and turn it into a <quote>temporal</quote> one by use of
triggers that implement the techniques described in
<citation>Developing Time-Oriented Database Applications in SQL
</citation> by <ulink url= "http://www.cs.arizona.edu/people/rts/">
- Richard T. Snodgrass</ulink>.
+ Richard T. Snodgrass</ulink>.</para></listitem>
-</itemizedlist>
+</itemizedlist></para>
<qandaset>
<qandaentry>
<question> <para> Where are the <quote>spool files</quote> for a
-subscription set generated?
+subscription set generated?</para>
</question>
<answer> <para> Any <link linkend="slon"> slon </link> node can
-generate them by adding the <option>-a</option> option.
+generate them by adding the <option>-a</option> option.</para>
</answer>
</qandaentry>
<qandaentry>
@@ -126,9 +126,9 @@
to rerun <application>slony1_dump.sh</application>:
<itemizedlist>
-<listitem><para><command> SUBSCRIBE_SET </command>
+<listitem><para><command> SUBSCRIBE_SET </command></para></listitem>
-</itemizedlist>
+</itemizedlist></para>
<para> A number of event types <emphasis> are </emphasis> handled in
such a way that log shipping copes with them:
@@ -136,23 +136,23 @@
<itemizedlist>
<listitem><para><command>SYNC </command> events are, of course,
-handled.
+handled.</para></listitem>
-<listitem><para><command>DDL_SCRIPT</command> is handled.
+<listitem><para><command>DDL_SCRIPT</command> is handled.</para></listitem>
-<listitem><para><command> UNSUBSCRIBE_SET </command>
+<listitem><para><command> UNSUBSCRIBE_SET </command></para>
<para> This event, much like <command>SUBSCRIBE_SET</command> is not
handled by the log shipping code. But its effect is, namely that
<command>SYNC</command> events on the subscriber node will no longer
-contain updates to the set.
+contain updates to the set.</para>
<para> Similarly, <command>SET_DROP_TABLE</command>,
<command>SET_DROP_SEQUENCE</command>,
<command>SET_MOVE_TABLE</command>,
<command>SET_MOVE_SEQUENCE</command> <command>DROP_SET</command>,
<command>MERGE_SET</command>, will be handled
-<quote>appropriately</quote>.
+<quote>appropriately</quote>.</para></listitem>
<listitem><para> The various events involved in node configuration are
irrelevant to log shipping:
@@ -163,7 +163,7 @@
<command>STORE_PATH</command>,
<command>DROP_PATH</command>,
<command>STORE_LISTEN</command>,
-<command>DROP_LISTEN</command>
+<command>DROP_LISTEN</command></para></listitem>
<listitem><para> Events involved in describing how particular sets are
to be initially configured are similarly irrelevant:
@@ -172,7 +172,7 @@
<command>SET_ADD_TABLE</command>,
<command>SET_ADD_SEQUENCE</command>,
<command>STORE_TRIGGER</command>,
-<command>DROP_TRIGGER</command>,
+<command>DROP_TRIGGER</command>,</para></listitem>
</itemizedlist>
</para>
@@ -183,7 +183,7 @@
could failover to. This would be quite useful if you were trying to
construct a cluster of (say) 6 nodes; you could start by creating one
subscriber, and then use log shipping to populate the other 4 in
-parallel.
+parallel.</para>
<para> This usage is not supported, but presumably one could add the
&slony1; configuration to the node, and promote it into being a new
@@ -206,7 +206,7 @@
your <application> psql</application> session will <command> ABORT
</command>, and then run through the remainder of that SYNC file
looking for a <command>COMMIT</command> or <command>ROLLBACK</command>
-so that it can try to move on to the next transaction.
+so that it can try to move on to the next transaction.</para>
<para> But we <emphasis> know </emphasis> that the entire remainder of
the file will fail! It is futile to go through the parsing effort of
@@ -217,21 +217,21 @@
<itemizedlist>
<listitem><para> Read the first few lines of the file, up to and
-including the <function> setsyncTracking_offline() </function> call.
+including the <function> setsyncTracking_offline() </function> call.</para></listitem>
-<listitem><para> Try to apply it that far.
+<listitem><para> Try to apply it that far.</para></listitem>
<listitem><para> If that failed, then it is futile to continue;
<command>ROLLBACK</command> the transaction, and perhaps consider
-trying the next file.
+trying the next file.</para></listitem>
<listitem><para> If the <function> setsyncTracking_offline()
</function> call succeeds, then you have the right next SYNC file, and
should apply it. You should probably <command>ROLLBACK</command> the
transaction, and then use <application>psql</application> to apply the
-entire file full of updates.
+entire file full of updates.</para></listitem>
-</itemizedlist>
+</itemizedlist></para>
<para> In order to support the approach of grabbing just the first few
lines of the sync file, the format has been set up to have a line of
@@ -243,7 +243,7 @@
start transaction;
select "_T1".setsyncTracking_offline(1, '744', '745');
--------------------------------------------------------------------
-</programlisting>
+</programlisting></para></listitem>
</itemizedlist>
</sect2>
Index: usingslonik.sgml
===================================================================
RCS file: /usr/local/cvsroot/slony1/slony1-engine/doc/adminguide/usingslonik.sgml,v
retrieving revision 1.8
retrieving revision 1.9
diff -Ldoc/adminguide/usingslonik.sgml -Ldoc/adminguide/usingslonik.sgml -u -w -r1.8 -r1.9
--- doc/adminguide/usingslonik.sgml
+++ doc/adminguide/usingslonik.sgml
@@ -283,15 +283,15 @@
pa_server, pa_client) from _slonycluster.sl_path;</command></para>
<para> The result of this set of queries is to regenerate
-<emphasis/and propagate/ the listen paths. By running the main
-<function/ _slonycluster.storelisten()/ function,
-<command/STORE_LISTEN/ events are raised to cause the listen paths to
-propagate to the other nodes in the cluster.
+<emphasis>and propagate</emphasis> the listen paths. By running the main
+<function> _slonycluster.storelisten()</function> function,
+<command>STORE_LISTEN</command> events are raised to cause the listen paths to
+propagate to the other nodes in the cluster.</para>
-<para> If there was a <emphasis/local/ problem on one node, and you
+<para> If there was a <emphasis>local</emphasis> problem on one node, and you
didn't want the updates to propagate (this would be an unusual
situation; you almost certainly want to fix things
-<emphasis/everywhere/), the queries would instead be:
+<emphasis>everywhere</emphasis>), the queries would instead be:</para>
<para> <command> select
slonycluster.droplisten_int(li_origin,li_provider,li_receiver) from
@@ -312,35 +312,35 @@
<listitem><para> The <quote>main</quote> function
(<emphasis>e.g.</emphasis> - without the <command>_int</command>
suffix) is called on a <quote>relevant</quote> node in the &slony1;
-cluster.
+cluster.</para>
<para> In some cases, the function may be called on any node, and it
can satisfactorily propagate to other nodes. That is true for <xref
linkend="function.storelisten-integer-integer-integer">, for instance.
In other cases, the function needs to be called on some particular
node because that is the only place where data may be reasonably
-validated. For instance, <xref linkend=
-"function.subscribeset-integer-integer-integer-boolean"> must be
-called on the receivernode.
+validated. For instance, <xref
+ linkend="function.subscribeset-integer-integer-integer-boolean"> must be
+called on the receivernode.</para></listitem>
<listitem><para> If that <quote>main</quote> function succeeds, then
the configuration change is performed on the local node, and an event
is created using <xref linkend="function.createevent-name-text"> to
cause the configuration change to propagate to all of the other nodes
-in the &slony1; cluster.
+in the &slony1; cluster.</para></listitem>
<listitem><para> Thirdly, the <command>_int</command> version of the
-function must be run.
+function must be run.</para>
<para> In some cases, where functions are idempotent, the node where
the <quote>main</quote> function runs can do this fairly early in
-processing.
+processing.</para>
<para> When the event propagates, the other nodes will run the
<command>_int</command> version, we rather hope with good
-success. </para>
+success. </para></listitem>
-</itemizedlist>
+</itemizedlist></para>
</sect1>
<!-- Keep this comment at the end of the file
- Previous message: [Slony1-commit] By cbbrowne: Normalized a bunch of stuff to deal with error messages
- Next message: [Slony1-commit] By cbbrowne: Added numerous omitted end tags
- Messages sorted by: [ date ] [ thread ] [ subject ] [ author ]
More information about the Slony1-commit mailing list