Chris Browne cbbrowne at lists.slony.info
Fri Oct 19 08:22:37 PDT 2007
Update of /home/cvsd/slony1/slony1-engine/tests/testlogship
In directory main.slony.info:/tmp/cvs-serv6908

Modified Files:
	README generate_dml.sh init_subscribe_set.ik moveset.sh 
	settings.ik 
Log Message:
Lots of changes to log shipping test based on the changes introduced
in 1.2.12


Index: settings.ik
===================================================================
RCS file: /home/cvsd/slony1/slony1-engine/tests/testlogship/settings.ik,v
retrieving revision 1.3
retrieving revision 1.4
diff -C2 -d -r1.3 -r1.4
*** settings.ik	27 Aug 2007 15:43:03 -0000	1.3
--- settings.ik	19 Oct 2007 15:22:35 -0000	1.4
***************
*** 1,6 ****
! NUMCLUSTERS=${NUMCLUSTERS:-"1"}
! NUMNODES=${NUMNODES:-"4"}
! ORIGINNODE=1
! WORKERS=${WORKERS:-"1"}
  ARCHIVE2=true   # Node #2 needs to run log archiving
! LOGSHIP3=true   # Node #3 receives data via log shipping
--- 1,8 ----
! NUMCLUSTERS=1
! NUMNODES=3    # These are the "regular" Slony-I nodes
!               # node #4 will also be populated as a log shipping node
! 
! ORIGINNODE=1  # at least initially - origin is later moved to node #3
! WORKERS=1
  ARCHIVE2=true   # Node #2 needs to run log archiving
! SLONCONF2=true

Index: init_subscribe_set.ik
===================================================================
RCS file: /home/cvsd/slony1/slony1-engine/tests/testlogship/init_subscribe_set.ik,v
retrieving revision 1.3
retrieving revision 1.4
diff -C2 -d -r1.3 -r1.4
*** init_subscribe_set.ik	27 Aug 2007 15:43:03 -0000	1.3
--- init_subscribe_set.ik	19 Oct 2007 15:22:35 -0000	1.4
***************
*** 3,7 ****
  sleep (seconds = 2);
  echo 'done sleeping...';
! subscribe set (id = 1, provider = 1, receiver = 4, forward = yes);
  echo 'sleep a couple of seconds...';
  sleep (seconds = 2);
--- 3,7 ----
  sleep (seconds = 2);
  echo 'done sleeping...';
! subscribe set (id = 1, provider = 1, receiver = 3, forward = yes);
  echo 'sleep a couple of seconds...';
  sleep (seconds = 2);

Index: README
===================================================================
RCS file: /home/cvsd/slony1/slony1-engine/tests/testlogship/README,v
retrieving revision 1.3
retrieving revision 1.4
diff -C2 -d -r1.3 -r1.4
*** README	20 Apr 2007 21:43:14 -0000	1.3
--- README	19 Oct 2007 15:22:35 -0000	1.4
***************
*** 28,30 ****
  table 4, adding two new columns, one to be populated via a default,
  for new tuples; the other has no default, but we assign the value 42
! to all tuples existing at the time that the DDL script runs.
\ No newline at end of file
--- 28,33 ----
  table 4, adding two new columns, one to be populated via a default,
  for new tuples; the other has no default, but we assign the value 42
! to all tuples existing at the time that the DDL script runs.
! 
! Note that node #2 (subscriber) has slon configuration managed via
! config file

Index: moveset.sh
===================================================================
RCS file: /home/cvsd/slony1/slony1-engine/tests/testlogship/moveset.sh,v
retrieving revision 1.2
retrieving revision 1.3
diff -C2 -d -r1.2 -r1.3
*** moveset.sh	27 Aug 2007 15:43:03 -0000	1.2
--- moveset.sh	19 Oct 2007 15:22:35 -0000	1.3
***************
*** 2,5 ****
  echo "
    LOCK SET ( ID = 1, ORIGIN = 1 );
!   MOVE SET ( ID = 1, OLD ORIGIN = 1, NEW ORIGIN = 4 );
  "
--- 2,5 ----
  echo "
    LOCK SET ( ID = 1, ORIGIN = 1 );
!   MOVE SET ( ID = 1, OLD ORIGIN = 1, NEW ORIGIN = 3 );
  "

Index: generate_dml.sh
===================================================================
RCS file: /home/cvsd/slony1/slony1-engine/tests/testlogship/generate_dml.sh,v
retrieving revision 1.4
retrieving revision 1.5
diff -C2 -d -r1.4 -r1.5
*** generate_dml.sh	27 Aug 2007 15:43:03 -0000	1.4
--- generate_dml.sh	19 Oct 2007 15:22:35 -0000	1.5
***************
*** 23,27 ****
  generate_initdata()
  {
!   numrows=$(random_number 50 1000)
    i=0;
    trippoint=`expr $numrows / 20`
--- 23,27 ----
  generate_initdata()
  {
!   numrows=$(random_number 125 150)
    i=0;
    trippoint=`expr $numrows / 20`
***************
*** 78,94 ****
    fi 
    status "data load complete - nodes are seeded reasonably"
- 
-   status "purge archive log files up to present in order to eliminate those that cannot be used"
-   for file in `/usr/bin/find ${mktmp}/archive_logs_2 -name "slony1_log_*.sql" -type f`; do
-     status "purge ${file}"
-     rm ${file}
-   done
-   sleep 5
    status "pull log shipping dump" 
    PGHOST=${HOST2} PGPORT=${PORT2} PGUSER=${USER2} ${SLTOOLDIR}/slony1_dump.sh ${DB2} ${CLUSTER1} > ${mktmp}/logship_dump.sql
-   status "load schema for replicated tables into node #3"
-   ${PGBINDIR3}/psql -h ${HOST3} -p ${PORT3} -U ${USER3} -d ${DB3} -f ${testname}/init_schema.sql
-   status "load log shipping dump into node #3"
-   ${PGBINDIR3}/psql -h ${HOST3} -p ${PORT3} -U ${USER3} -d ${DB3} -f ${mktmp}/logship_dump.sql
    
    status "generate more data to test log shipping"
--- 78,83 ----
***************
*** 116,120 ****
    wait_for_catchup
  
!   status "move set to node 4"
  
    init_preamble
--- 105,109 ----
    wait_for_catchup
  
!   status "move set to node 3"
  
    init_preamble
***************
*** 125,147 ****
  
    generate_initdata
!   eval db=\$DB4
!   status "loading extra data to node $db"
!   $pgbindir/psql -h $host -p $port -U $user -d $db < $mktmp/generate.data 1> ${mktmp}/even_more_data.log 2> ${mktmp}/even_more_data.log2
  
    wait_for_catchup
! 
  
    status "final data load complete - now load files into log shipped node"
!   firstseq=`(cd ${mktmp}/archive_logs_2; /usr/bin/find -name "*.sql") | cut -d "_" -f 4 | cut -d "." -f 1 | sort -n | head -1`
    lastseq=`(cd ${mktmp}/archive_logs_2; /usr/bin/find -name "*.sql") | cut -d "_" -f 4 | cut -d "." -f 1 | sort -n | tail -1`
    status "Logs numbered from ${firstseq} to ${lastseq}"
    currseq=${firstseq}
    while : ; do
- #      00000000000000000000
- #      12345678901234567890
      cs=`printf "%020d" ${currseq}`
      status "current sequence value: ${cs}"
      for logfile in `/usr/bin/find ${mktmp}/archive_logs_2 -name "slony1_log_*_${cs}.sql" -type f`; do
!       $pgbindir/psql -h ${HOST3} -p ${PORT3} -d ${DB3} -U ${USER3} -f ${logfile} >> $mktmp/logshipping_output.log 2>> $mktmp/logshipping_errors.log
        status "load file ${logfile} - ${?}"
      done
--- 114,145 ----
  
    generate_initdata
!   status "loading extra data to node 3"
!   $pgbindir/psql -h $HOST3 -p $PORT3 -U $USER3 -d $DB3 < $mktmp/generate.data 1> ${mktmp}/even_more_data.log 2> ${mktmp}/even_more_data.log2
  
    wait_for_catchup
!   status "done"
  
    status "final data load complete - now load files into log shipped node"
!   status "set up database for log shipped node"
!   ${PGBINDIR4}/createdb -p ${PORT4} -U ${USER4} ${DB4}
!   ${PGBINDIR4}/createlang plpgsql ${DB4}
! 
!   status "load schema for replicated tables into node #4"
!   ${PGBINDIR4}/psql -h ${HOST4} -p ${PORT4} -U ${USER4} -d ${DB4} -f ${testname}/init_schema.sql
!   status "load log shipping dump into node #4"
!   ${PGBINDIR4}/psql -h ${HOST4} -p ${PORT4} -U ${USER4} -d ${DB4} -f ${mktmp}/logship_dump.sql
!   
!   
!   firstseq=`psql -At -d ${DB4} -p ${PORT4} -c 'select at_counter from _slony_regress1.sl_archive_tracking ;'`
    lastseq=`(cd ${mktmp}/archive_logs_2; /usr/bin/find -name "*.sql") | cut -d "_" -f 4 | cut -d "." -f 1 | sort -n | tail -1`
    status "Logs numbered from ${firstseq} to ${lastseq}"
    currseq=${firstseq}
    while : ; do
      cs=`printf "%020d" ${currseq}`
      status "current sequence value: ${cs}"
+     firstseq=`psql -At -d ${DB4} -p ${PORT4} -c 'select at_counter from _slony_regress1.sl_archive_tracking ;'`
+     status "archive tracking ID: ${firstseq}"
      for logfile in `/usr/bin/find ${mktmp}/archive_logs_2 -name "slony1_log_*_${cs}.sql" -type f`; do
!       ${PGBINDIR4}/psql -h ${HOST4} -p ${PORT4} -d ${DB4} -U ${USER4} -f ${logfile} >> $mktmp/logshipping_output.log 2>> $mktmp/logshipping_errors.log
        status "load file ${logfile} - ${?}"
      done
***************
*** 152,155 ****
      fi
    done
!   status "done"
  }
--- 150,157 ----
      fi
    done
!   status "done loading data into log shipping node"
! 
!   NUMNODES=4
!   status "Changed number of nodes to 4 to reflect the log shipping node"
! 
  }



More information about the Slony1-commit mailing list