Tue Jul 11 09:06:12 PDT 2006
- Previous message: [Slony1-general] Slony with PostgreSQL mainline (fwd)
- Next message: [Slony1-general] 1.2 RC1 RSN
- Messages sorted by: [ date ] [ thread ] [ subject ] [ author ]
We've got a test that is intended to seek out problems in handling of UTF8 / multibyte values. Some may recall that there was a condition in 1.1.0 where it was possible to cut off the final byte of such a string... The "testutf8" test specifically goes after that condition, using the queries listed below. echo "INSERT INTO utf8table (string) values ('${txtb} - \303\241');" >> $GENDATA echo "INSERT INTO utf8table (string) values ('${txtb} -- \303\241');" >> $GENDATA echo "INSERT INTO utf8table (string) values ('\303\241 - ${txtb}');" >> $GENDATA echo "INSERT INTO utf8table (string) values ('\303\241 -- ${txtb}');" >> $GENDATA echo "INSERT INTO utf8table (string) values ('t3 -- \303\241 - ${txtb}');" >> $GENDATA echo "INSERT INTO utf8table (string) values ('t3 - \303\241 -- ${txtb}');" >> $GENDATA While that probably is quite satisfactory for making sure that we do not regress back into that particular failure, I expect that we might get something out of adding some additional UTF8 characters and doing more with them. I daresay I have no personal call to use UTF8 at this time, and a search through PG docs aren't finding me much in terms of how to comprehensively create full sets of UTF8 characters. I'd like, if I could, to create some sample data that would comprehensively create each UTF8 character that exists, and generate some patterns thereof. Does anyone have a link to a web page or relevant resource ("read ISO standard Foo" would not be welcome ;-)) that might help educate me a little on this?
- Previous message: [Slony1-general] Slony with PostgreSQL mainline (fwd)
- Next message: [Slony1-general] 1.2 RC1 RSN
- Messages sorted by: [ date ] [ thread ] [ subject ] [ author ]
More information about the Slony1-general mailing list