If you don't care to store those characters then maybe you should change the database character set to use SQL_ASCII instead of UTF-8. I believe ASCII will quietly discard those characters when converting from UTF-8. Mike On Wed, 2006-03-15 at 23:20 -0600, Wes wrote: > Based on a couple of other posts, I think I'm out of luck, but I'm hoping > something might have changed recently. > > I'm loading a very high volume of data with COPY using libpq - about 100+ > million rows per day. The problem is that the COPY sometimes aborts with > > invalid UTF-8 byte sequence detected > > I'd prefer not to add to the overhead by pre-validating every string, since > PostgreSQL validates it already. Is there a way to get the server to > replace invalid characters with something like blank instead of generating a > fatal error? > > Wes > > > > ---------------------------(end of broadcast)--------------------------- > TIP 1: if posting/reading through Usenet, please send an appropriate > subscribe-nomail command to majordomo@xxxxxxxxxxxxxx so that your > message can get through to the mailing list cleanly