Hi, The difference to my test is that we use the ODBC interface in our C program. Could it be that the difference in the runtimes is caused by the ODBC? BR Ingo -----Original Message----- From: ext Merlin Moncure [mailto:mmoncure@xxxxxxxxx] Sent: Thursday, October 07, 2010 7:17 PM To: Sander, Ingo (NSN - DE/Munich) Cc: ext Craig Ringer; pgsql-performance@xxxxxxxxxxxxxx Subject: Re: Runtime dependency from size of a bytea field On Thu, Oct 7, 2010 at 10:49 AM, Merlin Moncure <mmoncure@xxxxxxxxx> wrote: > On Thu, Oct 7, 2010 at 12:11 AM, Sander, Ingo (NSN - DE/Munich) > <ingo.sander@xxxxxxx> wrote: >> As written before I have rerun the test a) without compression and b) >> with enlarged BLOCK_SIZE. Result was the same. > > Using libpqtypes (source code follows after sig), stock postgres, > stock table, I was not able to confirm your results. 4000 bytea > blocks, loops of 1000 I was able to send in about 600ms. 50000 byte > blocks I was able to send in around 2 seconds on workstation class > hardware -- maybe something else is going on?. I re-ran the test, initializing the bytea data to random values (i wondered if uninitialized data getting awesome compression was skewing the results). This slowed down 50000 bytea case to around 3.5-4 seconds. That's 12-15mb/sec from single thread which is IMNSHO not too shabby. If your data compresses decently and you hack a good bang/buck compression alg into the backend like lzo you can easily double that number. merlin -- Sent via pgsql-performance mailing list (pgsql-performance@xxxxxxxxxxxxxx) To make changes to your subscription: http://www.postgresql.org/mailpref/pgsql-performance