Re: Best COPY Performance

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Oct 24, 2006 at 10:36:04PM -0700, Craig A. James wrote:
> Jim C. Nasby wrote:
> >Well, given that perl is using an entire CPU, it sounds like you should
> >start looking either at ways to remove some of the overhead from perl,
> >or to split that perl into multiple processes.
> 
> I use Perl for big database copies (usually with some 
> processing/transformation along the way) and I've never seen 100% CPU usage 
> except for brief periods, even when copying BLOBS and such.  My typical 
> copy divides operations into blocks, for example doing
> 
>  N = 0
>  while (more rows to go) {
>     begin transaction
>     select ... where primary_key > N order by primary_key limit 1000
>     while (fetch a row)
>        insert into ...
>     N = (highest value found in last block)
>     commit
>   }
> 
> Doing it like this in Perl should keep Postgres busy, with Perl using only 
> moderate resources.  If you're seeing high Perl CPU usage, I'd look first 
> at the Perl code.

Wait... so you're using perl to copy data between two tables? And using
a cursor to boot? I can't think of any way that could be more
inefficient...

What's wrong with a plain old INSERT INTO ... SELECT? Or if you really
need to break it into multiple transaction blocks, at least don't
shuffle the data from the database into perl and then back into the
database; do an INSERT INTO ... SELECT with that same where clause.
-- 
Jim Nasby                                            jim@xxxxxxxxx
EnterpriseDB      http://enterprisedb.com      512.569.9461 (cell)


[Postgresql General]     [Postgresql PHP]     [PHP Users]     [PHP Home]     [PHP on Windows]     [Kernel Newbies]     [PHP Classes]     [PHP Books]     [PHP Databases]     [Yosemite]

  Powered by Linux