I have a Postgis table with about 2 million polyline records. The most number of points I have in the geometry field is about 500. I have a simple DBD::Pg Perl program that does a select for most of these records and do some processing with them before writing them to a file. Unfortunately, I seem to keep getting this error: DBD::Pg::st execute failed: out of memory for query result DBD::Pg::st fetchrow_array failed: no statement executing This program works fine with less than a million records for sure, my next largest table. I believe the program is failing at the first execute after prepare on this table. Any ideas on how I can do this? Thanks, Tim ---------------------------(end of broadcast)--------------------------- TIP 9: the planner will ignore your desire to choose an index scan if your joining column's datatypes do not match