Thanks guys- I think I'll have to do multiple queries using LIMIT as Geoff suggested since apparently mysql_unbuffered_query() would lose the result set of the "select * from" query once I run the insert query. I'm still not sure why the MYSQL_CLIENT_COMPRESS didn't seem to have an effect, however. David -----Original Message----- From: Chris [mailto:dmagick@xxxxxxxxx] Sent: Monday, February 06, 2006 4:16 PM To: David Yee Cc: 'php-general@xxxxxxxxxxxxx' Subject: Re: PHP mysql data result set compression Hi David, See http://www.php.net/mysql_unbuffered_query It won't load the whole lot into memory before returning it to php. David Yee wrote: > Hi all- is there a way have a large data result set from MySQL compressed? > E.g. I have a table with over a million rows of data that I want to do a > "select * from " on and then take that result, do some field/data > manpulation, and then insert row-by-row to another table. The problem is > the result of the query is so big that it's casuing PHP to swap to disk, > causing things to slow to a crawl. Doing a "show processlist" on the mysql > console shows that "Writing to net" is the state of the running "select * > from " query. I tried adding the flag "MYSQL_CLIENT_COMPRESS" to both > mysql_pconnect() and mysql_connect() but it doesn't seem to do any > compression (I can tell by the size of the running php memory process). Any > ideas would be appreciated- thanks. > > David > -- PHP General Mailing List (http://www.php.net/) To unsubscribe, visit: http://www.php.net/unsub.php