> Paul Scott wrote: > > I have a very large text file that gets dumped into a directoory every > > now and then. It is typically around 750MB long, at least, and my > > question is: > > > > What is the best method to parse this thing and insert the data into a > > postgres db? > > > > I have tried using file(), fget*() and some others, all with limited > > success. It goes through OK (I am sending it to a background process on > > the server and using a callback function to email me when done) but it > > is really knocking the machine hard, besides taking a real long time to > > finish. > > > > Is there a better way of approaching this? Any help would be greatly > > appreciated. > > First, which is your approach? I suspect that you are doing this with a > cron job through php-cli. > > Now, to avoid using to many resources, try with fopen() and fgets(). > Also work with persistent connections, so you don't have that overhead. > > The problem with file() is that it will load all the file to memory, and > you don't want 700+MB in memory. In addition to Martin's good suggestions (and also assuming you're running php-cli via cron), you could use nice to stop it consuming too many resources: http://en.wikipedia.org/wiki/Nice_%28Unix%29 Edward -- PHP General Mailing List (http://www.php.net/) To unsubscribe, visit: http://www.php.net/unsub.php