On 4/25/05, Abu Hurayrah <abu_hurayrah@xxxxxxxxxxxxx> wrote: > Actually, I think I discovered my own solution. It turns out by hijacking > Apache's normal role of serving files and delegating it to my own PHP > script, I've caused a bit of a confusion, and Apache still attempts to > answer requests for partial content, while my script is only adept at > deliver the whole thing from start to finish. > > What ends up happening, I think, is that Apache will load the entire file > into memory, and then take care of the partial-content requests - I don't > know if it is successfully, because these requests are often repeated, much > to the dismay of my memory controller! This would explain why the processes > were maxing out at 140+MB - that is, coincidentally, around the same size as > the largest files I'm serving up. > > What I have to do is rewrite my entire routine to handle the HTTP requests > themselves, which shouldn't be amazingly hard - but it does through a wrench > in the works. I'll try and update this thread if I finish my solution. That is a possible explanation. There was some discussion of this issue on the dev list a while back, but I'm not sure what ever happend to it. That is certainly not good behavior on apache's part. One other possibility for you is to try to use some of the BrowserMatch variables (downgrade-1.0, force-response-1.0) to get apache to stop doing byte-range responses. Joshua. --------------------------------------------------------------------- The official User-To-User support forum of the Apache HTTP Server Project. See <URL:http://httpd.apache.org/userslist.html> for more info. To unsubscribe, e-mail: users-unsubscribe@xxxxxxxxxxxxxxxx " from the digest: users-digest-unsubscribe@xxxxxxxxxxxxxxxx For additional commands, e-mail: users-help@xxxxxxxxxxxxxxxx