On 23.12.09 12:47, Manjusha Maddala wrote: > I'm totally puzzled as to how the LRU cache replacement policy has been > implemented in Squid. > > Here's my situation. [...] > $ curl -H "User-Agent: some useragent" -H > "Accept-Encoding: gzip,deflate" "http://some-url" > > Squid access.log - > > xxx.xxx.xxx.xxx - - [23/Dec/2009:11:03:31 -0800] "GET http://some-url > HTTP/1.1" 200 19114 "-" "-" "-" "gzip,deflate" > TCP_MISS/FIRST_UP_PARENT 446 > > $ ll 001000FF > ls: 001000FF: No such file or directory > ............................................................ > > > A page that is just 3 days old encountered a cache miss, even though the > cached page is physically present on the disk. Could somebody explain > why this has happened? The exact same URL is requested from the command > line and Squid has removed its copy of the page after the command > execution. If the copy has become stale, going by the refresh_pattern > rules, a TCP_REFRESH_MISS should have got logged instead of TCP_MISS. > Its really weird that the copy on the disk was deleted, and a TCP_MISS > was reported. did you check if the page is cacheable? Did you try fetching it with "Cache-Control: only-if-cached" header? It's quite possible that the page is not cacheable, so squid won't cache it, or the cache has expired... -- Matus UHLAR - fantomas, uhlar@xxxxxxxxxxx ; http://www.fantomas.sk/ Warning: I wish NOT to receive e-mail advertising to this address. Varovanie: na tuto adresu chcem NEDOSTAVAT akukolvek reklamnu postu. "They say when you play that M$ CD backward you can hear satanic messages." "That's nothing. If you play it forward it will install Windows."