Search squid archive

RE: offline mode issue

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Thanks for the header_access suggestion, but unfortunately it is not
giving the desired results (yet).  I know this is a bongo request, but
my customer wants to have as much as possible (even stuff that shouldn't
be) cached.  Also, thanks for the pointer to that ircache web site.  I
wonder if some other element of the page could be "ignored" to be able
to cache this site (and others) while offline?  It gives this
information regarding google home page:
This object will be considered stale, because it doesn't have any
freshness information assigned. It doesn't have a validator present. It
won't be cached by proxies, because it has a Cache-Control: private
header. This object requests that a Cookie be set; this makes it and
other pages affected automatically stale; clients must check them upon
every request. It doesn't have a Content-Length header present, so it
can't be used in a HTTP/1.0 persistent connection.

I have also tried using a wwwoffle server inline (as a parent proxy) but
I'm getting REALLY unstable results that way. I'm thinking of looking
into some type of mirroring for a few key sites, but not sure how to
redirect requests from squid to the mirror contents. Any other ideas are
very appreciated!


-----Original Message-----
From: Chris Robertson [mailto:crobertson@xxxxxxx] 
Sent: Monday, October 02, 2006 12:51 PM
To: Squid Users
Subject: Re:  offline mode issue

Shaun Skillin (home) wrote:
> I'm happy to provide whatever documentation I can, but I'd like to
make
> sure it's working first and this outstanding issue still has me
stumped.
> Here's a quick snip from the access log, going to a few sites. Notice
> that all of them are misses.  Google is going to 3 different servers,
so
> I could maybe see a miss the first time it visits each site, but these
> are all misses every time. So, in the present setup, I cannot support
an
> offline_mode at all.  Any ideas on this?
>
> 1159796708.843    555 127.0.0.1 TCP_MISS/200 31444 GET
> http://www.microsoft.com - DIRECT/207.46.225.60 text/html
>   
http://www.ircache.net/cgi-bin/cacheability.py?query=http%3A%2F%2Fwww.mi
crosoft.com%2F
> 1159796710.853    129 127.0.0.1 TCP_MISS/200 31444 GET
> http://www.microsoft.com - DIRECT/207.46.225.60 text/html
> 1159796712.730     84 127.0.0.1 TCP_MISS/200 31444 GET
> http://www.microsoft.com - DIRECT/207.46.225.60 text/html
> 1159796718.040    175 127.0.0.1 TCP_MISS/200 3984 GET
> http://www.google.com - DIRECT/66.102.7.104 text/html
>   
http://www.ircache.net/cgi-bin/cacheability.py?query=http%3A%2F%2Fwww.go
ogle.com%2F
> 1159796719.611     88 127.0.0.1 TCP_MISS/200 3984 GET
> http://www.google.com - DIRECT/66.102.7.147 text/html
> 1159796720.844     91 127.0.0.1 TCP_MISS/200 3984 GET
> http://www.google.com - DIRECT/66.102.7.99 text/html
> 1159796851.780     90 127.0.0.1 TCP_MISS/200 3984 GET
> http://www.google.com - DIRECT/66.102.7.104 text/html
> 1159796853.392     84 127.0.0.1 TCP_MISS/200 3984 GET
> http://www.google.com - DIRECT/66.102.7.147 text/html
> 1159796854.326     87 127.0.0.1 TCP_MISS/200 3984 GET
> http://www.google.com - DIRECT/66.102.7.99 text/html
>
>
>   
In both cases the page is set as "private" and not cache-able.  You 
could try using  "header_access Cache_Control deny" with ACLs that 
stipulate the sites you wish to cache, but that may have adverse effects

(actual private data, such as emails and custom search pages might be 
cached).

Chris


[Index of Archives]     [Linux Audio Users]     [Samba]     [Big List of Linux Books]     [Linux USB]     [Yosemite News]

  Powered by Linux