I was able to get things to work in Firefox. Your tip about refresh always getting a MISS or REFRESH_HIT helped solve it. However, I am still seeing some different behavior in my curl. When I hit the page with Firefox I see that the page is being cached appropriately. However, when I hit the same page with curl, it does not seem to get the cached version. My curl is now returning the page correctly, however every time I connect it is not getting the cached version. I have confirmed that it is going through the proxy (my webserver will not accept traffic that isn't through the proxy), so I am a little confused. Thanks! -hortitude On Tue, Mar 10, 2009 at 9:04 PM, Amos Jeffries <squid3@xxxxxxxxxxxxx> wrote: > Hortitude Eyeball wrote: >> >> I am trying to setup Squid to be a simple proxy-cache. >> I am seeing two strange behaviors. >> I have 3 machines. I am using one as my web browser, one as my >> proxy-cache and the third as my web server. >> When I configure my web browser (Firefox) to connect through my >> proxy-cache to my web server I see content as expected, however it >> does not seem to be cached. The web page I am using is at the bottom >> of this post. When I view the web page I keep seeing the time change, >> so I know that it is not being cached. > > NP: pressing the refresh button n a web browser sends a special header to > refresh the page, at minimum forcing squid to check for an updated version. > Your web server as ircache reports sends a new object when asked about > modification. This will result in REFRESH_HIT or MISS. > >> Furthermore, when I use curl to through the proxy and look at the >> headers using the -D option I see a 400 error from the proxy server >> and then a 200 from the web server? I also see a message from the >> squid server of "Invalid Request" > > From that description I'd guess you have a domain name which resolves to > your Squid box AND the web server? > > 400 from squid is probably a forwarding loop? > > >> >> When I run my web page through >> http://www.ircache.net/cgi-bin/cacheability.py it says" >> >> This object will be fresh for 20 hr 22 min. It has a validator >> present, but when a conditional request was made with it, the same >> object was sent anyway. >> >> Can anyone help? >> Thanks! >> I am running SQUID 2.7.STABLE3 on Ubuntu. >> >> >> I have not changed the config much at all. I did a grep of all >> options that are set in the config file and have included them here: >> >> acl all src all >> acl manager proto cache_object >> acl localhost src 127.0.0.1/32 >> acl to_localhost dst 127.0.0.0/8 >> acl localnet src 10.0.0.0/8 # RFC1918 possible internal network >> acl localnet src 172.16.0.0/12 # RFC1918 possible internal network >> acl localnet src 192.168.0.0/16 # RFC1918 possible internal network >> acl SSL_ports port 443 # https >> acl SSL_ports port 563 # snews >> acl SSL_ports port 873 # rsync >> acl Safe_ports port 80 # http >> acl Safe_ports port 21 # ftp >> acl Safe_ports port 443 # https >> acl Safe_ports port 70 # gopher >> acl Safe_ports port 210 # wais >> acl Safe_ports port 1025-65535 # unregistered ports >> acl Safe_ports port 280 # http-mgmt >> acl Safe_ports port 488 # gss-http >> acl Safe_ports port 591 # filemaker >> acl Safe_ports port 777 # multiling http >> acl Safe_ports port 631 # cups >> acl Safe_ports port 873 # rsync >> acl Safe_ports port 901 # SWAT >> acl purge method PURGE >> acl CONNECT method CONNECT >> http_access allow manager localhost >> http_access deny manager >> http_access allow purge localhost >> http_access deny purge >> http_access deny !Safe_ports >> http_access deny CONNECT !SSL_ports >> http_access allow localhost >> http_access allow localnet >> http_access deny all >> icp_access allow localnet >> icp_access deny all >> http_port 3128 >> hierarchy_stoplist cgi-bin ? >> access_log /var/log/squid/access.log squid >> refresh_pattern ^ftp: 1440 20% 10080 >> refresh_pattern ^gopher: 1440 0% 1440 >> refresh_pattern -i (/cgi-bin/|\?) 0 0% 0 >> refresh_pattern . 0 20% 4320 >> acl apache rep_header Server ^Apache >> broken_vary_encoding allow apache >> extension_methods REPORT MERGE MKACTIVITY CHECKOUT >> hosts_file /etc/hosts >> coredump_dir /var/spool/squid >> >> >> ------------------------------------------------------------------------------------------- >> >> Here is the web page I am using >> >> <?php >> // the time we got hit and generated content >> $now = time(); >> $generatedAt = gmdate('D, d M Y H:i:s T', $now); >> >> // the last modified date (midnight on the same day of generation, as >> // per your business-rule) >> $lastModified = gmdate('D, d M Y 00:00:00 T', $now); >> >> // date of expiry (24 hours after the last modified date, as per your >> // business-rule) >> $expiresAt = gmdate('D, d M Y H:i:s T', strtotime($lastModified) + 86400); >> >> // the minimum required http headers to make Squid do what you asked is >> // Last-modified and Cache-control. We need to give Cache-control the >> // expiry time in terms of "age" (in seconds) so we calculate that below. >> // Optionally you could also provide the "Expires: $expiresAt" header to >> // tell the browser/client the same information, just in a different way. >> // This is not required for Squid though. >> $maxAge = strtotime($expiresAt) - strtotime($generatedAt); >> header('Last-modified: ' . $lastModified); >> header('Cache-control: max-age=' . $maxAge); >> header ('Expires: '.$expiresAt); >> >> // The rest is simply informational >> header('Content-type: text/plain'); >> echo "The content of this page was last modified at $lastModified\n"; >> echo "This page was generated at $generatedAt and will be cached by >> Squid for $maxAge seconds until $expiresAt\n"; >> ?> > > > Amos > -- > Please be using > Current Stable Squid 2.7.STABLE6 or 3.0.STABLE13 > Current Beta Squid 3.1.0.6 >