Search squid archive

Squid 3.4.6 is not caching anything

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I have tried deleting the cache and setting its size to 10GB. I ran squid -z
again and it created the directories before squid -z froze. The maximum
object size is set to 5GB, and I have checked some sites using redbot.org to
see if the can be cached or not. It says that they can, and I have had the
squid proxy running for about 48 hrs now with about 50 clients connected. I
have scanned the access.log and there is not a single hit. Even if the same
page is requested many times.

Are there some settings that I am missing in squid.conf that is stopping the
cache from working? Do you know where I can obtain a already compiled x86
package for Debian 7 with --enable-ssl and --enable-ssl-crtd?

Thanks for your help so far.

2014-06-30 00:46: Eliezer Croitoru sade:
> Well it's a bit hard to say that squid is not caching anything by only 
> this.
> 
> It depends on what you expect allow and config it to cache.
> Since squid uses a Store-ID to identify content you may think that two 
> request for the same "file" so to speak will be the same but in many 
> cases it is not due to couple additions in the url by the user or the 
> client.
> 
> There is also the possibility that you haven't configured squid to 
> cache big size files.
> Start with making the cache_dir in a size of more then 100MB(unless 
> it's good for you) also try to take a look\change:
> http://www.squid-cache.org/Doc/config/maximum_object_size/
> 
> and other object_size affecting config directives that you can see
> here:
> http://www.squid-cache.org/Doc/config/
> 
> Note that it is possible that you yet do not understand the complexity 
> of caching and I can redirect you if you want to couple articles or 
> config directives that can help you understand more about it.
> 
> I have seen a nice site you can try to see that is cached well:
> http://www.djmaza.info/
> 
> I do not know that much examples of a nice and cachable websites these 
> days due to the basic nature of dynamic content and SLL encrypted 
> sites out there in the wide web.
> 
> If you have specific site you want to analyze you can try redbot at:
> https://redbot.org/
> 
> You can insert the url and the interface will verify if the site\url 
> is cachable.
> 
> Eliezer
> 
> On 06/29/2014 09:54 AM, liam@xxxxxx wrote:
>> I have removed all comments from my config file:
>> http://pastebin.com/kqvNszyp
>> 
>> And here is a short excerpt from my access.log - don't think it will 
>> be too helpful though. I have removed some IP addresses and URLs.
>> 
>> http://pastebin.com/bZiZ3tUN
>> 
>> Note that I do get some TCP_MEM_HIT/200 sometimes.
>> 
>> I have tried using Firefox 29, Internet Explorer 11 and the latest 
>> version of Chrome for Debian 7 stable.





[Index of Archives]     [Linux Audio Users]     [Samba]     [Big List of Linux Books]     [Linux USB]     [Yosemite News]

  Powered by Linux