Search squid archive

Re: Squid Blocking non-listed websites

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Go Wow wrote:
I have attached 3 files named squid.conf sites.txt (whitelist) and
blacklisted_sites.txt (Blacklist)

okay i will paste my squid.conf here to will be helpful for others

auth_param basic children 5
auth_param basic realm Squid proxy-caching web server
auth_param basic credentialsttl 2 hours
auth_param basic casesensitive off

It looks like auth. But is never activated or used...

acl all src 0.0.0.0/0.0.0.0
acl manager proto cache_object
acl localhost src 127.0.0.1/255.255.255.255
acl to_localhost dst 127.0.0.0/8
acl SSL_ports port 443
acl Safe_ports port 80		# http
acl Safe_ports port 21		# ftp
acl Safe_ports port 443		# https
acl Safe_ports port 70		# gopher
acl Safe_ports port 210		# wais
acl Safe_ports port 1025-65535	# unregistered ports
acl Safe_ports port 280		# http-mgmt
acl Safe_ports port 488		# gss-http
acl Safe_ports port 591		# filemaker
acl Safe_ports port 777		# multiling http
acl CONNECT method CONNECT
acl BadSites url_regex -i "/etc/blacklisted_sites.txt"
acl GoodSites url_regex -i "/etc/sites.txt"
acl lpo_BadSites url_regex -i "/etc/lpo_blacklisted_sites.txt"

Your first thing are these regexes.

If they match ANYWHERE in the URI the site will either be allowed/denied. You have some very short matches in there that could be doing strange things to sites passing query parameters in the URI.

Secondly regex is slow and will never scale to anything large enough to do a content-filter job properly. A list of dstdomain would be better in some of those cases. It matches just the requested domain name against the ACL. Some of your matches are for domains.

Thirdly, their positioning at the top of the access list means that the allow GoodSites has your proxy open to abuse for anyone who can/wants to fiddle the URIs or use your permitted sites. ie check their gmail through you.

Consider; does it really matter that people can get to porn through you?
If you are trying to make a child protection filter there is much better software out there to do it with than squid.


acl home_network src 192.168.10.10-192.168.10.120
acl lpo_network src 192.168.10.60-192.168.10.72

Right at the top you are missing a:
    http_access deny !home_network

that will stop most people getting to your proxy to abuse it.
The rest should be about stopping internal abuses.

http_access deny lpo_BadSites lpo_network
http_access deny lpo_network

Okay, so the lpo part of the network is not allowed to use the proxy.
In that case checking BadSites here specially is a wast of time. They never get through to any of those anyway.

http_access allow GoodSites

Now anyone can get in if they are visiting a site you like (ie gmail) or want to fiddle the URI to contains certain easily detected things.

http_access deny BadSites

And the effects of a short-and-anywhere-without-case uri match makes itself felt...

Turning BadSites into a list of bad domains and a dstdomsin ACL would do a lot here.

http_access allow home_network

This should be the second-to-LAST thing you do.

http_access allow manager localhost
http_access deny manager
http_access deny !Safe_ports
http_access deny CONNECT !SSL_ports

These above denies are really useless unless they are done before the permits that override them.

The next line will be doing a global deny anyway if they get this far.

http_access deny all

http_reply_access allow all
icp_access allow all
http_port 192.168.10.1:3128 transparent
hierarchy_stoplist cgi-bin ?
acl QUERY urlpath_regex cgi-bin \?
cache deny QUERY
cache_mem 16 MB
access_log /usr/local/squid/var/logs/access.log squid
cache_log /usr/local/squid/var/logs/cache.log
cache_store_log /usr/local/squid/var/logs/store.log

This log is not that useful unless you are debugging storage.
You can gain a bit by setting:
  cache_store_log none

mime_table /usr/local/squid/etc/mime.conf
pid_filename /usr/local/squid/var/logs/squid.pid
refresh_pattern ^ftp:		1440	20%	10080
refresh_pattern ^gopher:	1440	0%	1440
refresh_pattern .		0	20%	4320
acl apache rep_header Server ^Apache
broken_vary_encoding allow apache
cache_effective_user squid
cache_effective_user squid

twice?

icon_directory /usr/local/squid/share/icons
error_directory /usr/local/squid/share/errors/English
hosts_file /etc/hosts
dns_testnames google.com
coredump_dir /usr/local/squid/var/cache

If you see my whitelist (sites.txt) I need to add words like
"examination" to access them although examination word is not added to
my blacklist. Let me know the problematic thing plz.

I'd say the sites are matching something in BadSites list. Probably 'ass' or 'tit'. (nee Examination ASSignment? embASSy? clASS examination? examination TITle?). The bad list has some very short words that are commonly part of other innocent words.
This is the universal problem with content filters.

You would probably have a better rate finding a list of porn websites and blocking those by IP address (dst) or domain name (dstdomain) than a short regex.

Amos
--
Please use Squid 2.6STABLE17+ or 3.0STABLE1+
There are serious security advisories out on all earlier releases.

[Index of Archives]     [Linux Audio Users]     [Samba]     [Big List of Linux Books]     [Linux USB]     [Yosemite News]

  Powered by Linux