Search squid archive

what are the Pros and cons filtering urls using squid.conf?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I have seen http://www.squidblacklist.org/ which is a very nice idea but I am wondering if squid.conf and other squid products are the good choice for any place.

For a mission critical proxy server you will need to prevent any "reload" of the proxy which can cause a *small* download corruption. I know that admins most of the time don't think about reload of a process but it can cause some department in the enterprise a problem.
From the ISP point of view you must prevent any problem to the client.

The current solutions for filtering are:helper,squid.conf,icap\ecap service.
A helper is a nice solution like squidguard but this also needs a reload of the whole squid service. On a small scale system it's easy to do but for a 24x7 filtering solution we can't just reload the squid process or recreate the DB.

There for I wrote my ICAP service for this specific issue.
The need for a more persistent solution which doesn't require down time from the DB proxy or the filtering service point of view.

Would you prefer a filtering based on a reload or a persistent DB like mongoDB or tokyo tyrant?

Since there were some bug fixes in squid ICAP based solutions can give more throughput and I got into a 8k persistent request on a INTEL ATOM based filtering system.

Regards,
Eliezer




[Index of Archives]     [Linux Audio Users]     [Samba]     [Big List of Linux Books]     [Linux USB]     [Yosemite News]

  Powered by Linux