Search squid archive

Re: Massive Squid Deployment

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



One performance-specific 2.7 question - I recall hearing mention of an issue with 2.6 where larger objects held in mem_cache required exponentially more CPU cycles to serve up (i.e. n cycles for a 4KB object, n*2 for an 8KB object, n*4 for a 12KB object, etc). Does anyone know if this issue is still present in 2.7 code?

Thanks,

-C

On Apr 1, 2009, at 12:42 AM, Amos Jeffries wrote:


Hello people, I'm having some bottlenecks on our squid deployement and I
was
wondering if anyone had any recommendations because I'm near out of ideas. Would someone change my architecture or does anyone have any experience
with
a squid deployment of this size? Basically we are pushing 600mb at
240,000k
connections. When we reach speeds on our around that that number, we start seeing slow performance and getting alot of page timeouts. We are running
20
squid 2.6 boxes running dsr behind a single foundry gte load balancer. I recently had 18 squid boxes and thought we had a squid bottleneck but no change. I was kinda leaning in the direction of splitting it in half and doing multiple load balancers. Does anyone have any experience pushing
this
much traffic?

I'd go to 2.7 if it was squid being slow. It has a number of small
performance boosters missing in 2.6.

But, it does sound like a load balancer bottleneck if you had zero change from adding two more Squid. The Yahoo and Wikimedia guys are the poster installs for these types of Squid deployment and they use CARP meshing.

PS: We are very interested in getting some real-world benchmarks from
high-load systems like yours. Can you grab the data needed for
http://wiki.squid-cache.org/KnowledgeBase/Benchmarks ?

Amos





[Index of Archives]     [Linux Audio Users]     [Samba]     [Big List of Linux Books]     [Linux USB]     [Yosemite News]

  Powered by Linux