Hi Pablo Actually, it is my case. The URL-hash is helpful to reduce the duplicated objects. However, once adding/removing squid server, load balancer needs to re-calculate the hash of URL which cause lot of TCP_MISS in squid server at the inital stage. Do you have same experience ? Thanks --- Pablo Garc燰 <malevo@xxxxxxxxx> 說: > I dealt with the same problem using a load balancer > in front of the > cache farm, using a URL-HASH algorithm to send the > same url to the > same cache every time. It works great, and also > increases the hit > ratio a lot. > > Regards, Pablo > > 2008/3/6 Siu Kin LAM <sklam2005@xxxxxxxxxxxx>: > > Dear all > > > > At this moment, I have several squid servers for > http > > caching. Many duplicated objects have been found > in > > different servers. I would minimize to data > storage > > by installing a large centralized storage and the > > squid servers mount to the storage as data disk. > > > > Have anyone tried this before? > > > > thanks a lot > > > > > > Yahoo! 網上安全攻略,教你如何防範黑客! > 請前往http://hk.promo.yahoo.com/security/index.html > 了解更多。 > > > Yahoo! 網上安全攻略,教你如何防範黑客! 請前往http://hk.promo.yahoo.com/security/index.html 了解更多。