Hey there, > It's a bit weird.. > Lets try it from another angle one at a time.. > First user the basic 256MB MEMcache which is less the 512MB. > Remove any cache_dir that do exits.. >>> *i started before without cache dir , but i had alot of "vary object >>> loop" logs in cache.log i put cache dir so that i dont want to see any suspicious logs in cache.log * > This way we have pushed squid into the CPU and mem land. > Having low memory is nothing to be afraid of.. > The main issue is why would you get into a position that squid that can > pump lots of users traffic is not responding to you very well. > If you have an unresponsive server to even SSH the next thing is basic > PING tests. > ARPING > PING > TCPING > SSH_TCP_PING (telnet\nc port 22) > HTTPING *>>>again , the server in this status is very slow and with i could hardly get access to last logs of cache manager due to slow responce of my machine* > maximum FD will be about 1k(1024) in use which should no be too much of > a problem for any tiny proxy. * >>> i raised it because im planning to put alot of users when caryy my work >>> to delr720 machine * > And now again, what OS are you using? why do you need 500k FD allowed > for a system that should use about 16k FD just for a test? 65k FS should > be enough for most basic small ISP setup. *>>i will try it , im using centos 6.4 with kernel rebuild to support tproxy,* > So only one instance of squid no workers at all no ROCK storage adding > users should show you how the service works.. > (I can calm you down that I took up and down a server that serves more > then 1k users live on it.) *>>>well i will make another test* > Indeed named can load on the server but since it's serving only squid I > assume you do have another local DNS server in place so point to it. > Using 8.8.8.8 public DNS will not solve your problems but rather make > them worse unless you do have 1-30ms response time from it. *>>>i know that , but just made a test to get problem of dns out and let it far from my current issue !!* You can also do DNS cache warm-up.. > Again what is the degradation you are talking about? > Try to download a static file from some couple random mirrors of let say > some linux distro or a mirror of another file you have in mind like > microsoft.. * >the degradation is as below : when start squid , and watching youtube , browse sites , no delay and youtube is excellent after some time , browsing is very slow ,youtube is interrupting and no buffering videos !!!* > I have looked at the graph and I do not understand what is the problem > while there is a degradation? *sorry, i ve modified the post above it was problem in display , here is the graph i re-uploaded it http://www2.0zz0.com/2013/11/07/21/928332022.png as u see , the traffic should be about 55-60 M but after sometime the traffifc become 30 M , which mean that there is slow and degredation occured !!* > Just a tiny test I want to add for all the above: > start any local http server that you like and prefer like apache, nginx, > lighthttpd, GoAhead-Webserver, micro_httpd or any other and put it on > the "lan" or "network segment" with the squid server and try to download > files of 512bytes 1KB 1MB 9MB 15MB 100MB and up. > also try to access the cache-mgr page using: > http://MY_CACHE_IP:zzzzzz/squid-internal-mgr/mem >> with 1 user , no problem and squid is very nice , at the same time there is a problem and after\before. > The graph that you supply doesn't really explains the problem in any of > the network measurements at least I can understand. *>look here http://www2.0zz0.com/2013/11/07/21/928332022.png* Try to take a small look at: http://wiki.squid-cache.org/ConfigExamples/UbuntuTproxy4Wccp2 > Which uses the tunnel method rather then the mac address rewritting > method 2 wccp forwarding. > (I am merely saying that I know it works very well for me) > What device does the WCCP interception?? it is cisco MLS 76xx There are bugs out-there in squid but I would like to see the bug.. I am not sure but what are the "free -m" status on this server while squid is not running at all? >>will make another test and tell u > I am until now try to think of a test that will show and explain the > problem in hands. > There is an issue that I remember about FD limit that is being forced by > the bash\startup script on squid master instance but I am not sure if > this is the source to the issue or it's another thing. > please add to squid init.d\startup script "ulimit -a" and the output of > that.. [root@squid ~]# ulimit -a core file size (blocks, -c) 0 data seg size (kbytes, -d) unlimited scheduling priority (-e) 0 file size (blocks, -f) unlimited pending signals (-i) 63772 max locked memory (kbytes, -l) 64 max memory size (kbytes, -m) unlimited open files (-n) 131072 pipe size (512 bytes, -p) 8 POSIX message queues (bytes, -q) 819200 real-time priority (-r) 0 stack size (kbytes, -s) 8192 cpu time (seconds, -t) unlimited max user processes (-u) 63772 virtual memory (kbytes, -v) unlimited file locks (-x) unlimited [root@squid ~]# ulimit -n 131072 [root@squid ~]# Are there any clues in the cache.log? at first , no logs , no errors , after sometime i have ""closing of due to life timeout in youtube"" videos Best Regards, Eliezer ----- Dr.x -- View this message in context: http://squid-web-proxy-cache.1019090.n4.nabble.com/is-there-any-thing-wrong-from-cache-manager-logs-tp4663156p4663172.html Sent from the Squid - Users mailing list archive at Nabble.com.