Search squid archive

Re: is there any thing wrong from cache manager logs ?!!

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hey there,

> It's a bit weird..
> Lets try it from another angle one at a time..
> First user the basic 256MB MEMcache  which is less the 512MB.
> Remove any cache_dir that do exits..

>>> *i started before without cache dir , but i had alot of "vary object
>>> loop" logs in cache.log
 i put cache dir so that i dont want to see any suspicious logs in cache.log
*


> This way we have pushed squid into the CPU and mem land.
> Having low memory is nothing to be afraid of..
> The main issue is why would you get into a position that squid that can 
> pump lots of users traffic is not responding to you very well.
> If you have an unresponsive server to even SSH the next thing is basic 
> PING tests.
> ARPING
> PING
> TCPING
> SSH_TCP_PING (telnet\nc port 22)
> HTTPING

*>>>again , the server in this status is very slow  and with i could hardly
get access to  last logs of cache manager due to slow responce of my
machine*




> maximum FD will be about 1k(1024) in use which should no be too much of 
> a problem for any tiny proxy.
*
>>> i raised it because im planning to put alot of users when caryy my work
>>> to delr720 machine
*

> And now again, what OS are you using? why do you need 500k FD allowed 
> for a system that should use about 16k FD just for a test? 65k FS should 
> be enough for most basic small ISP setup.

*>>i will try it , im using centos 6.4 with kernel rebuild to support
tproxy,*



> So only one instance of squid no workers at all no ROCK storage adding 
> users should show you how the service works..
> (I can calm you down that I took up and down a server that serves more 
> then 1k users live on it.)

*>>>well i will make another test*


> Indeed named can load on the server but since it's serving only squid I 
> assume you do have another local DNS server in place so point to it.
> Using 8.8.8.8 public DNS will not solve your problems but rather make 
> them worse unless you do have 1-30ms response time from it.

*>>>i know that , but  just made a test  to get problem of dns out and let
it far from my current issue !!*

You can also do DNS cache warm-up..


> Again what is the degradation you are talking about?
> Try to download a static file from some couple random mirrors of let say 
> some linux distro or a mirror of another file you have in mind like 
> microsoft..

*
>the degradation is as below :
when start squid , and watching youtube , browse sites , no delay and
youtube is excellent
after some time , browsing is very slow ,youtube is interrupting  and no
buffering videos !!!*


> I have looked at the graph and I do not understand what is the problem 
> while there is a degradation?

*sorry, i ve modified the post above it was problem in display , here is the
graph i re-uploaded  it
http://www2.0zz0.com/2013/11/07/21/928332022.png
as u see ,  the traffic should be about 55-60 M but after sometime the
traffifc become 30 M , which mean that there is slow and degredation occured
!!*


> Just a tiny test I want to add for all the above:
> start any local http server that you like and prefer like apache, nginx, 
> lighthttpd, GoAhead-Webserver, micro_httpd or any other and put it on 
> the "lan" or "network segment" with the squid server and try to download 
> files of 512bytes 1KB 1MB 9MB 15MB 100MB and up.
> also try to access the cache-mgr page using:
> http://MY_CACHE_IP:zzzzzz/squid-internal-mgr/mem

>> with 1 user  , no problem and squid is very nice ,

at the same time there is a problem and after\before.


> The graph that you supply doesn't really explains the problem in any of 
> the network measurements at least I can understand.

*>look here http://www2.0zz0.com/2013/11/07/21/928332022.png*



Try to take a small look at:
http://wiki.squid-cache.org/ConfigExamples/UbuntuTproxy4Wccp2

> Which uses the tunnel method rather then the mac address rewritting 
> method 2 wccp forwarding.
> (I am merely saying that I know it works very well for me)
> What device does the WCCP interception??

it is cisco MLS 76xx

There are bugs out-there in squid but I would like to see the bug..
I am not sure but what are the "free -m" status on this server while 
squid is not running at all?
>>will make another test and tell u


> I am until now try to think of a test that will show and explain the 
> problem in hands.
> There is an issue that I remember about FD limit that is being forced by 
> the bash\startup script on squid master instance but I am not sure if 
> this is the source to the issue or it's another thing.
> please add to squid init.d\startup script "ulimit -a" and the output of 
> that..

[root@squid ~]# ulimit -a
core file size          (blocks, -c) 0
data seg size           (kbytes, -d) unlimited
scheduling priority             (-e) 0
file size               (blocks, -f) unlimited
pending signals                 (-i) 63772
max locked memory       (kbytes, -l) 64
max memory size         (kbytes, -m) unlimited
open files                      (-n) 131072
pipe size            (512 bytes, -p) 8
POSIX message queues     (bytes, -q) 819200
real-time priority              (-r) 0
stack size              (kbytes, -s) 8192
cpu time               (seconds, -t) unlimited
max user processes              (-u) 63772
virtual memory          (kbytes, -v) unlimited
file locks                      (-x) unlimited
[root@squid ~]# ulimit -n
131072
[root@squid ~]# 

Are there any clues in the cache.log?
at first , no logs , no errors , after sometime i have ""closing of due to
life timeout in youtube"" videos 

Best Regards,
Eliezer





-----
Dr.x
--
View this message in context: http://squid-web-proxy-cache.1019090.n4.nabble.com/is-there-any-thing-wrong-from-cache-manager-logs-tp4663156p4663172.html
Sent from the Squid - Users mailing list archive at Nabble.com.




[Index of Archives]     [Linux Audio Users]     [Samba]     [Big List of Linux Books]     [Linux USB]     [Yosemite News]

  Powered by Linux