Search squid archive

Re: throughput limitation from cache

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Peter Smith wrote:

Richard Mittendorfer wrote:

Hello *,

When downloading a cached file from the local squid, I just get about
250 - 280kB/s. Even on localhost. Is this a limitation with diskd
serving files from cache or some intern limit? I also tried aufs, but
didn't get a better rate. I found a thread here about this, but it got
more into a diskd/aufs discussion :) and didn't provide a solution or
explains it. It doesn't look like it's disk/system related, vmstat/top doesn't show
high load.

linux 2.6 celeron 600MHz
squid-2.5.STABLE12 Debian GNU/Linux
2 x cache_dir diskd 1G 8 128 on 2 x SCSI reiserfs
32M cache_mem, acls, .. need more info?

TIA, ritch

I too have experience this. I believe this is because Squid has built-in safeguards to keep users from flooding the disk and to even the load across all disks for all users. Another reason is Squid handles all requests in sequential order--client requests, server requests, and memory/disk cache reads/writes.. It is like a virtualized network handler or VM. It really doesn't want one client to kill all others..

HTH,
Peter


Btw, it still does have a lot to do with diskd/aufs... Read here http://squid-docs.sourceforge.net/latest/html/x220.html#AEN318

Peter

[Index of Archives]     [Linux Audio Users]     [Samba]     [Big List of Linux Books]     [Linux USB]     [Yosemite News]

  Powered by Linux