Search squid archive

Re: latency issues squid2.7 WCCP

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Don't set the cache_mem value to 30gb (you might've done that); I meant
to try raising the cache_mem value in increments until the Squid process
size reaches about 30gb (use 'top' to find it's virtual size).  Also
don't use 2 partitions for each drive (seems like you were saying that);
use a single one for each drive, and make sure the filesystem usage
stays below about 85%.

Ryan Goddard wrote:
> have been testing the cache running strictly off memory and no disk fs.
> When Squid is set to use 16gb memory, connections are very fast.
> When I bumped the memory use to anything near 30 gb, latency becomes an
> issue.
> The cache doesn't report any errors in cache.log and watching the
> access.log I don't see anything
> out of the ordinary.
> I never have seen the cachemgr report any "Store Disk file open" values.
> Again, this is a dedicated squid server with 32gb ram and 4 dedicated
> 280+gb 15k sata drives
> running transparently using wccp.
> Presently the 4 drives are partitioned in 1/2, allocated to squid0-squid7.
> Any recommendations on repartitioning the 4 drives for quicker reads?
> 
> here's a general runtime report fwiw:
> 
> Squid Object Cache: Version 2.7.STABLE4-20080926
> 
> Start Time:    Tue, 07 Oct 2008 18:21:10 GMT
> Current Time:    Wed, 15 Oct 2008 15:51:20 GMT
> 
> Connection information for squid:
>     Number of clients accessing cache:    5393
>     Number of HTTP requests received:    220293340
>     Number of ICP messages received:    0
>     Number of ICP messages sent:    0
>     Number of queued ICP replies:    0
>     Request failure ratio:     0.00
>     Average HTTP requests per minute since start:    19374.7
>     Average ICP messages per minute since start:    0.0
>     Select loop called: -454862289 times, -1.500 ms avg
> Cache information for squid:
>     Request Hit Ratios:    5min: 32.8%, 60min: 32.3%
>     Byte Hit Ratios:    5min: 14.1%, 60min: 15.9%
>     Request Memory Hit Ratios:    5min: 56.9%, 60min: 56.5%
>     Request Disk Hit Ratios:    5min: 0.1%, 60min: 0.1%
>     Storage Swap size:    0 KB
>     Storage Mem size:    16777692 KB
>     Mean Object Size:    0.00 KB
>     Requests given to unlinkd:    0
> Median Service Times (seconds)  5 min    60 min:
>     HTTP Requests (All):   0.06640  0.06286
>     Cache Misses:          0.10281  0.09736
>     Cache Hits:            0.00000  0.00000
>     Near Hits:             0.05046  0.05046
>     Not-Modified Replies:  0.00000  0.00000
>     DNS Lookups:           0.04854  0.02941
>     ICP Queries:           0.00000  0.00000
> Resource usage for squid:
>     UP Time:    682209.558 seconds
>     CPU Time:    133086.580 seconds
>     CPU Usage:    19.51%
>     CPU Usage, 5 minute avg:    15.85%
>     CPU Usage, 60 minute avg:    15.31%
>     Process Data Segment Size via sbrk(): 18876028 KB
>     Maximum Resident Size: 0 KB
>     Page faults with physical i/o: 4381750
> Memory usage for squid via mallinfo():
>     Total space in arena:  -2095492 KB
>     Ordinary blocks:       2096010 KB  22166 blks
>     Small blocks:               0 KB      0 blks
>     Holding blocks:         67624 KB     12 blks
>     Free Small blocks:          0 KB
>     Free Ordinary blocks:    2801 KB
>     Total in use:          -2030670 KB 100%
>     Total free:              2801 KB 0%
>     Total size:            -2027868 KB
> Memory accounted for:
>     Total accounted:       18253692 KB
>     memPoolAlloc calls: 4284254700
>     memPoolFree calls: 4256906586
> File descriptor usage for squid:
>     Maximum number of file descriptors:   32000
>     Largest file desc currently in use:   8829
>     Number of file desc currently in use: 5500
>     Files queued for open:                   0
>     Available number of file descriptors: 26500
>     Reserved number of file descriptors:   100
>     Store Disk files open:                   0
>     IO loop method:                     epoll
> Internal Data Structures:
>     932205 StoreEntries
>     932205 StoreEntries with MemObjects
>     931937 Hot Object Cache Items
>          0 on-disk objects
> thanks
> -Ryan
> 
> 
> 
> Ryan Thoryk wrote:
>> The latency should be your disk caches, and I'm also assuming that the
>> sheer size of them could be contributing to it.  Also remember that
>> RAID-0 (or similar) won't help improve performance (since it's access
>> times that you need, not throughput).  We have a much smaller load here
>> (one machine peaks at around 1100 users), and switching from 7200rpm
>> SATA drives to 15k SCSI drives solved a lot of latency issues.
>>
>> One thing you can do is use the max_open_disk_fds value; we found that
>> our SATA machines had major performance issues when over 50 file
>> descriptors were currently open.  That parameter tells Squid to bypass
>> the disk cache if the number of open fd's is over that value (that would
>> definitely help during peak times).  You can find the current number of
>> open fd's in the "Store Disk file open" value in your cachemgr general
>> runtime page.
>>
>> Also I'd recommend greatly decreasing the size of your disk caches, and
>> increasing the cache_mem value (since you have 32 gigs of RAM, I'd
>> probably try to get the Squid process up to around 30 gigs).
>>
>> Ryan Thoryk
>>
>> Ryan Goddard wrote:
>>> Thanks for the response, Adrian.
>>> Is recompile required to change to internal DNS?
>>> I've disabled ECN, pmtu_disc and mtu_probing.
>>> cache_dir is as follows:
>>> (recommended by Henrik)
>>>> cache_dir aufs /squid0 125000 128 256 cache_dir aufs /squid1 125000
>>>> 128 256
>>>> cache_dir aufs /squid2 125000 128 256
>>>> cache_dir aufs /squid3 125000 128 256
>>>> cache_dir aufs /squid4 125000 128 256
>>>> cache_dir aufs /squid5 125000 128 256
>>>> cache_dir aufs /squid6 125000 128 256
>>>> cache_dir aufs /squid7 125000 128 256
>>> No peak data available, here's some pre-peak data:
>>> Cache Manager menu
>>> 5-MINUTE AVERAGE
>>> sample_start_time = 1222199580.85434 (Tue, 23 Sep 2008 19:53:00 GMT)
>>> sample_end_time = 1222199905.507274 (Tue, 23 Sep 2008 19:58:25 GMT)
>>> client_http.requests = 268.239526/sec
>>> client_http.hits = 111.741117/sec
>>> client_http.errors = 0.000000/sec
>>> IOSTAT shows lots of idle time - I'm unclear what you mean by
>>> "profiling" ?
>>> Also, have not tried running w/out any cache - can you explain
>>> how this is done?
>>>
>>> appreciate the assistance.
>>> -Ryan
>>
>>
> 


[Index of Archives]     [Linux Audio Users]     [Samba]     [Big List of Linux Books]     [Linux USB]     [Yosemite News]

  Powered by Linux