Search squid archive

Re: decreased requests per second with big file size

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



You now got my attention!
Depends on what you want you might be able to use external logging helper for that. I am unsure if it is possible to use two access log directives in the configuration and Amos or others can answer that. It is pretty simple to implement since the input data will flow like any access log lines but to a program.
Then you can update some DB live.
There is an option to even use the same log daemon for both logging and http interface for the statistics but I will not go this way now.
...
I have tested and it seems possible to use two access_log directives.
I am unsure how to implement the idea of both access.log and external logger combined.
But there is sure the option of:
"access_log tcp://host:port"
Which if you will write a TCP service will make your life easy.
I am unfamiliar with the logging protocol but it seems like the wiki can help with that.

*I am willing to write an example tcp log daemon for the squid in golang\ruby\python for the squid project if one is not present in these languages.*

Eliezer

* Testing 3.5.10 RPMs is in progress.

On 14/10/2015 09:00, Ambadas H wrote:
Hi Eliezer,

Its mostly like a live feed.

I am writing these sites+(a client tracking parameter) to a flat file via
squid, from where another process reads it & does further processing (eg.
analyze top sites used by any particular client).

And that is why i was working on getting just the urls entered by clients.


Ambadas


On Tue, Oct 13, 2015 at 2:01 PM, Eliezer Croitoru <eliezer@xxxxxxxxxxxx>
wrote:

Hey Ambadas,

I was wondering if you want it to be something like a "live feed" or just
for logs analyzing?

Eliezer

On 09/10/2015 15:47, Ambadas H wrote:

Hi,

I am using below setup:
Squid proxy 3.5.4.
CentOS 7.1

I am trying to analyze the most used websites by the users via Squid
proxy.
I just require the first GET request for that particular browsed page page
& not the proceeding GETs of that same page.

Eg:
1) user enters *http://google.com <http://google.com>* in client
(mozilla)
2) client gets page containing some other urls
3) client initiates multiple GETs for same requested page without users
knowledge

I myself tried a logic where I assumed if "Referer" header is present,
then
its not the first GET but a proceeding one for same requested page.

I know i cant rely on "Referer" header to be always present as its not
mandatory. But
I want to know if my logic is correct? & also if there's any alternative
solution?



_______________________________________________
squid-users mailing list
squid-users@xxxxxxxxxxxxxxxxxxxxx
http://lists.squid-cache.org/listinfo/squid-users


_______________________________________________
squid-users mailing list
squid-users@xxxxxxxxxxxxxxxxxxxxx
http://lists.squid-cache.org/listinfo/squid-users



_______________________________________________
squid-users mailing list
squid-users@xxxxxxxxxxxxxxxxxxxxx
http://lists.squid-cache.org/listinfo/squid-users




[Index of Archives]     [Linux Audio Users]     [Samba]     [Big List of Linux Books]     [Linux USB]     [Yosemite News]

  Powered by Linux