Search squid archive

Re: problem with google captcha

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hey Dmitry,

Sometimes it's because there is a sign of a proxy in the middle in the headers or that Google Thinks that there is too much traffic for you network and which is not typical. I do not know how google can allow identification of a network behind a proxy but it seems to me that adding some headers are the right direction.

You can use the "via off" option and also the removal of x-forward-for Headers on the proxy using:
http://www.squid-cache.org/Doc/config/forwarded_for/
with "delete".

it a basic try and find out..
and also fill this form:
https://support.google.com/websearch/contact/ban

One thing that I think you should check is the PTR record for you IP and also RBL checks on your IP.
You can use the tool I have modified at:
http://www1.ngtech.co.il/rbl/rblcheck.rb

You can try to add some google apps headers that restricts some access to google apps and which might block some traffic which should not be there in the first place(as a test only to see if it helps)

Eliezer


On 07/03/2014 03:43 PM, Дмитрий Шиленко wrote:
Hello. I am using SQUID 3,3,8 transparently on opensuse 13.1. Here is
the configuration:

visible_hostname koreamotors.com.ua
acl localnet src 192.168.0.0/24 # RFC1918 possible internal network
acl SSL_ports port 443
acl Safe_ports port 80          # http
acl Safe_ports port 21          # ftp
acl Safe_ports port 443         # https
acl Safe_ports port 70          # gopher
acl Safe_ports port 210         # wais
acl Safe_ports port 1025-65535  # unregistered ports
acl Safe_ports port 280         # http-mgmt
acl Safe_ports port 488         # gss-http
acl Safe_ports port 591         # filemaker
acl Safe_ports port 777         # multiling http
acl CONNECT method CONNECT

#custom Global-it's ACL's
acl AdminsIP src "/etc/squid/AccessLists/AdminsIP.txt"
acl RestrictedDomains dstdomain
"/etc/squid/AccessLists/RestrictedDomains.txt"
acl ad_group_rassh urlpath_regex -i
"/etc/squid/AccessLists/rasshirenie.txt"

http_access allow localhost
http_access deny !Safe_ports
# Deny CONNECT to other than SSL ports
http_access deny CONNECT !SSL_ports

#custom Global-it's settings
http_access allow AdminsIP
http_access deny RestrictedDomains
http_access deny ad_group_rassh
http_access allow localnet
http_access deny all

icp_access allow localnet
icp_access deny all

http_port 192.168.0.97:3128
http_port 192.168.0.97:3129 intercept

cache deny all

# Add any of your own refresh_pattern entries above these.
#
refresh_pattern ^ftp:           1440    20%     10080
refresh_pattern ^gopher:        1440    0%      1440
refresh_pattern -i (/cgi-bin/|\?) 0     0%      0
refresh_pattern .               0       20%     4320
. As soon as the squid and wrap it traffic, Google immediately starts to
request captcha. What should I do to solve this problem?

Dmitry





[Index of Archives]     [Linux Audio Users]     [Samba]     [Big List of Linux Books]     [Linux USB]     [Yosemite News]

  Powered by Linux