Skip to content

Instantly share code, notes, and snippets.

@samueltc
Created May 9, 2014 19:18
Show Gist options
  • Star 1 You must be signed in to star a gist
  • Fork 0 You must be signed in to fork a gist
  • Save samueltc/2dad265b519f6ecb4e95 to your computer and use it in GitHub Desktop.
Save samueltc/2dad265b519f6ecb4e95 to your computer and use it in GitHub Desktop.
squid cache_peer with crawlera/proxyhub
Change: $scrapinhub/crawlera_hostname$, $username$, $password$ and you're ready to go.
Tested with squid3
hierarchy_stoplist cgi-bin ?
acl QUERY urlpath_regex cgi-bin \?
no_cache deny QUERY
hosts_file /etc/hosts
refresh_pattern ^ftp: 1440 20% 10080
refresh_pattern ^gopher: 1440 0% 1440
refresh_pattern . 0 20% 4320
acl all src all
acl localhost src 127.0.0.1/255.255.255.255
acl to_localhost dst 127.0.0.0/8
acl purge method PURGE
acl CONNECT method CONNECT
cache_mem 1024 MB
http_access allow manager localhost
http_access deny manager
http_access allow purge localhost
http_access deny purge
acl lan src 192.168.1.1 192.168.2.0/24
http_access allow localhost
http_access allow lan
http_access deny all
http_reply_access allow all
icp_access allow all
visible_hostname myclient.hostname.com
coredump_dir /var/spool/squid
http_port 3128
cache_peer $scrapinhub/crawlera_hostname$ parent 8010 0 login=$username$:$password$ no-delay no-query proxy-only
never_direct allow all
request_header_access From deny all
request_header_access Via deny all
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment