On 10.07.2012 00:58, Muhammed Shehata wrote:
>> Hi All,
>> hope you all are doing well
>> actually I followed the replies on squid users-mail-list about
>> sslbump issues with showing up some websites inline without images or
>> css style sheet
>> like https://gmail.com and https://facebook.com
>> I do use broken sites acl to exclude them
>> however I see that method is CONNECT for those excluded website
>> not Get as all other bumped sites but it still the same result
All sites, bumped or not should be logging a CONNECT request. We are
still sorting out exactly how the bumped ones gets displayed though.
For sites which you *prevent* bumping (broken_sites), of course you
will never see the GET. The GET is *inside* the encrypted portion of the
request which "ssl_bump deny ..." prevents Squid decrypting.
>> 1341837646.893 45801 x.x.x.x TCP_MISS/200 62017 CONNECT
>> twitter.com:443 - DIRECT/199.59.150.7
>> I'm using squid 3.1.19
>> acl broken_sites dstdomain .twitter.com
>> acl broken_sites dstdomain .facebook.com
>> ssl_bump deny broken_sites
>> ssl_bump allow all
>> http_port 192.168.0.1:3128 ssl-bump generate-host-certificates=on
>> dynamic_cert_mem_cache_size=40MB
>> cert=/etc/pki/tls/certs/sslintercept.crt
>> key=/etc/pki/tls/certs/sslintercept.key
>>
Both sites include embeded objects (images, scripts, CSS) using domains
which never appear in the browser address bar. You may need to add those
alternative domains into your broken_sites ACL to keep the same
behaviour for the main HTML portion of the site and its other objects.
Other issues:
Twitter has some serious cacheability problems, and a number of smaller
annoying bugs affecting proxy responses:
http://redbot.org/?descend=True&uri=https://twitter.com/
* A gzip-compressed response was sent when it wasn't asked for.
* The resource doesn't send Vary consistently.
* The ETag doesn't change between negotiated representations.
* There was a problem checking for Last-Modified validation support.
* A ranged request returned another representation.
Facebook has improved noticably over the last year or so, but the main
site still has one major Vary issue that breaks cacheability:
http://redbot.org/?descend=True&uri=https://www.facebook.com/
* This response is negotiated, but doesn't have an appropriate Vary
header.
* The response body is different when content negotiation happens.
* Content negotiation for gzip compression makes the response 19%
larger.
These problems are operating on both the HTTP and HTTPS accesses to the
site.
Amos
Received on Mon Jul 09 2012 - 22:33:34 MDT
This archive was generated by hypermail 2.2.0 : Tue Jul 10 2012 - 12:00:02 MDT