Hi to all, I have a question concerning the usage of squid to dump the content of the cache, keep a copy of the cache or block this site (and its bound content) from being deleted from squid´s cache. There are some sites in the net that I would like to ensure that I can surf them in the future even if the site goes offline or gets deleted/ modified. wget is not really useful for this since it does not interpret js and may offer a different result than when surfing with the browser - robots.txt and similar nuisances. So, I would like to have a secured copy of the website I surfed. squid does this, but how do I secure the cached content? Am I missing something in the manuals? Best regards Luigi Monaco