Hi Doug,
So does that prevent crawling and browsing, but does allow if I click a link or include a file it will work?No, it prevents the directory and all files within from being served by the webserver at all. Anything else is either half-secure or half- brokenI don't understand what you mean by "crawling", "browsing", and "include a file". They're really all the same thing: A client (be it Firefox or GoogleBot) is asking the webserver for something.If you want to prevent the nice robots from asking for something, you can use a robots.txt file. This will not prevent naughty robots from asking for something.
Lets think about this a different way.Say I have a directory of files that contain my MySQL connection information, queries, etc, etc.
How do I prevent people from browsing the directory but allow the files to still be used when I include them in a page. Say to connect to MySQL.
-Jason --------------------------------------------------------------------- The official User-To-User support forum of the Apache HTTP Server Project. See <URL:http://httpd.apache.org/userslist.html> for more info. To unsubscribe, e-mail: users-unsubscribe@xxxxxxxxxxxxxxxx " from the digest: users-digest-unsubscribe@xxxxxxxxxxxxxxxx For additional commands, e-mail: users-help@xxxxxxxxxxxxxxxx