- If you want to maintain and keep your confidential data in your server, then you can place them in a password protected directory. Since Google bot and other spiders won't be able to access the files in those protected directories. For eg., If you'reusing Apache Web Server, you can edit your .htaccess file to password-protect the directory on your server. There are a lot of tools on the web that will let you do this easily.
This is one of simplest and effective way to prevent bots from crawling and indexing all our secret and confidential informations. - By using Robots.txt file to control the access to files and directories in your server. The Robot.txt file tells the bots and other crawlers which files and directories in our server should not be crawled. We need root access of our host in order to use the Robot.txt file.
- We can also use Noindex meta tags to avoid our site single pages from getting indexed from bots and other spiders. The no index meta tag has to be placed in the header portion of the html page.
- To block images from getting indexed then use, < meta name="robots" content="noimageindex" > in the head of the page.
0 comments:
Post a Comment