There are a lot of sites on the Internet whose site owners that like to hide some of the site's ages or even the entire site from the search engines. You can now find those sites with robots.txt.
Robots.txt is a text file present in the root directory of a site which is used to control which pages are indexed by a robot. If you use the 'disallow' word you can block parts of your sites to be found by search engines.
Step 1. Open the link http://www.google.com/ and search for the keyword :
"robots.txt" "disallow:" filetype:txt
Step 2. You will find the robots.txt file from sites that uses disallow command in it.




This is the hidden page from WhiteHouse.

No comments:
Post a Comment