What Is A Robot.txt?
What is a Robots.txt file?
The robots exclusion protocol (REP), or robots.txt is a text file webmasters create to instruct robots (typically search engine robots) how to crawl and index pages on their website. In short Web site, owners use the /robots.txt file to give instructions about their site to web robots; this is called The Robots Exclusion Protocol.
- To exclude all robot for an entire server:
user-agent:*
Disallow:/
- To allow all robots complete access:
user-agent:*
Disallow:
- To exclude all robot from part of Server:
User-agent:*
Disallow:/cgi-bin/
Disallow:/tmp/
Disallow:/junk/
- To exclude Single Robot:
User-agent:*
Disallow:user-agent:*
Disallow:/
- To exclude all file except one:
User-agent:*
Disallow:/-joe/stuff/
- Alternatively You can Explicity Disallow pages:
user-agent:*
Disallow:/-joe/junk.html
Disallow:/-joe/foo.html
Disallow:/-joe/bar.html
Comments
Post a Comment