What Is A Robot.txt?


What is a Robots.txt file?
The robots exclusion protocol (REP), or robots.txt is a text file webmasters create to instruct robots (typically search engine robots) how to crawl and index pages on their website. In short Web site, owners use the /robots.txt file to give instructions about their site to web robots; this is called The Robots Exclusion Protocol.

- To exclude all robot for an entire server:

  user-agent:*
  Disallow:/

- To allow all robots complete access:

  user-agent:*
  Disallow:

- To exclude all robot from part of Server:

   User-agent:*
   Disallow:/cgi-bin/
   Disallow:/tmp/
   Disallow:/junk/

- To exclude Single Robot:

  User-agent:*
  Disallow:user-agent:*
  Disallow:/

- To exclude all file except one:

  User-agent:*
  Disallow:/-joe/stuff/

- Alternatively You can Explicity Disallow pages:

  user-agent:*
  Disallow:/-joe/junk.html
  Disallow:/-joe/foo.html
  Disallow:/-joe/bar.html

Comments

Popular posts from this blog

What Is Reach ?

Some Of The Top Keywords For Digital Marketing