Robots.txt : It is a file created by webmasters is a set of instructions for the search engines robots that index the web pages. This is called The Robots Exclusion Protocol.
It includes User-Agent and Disallow. Here is the meaning of these words as these are very important part of robots.txt file :
User-Agent : It is used for the search engine robots. It gives the permission to a particular or all robots to index the pages. Here is the syntax :
User-Agent :* ("*" means all robots have permission to crawl all the files/pages)
Disallow : It denies access to robots to directories/files. Here is the syntax :
Disallow : /images/ (it denies access to images directory.)
A simple robots.txt file should be :
User-Agent: *
Disallow:
It means we give permission to all the robots to visit and the directories and files.
It is very easy to create this file. Open notepad editor. Insert these instructions and save it by name robots in text format. Put it in top-level directory and upload it. Suppose your website name is http://www.abctest.com/ then your robots file's path should be here http://www.abctest.com/robots.txt
You can check your robots.txt file here http://tool.motoricerca.info/robots-checker.phtml.
If you need help to create robots.txt, use robots.txt generator tool here : http://www.seochat.com/seo-tools/robots-generator/
Friday, December 28, 2007
Robots.txt - Its definition and use
at 11:17 AM
Subscribe to:
Post Comments (Atom)
No comments:
Post a Comment