The robots.txt file is a complete text file used for giving instructions to the search engine spiders as to on which files it has to crawl and which files to avoid.
Create a robots.txt file using any text editor and save with extension .txt. Specify the xml sitemap’s absolute path in live robots.txt file for identifying crawler or spider (boot) fast.
You can give instructions to the search console (Webmaster Tool) by submitting the robots.txt file.
The Structure of robots.txt file
Here user-agent: * indicates it is applicable to all search engine agents and disallows this instruction to block all files inside.
Use of robots.txt file in SEO
Robots.txt is a text file that is uploaded to a website’s root directory (inside public html folder). The robots.txt file contains information as to which page to index or not to index by spiders.
All search engines support these extensions.
To block the complete website from crawler, use the Disallow: option followed by a slash:
To allow everything:
To block specific directories use:
You can block individual search engine wise also
If the domain is digitalakash.in, then the robots.txt URL should be like:
Note: The filename is case sensitive (it should be robots.txt)
Benefits of robots.txt file:
- You can block unrelated pages (Like Admin Access)
- You can block individual search engine wise
- simple text file (no html)
|Bing||Images & Video||msnbot-media|
Check your robots.txt file at web master tool: