What Are Robots.Txt Files?
A robots.txt file is part of the Robots Exclusion Protocol (REP) and will inform search engines which of your URLs they can use to access your website. Essentially, it is a list of instructions created by webmasters to guide search engine bots on just how to crawl their website, highlighting which pages need to be indexed and which ones they can avoid. You’ll find robots.txt files in the website’s root domain. It is possible to edit your robots.txt file.
Every robots.txt file is a plain text file – this follows the Robots Exclusion Standard – that contains at least 1 blog of directives or rules (it can contain more), a specified user-agent (a search engine bot), and either an allow or disallow instruction.
A term you may come across when looking into your robots.txt is disallow. When you add this to your file, you are telling the bots to stay away from a certain page. If you disallow all robots to your site, you do risk your site being removed from the search engines altogether, so you must be careful about what you are entering. If you get removed, it means any traffic will be wiped, and you could see a drop in revenue because of it. The alternative to disallowing is allowing, so use this where you can to ensure you don’t lose your current position on the search results pages.
