What is: Robots.txt
What is: Robots.txt 

Robots.txt is a content record which enables a site to give directions to web creeping bots.

Web indexes like Google utilize these web crawlers, some of the time called web robots, to document and order sites. Mosts bots are arranged to look for a robots.txt record on the server before it peruses some other document from the site. It does this to check whether a site's proprietor has some extraordinary directions on the most proficient method to creep and record their site.

The robots.txt document contains an arrangement of guidelines that demand the bot to overlook particular records or registries. This might be with the end goal of protection or in light of the fact that the site proprietor trusts that the substance of those documents and catalogs is insignificant to the classification of the site in web crawlers.

On the off chance that a site has in excess of one subdomain, each subdomain must have its own particular robots.txt document. It is critical to take note of that not all bots will respect a robots.txt record. Some vindictive bots will even read the robots.txt document to discover which records and indexes they should target first. Likewise, regardless of whether a robots.txt document educates bots to disregard a particular pages on the site, those pages may even now show up in list items of they are connected to by different pages that are crept.