Skip to content

Simple robots.txt template. Keep unwanted robots out (disallow). White lists (allow) legitimate user-agents. Useful for all websites.

License

Notifications You must be signed in to change notification settings

delpers/robots.txt

 
 

Repository files navigation

Robots.txt File Template

Robots.txt is a repository that contains 2 robots.txt file templates to help webmasters keep unwanted web robots (e.g. scraper bots, people search engines, seo tools, marketing tools, etc.) away from their websites but allow legitimate robots (e.g. search engine crawlers).

To be legitimate and get listed, robots must fully obey the Robots Exclusion Standard. The robots.txt file templates contain a white list. Unlisted robots (User-agents) are, by the conventions of the Robots Exclusion Standard, not allowed to access.

Files

Template Files

The robots.txt template files contain an alphabetically ordered white list of legitimate web robots. In the commented version, each bot is shortly described in a comment above the (list of) user-agent(s). Uncomment or delete bots (User-agents) you do not wish to allow to access your website.

There are two robots.txt file versions:

  1. The regular file (with comments)
    /robots.txt
  2. The minified file (no comments)
    /robots.min.txt

If you use the minified version, do not forget to rename it to robots.txt to be effective.

Mirrors

You can find this repository at:


Warranty and Liability

Compared to the vastness of the topic of blocking bots, robots.txt is a very small, private project! The author makes absolutely no claims and representations to warranties regarding the accuracy or completeness of the information provided. However, you can use the files in this repository AT YOUR OWN RISK.

The descision which bot is wanted/unwanted is done by the author, who is very conservative and opinionated when it comes to blocking bots. However, the author's decisions should be sufficient for many. Do not forget to adjust the list of allowed/forbidden directories to your needs.

License

Robots.txt by Jonas Jacek is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License. Permissions beyond the scope of this license may be available upon request at https://www.jonas.me/.

Contribute

Found a mistake? Want to help? Send a pull request!

About

Simple robots.txt template. Keep unwanted robots out (disallow). White lists (allow) legitimate user-agents. Useful for all websites.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published