ai.robots.txt

ai.robots.txt

A list of AI agents and robots to block.

Stars: 692

Visit
 screenshot

ai.robots.txt is an open list of web crawlers associated with AI companies and the training of LLMs to block. Users are encouraged to contribute to and implement this list on their own site. The list includes crawlers sourced from Dark Visitors, and contributors can add information about a crawler by making a pull request with the bot name added to `robots.txt`, `ai.txt`, and any relevant details in `table-of-bot-metrics.md`.

README:

ai.robots.txt

This is an open list of web crawlers associated with AI companies and the training of LLMs to block. We encourage you to contribute to and implement this list on your own site. See information about the listed crawlers.

A number of these crawlers have been sourced from Dark Visitors and we appreciate the ongoing effort they put in to track these crawlers.

If you'd like to add information about a crawler to the list, please make a pull request with the bot name added to robots.txt, ai.txt, and any relevant details in table-of-bot-metrics.md to help people understand what's crawling.

Contributing

A note about contributing: updates should be added/made to robots.json. A GitHub action, courtesy of Adam, will then generate the updated robots.txt and table-of-bot-metrics.md.

Subscribe to updates

You can subscribe to list updates via RSS/Atom with the releases feed:

https://github.com/ai-robots-txt/ai.robots.txt/releases.atom

You can subscribe with Feedly, Inoreader, The Old Reader, Feedbin, or any other reader app.

Additional resources


Thank you to Glyn for pushing me to set this up after I posted about blocking these crawlers.

For Tasks:

Click tags to check more tools for each tasks

For Jobs:

Alternative AI tools for ai.robots.txt

Similar Open Source Tools

For similar tasks

For similar jobs