Skip to content

Instantly share code, notes, and snippets.

@eksiscloud
Last active April 7, 2020 17:01
Show Gist options
  • Save eksiscloud/e95dc3aa72ada44072dc88f0e1c7d755 to your computer and use it in GitHub Desktop.
Save eksiscloud/e95dc3aa72ada44072dc88f0e1c7d755 to your computer and use it in GitHub Desktop.
Another robots.txt for good bots, spiders and crawlers
User-agent: AffiliateLabz
User-agent: AhrefsBot
User-agent: AspiegelBot
User-agent: coccocbot
User-agent: IAS crawler
User-agent: MJ12bot
User-agent: seewithkids.com
User-agent: SemrushBot/6~bl
User-agent: TTD-Content
User-agent: YandexBot
Disallow: /
User-agent: *
Disallow: /wp-admin/
Allow: /wp-admin/admin-ajax.php
Disallow: /wp-login.php
Disallow: /xmlrpc.php
Sitemap: https://www.example.tld/sitemap_index.xml
@eksiscloud
Copy link
Author

Just an example but working one.

There is no point what so ever even bother to try stop bad bots using robots.txt. The list will be too long and difficult to read. Actually, you should use robots.txt just to limit those bots who have access, like Google etc, and stop everybody else (good and bad ones) usein webserver or reverse proxy.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment