summaryrefslogtreecommitdiff
path: root/internal/web/robots.go
diff options
context:
space:
mode:
authorLibravatar Daenney <daenney@users.noreply.github.com>2024-04-22 11:01:37 +0200
committerLibravatar GitHub <noreply@github.com>2024-04-22 11:01:37 +0200
commitdcab555a6b46de69a2c2a26f1fe409017ad229f3 (patch)
tree9d436a1c30a2198bea8d6043df13e1724d5e722b /internal/web/robots.go
parent[chore]: Bump github.com/KimMachineGun/automemlimit from 0.5.0 to 0.6.0 (#2859) (diff)
downloadgotosocial-dcab555a6b46de69a2c2a26f1fe409017ad229f3.tar.xz
[chore] Update robots.txt (#2856)
This updates the robots.txt based on the list of the ai.robots.txt repository. We can look at automating that at some point. It's worth pointing out that some robots, namely the ones by Bytedance, are known to ignore robots.txt entirely.
Diffstat (limited to 'internal/web/robots.go')
-rw-r--r--internal/web/robots.go64
1 files changed, 31 insertions, 33 deletions
diff --git a/internal/web/robots.go b/internal/web/robots.go
index 2511ee1d3..58b541413 100644
--- a/internal/web/robots.go
+++ b/internal/web/robots.go
@@ -29,45 +29,43 @@ const (
robotsTxt = `# GoToSocial robots.txt -- to edit, see internal/web/robots.go
# More info @ https://developers.google.com/search/docs/crawling-indexing/robots/intro
-# Before we commence, a giant fuck you to ChatGPT in particular.
-# https://platform.openai.com/docs/gptbot
-User-agent: GPTBot
-Disallow: /
-
-# As of September 2023, GPTBot and ChatGPT-User are equivalent. But there's no telling
-# when OpenAI might decide to change that, so block this one too.
-User-agent: ChatGPT-User
-Disallow: /
-
-# And a giant fuck you to Google Bard and their other generative AI ventures too.
-# https://developers.google.com/search/docs/crawling-indexing/overview-google-crawlers
-User-agent: Google-Extended
-Disallow: /
-
-# Block CommonCrawl. Used in training LLMs and specifically GPT-3.
-# https://commoncrawl.org/faq
+# AI scrapers and the like.
+# https://github.com/ai-robots-txt/ai.robots.txt/
+User-agent: AdsBot-Google
+User-agent: Amazonbot
+User-agent: anthropic-ai
+User-agent: Applebot
+User-agent: AwarioRssBot
+User-agent: AwarioSmartBot
+User-agent: Bytespider
User-agent: CCBot
-Disallow: /
-
-# Block Omgilike/Webz.io, a "Big Web Data" engine.
-# https://webz.io/blog/web-data/what-is-the-omgili-bot-and-why-is-it-crawling-your-website/
-User-agent: Omgilibot
-Disallow: /
-
-# Block Faceboobot, because Meta.
-# https://developers.facebook.com/docs/sharing/bot
+User-agent: ChatGPT-User
+User-agent: ClaudeBot
+User-agent: Claude-Web
+User-agent: cohere-ai
+User-agent: DataForSeoBot
User-agent: FacebookBot
+User-agent: FriendlyCrawler
+User-agent: Google-Extended
+User-agent: GoogleOther
+User-agent: GPTBot
+User-agent: ImagesiftBot
+User-agent: magpie-crawler
+User-agent: Meltwater
+User-agent: omgili
+User-agent: omgilibot
+User-agent: peer39_crawler
+User-agent: peer39_crawler/1.0
+User-agent: PerplexityBot
+User-agent: PiplBot
+User-agent: Seekr
+User-agent: YouBot
Disallow: /
# Well-known.dev crawler. Indexes stuff under /.well-known.
# https://well-known.dev/about/
-User-agent: WellKnownBot
-Disallow: /
-
-# Block Amazonbot, because Amazon.
-# https://developer.amazon.com/amazonbot
-User-agent: Amazonbot
-Disallow: /
+User-agent: WellKnownBot
+Disallow: /
# Rules for everything else.
User-agent: *