mirror of
https://github.com/ai-robots-txt/ai.robots.txt.git
synced 2025-04-03 19:03:57 +00:00
.htaccess: Allow robots access to /robots.txt
This commit is contained in:
parent
abfd6dfcd1
commit
c0d418cd87
3 changed files with 3 additions and 3 deletions
|
@ -1,3 +1,3 @@
|
|||
RewriteEngine On
|
||||
RewriteCond %{HTTP_USER_AGENT} ^.*(AI2Bot|Ai2Bot-Dolma|Amazonbot|anthropic-ai|Applebot|Applebot-Extended|Brightbot\ 1.0|Bytespider|CCBot|ChatGPT-User|Claude-Web|ClaudeBot|cohere-ai|cohere-training-data-crawler|Crawlspace|Diffbot|DuckAssistBot|FacebookBot|FriendlyCrawler|Google-Extended|GoogleOther|GoogleOther-Image|GoogleOther-Video|GPTBot|iaskspider/2.0|ICC-Crawler|ImagesiftBot|img2dataset|ISSCyberRiskCrawler|Kangaroo\ Bot|Meta-ExternalAgent|Meta-ExternalFetcher|OAI-SearchBot|omgili|omgilibot|PanguBot|PerplexityBot|PetalBot|Scrapy|SemrushBot-OCOB|SemrushBot-SWA|Sidetrade\ indexer\ bot|Timpibot|VelenPublicWebCrawler|Webzio-Extended|YouBot).*$ [NC]
|
||||
RewriteRule .* - [F,L]
|
||||
RewriteRule !^/?robots\.txt$ - [F,L]
|
||||
|
|
|
@ -142,7 +142,7 @@ def json_to_htaccess(robot_json):
|
|||
robots = map(lambda el: el.replace(" ", "\\ "), robot_json.keys())
|
||||
htaccess += "|".join(robots)
|
||||
htaccess += ").*$ [NC]\n"
|
||||
htaccess += "RewriteRule .* - [F,L]"
|
||||
htaccess += "RewriteRule !^/?robots\\.txt$ - [F,L]\n"
|
||||
return htaccess
|
||||
|
||||
|
||||
|
|
|
@ -1,3 +1,3 @@
|
|||
RewriteEngine On
|
||||
RewriteCond %{HTTP_USER_AGENT} ^.*(AI2Bot|Ai2Bot-Dolma|Amazonbot|anthropic-ai|Applebot|Applebot-Extended|Bytespider|CCBot|ChatGPT-User|Claude-Web|ClaudeBot|cohere-ai|Diffbot|FacebookBot|facebookexternalhit|FriendlyCrawler|Google-Extended|GoogleOther|GoogleOther-Image|GoogleOther-Video|GPTBot|iaskspider/2.0|ICC-Crawler|ImagesiftBot|img2dataset|ISSCyberRiskCrawler|Kangaroo\ Bot|Meta-ExternalAgent|Meta-ExternalFetcher|OAI-SearchBot|omgili|omgilibot|PerplexityBot|PetalBot|Scrapy|Sidetrade\ indexer\ bot|Timpibot|VelenPublicWebCrawler|Webzio-Extended|YouBot).*$ [NC]
|
||||
RewriteRule .* - [F,L]
|
||||
RewriteRule !^/?robots\.txt$ - [F,L]
|
||||
|
|
Loading…
Reference in a new issue