ai.robots.txt/code/dark_visitors.py
2024-08-06 17:12:26 +01:00

54 lines
No EOL
1.8 KiB
Python

import json
from pathlib import Path
import requests
from bs4 import BeautifulSoup
session = requests.Session()
response = session.get("https://darkvisitors.com/agents")
soup = BeautifulSoup(response.text, "html.parser")
existing_content = json.loads(Path("./robots.json").read_text())
added = 0
to_include = [
"AI Assistants",
"AI Data Scrapers",
"AI Search Crawlers",
"Archivers",
"Developer Helpers",
"Fetchers",
"Intelligence Gatherers",
"Scrapers",
"Search Engine Crawlers",
"SEO Crawlers",
"Uncategorized",
"Undocumented AI Agents"
]
for section in soup.find_all("div", {"class": "agent-links-section"}):
category = section.find("h2").get_text()
for agent in section.find_all("a", href=True):
name = agent.find("div", {"class": "agent-name"}).get_text().strip()
desc = agent.find("p").get_text().strip()
if name in existing_content:
continue
# Template:
# "Claude-Web": {
# "operator": "[Anthropic](https:\/\/www.anthropic.com)",
# "respect": "Unclear at this time.",
# "function": "Scrapes data to train Anthropic's AI products.",
# "frequency": "No information. provided.",
# "description": "Scrapes data to train LLMs and AI products offered by Anthropic."
# }
existing_content[name] = {
"operator": "Unclear at this time.",
"respect": "Unclear at this time.",
"function": f"{category}",
"frequency": "Unclear at this time.",
"description": f"{desc} More info can be found at https://darkvisitors.com/agents{agent['href']}"
}
added += 1
print(f"Added {added} new agents, total is now {len(existing_content)}")
Path("./robots.json").write_text(json.dumps(existing_content, indent=4))