Cookie Consent by Free Privacy Policy Generator User Agent Directory | Bot Database | CL SEO

User Agent Directory

A comprehensive database of 124 verified user agents crawling the web. Identify AI bots, SEO crawlers, search engine spiders and understand their behavior patterns.

124
Total User Agents
17
AI Crawlers
36
Search Engines
24
SEO Tools
Vendor: OpenAI
Mozilla/5.0 AppleWebKit/537.36 (KHTML, like Gecko; compatible; GPTBot/1.0; +https://openai.com/gptbot)
#ai-training #chatgpt #crawler #gpt
robots.txt: GPTBot
Vendor: OpenAI
Mozilla/5.0 AppleWebKit/537.36 (KHTML, like Gecko; compatible; ChatGPT-User/1.0; +https://openai.com/bot)
#ai #chatgpt #browsing #plugin
robots.txt: ChatGPT-User
Vendor: OpenAI
Mozilla/5.0 AppleWebKit/537.36 (KHTML, like Gecko; compatible; OAI-SearchBot/1.0; +https://openai.com/searchbot)
#ai #search #openai #crawler
robots.txt: OAI-SearchBot
Vendor: Anthropic
Mozilla/5.0 AppleWebKit/537.36 (KHTML, like Gecko; compatible; Claude-Web/1.0; +https://www.anthropic.com)
#ai #claude #anthropic #crawler
robots.txt: Claude-Web
Vendor: Anthropic
ClaudeBot/1.0; +https://www.anthropic.com
#ai #claude #anthropic #training
robots.txt: ClaudeBot
Vendor: Common Crawl
CCBot/2.0 (https://commoncrawl.org/faq/)
#dataset #ai-training #crawler #open-data
robots.txt: CCBot
Vendor: Perplexity AI
Mozilla/5.0 AppleWebKit/537.36 (KHTML, like Gecko; compatible; PerplexityBot/1.0; +https://perplexity.ai/bot)
#ai #search #answer-engine #crawler
robots.txt: PerplexityBot
Vendor: Google
Mozilla/5.0 AppleWebKit/537.36 (KHTML, like Gecko; compatible; Google-Extended/1.0; +https://developers.google.com/search/docs/crawling-indexing/google-common-crawlers)
#ai #bard #google #training
robots.txt: Google-Extended
Vendor: Meta
Meta-ExternalAgent/1.0 (+https://developers.facebook.com/docs/sharing/bot)
#ai #meta #facebook #training
robots.txt: Meta-ExternalAgent
Vendor: Amazon
Amazonbot/0.1 (+https://developer.amazon.com/support/amazonbot)
#amazon #alexa #ai #crawler
robots.txt: Amazonbot
Vendor: Google
Mozilla/5.0 (compatible; Googlebot/2.1; +http://www.google.com/bot.html)
#search #google #crawler #indexing
robots.txt: Googlebot
Vendor: Google
Googlebot-Image/1.0
#search #google #images #crawler
robots.txt: Googlebot-Image
Vendor: Google
Googlebot-Video/1.0
#search #google #video #crawler
robots.txt: Googlebot-Video
Vendor: Google
Googlebot-News
#search #google #news #crawler
robots.txt: Googlebot-News
Vendor: Google
AdsBot-Google (+http://www.google.com/adsbot.html)
#ads #google #quality-check #crawler
robots.txt: AdsBot-Google
Vendor: Google
Mediapartners-Google
#adsense #google #ads #crawler
robots.txt: Mediapartners-Google
Vendor: Microsoft
Mozilla/5.0 (compatible; bingbot/2.0; +http://www.bing.com/bingbot.htm)
#search #bing #microsoft #crawler
robots.txt: bingbot
Vendor: Microsoft
Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/534+ (KHTML, like Gecko) BingPreview/1.0b
#search #bing #preview #snapshot
robots.txt: BingPreview
Vendor: DuckDuckGo
DuckDuckBot/1.0; (+http://duckduckgo.com/duckduckbot.html)
#search #privacy #duckduckgo #crawler
robots.txt: DuckDuckBot
Vendor: DuckDuckGo
Mozilla/5.0 (compatible; DuckAssistBot/1.0; +https://duckduckgo.com/duckassist)
#ai #assistant #duckduckgo #privacy
robots.txt: DuckAssistBot
Vendor: Yandex
Mozilla/5.0 (compatible; YandexBot/3.0; +http://yandex.com/bots)
#search #yandex #russian #crawler
robots.txt: YandexBot
Vendor: Baidu
Mozilla/5.0 (compatible; Baiduspider/2.0; +http://www.baidu.com/search/spider.html)
#search #baidu #chinese #crawler
robots.txt: Baiduspider
Vendor: Apple
Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_5) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/13.1.1 Safari/605.1.15 (Applebot/0.1; +http://www.apple.com/go/applebot)
#search #apple #siri #spotlight
robots.txt: Applebot
Vendor: Apple
Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_5) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/13.1.1 Safari/605.1.15 (Applebot-Extended/0.1; +http://www.apple.com/go/applebot)
#ai #apple #training #crawler
robots.txt: Applebot-Extended

What are User Agents?

User agents are strings that identify the software making requests to your website. They help servers understand what type of client is accessing the content - whether it's a browser, search engine crawler, SEO tool, or AI bot.

Why This Matters

  • Control which bots can access your content
  • Identify AI crawlers harvesting data
  • Monitor SEO tools analyzing your site
  • Understand your traffic sources better

How to Use This Data

  • Create robots.txt rules: Block or allow specific bots
  • Server configuration: Set up rate limiting for aggressive crawlers
  • Analytics filtering: Exclude bot traffic from reports
  • Security monitoring: Identify suspicious crawler activity