Cookie Consent by Free Privacy Policy Generator Mail.RU_Bot User Agent - Mail.Ru Bot Details | CL SEO

Mail.RU_Bot

Mail.Ru Since 2008
Search Respects robots.txt
#search #russian #mail-ru #crawler
Quick Actions
Official Docs

What is Mail.RU_Bot?

Mail.RU_Bot is the web crawler for Mail.Ru, one of Russia's largest internet companies and email providers. The bot powers Mail.Ru's search functionality and various other services within their ecosystem. As a major player in the Russian internet (Runet), Mail.Ru's crawler is important for websites targeting Russian-speaking audiences. The bot indexes content for both web search and specialized searches within Mail.Ru's properties, including social networks and email services.

User Agent String

Mozilla/5.0 (compatible; Mail.RU_Bot/2.0; +http://go.mail.ru/bot)

How to Control Mail.RU_Bot

Block Completely

To prevent Mail.RU_Bot from accessing your entire website, add this to your robots.txt file:

# Block Mail.RU_Bot User-agent: Mail.RU_Bot Disallow: /

Block Specific Directories

To restrict access to certain parts of your site while allowing others:

User-agent: Mail.RU_Bot Disallow: /admin/ Disallow: /private/ Disallow: /wp-admin/ Allow: /public/

Set Crawl Delay

To slow down the crawl rate (note: not all bots respect this directive):

User-agent: Mail.RU_Bot Crawl-delay: 10

How to Verify Mail.RU_Bot

Verification Method:
Mail.Ru IP ranges

Learn more in the official documentation.

Detection Patterns

Multiple ways to detect Mail.RU_Bot in your application:

Basic Pattern

/Mail\.RU_Bot/i

Strict Pattern

/^Mozilla/5\.0 \(compatible; Mail\.RU_Bot/2\.0; \+http\://go\.mail\.ru/bot\)$/

Flexible Pattern

/Mail\.RU_Bot[\s\/]?[\d\.]*?/i

Vendor Match

/.*Mail\.Ru.*Mail\.RU_Bot/i

Implementation Examples

// PHP Detection for Mail.RU_Bot function detect_mail_ru_bot() { $user_agent = $_SERVER['HTTP_USER_AGENT'] ?? ''; $pattern = '/Mail\\.RU_Bot/i'; if (preg_match($pattern, $user_agent)) { // Log the detection error_log('Mail.RU_Bot detected from IP: ' . $_SERVER['REMOTE_ADDR']); // Set cache headers header('Cache-Control: public, max-age=3600'); header('X-Robots-Tag: noarchive'); // Optional: Serve cached version if (file_exists('cache/' . md5($_SERVER['REQUEST_URI']) . '.html')) { readfile('cache/' . md5($_SERVER['REQUEST_URI']) . '.html'); exit; } return true; } return false; }
# Python/Flask Detection for Mail.RU_Bot import re from flask import request, make_response def detect_mail_ru_bot(): user_agent = request.headers.get('User-Agent', '') pattern = r'Mail.RU_Bot' if re.search(pattern, user_agent, re.IGNORECASE): # Create response with caching response = make_response() response.headers['Cache-Control'] = 'public, max-age=3600' response.headers['X-Robots-Tag'] = 'noarchive' return True return False # Django Middleware class MailRU_BotMiddleware: def __init__(self, get_response): self.get_response = get_response def __call__(self, request): if self.detect_bot(request): # Handle bot traffic pass return self.get_response(request)
// JavaScript/Node.js Detection for Mail.RU_Bot const express = require('express'); const app = express(); // Middleware to detect Mail.RU_Bot function detectMailRU_Bot(req, res, next) { const userAgent = req.headers['user-agent'] || ''; const pattern = /Mail.RU_Bot/i; if (pattern.test(userAgent)) { // Log bot detection console.log('Mail.RU_Bot detected from IP:', req.ip); // Set cache headers res.set({ 'Cache-Control': 'public, max-age=3600', 'X-Robots-Tag': 'noarchive' }); // Mark request as bot req.isBot = true; req.botName = 'Mail.RU_Bot'; } next(); } app.use(detectMailRU_Bot);
# Apache .htaccess rules for Mail.RU_Bot # Block completely RewriteEngine On RewriteCond %{HTTP_USER_AGENT} Mail\.RU_Bot [NC] RewriteRule .* - [F,L] # Or redirect to a static version RewriteCond %{HTTP_USER_AGENT} Mail\.RU_Bot [NC] RewriteCond %{REQUEST_URI} !^/static/ RewriteRule ^(.*)$ /static/$1 [L] # Or set environment variable for PHP SetEnvIfNoCase User-Agent "Mail\.RU_Bot" is_bot=1 # Add cache headers for this bot <If "%{HTTP_USER_AGENT} =~ /Mail\.RU_Bot/i"> Header set Cache-Control "public, max-age=3600" Header set X-Robots-Tag "noarchive" </If>
# Nginx configuration for Mail.RU_Bot # Map user agent to variable map $http_user_agent $is_mail_ru_bot { default 0; ~*Mail\.RU_Bot 1; } server { # Block the bot completely if ($is_mail_ru_bot) { return 403; } # Or serve cached content location / { if ($is_mail_ru_bot) { root /var/www/cached; try_files $uri $uri.html $uri/index.html @backend; } try_files $uri @backend; } # Add headers for bot requests location @backend { if ($is_mail_ru_bot) { add_header Cache-Control "public, max-age=3600"; add_header X-Robots-Tag "noarchive"; } proxy_pass http://backend; } }

Should You Block This Bot?

Recommendations based on your website type:

Site Type Recommendation Reasoning
E-commerce Allow Essential for product visibility in search results
Blog/News Allow Increases content reach and discoverability
SaaS Application Block No benefit for application interfaces; preserve resources
Documentation Allow Improves documentation discoverability for developers
Corporate Site Allow Allow for public pages, block sensitive areas like intranets

Advanced robots.txt Configurations

E-commerce Site Configuration

User-agent: Mail.RU_Bot Crawl-delay: 5 Disallow: /cart/ Disallow: /checkout/ Disallow: /my-account/ Disallow: /api/ Disallow: /*?sort= Disallow: /*?filter= Disallow: /*&page= Allow: /products/ Allow: /categories/ Sitemap: https://example.com/sitemap.xml

Publishing/Blog Configuration

User-agent: Mail.RU_Bot Crawl-delay: 10 Disallow: /wp-admin/ Disallow: /drafts/ Disallow: /preview/ Disallow: /*?replytocom= Allow: /

SaaS/Application Configuration

User-agent: Mail.RU_Bot Disallow: /app/ Disallow: /api/ Disallow: /dashboard/ Disallow: /settings/ Allow: / Allow: /pricing/ Allow: /features/ Allow: /docs/

Quick Reference

User Agent Match

Mail.RU_Bot

Robots.txt Name

Mail.RU_Bot

Category

search

Respects robots.txt

Yes
Copied to clipboard!