Operated by Google
Googlebot-Discovery is used by Google to crawl content specifically for the Google Discover feed on mobile devices.
Googlebot-Discovery is used by Google to crawl content specifically for the Google Discover feed on mobile devices.
Googlebot-Discovery is one of Google's specialised crawlers, distinct from the general Googlebot. It serves a specific Google product (Images, Video, News, etc.) and uses the user-agent Googlebot-Discovery. Selectively blocking it disables the corresponding Google feature for your site (e.g., blocking Googlebot-Image removes your images from Google Image Search). Always verify which Google product is affected before blocking.
<code>User-agent: Googlebot-Discovery</code> — Matching is case-insensitive. Robots.txt is fetched from the root of each subdomain separately.
Googlebot-Discovery is verifiable via reverse-DNS lookup on the crawling IP addresses. You can safely allow it unless you have a specific reason to block (e.g., AI training opt-out or SEO tool visibility).Understanding Googlebot-Discovery's purpose helps you decide whether to allow or block it.
Googlebot-Discovery. This is the exact string you must use in robots.txt, Nginx, Apache, or Cloudflare firewall rules to target this bot. User-agent matching in robots.txt is case-insensitive, but the string must be spelled correctly. You can verify that a request genuinely comes from Googlebot-Discovery by performing a reverse-DNS lookup on the source IP — legitimate bots resolve back to their operator's domain.Googlebot-Discovery is verifiable via reverse-DNS lookup on the crawling IP addresses. You can safely allow it unless you have a specific reason to block (e.g., AI training opt-out or SEO tool visibility)./robots.txt file:
User-agent: Googlebot-Discovery Disallow: /This instructs Googlebot-Discovery not to crawl any path on your site. The Disallow: / directive covers the entire domain including subfolders. To only block specific sections, replace / with the path (e.g.,
Disallow: /blog/). Note: robots.txt is publicly readable — any bot or human can inspect it at yourdomain.com/robots.txt.Googlebot-Discovery (case-insensitive grep: grep -i "Googlebot-Discovery" /var/log/nginx/access.log). You can also check Google Search Console → Coverage → Crawl Stats for Googlebot variants. For Googlebot-Discovery specifically, filter by user-agent in your log analysis tool (GoAccess, AWStats, etc.).Disallow: / you can restrict Googlebot-Discovery to specific paths:
User-agent: Googlebot-Discovery Disallow: /private/ Disallow: /staging/ Allow: /This allows Googlebot-Discovery everywhere except the listed paths. Path matching in robots.txt uses prefix matching —
Disallow: /private/ blocks /private/page.html but NOT /public/private/.Check instantly with our free AI Bot Checker
Check Your Website