Give us a call: (800) 252-6164
SEO · Crawl Logs · Google User Agents

HOW MANY GOOGLEBOT
Crawlers Are There? (2026 Update)

The “number of Googlebot crawlers” depends on what you count. Google documents multiple crawler and fetcher user agents across Search, Ads, Shopping, and user-triggered tools. This page shows what’s current, what you’ll see in logs, and how to verify real Google traffic.

  • Updated to Google’s current docs
  • Log-ready tokens & examples
  • Verification steps to prevent spoofing
  • Robots.txt guidance that works

Quick answer: there isn’t one “official number”

Most people mean “How many Google user agents might hit my site?” Google’s current documentation groups them into three buckets: common crawlers, special-case crawlers, and user-triggered fetchers. Each list is described as not exhaustive, because Google runs many requestors and product-specific clients.

Practical take: If you want a workable 2026 list for logging, filtering, and robots.txt, use the documented user-agent tokens below (and verify suspicious traffic before trusting the UA string).

What people call “Googlebot” vs. Google’s full crawler ecosystem

In logs, “Googlebot” is usually one of the two core Search crawlers (desktop or smartphone). But Google also runs other crawlers for images, video, Shopping, Ads, abuse detection, and user-triggered tools (like Search Console verification).

Category What it does How it behaves
Common crawlers Core crawling for Google products (including Search) Respects robots.txt for automatic crawls
Special-case crawlers Product-specific crawling with an agreement (ex: Ads) May ignore global * group rules
User-triggered fetchers Fetches initiated by users (ex: site verification) Often ignore robots.txt because the user requested the fetch
SEO note: The “count” question is less important than correctly identifying which agent you’re seeing, and which control surface applies (robots.txt vs. verification vs. rate-limiting).

Common Google crawlers you’ll see in logs (2026 list)

These are the “workhorse” crawlers Google documents for common products. In practice, these are the user agents most site owners run into when investigating crawl spikes or indexing behavior.

Crawler / token What it’s for What it looks like in logs (example)
Googlebot
Googlebot
Core Search crawler (Desktop + Smartphone variants) … (compatible; Googlebot/2.1; +http://www.google.com/bot.html)
Googlebot-Image
Googlebot-Image
Fetches image URLs for Google surfaces that show images Googlebot-Image/1.0
Googlebot-Video
Googlebot-Video
Video-focused crawling for video-related features Googlebot-Video/1.0
Googlebot-News
Googlebot-News
News crawling (token exists even though it may not have a unique UA string) Crawling can occur via various Googlebot strings
Storebot-Google
Storebot-Google
Shopping surfaces (StoreBot) … Storebot-Google/1.0 …
Google-InspectionTool
Google-InspectionTool
Search testing tools (Rich Results Test, URL Inspection) … (compatible; Google-InspectionTool/1.0;)
GoogleOther
GoogleOther
Generic crawler used by product teams (not Search ranking) … (compatible; GoogleOther)
GoogleOther-Image
GoogleOther-Image
GoogleOther optimized for image URLs GoogleOther-Image/1.0
GoogleOther-Video
GoogleOther-Video
GoogleOther optimized for video URLs GoogleOther-Video/1.0
Google-CloudVertexBot
Google-CloudVertexBot
Crawls requested by site owners building Vertex AI Agents Google-CloudVertexBot
Google-Extended
Google-Extended
Control token for whether content can be used for future Gemini training / grounding Token only; uses other Google UAs for crawling
Important: Google-Extended does not affect inclusion or ranking in Google Search; it’s a separate control token. (If you’re managing AI usage policies, this is the one stakeholders ask about.)

Special-case crawlers (Ads / AdSense / abuse)

Special-case crawlers are associated with products where there’s an agreement about the crawl process. Google notes these may ignore the global * robots group.

Token Typical purpose Example user agent
APIs-Google
APIs-Google
Push notification delivery for Google APIs APIs-Google (+https://developers.google.com/webmasters/APIs-Google.html)
AdsBot-Google-Mobile
AdsBot-Google-Mobile
Ad quality checking (mobile web) … (compatible; AdsBot-Google-Mobile; +http://www.google.com/mobile/adsbot.html)
AdsBot-Google
AdsBot-Google
Ad quality checking AdsBot-Google (+http://www.google.com/adsbot.html)
Mediapartners-Google
Mediapartners-Google
AdSense crawler for relevant ad provisioning (compatible; Mediapartners-Google/2.1; +http://www.google.com/bot.html)
Google-Safety
Google-Safety
Abuse-specific crawling (ex: malware discovery) Google-Safety
Operational note: If you’re seeing crawl spikes from AdsBot or Mediapartners, treat it as separate from core SEO crawling. Your solution might be ad policy / page quality / performance, not crawl budget.

User-triggered fetchers (you’ll see these during setup & tools)

User-triggered fetchers often show up when someone is configuring something (verification, feeds, tools). Because the fetch is requested by a user, Google notes these generally ignore robots.txt rules.

User agent When you see it Example
FeedFetcher-Google RSS/Atom feed retrieval (News / WebSub) FeedFetcher-Google; (+http://www.google.com/feedfetcher.html)
GoogleProducer Publisher Center feeds for Google News landing pages GoogleProducer; (+https://developers.google.com/search/docs/crawling-indexing/google-producer)
Google-Read-Aloud Read Aloud (TTS) requests … (compatible; Google-Read-Aloud; +https://support.google.com/webmasters/answer/1061943)
Google-Site-Verification Search Console verification token fetch Mozilla/5.0 (compatible; Google-Site-Verification/1.0)
Google-CWS Chrome Web Store metadata URLs Mozilla/5.0 (compatible; Google-CWS)
Google-NotebookLM NotebookLM user-provided sources Google-NotebookLM
Google-Pinpoint Pinpoint user-specified sources Google-Pinpoint

How to verify “Googlebot” is real (spoofing is common)

User agent strings are easy to fake. Google recommends verifying suspicious requests using reverse DNS and forward DNS checks. If you’re doing this at scale, match IPs against Google’s published IP range JSON files.

1

Reverse DNS lookup

Take the IP from your logs and run a reverse lookup (ex: host 66.249.66.1).

2

Check the domain

Confirm the hostname ends in googlebot.com, google.com, or googleusercontent.com.

3

Forward DNS lookup

Resolve that hostname back to an IP and confirm it matches your original log IP.

4

Scale it (optional)

For automation, compare request IPs to Google’s published crawler IP ranges (JSON lists).

Why it matters: Filtering “Googlebot” by UA string alone can block legitimate Google access while still allowing spoofed bots through.

FAQ: Googlebot crawlers, user agents, and robots.txt

Common questions site owners ask when trying to identify Google crawlers, control crawling, and interpret server logs.

So… how many Googlebot crawlers are there in 2026? +

There isn’t a single fixed “official number.” Google documents multiple crawler and fetcher user agents across Search, Shopping, Ads, abuse/security, and user-triggered tools. The practical approach is to use the documented tokens above (Common / Special-case / User-triggered) and treat the lists as a maintained reference rather than a permanently correct count.

Tip: If your goal is log analysis, focus on tokens and verification—not the total count.
Which Google user agents should I allow in robots.txt? +

Most sites manage crawling by setting rules for Googlebot (and optionally Googlebot-Image / Googlebot-Video). Special-case crawlers (AdsBot, Mediapartners) may behave differently and can ignore the global * group.

If your intent is to manage AI training usage, use Google-Extended as the control token.

Why do I see “GoogleOther” in logs? +

GoogleOther is documented as a generic crawler that may be used by various Google teams to fetch publicly accessible content. It’s not the same as core Search crawling behavior, and it doesn’t imply ranking impact by itself.

How do I know a request is really from Google? +

Verify it. Google recommends a reverse DNS lookup, validate the domain, then forward DNS back to the original IP. For scale, match IPs against Google’s published JSON IP ranges for crawlers/fetchers.

Shortcut: Don’t trust UA strings alone—spoofing is common.
Can Potent Pages help me reduce bot load without harming SEO? +

Yes. We help teams classify bot traffic, verify authentic Google crawling, and implement safer controls (rate limiting, caching, edge rules, monitoring) so legitimate indexing continues while unnecessary load is reduced.

David Selden-Treiman, Director of Operations at Potent Pages.

David Selden-Treiman is Director of Operations and a project manager at Potent Pages. He specializes in custom web crawler development, website optimization, server management, web application development, and custom programming. Working at Potent Pages since 2012 and programming since 2003, David has extensive expertise solving problems using programming for dozens of clients. He also has extensive experience managing and optimizing servers, managing dozens of servers for both Potent Pages and other clients.

Web Crawlers

Data Collection

There is a lot of data you can collect with a web crawler. Often, xpaths will be the easiest way to identify that info. However, you may also need to deal with AJAX-based data.

Development

Deciding whether to build in-house or finding a contractor will depend on your skillset and requirements. If you do decide to hire, there are a number of considerations you'll want to take into account.

It's important to understand the lifecycle of a web crawler development project whomever you decide to hire.

Web Crawler Industries

There are a lot of uses of web crawlers across industries to generate strategic advantages and alpha. Industries benefiting from web crawlers include:

Building Your Own

If you're looking to build your own web crawler, we have the best tutorials for your preferred programming language: Java, Node, PHP, and Python. We also track tutorials for Apache Nutch, Cheerio, and Scrapy.

Legality of Web Crawlers

Web crawlers are generally legal if used properly and respectfully.

Hedge Funds & Custom Data

Custom Data For Hedge Funds

Developing and testing hypotheses is essential for hedge funds. Custom data can be one of the best tools to do this.

There are many types of custom data for hedge funds, as well as many ways to get it.

Implementation

There are many different types of financial firms that can benefit from custom data. These include macro hedge funds, as well as hedge funds with long, short, or long-short equity portfolios.

Leading Indicators

Developing leading indicators is essential for predicting movements in the equities markets. Custom data is a great way to help do this.

Web Crawler Pricing

How Much Does a Web Crawler Cost?

A web crawler costs anywhere from:

  • nothing for open source crawlers,
  • $30-$500+ for commercial solutions, or
  • hundreds or thousands of dollars for custom crawlers.

Factors Affecting Web Crawler Project Costs

There are many factors that affect the price of a web crawler. While the pricing models have changed with the technologies available, ensuring value for money with your web crawler is essential to a successful project.

When planning a web crawler project, make sure that you avoid common misconceptions about web crawler pricing.

Web Crawler Expenses

There are many factors that affect the expenses of web crawlers. In addition to some of the hidden web crawler expenses, it's important to know the fundamentals of web crawlers to get the best success on your web crawler development.

If you're looking to hire a web crawler developer, the hourly rates range from:

  • entry-level developers charging $20-40/hr,
  • mid-level developers with some experience at $60-85/hr,
  • to top-tier experts commanding $100-200+/hr.

GPT & Web Crawlers

GPTs like GPT4 are an excellent addition to web crawlers. GPT4 is more capable than GPT3.5, but not as cost effective especially in a large-scale web crawling context.

There are a number of ways to use GPT3.5 & GPT 4 in web crawlers, but the most common use for us is data analysis. GPTs can also help address some of the issues with large-scale web crawling.

Scroll To Top