All Articles
4 min read

GOOGLE-AGENT: THE WEB'S NEW VISITOR JUST GOT AN IDENTITY

GoogleAI AgentsAXOAgentic BrowsersTechnical SEO
AUTHOR
Slobodan "Sani" Manic

SLOBODAN "SANI" MANIC

No Hacks

CXL-certified conversion specialist and WordPress Core Contributor helping companies optimise websites for both humans and AI agents.

On March 20, 2026, Google quietly added a new entry to its official list of web fetchers. Not a crawler. Not a training bot. An agent.

Google-Agent is the user agent string for AI systems running on Google infrastructure that browse websites on behalf of users. When someone asks an AI assistant to research a product, fill out a form, or compare options across websites, Google-Agent is the thing that actually visits the page. Project Mariner, Google's experimental AI browsing tool, is the first product using it.

This is not Googlebot. Googlebot crawls the web continuously, indexing pages for search. Google-Agent only shows up when a human asks it to. That distinction changes everything about how it operates.

GET WEEKLY WEB STRATEGY TIPS FOR THE AI AGE

Practical strategies for making your website work with AI agents. No spam.

Robots.txt Does Not Apply

Google classifies Google-Agent as a user-triggered fetcher. The category includes tools like Google Read Aloud (text-to-speech), NotebookLM (document analysis), and Feedfetcher (RSS). All of them share one property: a human initiated the request. Google's position is that user-triggered fetchers "generally ignore robots.txt rules" because the fetch was requested by a person.

The logic: if you type a URL into Chrome, the browser fetches the page regardless of what robots.txt says. Google-Agent operates on the same principle. The agent is the user's proxy, not an autonomous crawler.

This is a meaningful departure from how OpenAI and Anthropic handle similar traffic. ChatGPT-User and Claude-User both function as user-triggered fetchers, but they respect robots.txt directives. If you block ChatGPT-User in robots.txt, ChatGPT won't fetch your page when a user asks it to browse. Google made a different call.

Website owners who relied on robots.txt as a universal access control mechanism now have a gap. If you need to restrict access from Google-Agent, you'll need server-side authentication or access controls. The same tools you'd use to block a human visitor.

Cryptographic Identity: Web Bot Auth

The more significant development is buried in a single line of Google's documentation: Google-Agent is experimenting with the web-bot-auth protocol using the identity https://agent.bot.goog.

Web Bot Auth is an IETF draft standard that works like a digital passport for bots. Each agent holds a private key, publishes its public key in a directory, and cryptographically signs every HTTP request. The website verifies the signature and knows, with cryptographic certainty, that the visitor is who it claims to be.

User agent strings can be spoofed by anyone. Web Bot Auth cannot. Google adopting this protocol, even experimentally, signals where agent identity is heading. Akamai, Cloudflare, and Amazon (AgentCore Browser) already support it. Google brings the critical mass.

This matters because the web is about to have an identity problem. As agent traffic increases, websites need to distinguish between legitimate AI agents acting on behalf of real users and scrapers pretending to be agents. IP verification helps, but cryptographic signatures scale better and are harder to fake.

What This Means for Your Website

Google-Agent creates a three-tier visitor model for the web:

  1. Human visitors browsing directly
  2. Crawlers indexing content for search and training (Googlebot, GPTBot, Google-Extended)
  3. Agents acting on behalf of specific humans in real time (Google-Agent, ChatGPT-User, Claude-User)

Each tier has different access rules, different intentions, and different expectations. A crawler wants to index your content. An agent wants to complete a task. It might be reading a product page, comparing prices, filling out a contact form, or booking an appointment.

Here's what to do now:

Monitor your logs. Google-Agent identifies itself with a user agent string containing compatible; Google-Agent. Google publishes IP ranges for verification. Start tracking how often agents visit, which pages they hit, and what they attempt to do.

Check your CDN and firewall rules. If your security tools aggressively block non-browser traffic, Google-Agent may be getting rejected before it reaches your server. Verify that Google's published IP ranges are permitted.

Test your forms and flows. Google-Agent can submit forms and navigate multi-step processes. If your checkout, booking, or contact forms rely on JavaScript patterns that confuse automated systems, agent visitors will fail silently. Semantic HTML and clear labels remain the foundation.

Accept that robots.txt is no longer a complete access control tool. For content you genuinely need to restrict, use authentication. robots.txt was designed for crawlers. The agent era needs different boundaries.

The Hybrid Web Isn't Coming. It's Logged.

A year ago, the idea that AI agents would browse websites alongside humans was a conference talk prediction. Today it has a user agent string, published IP ranges, a cryptographic identity protocol, and an entry in Google's official documentation.

The web didn't split into human and machine. It merged. Every page you publish now serves both audiences simultaneously, and Google just made it possible to see exactly when the non-human audience shows up.

QUESTIONS ANSWERED

What is Google-Agent?

Google-Agent is a user-triggered fetcher that Google added to its official crawler list on March 20, 2026. It identifies requests from AI agents running on Google infrastructure, including Project Mariner, that browse the web on behalf of users. Unlike Googlebot, which crawls continuously in the background, Google-Agent only activates when a person directs an AI assistant to perform a specific task.

Does Google-Agent respect robots.txt?

No. Google classifies Google-Agent as a user-triggered fetcher, not a crawler. Because the request is initiated by a human user, Google treats it like a browser visit rather than an automated crawl. Blocking Google-Agent in robots.txt will typically be ignored.

What is Web Bot Auth?

Web Bot Auth is an IETF draft standard that enables cryptographic verification of AI agents. Each agent holds a private key, publishes its public key in a directory, and signs every HTTP request. Websites can verify the signature to confirm the agent's identity. Google is experimenting with Web Bot Auth using the identity https://agent.bot.goog.

How can I identify Google-Agent traffic in my server logs?

Google-Agent identifies itself through a specific user agent string containing 'compatible; Google-Agent' in the HTTP request header. Google also publishes IP ranges in a user-triggered-agents.json file that can be used for verification, since user agent strings can be spoofed.

NEW TO NO HACKS?

Weekly strategies for making your website work with AI agents. From SEOs, developers, and AI researchers.