All Articles
5 min read

THE AGENT RUNTIME WARS STARTED THIS WEEK

AXOAgent Experience OptimizationAI AgentsOpenAIAgentic BrowsersAgent Runtime
AUTHOR
Slobodan "Sani" Manic

SLOBODAN "SANI" MANIC

No Hacks

CXL-certified conversion specialist and WordPress Core Contributor helping companies optimise websites for both humans and AI agents.

The agent runtime is the new browser layer, and your website is going to be evaluated against the runtime, not against any individual model.

That's a shift web professionals have not yet made. The conversation is still framed around models. Which model writes better. Which one cites more accurately. Which one's API is cheaper this month. The model conversation is loud because new models ship every few weeks and every release is theatrical.

The interesting story is the one underneath it. The foundation is being rebuilt. This week made it impossible to ignore.

GET WEEKLY WEB STRATEGY TIPS FOR THE AI AGE

Practical strategies for making your website work for AI agents and the humans using it. Podcast episodes, articles, videos. Plus exclusive tools, free for subscribers. No spam.

The runtime stack shipped this week

On April 15, Cloudflare shipped Project Think, a new Agents SDK built around durable execution with crash recovery and checkpointing, sub-agents that run as isolated children, persistent sessions with tree-structured messages, and sandboxed code execution running on Dynamic Workers. Within hours of the same day, OpenAI shipped the next evolution of its Agents SDK with native sandbox execution and a model-native harness. Two of the largest infrastructure operators on the web shipped competing answers to the same question, and the question was: how does a long-running AI agent actually run in production.

Then today (April 16) Cloudflare added five more pieces. AI Platform: a vendor-agnostic inference layer that routes models for agents. AI Search: a vector index plus chunking pipeline shipped as a managed product specifically for agent retrieval, competing with Pinecone and Algolia in the agent-side RAG layer rather than with Google AI Mode. Email Service in public beta, designed so agents can use the most universal interface in the world as a channel. PlanetScale Postgres and MySQL inside Workers. And the engineering foundation for hosting very large open-source LLMs like Kimi K2.5 directly on Cloudflare's network.

Sundar Pichai described the same shift a week earlier. On the April 7 Cheeky Pint podcast with Stripe co-founder John Collison, he called Search itself an "agent manager": "A lot of what are just information-seeking queries will be agentic in Search. You'll be completing tasks. You'll have many threads running." Many threads per query is a runtime description of Search. Google's CEO is pointing at the same substrate Cloudflare and OpenAI shipped this week.

If OpenClaw was the agentic web for consumers (a playable demo, an interesting prototype, something to gesture at), this is the agentic web for adults. Durable. Sandboxed. Auditable. The kind of infrastructure you would actually run a business on.

The pattern across all of it is one thing: the runtime. Not the model. Not the consumer chat app. Not the keynote slide. The runtime is the layer where agents are spun up, persisted across hours and days, given filesystem access, given network access, given memory. The runtime is the layer that decides whether an agent's session survives a crash, whether its sub-agents can be reasoned about, whether its code execution is contained.

The wrong question and the new one

Web professionals have spent the last 18 months asking the wrong question. The question was: which AI model should we optimize for. ChatGPT or Claude or Gemini or Perplexity. Whose citations matter more. Whose crawler should we let through. That conversation made sense when the models read your website directly.

They don't anymore. The model reads what the runtime hands it. The runtime fetched your page. The runtime parsed it. The runtime executed (or did not execute) your JavaScript. The runtime resolved your structured data. The runtime negotiated authentication. By the time the model sees anything from your website, it is seeing the runtime's interpretation of it.

The new question, if you take this week seriously, is which agent runtime your website is legible to. Three things to test before next week:

  1. Do your most important endpoints return machine-readable structured responses, or do they only render correctly inside a full browser session?
  2. Is your authentication scoped so an agent acting on a user's behalf can hold a session across multiple calls, or does it only support one-shot human logins?
  3. Does your structured data still mean the same thing if a runtime that did not execute your JavaScript tried to read it?

These are runtime-readability questions. The model has nothing to do with them. The runtime decides whether your answer is even in the model's context window, and the model picks from whatever the runtime hands over.

The web's plumbing is being rebuilt. Every model in the next two years will see your website through one of these runtimes, not directly. Your website's job, starting now, is to be legible to the runtime.

The model conversation will keep happening on conference stages and in keynote slides. The runtime conversation is happening in product changelogs from infrastructure companies. The companies that ship the runtime will decide which websites get reached by AI search and AI commerce. Stop asking which model. Start asking which runtime.

QUESTIONS ANSWERED

What is the agent runtime?

The agent runtime is the layer of infrastructure that runs an AI agent in production. It handles fetching pages, executing JavaScript or not, parsing responses, managing sessions, sandboxing code execution, recovering from crashes, and handing whatever it found to the model. The model itself does not read the web directly; it reads what the runtime gives it.

What did Cloudflare and OpenAI ship on April 15, 2026?

Both companies shipped competing Agents SDKs within hours of each other. Cloudflare's Project Think provides primitives for long-running agents including durable execution with checkpointing, sub-agents, persistent sessions, and sandboxed code execution. OpenAI's next-evolution Agents SDK adds native sandbox execution and a model-native harness for secure long-running agents across files and tools.

How should websites prepare for the agent runtime layer?

Test three things this week. First, whether your most important endpoints return machine-readable structured responses without requiring a full browser session. Second, whether your authentication can hold a session across multiple agent calls rather than only one-shot human logins. Third, whether your structured data still works for a runtime that did not execute your JavaScript.

NEW TO NO HACKS?

Practical strategies for making your website work for AI agents and the humans using it. Read by SEOs, developers, and AI researchers. Exclusive tools, free for subscribers.