222: AI VISIBILITY IS A VANITY METRIC WITH WIL REYNOLDS

WIL REYNOLDS
Seer Interactive
Founder of Seer Interactive and VP of Innovation, known for pioneering big data approaches to SEO and coining 'real company shit' as a marketing philosophy.
Seer Interactive lost 80% of its organic traffic over two years. The pipeline went up anyway. Wil Reynolds discovered his team had been tracking the wrong metrics for years, chasing traffic numbers while social channels converted at 5x the rate of organic search. The correlation between search traffic and revenue has broken for many businesses, and most marketers haven't noticed yet.
This breakdown reveals something uncomfortable about the AI visibility gold rush. When ChatGPT doubled the length of its answers, visibility scores jumped without a single additional human seeing your brand. Wil found that 44% of LLM users include brand names in their prompts, meaning the optimization target was never generic category queries. The real moat isn't showing up in AI answers. Michelin built a restaurant guide in 1900 that still drives pricing power today. Trust compounds. Hacks mortgage your brand.
KEY TAKEAWAYS
- Track AI visibility against pipeline metrics, not in isolation. When ChatGPT doubled answer length, visibility scores increased without any additional humans seeing your brand.
- Focus optimization efforts on brand comparison prompts rather than generic category queries. 44% of observed LLM users include specific brand names when searching.
- Audit your business phone numbers across AI platforms immediately. LLMs hallucinate contact information, and scammers are exploiting this vulnerability.
- Build trust through people, not listicles. Conference speaking, team thought leadership, and engaged newsletter audiences create belief that AI answers cannot replicate.
- Accept the 16-month training data lag as reality. Work done since January 2025 has zero provable impact on Gemini's training-data-based answers.
SHOW NOTES
The Traffic Correlation Has Broken
Seer Interactive's marketing team went from four people to zero over 18 months. Wil Reynolds found himself sending company tweets while watching organic traffic crater by 80%. Then something unexpected happened: the pipeline went up. Social traffic converted at five times the rate of organic search. Branded search increased. Direct traffic climbed. The leading indicator everyone trusted for two decades stopped leading anywhere useful.
The bifurcation appeared in the data around the same time ChatGPT gained traction. Traffic and pipeline had tracked together for years, then suddenly split in opposite directions. Wil overlaid the metrics and realized the correlation that built careers in SEO had fundamentally broken.
Why AI Visibility Numbers Lie
ChatGPT doubled the length of its answers in November 2024. Every AI visibility tracking tool showed improvements. But did cash registers ring more often? The same number of humans saw longer answers containing more brand mentions. Visibility scores climbed while actual exposure remained flat.
Google's AI Overviews compound the measurement problem. Clicking a brand link in an AI Overview triggers a Google search for that brand rather than visiting the company website. Google's traffic numbers stay healthy because every click generates a new search query. The 17 brand links Wil tested for time tracking software sent 15 of them back to Google search results.
This creates perfect conditions for selling vanity metrics. Sophisticated buyers learned decades ago that rankings meant nothing without traffic, and traffic meant nothing without conversions. The same lesson applies to AI visibility, but the pressure on marketing budgets has increased dramatically. When search represents 30% of spend and craters overnight, executives don't want to hear that everyone is still learning.
Real Humans Use Brands in Prompts
Seer's team watched actual people interact with LLMs rather than relying on scraped data and proxy metrics. The finding that 44% of users include brand names in their prompts changes the optimization target entirely. People aren't asking "what's the best project management software" as often as marketers assume. They're asking "how does Asana compare to Monday" or "is Notion better than Confluence for documentation."
Head-to-head brand comparisons represent the prompts worth tracking. Generic category queries generate visibility scores but rarely generate customers.
The Training Data Reality Check
Gemini 3.1 launched with training data from January 2025. Every optimization effort since then has zero provable impact on answers generated from training data. Marketers cannot call Sundar Pichai and request model updates because a client is upset. This 16-month experimental lag means organizations are running optimization programs without any feedback mechanism.
The three types of AI search require different approaches: search-led systems use web indexes plus AI, answer-led hybrids employ tool usage and web access, and fully generative systems rely entirely on training data. Most marketers treat all AI answers identically.
Phone Numbers and Fraud Exposure
LLMs hallucinate phone numbers for businesses. The wrong number appearing in an AI answer creates immediate fraud risk. Financial services companies targeting elderly customers face particularly dangerous exposure when scammers answer calls intended for legitimate advisors. This security vulnerability receives far less attention than visibility metrics.
Trust Compounds, Hacks Mortgage Your Brand
Michelin built a restaurant guide in 1900 that still drives foot traffic and pricing power 125 years later. RAMP launched 52 AI-generated restaurant pages in a single day. Both involve food content. One represents rotisserie chicken prepared by a chef. The other resembles a chicken nugget. Only one builds lasting trust.
The framework Wil proposes moves through three stages: be seen, be believed, be chosen. Visibility gets brands into the room. But when prospects search for your team members and find no shared content, no conference appearances, no engaged newsletter readers, belief collapses. Trust transfers from people who demonstrate expertise publicly. Listicles don't create belief. The humans behind the brand do.
WATCH ON YOUTUBE
QUESTIONS ANSWERED
Is AI visibility a vanity metric?
According to Wil Reynolds, AI visibility becomes a vanity metric when tracked in isolation from business results. When ChatGPT doubled answer length in November 2024, visibility scores increased without additional human exposure. Wil recommends tracking AI visibility against pipeline metrics to avoid false confidence in marketing performance.
How much do brand names appear in LLM prompts?
Seer Interactive's research found that 44% of observed LLM users include specific brand names in their prompts. This finding suggests marketers should optimize for head-to-head brand comparisons rather than generic category queries, as users often ask questions like 'how does Asana compare to Monday' rather than 'what is the best project management software.'
What is the training data lag for Gemini AI?
Gemini 3.1 launched with training data from January 2025, creating a 16-month gap between current optimization work and measurable impact on training-data-based answers. Wil Reynolds notes this means marketers cannot determine whether recent optimization efforts help or hurt AI-generated responses that rely on training data rather than live web access.
Why is organic search traffic dropping but revenue staying stable?
Seer Interactive lost 80% of organic traffic over two years while pipeline increased. Wil Reynolds discovered social traffic converted at 5x the rate of organic search, and branded search plus direct traffic both grew. The correlation between search traffic and revenue has broken for many businesses since ChatGPT's emergence, making traffic a less reliable leading indicator.
Can LLMs give wrong phone numbers for businesses?
Yes, LLMs hallucinate phone numbers for businesses, creating fraud exposure. Scammers can exploit incorrect phone numbers appearing in AI answers to intercept calls intended for legitimate companies. Wil Reynolds identifies this as particularly dangerous for financial services companies targeting elderly customers who may trust AI-provided contact information.
What are the three types of AI search engines?
Seer Interactive identified three types: search-led systems that combine web indexes with AI like Perplexity, answer-led hybrids that use tool access and web browsing, and fully generative systems that rely entirely on training data with no web access. Each type requires different optimization approaches because the answer sources differ fundamentally.
RELATED ARTICLES
GOOGLE-AGENT: THE WEB'S NEW VISITOR JUST GOT AN IDENTITY
On March 20, 2026, Google added Google-Agent to its official fetcher list. It ignores robots.txt, signs requests with cryptographic keys, and browses websites on behalf of users. The hybrid web is no longer theoretical.
YOUR WEBSITE IS A SOURCE, NOT A MEGAPHONE
Your website used to be the destination. Now it's the source. AI agents summarize, recontextualize, and relay your content without your layout, fonts, or scroll animations. If your message only works inside your design, you don't have a message. You have a brochure.
ANSWER ENGINE OPTIMIZATION: HOW TO GET YOUR CONTENT INTO AI RESPONSES
A practical guide to Answer Engine Optimization (AEO). How AI search engines parse content, what gets cited, and what Google, Microsoft, and OpenAI actually recommend.
ENJOYING THIS EPISODE?
Get weekly web strategy tips for the AI age. From SEOs, developers, and AI researchers.
