The attribution problem nobody has solved
94% of B2B buyers use LLMs during purchase research. AI referrals account for roughly 1% of tracked website traffic. The gap between those two numbers is where pipeline disappears.
April 2026
The gap
The 6sense 2025 Buyer Experience Report surveyed 4,510 B2B buyers and found 94% used an LLM during their last purchase. Conductor and BrightEdge independently estimate that AI-attributed traffic accounts for roughly 1% of tracked website visits.
Both numbers are credible. The distance between them is the problem. Buyers are using AI constantly during research. Almost none of that usage shows up in analytics. And attribution is how marketing leaders get budget: a channel you can measure gets funded, a channel you cannot measure gets questioned. The fastest-growing influence on B2B buying is the one CMOs are least equipped to justify spending on.
For the technical explanation of why AI platforms don't generate visits, see why your organic traffic is down. The rest of this piece focuses specifically on why the influence AI does exert can't be traced back to its source.
Why the gap is structural
The attribution gap is not a tooling limitation that better software will fix. It is a consequence of how AI platforms interact with web content.
No JavaScript execution. GA4, and virtually every analytics platform in use today, depends on client-side JavaScript to register a session. When ChatGPT, Claude, Gemini, or Perplexity serve content to a user, they pull from either the model's training data or a retrieval cache built by crawlers. Those crawlers fetch raw HTML without executing JavaScript. No JS execution means no analytics tag fires, no session is recorded, and no conversion event is tracked.
No referrer headers. When a buyer clicks a link inside an AI chat response, the referrer header is often stripped or unrecognized by GA4. Workshop Digital analyzed 181.6 million GA4 sessions and found 22% of ChatGPT sessions were misclassified as "(not set)" and 32% of Perplexity sessions ended up in the same bucket. AI mobile apps are particularly inconsistent about passing referral data. Loamly found 70.6% of identifiable AI traffic landed as "Direct" in GA4.
No cookies. Every layer of marketing measurement that depends on browser cookies is absent from AI chat interactions. Session tracking, retargeting pixels, conversion attribution, audience segmentation, all of it requires a cookie jar maintained by a browser. When a crawler fetches your page, no cookies are set. When a user reads your content inside a ChatGPT response, no cookie jar exists for that interaction. The user is consuming your information in an environment that has no concept of the tracking infrastructure marketers rely on.
Temporal separation between influence and visit. Even when AI chat does lead a buyer to your site, the influence and the visit are separated in time. A buyer researches in ChatGPT on Tuesday, learns about your product, and Googles your brand name on Thursday. GA4 records an organic session. A buyer reads a Perplexity summary, forms an impression, and types your URL directly a week later. GA4 records a direct visit. In both cases the AI interaction created the awareness, but the visit that resulted from it carries no trace of its origin. The attribution goes to whatever channel delivered the final click.
Deep research agents widen the gap
A growing category of AI usage involves agents that conduct extended, multi-step research autonomously. OpenAI, Google, and Perplexity all ship "deep research" features that browse hundreds of web pages over minutes, cross-reference findings, and produce structured reports with citations. A single deep research query can touch 100-200+ pages across dozens of domains.
These agents operate primarily on cached content, the same way the consumer chat platforms described in the previous section do. The buyer reads the synthesized report and forms a shortlist. They may never visit any of the cited websites. If they do, they arrive via brand search or direct URL with no referral trail. The research that shaped their decision happened entirely inside the agent's session, invisible to every analytics tool on the cited sites.
Browser-based agents partially change the picture
Not all agents operate on cached content. A growing number of agentic systems run real browsers: ChatGPT's Agent Mode, Claude in Chrome, Manus, OpenClaw, Perplexity Computer. When an agent operates a full browser, it executes JavaScript, which means analytics tags fire. From the perspective of your analytics, a browser-based agent visiting your site looks much more like a human visit than a traditional crawler does.
But not all browser-based agents are equal for attribution. Some, like Claude in Chrome, run inside the user's own browser with the user's real cookies and auth state, which means session tracking and conversion attribution work normally. Others, like ChatGPT Agent Mode and Manus, run sandboxed browsers hosted in the cloud that start fresh each session with no persistent cookie state. They fire your analytics tags, but retargeting and cross-session attribution do not carry over.
Industry telemetry shows agentic traffic up 6,900% year-over-year in 2025. The volume of agent-driven page loads is growing rapidly, and a meaningful portion of that growth is coming from browser-based systems that do register in analytics.
But browser-based agents are not a single category. Some take screenshots of pages and use visual understanding to navigate. Others parse the DOM directly. Some send identifiable request headers; many do not. Without the headers, distinguishing a browser-based agent from a human visitor requires sophisticated detection that most analytics setups are not built for. If agent traffic detection is relevant to your situation, talk to us.
The full landscape of agent types, how each interacts with web pages, and what the differences mean for site owners is covered in detail in how AI systems actually browse the web.
Why the attribution gap is different from previous dark funnels
Marketing has always had untrackable channels: word of mouth, events, Slack communities, podcast mentions. The standard response is that dark funnels are not new.
Previous dark funnel channels were additive. They generated awareness on top of a measurable search and content infrastructure that remained intact. AI is substitutive. Conductor found 93% of Google AI Mode searches end without a single website click. The measurable base of organic search traffic is shrinking. The channel replacing it produces no attribution data. The old system was imperfect but functional. The replacement, so far, is invisible.
McKinsey's October 2025 research found that just 16% of brands systematically track AI search performance.
What measurement exists
No single tool covers the full picture. Each approach captures something and misses something else.
GA4 referral tracking captures visits where AI referrer headers are present. Claude and Gemini attribute correctly in most cases. But 22% of ChatGPT sessions are misclassified, 70.6% of AI traffic lands as "Direct," and no JS-free visits are recorded at all.
Microsoft Clarity bot activity provides server-side detection of AI crawler requests, categorized by operator (AI Assistant, AI Crawler, AI Search). But Clarity shows crawling, not citation or recommendation. Microsoft's own disclaimer states that bot activity data "doesn't indicate content was retrieved, cited, or surfaced in AI responses."
Server log analysis captures all HTTP requests, including JS-free agent visits identifiable by user-agent strings (GPTBot, ClaudeBot, PerplexityBot). But some agents do not identify themselves, and there is no connection between a crawl event and a downstream citation or conversion.
AI visibility monitoring (like OpenLens) captures what ChatGPT, Claude, Gemini, and Perplexity say about your brand when asked category-level questions. But the data is probabilistic: Rand Fishkin found ChatGPT has less than a 1-in-100 chance of returning the same brand list for similar prompts (SparkToro, Jan 2026). Directional trends over hundreds of queries are meaningful. Individual spot checks are not.
Brand vs. non-brand search segmentation captures whether brand awareness is growing even as non-brand organic declines. But it cannot attribute the cause. Brand growth could come from AI, events, word of mouth, or anything else outside tracked channels.
In practice, covering the full picture means combining several of these: server-side agent traffic detection, AI visibility monitoring across platforms, and brand search analysis. We work on this problem. If any of it is relevant to your situation, talk to us.
The fingerprint
If your non-brand organic search traffic is declining while your brand search and direct traffic remain stable or grow, the pattern is consistent with AI replacing the informational queries that used to bring people to your site while simultaneously creating brand recognition that drives direct visits. Andy Crestodina at Orbit Media and the Discovered Labs "Dark Search Funnel" framework have both documented the pattern as a proxy signal.
Correlation is not attribution. But it is the closest thing to a signal that standard tools produce today.
Where agents might help
There is an irony in the attribution problem. The same AI agents that make buyer research invisible today could, in a structured interaction model, make it more visible than traditional web traffic ever was. When an agent interacts with your website through a declared protocol like WebMCP, the interaction is structured, logged, and attributable. The agent calls specific tools and receives specific responses. A tool call like "check pricing" is an implicit declaration of intent. A lead form submitted through a declared tool carries the agent's identity. That is richer attribution data than a human clicking through Google ever produced.
The transition from agents that silently scrape and synthesize to agents that interact through declared interfaces is the transition from invisible influence to measurable interaction. That transition depends on standards adoption that is still early (see our analysis of how agents browse the web and what WebMCP is). But the endgame of the agent era is not permanent attribution darkness. It is a different, potentially better, attribution model for the businesses that build for it.
Sources
- 6sense, "2025 B2B Buyer Experience Report" (Nov 2025, n=4,510). 6sense.com
- Conductor, "2026 AEO/GEO Benchmarks Report" (Nov 2025, 3.3B sessions, 13K+ domains). BusinessWire
- BrightEdge, "AI Search Visits Surging in 2025" (Sep 2025). brightedge.com
- Workshop Digital, "The AI Referral Gap" (181.6M GA4 sessions). workshopdigital.com
- Loamly, "The AI Traffic Attribution Crisis" (446,405 visits). loamly.ai
- HUMAN Security, AI agent traffic telemetry (6,900% YoY), 2025
- Microsoft Clarity, "AI Bot Activity" (Jan 2026). learn.microsoft.com
- Rand Fishkin, AI Recommendation Consistency (Jan 2026). sparktoro.com
- McKinsey, "New Front Door to the Internet" (Oct 2025). mckinsey.com
- Andy Crestodina / Orbit Media, branded search proxy methodology (2025-2026). orbitmedia.com
- Discovered Labs, "The Dark Search Funnel" (2025-2026). discoveredlabs.com