How We Grew AI-Referred Traffic 5–6× in 9 Months - Refine Labs

At the start of 2025, we had what most B2B marketing teams have: a decent content library, reasonable search rankings, and essentially zero traffic from AI tools. ChatGPT, Perplexity, Gemini - they were crawling our site, but not surfacing it. We showed up in neither the citations nor the answers.

That changed in April. Not because of an algorithm update. Not because we got lucky. Because we made a deliberate decision to change how we write - specifically, to write the way language models retrieve, not the way search engines rank.

Nine months later, the chart looks completely different. This post walks through exactly what we did, quarter by quarter, and what the data says about what worked.

5–6×
Peak AI session growth
Dec 2025 vs. Jan 2025
4–5×
Sustained lift
Still running Feb 2026
~80%
ChatGPT share
Dominant source throughout
$0
Paid to acquire
Zero paid media involved

The Data First

Here's what AI-referred sessions looked like across the full period. Every bar is one month. The green is ChatGPT. The rest are Perplexity, Gemini, Claude, Copilot, and OpenAI direct.

AI-Referred Website Sessions by Source — Jan 2025 to Feb 2026
ChatGPT
Perplexity
Gemini
Claude
Copilot / Other

The April inflection is not subtle. Before that point - three full months of essentially flat activity. After it, a consistent upward trend that compounds through the rest of the year and peaks in December at approximately 5-6× the January baseline. Importantly, January and February 2026 are still running at 4-5× baseline. This isn't a spike that collapsed. It's a new floor.

Context on the data

We're reading session counts directionally from analytics — exact numbers aren't published here. The growth ratios are what matter. If AI traffic represented roughly 2–3% of total sessions in Q1 2025, our estimate puts it at 8–12% now. That's material, and it happened with no paid investment in this channel.

Q1 2025: What Flat Looked Like

January through March weren't failures - they were baseline. The bars in the chart for those months are barely visible. ChatGPT was sending occasional sessions. Perplexity even less. Gemini essentially nothing. Month-over-month, there was no meaningful growth in any direction.

What this told us: we were not a source LLMs were choosing to cite. Our content was crawlable. It existed. It wasn't bad. But it wasn't structured the way language models retrieve information and that gap was the whole problem.

What we diagnosed in Q1

Our content had low semantic clarity — scattered across many topics rather than deeply owning any specific ones. Pages were written for narrative engagement rather than answer extraction. There were no frameworks, models, or structured breakdowns that LLMs could pull from cleanly. We were optimized for a channel we understood (search) and not for one that was rapidly growing (AI referral).

Q2 2025: The Pivot - April Through June

April is where the work started. The first thing we changed wasn't what we wrote - it was how we thought about why we were writing it.

Traditional SEO is about signals: keyword density, backlinks, domain authority, crawl budget. All of those still matter. But LLM discovery works on a different logic. A language model isn't ranking pages - it's deciding what to cite. It's looking for content that is structurally clear, topically authoritative, and easy to extract a precise answer from.

Once you see that distinction, the content changes you need to make become obvious.

Stop writing for algorithms. Start writing for retrieval. That's the whole strategic pivot — and it's more concrete than it sounds.

What we actually changed

01
Rebuilt content around clear topic ownership
Instead of publishing broadly on B2B marketing, we picked specific themes and went deep: Brand vs. Demand vs. Expand as distinct pillars, GTM maturity models, revenue alignment frameworks. LLMs respond to structured topic authority — not keyword density. Owning a topic means having multiple, interconnected, comprehensive pages on it.
02
Rewrote openings for answer extraction
Every page got a new first 100 words. Instead of narrative warmups or rhetorical questions, we opened with a direct definition: what this topic is, why it matters, and what the page covers. LLMs extract clean sections — they don't read narrative introductions the way humans do. The opening is the single highest-leverage thing you can change.
03
Rebuilt H-tag hierarchies as standalone questions
Every H2 was rewritten to function as a question a buyer might ask. The test: can you read just this section and get a complete answer without surrounding context? If not, it got rewritten. This single change likely drove more LLM citations than anything else we did in Q2.
04
Replaced narrative prose with frameworks
Any concept that could be expressed as a model, stage-by-stage process, or framework was. LLMs cite frameworks constantly — they're extractable, attributable, and memorable. Abstract opinion pieces get ignored. Structured frameworks get cited.

Q3 2025: Building Depth - July Through September

Q2 proved the thesis. Q3 was execution at scale. We shifted from rewriting existing pages to publishing new assets built specifically for LLM retrieval. The goal was to become the definitive reference on specific topics within our space - the source an AI would reach for when someone asked a relevant question.

The cornerstone content push

We published a series of long-form deep dives. Not blog posts - reference documents. Each one structured like a comprehensive guide rather than a thought leadership piece.

Content Asset
Primary AI Audience
Why it gets cited
Brand Deep Dive (3,000+ words)
ChatGPT, Gemini
Comprehensive coverage = clear topic authority signal
Demand Deep Dive (3,000+ words)
ChatGPT, Claude
Framework-heavy structure maps to how LLMs structure answers
GTM Maturity Model
All LLMs
Named frameworks get cited — they're specific and attributable
Revenue Alignment Frameworks
ChatGPT, Claude
Practical and specific — LLMs prefer actionable over abstract
AI for Marketers (101 primer)
All LLMs
Definitional content performs consistently across every AI source

Q4 2025: Compound Growth - October Through December

Q4 is where compounding becomes visible. The cornerstone content from Q2 and Q3 had been indexed, cited, and reinforced. New content published in Q4 picked up traction faster because we already had established authority in these topic clusters. Each new piece benefited from the infrastructure we'd built over the previous six months.

Q1 2025
Baseline — Essentially Zero
AI traffic present but negligible. No month-over-month growth. Content crawlable but not citation-worthy.
AI traffic share: ~2–3% of total sessions
Q2 2025
The Pivot — April Inflection
Content restructured for LLM retrieval. Topic ownership established. ChatGPT traffic roughly doubles from April to June.
Growth vs. Q1: ~2–3× by end of June
Q3 2025
Depth — Cornerstone Publishing
7 deep-dive pages published. Technical structure tightened. Perplexity +3–4×. Gemini and Claude appear for the first time.
Growth vs. Jan baseline: ~3–4×
Q4 2025
Compound — Peak in December
Format diversification unlocks more sources. All channels growing. December peaks at 5–6× January baseline.
December peak: 5–6× January baseline

Source breakdown - where the sessions came from

ChatGPT
Comprehensive deep dives
~80%
Perplexity
Structured citation posts
~12%
Gemini
SEO-indexed educational
~6%
Claude + Copilot
Emerging, growing share
~2%

The Numbers, Laid Out

Comparison Point Estimated Lift Primary Driver
April 2025 vs. January 2025 ~3× Initial content restructuring + topic ownership pivot
ChatGPT alone, May–Aug window ~100–150% Answer-extraction formatting + cornerstone publishing
October 2025 vs. April 2025 ~3–4× Deep-dive pages accumulating citation authority
December 2025 vs. January 2025 ~5–6× Compound effect — all channels elevated simultaneously
February 2026 vs. January 2025 ~4–5× Post-peak floor well above original baseline — sustained
Estimated AI traffic share (Feb 2026) ~8–12% Four sources now contributing meaningfully

Why This Worked

LLMs don't rank pages. They decide what to cite. That's a different optimization target — and most content teams haven't made the switch yet.

There are a few things we'd flag as genuinely explanatory versus coincidental.

Topic ownership compounds. The biggest driver wasn't any single piece - it was the cluster effect. Once we had three or four deeply structured pages on Brand, Demand, and Expand, new content in those areas picked up traction faster. LLMs have a model of what topics different sources are authoritative on. Once you're established in that model, staying there is easier than getting there.

Structure is a feature. The single highest-leverage change we made was treating H-tags as answer containers rather than navigational labels. Every H2 should be able to stand alone as a question and its section as the answer. This one change likely drove more LLM citations than any other tactical edit we made.

Depth beats volume. We published fewer pieces but made each one comprehensive. A 2,500-word deep dive consistently outperforms five 500-word takes on the same topic for LLM citation purposes. The publishing cadence slowed - the citation rate increased.

Technical structure multiplies content quality. All the content work in the world gets discounted if page structure is messy. Clean URLs, proper H-tag hierarchy, logical internal linking - these determine whether LLMs can correctly attribute content to your domain.

The Bottom Line

December was a peak. February is the new floor. That's the difference between a spike and infrastructure.

This started as a hypothesis in April 2025: if we write content the way language models want to retrieve it, they'll cite us more. The data confirmed it. But the more useful takeaway is that this playbook is not exotic or out of reach. It doesn't require a massive budget, a specialized team, or technical knowledge most marketing organizations don't have.

It requires clarity of thought, consistency of execution, and patience to let the compounding happen. We built the infrastructure over 9 months. The chart reflects that.

If you're running content for a B2B company and AI referral traffic is still flat, the gap is almost certainly structural - not creative. The question is whether an LLM can find a clear answer in your content and attribute it to you with confidence. Most sites can't pass that test yet.

Work With Refine Labs
Curious what this could look like for your company?

Refine Labs works with B2B marketing teams to build strategies that drive real pipeline. If the way we think about content and growth feels like a fit, we're happy to have a conversation.

Find out how Refine Labs can help →
No pitch. Just a conversation.


Related posts

View more

Ready to Grow Your Revenue?

Book a Strategy Call