TL;DR — Key Takeaways
- GPT-5.4 sends 56% of citations to brand websites. GPT-5.3 sends only 8% — routing 92% of traffic to Forbes, TechRadar, and Reddit instead.
- The two models share only 7% citation overlap for the same prompts. Being visible on one gives you no advantage on the other.
- GPT-5.4 sends 8.5x more search queries per prompt than GPT-5.3, using site: operators and domain restrictions to verify brands directly.
- 75% of GPT-5.4’s cited domains don’t appear in Google or Bing results for the same query. SEO rankings don’t determine GPT-5.4 citations.
- GPT-5.4 cites pricing pages 35x more than GPT-5.3, making commercial page optimization critical for premium ChatGPT users.
- One-size-fits-all GEO is dead. Model-specific strategy is the new requirement.
The Study That Should Alarm Every GEO Practitioner
On March 9, 2026, Writesonic published a study that should fundamentally change how we think about Generative Engine Optimization. Their team tested 50 prompts across ChatGPT’s two newest models — GPT-5.3 Instant (the default) and GPT-5.4 Thinking (the premium) — and extracted every fan-out query, web search result, and citation from 119 conversations.
The headline finding: GPT-5.4 sends 56% of its citations to brand websites. GPT-5.3 sends only 8%. Same questions. Same search index. Completely different outcomes.
This isn’t an edge case. Across comparison prompts like “HubSpot vs Salesforce vs Pipedrive,” GPT-5.3 cited zero brands — sending users to emailtooltester.com and salesflare.com instead. GPT-5.4 cited the actual brands 100% of the time: hubspot.com, pipedrive.com, salesforce.com.
The implications for GEO strategy are massive. And most practitioners haven’t caught up yet.
Source: Writesonic, “56% of GPT-5.4’s Citations Go to Brand Websites. Only 8% of GPT-5.3’s Do.”, March 9, 2026. Study analyzed 1,161 citations across 119 conversations, 532 fan-out queries, and 7,896 web search results.
Two Models, Two Completely Different Citation Worlds
The most striking finding isn’t the 56% vs 8% split. It’s this: the average citation overlap between GPT-5.3 and GPT-5.4 for the same prompt is just 7%. On 22 of 50 prompts, the overlap was exactly zero.
What does that mean in practice? A brand that dominates GPT-5.3 citations might be completely invisible on GPT-5.4, and vice versa. Any GEO audit that tests only one model is seeing less than half the picture.
Here’s what happens when you ask both models about CRM software:
| What you asked | GPT-5.3 cites | GPT-5.4 cites |
|---|---|---|
| ”Best CRM for B2B SaaS” | designrevision.com, techradar.com | hubspot.com, salesforce.com, attio.com |
| ”HubSpot vs Salesforce vs Pipedrive” | emailtooltester.com, salesflare.com | hubspot.com, pipedrive.com, salesforce.com |
GPT-5.3 sends users to blogs about your product. GPT-5.4 sends them to your actual product. Same question. Opposite outcomes.
This pattern holds across nearly every category tested — SaaS, ecommerce, healthcare, finance, travel, education, fitness. Head-to-head comparison prompts showed the biggest gap: 0% brand citations on GPT-5.3 versus 83% on GPT-5.4.
Why the Split Exists: Different Search Architectures
The citation split isn’t random. It’s architectural. The two models search the web in fundamentally different ways.
GPT-5.3 sends roughly 1 query per prompt — typically the raw user question — and gets about 27 web results. It picks citations from whatever surfaces in that single broad search. Since review sites and media outlets dominate broad search rankings, that’s where 92% of citations come from.
GPT-5.4 sends an average of 8.5 queries per prompt, using domain restrictions and site: operators to target specific brand websites directly. It follows a two-phase pattern:
- Phase 1 — Brand verification: Sends domain-restricted queries directly to brand sites (e.g., “site:klaviyo.com pricing email marketing 2026”)
- Phase 2 — Third-party validation: Checks review platforms like G2, Capterra, and app stores to confirm brand claims
Across 50 prompts, GPT-5.4 sent 304 targeted queries — 142 domain-restricted to brand sites and 156 using site: operators. No other model uses site: operators at all.
| Metric | GPT-5.3 | GPT-5.4 |
|---|---|---|
| Avg queries per prompt | 1.0 | 8.5 |
| Avg web results | 27.3 | 109.4 |
| Avg citations per response | 5.8 | 14.8 |
| Avg response length | 548 words | 769 words |
| Domain-restricted queries | 0 | 142 |
| site: operator queries | 0 | 156 |
Same search index, different decomposition. The fan-out strategy is the entire difference.
Source: Writesonic citation study methodology, March 2026.
The “Kingmaker” Problem on GPT-5.3
Because GPT-5.3 routes 92% of citations to third-party sites, a small number of media domains become gatekeepers for AI visibility:
| Domain | Citations | Role |
|---|---|---|
| forbes.com | 15 | Media/reviews |
| techradar.com | 10 | Tech reviews |
| tomsguide.com | 10 | Tech reviews |
| reddit.com | 7 | Forum/UGC |
| money.com | 5 | Finance media |
If Forbes or TechRadar hasn’t written about your product, you’re likely invisible to the default ChatGPT model. Your GEO strategy for GPT-5.3 isn’t about optimizing your website — it’s about getting covered by these kingmaker publications.
GPT-5.4’s top cited domains? The brands themselves: hubspot.com (18 citations), shopify.com (16), salesforce.com (14), quickbooks.intuit.com (10).
Two different visibility strategies for the same platform. That’s the new reality.
GPT-5.4 Cites Pricing Pages 35x More
The models don’t just cite different websites — they cite different types of pages.
| Page type | GPT-5.3 | GPT-5.4 |
|---|---|---|
| Blog/article pages | 92 (32%) | 61 (8%) |
| Pricing pages | 4 (1%) | 138 (19%) |
| Homepage/root | 42 (15%) | 161 (22%) |
| Product/feature pages | 13 (5%) | 73 (10%) |
GPT-5.3 is a “blog reader” — one-third of its citations point to articles and blog posts. GPT-5.4 is a “pricing page checker” — 51% of its citations land on commercial pages (pricing + homepage + product pages combined).
The implication: if your pricing page says “contact sales” instead of showing actual numbers, GPT-5.4 will find the gap. Transparent, detailed pricing pages are now a GEO asset.
4 pricing page citations on GPT-5.3 across 49 conversations. 138 on GPT-5.4. That’s a 35x difference.
Source: Writesonic page-type classification across 1,023 citations, March 2026.
Google Rankings Don’t Predict GPT-5.4 Citations
Here’s where it gets really interesting for SEO professionals.
Writesonic checked whether GPT-cited domains also appeared in Google or Bing results for the same queries. The results were dramatically different by model:
GPT-5.3:
- 47% of cited domains also rank on Google
- 27% also appear on Bing
- 44% don’t appear on either search engine
GPT-5.4:
- Only 25% of cited domains appear on Google or Bing
- 75% don’t appear in any traditional search results for the same prompt
Why? Because GPT-5.4 doesn’t discover brands through search. It already knows them from training data. It then sends targeted queries directly to those brand websites to verify current information.
When you ask GPT-5.4 about running shoes, it doesn’t search “best marathon running shoes” and hope Nike ranks. It sends domain-restricted queries to nike.com, asics.com, and hoka.com directly.
The takeaway: for GPT-5.3, invest in SEO. For GPT-5.4, invest in brand recognition and first-party content quality. Traditional rankings are almost irrelevant to the premium model.
This Is Happening Against a Bigger Backdrop
The citation split matters more when you zoom out.
AI assistants now generate roughly 45 billion monthly sessions, equaling approximately 56% of global search engine volume — and 34% in the U.S. Much of this growth comes from mobile apps like ChatGPT, Gemini, and Perplexity.
Source: Search Engine Land, “AI assistants now equal 56% of global search engine volume: Study”, March 9, 2026.
Meanwhile, Google AI Overviews now appear in 25.11% of all Google searches, up from 13.14% in March 2025, based on Conductor’s analysis of 21.9 million queries. AI Overviews reduce clicks to underlying websites by 34.5%.
Source: Conductor 2026 AEO/GEO Benchmarks; SE Ranking AI Statistics.
And a critical conversion stat: AI-driven visitors convert at 4.4x the rate of standard organic visitors. LLM visitors convert at 2x the rate in one-third of the sessions.
Source: Semrush AI Search SEO Traffic Study; Conductor/Knotch 2026.
So the traffic from AI citations isn’t just growing — it’s more valuable per visit than traditional organic. And which model your visitors use determines whether that traffic lands on your site or on Forbes.
The Attribution Problem No One Is Solving
GPT-5.4 appends utm_source=chatgpt.com to 87% of cited URLs. Combined with its 56% first-party citation rate, that means roughly 49% of all GPT-5.4 citation traffic is trackable in GA4.
| Model | First-party rate | UTM coverage | Trackable brand traffic |
|---|---|---|---|
| GPT-5.3 Instant | 8% | 96% | ~8% of citations |
| GPT-5.4 Thinking | 56% | 87% | ~49% of citations |
Source: Writesonic UTM analysis across 119 conversations, March 2026.
This is a genuine attribution breakthrough — for the first time, a thinking model makes AI search attribution comparable to paid search.
But here’s the gap nobody’s talking about: the UTM only captures the last click. It doesn’t capture:
- Dark traffic — users who read the AI answer, remember the brand name, then open a new tab and type the URL directly. This shows as “Direct” in GA4.
- Branded search lift — users who see the brand cited, then Google the brand name later. This shows as “Organic” in GA4.
- Cross-model attribution — a user might see your brand cited in GPT-5.3 (via a Forbes article), then ask GPT-5.4 the same question later and click through to your site. The UTM credits GPT-5.4, but GPT-5.3 started the journey.
Industry estimates suggest 40-60% of AI-influenced website visits are “dark” — invisible in analytics because users don’t click the citation link directly. They absorb the recommendation and act on it later through other channels.
Every GEO tool on the market tracks citations. None of them answer the question that actually matters: “How much revenue did those citations drive?”
For more on how to measure what matters, see our guide on The ROI of GEO: Calculate AI Search Value and AI Citations ≠ Traffic: The Citation Gap.
What This Means: Model-Specific GEO Is Now Required
The era of “optimize for AI search” as a single strategy is over. Here’s what the data demands:
For GPT-5.3 (the default model — most users)
GPT-5.3 relies on broad search rankings and third-party authority. Your strategy:
- Get featured on kingmaker sites — Forbes, TechRadar, Tom’s Guide, Reddit. If they haven’t reviewed your product, pitch them.
- Invest in traditional SEO — 47% of GPT-5.3 citations correlate with Google rankings. Strong organic presence feeds the default model.
- Create comparison and review content — GPT-5.3 loves third-party reviews. Encourage customers to post on G2, Reddit, and industry forums.
- Don’t expect brand website traffic — 92% of citations go elsewhere. Measure GPT-5.3 success by brand mentions, not clicks.
For GPT-5.4 (premium model — high-intent users)
GPT-5.4 bypasses search rankings entirely. Your strategy:
- Make your website the definitive source — GPT-5.4 sends domain-restricted queries directly to your site. If the answer isn’t there, you get skipped.
- Publish transparent pricing — 138 pricing page citations vs 4 on GPT-5.3. Show real numbers.
- Optimize product and feature pages — 10% of GPT-5.4 citations go to product pages (vs 5% on GPT-5.3). Detailed specs, comparisons, and use cases matter.
- Build brand recognition — GPT-5.4 selects brands from training data before searching. If your brand isn’t in the consideration set, no amount of SEO helps.
- Set up GA4 tracking now — create a segment for
utm_source=chatgpt.com. As GPT-5.4 adoption grows, you’ll see this traffic appear.
For your GEO audits
If your GEO tool or audit process only tests one model, you’re seeing less than half the picture. Model-specific testing is no longer optional — it’s the baseline for credible GEO analysis.
For detailed guidance on each platform’s citation mechanics, see our guides on How to Get Cited by ChatGPT, How to Get Cited by Perplexity AI, and How Microsoft Copilot Chooses What to Recommend.
The Bigger Strategic Question
The GPT citation split reveals something uncomfortable about the current GEO landscape: most GEO tools are built for a world where AI search is one channel with one set of rules. That world doesn’t exist anymore.
When citation overlap between two models on the same platform is 7%, tracking “total citations” as a single metric becomes almost meaningless. Knowing you were cited 47 times is useless if you can’t answer: by which model, to which user segment, driving what behavior?
The GEO tools market — currently valued at $848 million and projected to reach $33.7 billion by 2034 at a 50.5% CAGR — is building dashboards for a simplified version of reality. The brands that win will be the ones that see the complexity clearly and adapt their strategy to each model’s specific mechanics.
Source: Dimension Market Research, GEO Market Report, 2025.
One-size-fits-all GEO is dead. Model-specific GEO is the new standard.
FAQ
Do GPT-5.3 and GPT-5.4 cite the same sources?
No. Writesonic’s study of 1,161 citations found only 7% overlap between the two models for the same prompts. On 22 of 50 test prompts, the overlap was exactly 0%. Being visible on one model gives you no advantage on the other.
Which GPT model is better for brand websites?
GPT-5.4 (Thinking) sends 56% of citations to brand websites. GPT-5.3 (Instant), the default model, sends only 8% to brands and routes 92% of traffic to third-party sites like Forbes, TechRadar, and Reddit.
Does Google ranking help you get cited by ChatGPT?
It depends on the model. 47% of GPT-5.3 citations come from domains that rank on Google. But 75% of GPT-5.4 citations come from domains that don’t appear in Google or Bing results for the same query. GPT-5.4 finds brands from training data, not search rankings.
What is model-specific GEO?
Model-specific GEO is the practice of optimizing content differently for each AI model, rather than treating all AI search as one channel. Each model has different citation logic, search architecture, and source preferences. The GPT-5.3 vs 5.4 citation split proves this approach is necessary.
Can you track traffic from ChatGPT citations?
GPT-5.4 appends utm_source=chatgpt.com to 87% of cited URLs, making nearly half of all citation traffic trackable in GA4. GPT-5.3 adds UTM to 96% of citations, but since only 8% go to brand sites, most trackable traffic comes from GPT-5.4.