← Back to blog🧪Experiments

I Ran 8 Queries Across ChatGPT, Perplexity, Claude, Gemini, and AI Overviews to See What Content Gets Cited.

by Kamila Olexa

Everyone's talking about optimizing for "AI search." But there is no single AI search.

ChatGPT, Perplexity, Claude, Gemini, and AI Overviews each decide differently whether to search, what to read, and what to cite. Some queries don't even trigger a search: the model answers from training data, but here you are betting on random content.

So I ran a little experiment to answer a simple question: when you ask each platform a query with a specific intent (commercial, advisory, comparison, navigational) what's the search logic, and what content types get cited?

I ran 8 queries spanning commercial, advisory, comparison, data, definitional, and navigational intent across all 5 platforms. Same queries, same day, fresh conversations.

⚠️ Disclaimer
This isn't a rigorous study but a practitioner experiment that could be outdated with the next model update. Confirmation bias is real. I'd love to see others replicate this with different queries and different verticals.

But I wanted to entertain a thought, and the patterns pointed to something most content strategies are missing entirely: the platform you're targeting determines whether your content type can even be cited.

What you'll learn today:

  • Which content intent types trigger search on which platforms (and which get ignored completely)
  • The one content strategy that worked across all 5 platforms
  • A step-by-step method to run this audit for your own queries and build a content-to-platform map

The Setup

The 8 queries I picked:

  1. "best CRM for startups" commercial broad
  2. "Asana vs Monday vs ClickUp" commercial comparison
  3. "Asana pricing 2026" commercial brand
  4. "best tools to reduce churn in SaaS" commercial niche
  5. "how to reduce churn in SaaS" advisory/strategy
  6. "SaaS churn benchmarks 2026" data/research
  7. "what is a CRM" definitional
  8. "Stripe billing documentation" navigational

Each query ran in a fresh conversation on every platform's web UI. No conversation history, no priming. I captured every source panel. Same queries, same day.

The platforms: ChatGPT, Perplexity, Claude, Gemini, and Google AI Overviews.

LLMs Don't Agree on Whether to Search

This was the biggest surprise. I expected different sources across platforms. I didn't expect them to make fundamentally different decisions about whether to search at all.

Search trigger comparison across 5 AI platforms

AI Overviews showed a featured snippet or pure organic results instead.

What I'd loosely characterize as each platform's approach:

Perplexity always searches (8/8). It grounds every single answer with web sources, no matter what.

Claude and Gemini skip definitional and advisory queries (6/8). They seem to reason, "I'm confident enough to answer this from what I know."

AI Overviews skips navigational and precise data queries (6/8). Google already has direct links and featured snippets for those.

ChatGPT is the most selective (5/8). It skips advisory, definitional, and navigational queries. It only searches when it thinks the information might be outdated.

Why pay attention to this?

If a platform doesn't search for your query type, your content cannot be cited.

"How to reduce churn in SaaS" strategy content? Invisible on ChatGPT, Claude, and Gemini. They answer from training data. But it absolutely gets cited on Perplexity and AI Overviews.

A Note on API vs. Web UI

Before comparing platforms, I tried to understand how ChatGPT's search actually works internally. I went through three approaches, and honestly, it's worth sharing because it shaped how I ran the rest of the experiment.

1. Browser DevTools ❌

Made a console script to intercept ChatGPT's internal API stream and extract the search queries it fires. It mostly failed: the script couldn't reliably capture the data, and what it did catch showed a single simple query ("best crm") with only 3 cited sources.

Boring, next.

Browser DevTools attempt to intercept ChatGPT search queries

2. OpenAI API ❌

This worked better. The API exposes the sub-queries ChatGPT generates. For "best CRM for startups," the API didn't just search the phrase, it sent:

  • "best CRM for startups 2026 official pricing HubSpot Zoho Pipedrive Freshsales Monday Salesforce Starter official docs"
  • "startup CRM comparison 2026 official plans HubSpot Zoho Pipedrive Freshsales Monday Salesforce"

Brand names were already baked into the queries. ChatGPT didn't search to discover which CRMs existed. It searched to validate and update. BUT the API only returned 2 cited sources for that query.

OpenAI API exposing ChatGPT sub-queries with brand names baked in

3. Web UI 🎉

This is what I settled on. The web UI returned 25-30+ sources, and more importantly, it's what users usually opt for. I didn't find it productive to base content strategy on API behavior, since it doesn't match the product people use.

💡 API vs Web UI mismatch
I can't confirm the web UI sends the same brand-embedded sub-queries as the API. They might be entirely different systems. So everything that follows is based purely on what's observable in the web UI: source panels, citation counts, and search triggers.

How Each Platform Searches (IF It Does.)

ChatGPT: Three Layers in One Answer

ChatGPT source panel showing 25+ sources

ChatGPT's web UI shows the most sources of any platform (25-30+ per query). But what makes it interesting is that a single response blends 3 systems:

  1. The model's own knowledge. In the API, we could see that ChatGPT's search queries already contained brand names (HubSpot, Salesforce, Zoho, Pipedrive) before any search happened. The web UI may work differently, but the API suggests the model already has opinions about which brands matter before it searches. So it's more of a confirmation search: validating and updating what it already knows, instead of discovering from scratch.
  2. Web search. Current details get pulled from live sources and are cited inline. When search triggers, the source panel is massive.
  3. UI enrichment. When the model mentions a well-known product, ChatGPT shows a knowledge card alongside the answer. These cards have their own sources: official websites, app store listings, LinkedIn company pages, review platforms (Trustpilot), developer docs, and startup directories.

This means that even on queries where ChatGPT seemingly doesn't search, brands can still get visibility through knowledge cards, but the content that feeds them is quite different from the content that feeds search results. So, there are essentially 3 different opportunities to get in.

When search triggers, the source types skew heavily toward listicles and vendor comparison content.

Search pattern:
Hard to map as its thinking process isn't visible.

Perplexity: Three-Step Progressive Research Chain

Perplexity progressive research chain

Perplexity (in Pro mode) runs a visible 3-step process:

  1. Broad discovery searches the query as-is
  2. Specific comparison narrows to top contenders
  3. Niche deep-dive tailors to specific use case scenarios

Perplexity 3-step search pattern

Search pattern:

Perplexity search pattern details

Claude: Single Query, Transparent Reasoning

Claude source panel

Claude runs one search with one query. Also has visible thinking, so you can see it decide:

"This is a current topic where recommendations may have changed, so I should search for up-to-date information."

"Simple factual question, no need for tools."

Its source mix looks like traditional search results. No Reddit, no YouTube, no Wikipedia. But it uniquely cited interactive tools (CostBench calculator, PM Toolkit calculator) that no other platform surfaced.

Claude's decision logic maps to "has this changed recently?" If your content is about something that evolves (pricing, tools, benchmarks), Claude will search for it. If it's stable knowledge (definitions, strategy), it won't.

Claude source types

Search pattern:
Claude’s decision logic maps to “has this changed recently?” If your content is about something that evolves (pricing, tools, benchmarks), Claude will search for it. If it’s stable knowledge (definitions, strategy), it won’t.

Claude search decision pattern

Gemini: Reasoning-First, Wildly Variable

Gemini source panel

Gemini in Pro mode gave 2 sources for "best CRM for startups" and basically answered from reasoning. But for pricing queries, it went to 20+ sources and 15+ for benchmarks.

Gemini source types

Search pattern:
Its thinking chain is the most detailed, with 3-4 reasoning steps, naming tools, and categories before showing results.

Gemini search pattern

⚠️ Gemini mode inconsistency
I switched from Pro to Thinking mid-testing due to rate limits. Thinking mode produces dramatically more sources. This is a real variable worth acknowledging.

AI Overviews: Google Search + Selective AI Synthesis

AI Overviews source panel

AI Overviews sits on top of Google's full search index. The question isn't whether to search, it's whether to generate an AI synthesis or let traditional results handle it. For "Stripe billing documentation" Google just showed organic links to Stripe's docs, because AI synthesis would've been worse than a direct link. For "SaaS churn benchmarks," it showed a featured snippet with the exact number. But for "what is a CRM," it generated a full AI Overview.

AI Overviews citation pattern

AI Overviews is the YouTube/Reddit platform. YouTube appeared in 5 out of 6 triggered queries, and for "Asana vs Monday vs ClickUp," a video was the answer, not just a source. Reddit dominated the advisory query.

Search pattern:

AI Overviews search pattern

Cheatsheet

This is what the experiment boiled down to. One table based on what content type was cited, on which platform, and for which intent.

AI search optimization cheatsheet: content types cited by platform and intent

What about the training data?

For queries where search doesn't trigger, the only real strategy is broad ecosystem presence:

  • Get mentioned on review platforms G2, Capterra, TrustRadius
  • Get discussed in communities Reddit
  • Get listed on aggregator/directory sites industry-specific directories in your niche
  • Get written about not just your own content, but other people's blog posts, newsletters, roundups mentioning you

The logic: training data is a snapshot of the web. The more places you appear, the more likely you are to be in that snapshot. But we're all guessing about the lag time. If anyone has seen research on this, I'd love to hear about it.

Fun experiments like this

If you run this experiment for your own space, I'd genuinely love to see what you find. The more data points we have across different verticals and geographies, the better everyone's mental model gets.

The AI search landscape isn't one landscape. It's five, and they're all moving. The question isn't "how do I optimize for AI?", it's "which AI platforms matter for my queries, and what does each one actually reward?".

Kamila Olexa

Written by

Kamila Olexa

Growth Marketing Lead

Kamila writes about SEO, content strategy, and growth marketing. Consulting and leading organic & paid growth at multiple startups from 0 to 100k monthly visitors and counting.