Meet the team. See what’s next.
The conference for marketers ready to win in 2026
Search, AI, strategy—live on stage this October in San Diego
Data & Studies

Only 12% of AI Cited URLs Rank in Google’s Top 10 for the Original Prompt

Louise Linehan
Louise is a Content Marketer at Ahrefs. Over the past ten years, she has held senior content positions at SaaS brands: Pi Datametrics, BuzzSumo, and Cision. By day, she writes about content and SEO; by night, you'll find her playing football or screaming down the mic at karaoke.
In a dataset of 15,000 prompts, we found that—on average—only 12% of links cited by ChatGPT, Gemini, and Copilot appear in Google’s top 10 results for the same prompt.

Perplexity is the outlier: nearly 1 in 3 of its citations point to pages that rank in the top 10 for the target query.

We analyzed four AI assistants—ChatGPT, Gemini, Copilot, and Perplexity—to measure how often the URLs they cite or reference overlap with what Google or Bing ranks for the same query.

On average, the citation overlap between AI assistants and Google and Bing’s top 10 stands at 11%.

((Google AI overlap total + Bing AI overlap total) / 10) = 10.96%

Ahrefs research using Brand Radar: Only 11% of AI citations overlap with Google’s and Bing’s top 10

Further down, we look at this broken down by individual search engine.

Using data from Ahrefs Brand Radar, our data scientist, Xibeijia Guan, searched 15,000 long-tail queries in Google and Bing, then asked those same questions to AI assistants.

Example queries included:

  • “How much does it cost to install a security camera?”
  • “How to get enough protein when you can’t eat dairy?”
  • “What is the cheapest full comprehensive insurance?”
  • “¿Cuál es el mejor suplemento para la recuperación muscular?”
Search for query "How much does it cost to install a security camera" in Ahrefs Brand Radar

Xibeijia then extracted every visible citation returned for each question across both search and AI indices.

In the search dataset, she recorded the ranking positions of the cited URLs—then categorized them based on whether they appeared in the top 10, or top 100 results.

The AI assistants tested include ChatGPT (both in-text citations and bottom-of-response references), Gemini, Copilot, and Perplexity.

According to research carried out by our data scientist, Xibeijia Guan, just 12% of citations in AI assistants also rank in Google’s top 10, on average.

((28.6 + 8 + 6.1 + 8.6 + 8.2) / 5) = 11.9%

Ahrefs research using Brand Radar: 12% of AI citations are present in Google’s top 10

80% of those citations don’t rank anywhere in Google for the original query—80% is essentially the average of all the “red” columns in the chart above.

The overlap between AI and search citations is smaller than we anticipated, but we have a couple of theories as to why that might be.

We’ll get into them a little later on.

Perplexity aligns closest to Google

Perplexity is an AI assistant that was literally built to cite—its goal is to support nearly every statement with a source.

Perplexity is the world’s first answer engine. It searches the internet in real time to deliver fast, clear answers to any question—with sources and citations included…Every answer comes with clickable citations, making it easy to verify the information and dig deeper if you want.

Perplexity Team, Oct 2024

ChatGPT and Gemini, on the other hand, don’t always link out.

Their citation process is very much query-dependent, and their internal systems trigger searches based on different criteria.

ChatGPT, for example, seems to use something called the sonic classifier—if the prompt “probability” exceeds a certain threshold (i.e. if the response is less easy to “predict” based on training data), then ChatGPT will perform a web search.

Perplexity consistently favors content that ranks well in Google, with 28.6% of its cited URLs landing in the top 10.

For the other AI assistants, that number hovers around 8%—and more than 80% of their citations come from pages that don’t rank at all for the target query.

Since Perplexity focuses heavily on citations, it seems logical that its results would align closely with Google’s.

But, unlike Gemini and other AI assistants, Perplexity doesn’t draw on Google or Bing’s index. In fact, it has its own search index, based on its crawler: perplexitybot.

AI assistants—even those that draw on Google and Bing—appear to query search indexes in a fundamentally different way. This explains why we’re seeing minimal crossover.

When measuring against Bing results, overlap remains low in the top 10—hitting roughly 10% on average.

((16.6 + 14 + 8.1 + 8.1 + 3.3) / 5) = 10.02%

Ahrefs research using Brand Radar: 10% of AI citations overlap with Bing’s top 10 results

Copilot rises to the top of the list, which is to be expected since it is a Microsoft-owned product that draws on Bing search results when generating its responses.

Perplexity still remains one of the most SEO-aligned AI assistants, while ChatGPT and Gemini show the least overlap.

This study builds on our earlier research into how often high-ranking pages get cited in Google’s AI Overviews.

Further reading:

In those studies, we found a moderate positive correlation between Google rankings and citations in AI Overviews, with 76% of cited URLs coming from pages in the top 10.

Ahrefs AI Overview research: 76% of AI Overview citations pull from top 10 pages

Based on the data, it seems as if AI Overviews function like an extension of traditional search—surfacing results that mostly reflect the SERP.

In other words, AI Overviews follow the SERPs—AI assistants don’t.

Even the assistants built by Google aren’t closely aligned with Google’s SERPs.

There are clearly other factors influencing what gets cited, beyond just search rankings.

Our data shows that the URLs cited by AI assistants rarely match what Google and Bing ranks.

In most cases, they don’t appear in the top 10 for the target prompt—often, they aren’t in the top 100 at all.

This is because AI assistants don’t rank results in the same way search engines do.

Instead of processing just one user query when deciding what to cite, they tend to retrieve pages based on multiple variations of that query—a technique known as “query fan-out.”

These variations then get merged using methods like Reciprocal Rank Fusion (RRF), where the pages that appear consistently are favored more.

For example, a page that ranks around #6 for “how to descale a coffee machine,” “cleaning a Nespresso machine,” and “remove limescale from coffee maker” might be cited over one that ranks #1 for only one of those.

In fact, it’s possible that AI assistants may never search for the actual prompt that the user inputs—instead it just acts as a catalyst for other fan-out queries.

This is why eventual citations won’t always rank for the original prompt.

Personalization can also play a role: the exact same search query run through AI can lead to wholly different citations, due to factors like the user’s prompt or conversation history.

All of this may explain why AI citations don’t always mirror the SERPs.

Tip

You can optimize for multiple, long-tail query variations using Parent Topics in Ahrefs Keyword Explorer.

Just search a keyword, head to the Matching Terms or Related Terms report, and check out the Parent Topics on the left, or hit the Clusters by Parent Topic tab.

Then hit the Questions tab for long-tail queries to target in your content…

And to see how much ownership you have over existing long-tail query permutations, add a Target filter for your domain.

Adding a target to Ahrefs Parent Topics report to see where a site ranks for related keywords. Screenshot shows a view of the Matching Terms report, with various different sized and coloured boxes representing the size of the keyword cluster group.

Recent tests by Aleksis Rylko suggests that SearchGPT citations closely resemble Google citations.

Rylko believes ChatGPT may now quietly pull from Google’s search index.

Aleyda Solís corroborated this—she discovered that ChatGPT uses Google as a fallback when Bing pages are inaccessible.

These findings came as a surprise to many, because ChatGPT’s built-in provider is Bing.

From our own dataset, we aren’t seeing much differentiation in how ChatGPT surfaces both Google and Bing results.

Google (top 10)Bing (top 10)Google (top 100)Bing (top 100)
ChatGPT (in-text citations)8.00%8.10%8.80%5.00%
ChatGPT (references)6.10%8.10%9.20%5.20%

That said, we collected this data in early July, so results may have since changed—especially given that OpenAI is planning to release its AI-first web browser, and potentially even build its own search index, going by the recent memo leak.

We’ll be doing further research to study any citation shifts.

Wrapping up

From our research, it seems that ranking well definitely does help, but not necessarily for the queries you might assume you’d need to rank well for.

We’ll keep studying citation overlap in traditional search and AI to see if anything changes.

In the meantime, optimizing for query variations, building topic clusters, and “owning the entity” seems to be the way to go if you are hoping to get cited in AI.