Meet the team. See what’s next.
The conference for marketers ready to win in 2026
Search, AI, strategy—live on stage this October in San Diego
AI Search

ChatGPT May Scrape Google, but the Results Don’t Match

Louise Linehan
Louise is a Content Marketer at Ahrefs. Over the past ten years, she has held senior content positions at SaaS brands: Pi Datametrics, BuzzSumo, and Cision. By day, she writes about content and SEO; by night, you'll find her playing football or screaming down the mic at karaoke.
It’s official: ChatGPT is scraping Google’s results. But our data shows the citations don’t match as closely as you’d expect.

We know that AI assistants like ChatGPT access search indices, like Google and Bing, to retrieve URLs for their response. But how, exactly?

To find out, we’ve run a series of experiments looking at the relationship between the URLs cited by AI assistants, and the results found in Google when searching for the same topics.

So far, we’ve tested long-tail prompts (very long, very specific queries just like those you’d enter into ChatGPT); fan-out queries (mid-length prompts that relate to the original long-tail prompt); and today we’re testing short-tail keywords—ultra-short, ultra-specific “head” terms.

Short-tail keywords offer the clearest illustration of how AI citations track with Google results.

Based on three separate studies, our conclusion is that ChatGPT (and similar systems) don’t just lift URLs directly from Google, Bing, or other indexes. Instead, they apply additional processing steps before citing sources.

Even when we examined fan-out queries—the actual search prompts these systems send to search engines—the overlap between AI and search engine citations was surprisingly low.

In other words, while ChatGPT may pull from Google’s search index, it still appears to apply its own selection layer that filters and reshuffles which links appear.

It’s therefore not enough to identify fan-out queries and rank well for them—there are additional factors influencing which URLs get surfaced, that are outside of a publisher’s control.

Different query types tell us different things about how AI assistants handle information.

In our earlier research, Ahrefs’ data scientist Xibeijia Guan analyzed citation overlap between AI and search results for informational long-tail and fan-out prompts, using Ahrefs Brand Radar.

A screenshot of Ahrefs Brand Radar dashboard, highlighting 15 AI mentions for the long-tail query "how much does it cost to install a security camera"

This time, she has taken a sample of 3,311 classic SEO-style head terms, covering informational, commercial, transactional, and navigational intent.

Example queryInformationalCommercialTransactionalNavigational
1cincinnati bearcats basketballbest credit card rewardspools for saleonedrive sign in
2protein in shrimpsoundbar for tvshop girls dressverizon customer support
3what is cybersecurityat home saunabuy a domaincostco toilet paper

Each keyword has been run through ChatGPT, Perplexity, and Google’s top 100 SERPs to analyze citation overlap between AI and search.

If anything were to align closely with Google’s results, you’d expect it to be short-tail queries—since that’s the classic way we search.

But that’s not quite the case.

Ahrefs' study of ~3K short-tail queries showing short-tail query overlap between ChatGPT and Google for both URLs and Domains. URL overlap 20.69% in SERPs top 100 10.00% in SERPs top 10 domain overlap 53.51% in SERPs top 100 31.80% in SERPs top 10

While the citation overlap for short-tail queries (10%) is slightly stronger than for fan-out queries (6.82%), it’s still much weaker than we’d expect if it were directly echoing the SERPs.

This is even more surprising, now we have confirmation that OpenAI and Perplexity have been scraping Google results via a third-party provider.

It’s possible we’d see more overlap if our study focused only on ‘real-time’ queries (e.g., news, sports, finance), since those are reportedly the kinds ChatGPT scrapes Google for.

Perplexity citations align closely with Google’s search results across short-tail queries.

Ahrefs' study of ~3K short-tail queries showing short-tail query overlap between Perplexity and Google for both URLs and Domains. URL overlap 72.85% in SERPs top 100 65.07% in SERPs top 10 domain overlap 91.84% in SERPs top 100 80.58% in SERPs top 10

Unlike ChatGPT, overlap isn’t just visible at the domain level—most of Perplexity’s cited pages are also the exact URLs ranking in Google’s top 10.

This mirrors the findings in our long-tail query study, where Perplexity responses most resembled Google’s results, reinforcing its design as a “citation-first” engine.

Domain overlap is consistently higher than URL overlap, suggesting that ChatGPT and Perplexity cite the same websites as Google—but not the exact same pages.

Ahrefs' study of ~3K short-tail queries showing short-tail query overlap between Perplexity, ChatGPT, and Google for both URLs and Domains in the top 10. ChatGPT: URL overlap - 10.00% in SERPs top 10 domain overlap - 31.80% in SERPs top 10 Perplexity: URL overlap - 65.07% in SERPs top 10 domain overlap - 80.58% in SERPs top 10

In ChatGPT, the domain-URL gap is especially wide—31.8% vs. 10%.

In other words, ChatGPT cites ranking domains ~3X more than ranking pages.

On the one hand, this could mean ChatGPT selects different pages from the same domains as Google.

For example, Google cites one page from ahrefs.com/writing-tools/, while ChatGPT finds a better “fit” on ahrefs.com/blog/ and cites another.

If true, this reinforces the value of creating cluster content—optimizing multiple pages for different topic intents, to have the best chance of being found.

Another possibility is that both lean on the same pool of authoritative domains, but disagree on arbitrary pages.

Assess your cluster content in AI and search

You can check the SEO performance of your cluster content in the Related Terms report in Ahrefs Keywords Explorer.

This will show you if and where you rank across an entire cluster of related keywords.

Just add a Parent Topic filter, and a Target filter containing your domain.

A screenshot of Ahrefs "Clusters by Parent Topic" tab within the Related Terms report. A parent topic filter has been applied for "Is: check google ranking", and a target filter has been applied for "ahrefs.com". An arrow points from the target filter to a pop-up report which shows the highlighted ranking positions for Ahrefs across the parent topic.

Once you’ve done that, head to Ahrefs Brand Radar to check on the AI performance of your cluster content.

Run individual URLs through the Cited Pages report in Ahrefs Brand Radar to see if your cluster content is being cited by AI assistants like ChatGPT, Perplexity, Gemini, and Copilot.

A screenshot of the Cited Pages report in Ahrefs Brand Radar, circling a "Page URL Contains:" filter, with a specific Ahrefs blog included. An arrow points to the circled filter, with the writing "Check specific domains, URLs, and subfolders being cited in AI" A trend chart shows the trended performance of the blog in ChatGPT.

Work out if any content is missing from either surface, then optimize until you’ve filled those gaps and enriched the overall cluster.

You can use topic gap recommendations in Ahrefs’ AI Content Helper to help with this.

A screenshot of Ahrefs AI Content Helper interface, with the AI generated "Recommendations" section circled, which provides suggestions on how to fill topic gaps.

Short-tail queries show closer SERP-AI alignment than natural language prompts—especially when it comes to Perplexity.

A double bar chart showing ChatGPT and Perplexity URL overlaps with Google's SERP citations, based on short-tail queries, and long-tail queries ChatGPT short-tail query overlap: 10% Perplexity short-tail query overlap: 65.1% ChatGPT long-tail query overlap: 7.05% Perplexity long-tail query overlap: 28.6%

But the ChatGPT citations generated by fan-out queries (first studied by SQ and Xibeijia) show the least overlap. They match only 6.82% of Google’s top 10 results.

A bar chart showing three bars representing short-tail (10%), long-tail (7.05%), and fan-out queries (6.82%). The chart is titled: ChatGPT URL overlap with SERP across query types (Ahrefs study of ~3K queries)

We’re not comparing apples-with-apples here. These percentages represent different studies, and different sized datasets.

But each study produces similar findings: the pages that ChatGPT cites don’t overlap significantly with the pages that Google ranks. And it’s largely the opposite for Perplexity.

One other thing we haven’t mentioned is intent. The greater citation overlap we see across short-tail queries could partly be explained by the relative stability of navigational, commercial, and transactional queries—which we didn’t assess in our previous studies.

Navigational, commercial, and transactional head terms have SERPs that don’t tend to change too often, because the set of relevant products, brands, or destinations is finite.

This stability means AI assistants and Google are more likely to converge on the same sources, meaning overlap is higher than it is for informational queries (where the pool of possible pages is far larger and more volatile).

Final thoughts

Across all three studies, the story is consistent: ChatGPT doesn’t follow Google’s sources, Perplexity does.

What’s surprising is that ChatGPT differs so much from Google, when we now know that OpenAI does scrape Google’s results.

My hunch is that ChatGPT does more than Perplexity to differentiate its results set from Google.

This theory from SQ seems the most probable one to me:

“ChatGPT likely uses a hybrid approach where they retrieve search results from various sources, e.g. Google SERPs, Bing SERPs, their own index, and third-party search APIs, and then combine all the URLs and apply their own re-ranking algorithm.”

Whatever the case, search and AI are shaping discovery side-by-side, and the best strategy is to build content that gives you a chance to appear on both surfaces.