The conference for marketers ready to win in 2026
Starts in --Days --Hours --Minutes --Seconds
General SEO

The SEO Process in a Nutshell (Updated for AI Search)

Mateusz Makosiewicz
Marketing researcher and educator at Ahrefs. Mateusz has over 15 years of experience in marketing gained in agencies, SaaS and hardware businesses. When not writing, he's composing music or enjoying long walks.
Article Performance
Data from Ahrefs
  • Organic traffic
  • Linking websites

The number of websites linking to this post.

This post's estimated monthly organic search traffic.

Search engine optimization (SEO) is the practice of growing a website’s traffic from organic search results. Here are the five general steps of the SEO process:
  1. Get your technicals right. Make sure your website is crawlable and indexable. 
  2. Find a keyword to target. The topic of your page and the main source of organic traffic.
  3. Create an optimized page.  Both users and Google want useful and interesting content. 
  4. Build links to the page. Links are one of the most important ranking factors.
  5. Optimize for AI search visibility. The same fundamentals plus a few moves that get you cited in ChatGPT, Perplexity, Gemini, and AI Overviews. When this guide first went live, AI search wasn’t a thing; that’s why I’m updating it.

In this article, I’ll explain and demonstrate how this process works using some of the SEO tactics Ahrefs used to grow into a 9-figure SaaS company.

Technical factors can impact your rankings or even prevent your site from appearing on Google’s search result pages.

To rank your content, Google needs to:

  1. Find and crawl your content. You won’t rank if your content is inaccessible to Google (this may be because of a disallowed Googlebot).
  2. Index your content. We’re talking about the master list of all pages that Google keeps in order to display them for relevant search queries. Google may not want to show certain pages if it “thinks” those pages are not the main version of the content (see canonicalization for more info) or if the Googlebot is blocked.

How Google builds its index

In most cases, unless you’ve specifically instructed Googlebot not to crawl and/or index your site, your pages are ready to show up on the search engine results pages (SERPs). Keep in mind that it may take some time before Google indexes your content.

The easiest solution to technical SEO issues is to get SEO auditing tools and fix any problems they report back to you.

Two tools we recommend are Google Search Console and Ahrefs Webmaster Tools (they’re free). You can also set up Bing Webmaster Tools if you want to monitor your performance on Bing.

For example, to find sitemap issues in Ahrefs Webmaster Tools:

  1. Open Site Audit 
  2. Go to the All issues report and scroll to the Sitemap category
  3. Click on the found issues and see the affected pages. You can also see the recommended solution

Sitemap issues reported by AWT.

While you’re at it, check your ‘robots.txt‘ for AI crawlers. GPTBot, PerplexityBot, ClaudeBot, and Google-Extended are the main ones. Block them only if you have a deliberate reason to. Most sites benefit from letting them in, since being part of an LLM’s training or retrieval set is how you show up in AI answers.

Start with ‘robots.txt‘. It lives at the root of your domain, so you can pull it up directly: ‘https://yourdomain.com/robots.txt‘. The file is plain text and groups rules by ‘User-agent‘. Look for any block that disallows the AI bots either explicitly or as part of a wildcard ‘User-agent: *‘ rule. The bots worth checking by name are:

  • GPTBot. OpenAI’s training crawler.
  • OAI-SearchBot. powers ChatGPT search results.
  • ChatGPT-User. Fires when a user clicks a link inside ChatGPT
  • PerplexityBot and Perplexity-User. Training and on-demand retrieval
  • ClaudeBot. Anthropic.
  • Google-Extended. Controls whether Google can use your content for Gemini and AI Overviews. Note this is separate from Googlebot, so blocking it does not affect classic search rankings.
  • Bingbot with the ‘nocache‘/‘noarchive‘ directives (relevant for Copilot).
  • Applebot-Extended. Apple Intelligence.

A permissive setup for AI typically just lets these bots through with no explicit ‘Disallow‘ rules. If you want to restrict crawling on specific paths (admin sections, search result pages, faceted URLs), do it under each bot rather than a blanket disallow.

‘robots.txt‘ is the most common place this gets misconfigured, but it’s not the only one. A few other places blocks can hide:

  • Meta robots tags and ‘X-Robots-Tag‘ headers on individual pages or templates. A stray ‘noindex‘ or ‘noai‘ directive in your CMS template, theme, or framework config can keep AI bots from using a page even when ‘robots.txt‘ looks fine.
  • CDN bot management rules. Cloudflare, Fastly, AWS WAF, and similar services often have a one-click “Block AI Bots” toggle, plus rule sets that match known AI user agents and challenge or 403 them. Check your firewall rules and bot management dashboard.
  • Server-level rate limits and IP blocks.  Aggressive throttling that returns 429 or 403 to the AI bots’ IP ranges effectively blocks them. Review your access logs filtered by the user agents above.
  • Cloudflare’s AI crawl controls. If you’re on Cloudflare, the dashboard has a dedicated AI Crawlers section that overrides whatever your ‘robots.txt‘ says. Don’t trust the file alone.

To get traffic from search engines, you need to create content about something people search for.

To illustrate, the content on our blog alone brings us an estimated 256.9K organic visits every month. That’s because we create content designed to rank for keywords with search demand.

Overview of Ahrefs' blog in Ahrefs' Site Explorer

This is where keyword research tools, such as Ahrefs’ Keywords Explorer, come in. With the help of such tools, you’ll easily find hundreds or even thousands of keyword ideas.

For example, entering the main focus of our blog, SEO, in Keywords Explorer reveals over 395K keyword ideas.

Matching terms report for "seo" in Ahrefs' Keywords Explorer

395K keywords are probably a bit too many to manage, and probably not all of them would be a good fit. So here’s what to consider when prioritizing keywords:

  • Search traffic potential. Search traffic potential (not to be confused with search volume) tells you how much traffic you can potentially get from a keyword.
  • Business potential. Topics with high business potential can convert a good portion of your visitors to customers. Topics with low business potential will make it tough for you to feature your product/service.
  • Ranking difficulty. The more backlinks the top-ranking pages have and the more renowned the competing brands are, the harder it will be for you to rank.
  • Search intent. The reason behind the search. Usually, it’s one of three things: finding a specific website, learning something, or buying something.

Here’s an example. One of the keyword research methods we use is to look for search demand for specific niches or industries in our area of business. For this, we filter for keywords that include the word “for.”

For instance, the ones highlighted display considerable Traffic Potential (TP), have no extreme ranking difficulty (KD), and have high business potential for us.

Matching terms report for "seo" in Ahrefs' Keywords Explorer; the word "for" added in Include filter

Keywords that tick all of the four things above are an ideal situation, but that doesn’t happen all of the time. Mostly, SEOs and content marketers need to go for compromises, e.g., targeting a keyword with high business potential but lower traffic potential.

The content of a page is something that allows Google to “connect you” with the searchers. 

The more interesting and useful your content is, the better. And that’s because quality content is something users expect and search engines need to provide. In fact,

Google admits that content is the most important ranking signal.

How Google exactly ranks content is kept a secret. But they actually provide a hint on the five things that determine which results will be shown for a given search query:

  • Meaning – How well a page matches searchers’ expectations.
  • Relevance – Does a page contain relevant information, e.g., words, phrases, and even pictures and videos relevant to what the searchers are looking for.
  • Quality – Content also needs to be helpful. To determine content quality, Google will take into account both factors occurring on the page (e.g., E-E-A-T, clear and organized form, freshness) and those occurring outside the page (backlinks, which we’ll talk more about later).
  • Usability – If your pages and your competitors’ are equal in every other way, Google may allocate a higher ranking to pages that it finds more accessible (e.g., mobile-friendly, secured with an SSL, fast loading).
  • Context and settings – Google may customize search results based on users’ search history and their current whereabouts.

Let’s take a look at an example of optimizing a page for search.

In the case illustrated below, the search intent seems to favor listicles for the keyword “free seo tools”. Because of that, we had little chance of ranking with a product landing page, even though we offer free SEO tools. So we decided to target the keyword with a content type more aligned with search intent. It now ranks #1 and brings about 1.3K organic visits each month.

List of results in SERP overview showing Google prefers articles listing free SEO tools

There are many techniques SEOs and content marketers use to adhere to Google’s guidelines - too many to explain in this short article. If you want to take a moment to learn about them, see the video below. Otherwise, let’s move on to the next point: building links.

https://www.youtube.com/watch?v=ZWiNz-7gZ24

You’ll need two types of links: internal links and backlinks. Both are ranking signals, with backlinks being one of the major signals in SEO.

Internal links

Internal links are links from other pages on the same website, e.g., a link from one article to another on our blog.

Their main roles in SEO are to help search bots crawl pages more efficiently and pass link equity from linking pages.

Because of the above reasons, you probably won’t find an article on our blog without at least one internal link pointing to another article or a product landing page.

By having internal links, we can create a situation where a page with a lot of backlinks can give a much-needed boost to newer pages (see the middleman method for more details).

Here’s an example interlinking tactic. Including a link to our guide on SEO in the blog’s navigation automatically creates an internal link from every blog post to that guide, helping it to rank higher.

Link to SEO guide in our blog's navigation

Internal Backlinks report for Ahrefs' guide to SEO in Ahrefs' Site Explorer
The Internal Backlinks report in Site Explorer shows different articles linking to the SEO guide, and some of them come from navigation of the site. 

Backlinks

Backlinks are links from external websites. They act as votes. The more “votes” you get, the higher your chance of outranking the competition.

Backlinks impact rankings, and rankings impact traffic. So, generally speaking, the more backlinks you get, the more organic traffic you can generate.

Correlation between referring domains and search traffic.

The difficulty here is that you can’t fully control backlinks. You can either earn them organically (wait for people to discover you and link to you) or build them (ask people to link to you). Let’s look at that in more detail.

In the picture below, you can see examples of our case studies that continue to earn backlinks organically. Publishing original, unique data is one of the best ways to attract backlinks.

Best by links' growth report for Ahrefs' blog in Ahrefs' Site Explorer

The Best by links growth report in Site Explorer allows you to see content that people like to link to and/or is a current target of a link building campaign (useful for competitive research).

And here’s an example of a content piece created specifically for an outreach campaign. Unlike the previous examples, it didn’t need any original studies: 63 SEO Statistics. Our process was:

  1. Researching most cited SEO statistics among the top-ranking articles.
  2. Finding and including their more up-to-date versions in our article.
  3. Asking people who linked to websites with outdated statistics to link to our article instead (that’s the outreach part).
Backlinks report for Ahrefs' blog article on SEO statistics in Ahrefs' Site Explorer
This article got so many backlinks mainly due to an outreach campaign.

Regarding backlinks, it’s important to know that not all links will carry the same weight.

Generally speaking, the best links you can get are “followed” links placed within the main content and those that come from relevant, authoritative websites.

What makes a good link

Head on to the guides listed below if you want to learn more about backlinks and link building.

Recommendation
In the AI era, unlinked brand mentions matter too. After analyzing 75,000 brands, we found that branded web mentions correlate strongly with how often ChatGPT brings your brand up. LLMs look for consensus. When many sources mention you in the right context, you start showing up in AI answers, with or without a backlink. So a podcast appearance, a Reddit thread, or a quote in a roundup post all count. 

The final step in modern SEO is making sure your brand shows up in AI assistants. A lot of the same signals still matter, but the way they’re used is different enough that it needs its own focus. 

1. Build more online mentions of your brand 

Our correlation study of 75,000 brands ranked search factors by how strongly each predicts AI visibility. The top of the list wasn’t backlinks or Domain Rating. It was YouTube mentions (~0.737), with general branded web mentions close behind (0.66–0.71). DR sat at 0.266. Backlink count was negligible.

AI assistants are scoring brand presence, the web’s consensus, not link authority. A brand discussed in many places (linked or not) is a brand the model trusts to recommend. YouTube carries extra weight because LLMs are partly trained on its transcripts, and in the case of Google’s AI Overviews, it seems to be a trusted layer of information, coming from their own product.

To turn this into action, do a quarterly *mention gap analysis*. Add your brand plus 3–5 competitors to Ahrefs Brand Radar, then use the “Others only” filter on the AI charts to surface prompts where competitors get name-checked and you don’t.

How to check AI mention gap in Ahrefs Brand Radar.

The Cited Pages report shows the third-party URLs feeding those mentions, which doubles as your outreach pipeline.

2. Keep consistent brand information across sources

Because AI synthesizes responses from many sources at once, contradictions between those sources produce wrong or muddled output. The model may pick an arbitrary version, or stitch details from incompatible sources into a hybrid that doesn’t match reality.

Preventing this is mostly bookkeeping. Standardize one version of your brand name, tagline, founding year, headquarters, and one-line description, then push that version everywhere it appears: LinkedIn, Crunchbase, G2, Capterra, ProductHunt, AngelList, Wikipedia, industry directories.

Mark up your homepage with Organization schema and use the ‘sameAs‘ property to link your verified social profiles together. Don’t forget founder profiles, since prompts about company leadership pull from them disproportionately.

3. Fill information gaps 

When there’s no authoritative information available, AI doesn’t typically say “I don’t know.” It grounds in whatever third-party content it can find, or it fabricates plausible-looking details. Both outcomes are bad for the brand at the center of the answer.

A made-up brand experiment we ran shows the failure mode clearly. We built a fake company, seeded the web with three deliberately conflicting third-party sources, and posed 56 questions to 8 different models. Most models repeated the planted fictions as fact, even when the brand’s own FAQ contradicted them. When an AI must choose between an unspecific truth and a specific fabrication, the specific fabrication usually wins.

The practical implication is that being silent about yourself is a competitive risk. Publish an FAQ specific enough to leave nothing to interpretation, including direct denials of rumors that float around your category (“we have not been acquired,” “we don’t disclose unit counts”).

Keep pricing, plan limits, and policies in plain HTML rather than behind JavaScript widgets. Publish original numbers when you can. Maintain head-to-head comparison pages so the framing of “you vs competitor” comes from your side, not someone else’s.

A Google search result page displaying Ahrefs data statistics, with a highlighted snippet answering "how big is ahrefs' data?".

4. Keep product documentation and help docs up to date 

When people ask AI about your product specifically, the model can reach for  your documentation, help center articles, and changelog pages on your own domain. Those pages carry strong authority for branded queries because they’re explicitly about your brand and they’re well indexed.

That dependency is also a liability. Old docs hand AI an authoritative source for inaccurate answers: pricing tiers that no longer exist, features you’ve since removed, integrations that shipped without ever making it into the docs. Run a docs audit on a quarterly cadence. Publish changelogs and release notes openly so recency signals follow updates. Roughly once a month, run a short query set against ChatGPT, Perplexity, and Gemini, and verify what they say about your product.

When something is wrong, find the page feeding the error and update it. Brand Radar’s Cited Pages report shows you which of your own URLs AI is actually reading.

5. Write for AI humans, but help machines understand the content

If your content is easy to read, technically accessible, clear, and specific, it’s more likely to be picked up and used by AI systems.

Use clear formatting

AI tools prefer content that’s easy to scan and understand.

  • Use H1, H2, H3 headings to show structure.
  • Break information into bullet points.
  • Keep sentences short and direct.
  • Avoid long, dense paragraphs.

Add technical optimizations

Make your site easy for AI systems to access and interpret.

  • Add structured data (schema markup)
    • Use FAQ schema for common questions.
    • Use product schema for product pages.
  • Improve performance
    • Ensure fast loading speeds.
    • Optimize images and reduce unnecessary scripts.
  • Design for all devices 
    • Make your site mobile-friendly.
  • Allow access to AI crawlers 
    • Don’t block bots in robots.txt
  • Limit heavy JavaScript 
    • Many AI crawlers don’t fully render JavaScript.
    • Keep key content in plain HTML.

Write for clarity

Clear writing improves both human understanding and AI accuracy.

  • Use simple, direct language.
  • Break up long sections into smaller chunks.
  • Include specific facts, numbers, and details.
  • Answer common questions clearly and directly.

Before your jump into the process, it’s good to have realistic expectations in terms of the time and money needed to do SEO successfully. 

When it comes to the results, SEO is usually a lengthy process. We learned from over 4k respondents that it typically takes between 3 - 6 months. How long does SEO take - poll results.

Here are a couple of other things to know when it comes to the time involved in the process:

  • Crawling a page can take Google anywhere from a few days to a few weeks (source). So it seems that there’s a “natural delay” in SEO that comes from the technology itself.
  • According to our study, only 5.7% of pages ranked in the top 10 (for at least one keyword) within no longer than one year.

As for the cost of SEO, the least costly option is doing SEO by yourself - all you need is time to learn and experiment and tools that fit your budget.

Alternatively, you can consider hiring an agency, a freelancer, or a consultant. Read our full study on SEO pricing to know what costs to expect.

SEO hourly rates by business type.

Final thoughts

Monitor your results on a regular basis because search engine rankings tend to change. For this, it’s best to use a tool that tracks your ranking history and shows how you stack up against competitors—see our Rank Tracker, for example.

Track AI visibility on the same cadence using a tool like Ahrefs Brand Radar.

Thanks for reading! Feel free to reach out on LinkedIn.

Article Performance
Data from Ahrefs
  • Organic traffic
  • Linking websites

The number of websites linking to this post.

This post's estimated monthly organic search traffic.