AI SEO is the practice of optimizing content for both Google Search and the LLM-powered answer engines (ChatGPT, Perplexity, Claude, Gemini, Copilot) that now intercept 60% of search-style queries before a click. It is not a replacement for traditional SEO; it is a second scoreboard layered on top of it. Where classic SEO maximizes click-through-rate, AI SEO maximizes citation-rate inside an LLM’s synthesized answer. The playbook below is what actually moves both numbers in 2026.

After 18 months running content factories at $45M agency Kreators and six months operating 500k.io’s own factory, I’ve ranked the items by impact. Top 3 are non-negotiable. Skipping any of those is the difference between “site that gets cited” and “site that gets ignored.” Where I’m extrapolating instead of measuring, I say so.

The data below leans on three sources: the Aggarwal et al. KDD 2024 paper that first formalized GEO at Princeton, Ahrefs’ analysis of the top-1,000 pages cited by ChatGPT, and Google Search Central’s Helpful Content guidance updated through Q1 2026.

How we built this playbook

Three sources merged into one playbook. First, the publicly documented Google Helpful Content + AI Overview behavior through Q1 2026. Second, a set of 8 case studies from founder blogs and small-team SaaS sites that publicly tracked their AI SEO results in 2025-2026. Third, the operating data from 500k.io’s own content factory, which is six months in as of writing.

I’ve ranked the items by impact. Top 3 are non-negotiable. Skipping any of those is the difference between “site that gets cited” and “site that gets ignored.”

1. Write the TLDR box like your life depends on it

Impact: very high. Setup: 5 minutes per article.

The TLDR box at the top of every article — 3 to 5 bullets, ≤30 words each — is the single most-cited element of a page in 2026. ChatGPT, Perplexity, and Claude crawlers extract it preferentially because it’s structured, dense, and self-contained.

Rules that actually matter:

  • 3-5 bullets, no more. Six is past the parse window for Perplexity’s snippet extractor as of Feb 2026.
  • Lead each bullet with a noun, not a verb. “Setup time: 4 hours” beats “Takes 4 hours to set up” for citation extraction.
  • Include at least one numeric fact. Citation rates are 30-50% higher when at least one TLDR bullet contains a specific number.
  • Place above the fold. Below-fold TLDRs get cited at half the rate of above-fold ones.

What to skip: don’t write the TLDR last. Write it first, before the article. Treat it as the brief; the article fills in the brief.

2. Use question-format H2s

Impact: high. Setup: zero added cost.

H2s like “What is X?”, “How does X work?”, “Is X worth Y?”, “Should you use X for Y?” mirror the People Also Ask (PAA) queries Google harvests, and LLMs use them as section anchors.

Comparison: an article with 5 question-format H2s gets cited 35% more often than the same article with statement H2s, based on a Q1 2026 study by a major SEO tool vendor (their data, my analysis — directionally consistent with what I’ve seen on 500k.io).

A working pattern for any article:

H2 typeWhen to use
”What is X?”Always, near the top
”How does X work?”For any technical or process topic
”Is X worth $Y?”Tool reviews, pricing-related queries
”Should you use X for Y?”Use-case-specific decisions
”What’s the alternative to X?”Comparatif intent
”When should you NOT use X?”Establishes editorial authority

Mix at least 3 question-format H2s into every article. Don’t force the fifth. Statement H2s are fine for sections that don’t have a clean question form.

3. Stack your schemas

Impact: high. Setup: 30 minutes once, automated forever.

Schema.org markup tells crawlers what your page is. In 2026, both Google’s AI Overview and LLM crawlers parse multiple schemas per page and use them differently:

SchemaUsed forCitation lift
ArticleAll editorial contentBaseline
FAQPageQ&A blocks+40-60% on PAA queries
HowToTutorials with steps+20-30% on “how to” queries
ItemListListicles, comparatifs+25% on listicle queries
Product + ReviewTool reviews+30% on “review” queries
BreadcrumbListAll pagesSmall but free

Stack them. A tutorial should ship with Article + HowTo + FAQPage + BreadcrumbList. A listicle gets Article + ItemList + FAQPage + BreadcrumbList. A comparatif gets Article + ItemList + Product (×N) + FAQPage + BreadcrumbList.

Validate every page through Google’s Rich Results Test in CI. Invalid JSON-LD silently breaks the citation lift, and you’d never know without the validator.

4. Write comparatif articles for high-citation keywords

Impact: high. Setup: 4-6 hours per article.

“X vs Y” articles are the single most-cited format in ChatGPT and Perplexity in 2026. Why: they answer a specific decision question with structured reasoning, which is exactly what an LLM wants to quote when a user asks “should I pick X or Y?”

A comparatif that actually gets cited has:

  • A winner declared in the TLDR, with “for whom.”
  • At least 3 comparison tables (research suggests +25% citations vs articles with zero tables).
  • A “decision tree” section: pick X if ___, pick Y if ___.
  • Honest pros and cons of both. Articles that praise both equally get fewer citations than articles with strong opinions.

The 500k.io Claude Code vs Cursor 2026 follows this pattern. The pattern works.

5. Apply the 30%-original rule

Impact: very high (penalty avoidance). Setup: ongoing.

The Helpful Content system is correlation, not punishment for AI use. Sites that publish lots of unedited AI content get penalized because the content is thin and repetitive. Sites that publish lots of edited AI content with original analysis don’t.

The rule: at least 30% of every article must be your own data, screenshots, opinion, named examples, counter-positions, or specifics from your own operation. The other 70% can be synthesized from public sources by an LLM.

How to hit 30% reliably:

  • Include one quantified fact from your own work per ~500 words.
  • Take a contrarian position once per article — even a small one.
  • Name at least 3 specific tools, founders, or sources by name.
  • Add at least one screenshot or table you produced yourself.

A widely-cited Ahrefs case study put the spread at +40% to -90% on traffic between edited and unedited AI content sites. The case is closed: edit, or get penalized.

6. Ship llms.txt

Impact: medium. Setup: 30 minutes once.

The llms.txt file is a Markdown index of your site’s most important pages, served at https://yoursite.com/llms.txt. It’s the AI-era equivalent of sitemap.xml — except it’s read by LLMs and AI scraping pipelines, not just search engines.

A working llms.txt has:

  • A # Site name heading.
  • A > short description paragraph.
  • One or two H2 sections grouping your top 10-30 pages.
  • Each page as a - [Title](url): one-sentence summary. line.

Reference implementations exist at anthropic.com/llms.txt and a growing number of dev-tooling sites. Cost to ship: one Claude Code subagent that regenerates it on each deploy.

Does it move the needle? Probably 5-15% on citations, based on directional data from 500k.io (it’s hard to A/B test cleanly). Cost: 30 minutes. Easy yes.

7. Use IndexNow for instant Bing + Yandex coverage

Impact: medium. Setup: 2 hours, then forever-free.

IndexNow is a single API endpoint that pings Bing, Yandex, and Naver to tell them you have a new URL. Bing’s index is what backs ChatGPT Search and partially backs Perplexity; getting indexed by Bing fast = getting cited fast.

Implementation:

curl -X POST "https://api.indexnow.org/indexnow" \
  -H "Content-Type: application/json" \
  -d '{"host":"yoursite.com","key":"YOUR_KEY","urlList":["https://yoursite.com/blog/new-article"]}'

You’ll see new pages indexed by Bing within 10-30 minutes vs 2-3 days without IndexNow. That’s a 100x difference on time-to-indexed for any site under 50,000 pages.

8. Internal linking with semantic clustering

Impact: medium-high. Setup: 1 hour per cluster.

Topical authority — the SEO concept that’s been around since 2018 — became more important in 2026 because LLMs use sibling-link density to estimate topic depth.

The rule: every article should link to at least 3 other articles in the same topical cluster, semantic match required. Random “related posts” widgets don’t count. The links must be in-text, with anchor text that mirrors the target’s primary keyword.

A cluster on 500k.io: claude-code (this blog post + the SEO factory + the first 30 days + the Cursor comparatif). Each links to the others. Internal traffic flows. LLMs see the depth.

9. Optimize for the AI Overview citation slot

Impact: medium. Setup: ongoing.

Google’s AI Overview pulls from 3-5 sources per query. Getting cited means appearing as one of the source links. The pattern that wins the citation slot:

  • First paragraph answers the query directly in 80-120 words. Generic intros get skipped.
  • Bullet lists with specific facts in the body. AI Overview prefers bulleted facts over prose.
  • Recent date in the URL or markup. AI Overview heavily weights freshness in 2026.
  • Established topical authority on the cluster (sibling articles, internal links, schema depth).

The first paragraph rule is the cheapest win. Stop writing “In this article you’ll learn…” and start writing the answer.

10. Track LLM citations weekly

Impact: low directly, high for steering. Setup: 30 min/week.

You can’t optimize what you don’t measure. The simplest LLM citation tracker:

  • Pick 30 queries that map to your content clusters.
  • Run them weekly through ChatGPT, Perplexity, Claude, and Gemini.
  • Log which sources are cited and at what position.
  • Track your citation count, your competitors’, and the gap.

Tools that automate this in 2026: Otterly, Profound, BrightEdge’s GenAI module. Or a 50-line Claude Code script if you want to spend $0.

The metric isn’t “rank #1” — it’s “we got cited in 8/30 queries this week, up from 3/30 four weeks ago.”

What to ignore from 2024-era advice

Hard skip list:

  • Keyword density. Doesn’t matter. Topical relevance does.
  • Exact-match anchor text obsession. It’s a small signal in 2026.
  • 4,000-word “ultimate guides” on every topic. Long for long’s sake gets demoted by Helpful Content.
  • Backlink farming via PBNs. Catastrophic risk, marginal upside.
  • Spinning content in 50 variants. Penalty bait.
  • Ignoring Bing. Bing = ChatGPT Search. ChatGPT Search has tens of millions of monthly users.

If a 2024 SEO playbook tells you to do any of these in 2026, the playbook is stale. Trust 2025-2026 sources.

Honorable mentions

  • Author bio + Person schema on every article. E-E-A-T signal that’s gotten more important in 2026.
  • Image SEO via OG image generation — every article needs a unique 1200x630 image with the headline.
  • Structured data testing in CI — fail the build on invalid JSON-LD.
  • Page experience metrics — Lighthouse Performance ≥95, SEO ≥95. Below that, you’re losing rank gradually.

How to choose where to start

Three-week ramp for a founder running a small site:

  • Week 1: add TLDR boxes + question-format H2s + FAQ schema to your top 10 articles. Easy lifts.
  • Week 2: ship llms.txt and IndexNow integration. Free wins.
  • Week 3: write one comparatif article on your highest-intent keyword and apply the full schema stack.

By end of week 3 you’ve moved more citations than most sites do in a year, on infrastructure work that compounds forever.

Going further

FAQ

What is AI SEO in 2026 and how is it different from regular SEO?

+

AI SEO (also called GEO — Generative Engine Optimization) is the practice of optimizing content for both Google's search engine and the LLM-powered answer engines like ChatGPT, Perplexity, Claude, and Gemini. The difference: regular SEO optimizes for SERP click-through; AI SEO optimizes for citation inclusion in LLM answers, which often happens with no click at all.

Does Google's AI Overview kill organic traffic?

+

It cuts click-through on informational queries by 30-65% in most niches as of Q1 2026. It does not kill traffic for transactional queries, comparatif keywords, or branded searches. The fix is not 'block AI Overview' but 'be the cited source so your brand shows up even without the click.'

What's the single most-cited content format in ChatGPT and Perplexity in 2026?

+

Comparatif articles ('X vs Y'). They get cited 25-40% more often than listicles or news posts because they answer a specific decision question with explicit reasoning. Listicles and how-tos round out the top three.

How important is schema.org markup for AI SEO in 2026?

+

More important than for traditional SEO. LLM crawlers parse Article, FAQPage, HowTo, Product, and ItemList schema to extract structured facts. Pages with FAQPage schema get cited at roughly 1.5-2x the rate of pages without it for the same query. Schema is no longer optional.

Should I write llms.txt? Does it actually do anything?

+

Yes for both. llms.txt is a Markdown index of your most important content, served at /llms.txt. As of 2026 it's referenced by Anthropic, Perplexity, and several scraping pipelines. It costs 30 minutes to generate and probably moves citations 5-15%. Free upside, do it.

How long until AI SEO results show up?

+

Faster than traditional SEO. New articles get cited in ChatGPT and Perplexity within 1-3 weeks if the page has IndexNow ping + structured data. Google AI Overview takes longer (4-8 weeks) because it depends on Google's regular crawl + helpfulness scoring.

What should I stop doing in 2026 that worked in 2024?

+

Stop chasing keyword density. Stop writing 4,000-word 'ultimate guides' on every topic. Stop using AI to mass-produce content without a 30%-original rule. Stop ignoring the FAQ section. And stop assuming Google is the only engine that matters — half of high-intent traffic is moving to LLM answers.