← All posts

April 17, 2026

The 7-dimension AI SEO checklist (2026)

An AI SEO checklist is a structured set of technical and content criteria that determine whether AI assistants like ChatGPT, Perplexity, and Google AI Overviews will surface and cite a given site.

Classic SEO optimizes for ranking algorithms. AI SEO optimizes for something more specific: being the source a model quotes when someone asks the question your content answers. These are related but not identical goals. You can rank #1 organically and still never appear in an AI citation.

Here are the 7 dimensions that matter, what each one measures, and the concrete action you can take on each this week.

Dimension 1: AI Crawler Accessibility

What it is: Whether your robots.txt allows the major AI crawlers to fetch your content.

The bots you need to permit include GPTBot (OpenAI), ClaudeBot (Anthropic), PerplexityBot, Google-Extended (for AI Overviews training), Bingbot, and at least 10 others active in 2026. A single disallow line in an old robots.txt file — added to block scraping — can prevent all of them.

OpenAI publishes its crawler documentation at platform.openai.com/docs/bots. Anthropic's crawlers are documented in their usage policies. Both respect robots.txt directives.

The mistake: Most sites don't explicitly block AI crawlers, but some do — particularly those that added aggressive bot-blocking after 2022 to fight scraping. If your robots.txt uses User-agent: * with a broad Disallow, you may be blocking every AI crawler by accident.

Action: Fetch yourdomain.com/robots.txt and search for your policy on GPTBot and ClaudeBot. If they're blocked, add explicit Allow rules for each crawler user-agent.

Dimension 2: llms.txt Presence

What it is: A plain-text file at /llms.txt that curates your best pages for LLMs to prioritize.

Proposed by Jeremy Howard of fast.ai and documented at llmstxt.org, the convention is now adopted by Anthropic and a growing number of developer-focused companies. It doesn't guarantee citations, but it gives models a structured signal about what you consider authoritative.

Action: Create /llms.txt with an H1 (your brand name), a one-sentence blockquote summary, and 10-30 markdown links grouped by topic with brief inline descriptions. Host it at your domain root with Content-Type: text/plain.

Dimension 3: Schema.org Completeness

What it is: JSON-LD structured data that makes your content machine-readable at a semantic level.

The schema types that matter most for AI citability are: Organization, WebSite, Article or BlogPosting, FAQPage (LLMs heavily index FAQ content), HowTo, and Product. Google uses this data to construct AI Overviews; other models use it to resolve entity ambiguity during retrieval.

The gap most sites have: They implement schema on the homepage and nowhere else. Article schema with dateModified is especially high-value and almost universally missing from blog content.

Action: Audit your top 5 pages with Google's Rich Results Test. Add FAQPage schema to any page with a Q&A section. Add Article schema with dateModified to every blog post.

Dimension 4: Passage-Level Citability

What it is: Whether individual paragraphs and sections on your pages are structured in a way that makes them easy to extract and quote.

LLMs don't cite pages — they cite passages. A 1,500-word blog post might have one genuinely quotable paragraph and 1,400 words of context. That paragraph needs to be: self-contained (understandable without surrounding text), specific (contains a number, definition, or clear claim), and accurate (can withstand attribution).

Structural patterns that get cited:

  • Definitions ("X is Y") in the first sentence of a section
  • Numbered lists with parallel structure
  • Comparison tables with named competitors
  • Quoted statistics with source attribution

Action: Take your 3 best-performing blog posts and rewrite the first paragraph of each major section to lead with a definition or specific claim.

Dimension 5: AI Overviews Presence

What it is: Whether your content is positioned to appear in Google's AI Overviews for queries your audience is actually asking.

Google's AI Overviews pull from a mix of top-ranked pages and schema-rich content. Sites that appear in AI Overviews tend to have: clear topical authority on a narrow subject, FAQ schema, and fresh publish dates. The correlation between AI Overview inclusion and FAQPage schema is high enough to treat schema as a prerequisite.

Action: Identify the 5 questions your target customers ask most in discovery calls. Check whether your site has a dedicated page (or section) answering each one directly, with an H2 phrased as the question and a direct answer in the first two sentences below it.

Dimension 6: Content Freshness

What it is: The combination of sitemap lastmod timestamps and visible on-page publish/update dates that AI crawlers use to assess recency.

Models trained on crawl data and tools like Perplexity doing live retrieval both weight freshness. A post from 2021 with no lastmod update competes poorly against a 2025 post on the same topic, even if the older post is more thorough.

Action: Audit your sitemap.xml and confirm that lastmod values reflect actual content updates. For your top 10 pages, add or update a visible "Last updated: [date]" line.

Dimension 7: Structured Patterns LLMs Cite

What it is: A content-level audit of whether your writing uses the specific patterns that language models extract and quote.

This is the most overlooked dimension and the one with the most immediate ROI. The patterns that reliably get cited:

  • Definitions at the top of sections — "X is a Y that does Z"
  • Statistics with source attributions — "According to [specific source], X% of Y"
  • Comparison tables — Named alternatives with specific attributes compared
  • Numbered steps — Especially in how-to content
  • Q-header pairs — H2 phrased as a question, answered directly in the paragraph below
  • Checklists — Like this one

Homepage marketing copy ("We help enterprises transform their digital journey") essentially never gets cited. It has no extractable claim, no specificity, and no quotable unit.

Action: Score your homepage and top 3 landing pages against this list. If none of these patterns appear above the fold, rewrite the first section to include at least a definition and one specific stat.

Action checklist

  1. Fetch your robots.txt and confirm GPTBot, ClaudeBot, and PerplexityBot are allowed — then create or update /llms.txt with curated links.
  2. Add FAQPage and Article (with dateModified) JSON-LD schema to your top content pages.
  3. Run a free CiteReady audit to get a scored baseline across all 7 dimensions before you prioritize further work.

FAQ

Is AI SEO different from regular SEO? Partially. Technical foundations overlap — crawlability, schema, freshness, and authority signals matter for both. The divergence is at the content level: classic SEO rewards keyword density and backlink volume; AI SEO rewards structured, self-contained, quotable passages.

Which AI tool should I prioritize — ChatGPT, Perplexity, or Google AI Overviews? All three pull from different data sources, but the 7 dimensions above improve your standing with all of them. If forced to pick one, optimize for Google AI Overviews first — it has the largest reach and the clearest documented signals.

How long does it take to see results from AI SEO changes? Schema changes and llms.txt deployment can be picked up by crawlers within days. Content-level changes take longer because retrieval indices need to update. Expect 4-8 weeks for measurable changes in AI citation frequency.

Run a free CiteReady audit on your site

Get your score across all 7 dimensions — with a Claude-Opus executive summary and prioritized fix plan. First full audit is free.