The Format-Intent Grid: which pages AI engines actually cite

The format-vs-format question is wrong. AI citations follow a 5 × 4 × 3 grid — five page formats, four query intents, three platforms. Most teams optimize on a 1×1. Here's the full grid, the data behind each cell, and the under-deployed format hiding inside it: glossary pages.

The most common content question I get is some version of: 'Should I publish listicles or educational content for AI search visibility?' It's the wrong question. Asked that way, the answer is always 'it depends' — which is true and useless.

The right question has three axes, not one. AI citations follow a grid: five page formats (listicle, comparison, how-to, research essay, glossary) × four query intents (informational, commercial, transactional, navigational) × three platforms (ChatGPT, Google AI Overviews, Perplexity). That's 60 cells. Most content teams optimize on a 1×1 — pick one format, point it at one intent, hope it works on every platform. The published research says it doesn't.

This article does three things. It defines the five formats precisely (because most studies collapse 'comparison page' into 'listicle' and lose the signal). It maps each format-intent-platform combination to the strongest available data — Wix Studio's 75K-answer dataset, Allsopp's 26,283-URL ChatGPT analysis, Profound's 680M citations, Seer's 5.47M-query AIO study, Princeton's GEO paper, and others. And it surfaces the format hiding inside the grid that almost no B2B brand is running: glossary pages.

Five formats — defined precisely

Most published studies treat content format as 2–3 buckets: 'listicle', 'article', 'product page'. That's why the answers people get are vague. To map the grid you need a finer taxonomy. These five formats cover almost every page that gets cited by an AI engine.

Listicle / Best-of roundup

listicle

Ranked or unranked enumeration of N items in a category.

e.g. “Best AI search visibility tools

Best for
Top-of-funnel commercial intent in software, SaaS, products.
Weak on
Informational queries; transactional finalisation; entity disambiguation.

43.8% of all ChatGPT-cited page types — 51.4% in software (26,283 URLs).Allsopp / Ahrefs

Comparison page (X vs Y)

comparison

Side-by-side evaluation of 2–5 named products across the same axes.

e.g. “Profound vs Peec AI vs TurboAudit

Best for
Bottom-of-funnel transactional intent; near-purchase decisions.
Weak on
Awareness queries; queries with no defined alternatives yet.

AI Overviews triggers on 95.4% of comparison queries (5.47M queries).Seer Interactive

How-to / tutorial

howto

Multi-step procedural guide with explicit ordered actions.

e.g. “How to set up GPTBot in your robots.txt

Best for
Informational intent with stable workflows; AIO citations.
Weak on
Fast-changing tooling that goes stale in 90 days.

Pages with HowTo schema + numbered lists earn 2.8× more AIO citations.Digital Applied

Research / data essay

essay

Original analysis built around proprietary statistics or first-party data.

e.g. “We analyzed 250M AI search results — here's what we found

Best for
Citation magnet; long-tail authority; cross-engine reuse.
Weak on
Commercial queries; quick decisions; transactional intent.

Pages with 3+ unique data points are 4× more likely to be cited in AIO.2026 Wix Studio research

Glossary / definition

glossary

Canonical entity-defining page that owns the 'what is X?' query.

e.g. “What is Generative Engine Optimization?

Best for
Entity disambiguation; informational + navigational; long-term moat.
Weak on
Direct conversion; commercial queries; narrow product comparisons.

Wikipedia is 47.9% of ChatGPT's top-10 sources — cited 80× more on ChatGPT than on AI Mode.Profound 680M citations · BrightEdge

Four intents AI engines route differently

AI engines do not weigh formats uniformly across queries. The intent of the query determines which format wins the citation. The classification below is the standard search-intent taxonomy adapted for generative engines — and the column you sit in matters more than the format you publish.

01

Informational

e.g. “what is generative engine optimization

02

Commercial

e.g. “best ai search visibility tools

03

Transactional

e.g. “profound vs peec ai pricing

04

Navigational

e.g. “turboaudit chatgpt monitoring

What the studies actually say

Five primary studies anchor the grid. Each one isolates a different slice — read them together, not in isolation.

Wix Studio AI Search Lab analysed 75,000 AI answers and over 1M citations across ChatGPT, Google AI Mode, and Perplexity. Their headline split: listicles get 21.9% of citations across all queries, articles 16.7%, product pages 13.7%. Inside that average, intent reverses everything. For informational queries, articles are cited 2.7× more than other formats and reach 45.5% share. For commercial queries, listicles jump to 40%. Format choice is intent-conditional — that's the empirical foundation of this article.

Glen Allsopp's 750-prompt, 26,283-URL ChatGPT study gave the strongest single number on listicles: 'Best X' lists were 43.8% of all cited page types and 51.4% in the software vertical. Allsopp's prompt set was top-of-funnel commercial, which is why this number is so high — not because ChatGPT loves all listicles, but because his prompts triggered the cell where listicles win.

Profound's 680M-citation analysis exposed the platform variance that makes the third axis necessary. Wikipedia is 7.8% of all ChatGPT citations and 47.9% of its top-10 sources. On Perplexity, Reddit is 46.7% of top-10 sources. ChatGPT cites Wikipedia roughly 80× more than Google AI Mode does. The same content does not perform the same way on different engines — the grid has to be 3-dimensional.

Seer Interactive tracked 5.47M queries and reported that AI Overviews triggers on 95.4% of comparison-format queries — the single highest trigger rate of any query class. Their related dataset showed ChatGPT's listicle citations declined 30% month-over-month December 2025 to January 2026, while AI Overviews' listicle citations stayed flat or rose. That's the cell-level resolution this article is about: 'listicle' is not one answer.

Princeton's GEO paper (KDD 2024, Aggarwal et al.) provides the format-agnostic baseline. Adding statistics improved citation visibility by 32.8%, citing sources by 27.7%, and adding quotations by 42.6%. None of those depend on format. Whichever cell you optimize for, these three modifications stack on top.

The 60-cell grid

The matrix below maps each format-intent-platform combination to a 1–5 fit rating, with a one-line note and a confidence flag. Five means the format wins that cell; one means it loses outright. High confidence means at least one named primary study supports the rating directly; medium means it's inferred from related data; low means it's reasoned from first principles or single-source.

Read the grid by row to find your format's best cells, or by column to find which format owns a query intent on a given platform. The largest concentration of fives sits along three diagonals: listicles → commercial, comparison pages → transactional, glossary pages → informational + navigational.

5 · Wins4 · Strong3 · Competitive2 · Weak1 · LosesConf: low / med / high
ChatGPT
AI Overviews
Perplexity
Info.
Comm.
Tran.
Navi.
Info.
Comm.
Tran.
Navi.
Info.
Comm.
Tran.
Navi.
Listicle / Best-of roundup
high2 / 5Loses to articles 2.7×.
high5 / 543.8% of cited types.
med3 / 5Beat by comparison pages.
high1 / 5Wrong format — brand pages win.
high2 / 5Articles dominate (45.5%).
high4 / 5Stable / rising in AIO.
high3 / 5Comparison wins on 'vs' queries.
high1 / 5Wrong format — brand pages win.
med2 / 5Reddit + research win.
med3 / 5Outperformed by comparison + UGC.
med2 / 5G2 / Capterra preferred.
high1 / 5Wrong format — brand pages win.
Comparison page (X vs Y)
med2 / 5Too narrow for awareness.
high4 / 5Strong, second to listicle.
high5 / 5Cleanest extraction at decision.
med2 / 5Brand pages preferred.
high2 / 5Articles + how-tos win.
high4 / 5AIO trigger 95.4% on 'vs'.
high5 / 5AIO triggers near-universally.
med2 / 5Brand pages preferred.
low3 / 5Cited if research-flavoured.
high4 / 5Top high-performing format.
high5 / 5G2-style pages dominate.
med2 / 5Brand pages preferred.
How-to / tutorial
high4 / 5Strong with clean H2 / Q-blocks.
med2 / 5Cited only when mid-funnel.
med2 / 5Wrong tier — implementation comes after.
med2 / 5Brand-specific docs preferred.
high5 / 5HowTo schema = 2.8× citations.
med2 / 5Listicle / comparison preferred.
med2 / 5Wrong tier.
med2 / 5Brand-specific docs preferred.
med4 / 5Implementation guides cited often.
low3 / 5Cited as supporting source.
med2 / 5Wrong tier.
med2 / 5Brand-specific docs preferred.
Research / data essay
high5 / 5Citation magnet with proprietary data.
med3 / 5Cited as evidence beneath listicles.
med2 / 5Too abstract at the decision.
high1 / 5Wrong format.
high5 / 53+ data points = 4× citations.
med3 / 5Cited as evidence beneath comparisons.
med2 / 5Too abstract at the decision.
high1 / 5Wrong format.
high5 / 5Citation behaviour favors research.
med3 / 5Cited as evidence beneath comparisons.
low3 / 5Cited if directly compares offerings.
high1 / 5Wrong format.
Glossary / definition
high5 / 5Wikipedia = 47.9% of top-10.
med3 / 5Cited when defining the category.
med4 / 5Owns entity-defining buyer queries.
med4 / 5Cited when naming-the-thing matters.
high4 / 5Strong on 'what is' queries.
med3 / 5Cited when defining the category.
low3 / 5Cited but not centred.
med4 / 5Cited when naming-the-thing matters.
med4 / 5Cited as definitional anchor.
med3 / 5Cited when defining the category.
low3 / 5Cited as supporting source.
med4 / 5Cited when naming-the-thing matters.

Three patterns are worth calling out. First, the navigational column is mostly twos — every format loses navigational queries to first-party brand pages, which is the correct outcome. Second, comparison pages are the only format with a complete diagonal of fives across all three platforms for transactional intent — they are the cleanest bottom-of-funnel format and the most under-built one. Third, glossary pages quietly score four-and-five across informational and navigational on every platform — and almost no B2B brand is running them.

The format hiding in the grid: glossary pages

The most under-deployed AI citation format is the glossary or definition page — the one that owns the 'what is X?' query for a category-defining term. It's the playbook Wikipedia has run for two decades and Investopedia has run for one. Almost no B2B SaaS brand runs it.

The reason it works for AI engines is structural. Generative models need to disambiguate entities before they can answer questions about them. The page that defines the entity authoritatively becomes the page they reach for first — and the brand attached to that page becomes a default reference in answers about the category. That's not a clever optimization. It's how retrieval works.

Wikipedia is 47.9% of ChatGPT's top-10 sources across 680M citations.

Profound · 2025

ChatGPT cites Wikipedia roughly 80× more than Google AI Mode does.

BrightEdge · 2025

Investopedia is the entity-defining citation for finance queries — classified as a 'Power Player' across 109K finance citations.

Goodie · 2025

Wikipedia citation share peaked at 14% on ChatGPT in March 2025.

BrightEdge · 2025
  1. Pick the entity you want to own. It should be a term whose definition affects whether your product is included in answers — for an AI search visibility tool, that term might be 'generative engine optimization' or 'AI citation rate'. If the definition lives somewhere other than your domain, you do not own the entity.
  2. Write a single canonical definition page. One H1 in the form 'What is X?'. The first sentence answers it in under 30 words, with no marketing language and no internal links. This sentence is what the model lifts.
  3. Add `DefinedTerm` schema. The Schema.org type exists; almost no B2B site uses it. It signals to AI engines that the page is making a definitional claim, not a marketing claim.
  4. Cite three to five external authorities inline. Princeton GEO showed citing sources improves AI visibility by 27.7%; on a definition page it also borrows credibility for the underlying claim.
  5. Update the page on a calendar, not a whim. Definitions decay; freshness signals (dateModified, visible 'last updated' line) are scored.
  6. Cross-link from every adjacent piece of content using the entity by name. The definition page becomes the internal authority for that term across the site, and it accumulates the citations that prove the claim.

Three positions worth taking seriously

The grid is a present-tense map. Three credentialed dissenters argue that any format-level map is the wrong abstraction. They are partly right — the grid is not the whole picture — and worth addressing directly.

  1. Format is downstream of site quality

    Glenn Gabe (GSQi)

    ArgumentAI engines will eventually run a Panda equivalent — a site-level quality system that flags thin or templated listicles regardless of format. Tactical format wins are short-lived; only durable site quality matters.

    RebuttalProbably true on a long enough horizon. The grid describes the present and the next two to three years; site-level quality scoring is a forecast, not a measurement. Build for the cell you're in now while keeping site quality high.

  2. Entities, not formats, decide citations

    Mike King (iPullRank)

    ArgumentAI retrieval is entity-driven. The format is a vehicle; the entity claim inside is what gets surfaced. Optimize entities, schema, and disambiguation — format is secondary.

    RebuttalAgree on direction, disagree on substitution. Format is the vehicle through which entity claims become extractable — that's why glossary pages work. Entity-first and format-by-intent are stacked, not competing.

  3. Freshness eats format

    Ahrefs freshness study; Surfer's December 2025 scraped-vs-API research

    Argument76.4% of ChatGPT's top-cited pages are updated within 30 days. Whichever cell you sit in, freshness is the dominant factor — the format-intent fit is a multiplier, not the engine.

    RebuttalFreshness applies to every cell equally. It is a multiplier; the grid tells you which cell to multiply. A fresh listicle aimed at an informational query still loses to an article. A stale comparison page still beats a fresh how-to at the moment of decision.

How to use the grid on your own site

Audit your existing content against the grid before you write a new piece. Most teams have a stack of pages aimed at the wrong cell — that's the cheapest fix.

Audit checklist
  1. List your top 20 pages by traffic or conversion intent. Pull from your CMS, GSC, or analytics. Don't filter by format yet.
  2. Classify each page in the 5 × 4 grid. Pick one of the five formats and one of the four intents per page. If a page does not fit any cell cleanly, that is a finding — the page is doing too much and probably winning none.
  3. Find your distribution. Most B2B sites cluster around two cells: listicle × commercial and how-to × informational. The under-built cells are usually comparison × transactional, glossary × informational, and research essay × informational — the three cells where the grid says you can earn the most citations.
  4. Pick one under-built cell to fill first. Comparison page if you sell a product with named competitors. Glossary if you operate in a category whose terminology is still being defined. Research essay if you have access to first-party data nobody else has.
  5. Apply the format-agnostic Princeton stack on every page. Add citations (+27.7%), statistics (+32.8%), and quotations (+42.6%). These work across every cell.
  6. Re-check after 30, 60, 90 days. AI citation share moves on a monthly cadence — Seer found 30% MoM swings on ChatGPT alone. The grid is not static; track which cells your pages actually land in, not which ones you intended.

Frequently asked questions

Q · 01

If I had to publish only one format, which one wins the most citations across the grid?

Comparison pages. They are the only format with a complete row of fives across the three platforms for transactional intent (Seer's 5.47M-query analysis shows AI Overviews triggers on 95.4% of comparison queries) and they hold strong fours across commercial intent on every platform. They lose informational and navigational, but those are the smallest cells for most B2B sites.

Q · 02

Are listicles dying in AI search?

No. Listicle citations on ChatGPT declined 30% month-over-month from December 2025 to January 2026 (Seer Interactive), but the same period saw AI Overviews listicle citations stay flat or rise. The decline was concentrated in self-promotional listicles (where the publisher ranked their own product first without disclosed methodology) — a Peec AI analysis of 232,000 citations found self-promotional listicles are only 11% of all citations. Honest listicles aimed at commercial intent still own the strongest single cell on the grid.

Q · 03

Why do you treat comparison pages as a separate format from listicles?

Because they fit different cells. A 'best 10 X' listicle scores its highest fives at commercial intent on ChatGPT. A 'X vs Y vs Z' comparison page scores its highest fives at transactional intent on every platform — and Google's AI Overviews triggers on 95.4% of comparison-format queries versus a much lower rate for top-of-funnel listicle queries. Most published studies collapse the two and lose the signal.

Q · 04

What's the strongest evidence that intent matters more than format?

Wix Studio's 75K-answer analysis found that for informational queries, articles get 45.5% of citations and listicles get 21.7%. For commercial queries, listicles jump to 40% and articles drop. Same content types, opposite outcomes — the only variable is intent. Print this on a wall.

Q · 05

Where does video fit in the grid?

Outside it. Video is its own citation tier — Surfer's 36M-AIO study found YouTube is the single most-cited domain across many verticals (23.3% of all AIO citations). The grid in this article is for text content. Treat video as a parallel system with its own grid.

Q · 06

If glossary pages are this good, why aren't B2B brands running them?

Three reasons. They don't convert directly — the page sells nothing, so it gets deprioritized. They take time to compound — Wikipedia spent 20 years becoming the default. And the playbook is unsexy: write a clean definition, add `DefinedTerm` schema, cite three external sources, update on a calendar. Most content teams will publish a 'best of' listicle this quarter instead. That asymmetry is the entire opportunity.

  • Wix Studio AI Search Lab — 75K AI answers, 1M+ citations, format × intent breakdownsearchengineland.com
  • Glen Allsopp / Detailed (Ahrefs) — Do self-promotional 'Best' lists boost ChatGPT visibility? 750 prompts, 26,283 URLsahrefs.com
  • Profound (Nick Lafferty) — AI Platform Citation Patterns across 680M citationstryprofound.com
  • Seer Interactive — The listicle window is closing in AI search (30% MoM decline)seerinteractive.com
  • Aggarwal et al. — GEO: Generative Engine Optimization (IIT Delhi + Princeton, KDD 2024)arxiv.org
  • Surfer — Scraped AI answers vs API results (1,000-prompt study, December 2025)surferseo.com
  • Surfer — AI Citation Report 2025 (36M AI Overviews, 46M citations)surferseo.com
  • BrightEdge — How Google AI Overviews and ChatGPT cite Wikipedia differentlybrightedge.com
  • Peec AI — Self-promotional listicles analysis from 232K citationspeec.ai
  • Goodie AI — Most cited B2B SaaS domains in AI search (5.7M citations)higoodie.com
  • Kevin Indig — ChatGPT citations: 44% come from the first third of content (1.2M responses, 18K citations)almcorp.com
  • Ahrefs — Short vs long content in AI Overviews (174K pages, 1.68M cited URLs)ahrefs.com
  • Schema.org DefinedTerm — the structured-data type for definition pagesschema.org

Audit your pages against the grid.

TurboAudit checks the structural and entity signals that determine which cell of the format-intent grid your pages actually land in — schema, author attribution, content extractability, freshness, citation density, and the on-page factors AI engines use to decide whether to include you. From $49/mo.

Try TurboAudit