AwesomeCMO Research · April 2026

Your next customer is asking ChatGPT, not Google.

If the AI never mentions your business, you don't exist. This is the playbook for getting cited by ChatGPT, Perplexity, Google's AI Overviews, Gemini, Claude, and Copilot — without hiring an agency.

TL;DR — AI search works differently from Google. Most sites are accidentally invisible to it. Fix three things — let the AI robots in, put your answer at the top, add the structure they can quote — and you show up. This page has the exact steps.
The acronyms: SEO is what you already know (ranking on Google). AEO is getting picked as the direct answer. GEO is getting cited by generative AI like ChatGPT. Same basic job, three flavors.
0.0×

Visitors who arrive via AI buy 4.4× more often than Google visitors.

0%

Of all the pages ChatGPT reads, only ~15% get quoted. Structure decides who.

0%

Of websites accidentally block the AI robots. They vanish from ChatGPT Search.

0.0×

Brands listed on Wikidata are 3.2× more likely to get cited by Google AI.

0

AI crawlers run zero JavaScript. If your site needs JS to show text, it's invisible.

Keep reading
The whole picture

What happens between a question and an AI answer

If you only read this panel, here's the whole thing. The rest of the page is the how.

A structured webpage on the left connects through a funnel of AI retrieval and reranking to a cited answer on the right.
Your site
AI picks a shortlist
You get quoted
Six glowing orbs interconnected, representing six AI search engines.

Six AI search engines, not one

ChatGPT, Google's AI Overviews, Perplexity, Gemini, Claude, and Microsoft Copilot. Each one picks its sources differently. A winning strategy works across all six, not just Google.

A structured document with upward-trending growth arrow and quote marks.

Structure beats volume

Adding real numbers to your page raises your citation rate by 41%. Expert quotes add another 28%. FAQ sections triple your chances. The AI rewards pages written to be quoted — not just read.

A radar sweep with two detected signal dots, representing citation tracking.

Measure what AI sees

AI traffic is only about 1% of your visits today — but it converts 4.4× better. You can't improve what you can't see, so track how often you're cited, not just how often you rank.

Keep scrolling for the evidence, the tactics, and the 90-day plan.

How AI answers get built

A four-step pipeline — same for every engine

Every AI search engine follows roughly the same recipe. Understanding it is the whole game. Here's what happens when someone asks ChatGPT about your industry.

1

The AI searches

It runs a normal web search. ChatGPT uses Bing, Gemini uses Google, Claude uses Brave. Your first job is to rank well on whichever search engine feeds the AI.

2

It picks a shortlist

The AI scores each result on freshness, trust, and how well-structured the page is. Most pages get dropped here — only the top handful move on.

3

It breaks pages into quotable chunks

The AI splits each winner into short passages, one topic per paragraph. Pages with clear headings and a summary at the top survive this step. Walls of text don't.

4

It writes the answer and credits you

The AI assembles its reply from the best chunks, and adds a little citation link next to the ones it used. Only about 15% of pages that made the shortlist actually get quoted.

Only ~15% of the shortlist gets quoted.

Getting on the shortlist is the easy part — Google already does that if you rank. Getting quoted is the new game. For ChatGPT specifically, the first 30% of your page drives almost half of all citations. Bury the lede and you lose.

15%

The AI searches. It runs a normal web search. ChatGPT uses Bing, Gemini uses Google, Claude uses Brave. Your first job is to rank well on whichever search engine feeds the AI. It picks a shortlist. The AI scores each result on freshness, trust, and how well-structured the page is. Most pages get dropped here — only the top handful move on. It breaks pages into quotable chunks. The AI splits each winner into short passages, one topic per paragraph. Pages with clear headings and a summary at the top survive this step. Walls of text don't. It writes the answer and credits you. The AI assembles its reply from the best chunks, and adds a little citation link next to the ones it used. Only about 15% of pages that made the shortlist actually get quoted.

What actually moves the needle

Five content moves with measured impact

Princeton researchers tested specific tactics and measured how much each one raised citation rates. Industry follow-ups replicated the findings. These are the ones worth doing.

Link out to authoritative sources

+30 to +115%

When you cite real sources in your content, AI treats your page as corroborated and quotes it more often. The effect is biggest for pages that don't already rank highly.

Use real numbers and statistics

+41%

AI treats numbers as facts it can safely quote. A page with 'trusted by thousands' gets ignored; a page with '3,247 customers' gets cited.

Include expert quotes

+28%

A direct quote from a credentialed person is the highest-confidence signal an AI can find. If you don't have experts on staff, interview customers or industry peers.

Show when the page was last updated

+18 to +22%

AI strongly prefers fresh content. Add a 'Last updated' line at the top and the matching `dateModified` tag in the page code. Then actually update — not just the date.

Put the answer in the first paragraph

+18%

Almost half of ChatGPT's citations come from the first 30% of a page. Lead with the direct answer in 40 to 60 words. Put your backstory below.

3.2×
Pages with FAQ sections get cited 3.2× more

AI loves Q&A format — it matches how people ask questions. Add 3–5 real customer questions with direct answers at the bottom of each important page.

Keep one real statistic per 150–200 words, sentences under 10 words on average, and the answer in the first 60 words. Pages below this density get judged as fluff and demoted.
First, unlock the door

Let the AI robots in

Before any of this matters, the AI has to be able to read your site. Many sites block the AI robots by accident — often via a default setting from years ago. Fixing it takes five minutes.

Training robots

These visit your site to help train the next model version. Blocking them protects your content from being absorbed. Allowing them means you might be remembered by future models as a default answer.

GPTBotOpenAI
ClaudeBotAnthropic
Google-ExtendedGoogle
BytespiderByteDance
CCBotCommon Crawl

Live-search robots

These visit your site the moment someone asks the AI about your topic. Block any of these and you disappear from that AI's answers — today. Always allow these.

OAI-SearchBotOpenAI
ChatGPT-UserOpenAI
Claude-WebAnthropic
Claude-SearchBotAnthropic
PerplexityBotPerplexity
Perplexity-UserPerplexity
Applebot-ExtendedApple
Meta-ExternalAgentMeta
robots.txt — copy this as your starting point
# Allow all retrieval crawlers — blocking these makes you invisible
# in ChatGPT Search / Claude / Perplexity / Apple Intelligence / Meta AI

User-agent: OAI-SearchBot
Allow: /
User-agent: ChatGPT-User
Allow: /
User-agent: Claude-Web
Allow: /
User-agent: Claude-SearchBot
Allow: /
User-agent: PerplexityBot
Allow: /
User-agent: Perplexity-User
Allow: /
User-agent: Applebot-Extended
Allow: /
User-agent: Meta-ExternalAgent
Allow: /

# Training crawlers — decide per-brand (distribution vs. IP protection)
User-agent: GPTBot
Allow: /
User-agent: ClaudeBot
Allow: /
User-agent: Google-Extended
Allow: /
The opportunity hiding in robots.txt Right now ~62% of sites block GPTBot and ~69% block ClaudeBot — usually by accident. In most industries, your competitors are invisible to the AI. The brands that allow these bots are disproportionately showing up in answers. First-mover window.
How each AI picks sources

Six engines, six slightly different games

The basics are the same everywhere. The differences are in which third-party sites each AI trusts most, and what tracking tools they give you. Here's the short brief for each.

Where it looks for answers

The AI answer Google writes above normal results comes from the top 10 organic results ~76% of the time. Your regular Google SEO still does almost all the work — AI Overviews is the ceiling, not a new foundation.

What earns the quote

  • Original research or data you own — anything the AI can't find elsewhere
  • Clean tables, numbered steps, and definition blocks (AI loves parsing these)
  • Code-level structured data (schema) that matches what visitors actually see on the page
  • Google's 'query fan-out' means a single question triggers multiple retrievals — so deep topic coverage wins

Sources it tends to cite

  • YouTube and Google Maps
  • Wikipedia
  • Established publishers

How to see your score

Google Search Console (deltas), Semrush AI Visibility Toolkit, Ahrefs Brand Radar

Tap a logo to swap playbooks. Your move doesn't change — just the evidence for it.

Where citations actually come from

You don't always win by writing more — you win by appearing in more places

For 'best X for Y' questions, third-party sites dominate: Reddit, YouTube, Wikipedia, G2. For feature and pricing questions, vendor docs dominate. Plan distribution across both. These percentages shift monthly — the point is the pattern, not the exact number.

Choose a question type

ChatGPT
  • Wikipedia48%
  • Established publishers22%
  • Reddit12%
  • YouTube9%
  • G2 / Capterra reviews5%
  • Other4%
What this means: ChatGPT trusts Wikipedia more than any other source. Getting a Wikipedia page for your brand (through earned media, not self-editing) is still the single highest-leverage move for ChatGPT visibility.
The performance floor

Slow pages get dropped before they're even considered

AI crawlers give your page 1 to 5 seconds to load before they give up. Google has published exact speed thresholds you need to clear. Past those thresholds, more speed doesn't buy more citations — but failing them kills you.

LCP
Largest Contentful Paint
≤ 2.5s
How long before the main thing shows up

This is the moment the biggest visible element (usually your hero image or headline) finishes loading. AI pages with an LCP under 2.5 seconds get cited 2.8× more.

INP
Interaction to Next Paint
≤ 200ms
How snappy the page feels when you click

Click a button, tap a menu, type in a field. INP measures how long the page takes to respond. Under 200ms feels instant. Above 500ms feels broken.

CLS
Cumulative Layout Shift
≤ 0.1
How much stuff jumps around while loading

If your page moves content around while loading (ads squeezing in, images loading with no size reserved), visitors rage-quit. CLS should be basically zero.

INP replaced FID in March 2024 as the interactivity metric. Guides still listing First Input Delay are out of date — ignore them.

The 90-day plan

What to actually do, week by week

Everything above, compressed into an ordered checklist. Ranked by leverage per hour of work. Click a phase to open its full list.

  • Open your robots.txt and allow every AI search bot (OAI-SearchBot, ChatGPT-User, Claude-Web, Claude-SearchBot, PerplexityBot, Perplexity-User, Applebot-Extended, Meta-ExternalAgent). Five-minute job.
  • Test your site with the AI's eyes: View Source on any page. If your headline, pricing, or product description is missing from the raw HTML, the AI can't see it either. Fix that first.
  • Create a Wikidata entry for your brand. List founders, category, founding date, social links. Free, takes 15 minutes. The single biggest lever for Google AI Overview citations.
  • Add Article + Person + Organization schema to every main page. Most CMSes have a one-click plugin — just turn it on.
  • Start counting citations. Pick 50–100 questions your customers actually ask, run them through ChatGPT and Perplexity weekly, track when your brand gets mentioned. This is your baseline.
Myths that waste money

Nine things the internet repeats that are wrong

When a blog post disagrees with Google, Apple, or OpenAI's own documentation — trust the primary source. Tap a card to flip it.

You need a special 'AI-friendly' schema. Reality: Google explicitly says there's no such thing. The same Article / Organization / FAQPage schema that helped with Google SEO is what the AI wants too. Just apply it consistently. Blocking GPTBot keeps you out of ChatGPT. Reality: Half true. GPTBot is the training robot — blocking it protects your content from being absorbed into future models. But ChatGPT Search uses a different robot (OAI-SearchBot). Block that one and you disappear from live answers. llms.txt is the new robots.txt. Reality: Not yet. Studies show zero measurable lift for marketing pages. If you run a developer product with heavy documentation, llms-full.txt can help — agents load your docs in one fetch. For everyone else, skip it. Canonical tags stop AI from citing duplicates. Reality: Only partially. AI retrievers often grab non-canonical URLs anyway. If two pages have near-identical content, consolidate them at the URL level — don't just add a canonical tag. AI crawlers read JavaScript like Googlebot. Reality: No. A study of 500 million AI crawler fetches found zero JavaScript execution. If your content only appears after JS runs, the AI sees an empty page. Server-side rendering is mandatory. Higher domain authority = more AI citations. Reality: Weakly true. Domain authority still matters, but 'entity authority' is replacing it fast. Brands that are on Wikidata, have Organization schema, and publish verified-author content beat high-DA sites that don't. GEO is so different that SEO doesn't matter. Reality: Backwards. Google AI Overviews cites the top-10 results 76% of the time. ChatGPT runs on Bing. Gemini is Google-grounded. Classic search engine optimization is the floor; AI optimization is the ceiling. For apps, ASO is all you need. Reality: For the App Store, yes. For 'best X app for Y' questions asked to ChatGPT or Perplexity, the AI barely reads App Store metadata — it cites tech reviews, Reddit threads, and YouTube. Apps need a public web presence. First Input Delay (FID) is still a Core Web Vital. Reality: No. Google replaced FID with INP (Interaction to Next Paint) in March 2024. Any guide still citing FID is at least two years out of date — be suspicious of anything else it says. Google and Bing snippet controls are different. Reality: They used to be. In 2025 Bing added support for the `data-nosnippet` attribute across both regular search and AI answers. Now one tag controls both — useful for hiding out-of-date prices or legal boilerplate from AI quotes.

How this maps to AwesomeCMO

The playbook, automated

Everything above is a checklist. AwesomeCMO runs it on your site weekly — scoring your structure, tracking your citations across six engines, and flagging when something drifts. Here's what each agent does.

Scan

Audits your site every week: are AI robots allowed, is server-side rendering working, is schema installed, are you on Wikidata. Week-over-week diff so you see drift.

SEO

Tracks your citation share across ChatGPT, Perplexity, Gemini, and Claude for the questions your customers actually ask. Target 15% share of voice; category leaders sit at 25–30%.

Content

Scores every draft against the Princeton tactics: enough statistics, expert quotes, first-paragraph answer, FAQ block, outbound citations. Rejects drafts below threshold.

Competitor

Watches your competitors' Wikidata edits, G2 profile changes, new Reddit threads, new YouTube videos. Alerts when a competitor earns ground you didn't.

Report

Weekly digest: how citations shifted, which pages earned or lost them, what AI engines are currently saying about your brand. One screen per week.

Hallucination

Samples common questions weekly, detects when the AI says something wrong about your brand (wrong pricing, missing features, invented incidents), and generates the correction checklist.

◉ system.init

Run a free GEO scan on your brand

AwesomeCMO checks every item in this guide against your live site and shows you what to fix first. Takes 5 minutes. No credit card.

Start free