How to Measure Your AI Citation Rate (Step-by-Step)
Your AI citation rate is the percentage of topic queries where AI search engines name your site as a source. You can measure it manually in under an hour: pick 20 queries in your niche, run them in ChatGPT Search and Perplexity, count how many return your domain, divide by 20. That number is your baseline. Most sites have never checked. When we ran this process across six well-known domains as part of building citability.dev, the results were clarifying.
What Does Real Baseline Data Look Like?
Before building the AVR (AI Visibility Readiness) framework, we ran citation rate tests across six domains with varying authority and content strategies. Here is what the baseline showed across 20 topic queries each on ChatGPT Search and Perplexity:
| Domain | Domain Type | Avg Citation Rate | Notes | |---|---|---|---| | ahrefs.com | SEO tool / content publisher | 68% | Dense structured guides, frequent updates | | semrush.com | SEO tool / content publisher | 61% | High query volume coverage, strong schema | | reddit.com | Community platform | 45% | Volume-driven, user-generated answers | | medium.com | Publishing platform | 32% | Mixed quality, no consistent schema | | vercel.com | Developer tool / docs | 28% | Strong on technical queries, thin on general topics | | chudi.dev | Personal site / consulting | 0% | No topic query citations at baseline |
The 0% for chudi.dev is not an edge case. It is the starting point for most personal sites and small businesses. The infrastructure audit revealed why: no per-page structured data, no FAQPage or HowTo schema, content that answered questions but not in AI-readable format.
The Ahrefs result (68%) reflects years of deliberately structured content: clear headings, definitive answers in the first paragraph, and consistent use of Article and FAQPage schema. They did not optimize for AI citations specifically. They optimized for clarity, and that clarity translated directly to AI citability.
Step 1: Define Your Topic Query Set
Do not test with branded queries ("chudi nnorukam consulting" or "citability.dev review"). Those test whether AI knows you exist, not whether AI cites you as an authority on your topics.
Instead, write down the 10-20 questions your target readers are asking that your content should answer. For a site like citability.dev, that means queries like:
- "how to improve AI citability"
- "why is my site not cited by ChatGPT"
- "what is answer engine optimization"
- "how do AI search engines choose sources"
- "how to get cited by Perplexity"
Group them into 3-5 clusters by intent. Topic clusters matter because citation rate can vary dramatically by intent type. Informational queries ("what is X") often show different citation patterns than procedural queries ("how to X") or comparative queries ("X vs Y").
Step 2: Run the Manual Query Test
Open ChatGPT with web search enabled (the globe icon). Run each query. For each response, look for cited sources in the footnotes or sidebar. Check whether your domain appears. Log it as yes or no.
Repeat for Perplexity (which shows sources prominently as cards) and Claude.ai (check for cited URLs in responses).
The logging format does not need to be complex:
Query: "how to measure AI citability"
ChatGPT: no citation
Perplexity: no citation
Claude: no citation
Query: "what is structured data for SEO"
ChatGPT: schema.org cited, google.com cited
Perplexity: search.google.com cited
Claude: no web citations returned
Twenty queries across three platforms gives you 60 data points per platform after normalization. Expect significant variation. Perplexity typically cites more sources per response than ChatGPT. Claude in standard mode does not browse the web by default, so limit those tests to Claude Search or note when web access is enabled.
Step 3: Calculate Your Citation Rate
The math is straightforward. Count how many of your queries returned your domain as a cited source. Divide by total queries. Multiply by 100.
queries_tested = 20
times_cited = 3
citation_rate = (times_cited / queries_tested) * 100
# Result: 15.0%
Calculate this separately per platform since rates diverge. A site might appear in 25% of Perplexity results and 5% of ChatGPT results, which points to different structural issues (Perplexity weighs recency and structured data; ChatGPT Search weighs Bing index signals and domain authority).
For a 95% confidence interval on your measured rate, use the Wilson score interval. At n=20 and 15% measured rate, your true rate is likely between 5% and 36%. That is a wide range. At n=50 with 15% measured, the CI narrows to 7%-28%. This is why sample size matters before making infrastructure changes.
Step 4: Run the Infrastructure Scan
Manual citation testing tells you your current rate. Infrastructure scanning tells you why.
The citability.dev free scan checks 10 signals across four categories:
- Crawl access: robots.txt permissions for GPTBot, ClaudeBot, PerplexityBot. Sitemap existence and validity.
- Rendering: Whether your content is visible in raw HTML (server-side rendered) or requires JavaScript execution to appear.
- Structured data: JSON-LD schema coverage at site level and per-page level. Presence of Article, FAQPage, HowTo markup.
- Content signals: Answer-first layout, heading hierarchy, freshness indicators (dateModified schema, recent publication dates).
Cross-referencing your citation rate with your infrastructure score reveals the pattern. Sites with 8+ infrastructure checks passing and low citation rates typically have a content structure problem: technically accessible but not answer-formatted. Sites with low infrastructure scores and low citation rates have the simpler problem: fix the infrastructure first, then reassess.
Step 5: Interpret the Gap Between Platforms
Citation rate gaps between platforms are diagnostic. Here is what the gaps typically mean:
High Perplexity, low ChatGPT: Your content is indexed but your Bing authority is weaker than your broader web footprint. Perplexity uses multiple sources including direct crawls; ChatGPT Search relies more heavily on Bing's index. Backlink building and Bing Webmaster Tools registration help close this gap.
High ChatGPT, low Perplexity: Your content ranks well in Bing but may lack the structured data and real-time freshness signals Perplexity weights. Adding FAQPage schema and updating content within the last 90 days helps.
Low on all platforms: Infrastructure or content structure problem. Run the scan before changing anything else.
Consistent 0% across all platforms: You are almost certainly blocked by robots.txt, not indexed by Bing, or your content is client-side rendered without SSR. These are fixable in hours.
Step 6: Repeat and Track
AI citation rates are not static. Major model updates (GPT-4o, Claude 3.5 Sonnet, Gemini 2.0) shift citation patterns. When OpenAI updated its web search integration in late 2025, dozens of sites saw citation rates shift by 15+ percentage points in either direction.
Set a quarterly calendar reminder. Run the same 20 queries each quarter using the same logging format. Track rate by platform and by topic cluster. Quarterly data over two or three cycles reveals which content types your site gets cited for and which gaps remain.
The query set should stay consistent so you are measuring the same thing each quarter. If you change your query set, note it in your log so you can account for the change when comparing periods.
What to Do With Your Baseline
A measured baseline is only useful if it triggers action. The decision tree is simple:
- Citation rate above 40%: Your citability infrastructure is working. Focus on expanding topic coverage.
- Citation rate 15-40%: Infrastructure is partially working. Run the scan, fix the failing checks, retest in 30 days.
- Citation rate below 15%: Treat this as a structural problem. Start with the infrastructure scan, fix crawl access and structured data first, then address content structure.
- Citation rate 0%: Diagnose before anything else. Check robots.txt for GPTBot blocks. Check if your content is SSR or client-rendered. Check Bing Search for your domain. Fix the blocking issue before any content work.
The guide I wish I had when I started this: measure first, then fix. Every hour spent "optimizing for AI" without a baseline is potentially wasted effort. The baseline tells you which lever to pull.
Run the free citability scan on your domain to get your infrastructure score alongside your manual citation rate baseline. The scan takes under two minutes and shows exactly which signals are failing and why they matter for AI citation likelihood. You will have a clearer picture of your AI visibility in the time it takes to read a single blog post.
The methodology behind the scan is documented at citability.dev/methodology if you want to understand the scoring criteria before running it.