A brand's measurable presence across AI platforms (ChatGPT, Perplexity, Gemini, AI Overviews). Replaces impressions as the key MR metric.
AI Visibility Score is a composite metric that measures how often and how strongly a brand appears across AI answer surfaces for a defined set of category queries. It replaces traditional impression counting — which captures almost none of the real value in AI-mediated discovery — with a durable performance baseline that teams can track over time, benchmark against competitors, and tie directly to pipeline influence.
The concept is simple: when buyers ask AI engines who to use for a given job, how often does your brand get mentioned? AI Visibility Score answers that question with a number that moves in response to the right interventions.
AI engines have replaced Google as the primary vendor research tool for enterprise buyers. Forrester's B2B Buyers' Journey Survey (2026) found that AI engine consultations now precede 73% of shortlist decisions in the enterprise software and professional services categories. That means a brand's position in AI-generated answers is the first filter buyers apply — before a website visit, before a sales call, before a demo request.
Without a score, teams are blind to that filter. They track organic traffic (which drops in a zero-click world), branded search (which is a lagging indicator), and media mentions (which measure output, not outcome). AI Visibility Score measures the actual output of the Machine Relations system: are AI engines recommending the brand or leaving it off the list?
It also creates accountability. Marketing and PR have historically struggled to connect earned media to pipeline. AI Visibility Score provides the intermediate measurement: earned media placements drive citation frequency, citation frequency drives AI Visibility Score, AI Visibility Score drives shortlist inclusion. The causal chain is traceable in a way that "impressions" never was.
A rigorous AI Visibility Score requires five components:
1. Query Set Definition Build a set of 30-50 queries that represent how a real buyer researches the category. Include problem-framing queries ("what is the best way to manage AI brand visibility"), comparison queries ("best AI visibility tools for B2B SaaS"), and vendor-direct queries ("who does [category] well"). The query set is the measurement instrument. A lazy query set produces a meaningless score.
2. Multi-Engine Sampling Run the query set across ChatGPT, Perplexity, Gemini, Claude, and Google AI Overviews. Engine behavior is not uniform — a brand can be strong on Perplexity and absent from ChatGPT. A single-engine score is not an AI Visibility Score; it is a single-engine appearance rate.
3. Citation Recording For each query-engine combination, record whether the brand is cited, how prominently (first mention, embedded, footnote), and what context it appears in (recommended, compared, cautioned against).
4. Scoring The simplest defensible formula: (Number of query-engine combinations with brand citation) / (Total query-engine combinations tested). A brand that appears in 90 out of 250 query-engine combinations holds a 36% AI Visibility Score against that query set.
More sophisticated models weight by engine authority (Perplexity citation may carry more buyer intent than Google AI Overview inclusion), by query intent (bottom-of-funnel queries weight higher), and by citation prominence.
5. Baseline + Trend A score in isolation is a snapshot. AI Visibility Score becomes useful as a trend line. Run the same query set weekly or monthly. The direction of movement — not the absolute number — tells you whether the Machine Relations system is compounding.
| Metric | What It Measures | Level | Use Case |
|---|---|---|---|
| AI Visibility Score | Overall citation presence across all engines and queries | Portfolio | Executive reporting, campaign baselines |
| Share of Citation | Single brand's citation rate | Brand | Performance tracking vs. own baseline |
| AI Share of Voice | Citation share vs. competitors | Category | Competitive positioning |
| Citation Velocity | Rate of new citation accumulation | Brand | Campaign impact, growth speed |
| Entity Resolution Rate | AI confidence in brand identity | Infrastructure | Entity health, Layer 2 diagnosis |
| Recommendation Rate | How often the brand is recommended vs. just cited | Conversion | Bottom-of-funnel AI influence |
AI Visibility Score is the summary metric. The others diagnose why it is moving in a given direction.
Failure mode 1: Single-engine tracking. Running AI Visibility Score only on ChatGPT (or only on Perplexity) is like measuring TV advertising ROI only during prime time on one network. Different engines serve different buyer populations and have different indexing behaviors. A brand invisible on Perplexity is invisible to every buyer who runs their vendor research there — regardless of how it scores on ChatGPT.
Failure mode 2: Optimizing for position, not presence. Appearing first in one AI response is far less valuable than appearing consistently across many responses. AI engine answers are not ranked lists with stable positions — they are dynamically generated responses where position varies by phrasing, session context, and model version. Presence across many query-engine combinations is a more durable and actionable target than first-position in a few.
Failure mode 3: Wrong query set. Using branded queries ("what is [brand name]") instead of category queries ("who should I use for X") produces a score that measures brand recognition in the AI's training data, not buyer discovery behavior. The query set must simulate how a buyer researches the category, not how someone already familiar with the brand would ask about it.
Failure mode 4: Treating it as a marketing metric instead of a business metric. AI Visibility Score is a leading indicator for pipeline, not a branding KPI. Teams that track it in isolation from deal sourcing and buyer journey data miss its purpose. The goal is shortlist inclusion — AI Visibility Score measures the indicator that predicts shortlist inclusion.
Failure mode 5: No cadence. A score measured once is not an operating metric. Without a consistent measurement cadence, teams cannot distinguish signal from noise, attribute changes to specific interventions, or catch competitive encroachment before it compounds.
AI Visibility Score is the summary measurement that Layer 5 of the MR Stack produces. It is the number that tells you whether the first four layers — Earned Authority, Entity Optimization, Citation Architecture, GEO/AEO Distribution — are working together or leaking.
A stagnant or declining AI Visibility Score is not a marketing problem. It is a system diagnosis: one or more layers of the MR Stack are underperforming. The diagnostic path runs: check Citation Velocity (is earned media producing new citations?) then Entity Resolution Rate (is the entity layer clean?) then Share of Citation (is the presence rate per query declining?). The summary score tells you something is wrong. The component metrics tell you where.
---
What is a "good" AI Visibility Score? This depends entirely on category size, query set breadth, and competitive intensity. The more useful framing is relative and directional: is the score growing month-over-month, and is it outpacing competitors? A 25% score growing 5 points monthly in a competitive category is far more valuable than a static 40% in an easy one.
How is AI Visibility Score different from traditional share of voice? Traditional share of voice counts media mentions. AI Visibility Score measures citation presence in AI-generated answers — a fundamentally different mechanism with different inputs and outputs. A brand can have strong traditional share of voice and low AI Visibility Score if its earned media appears in publications that AI engines do not trust or index frequently.
Can AI Visibility Score be gamed? Short-term manipulation is possible — flooding the web with keyword-stuffed content may spike shallow scores temporarily. Durable AI Visibility Scores require the same things that drive actual buyer trust: credible earned media, clean entity signals, and consistently cited expertise. Manipulation that does not produce genuine authority erodes quickly as AI engines refresh their source weighting.
How often should AI Visibility Score be measured? Weekly for active campaigns, monthly for baseline tracking. Daily measurement is unnecessary noise. The right cadence matches the pace at which earned media and entity changes take effect — typically 2-4 weeks per intervention.
AI Share of Voice is the proportion of AI-generated responses where a brand is mentioned, cited, or recommended relative to competitors for a defined set of category queries across ChatGPT, Perplexity, Gemini, Claude, and Google AI Overviews. Distinct from traditional share of voice (media mentions) and search share of voice (ranking visibility), AI Share of Voice measures competitive position in the AI discovery layer.
Citation Decay is the rate at which AI engine citations of a brand decrease over time without sustained earned media activity. AI engines continuously re-evaluate source freshness and authority, and brands that stop generating new high-quality signals see their citation presence erode as competitors produce newer, more relevant content.
The delta between a brand's traditional search ranking and its AI citation frequency. A brand can rank #1 on Google but appear in 0% of ChatGPT answers.
Citation Velocity is the rate at which new AI engine citations accumulate for a brand, typically measured as new citation appearances per week across a monitored query set. Higher velocity indicates active authority growth. Citation Velocity is the offensive counterpart to Citation Decay in the Machine Relations measurement framework.