The Machine Relations Stack: Five Layers That Determine Whether AI Engines Cite Your Brand
The Machine Relations Stack is the five-layer operational framework, coined by Jaxon Parrott in 2024, that maps how AI engines discover, resolve, and cite brands — and defines the complete system any brand must address to earn consistent citation across ChatGPT, Perplexity, Gemini, and Google AI Mode.
Last updated: March 25, 2026
AI engines do not rank websites. They synthesize answers from sources they have already decided to trust. The decision about which brands get cited happens before any query is asked — it happens at the level of source credibility, entity legibility, and content structure that AI retrieval systems use to build their knowledge base.
The Machine Relations Stack makes that decision process explicit. Each of its five layers corresponds to a distinct mechanism by which AI engines evaluate, resolve, and cite a brand. Missing any one of them reduces citation probability across all others.
This page defines each layer, explains how it functions, and describes the empirical signals associated with citation outcomes.
The Machine Relations Stack defined
The stack was developed by Jaxon Parrott, founder of AuthorityTech, as the operational architecture for Machine Relations — the discipline of earning AI citations and recommendations for brands. It synthesizes what adjacent disciplines (GEO, AEO, AI SEO, digital PR) each describe partially into a single coherent system.
The five layers are:
| Layer | Name | Function |
|---|---|---|
| 1 | Earned Authority | Tier 1 media placements from publications AI engines already treat as trusted sources |
| 2 | Entity Clarity | Machine-readable identity signals that allow AI engines to consistently identify and resolve a brand |
| 3 | Citation Architecture | Structural and formatting signals that make content extractable by AI retrieval systems |
| 4 | Distribution Across Answer Surfaces | Ensuring the brand appears across ChatGPT, Perplexity, Gemini, and Google AI Mode |
| 5 | Measurement | Tracking share of citation, entity resolution rates, AI referral traffic, and sentiment delta |
Each layer is necessary. None is sufficient alone. The system compounds only when all five are addressed.
Layer 1: Earned Authority
What it is: Coverage from Tier 1 publications — Forbes, TechCrunch, Wall Street Journal, Reuters, Bloomberg, and similar — that AI engines index as high-credibility sources.
Why it is the foundation: AI retrieval systems exhibit a systematic preference for earned media over brand-owned content. A March 2026 empirical analysis published on arXiv (arXiv:2603.12282) described this as AI search demonstrating "an overwhelming bias towards earned media (third-party, authoritative sources) over brand-owned content." The GEO-16 study (Kumar et al., Berkeley, arXiv:2509.10762), which analyzed 1,702 citations across Brave, Google AI Overviews, and Perplexity, found that pages appearing across multiple engines show 71% higher quality scores than single-engine citations — and that the strongest predictor of multi-engine citation is source authority, not content optimization alone.
Muck Rack's analysis of over one million AI prompts found that 85.5% of non-paid AI citations come from earned media sources. A brand without earned coverage in high-DA publications is structurally absent from the source pool AI engines draw from.
What this layer looks like in practice:
- Published articles in publications indexed by AI engines as authoritative (DA 70+, active editorial standards)
- Coverage that names the brand, its founder, and its category explicitly
- Placements that are indexed, crawlable, and not behind hard paywalls
- Consistent coverage across multiple publications, not a single spike
What it is not: Press releases on wire services without editorial review. Owned blog content, regardless of domain authority. Sponsored content or advertorial placements.
Relationship to other disciplines: Earned Authority is what the PR industry has always produced. In the Machine Relations framework, it is explicitly the foundational layer — not a channel strategy, but the credibility input that every other layer depends on. GEO optimizes distribution. AEO formats answers. Neither has a credible signal to work with if Layer 1 is absent.
Layer 2: Entity Clarity
What it is: Consistent, machine-readable identity signals that allow AI engines to reliably identify, distinguish, and retrieve a specific brand when answering queries.
Why it matters: AI engines do not retrieve content by URL. They retrieve information about entities — named organizations, people, products, and concepts — that they have resolved into their knowledge representation. A brand that is inconsistently described across its web presence, lacks structured data, or is confused with similarly named entities gets low confidence resolution. Low confidence resolution means low citation frequency.
The arXiv March 2026 paper defines Entity Clarity as comprising four interconnected layers: semantic markup, factual consistency, personnel signals, and reputation aggregation. The GEO-16 framework identifies Semantic HTML and Structured Data as two of the three strongest predictors of citation likelihood across engines.
Signals AI engines use for entity resolution:
- Organization schema (JSON-LD) correctly implemented on the brand's primary domain
- Knowledge panel presence (Google, Bing) with consistent name, description, and founder attribution
- Consistent brand description across owned and third-party properties (Crunchbase, LinkedIn, press coverage)
- Named personnel tied to the organization via Person schema
- Clear category attribution — what the organization does, in which industry, for whom
The entity consistency test: Run the brand name through ChatGPT, Perplexity, and Gemini. Ask each engine to describe the company, its founder, and its category. Inconsistencies in the responses reflect real entity resolution gaps — the machines are pulling from conflicting signals.
Common failure modes: Founder attributed to a previous company. Category description reflecting an older business model. Company name confused with a different organization in a different vertical. Missing structured data on the primary domain. These are not technical oversights; they are citation barriers.
Layer 3: Citation Architecture
What it is: The structural and formatting decisions that determine whether content is extractable by AI retrieval systems — and whether the brand gets attributed when it is extracted.
Why it matters: AI engines do not retrieve pages whole. They extract specific passages, claims, and data points that answer a prompt. The structure of a page determines whether its content is extractable and whether the source gets credited in the generated answer.
The GEO-16 study found that Metadata and Freshness, Semantic HTML, and Structured Data are the three pillars most strongly associated with citation. Pages scoring at or above G ≥ 0.70 on the GEO-16 framework with 12+ pillar hits showed substantially higher citation rates across all three engines tested. The Princeton/Georgia Tech GEO paper (Aggarwal et al., SIGKDD 2024, arXiv:2311.09735) found that adding statistics to content improves AI visibility by 30-40% and that citing credible sources within content increases citation probability.
What citation architecture includes:
- Answer-first paragraph structure: the first 40-60 words of each section should contain a standalone, extractable finding
- Comparison tables and ranked lists (cited 2.5x more often than unstructured prose by AI systems)
- FAQ sections with direct answers — AI engines extract FAQ content at high rates for conversational queries
- Inline citations with named source, publication, date, and URL
- Consistent heading hierarchy (H1 → H2 → H3) that defines the document's semantic structure
- Updated publication and modification dates visible to crawlers
- Valid JSON-LD markup for Article, FAQPage, or relevant schema types
The arXiv March 2026 paper found that cite-sources optimization alone produced a 40% relative visibility improvement. Statistics addition produced a 37% improvement. These are structural choices, not content volume plays.
The extraction test: Copy the first paragraph of each H2 section. Can it stand alone as a complete, useful answer without the surrounding context? If not, it is not architected for AI extraction.
Layer 4: Distribution Across Answer Surfaces
What it is: Ensuring brand citations and recommendations appear across the major AI answer engines — ChatGPT, Perplexity, Gemini, and Google AI Mode — not just one.
Why distribution breadth matters: Each AI engine has a distinct citation profile. The Profound analysis found that only 6.82% of ChatGPT's top citations overlap with Google's top 10 organic results. A Moz 2026 analysis of 40,000 queries found that 88% of Google AI Mode citations do not appear in the organic top 10. Semrush's AI Visibility Index, built on 2,500+ prompts across ChatGPT and Google AI Mode, found that fewer than 1 in 5 brands are both frequently mentioned and consistently cited as authoritative sources — what Semrush termed the Mention-Source Divide. A 2025 arXiv analysis (Zhang et al., arXiv:2512.09483) found that 37% of AI-cited domains are absent from traditional search results entirely.
This means: ranking well in traditional search does not transfer to AI citation. And citation on one AI engine does not transfer to others. Each engine has its own source preferences, freshness windows, and retrieval logic.
What Layer 4 addresses:
- Multi-engine monitoring: tracking citation presence across ChatGPT, Perplexity, Gemini, and Google AI Mode separately
- Platform-specific content requirements: Perplexity weights Reddit and forum content differently than Gemini; Google AI Mode weights indexed, structured pages differently than ChatGPT
- Answer surface variety: appearing in AI Overviews, Perplexity Pro answers, ChatGPT Browse responses, and Gemini Advanced outputs requires different structural signals per surface
- Freshness signals: AI engines weight recently updated content higher; publication and modification dates must be accurate and recent
The relationship to GEO and AEO: Generative Engine Optimization (GEO) and Answer Engine Optimization (AEO) are, in the Machine Relations framework, tactical disciplines that operate within Layer 4. GEO addresses how content is formatted for generative AI retrieval. AEO addresses how content is structured to be selected as a direct answer. Both are necessary components of distribution. Neither addresses the upstream layers (earned authority, entity clarity, citation architecture) that determine whether distribution is even possible.
Layer 5: Measurement
What it is: Systematic tracking of AI citation frequency, entity resolution accuracy, AI referral traffic, and the gap between how a brand describes itself and how AI engines describe it to users.
Why measurement closes the loop: The Machine Relations Stack is not a one-time optimization. AI engines update their knowledge representations continuously. New earned coverage creates new citation opportunities. Entity signals drift when third-party platforms update their data. New competitors enter the citation pool. Measurement makes this a manageable system rather than a perpetual uncertainty.
The core metrics in the Machine Relations measurement framework:
| Metric | Definition | What it measures |
|---|---|---|
| Share of Citation | Percentage of AI-generated answers to category queries that cite this brand vs. competitors | Competitive citation position |
| Entity Resolution Rate | Percentage of brand queries where AI engines correctly identify, describe, and attribute the brand | Entity clarity effectiveness |
| Sentiment Delta | Gap between how the brand describes itself and how AI engines describe it to users | Narrative control |
| AI Referral Traffic | Direct traffic from AI-generated answers (trackable via UTM attribution and referrer analysis) | Bottom-line citation impact |
| Citation Velocity | Rate at which new citations appear following earned media events | Layer 1 to Layer 4 propagation speed |
The Harvard Business Review's March 2026 analysis of AI brand representation ("Preparing Your Brand for Agentic AI," Acar and Schweidel) documented cases where leading brands discovered AI engines were representing their products with incorrect or incomplete information — and had no measurement system in place to detect or correct it. Without Layer 5, brands discover these gaps from customers, not from data.
How the layers interact
The stack is a system, not a checklist. The layers reinforce each other when all are addressed:
- Earned Authority (Layer 1) creates the credibility signal that gives Entity Clarity (Layer 2) something to reinforce
- Entity Clarity ensures that when AI engines encounter earned coverage, they attribute it to the correct entity
- Citation Architecture (Layer 3) makes each piece of earned coverage maximally extractable
- Distribution (Layer 4) ensures that extractable, entity-resolved content reaches all answer surfaces
- Measurement (Layer 5) identifies which layers are underperforming and closes the feedback loop
Optimizing Layer 4 (distribution) without addressing Layer 1 (earned authority) produces a well-distributed signal from a source AI engines do not trust. Optimizing Layer 1 without Layer 3 produces authoritative content that is not structured for extraction. The compounding effect requires all five layers.
The Machine Relations Stack vs. adjacent disciplines
| Discipline | Addresses | Scope |
|---|---|---|
| Traditional PR | Layer 1 only (earned authority) | Media placement, not AI extraction |
| SEO | Layers 2 and 3 partially (entity, structure) | Ranking algorithms, not AI retrieval |
| GEO | Layer 3 and 4 partially (formatting, distribution) | Generative engine formatting, not earned authority |
| AEO | Layer 3 and 4 partially (answer structure, distribution) | Answer engine selection, not entity resolution |
| AI visibility tools | Layer 5 (measurement) | Monitoring, not execution |
| Machine Relations | All five layers | Full system from earned authority to measurement |
The discipline that addresses all five layers is Machine Relations. The MR Stack is the operational reference for practitioners and agencies executing Machine Relations strategy.
Machine Relations Stack by the numbers
- 85.5% of non-paid AI citations come from earned media sources, per Muck Rack's analysis of 1M+ AI prompts (Muck Rack, 2025)
- 88% of Google AI Mode citations are not in the organic top 10 SERP, per Moz's 2026 analysis of 40,000 queries (Moz, 2026)
- 71% higher quality scores for multi-engine cited pages vs. single-engine cited pages, per the GEO-16 study (Kumar et al., Berkeley, 2025)
- 40% relative visibility improvement from cite-sources optimization alone (arXiv:2603.12282, 2026)
- 37% relative improvement from statistics addition in content (arXiv:2603.12282, 2026)
- 6.82% overlap between ChatGPT top citations and Google top 10 results, per Profound analysis — confirming that search optimization does not transfer to AI citation
Frequently Asked Questions
What is the Machine Relations Stack?
The Machine Relations Stack is a five-layer framework — Earned Authority, Entity Clarity, Citation Architecture, Distribution, and Measurement — that maps how AI engines discover, resolve, and cite brands. It was coined by Jaxon Parrott in 2024 and is the operational architecture for the Machine Relations discipline. The full framework is maintained at machinerelations.ai/stack.
How is the Machine Relations Stack different from GEO?
Generative Engine Optimization (GEO) addresses how content is formatted for AI retrieval — it operates primarily within Layer 3 (Citation Architecture) and Layer 4 (Distribution) of the Machine Relations Stack. GEO does not address Layer 1 (Earned Authority) or Layer 2 (Entity Clarity), which determine whether AI engines treat a source as credible in the first place. The Machine Relations Stack is the complete system; GEO is one component within it.
Why does earned media matter so much for AI citations?
AI engines are trained on and retrieve from sources they have determined to be credible. Third-party editorial coverage in established publications (Tier 1 earned media) carries the trust signals — editorial review, domain authority, consistent sourcing standards — that AI engines weight heavily. Brand-owned content does not carry equivalent trust signals regardless of its quality or volume. The empirical data is consistent: 85-88% of AI citations come from earned, third-party sources.
Which AI engines does the stack apply to?
All major AI answer engines: ChatGPT (Browse mode and GPT-4o), Perplexity (Pro and standard), Gemini (Advanced and standard), and Google AI Mode/AI Overviews. Each engine has distinct citation preferences and source weighting, which is why Layer 4 (Distribution) tracks them separately.
How do brands measure their Machine Relations Stack performance?
The primary metrics are Share of Citation (how often the brand is cited vs. competitors for category queries), Entity Resolution Rate (how accurately AI engines describe the brand), Sentiment Delta (the gap between self-description and AI description), and AI Referral Traffic. A measurement methodology is maintained at machinerelations.ai/glossary/share-of-citation.
Can a brand address all five layers without an agency?
Individual layers are addressable internally. Earned Authority requires relationships with editors at Tier 1 publications — these take years to build and are the primary barrier to entry for self-execution. Entity Clarity, Citation Architecture, and Measurement are technical and operational functions that in-house teams can own. Distribution monitoring is a tooling function. Most brands that attempt Layer 1 internally find that it is the one layer where execution quality has the largest variance.
This framework is maintained by machinerelations.ai, the category site for Machine Relations research and methodology. For the complete framework reference, see the MR Stack. For the full taxonomy of Machine Relations terms, see the MR Glossary.