Methodology

How we measure Agent Citability

The Agent Citability Index (ACI) quantifies how likely AI systems are to discover, understand, trust, and cite a given website. This document describes what we measure, how scores are structured, and what the data reveals.

Published 17 April 2026 6 dimensions 500-site benchmark

What is the Agent Citability Index?

As AI systems — large language models, autonomous agents, shopping assistants, and search overviews — become primary intermediaries between consumers and businesses, traditional SEO metrics no longer capture the full picture of digital visibility.

A site can rank well in Google search results and still be effectively invisible to AI-driven discovery. The reasons are structural: AI systems evaluate content differently from keyword-matching algorithms. They assess whether content is machine-readable, structurally coherent, semantically clear, demonstrably authoritative, and fresh enough to be reliable.

The Agent Citability Index was developed to make these signals measurable. A high ACI score indicates that a site is well-positioned to appear in AI-generated responses, be cited in LLM outputs, and transact with AI-driven commerce agents. A low score reveals specific, fixable gaps in how the site presents itself to non-human systems.


The six dimensions

Each page or site is evaluated across six independent dimensions. Scores for each dimension range from 0 to 100. The composite ACI score is a weighted combination of all six, calibrated against observed citation behaviour across major AI systems.

Dimension 01
Schema
Measures the presence, correctness, and richness of structured data markup. AI systems use schema to identify what a page is about, who published it, and how its content relates to other entities — without needing to parse prose.
  • JSON-LD presence and validity
  • Type specificity (Article, Product, Organization, etc.)
  • Entity completeness and interlinking
  • Coverage across page types
Dimension 02
Structure
Evaluates how well the page's information architecture serves automated comprehension. Content that is logically ordered, properly headed, and semantically marked up is far easier for AI systems to parse, summarise, and excerpt accurately. Structural issues are consistently the most surprising low-score driver — pages with strong content but wrong markup score poorly because AI agents cannot detect structure they cannot read.
  • Real heading hierarchy — <h2> tags, not styled <p> elements or <details> accordions
  • Structured lists — <ul>/<ol> elements, not comma-separated prose
  • Semantic wrappers — <section> and <article> to define content boundaries
  • Answer-first pattern — direct answer in the opening paragraph
  • Content depth — minimum word count for citation eligibility
Dimension 03
Freshness
AI systems are trained to prefer recent, actively maintained sources over stale ones. Freshness measures both how recently content was published or updated and how clearly those dates are communicated to automated systems.
  • datePublished and dateModified in structured data
  • Machine-readable datetime attributes
  • Publication cadence signals
  • Recency of cited data and claims
Dimension 04
E-E-A-T
Experience, Expertise, Authoritativeness, and Trustworthiness. AI citation models weight sources by perceived credibility. This dimension assesses the signals that establish a site as a trustworthy, authoritative source worth referencing.
  • Author attribution and credentials
  • About and contact information
  • External references and citations
  • Transparency signals (methodology, data sourcing)
Dimension 05
Technical
Covers the foundational technical requirements that determine whether AI crawlers can reliably access and process a site. Strong technical scores are necessary but not sufficient — they enable the other dimensions to be evaluated at all.
  • Crawlability and robots.txt configuration
  • Page speed and Core Web Vitals
  • HTTPS and security headers
  • llms.txt and AI-specific accessibility
Dimension 06
Agent Correlation
The only dimension that draws on live traffic data. Measures the degree to which a site is already being visited by known AI agent systems — crawlers, LLM training bots, and commerce agents — as a proxy for existing discoverability in AI pipelines.
  • AI crawler visit frequency
  • Agent type diversity
  • Recency of agent visits
  • Coverage across page types

Score bands and what they mean

The composite ACI score runs from 0 to 100. Scores are normalised against observed performance across a continuously updated reference set. The bands below reflect meaningful thresholds in real-world citability outcomes.

Band Score range What it means
Critical 0 – 29 Significant structural barriers to AI discovery. Likely invisible to most AI citation systems. High-priority remediation needed before agentic commerce becomes viable.
Needs Work 30 – 54 Partial visibility. Some AI systems may encounter the site, but inconsistent signals reduce citation likelihood. Targeted improvements to the lowest-scoring dimensions will have the highest impact.
Good 55 – 74 Solid foundation. The site is legible to most AI systems and reasonably likely to appear in relevant AI-generated responses. Optimisation at this level focuses on competitive differentiation.
Excellent 75 – 100 High citability. The site is well-structured, trusted, fresh, and technically accessible to AI systems. Well-positioned for agentic commerce and AI-driven discovery at scale.

Shopify e-commerce: April 2026

In April 2026, COVEN AI conducted an ACI analysis of 500 Shopify-powered e-commerce sites spanning multiple industries and regions. The dataset represents a cross-section of active online retailers and provides the most comprehensive public benchmark of agent citability for e-commerce currently available.

The results reveal that the majority of online businesses are significantly underprepared for AI-driven discovery — a gap that will widen as agentic commerce adoption accelerates.

500
Shopify sites analysed
29.4
Mean ACI score (0–100)
Critical
Median score band
55%
Agent traffic share among scanned sites

The mean score of 29.4 places the average Shopify store in the Critical band. The primary drivers of low scores are weak Schema markup, absent freshness signals, and minimal E-E-A-T infrastructure — all of which are addressable with targeted effort.

Score distribution by industry

Average ACI score across verticals in the April 2026 Shopify benchmark.

Industry Avg. ACI score
Health & Wellness
38.0
Home & Garden
34.0
Fashion & Apparel
31.0
Food & Beverage
29.0
Electronics & Tech
27.0
Beauty & Personal Care
25.0

What we analyse

ACI scores are derived from live page analysis conducted by COVEN AI's scanning infrastructure at the time of assessment. Scores reflect the state of a site at scan time and will change as the site evolves. Scores are not cached indefinitely — monitoring subscribers receive updated scores on a regular cadence.

The analysis draws on signals from the publicly accessible version of each page as seen by a standard web client, structured data validators, crawl behaviour logs (for Agent Correlation), and heuristic evaluation of content quality signals. No proprietary or authenticated data is used; ACI scores reflect only what AI systems themselves can observe.

Industry benchmark data is aggregated and anonymised. Individual site scores are not disclosed in public reports.


How to improve your ACI score

Because ACI scores are decomposed into six independent dimensions, improvement is systematic rather than speculative. The highest-impact actions are almost always in the lowest-scoring dimensions — and the dimensions with the most room to improve tend to be Schema, Freshness, and E-E-A-T, as the Shopify benchmark data confirms.

Structure is consistently the most surprising high-impact area. Pages with genuinely good content often score poorly on structure because of invisible markup issues: FAQ sections built with <details> accordions register zero headings to an AI agent; section titles styled as <p> tags are indistinguishable from body copy; comma-separated item lists score nothing where a <ul> would score full points. These are low-effort fixes with outsized score impact. Read the full breakdown →

Run a free scan to see your site's current scores across all six dimensions, or commission a full Agent Readiness Audit for a prioritised action plan with copy-paste implementation guidance.

See your ACI score now

Find out how your site scores across all six dimensions — free, no account required.