How do you measure AI readiness?
+
We score each brand across five proprietary dimensions, each worth up to 20 points for a maximum of 100:
A1 · STR
Structural Readiness
Can machines find and parse the brand? Schema markup, retail distribution, product feed density, and data infrastructure.
A2 · SEM
Semantic Clarity
Do machines understand the brand as intended? Positioning clarity, differentiation specificity, and identity consistency across training data.
A3 · SYN
Synthetic Customer Test
Who wins when an AI agent shops? Live query testing across ChatGPT, Perplexity, Gemini, and Claude on real purchase occasions.
A4 · EMO
Emotional Residue
What did brand building leave in the data? Sentiment, cultural footprint, and trust signals in AI training material.
A5 · VOI
Voice & Agentic Readiness
What happens in a zero-visual world? Phonetic clarity, voice resolution, and personality consistency across AI outputs.
How does ARA work technically?
+
Each dimension uses a distinct technical methodology:
A1 — Structural Readiness is assessed through automated auditing of schema.org markup completeness, product feed density and accuracy across retail platforms, review corpus volume and recency, and the technical infrastructure that makes a brand machine-parseable.
A2 — Semantic Clarity is measured by probing AI models directly with identity and positioning questions — "What is [Brand]?", "What does [Brand] stand for?", "Who is [Brand] for?" — and comparing the outputs against the brand's stated positioning. Specificity, accuracy, and differentiation are all scored.
A3 — Synthetic Customer Test runs a defined set of purchase-occasion queries through ChatGPT, Perplexity, Gemini, and Claude under a standardised protocol: fresh session, no system prompt, default model settings, verbatim response recording. Brand appearances are coded against a consistent rubric and mapped to a presence heatmap.
A4 — Emotional Residue analyses the sentiment distribution of brand-adjacent content across editorial, social, and review corpora likely to have influenced AI training. It also probes how models characterise a brand's emotional identity — the vocabulary, associations, and affect they reproduce unprompted.
A5 — Voice & Agentic Readiness evaluates phonetic clarity and name resolution in voice contexts, consistency of personality descriptors across AI outputs, and how reliably a brand's tone is reproduced when AI systems are asked to write in its voice.
What do the tier labels mean?
+
Tiers describe a brand's overall ARA standing within a scored category. They are relative to the study cohort, not absolute thresholds. The four tiers are:
Awesome — category-leading machine presence. AI consistently finds, understands, and recommends this brand.
Strong — above-average AI readiness with identifiable gaps. Competitive but not dominant.
Average — AI knows the brand but doesn't reliably recommend it. Structural or semantic work needed.
Weak — AI either misrepresents the brand or passes it over. Significant intervention required.
Is the score comparable across categories?
+
Category scores are benchmarked within their cohort, not across the full index. A score of 72 in luxury fragrance means something different from 72 in fast food — the competitive dynamics, AI query patterns, and structural data norms differ by category.
Across-category comparisons should be treated as directional rather than exact. Where we have run multiple waves in the same category, wave-on-wave comparisons are fully valid.
How do you ensure the A3 queries are unbiased?
+
Each A3 query set is designed before testing begins and held constant across all platforms and brands in the study. Queries represent real purchase occasions derived from category-specific consumer research — not prompts engineered to produce particular outputs.
We run each query through a standardised session protocol (fresh context, no system prompt, default model settings) and record responses verbatim. Brand mentions are coded by a consistent rubric: present, absent, or not applicable based on query relevance. Platform totals and query totals are reported transparently in every deliverable.
Do AI models change their answers over time?
+
Yes. AI models are updated continuously — some through fine-tuning, some through retrieval augmentation, and some through full retraining. This means that A3 results for a given brand can shift between waves without any action on the brand's part, simply because the model has ingested more (or different) training data.
This is precisely why re-auditing matters. A brand that earns strong A3 performance in Q2 cannot assume that result holds in Q4 without re-testing. Sustained machine presence requires ongoing investment in the inputs that inform AI training.
How is Emotional Residue (A4) measured?
+
A4 is one of the more nuanced dimensions. It evaluates what brand-building activity has left behind in the data AI systems trained on — not just whether a brand is known, but whether it is known
well.
We assess the sentiment distribution of brand-adjacent content across editorial, social, and review corpora. We also probe how AI models characterise the brand when asked open-ended questions: the vocabulary, associations, and affect they reproduce. High A4 scores reflect brands that have generated rich, positive, and specific cultural material — not just reach.
Authentic brand love leaves a measurable linguistic signature. Real advocates write differently about a brand than manufactured marketing does. AI can tell the difference.
Does the ARA score factor in paid advertising?
+
Not directly. Paid media does not influence how AI language models represent a brand in conversational contexts. AI systems are not advertising platforms — they do not sell placements, and paid spend has no bearing on whether or how an AI recommends a brand.
That said, paid media can indirectly affect AI readiness over time — it drives traffic, review volume, and earned media coverage that does enter training data. We account for the structural and semantic effects of this activity, not the spend itself.