Auditor AI Team
After every release, sixteen specialist sub-agents audit the live site in parallel. Each scores its own dimension one to ten and ships pasteable patches with priorities P0, P1, or P2. Every dispatch prompt is public. Every audit report is public. The pipeline that produces the data is reviewed by a separate pipeline that critiques the rendering.
Why a 16-specialist team?
A single reviewer always misses dimensions outside their lane. A research site serves vendors, donors, investors, government officials, journalists, contributors, and AI search crawlers — each judges different things. The specialists below are organised so each pair of dimensions has at least one dedicated auditor.
What did the audit catch in v1.0.0?
- 17 HIGH severity issues + 16 MEDIUM + 2 LOW.
- Four wrong Tier-1 identities (IT Park UZ CEO, KG UDP head, KG Min Health, KG last Минцифры minister) — caught by paid Sonar Pro re-verification, corrected before publication.
- WCAG contrast failure (#00aa44 at 2.89:1) before audit; replaced with #005c27 at 4.5:1 after.
- OG image referenced .png but only .svg existed — every social share card was broken until this audit caught it.
- 169 Cyrillic cells without lang= attribution; auto-wrapped via the new ru() helper.
- Nav buried H1 below the fold at 320 px — fixed with a CSS-only hamburger pattern and tap targets ≥ 44 px.
The team
Each card links to the dispatch prompt (so you can re-run any specialist yourself) and the actual audit report it produced.
01 · Reference Benchmarker
Score the audit site vs the padel-market-analysis reference dimension-by-dimension; identify copyable patterns and structural advantages.
02 · Information Architect
Audit nav, hierarchy, breadcrumbs, internal-link graph, and scent of information across all rendered pages.
03 · Content Voice Editor
Plain English, anti-jargon, third-person professional, Flesch-Kincaid grade ≤ 10. No marketing badges, no hedging fluff.
04 · Citation / Provenance
Every numeric claim and named entity traceable to a source. Russian / Uzbek / Kyrgyz share ≥ 30%. Dead-link health.
05 · GEO / AIO / AEO / LLMO
Maximise discoverability for ChatGPT Search, Claude, Perplexity, Gemini, AI Overviews. llms.txt, JSON-LD @graph, FAQPage, HowTo, Dataset schemas.
06 · Visual / Typography
Beautiful, calm, scannable typography. Fluid type scale, 60–80ch line length, WCAG AA contrast, whitespace rhythm.
07 · Mobile-First QA
Flawless on a 320 px iPhone SE. Tap targets ≥ 44 px (WCAG 2.5.5), no horizontal scroll, hamburger nav, card layout for narrow viewports.
08 · Accessibility (WCAG 2.2 AA)
Skip-link, focus-visible, scope=col, lang attribution, semantic HTML5. Screen-reader-friendly + keyboard-only parity.
09 · Performance Engineer
Core Web Vitals green: LCP < 1.5 s, FCP < 0.4 s, CLS < 0.05, INP < 100 ms. Lighthouse Performance ≥ 97.
10 · HTML Code Quality
Semantic HTML5, W3C-validating, proper landmark structure. thead / tbody / scope=col, abbr, time, cite, figure.
11 · CSS Architect
Custom properties, fluid clamp() type, logical properties, prefers-color-scheme dark mode, prefers-reduced-motion, complete print stylesheet.
12 · Data Visualization
Charts that justify their existence. Sortable tables ≥ 10 rows. Inline SVG. No chartjunk.
13 · Trust & Brand
Establish trust at first scroll. Distinguish from generic SEO content farms and Big-4 boilerplate. Author surface, license badge, methodology openness, ethics statement.
14 · Conversion / CTA
Every page has ONE primary action (≤ 2 secondary). Persona-specific routing. Cite-this-research widget. Privacy-respecting share row.
15 · Internationalization
Cyrillic content renders correctly, is searchable, and screen-reader-friendly with proper lang= attribution. Hreflang for any /ru/ mirror pages.
16 · Dev-Ex / Reproducibility
Anyone can clone the repo and reproduce or extend. README clarity, mermaid architecture, CI badges, CONTRIBUTING, issue templates, citation file.
How to re-run the full audit
The audit dispatches 16 Claude sub-agents in parallel using subagent_type=general-purpose and run_in_background=true. Total wall-clock per pass: about 30 minutes. Anthropic spend on Sonnet: about USD 4–8. Zero paid OpenRouter calls — the audit team reads only the public web and local files.
# 1. Fan out 16 specialists in parallel from a single Claude Code session:
for spec in .claude/audit-team/*.md; do
echo "Dispatch: $spec"
done
# 2. Each writes to state/audit/team/<NN>_*.md
# 3. Synthesise, apply patches, re-render, re-deploy
Where do the prompts live?
Two mirrored locations — Claude Code agent specs at .claude/audit-team/ and a public flat-file index at prompts/audit-team/. The actual reports live at state/audit/team/.