Perly Consulting │ Beck Eco

The State of Play

A living index of AI adoption across industries — where established practice meets the bleeding edge
UPDATED DAILY

The AI landscape doesn't move in one direction — it lurches. Some techniques leap from experiment to table stakes in a single quarter; others stall against regulatory walls, technical ceilings, or organisational inertia that no amount of hype can dislodge. Knowing which is which is the hard part. The State of Play cuts through the noise with a rigorously maintained index of AI techniques across every major business domain — classified by maturity, evidenced by real-world adoption, and updated daily so you always know where you stand relative to the field. Stop guessing. Start knowing.

The Daily Dispatch

A daily newsletter distilling the past two weeks of movement in a domain or two — delivered to your inbox while the index updates in the background.

AI Maturity by Domain

Each dot marks the weighted maturity of practices within a domain — hover for a brief summary, click for more detail

DOMAIN
BLEEDING EDGEESTABLISHED

Specialist content — events, technical & product documentation

LEADING EDGE

TRAJECTORY

Stalled

AI that generates event materials, sales collateral, technical documentation, and product content for specific professional contexts. Includes white paper drafting and event programme generation; distinct from general long-form or short-form content which targets broader audiences.

OVERVIEW

AI-generated specialist content — white papers, technical documentation, event materials, and sales collateral — has moved into mainstream production adoption, but the evidence reveals a persistent bifurcation: discrete, bounded deployments (technical docs at PostHog and Booking.com with 60–80% quality on first-draft; event activation AI generating 400+ branded posts per event; API documentation tools achieving 75% error reduction through RAG grounding) deliver measurable efficiency. Broader deployments fail: Snowflake, Amazon, and Cloudflare eliminated thousands of documentation and content roles while adoption metrics increased, indicating scale without productivity. The fundamental constraint is not model capability but infrastructure maturity. Documentation professionals have shifted from writing to editing/validation (56% YoY increase), and compliance failures are mounting—CFPB fines, product recalls due to AI-generated instructions, FDA enforcement escalation (200+ letters in 2025)—all stemming from organisations deploying AI-generated content without verification systems. Hallucination rates remain at 3–10% on best models, worsening on complex reasoning (33–51%); the April 2026 ICLR crisis (50+ peer-reviewed papers with fabricated citations) and pharmaceutical deployment risks (44% of organisations reporting negative consequences) demonstrate that fluent outputs trigger expertise paradoxes where domain experts skip verification. The practice has hardened into disciplined boundaries: AI as drafting assistant with mandatory human review where ROI is demonstrable (logistics, routine generation, compliance filtering) versus rejection of automation-grade solutions for critical-path content. Event marketing leads adoption; technical and regulated documentation proceed with institutional scepticism.

CURRENT LANDSCAPE

Adoption has reached mainstream scale. The State of Docs 2026 survey reports 76% of documentation professionals now use AI regularly (16-point YoY increase), with 56% reallocating time from drafting to editing and validation. Ahrefs data: 74.2% of new web pages contain AI-generated content; Siege Media survey: 97% of content marketers plan AI use in 2026. The AI document generator market grew from $1.5B (2023) to $5.6B (2025, 273% increase), and deployment cases are concrete: Moderna deployed AI agents for Medical/Legal/Regulatory content with 38% of MLR projected AI-driven by 2028; Wonderchat platform supporting technical documentation at ESAB, Jortt (92% autonomous resolution), and Keytrade Bank; Captured Celebrations running AI photo booth content generation across 500+ corporate events (Adidas, Sony Music, Four Seasons). However, scale without quality persists. Only 10% of organisations are "fully prepared" for AI-driven content (OpenText), and 72% struggle with daily integration; 80% of enterprise GenAI projects report no measurable impact on P&L (Deloitte/MIT analysis). Hallucination consequences are documented: healthcare organisations (44% reporting negative consequences, averaging $4.4M loss per incident); regulatory domain (FDA 200+ enforcement letters in 2025; Deloitte refunded AU$440,000 for a government report with fabricated references); academic publishing (50+ ICLR papers with AI-hallucinated citations passing peer review). The compliance gap widened with EU AI Act enforcement imminent (87 days as of May 2026); technical documentation demands absolute precision, yet 18% of AI-generated discharge summaries contained incomplete/misleading information (JAMA). Paradoxically, tech writer roles declined at scale (Snowflake 47–70 roles, Amazon 16,000+, Cloudflare 1,100) despite 80% of buyers reviewing documentation pre-purchase, indicating cost-cutting disconnected from customer impact. Practitioner consensus is firm: AI as editing/verification tool under expert oversight (PostHog, Booking.com, API platforms with RAG achieve 60–90% quality on bounded tasks) versus rejection of automation-grade solutions. Governance infrastructure is the bottleneck: eight production patterns (RAG, schema validation, hard guardrails, citation verification, human-in-the-loop) combined reduce hallucination but require intensive engineering. Event marketing continues leading adoption, with AI for personalisation, chatbot automation, and submission scoring; technical and regulated documentation proceed with hardened institutional scepticism.

TIER HISTORY

ResearchJan-2023 → Jul-2024
Bleeding EdgeJul-2024 → May-2026
Leading EdgeMay-2026 → present

EVIDENCE (77)

— Healthcare organizations scaling AI for medical writing; 44% experienced negative consequences from GenAI, averaging $4.4M loss per incident; JAMA: 18% of AI-generated discharge summaries contained incomplete/misleading information—documents specialist domain deployment risks.

— 1,100+ docs professionals surveyed: 76% use AI regularly in workflows (16-point YoY increase), 56% shift from drafting to editing; 70% factor AI into information architecture—mainstream production adoption confirmed.

— AI photo booth content generation deployed at 500+ corporate events (Adidas, Four Seasons, Sony Music); 400+ branded posts per event, 15K–25K impressions within 72 hours, 15–20 min engagement—quantified event content outcomes at scale.

— Snowflake (47–70 roles eliminated), Amazon (16,000+), Cloudflare (1,100) demonstrate AI-driven tech writer reduction; 80% of buyers review docs pre-purchase yet companies cut documentation staff—documents adoption paradox and structural maturity gaps.

— Ahrefs analysis of 900,000 new web pages: 74.2% contain AI-generated content; Siege Media survey of 1,000+ content marketers: 97% plan AI use in 2026; market grew from $1.5B (2023) to $5.6B (2025)—quantifies scale of deployment.

— High-volume content workflows identified as second major GenAI success category; Quilter estimates 13,000+ hours/month saved via M365 Copilot for post-call documentation. Limitation: 80% report no measurable impact on enterprise EBIT—adoption scale without ROI.

— Eight production patterns for clinical AI reliability (RAG, schema validation, hard guardrails, citation verification, human-in-the-loop); no single pattern sufficient; combination reduces hallucination to operationally acceptable levels—transferable to technical documentation.

— ESAB (20,000+ product catalog), Jortt (92% autonomous resolution), Keytrade Bank deployments demonstrate technical documentation specialist content production; source-attribution and black-box elimination required for enterprise deployment.

HISTORY

  • 2023-H2: ChatGPT white paper tests showed B-minus quality requiring major revision. Hallucination research documented systematic accuracy failures across healthcare and legal domains. No evidence of production adoption at scale; practice remains experimental.

  • 2024-Q1: Foundational research (Northwestern, NUS, Stanford) proved hallucinations are inevitable and quantified failure rates at 30-88% across document, legal, and QA tasks. Early commercial products emerged (Storydoc at 2,500+ customers) but MIT data showed 95% of enterprise GenAI pilots delivered zero ROI. Adoption severely constrained by reliability barriers.

  • 2024-Q2: Hallucination detection advances (Oxford Nature study on semantic entropy) offered mitigation hope, but deployment challenges persisted: JMIR medical study confirmed 28.6% hallucination rates for GPT-4 on systematic reviews, recommending against primary use. Survey data revealed 61% of enterprises experienced accuracy issues with in-house solutions (only 17% rated excellent). Early commercial traction in sales collateral (Highspot showing 2.3x view lift), but practice remains research-stage with high human oversight requirements.

  • 2024-Q3: Broad workplace adoption accelerated—39% of U.S. adults and 24% of workers using GenAI weekly (Federal Reserve survey). Yet specialist content barriers sharpened: JMIR study found ChatGPT 3.5 and Bing at critical hallucination levels for medical documentation; Northwestern research reaffirmed hallucinations are fundamental LLM features, not fixable by architecture; practitioners at law firms adopted tools cautiously, noting they remained "not necessarily that great at legal research." Microsoft's Correction tool faced expert skepticism. Technical writers pragmatically integrated AI for time-saving assistant tasks but rejected it as primary content generator. Practice shows simultaneous adoption momentum and intractable reliability barriers.

  • 2024-Q4: Specialist content adoption shifted into pragmatic integration phase. GitHub Copilot Metrics API reached general availability (October 2024), enabling production tracking of AI-assisted documentation. Real-world deployments confirmed: Rehab for JAPAN measured GitHub Copilot acceptance rates in technical documentation at 30% overall (40% for Ruby, 14% for Java), showing language-dependent effectiveness. Professional Copilot adoption jumped to 89% weekly use by December 2024, with 50,000+ organizations adopted globally. Event marketing emerged as adoption leader: 57% of event marketers expect AI to fundamentally reshape planning and execution, with live case studies of AI-generated event content (Coca-Cola Spiced Shop). Hallucination mitigation matured: AWS shipped production-ready RAG + human-in-the-loop tutorial patterns in November 2024. Critical counterpoint: academic analysis emphasized AI-generated writing remains generic and unsuitable for creative, discovery-oriented specialist content. Practice consolidates around collaboration model: AI as time-saving assistant and personalization engine, human as final validator.

  • 2025-Q1: Specialized reliability research deepened. FailSafeQA benchmark found LLMs hallucinate in 41% of finance-related documentation queries, with open-source models at 22% error rate vs. 5% for commercial—stratifying tool reliability by capability and input sensitivity. Harvard Data Science Review published theoretical framework on AI errors, situating hallucinations as inevitable consequences of design and training data structure, not fixable by model-only approaches. Ecosystem maturity continued: Microsoft's Copilot Usage Dashboard shipped in February 2025, enabling enterprise teams to track acceptance rates and ROI. Security analysis documented code hallucination rates of 48% in AI-generated code with emerging 'slopsquatting' vulnerabilities. Practice remains in pragmatic collaboration phase with increased instrumentation and validation rigor.

  • 2025-Q3: Practitioner adoption accelerated despite persistent reliability barriers. Peer-reviewed study of 83 technical writers confirmed AI time-saving benefits for routine tasks but documented systemic accuracy limitations and ethical concerns—revealing adoption bounded by domain-specific verification requirements. Hallucination benchmarks continued to worsen: industry research found 17-45% hallucination rates across general-purpose LLMs with case studies of critical failures (e.g., financial services AI fabricating brand themes). Conference practitioners and technical writers reinforced consensus: AI works best as time-saving assistant with strict human oversight, not as primary content generator. Survey of 400 B2B marketing executives found adoption lagging—only 11% optimizing content for AI discovery, indicating unreadiness among specialist content creators. Event marketing remained adoption leader with AI applications in personalization and measurement. Practice consolidates around collaboration model with increased validation rigor and measured skepticism of vendor claims about accuracy improvement.

  • 2025-Q4: Adoption reaches maturity with hardened expectations. Wharton survey (November 2025) of 800 senior leaders confirms 46% daily GenAI use (up 17 points), 75% reporting positive ROI, but paradox sharpens: advanced models (o3, o4-mini) hallucinate at 33-79% on benchmarks. Legal domain hallucinations persist at 17-33% despite RAG with 508 cases documented globally. MIT analysis: 95% of AI pilots fail to deliver value. Technical writing community crystallizes durable framework distinguishing "writing with AI" (time-saving assistant) from "writing for AI" (content for AI consumption); emphasizes structured CCMS and human-required validation. Practice transitions from experimental to mainstream augmentation constrained by acceptance of technical barriers—AI integrated where scope is narrow and oversight is systematic, rejected where accuracy is critical-path.

  • 2026-Jan: Healthcare and technical documentation domains show pragmatic adoption amid persistent hallucination concerns. HTA scoping review of hospital documentation systems documents variable accuracy in AI scribes and required human oversight for AI-generated clinical documentation. API documentation platforms (Apidog, Treblle) achieve production deployments with measured gains (75% error reduction, 90% accuracy improvements), but data analysis reveals hallucination improvement only in grounded tasks (0.7-1.5%) with surge in complex reasoning (33-51% for o3). Event content management shifts from creation toward AI-assisted logistics and compliance workflows. GitHub Copilot usage dashboard advances as ecosystem instrumentation. Specialists affirm that hallucinations remain fundamental LLM limitation unsolvable by model scaling, reinforcing adoption boundary that separates "writing with AI" (efficiency) from automation-grade solutions.

  • 2026-Feb: Ecosystem instrumentation accelerates as GitHub expands Copilot metrics to org level and CLI telemetry, enabling enterprises to track adoption patterns. Yet integration barriers persist: Gartner data shows 72% of organizations struggle with daily tool integration, only 6% achieve enterprise-wide rollout; OpenText analysis finds only 10% of organizations fully prepared for AI with 40% of agentic projects likely to be canceled by 2027. Healthcare adoption documented by American Hospital Association emphasizing clinical oversight and hallucination risks in ambient documentation tools. Critical counterevidence: Deloitte refunded AU$440,000 for government report with AI-fabricated references (published with corrections Feb 3), demonstrating real-world specialist content failures. Enterprise adoption data: 15M paid M365 Copilot seats (160% YoY), 4.7M GitHub Copilot subscribers (75% YoY), but quality integration remains bottleneck not scale.

  • 2026-Q2: Manufacturing sector shows pragmatic adoption of AI for specialist technical content. STADLER (230-year-old European manufacturer) deployed ChatGPT Enterprise across 650 employees with 125+ custom GPTs for technical documentation and engineering specifications, achieving 30-40% time savings. ENEOS Materials (Japanese producer) created 1,000+ custom GPTs for plant design and multilingual documentation translation with 80% workflow improvements, demonstrating sustainable patterns in conservative sectors. Event industry benchmark (Tree-Fan Events) documents 71% of workflows AI-capable but only 22% deployed (49-point implementation gap), with 39-43% creating content using AI; identifies critical infrastructure requirements (review workflows, guardrails, role-based permissions) separating experiments from operational systems. Expertise paradox emerges: NRC media editor embedded 15 fabricated quotes despite years of warning about hallucinations, revealing that polished outputs and fluency trust lead experts to skip verification—a structural vulnerability in specialist content workflows. World Bank evaluation synthesis case study documents complete failure (fabricated all evidence) but documents remediation pathway: corrected methodology (summarize components, validate each, synthesize, mandate citations, allow unknown, require line-by-line human review) achieved 100% faithfulness in 2024-2025 follow-up. Regulatory frameworks mature: FDA, EMA, MHRA, ISPE publish guidance on AI-assisted laboratory documentation (protocol drafting, regulatory submissions) with risk-based validation and human-centric design. Named tech firms (PostHog, Airbyte, dbt Labs, Booking.com) confirmed AI agents generating first-draft documentation PRs with 60% completion on first try via context engineering and agentic QA loops. Hallucination risks escalate in high-stakes contexts: ICLR 2026 exposed 50+ peer-reviewed papers with AI-hallucinated citations and datasets passing peer review; JMIR study documented five clinically relevant error categories in AI-transformed psychiatric notes despite stylistic improvement. MIT data from 300 deployments shows 95% of GenAI projects deliver zero P&L impact, attributing failure to organisational readiness gaps rather than model capability. Practice consolidates around infrastructure-intensive reliability models—domain expertise remains non-delegable, but structured workflows increasingly enable safe deployment in bounded contexts.

  • 2026-May: Mainstream production adoption confirmed at scale but governance failures accumulate. The State of Docs 2026 survey (1,100+ professionals) reports 76% now use AI regularly — a 16-point YoY increase — with 56% shifting from drafting to editing and 70% factoring AI into information architecture decisions. Ahrefs data shows 74.2% of new web pages contain AI-generated content; the AI document generation market reached $5.6B in 2025. Event content deployment is documented at corporate scale: AI photo booths across 500+ events (Adidas, Four Seasons, Sony Music) generate 400+ branded posts per event with 15K-25K impressions within 72 hours. However, the compliance and governance crisis deepens in parallel: CFPB fines, FDA enforcement (200+ letters in 2025), and JAMA research showing 18% of AI-generated discharge summaries contained incomplete or misleading information document the regulated-domain risks. Tech writer role elimination accelerates (Snowflake 47-70, Amazon 16,000+, Cloudflare 1,100 roles) while 80% of buyers review documentation pre-purchase — an adoption paradox that points to cost-cutting disconnected from quality impact. Moderna's AI-powered pre-review agents in Veeva PromoMats signal pharmaceutical sector adoption but underscore the dual-sided risk: 38% of MLR projected AI-driven by 2028, yet hallucination-driven "semantic drift" creates unique enforcement exposure. Enterprise-wide production patterns remain elusive: Quilter estimates 13,000+ hours/month saved via Copilot but 80% of projects report no measurable P&L impact.

TOOLS