The AI landscape doesn't move in one direction — it lurches. Some techniques leap from experiment to table stakes in a single quarter; others stall against regulatory walls, technical ceilings, or organisational inertia that no amount of hype can dislodge. Knowing which is which is the hard part. The State of Play cuts through the noise with a rigorously maintained index of AI techniques across every major business domain — classified by maturity, evidenced by real-world adoption, and updated daily so you always know where you stand relative to the field. Stop guessing. Start knowing.
A daily newsletter distilling the past two weeks of movement in a domain or two — delivered to your inbox while the index updates in the background.
Each dot marks the weighted maturity of practices within a domain — hover for a brief summary, click for more detail
AI that assists with drafting documents and improving writing quality through editing, style, and tone suggestions. Includes document drafting from outlines and readability improvement; distinct from content generation in Marketing which targets published content rather than personal professional writing.
AI writing assistance is a proven productivity category with mature tooling, validated ROI, and mainstream adoption — but its real-world impact remains tightly constrained and adoption paradoxes suggest the category has reached a stable equilibrium rather than further growth. The ecosystem spans dedicated vendors like Grammarly (40M daily users, $13B+ valuation, $200M+ ARR), platform-native offerings like Microsoft 365 Copilot (90%+ Fortune 500 adoption), and general-purpose LLM interfaces. Randomised controlled trials and field studies confirm bounded gains: Microsoft's RCT of 7,137 workers across 66 firms found 3.6 hours saved weekly on email alone, and practitioner analysis shows the category succeeds precisely because writing tasks have clear inputs and measurable outputs, unlike broader productivity claims. Consumer uptake is broad (63% of U.S. adults use AI tools monthly, writing assistance is the top use case) and sector-specific adoption is strong (73% college students, 89% marketing teams, 94% content agencies). Yet UC Berkeley peer-reviewed research (April 2026) reveals a critical paradox: heavy AI writing assistance use reduces argument coherence by 70% and erodes writer voice, while users report equal or higher satisfaction — suggesting a systematic misalignment between perceived writing quality and actual quality outcomes.
The structural barriers preventing tier advancement are both technical and behavioural. Hallucination rates in professional writing are escalating (18% in news-related prompts in Aug 2024 rose to 35% by Aug 2025, resulting in 12,842 articles pulled from circulation in Q1 2025), with legal domain showing 69-88% failure rates and 486+ court cases involving AI hallucinations. Enterprise adoption remains stuck in pilot limbo: only 16% of pilots reach production, and 73% of regulated-sector organisations have paused rollouts over security, data-privacy, and measurement gaps. Specialist vendors face intensifying pressure: Jasper AI collapsed with 40% layoffs as users migrated to free platforms; Grammarly's Expert Review feature (Aug 2025–Mar 2026) was disabled after 7 months following a class action lawsuit for using journalists' identities without consent; Writer.com abandoned SMB for enterprise-only positioning. Market consolidation data (22.5% CAGR, $4.8B→$20B 2025–2034) masks a permanent bifurcation: Fortune 500 firms and content professionals extract value in bounded tasks (drafting, summarizing, light editing), while mid-market and regulated organisations remain structurally blocked by accuracy, compliance, and measurement immaturity.
The vendor landscape is consolidating around two poles while specialist competitors exit or retreat. Microsoft continues expanding Copilot with Agent Mode for Word/Excel/PowerPoint and new draft-generation agents, reporting 90%+ Fortune 500 adoption. Grammarly ($700M+ revenue, 50,000 enterprise customers, 40M daily active users) is specialising vertically — shipping dedicated products for HR and customer-support teams with features like style-guide enforcement, term banks, and multilingual support. Both platforms are maturing: Grammarly reports 283% return for enterprise customers; Writer.com released 200+ Skills, observability/Datadog integration, and NVIDIA NIM support (Mar-Apr 2026). Specialist vendors face structural pressure: Jasper AI (once valued at $1.5B) collapsed with 40% layoffs in early 2026 as users migrated to free ChatGPT/Claude; Grammarly's Expert Review feature was disabled in March 2026 following a class action lawsuit for using journalists' identities without consent; Writer.com abandoned SMB markets and now targets enterprise exclusively at $18,000/year minimums. Market consolidation signals that writing assistance is moving toward general-purpose platforms and vertical specialization.
Adoption breadth masks deployment stall. Writing and communications account for 80% of all workplace generative AI use — the single largest application category. Sector-specific metrics show strength: 73% of college students use AI for writing, 89% of marketing teams use AI (up from 67% in 2024), 94% of content agencies embed it in workflows, and 74% of content marketers use AI tools. In professional services (legal, tax), 40% have deployed generative AI with 30% drafting time reductions. Yet enterprise conversion remains the critical bottleneck: only 19% of content marketing teams track AI-specific KPIs despite reporting 3.4x velocity and 67% cost reduction, indicating measurement immaturity; Copilot rollouts routinely stall at 20% adoption within organisations due to absent manager modelling, poor workflow integration, and unclear permissions. EU enterprises face additional compliance barriers: Grammarly Enterprise requires Data Processing Agreement (Standard Contractual Clauses only, no EU residency), DPA unavailable for Free/Premium tiers, and Works Council participation rights under German labor law, raising deployment friction for regulated sectors.
Accuracy and reliability remain limiting factors showing escalation. Hallucination rates in professional writing are worsening: news-related prompts doubled from 18% (Aug 2024) to 35% (Aug 2025), with 12,842 AI-generated articles pulled from circulation in Q1 2025. Enterprise testing of 6 AI content tools (March 2026) documented 15-25% hallucination rates. Legal domain shows systematic failure: 69-88% hallucination rate on specific legal queries, with 486+ court cases now involving AI hallucinations and 128 lawyers sanctioned (including Mata v. Avianca where ChatGPT fabricated 6 cases). Medical domain (Mount Sinai 2025): 64% hallucination rate on long clinical cases without mitigation. The paradox: UC Berkeley research shows heavy AI use reduces argument coherence by 70% and erodes writer voice, yet users report equal satisfaction—suggesting confidence masking actual quality loss. The market's next phase hinges less on feature expansion than on solving reliability (acceptable hallucination rates for high-stakes contexts), measurement (demonstrating ROI beyond early-adopter cohorts), and user-preference alignment (actual vs. perceived writing quality).
May 2026 platform developments reveal the "verification tax" shaping deployment reality. Google shipped tone/style personalization in Gmail (May 7, 2026), and Microsoft deployed Writing Tools natively in Notepad, confirming platform convergence on integrated drafting—yet field research shows this integration paradox: MIT's 1,258-person experiment validated 137% message volume and 20% editing time reduction, but independent studies document that 80% of recovered time is reabsorbed into review and quality control. Leaders report spending more time validating polished AI output than writing from scratch, and researchers identify that improved fluency masks errors and hallucinations rather than reducing them. In regulated domains (legal, healthcare, finance), verification burden blocks adoption entirely: 58% hallucination on legal queries and institutional shifts to sampling-based triage confirm that polished AI writing requires deeper human scrutiny, not less. The category has completed the "generation" phase and entered the "verification" phase—adoption now depends on whether organizations can afford the human overhead of validation rather than whether tools generate plausible text.
— Comprehensive hallucination benchmarks: 0.7–4.6% on basic summarization, 18.7% on legal queries, 15.6% on medical—MIT research shows models are 34% more confident when false—establishing that hallucination is fundamental ceiling on trustworthiness.
— Google released tone/style personalization for Gmail writing, inferring writer voice from email history and integrating context from Drive—confirming major platform production deployment of learning-based style adaptation.
— Research-backed analysis revealed bimodal productivity distribution: power users reclaim 9–20+ hours/week from drafting/summarization/translation, while casual users see negligible gains—confirming writing assistance ROI is context-dependent and adoption barriers are structural.
— SSHRC-funded research identifies psychological risks: deferring writing to AI shifts users from active contributors to passive reviewers, eroding confidence in own writing abilities—documenting fundamental adoption barrier rooted in skill erosion and loss of authorship agency.
— Independent research of 90+ leaders: ~80% reported recovered writing time reabsorbed into reviewing/prompting/QC; polished AI output is harder to fact-check than rough drafts—confirming that writing assistance creates 'different work' rather than productivity gains at scale.
— Carl Benedikt Frey identifies the 'verification tax': field study of experienced developers showed AI access made them 19% slower (vs. 14% gains in customer support); Sullivan & Cromwell fabricated citations case shows net productivity depends on error cost—explaining why writing assistance stalls in high-stakes domains.
— Real-world adoption (42% of tenants use AI for legal interpretation) met with institutional friction: 58% hallucination on legal queries; organizations shifted from linear reading to triage/sampling/cross-checking because polished AI output conceals fabricated authorities—evidence of high adoption and high institutional friction.
— MIT field experiment (n=1,258 teams) showed AI writing agents reduced editing time 20% while increasing message volume 137%; paid X campaigns validated quality, providing rare RCT-grade evidence of productive human-AI writing collaboration.