The AI landscape doesn't move in one direction — it lurches. Some techniques leap from experiment to table stakes in a single quarter; others stall against regulatory walls, technical ceilings, or organisational inertia that no amount of hype can dislodge. Knowing which is which is the hard part. The State of Play cuts through the noise with a rigorously maintained index of AI techniques across every major business domain — classified by maturity, evidenced by real-world adoption, and updated daily so you always know where you stand relative to the field. Stop guessing. Start knowing.
A daily newsletter distilling the past two weeks of movement in a domain or two — delivered to your inbox while the index updates in the background.
AI for individual productivity, communication, organisation, and self-directed learning. The most polarised domain: writing assistance and meeting summarisation are good practice, but nearly half the practices are bleeding-edge — personal AI agents, life planning, and autonomous scheduling lack reliable implementations. Most trajectories are stalled, reflecting a gap between consumer hype and sustained daily utility.
Personal effectiveness is the domain where AI adoption is widest and organisational impact is weakest. The numbers are not subtle: 63% of U.S. adults use AI tools monthly, writing assistance is the single largest application category (80% of all workplace generative AI use), and platforms like Gmail, Outlook, and Microsoft 365 ship AI features to billions of users as default infrastructure. Translation has matured into proven enterprise tooling with 79% of organisations integrating it into workflows. Accessibility support runs at production scale across healthcare, broadcasting, and emergency services. Writing assistance generates $700M+ in annual revenue for Grammarly alone, and Microsoft reports 90%+ Fortune 500 adoption of Copilot. The tooling layer is not experimental -- it is shipping, monetised, and embedded in the platforms people already use.
Yet the domain's defining characteristic is the gap between individual adoption and organisational value. Across practices, the same pattern repeats: individuals extract real time savings in bounded tasks while enterprises struggle to move beyond pilots. Only 16% of AI writing pilots reach production. Copilot rollouts routinely stall at 20% adoption within organisations. A Workday study of 3,200 executives found 37% of time savings lost to rework, with only 14% achieving consistent net productivity gains. PwC's survey of 4,454 CEOs found 56% reporting no significant AI ROI. The practices that have crossed into mainstream deployment -- translation, writing assistance, accessibility -- did so by solving narrow, well-defined problems with clear inputs and measurable outputs. Everything else remains stuck in a structural bind where individual productivity gains do not aggregate into organisational transformation.
The deeper issue is that AI tools in this domain are now generating their own counter-effects. UC Berkeley research shows heavy AI writing use reduces argument coherence by 70% while users report equal satisfaction -- a systematic misalignment between perceived and actual quality. BCG/Harvard found 14% of employees report "AI brain fry" from cognitive overload, with 39% more errors when using four or more tools. ActivTrak's analysis of 163,000 employees found AI adoption surged to 80% but focus time declined 9% and multitasking rose 12%. AI email tools increase reply volume by 38% but 61% of those replies contain no new information, leading to 22% longer resolution times. The domain has entered a phase where the costs of AI assistance -- cognitive dependency, quality erosion, workload intensification -- are becoming as measurable as the benefits.
No practices changed tier or trend in this scan cycle. The domain's structure remained stable, which itself is a signal: thirteen practices scanned, all holding position, most marked "stalled." The evidence gathered during this window reinforced existing dynamics rather than disrupting them.
The most significant new evidence centres on three themes. First, deployment reality collided with efficacy claims in education: Khan Academy founder Sal Khan publicly admitted Khanmigo was "a non-event" for most students despite reaching 700K+ users across 380+ districts, with only 15% engagement. A named teacher at a flagship school stopped using it. A 1,222-student RCT showed AI assistance reduces persistence and impairs unassisted performance after just 10 minutes -- directly undermining the case for AI-powered deliberate practice. Second, vendor pricing and adoption friction intensified: Microsoft disabled free Copilot Chat in Office apps on April 15, forcing a $30/user/month paywall after only 3% voluntary conversion, affecting spreadsheet automation, presentation creation, and writing workflows simultaneously. Third, the hallucination evidence deepened across multiple practices: independent testing found 17-44% factual accuracy failures in presentation tools, Oumi's analysis of Google AI Overviews found only 39% of summaries both correct and source-supported, and OpenAI's o4-mini hallucinated at 80% on general knowledge questions. These findings confirm that reliability barriers are not converging toward resolution but in several cases worsening.
Individual productivity vs. organisational value. The single most persistent tension across the domain. Individual knowledge workers report genuine time savings -- 45 minutes daily on email triage, half a day weekly on writing tasks, 3.5x meeting scheduling capacity. But organisational deployment tells a different story: only 14% of executives achieve consistent net productivity gains (Workday, 3,200 executives), and Deloitte found that despite 60% of workers now having access to sanctioned AI tools, utilisation rates remain unchanged year-over-year. The access-to-activation gap is structural, rooted in weak governance, absent manager modelling, and unclear permissions, not in software capability.
Speed vs. quality erosion. AI tools systematically trade speed for hidden quality costs that users do not perceive. UC Berkeley's peer-reviewed research found heavy AI writing use reduces argument coherence by 70% while users remain equally satisfied. AI email replies increase volume 38% with 61% containing no new information. Brainstorming tools converge 94% of ideas toward common themes. Presentation tools fabricate statistics at 17-44% rates. In each case, the speed gain is visible and the quality loss is invisible to the person using the tool, creating a confidence-accuracy gap that compounds over time.
Cognitive dependency and workload intensification. Rather than reducing cognitive load, AI tools in personal effectiveness are increasingly documented as amplifying it. BCG/Harvard's survey of 1,488 employees found 14% experiencing "AI brain fry" with 33% more decision fatigue. ActivTrak's 163,000-employee dataset showed AI adoption correlating with 9% less focus time and 12% more multitasking. A 1,222-student RCT showed AI assistance impairs unassisted performance after just 10 minutes. MIT EEG studies document progressively weaker brain engagement over time. The tools work as intended in the moment but may be degrading the cognitive capacities they were designed to augment.
Vendor economics vs. adoption reality. Microsoft's April 15 decision to paywall Copilot Chat -- after only 3% of 450 million Microsoft 365 users converted to paid tier voluntarily -- reveals the gap between vendor ambition and user willingness to pay. Jasper AI collapsed with 40% layoffs as users migrated to free alternatives. Gamma reaches $100M ARR at $2.1B valuation but users routinely discard its output. Grammarly's Expert Review was withdrawn after a class action lawsuit. The vendor ecosystem is consolidating not because the market is maturing but because the economics of standalone AI productivity tools are proving unsustainable against free general-purpose alternatives.
The translation exception and what it reveals. Translation is the only practice advancing in trend, and its success pattern is instructive: it solves a well-defined problem (language conversion) with measurable output quality, operates in workflows where 70-80% accuracy is acceptable for most use cases, and has a natural human-in-the-loop fallback (post-editing) that enterprises already understand. United Wholesale Mortgage has processed 14,000+ loans via real-time AI translation; DeepL Voice achieves 96.4/100 quality ratings. The practice works precisely because it does not require AI to exercise judgment, maintain voice, or generate novel content -- the tasks that remain structurally blocked across the rest of the domain.
Khan Academy "Learning in the Open" (case-study) — Khanmigo's own data shows 15% student engagement despite 700K+ user access and 108M+ interactions, making this the clearest published admission that reach and impact are not the same metric. https://blog.khanacademy.org/learning-in-the-open-what-ai-is-and-isnt-changing/
"Microsoft Is Betting Scarcity Will Help It Sell Copilot" (news-coverage) — The April 15 paywall decision, triggered by a 3% voluntary conversion rate among 450 million Microsoft 365 users, quantifies exactly how far vendor adoption narratives have outrun actual user willingness to pay for AI productivity tools. https://www.reworked.co/digital-workplace/microsoft-is-betting-scarcity-will-help-it-sell-copilot/
"Why AI emails can quietly destroy trust at work" (research-paper) — A 1,100-person survey finding that managerial email trust drops from 83% to 40-52% in high-AI-use teams puts hard numbers on the speed-versus-quality-erosion tension and explains why email AI adoption stalls at the team level even when individuals benefit. http://www.sciencedaily.com/releases/2025/08/250811104226.htm
"Oumi's Study Finds 50% of AI Overviews Untrustworthy" (case-study) — Independent evaluation of Google AI Overviews at 100M+ monthly users finding only 39% of summaries both correct and source-supported is the most concrete evidence that hallucination rates are not converging toward resolution at production scale. https://oumi.ai/blog/oumis-study-finds-50-of-ai-overviews
"80% Reject AI at Work: CRE Investor Impact 2026" (adoption-metric) — The WalkMe study of 3,750 professionals finding only 9% trust AI for complex decisions versus 61% executive claims is the starkest single data point on the individual-productivity-versus-organisational-value gap that runs through the entire domain. https://www.theaiconsultingnetwork.com/blog/enterprise-ai-adoption-resistance-80-percent-workers-reject-cre-investors-2026
"Cognitive debt AI: 48% boost, then 17% crash" (opinion) — This synthesis of peer-reviewed studies documenting that AI tutoring assistance reverses 48% practice gains to a 17% exam failure when the tool is removed is the most vivid quantification of the cognitive dependency tension and directly undermines the case for AI-powered deliberate practice. https://www.outlierreport.com/en/news/ai-boosted-student-scores-48-then-crashed-them-17
"AI and the Translators Left Behind: When Good Enough Wins" (news-coverage) — The IMF reducing translation staff from 200 to 50 and 28,000+ positions eliminated industry-wide since 2010 illustrates why translation is the domain's success story and its most uncomfortable one simultaneously: it works precisely by displacing the humans it replaced. https://smarterarticles.co.uk/ai-and-the-translators-left-behind-when-good-enough-wins
"5 key takeaways from DeepL Spring Launch" (product-ga) — DeepL Voice achieving 96% linguist preference over Google, Microsoft, and Zoom across 40+ languages is the strongest positive benchmark in the domain and explains why translation is the only practice marked advancing in trend while everything else stalls. https://www.deepl.com/en/blog/key-takeaways-spring-launch
"12 Common Mistakes with AI-Generated Presentations" (adoption-metric) — Analysis of 300+ Q1 2026 AI-generated decks finding fabricated statistics in 41%, generic openings in 78%, and uniform bullet rhythm in 62% documents how speed gains in presentation creation come with hidden quality costs that users systematically fail to catch before distribution. https://2slides.com/blog/common-mistakes-ai-generated-presentations
Microsoft Copilot agentic capabilities GA in Word, Excel, PowerPoint (product-ga) — The same week Microsoft paywalled free Copilot Chat, it announced agentic GA with +11% tries per user per week and +25% satisfaction — the juxtaposition of forced monetization alongside genuine capability improvement captures the vendor economics tension in a single product cycle. https://www.microsoft.com/en-us/microsoft-365/blog/2026/04/22/copilots-agentic-capabilities-in-word-excel-and-powerpoint-are-generally-available/