The AI landscape doesn't move in one direction — it lurches. Some techniques leap from experiment to table stakes in a single quarter; others stall against regulatory walls, technical ceilings, or organisational inertia that no amount of hype can dislodge. Knowing which is which is the hard part. The State of Play cuts through the noise with a rigorously maintained index of AI techniques across every major business domain — classified by maturity, evidenced by real-world adoption, and updated daily so you always know where you stand relative to the field. Stop guessing. Start knowing.
A daily newsletter distilling the past two weeks of movement in a domain or two — delivered to your inbox while the index updates in the background.
AI for turning raw data into queryable, analysable, actionable insight. Streaming analytics, MLOps, and feature engineering are good practice with proven deployments at scale. The bulk sits at leading-edge, held back not by tooling but by data quality and governance gaps — 60% of AI projects stall on data readiness. Nearly all practices are stalled in trajectory.
Data and analytics is the domain where AI's promise is most visibly ahead of its delivery. Across sixteen practices -- from data quality automation to natural-language querying, from anomaly detection to synthetic data generation -- the tooling is mature, the vendor ecosystem is consolidated, and the cloud platforms have shipped. The bottleneck is not technology. It is the organizational capacity to absorb what the technology demands: governed metadata, clean pipelines, skilled teams, and executive commitment to data foundations over dashboards. PwC's 2026 study of 1,217 executives found that 74% of AI's measurable value accrues to just 20% of companies, and the differentiator is data infrastructure maturity, not model sophistication. Deloitte's survey of 3,235 leaders across 24 countries tells the same story from the demand side: AI tool access has expanded to 60% of organizations, but active usage sits below that threshold, and only 25% have moved pilots to production.
The domain's maturity distribution makes the point structurally. Five practices -- data cataloguing, data quality automation, feature engineering and AutoML, MLOps, and real-time streaming -- have reached the point where production deployment is proven, economically justified, and accessible to any organization willing to invest in foundations. Seven more -- exploratory data analysis, pipelines and dashboards, privacy automation, causal inference, geospatial analytics, graph analytics, and narrative generation -- sit at the leading edge, where forward-leaning teams extract real value but the broader market has not followed. One practice, anomaly detection, remains stuck at the experimental fringe despite a $6 billion market. The overall trend is stalled: fourteen of sixteen practices show no directional movement, and the two that are advancing (MLOps and real-time streaming) are doing so on the strength of platform consolidation rather than novel capability. This is not a domain in crisis -- it is a domain in plateau, where the gap between what is possible and what is deployed has become the defining feature of the landscape.
What distinguishes the leaders is unglamorous. Organizations extracting value from AI analytics -- Uber processing a trillion daily Kafka messages through AthenaX, Snowflake's 9,100+ weekly active Cortex AI accounts, Meta scaling agentic analytics company-wide in six months -- share a common trait: they invested years in data engineering before the AI layer arrived. The rest face a structural catch-up problem. Gartner's 2026 Data & Analytics Summit confirmed that 60% of enterprise AI projects are abandoned due to data readiness failures, with only 7% of enterprises reporting fully AI-ready infrastructure. The uncomfortable truth is that most organizations are not blocked by what AI can do; they are blocked by what they have not done to prepare for it.
The most significant movement this cycle was synthetic data generation's advance from experimental to leading-edge maturity. The UK Financial Conduct Authority published its Synthetic Data and Anti-Money Laundering project, deploying fully synthetic datasets with embedded money laundering typologies for regulatory testing -- a concrete signal that regulators are not just permitting synthetic data but actively building with it. The EU Data Protection Supervisor followed with authoritative governance guidance. Google's Simula framework demonstrated independent control over quality, diversity, and complexity in synthetic generation, and the CHIMERA framework showed 9,000 high-quality synthetic samples outperforming larger models on reasoning tasks. The practice remains bounded to high-governance verticals (finance, pharma, healthcare, government), but the combination of regulatory endorsement and research maturation warranted the tier shift.
Elsewhere, the domain held position. Snowflake reported 200% growth in AI workloads, and its Cortex AI account base reached 9,100+ weekly active users -- confirming natural-language querying momentum at the platform level. Databricks shipped native data profiling in SQL Editor and Notebooks, pushing automated exploratory analysis deeper into default workflows. Microsoft forced Power BI Copilot narrative mode for licensed users, while hallucination benchmarking across 40+ models documented 15.6-18.7% error rates in medical and legal narrative domains -- sharpening the case for mandatory human review. Uber published three significant production case studies: AthenaX for trillion-message streaming, D3 for automated drift detection across ML pipelines, and Bayesian neural networks for demand forecasting with principled uncertainty quantification. These are signals of deepening operational maturity at the vanguard, not broadening adoption across the market. The structural picture remains unchanged: the gap between what leading organizations deploy and what the median enterprise can absorb continues to define this domain.
The data readiness wall. Every major consulting survey in this cycle -- PwC (1,217 executives), Deloitte (3,235 leaders), Capgemini (1,500+ leaders), RAND (65 project meta-analysis) -- converges on the same finding: data quality and governance maturity, not model capability, determine whether AI analytics delivers value. Gartner reports 72% of enterprise AI projects fail, with seven in ten failures tracing to poor data quality. High-performing organizations allocate 60% of AI spend to data foundations rather than analytics tooling. The implication is that most AI analytics investment is sequenced backwards -- organizations buy the dashboard before they build the pipeline.
Platform embedding versus standalone viability. First-generation standalone AI analytics products are failing commercially. AWS Lookout for Equipment joins Azure Anomaly Detector in end-of-life despite named enterprise customers. Synthetic data startups (Datagen, Synthesis AI, AI.Reverie) shut down or were acqui-hired; survivors embed into broader platforms. The same pattern holds across the domain: Databricks, Snowflake, and Microsoft are absorbing capabilities (profiling, anomaly detection, narrative generation, text-to-SQL) that were once separate product categories. The strategic consequence is that analytics AI is becoming a platform feature, not a purchasing decision -- and organizations locked into legacy stacks face rising switching costs.
The agentic reliability gap. Agentic analytics -- AI systems that autonomously generate queries, explore data, and surface insights without human prompting -- is the next frontier across multiple practices (exploratory analysis, natural-language querying, causal inference). Meta, OpenAI, and Ramp run production agentic analytics. But peer-reviewed research documents a 55% failure rate on real-world datasets, with systems reaching unsupported conclusions. IDC confirms 96% of generative AI deployments face unexpected cost overruns, rising to 92% for agentic workflows specifically. The gap between demo and deployment remains wide enough to stall adoption beyond the vanguard.
Privacy automation in an adversarial environment. Data privacy tooling has matured -- OpenAI shipped a 1.5-billion-parameter open-weight PII redaction model achieving 96-97% F1 scores, Snowflake reached GA with automated PII/PCI/PHI classification, and Protegrity demonstrated production-scale tokenization. But the threat has commoditized faster than the defense: a synthesis of 12 peer-reviewed studies shows LLM-based deanonymization achieves 68% accuracy at $1-4 per profile, and ACL 2026 research confirms that 90%+ span-level masking still exposes 67% of personal information through contextual inference. Enterprise telemetry finds 47.9% of secrets and 36.3% of financial data leaking through AI tools. Privacy automation is running to stand still.
The benchmark-to-production collapse. Across natural-language querying, causal inference, anomaly detection, and time series forecasting, the same pattern recurs: impressive benchmark performance collapses in production conditions. Text-to-SQL achieves 86.6% accuracy on Spider 1.0 but drops to 10% on complex enterprise schemas. 62% of causal treatment-effect models underperform trivial baselines on real-world heterogeneous data. Anomaly detection shows 7-day detection lag and 40% conversion drops at production false-positive rates. Time series foundation models show competitive zero-shot performance on standard datasets but exhibit method-selection instability and market-regime blindness in operational environments. The consistent finding is that controlled evaluations systematically overstate production readiness, and organizations that deploy based on benchmark claims face painful recalibration.
FCA Synthetic Data and Anti-Money Laundering project report (case-study) — Regulators are not just permitting synthetic data but building production systems with it, making this the clearest single signal behind the practice's tier shift from experimental to leading-edge. https://www.tlt.com/insights-and-events/insight/fca-synthetic-data-and-anti-money-laundering-project-report-key-points-for-financial-services-firms
AI Data Maturity in the Midmarket 2026: Five Must-Dos Before Your First Production Agent (adoption-metric) — The Gartner finding that 72% of enterprise AI projects fail and seven in ten trace to poor data quality, not model problems, is the single quantitative anchor for the summary's central argument that organizations are investing in the wrong layer. https://mybusinessfuture.com/en/ai-data-maturity-in-the-midmarket-2026-five-must-dos-before/
AI adoption in practice: What real enterprise usage data reveals about risk and governance (adoption-metric) — Enterprise telemetry showing 47.9% of secrets and 36.3% of financial data leaking through AI tools makes the privacy automation tension concrete: tools are in production, but the sensitive data flowing through them is not under control. https://www.nudgesecurity.com/post/ai-adoption-in-practice-what-real-enterprise-usage-data-reveals-about-risk-and-governance
Subject-level Inference for Realistic Text Anonymization Evaluation (research-paper) — The ACL 2026 finding that 90%+ span-level masking still exposes 67% of personal information through contextual inference explains why privacy automation is running to stand still: the defense metric (span recall) does not measure the actual attack surface. https://arxiv.org/abs/2604.21211
OpenAI Releases Privacy Filter: A 1.5B-Parameter Open-Source PII Redaction Model (news-coverage) — An open-weight 96-97% F1 PII redaction model from OpenAI illustrates the asymmetry between maturing defenses and commoditizing attacks documented in the privacy tension; capable tooling is now freely available yet re-identification costs continue to fall faster. https://www.marktechpost.com/2026/04/28/openai-releases-privacy-filter-a-1-5b-parameter-open-source-pii-redaction-model-with-50m-active-parameters/
AI Hallucination Statistics: Research Report 2026 (industry-report) — Cross-model benchmarking documenting 15.6-18.7% error rates in medical and legal narrative domains directly supports the narrative generation tension: platform-embedded narration (Power BI Copilot forced-default) is rolling out at scale while reliability constraints in high-stakes domains remain unresolved. https://suprmind.ai/hub/insights/ai-hallucination-statistics-research-report-2026/
Do Contemporary Causal Inference Models Capture Real-World Heterogeneity? (research-paper) — The benchmark-to-production collapse is sharpest in causal inference: 62% of contemporary treatment-effect models underperform a trivial baseline on real-world heterogeneous data, the same data organizations actually have, not the clean datasets benchmarks assume. https://chatpaper.com/fr/chatpaper/paper/112416
Predictive Maintenance — Amazon Lookout for Equipment (product-ga) — AWS Lookout's October 2026 end-of-life despite named enterprise customers (Koch Ag, CEPSA) exemplifies the platform-embedding tension: first-generation standalone analytics AI products are failing commercially even after achieving production deployments. https://aws.amazon.com/lookout-for-equipment/
Grow payment conversion with AI — Adyen Uplift (product-ga) — Adyen's 10% conversion lift on trillions of payment transactions is the kind of large-scale production causal inference that the 62% benchmark-failure finding makes harder to replicate; it requires both clean transactional data infrastructure and years of A/B testing history most organizations lack. https://www.adyen.com/uplift
Deloitte State of AI in the Enterprise 2026 (industry-report) — The 3,235-leader survey showing data management maturity at 40% versus technology infrastructure at 43% and talent at 20% puts numbers on the structural catch-up problem: organizations have the platforms before they have the data discipline to use them. https://mybusinessfuture.com/en/deloitte-ai-enterprise-report-execution-gap/