The AI landscape doesn't move in one direction — it lurches. Some techniques leap from experiment to table stakes in a single quarter; others stall against regulatory walls, technical ceilings, or organisational inertia that no amount of hype can dislodge. Knowing which is which is the hard part. The State of Play cuts through the noise with a rigorously maintained index of AI techniques across every major business domain — classified by maturity, evidenced by real-world adoption, and updated daily so you always know where you stand relative to the field. Stop guessing. Start knowing.
A daily newsletter distilling the past two weeks of movement in a domain or two — delivered to your inbox while the index updates in the background.
Each dot marks the weighted maturity of practices within a domain — hover for a brief summary, click for more detail
AI that generates financial forecasts and enables rapid scenario modelling across revenue, cost, and cashflow projections. Includes driver-based forecasting and automated scenario comparison; distinct from sales forecasting which predicts pipeline-level revenue rather than company-level financials.
AI-driven financial forecasting and scenario modelling is a proven capability held back by an execution gap. The tooling is mature: Workday Adaptive Planning and Anaplan have established large production footprints, analyst firms recognise them as category leaders, and documented deployments deliver 30-50% cycle-time reductions with rapid scenario reforecasting. Broadridge data shows 80% of financial services firms now use generative or predictive AI in some capacity. The practice has earned its place as good-practice -- the question is no longer whether it works, but how to operationalise it reliably. That rollout question, however, is not trivial. Only 27% of deployments produce measurable business benefits, and most finance teams still treat algorithmic forecasts as decision-support rather than primary authority. Explainability demands, data governance debt, model validation complexity, and recurring implementation failures keep adoption concentrated among well-resourced organisations. The defining tension is not technological but organisational: platforms that can deliver value at scale, constrained by institutions that cannot yet absorb it—and recurringly struggling with operational execution despite high investment intent.
Q1 2026 vendor activity demonstrates continued platform maturity with expanding ecosystem: Workday Adaptive Planning 2026 R1 (March) expanded Predictive Forecaster to 10M cells, added ML-driven anomaly detection, launched Planning Hubs; Anaplan launched CoModeler, Custom Analyst agents with LLM-deterministic planning engine architecture (March, confirmed with named customer testimonials from Sky and Virgin Media O2); Board released FP&A Agent with econometric forecasting claiming 50% accuracy improvement; Oracle released Advanced Predictions ML feature in core EPM platform. All major platforms—Workday, Anaplan, Board, Oracle—hold sustained Leader/strong analyst recognition. Workday Adaptive Planning serves 7,000+ customers globally; Anaplan’s Intelligence portfolio supports production deployments across financial services and manufacturing. JPMorgan case demonstrates Fortune 500 production scale: 450+ GenAI use cases in production including treasury stress scenarios and scenario analysis with firmwide CDO governance. Finastra survey (1,509 financial executives) documents 65% of US institutions in active AI deployment with 42% planning >50% investment increase in 2026.
Yet the value-delivery gap persists and is widening. Only 27% of financial services firms report measurable business benefits from AI investments (Broadridge); peer-reviewed Duke/Federal Reserve research confirms CFOs report 1.8% productivity gains while actual revenue outcomes lag far behind; MIT analysis puts enterprise AI pilot success at 5%; Fullstack Labs analysis of 140 GenAI implementations shows 73% fail to deliver ROI with organizational causes (77%) outweighing technical factors (23%). Production-specific implementation failures are common—LLM hallucinations remain fundamental, not engineering defects: peer-reviewed May 2026 research proves hallucinations are mathematically inevitable (OpenAI/Georgia Tech papers on epistemic uncertainty and computational intractability); independent benchmarks document 4.2%-19.1% hallucination rates across frontier models with citation accuracy worst-performing at 12.4% average. Workday Adaptive Planning Forrester TEI (May 2026) documents real deployment value—242% ROI, $6.3M 3-year benefits, 35% FP&A productivity gains—but commissioned studies mask adoption reality: NYSE/Oliver Wyman survey of 500 CFOs (12% of global market cap, April 2026) reveals only 8% have deployed AI agents at scale, 74% still in planning/piloting, 68% expect increased analytics involvement but lack execution pathways. Regulatory maturity accelerates: NIST AI 600-1 framework (July 2024) now formally treats confabulation as tier-1 financial services risk, establishing pre-deployment testing mandates that layer governance overhead onto already-complex deployments. Practitioner surveys reveal structural barriers: only 14% of mid-market CFOs trust AI for forecasting accuracy without human oversight; 64% find scenario planning extremely challenging due to data consolidation complexity; 78% of finance teams cannot run scenarios within a day; only 3% of organizations achieve real-time scenario capability. Bain survey (April 2026) shows 56% of CFOs planning >15% AI investment increases despite only 31% reporting strongly positive results—investment intent rising while execution success stalls. Gartner projects over 40% of agentic AI projects will be cancelled by 2027; two-thirds of finance AI buyers report post-purchase regret with major failure modes including "Me Too Trap" (copying other organizations’ use cases rather than solving own problems), talent gaps, and cost volatility. The result is market segmentation: well-resourced institutions with enterprise-grade data governance and deterministic governance patterns operationalise AI-driven scenario modelling for competitive advantage (Federal Reserve CFO Survey 2001–2026 validates firm-level forecasting accuracy as leading indicator, underpinning scenario modelling ROI), while mainstream finance organizations stall on hallucination risk, regulatory burden, data quality debt, model reliability risk (drift, miscalculation), and recurring execution friction despite accelerating investment intent.
— Federal Reserve research on CFO forecasting behavior and price expectations using quarterly CFO Survey data (2001–2026); validates CFO firm-level forecasting as accurate inflation signal.
— Forrester TEI study (commissioned by Workday) showing 242% ROI, payback <6 months, $6.3M 3-year benefits, 35% FP&A productivity gains. Tier-1 evidence type with specific financial impact metrics for Adaptive Planning deployment.
— Independent vendor review of Anaplan, detailing scenario modeling, FP&A capabilities, AI forecasting agents, and deployment complexity at enterprise scale.
— NYSE + Oliver Wyman Forum survey of 500 CFOs (12% of global market cap) showing broad priority shift toward AI-driven financial planning, scenario analysis, and continuous planning with 68% expecting increased analytics involvement.
— Technical analysis citing Sept 2025 OpenAI/Georgia Tech paper proving LLM hallucinations are mathematically inevitable (not engineering defect). Proposes 'LLM Sandwich' architecture (deterministic layers wrapping LLM) as production pattern. Finding: only 14% of CFOs completely trust AI for accurate accounting.
— Peer-reviewed research (ACL 2026 Industry Track) on mitigating financial AI hallucinations. FinGround three-stage pipeline reduces hallucination by 68% vs. baseline, 78% with full pipeline. Addresses EU AI Act enforcement deadline (Aug 2026). Existing detectors miss 43% of computational errors.
— Authoritative documentation of NIST AI 600-1 regulatory framework treating confabulation as a tier-1 risk for financial services GenAI. Establishes formal governance expectations for confabulation testing pre-deployment. Critical institutional signal of regulatory maturity.
— Independent benchmark study (5,000 prompts across 5 frontier models) documenting persistent hallucination rates (4.2%-19.1%) across task families. High methodological rigor with automated+human grading. Critical evidence of technical limitations affecting forecast reliability.