The AI landscape doesn't move in one direction — it lurches. Some techniques leap from experiment to table stakes in a single quarter; others stall against regulatory walls, technical ceilings, or organisational inertia that no amount of hype can dislodge. Knowing which is which is the hard part. The State of Play cuts through the noise with a rigorously maintained index of AI techniques across every major business domain — classified by maturity, evidenced by real-world adoption, and updated daily so you always know where you stand relative to the field. Stop guessing. Start knowing.
A daily newsletter distilling the past two weeks of movement in a domain or two — delivered to your inbox while the index updates in the background.
Each dot marks the weighted maturity of practices within a domain — hover for a brief summary, click for more detail
AI that generates and maintains operational runbooks and produces post-incident review reports. Includes automated playbook creation and blameless post-mortem drafting; distinct from incident response automation which executes actions rather than documenting them.
AI-generated runbooks and post-incident reports have crossed from experimental to vendor-ready, but organisational adoption remains selective and constrained by maturity gaps. The technology addresses a genuine pain point: manual runbooks decay as systems change, and post-mortems are routinely delayed or incomplete because documentation competes with recovery for engineers' time. Forward-leaning organisations are getting measurable value -- incident.io reports 37% MTTR reduction and $29,700 annual savings, SolarWinds data across 2,000+ ITSM systems shows 17.8% resolution time cuts, and financial services deployments (Danske Bank) report 300% efficiency improvements in runbook automation. The vendor ecosystem has solidified: BMC HelixGPT, PagerDuty Advance, incident.io, Rootly, ServiceNow, and others all ship production-ready features for automation and postmortem drafting. Yet April 2026 evidence reveals a darker reality: operational toil actually increased 30% in 2025 despite AI investment, teams deployed AI agents without runbook discipline, and post-mortem practices are plagued by systemic root-cause analysis failures—most AI incident postmortems blame model hallucination when the real cause is permission misconfiguration. Adoption stalls at organisational boundaries, not technical ones. Governance gaps, poor incident-data quality, and unclear ROI measurement constrain deployment more than capability shortfall. Accuracy risks are acute: industry-average hallucination rates of ~20% make AI postmortems unreliable for high-stakes contexts; single-model AI postmortems hallucinate details; forensic readiness gaps prevent effective incident investigation; and high-stakes deployments (police report generation) have surfaced false attributions and evidence distortions. The practice is vendor-mature and proven in select deployments, but the organisational prerequisites are steep -- mature incident-reporting cultures, strong data hygiene, runbook discipline, and governance frameworks -- and most teams haven't cleared them.
The vendor ecosystem has crystallised into mature, GA-ready offerings. BMC HelixGPT (26.1), PagerDuty Advance, ServiceNow automated post-incident review agents, Rootly, Datadog Bits AI, and incident.io all ship production-ready features for runbook automation and postmortem generation. incident.io's March 2026 GA launch includes one-click AI draft generation from Slack/Teams, timeline reconstruction from event data, and AI accuracy validation that flags missing or contradictory details. PagerDuty's Scribe agent drafts structured post-mortem summaries in Teams; Datadog generates postmortems in under a minute (after 100+ hours tuning); BMC's Helix performs 5-Why analysis and root cause extraction. Real deployments deliver quantified returns in financial services and IT operations: Danske Bank achieved 300% resilience efficiency gains in runbook automation; SolarWinds measured 17.8% incident resolution time reduction across 2,000+ ITSM systems; incident.io customers report 37% MTTR reduction and $29,700 annual savings. Real-world deployments also reveal acute failure modes: Runcycles documented 20+ AI agent incidents with costs ranging from $1.40 to $12,400 in direct spend and up to $50K+ business impact—exactly the failures runbooks should prevent. Operator discipline is collapsing: April 2026 evidence shows operational toil increased 30% despite AI investment because teams deployed agents without runbook discipline; 69% of AI-powered decisions still require human verification, creating a "messy middle" where the automation layer was added but the manual layer wasn't removed. Post-mortem quality is systemically broken: most AI incident postmortems miss root causes by focusing on model hallucination when the real cause is credential misconfiguration—a systematic failure pattern in how teams analyze incidents. Large-firm AI adoption in IT operations has stalled at 12%, with only 14% of enterprises successfully scaling pilots to production. Hallucination remains endemic: industry-average rate of ~20% (1 error per 5 queries) and legal AI research showing 65–43% accuracy rates highlight accuracy risks in high-stakes operational documentation. The binding constraints are organisational. Incident-reporting systems remain underused due to blame culture and reporting friction, starving AI models of training data. Most AI deployments lack the telemetry infrastructure (model versions, prompt logs, retrieval context, embedding versions) needed for effective forensic postmortems. In regulated and high-stakes contexts, accuracy risks are acute: police AI report generation has produced hallucinated officer attributions and evidence distortions; compliance-critical environments face a fundamental tension between probabilistic AI outputs and deterministic documentation requirements. Governance frameworks (terminology control, human review workflows, audit trails) are emerging as essential, not optional. Runbook discipline—structured, testable, maintainable procedures—is foundational. Successful deployments cluster where blameless postmortem cultures and strong incident-data hygiene already exist -- the AI amplifies mature practices rather than compensating for absent ones.
— Peer-reviewed Amazon Science research on AI agents improving runbooks in operational incident resolution systems; validates agentic runbook automation at academic credibility level and demonstrates core practice maturity in research publications.
— Comparison of 10 production AI incident management platforms documenting ecosystem maturity; catalogs AI-specific runbook and postmortem capabilities (drift monitoring, hallucination detection, prompt tracing, governance) as standard platform features.
— Comprehensive runbook design guide for LLM incident response covering severity classification, detection signals, containment primitives, and RCA templates adapted for AI failure classes; foundational operational documentation for AI system reliability.
— Detailed production postmortem documenting AI incident RCA with four root causes, 3-layer guardrail remediation, and validation metrics (0 incidents over 4.2M requests); demonstrates mature post-incident documentation practices adapted for AI system failures.
— Systematic catalog of 12 AI agent failure modes from incident analysis with detection signals and operational containment procedures; directly informs runbook design for identifying and mitigating production failures.
— Independent analysis of vendor positioning and postmortem/runbook data as training-signal moat; identifies incident meeting transcripts as emerging corpus shift and highlights verification gap in AI-generated incident documentation.
— Analysis of two named AI agent incidents (PocketOS database wipe, auth system rewrite) with root cause analysis and three operational controls (snapshots, least-privilege, mandatory checkpoints) that runbooks must enforce to prevent catastrophic failures.
— Technical framework for designing reversible effects and rollback procedures in AI automation workflows; core architectural pattern for incident recovery documentation and runbook design in AI systems.