The AI landscape doesn't move in one direction — it lurches. Some techniques leap from experiment to table stakes in a single quarter; others stall against regulatory walls, technical ceilings, or organisational inertia that no amount of hype can dislodge. Knowing which is which is the hard part. The State of Play cuts through the noise with a rigorously maintained index of AI techniques across every major business domain — classified by maturity, evidenced by real-world adoption, and updated daily so you always know where you stand relative to the field. Stop guessing. Start knowing.
A daily newsletter distilling the past two weeks of movement in a domain or two — delivered to your inbox while the index updates in the background.
Each dot marks the weighted maturity of practices within a domain — hover for a brief summary, click for more detail
AI that monitors infrastructure configurations for drift from desired state and can automatically remediate deviations. Includes policy-as-code enforcement and drift alerting; distinct from change risk assessment which evaluates planned changes rather than detecting unplanned ones.
Configuration drift detection and remediation is a mature, proven practice with GA tooling from every major cloud vendor and a growing ecosystem of specialized platforms. The question for infrastructure teams is no longer whether to detect drift but how to remediate it safely and automatically at scale. Drift detection itself — comparing live resources against IaC definitions — reached commodity status by 2024 across AWS, Azure, GCP, Oracle, and Kubernetes. The frontier has shifted to AI-assisted remediation, policy-to-code workflows, and continuous governance that translate detected deviations directly into versioned fixes. Documented deployments show concrete ROI: reduced MTTR, six-figure annual savings, and significant cuts to cloud waste. Yet a persistent adoption paradox constrains the practice's impact. Surveys show only 6% of organisations achieve full cloud codification despite 89% claiming IaC adoption, and fewer than a third proactively monitor for misconfigurations. The core tension remains operational: ClickOps — manual console changes during incidents — continues because enforcing IaC discipline conflicts with incident-response speed. Tooling has outpaced organisational readiness, making culture and change management the binding constraint rather than technical capability.
The vendor ecosystem now treats drift remediation — not just detection — as a core platform capability. AWS's Managed Services Trusted Remediator shipped with 116 automated remediations and claims a 95% reduction in remediation time. CloudFormation's drift-aware change sets, independently validated in production, offer three-way comparisons between templates, prior state, and live resources, letting operators revert drift without rewriting templates. Firefly, env0, and Devonair have each released AI-assisted remediation features that translate policy violations into IaC code fixes across multi-cloud environments. May 2026 vendor signals confirm ecosystem acceleration: IBM/HashiCorp's HCP Terraform public preview integrates Infragraph knowledge graphs for unified drift management across multi-cloud deployments with real-time asset state tracking; Pulumi shipped Helm Chart v4 with enhanced drift remediation across all SDKs (TypeScript, Go, Python, .NET, Java, YAML). AWS DevOps Agent (GA March 2026) demonstrates autonomous security remediation with architecture claims of 75% MTTR reduction via topology-aware agents and Model Context Protocol integration. Firefly customer case studies document measurable outcomes: Comtech reports $180K in annual savings, Basis Technologies cut cloud waste by 83% through continuous governance.
Evidence from May 2026 shows the practice remains grounded in organisational reality, not just vendor momentum. A Qualys analyst report (250+ enterprise survey, May 2026) identifies a critical remediation bottleneck: 49.4% of organisations still rely on monitoring + manual response workflows, leaving organisations vulnerable to remediation delays. A separate survey of 250 security professionals across FinServ, Retail, Public Sector, Healthcare, and Critical National Infrastructure found that 97% of organisations experienced drift-related incidents in the past 12 months, yet remediation cycles average 8+ days, leaving organisations in exploitable exposure windows. Platform engineering practitioners note that the gap is no longer detection (which is universal) but safe remediation: teams can identify drift but struggle to correct it without infrastructure ontology encoding resource relationships, policies, and ownership. Drift detection coverage across IaC frameworks (Terraform, OpenTofu, CloudFormation, Kubernetes) has become a baseline procurement criterion, actively driving platform switching decisions.
Emerging operational patterns highlight new drift vectors. Particle41 consulting firm documents AI agents making direct infrastructure changes that bypass IaC pipelines, creating untracked drift (e.g., resource right-sizing creating IaC-reality divergence). Client case studies show one organisation reduced infrastructure audit time from 40 hours/quarter to 4 hours through enforced IaC gates for agent outputs; another caught security misconfiguration before agent deployment through continuous drift monitoring. Recovery and disaster-recovery testing surfaces detection gaps: NTCTech documented a quarterly recovery drill that exposed four months of silent drift (service endpoints changed via manual updates, certificate trust paths rotated, security policies tightened without runbook updates) — the backup was consistent but the recovery environment was not. Organisational adoption barriers remain despite mature tooling. Firefly's 2025 IaC Report found that fewer than a third of organisations proactively monitor and remediate misconfigurations, and only 6% have codified their full cloud footprint — despite near-universal claims of IaC adoption. Real-world deployment data from April–May 2026 confirms these constraints persist: a practitioner case study documents 47 drifted resources accumulating silently over 4 months across 3 AWS accounts from incident-response console changes; remediation consumed 3 engineers for 2 full days. A critical failure case (GitLab.com incident April 2026, root cause July 2023) shows how stale Terraform plans can execute against live production with catastrophic results (130+ minute site outage, 617 resources marked for destruction). The gap is not tooling but discipline: practitioners still resort to manual console changes during incidents because IaC enforcement introduces friction when speed matters most. A 2025 breach analysis (Secure.com) found that 55% of cloud breaches trace to drift/misconfiguration and 82% of configuration errors originate from manual changes — evidence that drift remains a primary breach driver even as detection maturity increases. The practice has arrived as good-practice; rolling it out is an organisational change management challenge, not a technology procurement one.
— Qualys analyst report (250+ enterprise survey): 49.4% of organizations rely on monitoring + manual response workflows vs. infrastructure-as-code, identifying remediation speed lag as critical operational risk and security control.
— Pulumi Helm Chart v4 GA: enhanced drift remediation for Kubernetes across all SDKs (TypeScript, Python, Go, .NET, Java, YAML) addressing prior chart resource inconsistencies and improving Helm deployment governance.
— 2026 guide on safe AI-assisted IaC workflows: drift detection (CloudQuery, Driftctl) positioned as mandatory control for AI agent outputs. Real case study: manufacturing company's drift detection caught legacy team's unauthorized database replica creation.
— NTCTech recovery drill incident: four months of silent drift accumulated between backup capture and recovery target (endpoint changes, certificate paths, network policies). Demonstrates drift detection gap in DR/recovery workflows.
— Infrastructure practitioner analysis with three named deployments: GitOps reduced mean time-to-detect from 48 hours to under 5 minutes; immutable infrastructure achieved 90% reduction in incidents and <10min MTTR vs 2 hours.
— Lavawall (ThreeShield) drift detection for M365/Entra/Azure: extends practice beyond IaC to identity and policy configurations. Demonstrates product-ready detection, severity assessment, attribution, and rollback workflows in regulated environments.
— AWS DevOps Agent (GA March 2026) autonomous security remediation: detects S3 bucket policy drift and other misconfigurations. Architecture claims 75% MTTR reduction via topology-aware agents, MCP integration, and immutable audit trails.
— 2025 breach analysis: 55% of cloud breaches trace to drift/misconfiguration; 82% of config errors from manual changes; half of audit failures involve configuration findings. Quantifies drift as systemic breach and compliance driver.