Perly Consulting │ Beck Eco

The State of Play

A living index of AI adoption across industries — where established practice meets the bleeding edge
UPDATED DAILY

The AI landscape doesn't move in one direction — it lurches. Some techniques leap from experiment to table stakes in a single quarter; others stall against regulatory walls, technical ceilings, or organisational inertia that no amount of hype can dislodge. Knowing which is which is the hard part. The State of Play cuts through the noise with a rigorously maintained index of AI techniques across every major business domain — classified by maturity, evidenced by real-world adoption, and updated daily so you always know where you stand relative to the field. Stop guessing. Start knowing.

The Daily Dispatch

A daily newsletter distilling the past two weeks of movement in a domain or two — delivered to your inbox while the index updates in the background.

AI Maturity by Domain

Each dot marks the weighted maturity of practices within a domain — hover for a brief summary, click for more detail

DOMAIN
BLEEDING EDGEESTABLISHED

Configuration drift detection & remediation

GOOD PRACTICE

TRAJECTORY

Stalled

AI that monitors infrastructure configurations for drift from desired state and can automatically remediate deviations. Includes policy-as-code enforcement and drift alerting; distinct from change risk assessment which evaluates planned changes rather than detecting unplanned ones.

OVERVIEW

Configuration drift detection and remediation is a mature, proven practice with GA tooling from every major cloud vendor and a growing ecosystem of specialized platforms. The question for infrastructure teams is no longer whether to detect drift but how to remediate it safely and automatically at scale. Drift detection itself — comparing live resources against IaC definitions — reached commodity status by 2024 across AWS, Azure, GCP, Oracle, and Kubernetes. The frontier has shifted to AI-assisted remediation, policy-to-code workflows, and continuous governance that translate detected deviations directly into versioned fixes. Documented deployments show concrete ROI: reduced MTTR, six-figure annual savings, and significant cuts to cloud waste. Yet a persistent adoption paradox constrains the practice's impact. Surveys show only 6% of organisations achieve full cloud codification despite 89% claiming IaC adoption, and fewer than a third proactively monitor for misconfigurations. The core tension remains operational: ClickOps — manual console changes during incidents — continues because enforcing IaC discipline conflicts with incident-response speed. Tooling has outpaced organisational readiness, making culture and change management the binding constraint rather than technical capability.

CURRENT LANDSCAPE

The vendor ecosystem now treats drift remediation — not just detection — as a core platform capability. AWS's Managed Services Trusted Remediator shipped with 116 automated remediations and claims a 95% reduction in remediation time. CloudFormation's drift-aware change sets, independently validated in production, offer three-way comparisons between templates, prior state, and live resources, letting operators revert drift without rewriting templates. Firefly, env0, and Devonair have each released AI-assisted remediation features that translate policy violations into IaC code fixes across multi-cloud environments. May 2026 vendor signals confirm ecosystem acceleration: IBM/HashiCorp's HCP Terraform public preview integrates Infragraph knowledge graphs for unified drift management across multi-cloud deployments with real-time asset state tracking; Pulumi shipped Helm Chart v4 with enhanced drift remediation across all SDKs (TypeScript, Go, Python, .NET, Java, YAML). AWS DevOps Agent (GA March 2026) demonstrates autonomous security remediation with architecture claims of 75% MTTR reduction via topology-aware agents and Model Context Protocol integration. Firefly customer case studies document measurable outcomes: Comtech reports $180K in annual savings, Basis Technologies cut cloud waste by 83% through continuous governance.

Evidence from May 2026 shows the practice remains grounded in organisational reality, not just vendor momentum. A Qualys analyst report (250+ enterprise survey, May 2026) identifies a critical remediation bottleneck: 49.4% of organisations still rely on monitoring + manual response workflows, leaving organisations vulnerable to remediation delays. A separate survey of 250 security professionals across FinServ, Retail, Public Sector, Healthcare, and Critical National Infrastructure found that 97% of organisations experienced drift-related incidents in the past 12 months, yet remediation cycles average 8+ days, leaving organisations in exploitable exposure windows. Platform engineering practitioners note that the gap is no longer detection (which is universal) but safe remediation: teams can identify drift but struggle to correct it without infrastructure ontology encoding resource relationships, policies, and ownership. Drift detection coverage across IaC frameworks (Terraform, OpenTofu, CloudFormation, Kubernetes) has become a baseline procurement criterion, actively driving platform switching decisions.

Emerging operational patterns highlight new drift vectors. Particle41 consulting firm documents AI agents making direct infrastructure changes that bypass IaC pipelines, creating untracked drift (e.g., resource right-sizing creating IaC-reality divergence). Client case studies show one organisation reduced infrastructure audit time from 40 hours/quarter to 4 hours through enforced IaC gates for agent outputs; another caught security misconfiguration before agent deployment through continuous drift monitoring. Recovery and disaster-recovery testing surfaces detection gaps: NTCTech documented a quarterly recovery drill that exposed four months of silent drift (service endpoints changed via manual updates, certificate trust paths rotated, security policies tightened without runbook updates) — the backup was consistent but the recovery environment was not. Organisational adoption barriers remain despite mature tooling. Firefly's 2025 IaC Report found that fewer than a third of organisations proactively monitor and remediate misconfigurations, and only 6% have codified their full cloud footprint — despite near-universal claims of IaC adoption. Real-world deployment data from April–May 2026 confirms these constraints persist: a practitioner case study documents 47 drifted resources accumulating silently over 4 months across 3 AWS accounts from incident-response console changes; remediation consumed 3 engineers for 2 full days. A critical failure case (GitLab.com incident April 2026, root cause July 2023) shows how stale Terraform plans can execute against live production with catastrophic results (130+ minute site outage, 617 resources marked for destruction). The gap is not tooling but discipline: practitioners still resort to manual console changes during incidents because IaC enforcement introduces friction when speed matters most. A 2025 breach analysis (Secure.com) found that 55% of cloud breaches trace to drift/misconfiguration and 82% of configuration errors originate from manual changes — evidence that drift remains a primary breach driver even as detection maturity increases. The practice has arrived as good-practice; rolling it out is an organisational change management challenge, not a technology procurement one.

TIER HISTORY

ResearchJan-2020 → Jan-2020
Bleeding EdgeJan-2020 → Apr-2024
Leading EdgeApr-2024 → Oct-2025
Good PracticeOct-2025 → present

EVIDENCE (107)

— Qualys analyst report (250+ enterprise survey): 49.4% of organizations rely on monitoring + manual response workflows vs. infrastructure-as-code, identifying remediation speed lag as critical operational risk and security control.

Example: Argo CdProduct Launches

— Pulumi Helm Chart v4 GA: enhanced drift remediation for Kubernetes across all SDKs (TypeScript, Python, Go, .NET, Java, YAML) addressing prior chart resource inconsistencies and improving Helm deployment governance.

— 2026 guide on safe AI-assisted IaC workflows: drift detection (CloudQuery, Driftctl) positioned as mandatory control for AI agent outputs. Real case study: manufacturing company's drift detection caught legacy team's unauthorized database replica creation.

— NTCTech recovery drill incident: four months of silent drift accumulated between backup capture and recovery target (endpoint changes, certificate paths, network policies). Demonstrates drift detection gap in DR/recovery workflows.

— Infrastructure practitioner analysis with three named deployments: GitOps reduced mean time-to-detect from 48 hours to under 5 minutes; immutable infrastructure achieved 90% reduction in incidents and <10min MTTR vs 2 hours.

How Drift Detection WorksProduct Launches

— Lavawall (ThreeShield) drift detection for M365/Entra/Azure: extends practice beyond IaC to identity and policy configurations. Demonstrates product-ready detection, severity assessment, attribution, and rollback workflows in regulated environments.

— AWS DevOps Agent (GA March 2026) autonomous security remediation: detects S3 bucket policy drift and other misconfigurations. Architecture claims 75% MTTR reduction via topology-aware agents, MCP integration, and immutable audit trails.

— 2025 breach analysis: 55% of cloud breaches trace to drift/misconfiguration; 82% of config errors from manual changes; half of audit failures involve configuration findings. Quantifies drift as systemic breach and compliance driver.

HISTORY

  • 2020: Configuration drift emerged as recognized operational and security problem; AWS and Oracle released native drift detection tooling; Accurics research found 90% of cloud resources drift post-deployment.
  • 2021: Drift detection achieved multi-platform coverage: Cloudskiff released driftctl for multi-cloud detection, Red Hat integrated proactive drift detection in OpenShift, and customer adoption (Cloud Posse) demonstrated real-world governance need. Tooling remains vendor-specific; automated remediation lags detection.
  • 2022-H1: Industry adoption metrics emerge (CSA survey: 43% of orgs experienced drift-related security incidents); driftctl matures to v0.38.0 with multi-cloud support; AWS auto-remediation via SSM advances remediation capabilities; Terraform implementation challenges surface (v1.1.0 drift detection regression). Adoption concentrated among sophisticated infrastructure teams; detection frequency remains low for most organizations (46% check monthly or less).
  • 2022-H2: Ecosystem expansion accelerates: Snyk launches GA drift detection (October), SpotOn fintech company adopts Spacelift and sees 86% reduction in infrastructure PRs; AWS advances automated drift remediation with EventBridge/SNS alerting and Config auto-remediation guides (September-December). Federal government analysis reveals widespread drift risk and inadequate detection frequency (November). Vendor tooling maturity increases while remediation remains mostly vendor-locked and not cross-platform.
  • 2023-H1: AWS formalizes drift management in Well-Architected Framework as disaster recovery best practice (April); Spacelift GA enhances drift detection and remediation tooling; Red Hat and Snyk continue ecosystem maturity. Oracle production deployments reveal implementation challenges (drift detection accuracy bugs). Vendor coverage becomes standard but cross-platform automated remediation and organizational adoption rates remain the limiting factors.
  • 2023-H2: Vendor maturity continues: Spacelift adds targeted replans feature for selective Terraform change application (July); AWS publishes CI/CD integration patterns for drift detection in CDK pipelines (August). Open-source driftctl faces production reliability issues with false negatives in multi-cloud scenarios (October). Ecosystem expansion and CI/CD integration advance the practice, but tool accuracy and cross-platform coverage remain adoption barriers.
  • 2024-Q1: Vendor ecosystem continues to expand with Kubernetes/Helm drift detection (Quali Torque, March) and AI-assisted root cause analysis (env0 Cloud Compass); security vendors add drift detection to posture management (Varonis, February). Snyk's GitHub Action for driftctl signals CI/CD integration maturity. Peer-reviewed research validates GitOps efficiency for drift remediation. Production deployment challenges emerge: AWS Config auto-remediation infinite loop bug reveals critical pitfall in parameter configuration, underscoring the need for careful validation before enabling automated remediation at scale.
  • 2024-Q2: Vendor ecosystem continues maturation: Spacelift positions automated drift discovery/remediation as core GA feature (Checkout.com deployment: 500+ deployments/day); HashiCorp's HCP Terraform integrates drift detection with OPA/Sentinel policy enforcement; Firefly's Series A funding reflects growing market demand (survey: 23% of practitioners manage 100+ cloud accounts, 2x YoY growth). IDC forecasts configuration management market at 24.3% CAGR through 2028. Open-source driftctl remains actively used but faces reliability challenges (permission configurations, false negatives). Drift detection expands beyond compute: StorageGuard adds 2000+ configuration checks for storage/backup systems, extending drift management across infrastructure domains.
  • 2024-Q3: Ecosystem expansion continues across container and multi-cloud layers: Microsoft enters market with binary drift detection in Defender for Containers (public preview); Spacelift releases OpenTofu v1.8 support with enhanced dashboard visibility; Firefly's Gartner SRE Hype Cycle inclusion signals analyst recognition of AI-assisted drift detection. Vendor tooling now spans compute, storage, containers, and orchestration platforms. Barriers persist: open-source driftctl reliability challenges remain, cross-platform automated remediation incomplete, organizational adoption still concentrated among sophisticated multi-cloud teams.
  • 2024-Q4: Analyst recognition and adoption metrics signal maturation: Gartner's 2024 Cool Vendors report includes Firefly for drift detection capabilities; market research shows 68% of enterprises report drift incidents annually and 72% of DevOps teams use automation for large-scale environments. AWS advances drift analysis with Bedrock LLM integration for root-cause diagnosis in Control Tower. Significant barriers persist: driftctl remains in maintenance mode with 133 open issues reflecting reliability gaps; critical adoption gap identified (89% claim IaC adoption but only 6% achieve full codification), revealing that ClickOps and manual changes continue to drive drift despite tooling availability. Drift detection is now table-stakes for infrastructure platforms, but cross-platform remediation and organizational commitment to IaC codification remain constraining factors.
  • 2025-Q1: Vendor ecosystem consolidation: AWS refreshes Well-Architected Framework guidance (February) embedding drift management as DR best practice; IBM Cloud releases GA drift detection in Schematics (January); Spacelift reaffirms drift as core platform feature; Microsoft continues container drift detection expansion. Industry data reveals persistent adoption gap: 73% of organizations have undetected drift despite tooling availability, and 68% of security incidents involve misconfigurations. Barrier analysis shows tooling maturity exceeds organizational adoption: the practice is vendor-complete but remains constrained by organizational culture (IaC discipline conflicts with incident response urgency), tool reliability (cross-platform remediation vendor-locked, open-source gaps), and change management complexity at scale. ClickOps remains prevalent.
  • 2025-Q2: Vendor ecosystem expands to container orchestration: Microsoft GA drift detection in Azure Kubernetes Fleet Manager (April); policy-to-code remediation platforms (Resourcely, Gomboc, Firefly) emerge with AI-assisted drift prevention. Industry paradox deepens: Firefly 2025 IaC Report documents drift as "growing operational problem" despite 89% claimed IaC adoption—revealing only 6% of organizations achieve full cloud codification. Drift detection is table-stakes across major platforms, but remediation remains vendor-locked and tool accuracy immature. Adoption barrier persists: ClickOps dominates production incident response because enforcing IaC discipline creates friction with incident speed requirements; organizational culture and incident-response change management remain the limiting factors, not tooling availability.
  • 2025-Q3: Ecosystem expansion continues into container runtime (Microsoft binary drift detection, July) and AI-assisted remediation (StackAnchor agent, July). Real-world deployments demonstrate drift's cost impact: Ziff Davis production case of server configuration standardization via EC2 Image Builder/Systems Manager; fintech example of forgotten cluster scaling creating IaC-reality divergence. Adoption paradox deepens: Firefly survey shows only 31% of organizations proactively monitor and remediate misconfigurations, yet 58% plan AI-driven capabilities within 6 months. Emerging pattern: combining drift detection with Just-In-Time access provisioning addresses incident-response friction. Fundamental shift in constraint analysis: drift detection is universal and mature; remediation is increasingly automated; organizational change management and incident-response culture remain the blocking factors, not tooling availability.
  • 2025-Q4: Vendor ecosystem focuses on safe, automated remediation at scale: AWS launches CloudFormation drift-aware change sets (November) enabling three-way template comparisons for production safety; Firefly introduces Cloud Resilience Posture Management with AI-driven policy-to-code remediation (November); env0 and Devonair both release AI-assisted continuous monitoring and auto-remediation (December). Real-world case study: enterprise AI workload deployment via AWS Config/Systems Manager demonstrating practical MTTR and reliability gains. Practice maturity signals a shift from "detecting drift" to "remediating drift safely and at scale"; however, organizational adoption barriers (IaC discipline vs. incident response speed) and change management challenges remain the limiting factors despite mature, readily available tooling across all major vendors.
  • 2026-Feb: Vendor ecosystem consolidates around automated remediation: AWS Managed Services Trusted Remediator achieves GA with 116 automated remediations reducing remediation time by 95% (February); independent technical validation of CloudFormation drift-aware change sets feature confirms production readiness (Classmethod, February); Firefly case studies document customer ROI with $180K annual savings and 83% waste reduction through continuous drift governance. Persistent tension evident: practitioner documentation continues to reveal Terraform state drift and manual remediation hacks required in production, indicating that despite mature tooling, operational complexity and organizational discipline remain constraining factors.
  • 2026-Mar: Real-world deployments validate practice maturity: Kubernetes production case study (14 of 47 deployments drifted; custom ArgoCD/Go controller detected unplanned changes; visibility enabled self-correction); European telco deployed hourly drift scans across 2000 AWS accounts, discovering 700 misconfigurations and auto-remediating 93% within 48 hours (€120K audit savings). Comprehensive operational guides (OpenTofu/Terraform drift detection, enterprise-scale Terraform management, cloud consulting firm detection strategies) document three-pillar maturity: detection universalized, remediation increasingly automated, organizational change management as binding constraint. Field evidence from regulated bank shows safe automated remediation with dry-run/rollback for governance; GitLab infrastructure team deploying Atlantis to improve Terraform drift remediation visibility. Practice signal: drift detection is commodity; safe remediation at scale remains the frontier.
  • 2026-Apr: Organisational adoption barriers persist despite mature tooling: survey of 250 security professionals across FinServ, Retail, Public Sector, and Healthcare confirms 97% of organisations experienced drift incidents yet remediation takes 8+ days on average; 72% of security budgets remain reactive rather than proactive. Platform engineering practitioners articulate critical remediation gap: teams detect drift reliably but cannot remediate safely without infrastructure ontology encoding resource relationships and ownership—a constraint that applies even as Driftctl deployments document 60% reductions in drift incidents where tooling is applied. Drift detection coverage across frameworks (Terraform, OpenTofu, CloudFormation, Kubernetes) has become a baseline procurement criterion actively driving platform switching decisions. AWS CloudFormation confirms detection as standard GA feature with dependency management and safety controls. Evidence signal: drift detection universalized to commodity; safe remediation at scale and organisational discipline remain the binding constraints.
  • 2026-May: High-profile failure cases reinforced the cost of gaps: a GitLab.com site-wide outage (130+ minute downtime, 617 resources marked for destruction) from a stale Terraform plan, and a NTCTech DR drill exposing four months of silent endpoint and certificate drift that made the backup environment unrecoverable. A Qualys survey (250+ enterprises) confirmed 49.4% still rely on monitoring plus manual response, while 55% of cloud breaches trace to drift/misconfiguration. Vendor tooling advanced on multiple fronts: IBM/HashiCorp HCP Terraform entered public preview with Infragraph knowledge-graph drift management across multi-cloud; Pulumi shipped Helm Chart v4 with enhanced Kubernetes drift remediation; and AWS DevOps Agent (GA March 2026) claims 75% MTTR reduction via topology-aware autonomous remediation. BMC Helix CMDB shipped GA change-correlation distinguishing authorised from unauthorised drift—signalling that the tooling frontier has moved from detection to governance-aware remediation, while ClickOps and organisational discipline remain the binding constraint.