Perly Consulting │ Beck Eco

The State of Play

A living index of AI adoption across industries — where established practice meets the bleeding edge
UPDATED DAILY

The AI landscape doesn't move in one direction — it lurches. Some techniques leap from experiment to table stakes in a single quarter; others stall against regulatory walls, technical ceilings, or organisational inertia that no amount of hype can dislodge. Knowing which is which is the hard part. The State of Play cuts through the noise with a rigorously maintained index of AI techniques across every major business domain — classified by maturity, evidenced by real-world adoption, and updated daily so you always know where you stand relative to the field. Stop guessing. Start knowing.

The Daily Dispatch

A daily newsletter distilling the past two weeks of movement in a domain or two — delivered to your inbox while the index updates in the background.

AI Maturity by Domain

Each dot marks the weighted maturity of practices within a domain — hover for a brief summary, click for more detail

DOMAIN
BLEEDING EDGEESTABLISHED

AI-assisted code review with auto-approve

BLEEDING EDGE

TRAJECTORY

Stalled

AI that autonomously approves and merges code changes meeting defined quality and safety thresholds. Includes automated merge for low-risk changes like dependency bumps; distinct from suggestion-mode review which always requires human sign-off.

OVERVIEW

Autonomous code approval -- where AI reviews, approves, and merges changes without a human gate -- remains organisationally stalled despite continued vendor maturity. The technical capability exists: Claude Code ships auto-merge, OpenAI deployed AI-reviewing-AI with 99.93% approval accuracy, and production case studies document bounded auto-approve (Ona, Augment, mabl). Yet every major deployment maintains human gates. The defining tension is not capability but safety: PanDev Metrics' study of 100 teams documented that AI-only auto-approve escapes 46% more defects and doubles severity-1 incident rates; independent benchmarks show review tools catch only 33% of production bugs. Auto-approve in production is consistently bounded (low-risk PRs under 1K LOC, no migrations or auth) with humans making final merge decisions. The practice remains bleeding-edge where vendor maturity meets hard organisational constraints: 96% of developers distrust AI code accuracy, and review has become the visible bottleneck (91% increase in review time despite code acceleration). Whether organisations will delegate the approve-and-merge decision to AI systems continues to receive negative signals.

CURRENT LANDSCAPE

Vendor ecosystem maturity has reached full GA on multiple fronts. Claude Code ships auto-merge, OpenAI's Auto-review system (deployed in Codex) achieves 99.93% approval accuracy with 99.3% prompt-injection blocking—the first frontier-lab production deployment of AI-reviewing-AI with quantified safety metrics. Amazon Q Developer and GitLab Duo offer integrated review automation. CodeRabbit has connected 2M repositories and analyzed 75M defects. Production deployments demonstrate organisationally-acceptable bounded auto-approve: Ona achieved 74% lead-time reduction (4.1h → 1.1h) with explicit low-risk scoping (<1K LOC, no migrations or auth changes); Augment's Cosmos agents achieved 3x code output with Intent Reviewer gates for high-judgment decisions; mabl scaled to 75+ repositories with strict governance that "there is no scenario where code auto-merges without human approval."

Yet this vendor maturity encounters hard safety and organisational limits. PanDev Metrics' empirical study of 100 B2B teams (23,847 PRs) found AI-only auto-approve configuration escapes 46% more defects with 18% post-merge rework rate and near-doubled severity-1 incident rate. Independent benchmark (Entelligence) on 67 production bugs shows CodeRabbit catches 33%, Copilot 22.6%—insufficient safety margins for autonomous approval. Developers spend 91% longer reviewing PRs despite AI acceleration, and 96% distrust AI code accuracy. Practitioner governance consistently rejects full automation: mabl, Google Cloud (after a YOLO-mode incident), and other deployments maintain human approval gates. Security research documented exploitation vectors: prompt injection can spoof git author identity and trigger auto-approve of malicious payloads (12,400+ public claude-code-action workflows exposed). The core pattern is unchanged: organisations adopt AI review as productivity augmentation while governance demands human final authority on merge decisions. Code-quality bottlenecks and trust deficits prevent delegation of autonomous approval.

TIER HISTORY

ResearchJun-2024 → Oct-2024
Bleeding EdgeOct-2024 → present

EVIDENCE (79)

— MSR 2026: 28.3% AI PRs merge instantly but many agents fail to converge under review. GitClear documents 9x higher code churn with AI. 66% of developers report AI outputs 'almost correct' but flawed.

— Independent benchmark on 67 production bugs: CodeRabbit 33% catch rate, Copilot 22.6%. Low safety margins undermine auto-approve assumptions; combined with PanDev's 46% defect escape, margin approaches zero.

— Plandek analyzed 2,000+ teams: code review became visible bottleneck; bottom-quartile teams take 35+ hours to merge, top teams 21 hours. AI exposes delivery system weakness, does not fix it.

— PanDev Metrics tracked 100 teams over 15 months analyzing 23,847 PRs: AI-only auto-approve escapes 46% more defects, generates 18% post-merge rework rate, doubles severity-1 incidents vs baseline.

— Ona deployed bounded auto-approve (low-risk: <1K LOC, no migrations/auth). Lead time dropped 74% (4.1h → 1.1h), deploys tripled (3.1x). Human always merges; governance via objective criteria.

— Augment deployed Cosmos agents with auto-approve for low-risk PRs (docs, configs). Code output 3x, merge time halved, bug rate per output stable. Intent Reviewer gates high-judgment decisions to humans.

— Security researchers exploited auto-approve in Claude Code with git identity spoofing + malicious payload. 12,400+ public workflows use claude-code-action. Documented supply-chain attack vector against auto-approve.

— LinearB + CircleCI 2026: PR review time increased 91% despite AI acceleration; 39-point perception gap between feeling fast and actual delivery. Review is now the critical constraint, not code generation.

HISTORY

  • 2024-Q2: Initial evidence gathered. Large-scale developer surveys (481 and 395 respondents) documented adoption barriers including trust deficits and policy constraints. Research papers (EASE 2024, grounded theory study) provided empirical data on developer perceptions and organizational adoption dynamics. Google's Gerrit plugin demonstrated enterprise investment in AI infrastructure. Production reliability concerns surfaced in Qodo PR-Agent bug reports. Practitioner assessments highlighted context awareness as a limiting factor.
  • 2024-Q4: Vendor ecosystem accelerated with GitHub Copilot code review reaching GA (1M+ users in preview) and AWS releasing Amazon Q code review automation. Qodo Merge reported 20k+ daily PRs and Fortune 500 adoption. Industrial study (ICSE 2025) of Qodo across 10 companies with 1,568 reviews revealed mixed outcomes: higher bug detection but also irrelevant comments and 2.5h longer PR closure times. Research identified "tunnel vision" effects and context-awareness gaps as persistent barriers to broader adoption. Auto-approve workflows remain limited to low-risk categories (docs, dependency bumps).
  • 2025-Q1: Major vendors accelerated GA deployments: GitHub expanded Copilot auto-review configuration for organization rulesets, AWS launched Amazon Q /review agent across all regions. Market growth projected at 9.2% CAGR to $750M. However, Stack Overflow survey (65k developers) showed adoption climbing to 84% but trust stalled: only 3% highly trust AI, 46% actively distrust accuracy, and only 31% use AI agents. Deployed systems showed reliability gaps (Copilot skipping "low risk" files), and peer research (CHASE 2025, 20 engineers) confirmed persistent trust and context-awareness barriers. Alert noise and false positives remained significant adoption friction. Auto-approve remained limited to low-risk categories despite vendor momentum.
  • 2025-Q2: Vendor ecosystem expanded further with AWS extending Amazon Q Developer to GitHub (May) and GitLab announcing partnership integration (June), signaling multi-platform maturity. Greptile reported processing 700k+ PRs monthly, indicating substantial scale in deployed auto-review workflows. However, adoption barriers persisted: trust remained fragile despite tool availability, context-awareness limitations continued to constrain scope to low-risk categories (dependencies, docs), and alert-noise issues remained unresolved. Industry assessment (Qodo, mid-2025) confirmed that while AI code review was mainstream at vendors and in pilot/early adoption at enterprises, autonomous approval workflows with minimal human gates remained limited to narrow, provably-safe change categories. Market confidence remained high ($750M projected revenue), but production adoption remained gated by reliability and trust gaps.
  • 2025-Q3: Vendor ecosystem continued advancing with AWS releasing Amazon Q Developer for GitHub preview (September 2025), but deployment evidence revealed critical safety and adoption gaps. Jellyfish's study of 1,000 reviews across 400 companies (May-July 2025) showed agents in only 22% of reviews with 18% leading to changes, indicating minimal practical impact. Apiiro's Fortune 50 research documented 10x more security findings and 322% spike in architectural flaws from AI-generated code, concluding that AI adoption must be paired with mandatory AppSec. Canva's survey of 300 tech leaders revealed 93% enforce peer review despite 92% tool adoption, confirming universal organizational blocks on autonomous approval. Graphite's internal testing found persistent false positives and hallucinations, concluding final approval should remain human. Enterprise analysis reported underwhelming ROI (~10%) and pervasive skepticism about autonomous production agents. The widening gap between vendor capability growth and actual autonomous deployment hardened around safety, governance, and credibility.
  • 2025-Q4: Vendor ecosystem reached GA maturity with Amazon Q Developer for GitHub (November) and GitLab Duo Code Review (December), confirming multi-platform automated code review availability. Production case study evidence emerged (Voithru: 50% bug reduction, 5x review volume, 80% test coverage), alongside practitioner reports of high initial noise (80%) requiring custom tuning. Adoption metrics surged: code review agents grew from 14.8% (Jan) to 51.4% (Oct 2025), and market projections reached $25.7B by 2030. However, developer trust declined sharply to 33% despite 84% adoption, with primary frustrations centered on code quality ("almost right but not quite") and debugging burden. The core dynamic remained unchanged: GA tooling and growing pilot deployments masked persistent governance, quality, and trust barriers to autonomous approval. Auto-approve remained limited to low-risk categories and required heavy human oversight.
  • 2026-Jan: Market adoption accelerated: CodeRabbit's platform reached 2M connected repositories and 75M defects analyzed, with NVIDIA as a named enterprise customer. Industry analyst report (Zylos) documented 84% developer adoption of AI tools and $750M market size with 9.2% CAGR growth. However, developer sentiment remained contradictory: 96% reported low trust in AI-generated code accuracy; AI-generated PRs had 32.7% acceptance vs 84.4% for manual code and faced 4.6x longer review times. Research analysis of integrated review systems revealed fundamental architectural failures: systems that generate and review code show 8x more duplicated code, 39.9% fewer refactors, and 37.6% higher vulnerabilities. Individual and enterprise deployments emerged (Leena Malhotra's multi-model review workflow, Platformr's six-month Amazon Q integration with custom rules), but successes remained contingent on extensive human oversight and project-specific tuning. Auto-approve workflows continued to require high human involvement despite vendor GA tooling maturity.
  • 2026-Feb: Vendor auto-merge capabilities expanded with Anthropic's Claude Code shipping autonomous merge features (auto-merge when all CI checks pass) and Amazon Q Developer extending GitHub integration with on-demand /q review command. End-to-end automation pipelines emerged in practice (Zenn case study using Claude Code + Copilot for full implementation-review-merge automation). However, developer trust remained a critical barrier: Stack Overflow's 2026 data showed only 29% of developers trust AI tools (down from 2024), underscoring persistent organizational reluctance to delegate approval authority to AI systems. The practice remained at bleeding-edge with product maturity on the vendor side but organizational adoption severely constrained by governance and trust deficits.
  • 2026-Mar: Product auto-approve capability further validated: GitHub v1.110 officially released /autoApprove and /yolo commands with terminal sandboxing; GitHub Copilot Code Review reached 60M cumulative reviews handling 20% of all PRs across 12K+ organizations. However, critical adoption barriers hardened. Amazon formally restricted AI coding tools post-incident (March 5 shopping outage), requiring mandatory senior review for all AI-assisted code—the first major tech company to formally gate AI tools due to production failures. METR research published peer-reviewed evidence of a 24-point gap between automated test pass rates (76%) and human reviewer approval (52%), proving automated signals cannot substitute for human judgment. Production case studies (HubSpot Sidekick AI reviewer, 90% feedback time reduction) confirmed that even at enterprise scale, autonomous approval remains undeployed—human gates persist via filtering agents. Practitioner analyses reinforced scope constraints: 400+ team study showed teams removing human review entirely experienced higher change failure rates; highest-performing teams use AI as an augmentation layer, not a replacement; practitioners propose limiting auto-approve to mechanical changes (dependency updates, migrations) while requiring human oversight for logic and architecture. The practice stayed at bleeding-edge with widening evidence that product maturity cannot overcome organizational risk aversion and quality concerns.
  • 2026-Apr (Week 1-3): Critical security vulnerabilities in auto-approve infrastructure exposed. IDEsaster vulnerability class (24 assigned CVEs, dozens more pending) demonstrated that prompt injection defeats auto-approval gates in Cursor, GitHub Copilot, Windsurf, Zed, and other tools, enabling data exfiltration and RCE. CVE-2026-30304 documented that AI Code's binary safe/unsafe command classification is vulnerable to prompt manipulation, invalidating assumption that AI can autonomously classify execution safety. Formal verification study (arXiv 2604.05292) of 3,500 code artifacts across 7 frontier LLMs quantified generation–review asymmetry: models identify 78.7% of vulnerabilities when reviewing vs 55.8% when generating, proving AI cannot reliably review its own code for autonomous approval. Security audit of 50+ production applications showed 92% contain critical vulnerabilities with 18-day average to exploitation—establishing baseline risk profile that makes autonomous approval unsafe. Meanwhile, adoption barriers hardened: developer trust collapsed to 29% despite 84% tool adoption; code quality metrics showed 1.7x more issues and 4x code duplication in AI-authored code; benchmark evaluation (200k+ PRs) found current tools at 50-60% F1 effectiveness with 96% of developers distrusting AI code. Vendor GA capabilities (Amazon Q, CodeRabbit, GitHub Autofix) matured in the same window, but organizational adoption stalled: enterprises continued restricting auto-approve to mechanical changes (dependency updates, migrations) requiring extensive human vetting and project-specific tuning.
  • 2026-Apr (Week 4): Vendor ecosystem continued maturing while security and governance concerns hardened. GitHub Copilot launched global auto-approve feature in JetBrains IDEs (April 24), automatically approving all tool calls including destructive actions. Cloudflare published production case study (April 20) of orchestrated 7-specialist AI code review system handling tens of thousands of MRs at scale, demonstrating organizational sophistication in auto-review orchestration. Amazon Q Developer confirmed code review automation as GA (April 20), integrated with GitHub Enterprise. Academic research (April 21) analyzed 33,000+ AI-generated PRs documenting acceptance patterns and flawed code still being merged. However, a critical supply-chain incident (Clinejection, April 27) exposed how autonomous workflows amplify risk: prompt injection in a GitHub issue title hijacked Cline's AI triage bot, leading to credential theft and malicious code deployment on 4,000 machines within 8 hours, with npm audit, code review, and provenance attestation all failing to detect the attack. Adoption metrics (April 24) showed review bottleneck worsening (review time up 91%) as code generation accelerates, driving demand for auto-approve but also revealing fundamental verification burden. Governance frameworks (April 20) proposed autonomy escalation models (shadow → advisory → co-pilot → autopilot) with quality gates. The gap between tooling maturity and organizational readiness continues to widen, now reinforced by concrete security incidents demonstrating that removing human review gates from automation multiplies vulnerability surface.
  • 2026-May: Production deployment patterns crystallized around bounded auto-approve with persistent human gates. PanDev Metrics' study of 100 B2B teams (23,847 PRs) showed AI-only auto-approve escapes 46% more defects, generates 18% post-merge rework, and doubles severity-1 incident rates; an independent benchmark of 67 production bugs found CodeRabbit catching 33% and Copilot 22.6% — margins too thin for autonomous approval. Successful bounded deployments (Ona: 74% lead-time reduction with <1K LOC scoping; Augment: 3x code output with Intent Reviewer gates) held governance constraints firm. OpenAI's AI-reviewing-AI system achieved 99.93% approval accuracy with 99.3% prompt-injection blocking, the first frontier-lab deployment with quantified safety metrics. A security incident (git identity spoofing) exploited auto-approve in Claude Code workflows affecting 12,400+ public workflows, and Plandek's analysis of 2,000+ teams confirmed review has become the primary delivery bottleneck — one that auto-approve alone cannot resolve without structural governance changes.