The AI landscape doesn't move in one direction — it lurches. Some techniques leap from experiment to table stakes in a single quarter; others stall against regulatory walls, technical ceilings, or organisational inertia that no amount of hype can dislodge. Knowing which is which is the hard part. The State of Play cuts through the noise with a rigorously maintained index of AI techniques across every major business domain — classified by maturity, evidenced by real-world adoption, and updated daily so you always know where you stand relative to the field. Stop guessing. Start knowing.
A daily newsletter distilling the past two weeks of movement in a domain or two — delivered to your inbox while the index updates in the background.
Each dot marks the weighted maturity of practices within a domain — hover for a brief summary, click for more detail
Multiple AI agents collaborating across development tasks such as planning, coding, reviewing, and testing in coordinated workflows. Includes orchestrated agent teams with specialised roles; distinct from single-agent agentic coding which uses one agent across the lifecycle.
Multi-agent development pipelines coordinate multiple specialised AI agents -- planner, coder, reviewer, tester -- across software engineering workflows, distributing tasks that single-agent systems handle monolithically. The premise is compelling: decompose complex development work the way human teams do, with role-specific agents handing off artifacts through an orchestrated pipeline.
By April 2026, infrastructure maturity has consolidated around standardized patterns. All major vendors (OpenAI Codex Subagents GA, Claude Code Agent Teams, GitHub Copilot /fleet, Devin 2.0, Cursor 3, VS 2026) shipped production multi-agent execution by Q2 2026. Framework ecosystem has stabilized: LangGraph (graph-based orchestration), CrewAI (role-based teams, 12M daily executions), AutoGen (now part of Microsoft Agent Framework 1.0), and AWS Strands Agents (2000+ stars) represent production-ready options. Real deployments demonstrating measurable value exist at organizational scale: 1inch's multi-agent CI pipeline (ticket-to-PR automation), Stripe Minions (1000+ unattended PRs/week), Google Agent Smith (25%+ of production code), OpenAI Harness (1M agent-written LOC). Yet enterprise adoption remains bifurcated: 11% in production vs 39-66% experimenting, with Gartner forecasting 40% project cancellation by 2027. The core bottleneck is not tooling but orchestration architecture: coordination overhead degrades sequential reasoning 39-70% (DeepMind/MIT 2026), specification gaps cause persistent 25-39pp accuracy loss independent of model capability, error cascades amplify false claims 17.2× without mitigation, and failure attribution across agent boundaries remains largely unsolved. Verified production failures include security vulnerabilities (86% XSS in design system), cost explosions (multi-agent costs 5-20× single-agent with comparable accuracy), reliability degradation (collaboration success 50% lower than solo agents), and cascading hallucinations where single false claim spreads through all agents within 3 rounds. Tightly scoped deployments with explicit orchestration topology (Klarna's structured LangGraph system, mabl's four-layer architecture, 1inch's parallel review pattern) deliver measurable value; scaling to enterprise workflows remains constrained by unresolved architectural brittleness.
Production deployments span both organizational-scale and process-scale implementations. At organizational scale: Stripe Minions produce 1000+ unattended PRs per week, Google Agent Smith reportedly handles >25% of new production code, OpenAI Harness shipped ~1M lines of agent-written code with zero manual writing, Ramp Inspect handles >50% of merged PRs. At process scale: 1inch deployed multi-agent CI pipeline with implement agent + seven parallel review perspective agents + synthesizer + fix agent, running ticket-to-PR autonomously with human approval; Klarna's structured LangGraph system processes 2.3 million conversations per month achieving $60M annual savings; mabl scaled to 39-60% AI-assisted commits across 75+ repos; Rakuten completed 12.5 million lines of code extraction in 7 hours with 99.9% accuracy using 24 parallel agents. FRE|Nxt's InterviewLM achieves 100+ concurrent sessions with 8 specialized agents and 40% cost savings. These successes share common patterns: tightly scoped domain definitions, explicit orchestration topology (supervisor+worker or hierarchical), human-in-the-loop approval gates, and disciplined skill/tool management via MCP.
However, the failure modes are now well-documented and structural. Peer-reviewed research shows two-agent systems lose 25-39 percentage points of accuracy when specification detail decreases -- a coordination gap that persists regardless of model capability. CooperBench's first multi-agent collaboration benchmark reveals agents achieve 50% lower success rates collaborating vs. solo. Research by DeepMind/MIT documents error amplification: 17.2× without explicit coordination, 4.4× with centralized orchestration; sequential reasoning degrades 39-70% across all multi-agent variants. Critical negative signal: peer-reviewed study found single false claim spreads to all agents within three rounds across six tested frameworks -- genealogy graph mitigation only raises defense success from 32% to 89%. Princeton research shows reliability lags accuracy by 50-86%; cascading failures reduce end-to-end system reliability to 74% even when components exceed 90%. A real design system deployment documented 86% generated components with XSS vulnerabilities, 70% missing accessibility, and cost variance $0.88–$146/component. Token consumption multiplies 5-20×, cost escalation unpredictable, and production failure research identifies 12 distinct failure modes: error propagation (silent cascades), non-determinism, state corruption, infinite loops, cost explosion, context exhaustion, retry complexity, orchestrator bypass, observability collapse, topology collapse, collusive validation, and reward hacking. Single critical negative signal analysis cites production data: three-agent document analysis costs $47k/month vs $22.7k single-agent with only 2.1pp accuracy difference.
Vendor platform maturity is accelerating (all major vendors shipped Q1 2026), MCP SDK reached 97M/month downloads, frameworks consolidated around LangGraph/CrewAI/AutoGen architectures. Yet enterprise adoption remains constrained: 11% in production vs 39-66% experimenting, with Gartner forecasting 40% project cancellation by 2027. The gap between pilot success (72% report pilot gains) and production viability (only 28% sustain post-deployment) points to unresolved infrastructure challenges: scheduling, lifecycle management, supervision hierarchies, and FinOps remain unsolved. Organizations attempting to scale encounter reliability requirements incompatible with current architectures, cost modeling failures, and governance gaps. Successful deployments limit coordination to 3 agents maximum, use human-in-the-loop checkpoints at stage boundaries, and implement explicit error as state (not exceptions) patterns. The practice demonstrates that tightly scoped, disciplined orchestration can deliver value, but general-purpose enterprise-scale adoption remains blocked by architectural brittleness and absence of proven scaling patterns.
— Synthesizes UC Berkeley MAST study analyzing 1,600+ execution traces documenting 41-86.7% failure rates across seven frameworks; identifies structural prevention patterns (scope hierarchy, authority attenuation, typed protocols) for production multi-agent reliability.
— Deriv deployed 50+ AI agents with registry-based orchestration and Operations Center architecture (Agent Officer pattern), demonstrating real-world scaling challenges (integration tax, capability explosion) and solution patterns for production multi-agent systems.
— GitLab's production architectural decision record evaluating five orchestration frameworks (LangGraph, Temporal, Prefect, Claude Agent SDK, Haystack) with explicit hard requirements and failure rationales, demonstrating framework selection criteria for enterprise multi-agent development systems.
— Official GA documentation for LangGraph orchestration runtime; confirms production trust by Klarna, Uber, and JPMorgan; establishes framework as de facto standard for stateful multi-agent orchestration with durable execution and human-in-the-loop support.
— Cursor built a browser in one week using 1M lines of code across 1,000 files with three-role hierarchical agent orchestration (Planner, Worker, Judge), demonstrating production multi-agent development pipeline at enterprise code scale with measured adoption metrics.
— StudioMeyer operates 40-agent fleet with three-layer observability architecture (Sentry, Langfuse, LangGraph); demonstrates stateful multi-agent workflows with Postgres checkpointing and resume-from-failure patterns at production scale.
— Peer-reviewed empirical study comparing in-context prompting vs. LangGraph orchestration across procedural tasks; shows orchestration failures (24% travel, 9% Zoom, 17% insurance) significantly exceed simpler in-context baseline, establishing critical negative signal constraining orchestration applicability.
— Research synthesis establishing Wave 1 (viability) vs Wave 2 (measurement) taxonomy; critical insight that single-agent systems with good interfaces often outperform multi-agent architectures (SWE-agent 10.7pp improvement from interface design), defining appropriate use cases.
2024-Q3: Early research phase with benchmark-driven validation. HyperAgent achieves SOTA on SWE-Bench and Defects4J. Enterprise case studies reveal $127M in failed deployments; critical blockers include inadequate testing and legacy system integration. Token efficiency concerns emerge from ChatDev analysis. Field consensus: feasible in research, barriers block production adoption.
2024-Q4: Production deployments emerge at scale: LinkedIn SQL Bot, Uber code migration, AppFolio copilot (10+ hrs/week savings), Elastic and Replit multi-agent systems all live in production. Framework infrastructure (LangGraph) matures. Adoption breadth grows (68% of companies deployed agents) but ROI gap widens (only 32% see significant value). Industry shift toward tightly scaffolded "intelligent workflows" signals recognition of autonomy limits. Conference engagement increases but practitioner analysis remains cautious: applications scarce, systems not yet human-assistant equivalents.
2025-Q1: Research now documents systematic failure modes: UC Berkeley peer-reviewed study identifies 18 failure patterns across 5 frameworks on 150+ tasks, with performance gains remaining minimal vs. single agents. New production case study: Build.inc's 25-agent LangGraph system reduces land diligence from 4 weeks to 75 minutes. Cloud vendors (AWS, Microsoft) release native multi-agent tutorials and orchestration capabilities. Critical gap emerges: practitioner analysis informed by Microsoft Research interviews surfaces underdeveloped debugging infrastructure, missing security/compliance standards, and tool immaturity as primary adoption barriers. Deployment breadth unchanged (68% companies) but ROI realization stalls (32% threshold). Bifurcation signal: domain-specific systems (land diligence, SQL conversion) demonstrate viability; general-purpose orchestration faces reliability and debuggability challenges.
2025-Q2: Framework infrastructure reaches maturity: LangGraph Platform reaches GA with 400 companies deploying to production. Anthropic releases production multi-agent system achieving 90.2% performance improvement over single-agent (though at 15x token cost). Enterprise adoption accelerates: KPMG survey shows 33% of organizations deployed agents, up from 11% in prior quarters. Simultaneously, research and practitioner evidence documents persistent technical barriers: failure attribution models achieve only 14.2% accuracy in pinpointing failure steps; Gartner forecasts 50% error rates in multi-agent systems; production failures documented (32% conversion drops in e-commerce). GitHub signals platform evolution with agentic workflow capabilities. Pattern emerges: rapid adoption momentum (infrastructure GA, framework maturation, vendor platform integration) coexists with unresolved technical fragility (failures, debugging gaps, token inefficiency), expanding deployment breadth while reliability concerns remain.
2025-Q3: Vendor platform expansion and concurrent risk signal convergence: AWS ships Strands Agents 1.0 with 2,000+ stars and multi-provider backing (Anthropic, Meta, OpenAI, Cohere, Mistral); Microsoft positions multi-agent systems as enterprise strategic imperative with architecture guides. Academic frameworks advance (Yale/Chicago/Oxford freephdlabor system for dynamic workflows). Deployment reports claim growth: 51% of teams in production (ZenML), millions of queries via Deutsche Telekom LMOS and Cognizant. Critical countervailing signals intensify: Gartner predicts 40% project cancellation by 2027; Carnegie Mellon benchmark shows 70% agent failure rate on standard tasks (Claude 3.7 Sonnet 26.3%, Gemini 2.5 Pro 30.3%, GPT-4o 8.6% success); HP infrastructure analysis documents 88% prototype failure cascade and unresolved cost/privacy/security roadblocks. Bifurcation sharpens: tooling maturity and platform integration accelerate while production viability signals worsen, suggesting the "adoption" metric reflects experiment breadth rather than production value realization.
2025-Q4: Framework maturation continued with LangGraph Platform confirming hundreds of production deployments and GitHub shipping Custom Agents for Copilot (October 2025). Analyst consensus hardened on adoption limits: Gartner reports less than 5% of enterprise applications deployed "real agents" by year-end (IntuitionLabs, November 2025). Deloitte warns 40% of agentic projects face abandonment by 2027. Production deployment evidence: JPMorgan Chase, NuvoBank, LinkedIn, and food manufacturing firms using LangGraph with human oversight; Cognizant and Deutsche Telekom processing millions of queries. Critical negative signals: Parallel AI documents $47,000 loss from coordination failures (November 2025); analysis attributes 95% deployment failures to architectural flaws in state management. Market forecasts project $35B autonomous agent market by 2030 yet production penetration remains confined to tightly scoped workflows. Orchestration complexity, failure recovery, and governance gaps persist as primary blockers to general-purpose enterprise adoption.
2026-Jan: Ecosystem maturity and failure mode documentation intensify in parallel. Large-scale analysis of 42K commits across 8 multi-agent systems (LangChain, CrewAI, AutoGen) reveals 40.8% perfective maintenance and 10% issues attributed to agent coordination. Production case: FRE|Nxt's InterviewLM with 8 specialized agents achieving 100+ concurrent sessions and 40% cost optimization. Adoption metrics show persistent gap: 66% experimenting, 38% in pilots, 14% ready to deploy, but only 11% live—with Gartner forecasting 40% project cancellation by 2027. Critical negative signals consolidate: 35% performance degradation in production systems from coordination overhead; MAST taxonomy documents 14 failure modes (41.8% specification, 36.9% misalignment, 21.3% verification); OWASP 2026 analysis identifies inter-agent trust and cascading failure vulnerabilities. Bifurcation persists: frameworks mature while deployment reliability and architectural robustness remain unresolved core challenges constraining viability to tightly scoped, heavily guarded workflows.
2026-Feb: Production infrastructure maturation coexists with persistent adoption-to-deployment gap. Dotzlaw case study demonstrates viable LangGraph scaling to 10k concurrent users with 60% cost savings. Durable execution emerges as critical infrastructure (Temporal $5B valuation). Ecosystem expands rapidly (97M MCP SDK downloads/month) but quality concerns intensify (average tool score 44.7/100). Adoption metrics reveal stalled enterprise transition: 11% in production vs 39% experimenting; Gartner forecasts 40% project cancellation by 2027. GitHub and practitioner analysis identify core failure patterns in orchestration and state management. Bifurcation sharpens: framework maturity and vendor backing accelerate while deployment reliability and operational overhead remain central blockers to enterprise-scale adoption.
2026-Mar: Vendor platform convergence confirmed: all major platforms (OpenAI Codex Subagents GA with manager-worker architecture, Dapr Agents v1.0 GA, Claude Code, GitHub, Devin, Grok) shipped multi-agent parallel execution by mid-March, and adoption metrics show high-adoption teams achieving 2.2x PR throughput (Jellyfish, 700+ companies). Concurrent failure evidence hardened: peer-reviewed study documents two-agent coordination accuracy drops from 58% to 25% under reduced specification detail (25-39pp gap independent of model capability); CooperBench confirms 50% lower collaboration success vs solo agents; Princeton research shows chained-agent reliability degrades to 74% combined even when components exceed 90%; a real 4-agent design system deployment documented 86% XSS vulnerabilities and $0.88–$146 cost variance per component. Enterprise adoption reality remains bifurcated: Deloitte confirms only 11% of companies use agents in production, with Klarna's structured LangGraph system (2.3M conversations/month, $60M savings) demonstrating that tightly scoped deployments with explicit orchestration topology deliver measurable ROI while general-purpose scaling remains constrained by coordination overhead and unresolved failure attribution.
2026-Apr: Organizational-scale production evidence accumulated: Stripe Minions (1000+ unattended PRs/week), Google Agent Smith (25%+ of production code), OpenAI Harness (~1M agent-written LOC), and 1inch's ticket-to-PR CI pipeline (implement + seven parallel review agents + synthesizer) documented as distinct deployment patterns; framework ecosystem consolidated around CrewAI (12M daily executions), LangGraph (Klarna/JPMorgan), and Microsoft Agent Framework 1.0 (AutoGen merger). Anthropic launched Claude Managed Agents public beta (April 8, 2026) as the first major vendor production-grade agent runtime with built-in orchestration, sandboxing, MCP integration, and state persistence. Enterprise orchestration evidence sharpened: Salesforce documented Alcon reaching 900+ agents in uncoordinated silos (security and governance crisis) versus RBC Advisor deploying 12+ specialized agents with an orchestrator-supervisor achieving 50% reduction in advisor prep time—illustrating that orchestration architecture, not agent count, determines enterprise viability; Gartner logged a 1445% inquiry surge for multi-agent topics while separate analysis shows 40% of multi-agent pilots fail in production. Production deployment economics reinforced coordination overhead as central viability constraint ($47k/month multi-agent vs $22.7k single-agent with only 2.1pp accuracy gain); Deloitte data confirmed 89% of multi-agent pilots fail at production deployment, with failures attributed to governance and integration maturity rather than technical limitations. Concurrent failure research hardened: peer-reviewed study shows single false claim spreads to all agents within three rounds across six frameworks (genealogy graph mitigation raises defense from 32% to 89%); structural limits synthesis documents 17.2x error amplification without coordination and 39-70% sequential reasoning degradation across all multi-agent variants. Research taxonomy established Wave 1 (viability) vs Wave 2 (measurement) framing, with key insight that single-agent systems with well-designed interfaces (10.7pp improvement from interface design alone in SWE-agent) often outperform multi-agent architectures for narrowly scoped tasks. Anthropic published five coordination patterns (generator-verifier, orchestrator-subagent, agent teams, message bus, shared state) with documented failure modes, establishing architectural standards for production deployment.
2026-May: Framework architecture selection matures with production decision guidance. GitLab published architectural decision record evaluating five orchestration frameworks (LangGraph, Temporal, Prefect, Claude Agent SDK, Haystack); LangGraph selected for native LLM token streaming, custom checkpoint backends, and human-in-the-loop support despite Temporal's superior durability model—signaling framework choice reflects production orchestration requirements, not capability parity. Real deployments demonstrate scaling patterns: Deriv operates 50+ agents with registry-based Operations Center architecture (Agent Officer pattern) solving integration tax challenges; Cursor built 1M-line browser in one week with three-role hierarchical orchestration (Planner/Worker/Judge), validating multi-agent development pipeline at enterprise code scale. Critical negative evidence intensifies: peer-reviewed study shows in-context prompting significantly outperforms LangGraph orchestration on procedural tasks (24% orchestration failures vs 11.5% in-context on travel, 9% vs 0.5% on Zoom support, 17% vs 5% on insurance); UC Berkeley MAST study analyzing 1,600+ production traces documents 41-86.7% failure rates across seven frameworks with structural prevention patterns (scope hierarchy, authority attenuation, typed protocols) as engineering requirement, not optional. Production infrastructure consolidation: StudioMeyer operates 40-agent fleet with three-layer observability (Sentry, Langfuse, LangGraph), demonstrating stateful workflows with Postgres checkpointing and resume-from-failure at scale. Ecosystem signal: Q2 2026 adoption metrics show 31% pilot-to-production conversion (2× Q1), MCP standardization (9.4k servers, +58% QoQ), and agentic funding ($20B of $42.6B total AI funding), indicating infrastructure maturity and market investment while structural reliability barriers remain unresolved at enterprise scale.