Perly Consulting │ Beck Eco

The State of Play

A living index of AI adoption across industries — where established practice meets the bleeding edge
UPDATED DAILY

The AI landscape doesn't move in one direction — it lurches. Some techniques leap from experiment to table stakes in a single quarter; others stall against regulatory walls, technical ceilings, or organisational inertia that no amount of hype can dislodge. Knowing which is which is the hard part. The State of Play cuts through the noise with a rigorously maintained index of AI techniques across every major business domain — classified by maturity, evidenced by real-world adoption, and updated daily so you always know where you stand relative to the field. Stop guessing. Start knowing.

The Daily Dispatch

A daily newsletter distilling the past two weeks of movement in a domain or two — delivered to your inbox while the index updates in the background.

AI Maturity by Domain

Each dot marks the weighted maturity of practices within a domain — hover for a brief summary, click for more detail

DOMAIN
BLEEDING EDGEESTABLISHED

Chat-based code assistance & debugging

LEADING EDGE

TRAJECTORY

Stalled

Conversational AI that answers coding questions, explains errors, and helps debug issues in a chat interface. Includes IDE chat panels, web-based coding assistants, and error explanation tools; distinct from inline autocomplete which operates without explicit prompting.

OVERVIEW

Chat-based code assistance reached full market maturity in May 2026 with a decisive market consolidation: Claude Code surpassed GitHub Copilot to become the dominant chat-based tool (46% "most loved" developer rating vs Copilot's 9%), capturing 63% developer preference within three months of enterprise launch. Yet the practice remains defined by the same structural tension: widespread adoption (85%+ regular use, 90% organizational exposure) paired with persistent distrust (29% confidence in AI code accuracy) and unresolved quality ceilings. New peer-reviewed security analysis (IOActive testing 27 models with 730 prompts; Veracode's 100+ LLM study) confirmed 45% OWASP Top 10 vulnerability rates and 59% baseline security performance -- metrics unchanged since 2024, suggesting systemic rather than scaling limitations. Real-world deployments showed scale but not maturation: Google 75% AI-generated code as of May 2026, Mercari 95% adoption with 64% output increase, yet CodeRabbit's repository analysis still documents 1.7x bug rates and disproportionately higher logic and security errors in AI-generated code. The emerging operational reality undercut vendor narratives: Claude Code's documented six-week regression (March 4 - April 20, 2026) degrading correctness by 15+ percentage points revealed production chat-based tools cannot yet serve as durable development infrastructure. For now, the practice delivers value in boilerplate, documentation, and routine tasks while remaining restricted from production-critical and security-sensitive paths due to unresolved quality and operational stability constraints.

CURRENT LANDSCAPE

Market consolidation accelerated in May 2026 with Claude Code surpassing GitHub Copilot as the leading chat-based assistant. Developer surveys (15,000+ across Pragmatic Engineer, JetBrains, Stack Overflow) show Claude Code at 46% "most loved" with 63% adoption growth in three months, vs Copilot's declining 9% satisfaction and Cursor's stable 18%. GitHub's Copilot family (Copilot Chat, Copilot Enterprise) remains at 20M+ users but market share concentrated in freemium tier; multi-model architecture (GPT-5.5, Claude Opus 4.7 backends) addresses capability but not quality ceiling. JetBrains ecosystem matured with AI-powered IDE integration across 11.4M recurring users and an enterprise analytics console. Real-world deployment data surfaced in May 2026: Google 75% AI-generated code (May); Stripe merging 1,300+ agent PRs/week; Mercari 95% adoption with 64% output increase; Fortune 1000 Cursor penetration at 70%. Yet security analysis quantified unresolved limitations: Veracode's study of 100+ LLMs on 80 curated tasks found 45% introduced OWASP Top 10 vulnerabilities; IOActive's peer-reviewed analysis (27 models, 730 prompts, 219 vulnerability categories) showed 59% baseline security performance. Operational fragility emerged as critical barrier: Anthropic's May 2026 postmortem documented Claude Code's six-week regression (March 4 - April 20) degrading code correctness by 15+ percentage points, demonstrating production chat-based tools cannot yet guarantee stable quality. Supply chain attacks accelerated: Q1 2026 incidents targeting Bitwarden, Lovable, and LiteLLM (malicious code injection into AI assistant context) revealed attackers systematically exploit development tools' legitimacy. Organizations face bounded deployment reality: full-stack agency case studies (TechVinta, Nipralo) document productivity gains (15-40% task speedup) but restricted to junior developers and routine tasks; senior and security-sensitive workflows remain human-led due to unresolved verification costs and quality constraints.

TIER HISTORY

ResearchNov-2022 → Nov-2022
Bleeding EdgeNov-2022 → Apr-2024
Leading EdgeApr-2024 → present

EVIDENCE (118)

— Postmortem documents Claude Code's six-week regression (March 4 - April 20, 2026) with specific capability degradation metrics; demonstrates operational fragility of production chat-based tools.

— First-party deployment data: Google 75% AI-generated code, Stripe 1,300+ agent PRs/week, Mercari 95% adoption with 64% output increase; Claude Code dominance (46% most-loved vs Copilot 9%).

— EASE 2026 peer-reviewed study of 1,000+ files across 100 GitHub repos reveals AI-code receives less frequent maintenance with divergent modification patterns vs human code.

— Veracode's 2025 study of 100+ LLMs on 80 coding tasks found 45% introduced OWASP Top 10 vulnerabilities; security performance unchanged despite model improvements.

— Market analysis showing Claude Code captured 63% developer preference with 46% 'most loved' rating; surpassed GitHub Copilot as leading chat-based coding assistant in 2026.

— Peer-reviewed analysis testing 27 AI models with 730 real-world prompts across 219 vulnerability categories; baseline 59% average security performance reveals systemic limitations.

— Q1 2026 incident report documenting three coordinated supply chain attacks on AI development infrastructure (Bitwarden, Lovable, LiteLLM); reveals new attack pattern: malicious code injection into AI assistant context.

— Full-stack agency's 18-month production case study across 50+ apps with three leading tools; documents specific productivity ranges and deployment recommendations by role.

HISTORY

  • 2022-H2: ChatGPT released to public (Nov 2022), sparking immediate developer experimentation for coding tasks. GitHub Copilot Chat emerged as IDE-integrated alternative. Academic research and independent practitioners documented adoption in learning and debugging workflows, but also identified critical limitations: variable code quality across languages, insecure code generation, and service reliability issues. Stack Overflow banned AI-generated answers due to low quality.

  • 2023-H1: Chat-based assistance transitioned to early mainstream adoption. GitHub announced Copilot X (March) with expanded conversational capabilities; VS Code shipped integrated Copilot Chat reaching 1M+ active users. Stack Overflow 2023 survey found 44% of developers actively using AI tools. Independent research showed 62-76% of developers report productivity gains, but trust deficits remained—only 42% trust accuracy. Service reliability and content filtering continued to cause user friction.

  • 2023-H2: Mainstream adoption solidified with GitHub announcing Copilot Chat GA and Copilot Enterprise (November); JetBrains released competing AI Assistant to GA across IDEs in December, expanding ecosystem beyond GitHub. Independent survey of 3,240 developers found highest adoption in Asia/Africa (80%+) and small companies. Real-world limitations surfaced: users reported quality degradation, content filtering false positives blocking legitimate code, and insufficient context for complex debugging. Practice remained in learning/prototyping phases; production deployment rare due to trust and operational barriers.

  • 2024-Q1: Chat-based assistance reached normalized platform scale. JetBrains reported 11.4M active users with AI Assistant GA across all IDEs; GitHub Copilot Chat reached GA for organizations and individuals (January). Large-scale independent survey of 100,000 workers found 50% adoption in exposed occupations; empirical analysis of GitHub PRs documented 580 shared ChatGPT conversations for code generation and debugging. Post-GA metrics showed strong user satisfaction (77% productivity gains, 3-8 hours/week time savings). However, critical quality research emerged: code churn increased from 3-4% pre-AI to 5.5% in 2023, with analysis suggesting AI-assisted development may degrade long-term code maintainability despite speed gains.

  • 2024-Q2: Real-world deployments confirmed productivity benefits but revealed persistent adoption barriers. DANA (Indonesian fintech) deployed Copilot to ~300 developers with 55% faster coding and 70% improved code understanding. Empirical evaluation on real projects documented 30-50% time savings in routine tasks, projecting 33-36% overall reduction. However, survey of 481 developers identified barriers: trust, insufficient project context, and company policies limiting use to test/documentation generation. Experienced developers showed no time gains and were reluctant to review large AI-generated code blocks. GitHub expanded Copilot Enterprise with context-aware features (PR summaries, discussion analysis), while testing expert critiques highlighted persistent hallucination and quality concerns undermining production confidence.

  • 2024-Q3: Chat-based assistance matured as platform standard but productivity expectations reset. GitHub expanded Copilot feature set (context window, chat IDE integration improvements) and Copilot Enterprise added support documentation awareness. Stack Overflow 2024 survey revealed hype correction: adoption continued but promised productivity gains disappointed relative to 2023 forecasts—developers reported improved "quality of time" rather than absolute speed savings. GitHub's broader enterprise survey (2,000 engineers) confirmed widespread multi-team interest but implementation barriers persisted. Comparative research (benchmarking ChatGPT vs. Codeium vs. Copilot) and industry analysis documented tool performance variations and inherent limitations. Developer experiences confirmed context challenges and over-reliance risks, positioning chat-based assistance firmly in supportive role rather than autonomous production capability.

  • 2024-Q4: Chat-based assistance expanded beyond conversation with GitHub Copilot Workspace technical preview (Dec) enabling agent-like autonomous task execution. Vendor investment continued with GitHub enhancements (VS Code model picker expansion) and JetBrains ecosystem maturation. However, Q4 brought concrete failure documentation: developer case studies and empirical analysis (Uplevel code study) reported zero productivity gains and introduced 41% more bugs, contradicting earlier vendor metrics. Security analysis surfaced production risks from generated code, reinforcing structural barriers to critical-path adoption despite technical maturity and widespread availability.

  • 2025-Q1: Productivity claims underwent reality correction as longitudinal research (800-developer study) documented increased code churn and developer trust collapsed (3% high-trust adoption down from 40% in 2024). Three GitHub Copilot Chat production incidents (Jan 29, March 18, 21) exposed service reliability constraints. Real-world studies show 80-90% adoption exposure but 66% of developers spending more time fixing AI code than saved; code reuse metrics declined through 2024. Enterprise adoption remains constrained by context limitations, data security concerns, and organizational governance, with deployment concentrated in documentation and test generation rather than production workflows.

  • 2025-Q2: Chat-based assistance reached pragmatic maturity with vendor ecosystem expansion (JetBrains free tier, GitHub context doubling) driving broader exposure to 82% daily/weekly use, but adoption-quality gap widened. New evidence shows 25% of AI suggestions contain hallucinations, Microsoft debugging benchmarks reveal low success rates (48% for Claude), and user satisfaction remains fractured despite productivity claims. Specialized tools like ChatDBG demonstrated higher-fidelity assistance (67-85% debugging success) through domain-specific implementation. Service reliability incidents persisted (April EU outage, June Free tier disruption), confirming operational constraints. Practice positioned as mature but bounded: strong in documentation/routine tasks, limited in production-critical and complex reasoning workflows.

  • 2025-Q3: Chat-based assistance stabilized at widespread but shallow adoption with developer trust collapsing to 29%. GitHub expanded Copilot Chat with autonomous repository operations (file/branch/PR management), signaling agentic evolution; JetBrains ecosystem matured but user ratings remained low (2.3/5). Independent research documented adoption-quality mismatch: Stack Overflow survey of 49,000+ developers found 80% exposure but 45% dealing with incorrect solutions, 66% spending more time fixing AI code than saved. Security research identified production risks (prompt injection, credential leakage). Specialized tools (ChatDBG, 75,000+ downloads) demonstrated higher-fidelity chat-based debugging (67-85% success), contrasting with general tools. Enterprise adoption remained concentrated in low-context tasks; production-critical work avoided due to governance, data security, and verification costs. Practice matured from aspirational to pragmatic: vendor investment continued, but realistic boundaries now evident—valuable for routine tasks and junior developers, limited for complex reasoning and senior workflows.

  • 2025-Q4: Chat-based assistance reached full market maturity with stabilized adoption metrics but persistent quality concerns. JetBrains survey of 24,534 developers confirmed 85% adoption globally; Jellyfish platform data showed growth from 49.2% (Jan) to 69% (Oct) organizational adoption. GitHub Copilot dominated with 20M+ users and expanded capabilities (model deprecations, feature additions). Final year-end data synthesized 2025 reality: 80% developer adoption but trust remained at 29%, with 45% routinely dealing with "almost-right" code and 66% spending more time fixing AI suggestions than saved. Analysis of the productivity paradox became explicit—measured studies (METR, CodeRabbit) showed 19% slowdown for experienced developers with 1.7x more bugs in AI-generated code, contradicting perceptual gains. Specialized tools demonstrated category leadership: GPT-5.1 debugging experiments reported 69%+ success rates for well-scoped bugs. Microsoft internal deployment (55% faster tasks) confirmed that disciplined organizational rollout with change management and context discipline achieves meaningful productivity, but broad-adoption deployments struggle due to insufficient context and verification overhead. Practice consolidated at leading-edge tier: vendors continue investment, adoption is ubiquitous in exposure, but production deployment remains bounded by governance, data security, and realistic quality-trust constraints. Senior and specialized-domain workflows favor selective or no AI assistance; boilerplate, documentation, and junior-developer support remain strongest use cases.

  • 2026-Jan: Chat-based code assistance expanded with vendor ecosystem maturation while quality evidence solidified concerns. GitHub released Copilot metrics dashboards with data residency for Enterprise Cloud (January 2026), signaling enterprise-grade adoption tracking; JetBrains integrated Codex as model option across IDEs (January 2026). Sonar survey (1,100 developers) documented 72% daily use but only 48% verification rate before commit, with 96% doubting AI code correctness. CodeRabbit analysis of 470 GitHub repos revealed AI produces 1.7x more bugs than humans, with 75% more logic errors and 1.5-2x higher security issues—quantifying quality degradation. Adoption reached 85% regular use (Zylos Research) but remained narrowly scoped: boilerplate, documentation, tests; production-critical code avoided due to demonstrated vulnerability patterns. Specialized chat-based debugging (ChatDBG, 75,000+ downloads) continued demonstrating higher-fidelity performance (67-85% bug fix rates). Practice remained bounded by fundamental quality-trust constraints despite ubiquitous exposure.

  • 2026-Feb: Chat-based assistance matured into vendor-standard feature with formalized enterprise governance. GitHub expanded Copilot metrics dashboards to include CLI telemetry (February 2026); JetBrains launched Console with AI management, analytics, and credit tracking across organizations. However, infrastructure fragility emerged as operational barrier: ChatGPT conversation loss case studies documented backend failures with ineffective support, while GitHub Copilot experienced 100% error rates due to OpenAI dependencies, confirming reliability constraints in production workflows. Developer adoption reached 92% daily use (US market) but concentrated in low-context tasks (boilerplate, documentation) with 41% of global code now AI-generated; 45% of AI-generated code failed security tests, maintaining quality-adoption mismatch. METR research corrected prior claims of productivity slowdown with selection-effect analysis. Practice consolidated at leading-edge tier: vendor infrastructure and governance capabilities matured, but fundamental quality and reliability constraints remained—positioning chat-based assistance as mature but bounded tool for supplementary workflows rather than autonomous production development.

  • 2026-Mar: Chat-based assistance solidified as category leader in adoption metrics but entered final reality-check phase on productivity claims. Large-scale studies (Jellyfish 700 companies, 200k engineers) confirmed 64% of teams now generate majority of code using AI in production—up from 49.2% in Jan 2026. However, peer-reviewed research (MSR '26 on Cursor, ICLR 2026 benchmarking) quantified hard ceiling on quality: AI-assisted development shows measurable short-term velocity gains (+2-8%) offset by persistent long-term code complexity and quality degradation (1.7x bug rates, 75% accuracy on structured outputs). METR's controlled trial with 16 experienced developers on 246 real issues documented the productivity paradox in full: 19% actual slowdown vs 20% perceived speedup (39-point gap between experience and perception). Security threat landscape expanded: documented CVEs (CVE-2025-53773 RCE in Copilot via malicious comments, CVE-2025-59536 API key exfiltration, MCP supply chain compromises) show attackers systematically exploiting tool legitimacy to bypass traditional security controls. Despite ubiquity (90% adoption rate, 20M Copilot users, 51% daily use), the practice remains constrained by verification-cost bottleneck: 62% of AI-generated code contains design flaws or vulnerabilities, yet most organizations lack governance structures to enforce review. Practice status: mature leading-edge platform feature with well-documented limitations; adoption-quality gap shows no sign of closing. Realistic deployment window continues narrowing to boilerplate, documentation, and junior-developer support. Senior developers, production-critical code, and security-sensitive paths remain human-led.

  • 2026-Apr: Chat-based assistance confirmed final market consolidation around highest-performing tools, with critical evidence reinforcing adoption-quality paradox. Market shift accelerated: Claude Code surpassed GitHub Copilot (41% vs 38% developer adoption) with vastly superior sentiment (46% "most loved" vs 9% for Copilot), signaling that capability and developer experience now outweigh ecosystem lock-in. Adoption metrics reached inflection: 84-85% regular use but only 29% trust accuracy, 3.1% high-confidence adoption—widening the perception-reality gap to 39 percentage points as METR data showed experienced developers 19% slower with AI. New peer-reviewed evidence (ICLR 2026) quantified structural limits: models achieve 75% accuracy on structured outputs (the 1-in-4 error rate affects all chat-based debugging), and large-scale behavioral study (11,579 real IDE sessions) showed conversational programming operates as "progressive specification"—iterative refinement rather than direct specification—exposing verification overhead as the bottleneck preventing productivity gains. Security risks escalated: CamoLeak vulnerability (CVE-2025-59145, CVSS 9.6) demonstrated silent code/credential exfiltration via prompt injection, with systemwide architectural pattern analysis revealing three dominant attack vectors (config-as-execution, localhost trust assumptions, untrusted input with privilege). Production deployments continued revealing hidden costs: real organizational case study documented 18% incident increase, $85K downtime failure, and 4-6 hours/week review overhead, with maintenance costs hitting 4x baseline by year two at >40% AI code share. Critical assessment emerged on optimization misdirection: industry optimized for speed (5% of dev time) while missing 95% of actual bottleneck (understanding/maintenance), suggesting the practice has matured to acknowledge its own limited scope. Specialized domain-specific tools (ChatDBG, 75k+ downloads) maintained 67-85% bug-fix rates vs 48% for general models on debugging—confirming that narrower scope achieves higher fidelity. Latest April data (April 14-28) reinforces maturity: JetBrains 10,000+ developer survey shows 90% AI tool adoption with Copilot 29%, Claude 18%, Cursor 18% market share; GitHub expanded Copilot Chat debugging on web with structured root-cause analysis; Cursor business analysis shows evolution to agent-first interface, reaching $2B ARR and 70% Fortune 1000 penetration; Microsoft peer-reviewed research documents 39% performance degradation in multi-turn conversations, core limitation of chat-based workflows. Most critically, AMD production deployment exposed Claude Code quality collapse (read-to-edit ratio 70% drop, accuracy 83.3% to 68.3%, $12 to $1,504/day API costs), and Fortune coverage of Anthropic's admission of engineering missteps documents market backlash. Practice consolidated at leading-edge maturity with realistic boundaries: mainstream adoption (90% exposure, 85% regular use) confirmed, but trust-adoption gap, quality ceiling (1.7x bugs, 75% accuracy), multi-turn conversation degradation (39% performance drop), and verification-cost bottleneck locked deployment to boilerplate/junior-dev support. No evidence of tier advancement pathway; mature but bounded.

  • 2026-May: Market consolidation crystallized around Claude Code's dominance, while peer-reviewed evidence confirmed persistent quality and security ceilings. First-party deployment data (Google 75% AI-generated code, Stripe 1,300+ agent PRs/week, Mercari 95% adoption with 64% output increase) demonstrates scale is real; developer sentiment surveys (15,000+) show Claude Code at 46% most-loved vs Copilot's 9%, confirming the market has shifted decisively toward highest-capability tools. However, Anthropic's own postmortem documented Claude Code's six-week regression (March 4 - April 20, 2026) degrading correctness by 15+ percentage points, directly evidencing that production chat-based tools cannot yet guarantee stable quality. Peer-reviewed security research (IOActive: 27 models, 730 prompts, 219 categories; Veracode: 100+ LLMs on 80 tasks) confirmed 59% baseline security performance and 45% OWASP Top 10 vulnerability rates—metrics unchanged from 2024, indicating systemic rather than scaling limitations. EASE 2026 study (1,000+ files, 100 GitHub repos) found AI-generated code receives less frequent maintenance with divergent modification patterns, suggesting emerging long-term maintenance liability. Supply chain attacks on AI development tools (Q1 2026: Bitwarden, Lovable, LiteLLM) established a new threat class: malicious code injection into AI assistant context. Practice remains bounded at leading-edge: scale and market consolidation confirmed, but quality ceiling (unchanged security metrics), operational fragility (documented regressions), and supply chain risks constrain deployment to boilerplate and junior-developer support.

TOOLS