Perly Consulting │ Beck Eco

The State of Play

A living index of AI adoption across industries — where established practice meets the bleeding edge
UPDATED DAILY

The AI landscape doesn't move in one direction — it lurches. Some techniques leap from experiment to table stakes in a single quarter; others stall against regulatory walls, technical ceilings, or organisational inertia that no amount of hype can dislodge. Knowing which is which is the hard part. The State of Play cuts through the noise with a rigorously maintained index of AI techniques across every major business domain — classified by maturity, evidenced by real-world adoption, and updated daily so you always know where you stand relative to the field. Stop guessing. Start knowing.

The Daily Dispatch

A daily newsletter distilling the past two weeks of movement in a domain or two — delivered to your inbox while the index updates in the background.

AI Maturity by Domain

Each dot marks the weighted maturity of practices within a domain — hover for a brief summary, click for more detail

DOMAIN
BLEEDING EDGEESTABLISHED

Customer support chatbots — LLM-powered conversational

BLEEDING EDGE

TRAJECTORY

Stalled

Large language model-powered chatbots that handle customer queries with natural conversation and contextual understanding. Includes RAG-based support bots and multi-turn conversation handling; distinct from autonomous resolution which takes actions rather than just conversing.

OVERVIEW

LLM-powered conversational chatbots occupy a persistent gap between vendor capability and production reliability. Since GPT-4 enabled the category in early 2023, platforms like Intercom, Zendesk, and Vonage have shipped RAG-grounded bots that resolve 50-70% of support tickets in controlled deployments. Organisational enthusiasm is strong: roughly two-thirds of enterprises report active adoption, and ROI figures of 148-200% within twelve months circulate widely. Yet the practice remains experimental. Hallucination rates on grounded tasks have fallen to 0.7-1.5%, but complex reasoning errors have worsened to 33-51% in recent benchmarks. High-profile failures — NEDA's bot dispensing harmful eating-disorder advice, a Chevrolet bot discounting a vehicle to one dollar, DPD's chatbot swearing at customers — illustrate governance risks that technical progress has not resolved. Consumer trust trails organisational confidence by a wide margin; a 2024 survey found only 50% of consumers positive about AI interactions versus 91% of business leaders. Three years into the category's life, the core tension is unchanged: vendors can demonstrate impressive metrics in scoped deployments, but reliability, bias, and governance barriers keep LLM-powered chatbots firmly in pilot territory for most organisations.

CURRENT LANDSCAPE

Vendor platforms continue consolidating around LLM-powered resolution. Zendesk's April 2026 announcement removing AI tier distinctions and democratizing agentic capabilities across all plans signals industry confidence in scaling—rollout begins April 27 with support ending for legacy tiers by August 2026. Intercom's Fin maintains 67% average resolution across 7,000+ customers, with internal deployment achieving 81% automation while absorbing 300% growth in customer inquiries without proportional headcount expansion, delivering $7.5-9M in annual cost savings. Scale milestones confirm production viability: 40M+ conversations resolved, 66% average rate, with realistic maturity data showing support teams improve from initial 41% to 51% through optimization cycles. Named customers demonstrate consistent outcomes—tado° at 90-95% CSAT with 70% workflow automation, Nuuly 95% CSAT, OPPO 83% resolution with 57% repurchase uplift—confirming the category's production deployment stage for well-scoped use cases.

Yet deployment reality reveals fundamental constraints unsolved by technical progress. A 2025 Gartner study cited by LoopReply found 67% of chatbot projects failed to meet expectations, with seven documented failure modes rooted not in technology but in implementation (ghost-town knowledge bases, poor escalation, wrong metrics, rigid flows, insufficient testing, platform misfit, treating chatbots as projects rather than ongoing products). Industry benchmarks remain modest: Comm100's 220M-interaction dataset shows 44.8% average resolution with significant variation (38-98%), and importantly, finds that high resolution rates do not correlate with satisfaction—proper human handoff quality matters more than deflection percentages. Real-world analysis from eesel suggests current average performance sits at 30-40%, with only 10% of teams achieving 50-60% maturity, exposing a wide gap between vendor aspirations (70-80%+) and organizational reality.

The failures continue at scale. In 2024, 39% of AI chatbot deployments were pulled back or reworked due to errors. Peer-reviewed research confirmed fundamental limits: LLM-generated content biases customer decisions 32% more than original reviews with 60% hallucination on out-of-training queries; human-AI collaboration studies show high-quality bot suggestions improve worker accuracy by 27 points but hit an "underreliance plateau" where improvements level off. ECRI ranks AI chatbot misuse as the number-one health technology hazard for 2026 (40M daily unvalidated queries). Deployment economics remain boundary-dependent: Fin effective for support deflection at $0.99/resolution but creates cost traps in revenue-generation contexts.

This remains classic bleeding-edge territory. Vendor platforms have crossed into production readiness and deliver measurable value for well-defined support deflection use cases, supported by robust internal deployments and adoption momentum (68% enterprise adoption rate). But the practice is constrained by hallucination in complex reasoning, governance frameworks lagging deployment pace, implementation barriers that technology alone cannot solve, and consumer trust—now at 61% comfort but masking persistent accuracy, bias, and human-fallback concerns. For organisations, these tools remain strategic bets, not routine procurements.

TIER HISTORY

ResearchMar-2023 → Mar-2023
Bleeding EdgeMar-2023 → present

EVIDENCE (90)

— Zendesk CEO reveals 70-80% autonomous resolution on simple/medium issues with white-box reasoning, outcome-based pricing accountability, and production operational requirements for scalable LLM chatbot deployment.

— Peer-reviewed CMR study documents 64% customer preference against AI and 53-77% negative experience rates, revealing persistent adoption barriers despite vendor capability gains and cost-per-interaction benefits.

— Hiver survey (700+ leaders): 90% uncomfortable with AI representing brand directly; documents critical trust gap between adoption metrics and customer/agent confidence—structural barrier limiting productivity gains realization.

— TIMEWELL analysis documents 40M+ Intercom Fin resolutions at 67%, Klarna at 2.3M/month with 82% faster resolution, showing vendor platforms and integration patterns delivering production-scale LLM chatbot performance.

— MIT NANDA analysis: 95% of AI pilots deliver no impact, 50% of projects abandoned after PoC; root causes are data infrastructure, governance, and operational integration—not technology or skills, highlighting implementation constraints.

— Digital Applied compilation (150+ data sources): 41.2% median deflection, 90%+ lower cost-per-resolution, 27% production deployment of agentic AI—balanced signal combining productivity gains with hallucination governance risks.

— Deloitte analysis: 83% of CX leaders see memory-rich AI as essential; enterprise adoption signals strong with 82% invested in AI (though only 10% at mature deployment), indicating organizational confidence despite implementation barriers.

— Forrester Wave (Q2 2026) analyst report evaluating 14 conversational AI platforms for customer service. Tier 1 signal of platform maturity, agentic capability adoption, enterprise integration challenges, and data security constraints.

HISTORY

  • 2023-H1: Major platforms (Intercom, Zendesk) shipped GPT-4-powered conversational bots with RAG safeguards. Gartner survey showed low customer adoption (8% usage, 25% repeat intent) despite positive perception for simple cases. Hallucination risks and need for human oversight identified as primary deployment barriers.

  • 2023-H2: Enterprise adoption momentum accelerated (92% of customer support teams planned or deployed chatbots), and vendor metrics improved (Intercom: 59% resolution, 50% instant). However, real-world gaps widened: deployment barriers emerged (cold responses, integration complexity, privacy concerns), documented production failures (Pak'nSave recipe hazards, legal hallucinations) exposed governance risks, and hallucination remained unresolved despite 1.5-2 year remediation timelines from industry leaders. Even major vendors (Google Gemini) faced maturity challenges with non-English reliability. Market exhibited classic bleeding-edge pattern: strong organizational interest, vendor availability, but constrained by technical limitations and modest customer adoption.

  • 2024-Q1: Enterprise deployment accelerated with real-world case studies (Frends: 59% resolution, 52.6% independent handling), and vendor progress on hallucination via structured reasoning (Vonage: 23.7% → 1.0% error rates). Zendesk reported 70% of CX leaders reimagining journeys with GenAI, 83% claiming positive ROI; Botco survey found 76% of contact centers actively using chatbots. However, critical adoption barriers persisted: a significant "AI Gap" emerged (91% of leaders vs. 50% of consumers positive about AI interactions), real-world failures documented (McDonald's AI drive-thru test ended after customer complaints), and customer willingness to use chatbots remained low (8% usage, 25% repeat intent unchanged from 2023). Organizational deployment faced governance, privacy, and user experience challenges that vendor technical improvements had not yet resolved.

  • 2024-Q2: Vendor platforms continued shipping improvements (Intercom's Fin AI Copilot boosting agent efficiency 31%, Zendesk expanding AI across retail and CX domains) and enterprise adoption momentum persisted. Yet governance failures crystallized: NYC's AI chatbot advisory system remained active despite documented evidence it was advising illegal business practices, exposing gaps between production deployment and risk mitigation. Peer-reviewed research confirmed hallucination remained a fundamental property of LLM-based systems even as vendors claimed technical progress. Customer adoption and willingness stayed stagnant (8% usage, 25% repeat intent), while organizational enthusiasm for GenAI deployment continued. The category remained characterized by strong vendor investment and adoption intent coupled with persistent customer trust deficits and unresolved governance challenges.

  • 2024-Q3: Vendor platforms shipped incremental improvements: Intercom expanded Fin AI to 45 languages in GA, and Gartner analysis reframed GenAI as redefining traditional conversational AI ROI expectations. Enterprise adoption signals remained strong (74% of companies implementing chatbots, 89% rating chatbots as most useful AI application). Real-world case studies demonstrated value (Vagaro resolved 44% of incoming requests, reduced handling time from 3h to 23min, improved CSAT 87%→92%). However, adoption barriers remained persistent and documented (job loss fears, data security concerns, integration complexity, skepticism about effectiveness). Governance failures continued: LAUSD shut down its 'Ed' chatbot after five months of deployment due to documented failures, exemplifying risks in regulated environments. Technical capability and organizational adoption intent coexisted with constrained customer trust and unresolved governance challenges—the category remained in bleeding-edge territory with technology ahead of organizational readiness.

  • 2024-Q4: Vendor platforms shipped measurable product improvements: Intercom's Fin 2 (powered by Claude) achieved 51% average resolution rate across thousands of customers, up from 23% for Fin 1. Large-scale deployments demonstrated business impact: Vodafone's TOBi resolved 70% of inquiries and cut cost-per-chat by 70%. Enterprise adoption momentum persisted. However, customer satisfaction deficits widened: Kapture CX survey found 43% of shoppers frustrated with chatbot ineffectiveness, and practitioner forums revealed mixed results (CSAT ~50% on Fin deployments). The core tension remained: product maturity and organizational deployment scaled, yet end-user trust and satisfaction stayed constrained, exemplifying classic bleeding-edge constraints where technical capability outpaced customer adoption willingness.

  • 2025-Q1: Organizational adoption broadened: CMSWire survey showed 51% of CX leaders deployed chatbots with speed and cost as primary drivers. Vendor platforms continued feature expansion (Fin 41% resolution, 20+ new capabilities; fintech deployments achieved 50-90% automation in Sharesies and Fundrise). However, production risks and governance failures accelerated: Fortune 500 retailer experienced $2.3M loss from chatbot hallucination before mitigation (58% escalation reduction); Air Canada held liable for legal damages from chatbot misinformation; Vrije Universiteit research documented LLM service failures as structural reliability concern. Data privacy emerged as largest adoption barrier (32% of leaders). The category's core tension deepened: organizational deployment and vendor investment continued despite documented production failures, legal liability precedents, and unresolved governance challenges.

  • 2025-Q2: Vendor platforms shipped incremental product improvements (Intercom, Zendesk) with expanded technical documentation on RAG architecture and hallucination mitigation. However, academic research confirmed fundamental reliability constraints: Phare benchmark showed 20% accuracy drops in critical tasks; meta-analysis documented 73% of scientific summaries containing exaggerations. Production failures accelerated: 34-hour ChatGPT outage disrupted customer service operations globally; OpenAI rolled back ChatGPT update due to excessive politeness requiring guardrails refinements. Deployment variability persisted: 35% of AI customer service projects never break even vs. 30% cost reduction and 70% containment for successful implementations. Organizational adoption continued (51% deployment rate), yet structural reliability gaps and operational SLA risks remained unresolved, exemplifying bleeding-edge immaturity where vendor capability claims diverge from real-world production stability.

  • 2025-Q3: Vendor platforms shipped incremental product improvements: Zendesk's internal AI deployment handled 60K+ requests per quarter with 120% improvement in response quality; Intercom research focused on production feedback classification and agent optimization. Organizational adoption continued (51% deployment rate maintained). However, critical constraints emerged: Qualtrics Q3 consumer survey (20K+ global respondents) showed 20% of AI customer service users saw no benefit—a 4x higher failure rate than other AI applications—with rising data privacy and human-exclusion concerns. Forrester analysis revealed systemic barriers: fragmented tech stacks, outdated systems, and metric misalignment trap customers in deflection loops rather than solving problems; current AI adoption mostly confined to efficiency gains rather than self-service transformation. Deployment quality remained inconsistent, with common patterns around hallucinations, escalation failures, knowledge-base decay, and compliance risks limiting real-world success. The category exhibited acute bleeding-edge tension: vendor capability and organizational investment accelerated while consumer satisfaction, customer willingness to engage, and reliable deployment outcomes remained fundamentally constrained by unresolved technical limitations and governance gaps.

  • 2025-Q4: Vendor platforms continued shipping production improvements: Zendesk demonstrated customer success (Unity: $1.3M savings, 8,000 ticket deflections, 80% automation); Intercom Fin maintained 60% resolution across hundreds of thousands of deployments. Industry adoption continued (95% of interactions expected to involve AI by year-end). However, critical liability and reliability risks materialized: Air Canada chatbot hallucination resulted in legal damages and established organizational liability precedent; Finova Bank required 89% hallucination reduction through complex RAG/validation layers. Ecosystem stability emerged as operational risk: ChatGPT December outage (30+ min) disrupted customer service operations globally. Consumer trust remained constrained (42% ethical AI confidence), indicating that vendor GA maturity and organizational deployment momentum coexisted with unresolved technical, governance, and consumer perception constraints.

  • 2026-Jan: Continued strong organizational momentum with Congruence MI projecting $6.2B market by 2032 (22.6% CAGR) and 68% enterprise adoption. Named vendor deployments at scale: Klarna handling equivalent of 700 human agents; Intercom Fin maintaining 60% resolution across hundreds of thousands; specific verticals (fintech, smart home) showing strong results. However, hallucination constraints hardened: grounded tasks improved to 0.7-1.5% but complex reasoning worsened to 33-51% error rates, with ECRI ranking chatbot misuse as #1 health technology hazard (40M daily ChatGPT users for unvalidated health information). Deployment cost boundaries clarified: Fin effective for support deflection (50%+) but $0.99/resolution creates cost traps for revenue use cases. Consumer comfort improved to 61% but masked persistent trust gaps. Bleeding-edge pattern sustained with adoption momentum coexisting with hardening technical and governance constraints.

  • 2026-Jan: Market growth acceleration continued with Congruence MI projecting $6.2B+ market by 2032 (22.6% CAGR) and 68% enterprise adoption. Vendor deployments at scale (Klarna 700-person equivalent workload, Intercom Fin 60% resolution), with specific-vertical success (fintech) but infrastructure barriers (Nigerian retailer case). Hallucination constraints hardened rather than resolved: grounded tasks improved to 0.7-1.5% but complex reasoning worsened to 33-51%; ECRI ranked chatbot misuse as #1 health tech risk. Deployment boundaries clarified: Fin AI effective for support (50%+ deflection) but limited for revenue use cases at $0.99/resolution. Consumer comfort improved to 61% but masked persistent trust gaps. Bleeding-edge pattern sustained: adoption momentum and vendor investment coexisting with hardening cost/capability boundaries and governance risks in regulated verticals.

  • 2026-Feb: Vendor platforms shipped incremental governance and visibility improvements (Zendesk's AI agent conversations feature GA, Intercom expanded reporting metrics). Named deployments maintained at scale: tado° achieving 90-95% CSAT with 70% workflow automation; Nuuly 95% CSAT; Lightspeed 72% resolution across production. Industry ROI metrics remained strong: 148-200% ROI within 12 months, up to 95% interaction handling potential, 84% of businesses reporting faster resolution, $3.50-4.13 per-dollar savings. However, deployment risks and failure rates hardened despite positive headlines: 39% of deployments were pulled back or reworked in 2024; specific production failures documented (NEDA harmful advice, Chevrolet deep discounting bot, DPD brand-damaging swearing). Research confirmed fundamental limitations: LLM-generated content biases customer decisions 32% more than original content (26.5% sentiment manipulation, 60% hallucination on out-of-training queries). Organizational adoption continued while production reliability constraints and user preference for human interaction remained structural barriers. Bleeding-edge category exhibited acute tension: vendor capability and org adoption momentum coexisting with documented deployment failures and persistent trust deficits.

  • 2026-Apr: Vendor platform consolidation accelerated: Zendesk announced major expansion (April 2, 2026), removing AI tier distinctions and unlocking agentic capabilities (reasoning, multi-step procedures, API integration) in base plans, with rollout April 27-May 18 and support ending for legacy AI tiers by August 31. Intercom Fin scaled milestone data: 40M+ conversations resolved at 66% average, improving to 67% on Zendesk integration, with trajectory showing teams improve from 41% initial to 51% optimized through continuous learning. Internal case study: Intercom's three-year Fin deployment achieved 81% automation while absorbing 300%+ customer demand growth without proportional headcount increase, delivering $7.5-9M annual cost savings. Deployment scale reinforced across multiple sources: TIMEWELL consulting documented Klarna at 2.3M conversations/month with 82% faster resolution (11min→2min) and Lightspeed at 65% end-to-end resolution; Deloitte analysis confirmed 82% of leaders invested in AI (though only 10% achieved mature deployment). Peer-reviewed research (arXiv March 2026) confirmed human-LLM collaboration dynamics: high-quality bot suggestions improve worker accuracy by 27 points but hit diminishing returns plateau. Deployment reality check: Gartner 2025 study cited by LoopReply found 67% of chatbot projects failed to meet expectations due to implementation issues (knowledge base quality, escalation design, metrics misalignment); Comm100 benchmark (220M conversations) showed 44.8% average resolution with finding that high resolution rates don't correlate with satisfaction; eesel analysis quantified realistic baseline at 30-40% current performance vs 70-80% vendor targets; Digital Applied compilation found 41.2% median deflection with 27% in full production. Fundamental adoption barrier crystallized: Hiver survey (700+ leaders) found 90% uncomfortable with AI representing brand directly; Berkeley CMR research documented 64% customer preference against AI and 53-77% reporting negative experiences despite business cost savings of ~$0.70/interaction. Implementation failures root cause identified: MIT NANDA analysis showed 95% of AI pilots deliver no measurable impact with root causes in data infrastructure, governance, and operational integration—not technology skill gaps. Named organization (DoorDash) built LLM conversation simulator reducing hallucination by ~90% before deployment, documenting production-grade testing methodology. Named deployment (OPPO) achieved 83% chatbot resolution, 94% positive feedback, 57% repurchase increase on large-scale seasonal operation. Organizational adoption continued (68% enterprise rate), infrastructure consolidation signaling vendor confidence, but implementation, trust, and reliability constraints remained unresolved. Category remained in acute bleeding-edge tension: platforms achieving production-scale resolution on well-scoped deflection use cases coexisting with documented 67% failure rate on organizational implementations, critical customer-organization trust gaps, and persistent barriers to broader adoption beyond pilot and optimization phases.

  • 2026-May: Consumer trust barriers sharpened as the dominant constraint. Berkeley CMR peer-reviewed research confirmed 64% customer preference against AI chatbots and 53-77% negative experience rates; Hiver survey (700+ leaders) found 90% uncomfortable with AI representing their brand directly. Zendesk CEO publicly confirmed 70-80% autonomous resolution on simple and medium issues but acknowledged outcome-based pricing accountability as a prerequisite for responsible scaling. MIT NANDA analysis documented that 95% of AI pilots deliver no measurable impact, with root causes in data infrastructure and governance rather than technology gaps. Despite these structural barriers, vendor platform deployment continued at scale—Intercom Fin at 40M+ resolved conversations—and Deloitte confirmed 82% of CX leaders invested in LLM chatbots, with the 10% mature deployment rate signalling that breadth of adoption has substantially outrun depth of execution.