Perly Consulting │ Beck Eco

The State of Play

A living index of AI adoption across industries — where established practice meets the bleeding edge
UPDATED DAILY

The AI landscape doesn't move in one direction — it lurches. Some techniques leap from experiment to table stakes in a single quarter; others stall against regulatory walls, technical ceilings, or organisational inertia that no amount of hype can dislodge. Knowing which is which is the hard part. The State of Play cuts through the noise with a rigorously maintained index of AI techniques across every major business domain — classified by maturity, evidenced by real-world adoption, and updated daily so you always know where you stand relative to the field. Stop guessing. Start knowing.

The Daily Dispatch

A daily newsletter distilling the past two weeks of movement in a domain or two — delivered to your inbox while the index updates in the background.

AI Maturity by Domain

Each dot marks the weighted maturity of practices within a domain — hover for a brief summary, click for more detail

DOMAIN
BLEEDING EDGEESTABLISHED

Agent assist — auto-draft with human review

GOOD PRACTICE

AI that automatically drafts full responses for agents to review, edit, and send during customer interactions. Includes tone-matched response generation and policy-aware drafting; distinct from response suggestion which offers options rather than complete drafts.

OVERVIEW

Auto-draft with human review has become the proven pattern for AI in customer support. The approach -- AI generates a full, tone-matched response draft; the agent edits and sends it -- is now a GA feature across tier-1 platforms, with documented ROI at enterprise scale. The question for most organisations is how to roll it out effectively, not whether it works.

What makes auto-draft durable is what it chose not to automate. Fully autonomous AI agents face high failure rates and mounting governance concerns; auto-draft sidesteps these by keeping the human in the approval loop. That architectural choice, once seen as a concession, has proven to be the practice's competitive advantage. Deployments that preserve agent judgment deliver measurable gains in handle time, resolution rate, and satisfaction. Those that skip the review gate face spiralling incident rates and stalled scaling.

CURRENT LANDSCAPE

Zendesk and Intercom ship auto-draft as standard platform infrastructure, not add-on pilots. Zendesk expanded its AI writing tools (tone controls, expand, simplify) to Professional-tier plans in early 2026, while its AI-generated procedure drafts give agents review-ready content at scale. March 2026 releases introduced auto-assist event logging for full audit trails and pre-approved action workflows, maturation signals that governance layers are now native to the product, not bolted-on. Intercom publishes a formal automation-rate KPI for its Fin agent, treating draft-and-review throughput as a production metric.

Deployment is approaching mainstream at the 55-60% adoption mark. Metrigy research across 656 companies (August-September 2025) shows 55% have deployed agent assist, 39% planning or evaluating (94% combined engagement), with two-thirds reporting improvements in agent quality and 59% noting increased sales. Real-world ROI evidence strengthens the case where implementation is mature. Nucleus Research documents measurable impact across 30+ Zendesk customers in production, showing resolution performance and effort reduction gains with human-supervised workflows at scale. Google's Agent Assist deployment guide documents 10-15% handle-time reductions; named enterprise customers report results ranging from 64% email automation with CSAT gains to 92% faster resolution.

However, deployment-execution gaps widen at scale. Intercom's 2026 survey of 2,400 service professionals found 82% had invested in AI for customer service, yet only 10% rated their deployment as mature—and the gap in quality outcomes between mature and early-stage teams was stark (87% vs. 43% reporting improvements). Qualtrics research across 20,000+ consumers in 14 countries reveals AI customer service fails at 4x the rate of other AI applications, with context loss and hallucination being common failure modes. This disparity validates auto-draft's human-review gate: Gravitee's survey of 900+ executives shows 81% deployed AI agents but only 14% had full security approval; 88% reported incidents. The practice's governance advantage is structural—human review prevents the silent failures that plague autonomous systems. The scaling challenge is now operational: change management, agent training, workflow integration, and governance rigor, not proof of concept.

TIER HISTORY

ResearchJun-2023 → Jul-2023
Bleeding EdgeJul-2023 → Oct-2024
Leading EdgeOct-2024 → Jan-2025
Good PracticeJan-2025 → present

EVIDENCE (63)

— Agent assist ROI framework quantifies direct cost savings (20% AHT reduction), quality improvements, and throughput gains; positioned as proven ROI category.

— Customer service automation leads with 620% average ROI within 18 months; AI agents handling Tier 1/2 resolve 78% without escalation in production.

— Survey of 700+ leaders: 90% uncomfortable with AI representing brand directly to customers, validating human review gate as essential control for adoption.

— Hybrid human-AI escalation model achieves 4.25/5 CSAT (vs 4.1 pure-AI), narrowing gap to human-only 4.3 by just 0.05 points; validates auto-draft architecture.

— Sinch deployment: AI Copilot for agents combined with autonomous agents achieved 47% faster resolution and doubled self-service automation from 17% to 32%.

AI AutoresponderProduct Launches

— LiveAgent's 'draft & approve' mode demonstrates standard GA implementation: AI generates response as private note, agent reviews and edits before sending.

— Customer service agents achieve 9x cost reduction per task, 8.7 hours saved weekly, 4.2x productivity multiplier, 4.1-month payback period across deployments.

— 40-person SaaS team expected 60% automation but achieved 23% after six months; knowledge-base indexing limits and intent classification gaps identified as root causes, illustrating steep configuration and tuning cliff in auto-draft deployments.

HISTORY

  • 2023-H2: SupportLogic, Maven, and Macha released production auto-draft features with agent review workflows. Evidence shows response generation with tone control and editable draft modes gaining traction in vendor roadmaps; concurrent critical coverage highlights risks of AI implementation without proper human involvement.
  • 2024-Q1: Auto-draft moved into mainstream platform adoption. Zendesk and Microsoft shipped GA auto-draft tools; Gartner reported 94% of customer service leaders exploring GenAI copilots for agent assist. Vendor implementations converge on draft-review workflows, but satisfaction gaps remain (80% see value, 41% satisfied); risks around hallucination and prompt injection documented.
  • 2024-Q2: Zendesk positioned Agent copilot as core platform feature with proactive guidance; Intercom launched Fin copilot for conversational response generation. Auto-draft consolidates as mainstream capability in tier-1 platforms, moving from experimental to standard agent-assist offering in major contact center stacks.
  • 2024-Q3: Zendesk and Intercom released GA auto-draft products with confirmed human review workflows; independent roundtable coverage confirms agent assist as central investment area across ecosystem (Avaya, AWS, Genesys, NICE, Talkdesk, Zoom). Telecom deployments show 25-90% improvements in agent productivity and troubleshooting. Adoption surveys show 80% positive sentiment on AI's impact, though critical assessments highlight persistent ROI verification challenges and low satisfaction gaps in real deployments.
  • 2024-Q4: Production auto-draft deployments accelerated across tier-1 platforms. Zendesk and Intercom reported specific customer outcomes: email automation (64% volume, 10-point CSAT lift), agent productivity multipliers (3x ticket throughput), and resolution rate benchmarks (51-65% autonomous resolution). Broad enterprise adoption (68%) contrasted with ROI realization challenges (32% see significant ROI), revealing adoption-execution gap. Ecosystem consolidation evident: Genesys deprecated earlier Agent Assist in favor of Agent Copilot. Critical assessments documented persistent implementation barriers: hallucinations, over-automation risks, need for specialized training, and human supervision requirements highlighted as prerequisites for safe deployment.
  • 2025-Q1: Auto-draft entered production maturity phase across enterprise tier-1 platforms. KPMG analyst validation confirmed enterprise-wide shift from experimentation to large-scale production deployment. Named customer case (Freedom Furniture) demonstrated 92% faster resolution and 17% CSAT improvement from agent copilot workflows. However, critical assessments reinforced that agent-assist technology delivers immediate ROI while autonomous systems remain premature; Zendesk production incident revealed reliability challenges in large-scale deployment, highlighting need for careful implementation and monitoring.
  • 2025-Q2: No new independent deployment evidence identified. Vendor announcements and agentic AI discussions dominated the window; no named customer case studies or adoption metrics specific to auto-draft in customer service operations during this period.
  • 2025-Q3: Auto-draft consolidates as the proven pattern within agentic AI. Broader agentic AI adoption faces friction: EY survey shows only 34% implementation despite 55% intent for customer support; CMU/Gartner research predicts 70% failure rate and 40% project cancellations by 2027 for autonomous agents. Google Agent Assist deployment guide documents 10-15% AHT improvement with proper implementation. Q3 evidence shows sharp bifurcation: human-in-the-loop auto-draft advancing into maturity and sustained ROI, while fully autonomous systems face mounting skepticism. Auto-draft's success hinges on maintaining human review gates and agent agency—validation that augmentation strategies outperform replacement automation.
  • 2026-Jan: GA feature launches demonstrate maturity: Zendesk releases AI-generated procedure drafts (3 per week), Intercom publishes automation rate KPI with production tracking. However, adoption-execution gap widens: Intercom survey shows 82% invested but only 10% mature (87% of mature teams see quality improvements vs 43% of explorers). Agentic AI failure rates spike: RAND/Gartner research documents 88% project failure, 11% production deployment, 40% cancellations forecast by 2027. Technical reliability concerns deepen: drift, inconsistency, and engineering cost barriers highlighted across industry analysis. Bifurcation sharpens: human-in-the-loop auto-draft advancing to enterprise scale with sustained ROI, while autonomous systems face escalating cancellations and skepticism.
  • 2026-Feb: Auto-draft accessibility expands: Zendesk rolls out capped AI writing tools (tone control, expand/simplify) to Professional+ plans in Feb 2026, broadening feature access from enterprise to mid-tier. Governance challenges surface: Gravitee's survey of 900+ executives finds 81% deployed AI agents but only 14.4% have security approval and 88% report incidents—validating mandatory human review as critical architectural control rather than limitation. Production deployments mature: Named customers (Telus 40 min/interaction, Suzano 95% query time reduction, Danfoss 80% automation) demonstrate enterprise-scale ROI, reinforcing that AI agents deliver value when properly implemented with human gates.
  • 2026-Apr: Governance tooling matures further: Zendesk's March 2026 release adds auto-assist event logging for full audit trails and pre-approved action workflows for low-risk tasks, confirming governance layers are now native to the product. Deployment breadth consolidates around 55% adoption (Metrigy, 656 companies), with Nucleus Research documenting measurable resolution and effort gains across 30+ production customers. Consumer risk evidence sharpens the case for human review: Qualtrics' survey of 20,000+ consumers in 14 countries documents AI customer service failing at 4x the rate of other AI applications, while Stanford-CMU research shows hybrid human-AI teams outperform autonomous systems by 68.7%—reinforcing the human-review gate as both a governance necessity and a performance advantage.
  • 2026-May: Quantified ROI consolidates across multiple benchmarks: Digital Applied documents hybrid escalation model at 4.25/5 CSAT with $0.62 per-resolution cost (vs $7.40 human-only); Balto ROI analysis documents 20% AHT reduction as a standard deployment outcome; enterprise AI benchmarking shows 620% average ROI within 18 months where tier-1 and tier-2 queries are handled with 78% autonomous resolution before escalation. Agent assist productivity metrics firm up: 9x cost reduction per task, 8.7 hours saved weekly, 4.2x productivity multiplier, 4.1-month payback period across deployments. Governance remains essential: Hiver survey of 700+ leaders finds 90% uncomfortable with AI representing brand directly to customers, confirming that the human-review gate is as much a trust requirement as a performance control.