Perly Consulting │ Beck Eco

The State of Play

A living index of AI adoption across industries — where established practice meets the bleeding edge
UPDATED DAILY

The AI landscape doesn't move in one direction — it lurches. Some techniques leap from experiment to table stakes in a single quarter; others stall against regulatory walls, technical ceilings, or organisational inertia that no amount of hype can dislodge. Knowing which is which is the hard part. The State of Play cuts through the noise with a rigorously maintained index of AI techniques across every major business domain — classified by maturity, evidenced by real-world adoption, and updated daily so you always know where you stand relative to the field. Stop guessing. Start knowing.

The Daily Dispatch

A daily newsletter distilling the past two weeks of movement in a domain or two — delivered to your inbox while the index updates in the background.

AI Maturity by Domain

Each dot marks the weighted maturity of practices within a domain — hover for a brief summary, click for more detail

DOMAIN
BLEEDING EDGEESTABLISHED

Knowledge base generation & maintenance

GOOD PRACTICE

TRAJECTORY

Stalled

AI that generates knowledge base articles from support history and autonomously maintains, updates, and identifies gaps in existing knowledge. Includes article drafting from resolved tickets and coverage gap detection; distinct from self-service content which creates user-facing experiences rather than internal knowledge.

OVERVIEW

AI-powered knowledge-base generation has reached proven, accessible maturity -- every major CX platform ships it as a GA feature, deployments number in the tens of thousands, and the ROI case is well documented. The practice has stalled not because it failed but because it hit an architectural ceiling: autonomous article drafting works, yet fully autonomous maintenance does not. Hallucination research consistently shows that AI amplifies knowledge-quality problems faster than organisations can fix them, which means human review gates remain structurally necessary. For teams evaluating this space, the question is no longer whether to adopt KB generation tooling but how to build the data-hygiene and governance discipline that makes it reliable. The tooling is commoditised; the operational wrapper around it is not.

CURRENT LANDSCAPE

Zendesk, ServiceNow, Freshworks, Microsoft, and HubSpot all offer GA knowledge-base generation features, and the market has fully commoditised. Zendesk Knowledge Builder powers over 50,000 active knowledge bases; Freshworks serves 73,000+ customers with Freddy AI, and named deployments show real results -- Qualia reached 91% help-centre usage with a 30% ticket reduction, while ServiceNow's internal deployment hit 54% deflection and $5.5M in annual savings. The AI knowledge-management market grew from $5.23B in 2024 to $7.71B in 2025, projected to reach $35.83B by 2029.

That scale, however, has not solved the accuracy problem. Comprehensive April-May 2026 research shows hallucination rates spanning 0.7%-88% depending on model and task (Suprmind benchmark), with data governance as the decisive lever: 52% of enterprise AI responses hallucinate on ungoverned data versus near-zero on governed data using the same model (Atlan). Industry data from 2024 shows 39% of AI customer service implementations were rolled back or reworked due to hallucinations, with 76% requiring human-in-the-loop review before production. Peer-reviewed research demonstrates knowledge base semantic quality improves accuracy by 17-23 percentage points across frontier models (Claude Opus 4.7, Claude Sonnet 4.6, GPT-5.4), proving governance is the critical upstream work. Real-world deployments (MBH Architects, Docker, Nokia, OpenAI using Kapa.ai) show AI-assisted gap detection and maintenance workflows work, but all remain semi-autonomous with human review gates. Customer sentiment remains cautious: 94% of IT leaders concerned about vendor lock-in; vendors democratized KB features to standard plans (Zendesk April 2026) but accuracy and governance constraints, not tooling gaps, remain the binding limitations on full autonomy.

TIER HISTORY

ResearchJan-2023 → Jul-2023
Bleeding EdgeJul-2023 → Jul-2024
Leading EdgeJul-2024 → Apr-2025
Good PracticeApr-2025 → present

EVIDENCE (88)

— Comprehensive benchmark aggregating hallucination rates (0.7%-88% depending on model and task) across frontier AI models; documents endemic hallucination problem limiting autonomous KB deployment without governance.

— AI-assisted KB maintenance deployed by Docker, Nokia, and OpenAI; demonstrates practical workflow for gap detection using RAG-powered analysis to identify coverage gaps and guide content creation.

— Adoption barrier data: 39% of AI customer service implementations were rolled back or reworked due to hallucinations in 2024; 76% of enterprises require human-in-the-loop review to catch hallucinations before deployment.

What's new in Zendesk – April 2026Product Launches

— Zendesk generalizes AI KB features to Suite plans (not premium-only) including generative article writing, unified RAG system for search and answers, signaling platform maturity and commoditization of KB AI.

— Peer-reviewed benchmark testing Claude Opus 4.7, Claude Sonnet 4.6, and GPT-5.4 shows semantic context improves accuracy by 17-23 percentage points; proves KB semantic quality is structural requirement, not model-dependent.

— Industry analysis: 52% of enterprise AI responses contain hallucinations on ungoverned RAG data vs. near-zero on governed data (same model); proves KB governance and maintenance—not tooling—is the critical lever for reliable AI systems.

— MBH Architects deployed firm-wide KB transformation spanning marketing proposals, practice knowledge, project data, and learning; demonstrates structured knowledge capture, AI-enabled retrieval, and scope-drafting agents with measured time savings.

— KCS is mature methodology for knowledge generation embedded in customer support workflows; reps create and refine knowledge in real-time during case resolution, with double-loop Solve/Evolve process for continuous KB improvement.

HISTORY

  • 2023-H1: Platform vendors (Zendesk, ServiceNow, Freshworks, Microsoft) began integrating generative AI capabilities. Early interest in AI-powered knowledge work emerged, but few production deployments of autonomous KB generation and maintenance were documented. Market was exploring application of LLMs to knowledge problems.
  • 2023-H2: Major platforms moved AI-powered KB capabilities to general availability: Zendesk deployed knowledge summarization for bot responses, Freshworks reported 80% agent time reduction on KB automations with customer expectations of 10%+ ticket deflection, and Front launched integrated KB with AI-powered responses. Parallel market discussion surfaced critical barriers: accuracy limitations requiring controlled inputs, lack of contextual understanding, and governance concerns. Semi-autonomous workflows with human review emerged as the viable production model; full autonomous generation remained largely theoretical.
  • 2024-Q1: All major platforms published official GA documentation for AI KB generation and maintenance (Zendesk, ServiceNow, Microsoft). Early deployments showed promise with 40-60% automatic ticket resolution from AI-trained knowledge bases; Gartner 2024 research confirmed 60% higher ticket deflection on AI-first support platforms. However, capability gap persisted: systems excelled at knowledge retrieval and synthesis but lacked autonomous generation of new articles and external data integration. The practice remained semi-autonomous, with human judgment required for knowledge creation and quality governance.
  • 2024-Q2: KB generation and maintenance became standard feature in all major CX platforms. Zendesk announced advanced generative tools at Relate 2024 conference; ServiceNow released Now Assist for Knowledge Management (May 2024) with limited automation requiring manual review; Freshservice transitioned from legacy suggestion tools to AI-powered alternatives. Independent adoption data showed 54% of enterprises using AI for CX, with vendor-trained platforms 3.5x more effective than in-house builds. Real-world deployments (e.g., Fivetran's 2,521-article migration) demonstrated technical complexity and ongoing security, quality, and governance challenges limiting full autonomy.
  • 2024-Q3: Production deployments moved into mainstream adoption phase. Total Expert achieved 248% ROI deploying KB-powered support; HubSpot launched Breeze with Customer Agent trained on knowledge bases; Microsoft expanded Dynamics 365 with Copilot KB generation features. However, academic and practitioner research highlighted persistent limitations: hallucination risks, data privacy and regulatory constraints (GDPR, EU AI Act), and need for human governance. The practice remained fundamentally semi-autonomous—tooling for synthesis and creation, human oversight for quality and accuracy.
  • 2024-Q4: All major platforms solidified KB generation as a core, mainstream capability. Zendesk published GA documentation for AI-powered KB with generative article creation; HubSpot forced migration of all legacy knowledge bases to new AI-enhanced infrastructure with Breeze Copilot; ServiceNow released Q&A Genius Results for KB summarization. Industry guidance proliferated: ICMI published best practices for GenAI-ready KBs, and practitioners increasingly deployed LLMs for autonomous gap detection and coverage analysis. However, real-world quality constraints persisted: 30% of KB-powered chatbots experienced user abandonment due to irrelevant responses, only 45% met customer expectations, and human oversight remained non-negotiable for production accuracy and governance.
  • 2025-Q1: Vendors accelerated KB generation feature releases across platforms. Zendesk GA'd Resolution Platform with Knowledge Builder auto-generating KBs from ticket data; Freshworks GA'd Freddy AI Copilot with Help Article Generator for IT teams; Microsoft expanded Dynamics 365 with native KB generation from resolved cases. Production constraints remained persistent: expert analysis revealed KB maintenance paradox where AI generation requires clean data but generating that data remains labor-intensive, and typical ServiceNow deployments showed only 15% deflection vs 50% targets due to minimal agent notes limiting content creation.
  • 2025-Q2: Knowledge base generation and maintenance fully commoditized across vendor ecosystem. Zendesk Knowledge Builder scaled to power 50,000+ active knowledge bases; Freshworks demonstrated named customer deployments (Hobbycraft, Dunzo) with 30-48% outcome improvements; ServiceNow, HubSpot, Microsoft all offered GA KB generation features. However, fundamental limitations remained non-negotiable: AI factual accuracy below 50% in controlled testing, stale knowledge requiring dual-layer mitigation architecture, and quality input data dependency creating persistent knowledge maintenance bottleneck. Practice demonstrated plateau in maturity—vendor tooling mature but autonomous deployment without human governance remained aspirational.
  • 2025-Q3: Continued vendor stabilization with deepened focus on accuracy and governance challenges. Zendesk expanded Knowledge Builder with named production deployments (Qualia 91% help center usage/30% ticket reduction, Squarespace 95% self-service success, Tesco 30%→73% self-service growth, Humi 57% automated resolutions); peer-reviewed research validated RAG architecture (JMIR Cancer study confirmed 0% hallucination rate with quality sources vs 40% without), but independent academic research (Harvard Kennedy School) documented persistent hallucinations in deployed AI systems with real consequences, highlighting that technical mitigations remain insufficient for full autonomous operation without human verification.
  • 2025-Q4: Vendor ecosystem solidified feature parity with autonomous KB maintenance. ServiceNow deployed AI-driven knowledge gap detection in Now Assist (Yokohama release, Nov 2025) and reported internal deployment achieving 54% ticket deflection with $5.5M annual savings; Microsoft GA'd Customer Knowledge Management Agent for Dynamics 365 Contact Center (Oct 31) automating KB article drafting from case analysis; Freshworks published comprehensive Freddy AI documentation with article generation, suggestion, and agent-training capabilities; Zendesk Knowledge Builder refined workflow with recognized limitations (30-day data window, single-source constraint) requiring complementary tooling. However, governance remained the persistent blocker: practitioner analysis confirmed hallucination root causes trace to knowledge base quality, fragmentation, and insufficient data hygiene rather than model capabilities. McKinsey data cited only 1% of enterprises achieving AI maturity, with accuracy and governance as primary barriers. The practice remained semi-autonomous by architectural necessity—vendors had commoditized generation but governance, data quality, and human oversight remained mandatory for production deployments.
  • 2026-Jan: Vendor momentum continued with Freshworks GA of Freddy Copilot Help Article Generator (Jan 24), multimodal image support, and Google Drive integration. However, critical independent research amplified concerns about deployment readiness: Guru study documented knowledge accuracy as the primary AI blocker (8-12% manual verification capacity vs needs of 5,000-15,000 pieces), while deployment failure data showed 88% of AI agents never reach production with data fragmentation and integration complexity as core barriers. Freshworks reported 6,000+ paying customers achieving 50-60% query deflection, but real-world implementation issues (hallucinations, weak search relevance, confidential data exposure) dominated practitioner forums. Market growth accelerated with AI knowledge management expanding from $5.23B (2024) to $7.71B (2025), projected $35.83B by 2029, but maturity remained blocked by accuracy and governance challenges requiring strict human oversight.
  • 2026-Feb: Vendor ecosystem pushed feature parity with KB generation updates across all major platforms (Zendesk improving Knowledge agent RAG accuracy/latency, ServiceNow extending KB creation to IT Operations, Freshworks continuing Help Article Generator expansion). However, market reception cooled markedly: Parallels survey (Feb 2026) of 540 IT professionals showed 94% feared vendor lock-in and only 29% willing to pay premium for AI, signaling customer caution over proprietary platforms. Critical new research on hallucinations emerged: 17% of AI citations documented as unverifiable with 5% fabricated, driving shift toward retrieval-augmented systems (OpenScholar, PaperQA); concurrent LLM research (OpenAI, Anthropic) revealed hallucinations as structural problem resistant to scale. Adoption remained robust in scale (Freshworks 73,000+ customers, Zendesk 50,000+ KB instances, 6,000+ deployments achieving 50-60% deflection) but vendors lost pricing power as accuracy barriers overcame customer confidence in AI-only knowledge maintenance.
  • 2026-Mar/Apr: Vendor momentum continued with major platform updates: ServiceNow GA'd Now Assist in Knowledge Management for automated KB article generation (Australia/Zurich releases, March 2026); Zendesk democratized AI KB features to all plans (expanded availability March 2026); Freshservice published Freddy AI documentation with Help Article Generator and Knowledge Content Recommendations (March 2026). Comprehensive hallucination research compiled from authoritative benchmarks (Suprmind report, March 2026) confirmed persistent accuracy barriers across all models (3-13% hallucination rates in recent testing), reinforcing that human verification remains architecturally necessary. Market adoption data showed Zendesk AI ARR hit $500M (150% YoY growth) but only 25% of organizations fully integrated AI while 75% remained in pilot/partial phases, documenting sustained implementation maturity gap. Real-world deployments demonstrated concrete outcomes (Qualia 91% help center usage, Squarespace 95% self-service success, Tesco grew adoption from 30% to 73%) but confirmed that tooling commoditization has not translated to autonomous KB maintenance. The practice remained structurally semi-autonomous—vendors had solved KB generation feature parity, but data quality, governance discipline, and human oversight remained non-negotiable for production reliability.
  • 2026-Apr/May: Evidence crystallized on KB governance as decisive factor in hallucination control. Peer-reviewed research (arXiv April 28) demonstrated knowledge base semantic quality improves LLM accuracy by 17-23 percentage points across Claude Opus 4.7, Claude Sonnet 4.6, and GPT-5.4—proving semantic structure matters more than model selection. Suprmind's comprehensive April 30 hallucination benchmark aggregated data across frontier models showing 0.7%-88% rates depending on task; Atlan research confirmed 52% of enterprise AI responses hallucinate on ungoverned RAG data vs. near-zero on governed data. Industry adoption metrics documented mature KB practices: Knowledge-Centered Service (KCS) methodology embedded in customer support workflows (Salesforce); AI-assisted gap detection deployed by Docker, Nokia, and OpenAI (Kapa.ai); Zendesk April 2026 release unified RAG system for search/answers across Suite plans. However, rollback data from 2024 shows 39% of AI customer service implementations were reworked due to hallucinations, with 76% requiring human review before production. Market reception remained cautious despite feature parity—governance and data quality had become the binding constraints on wider adoption, not tooling availability. The practice remained in sustained maturity plateau: vendors shipped commoditized KB generation, but deployment obstacles centered on organizational change (KCS adoption, data hygiene discipline, governance rigor) rather than capability gaps.

TOOLS