The AI landscape doesn't move in one direction — it lurches. Some techniques leap from experiment to table stakes in a single quarter; others stall against regulatory walls, technical ceilings, or organisational inertia that no amount of hype can dislodge. Knowing which is which is the hard part. The State of Play cuts through the noise with a rigorously maintained index of AI techniques across every major business domain — classified by maturity, evidenced by real-world adoption, and updated daily so you always know where you stand relative to the field. Stop guessing. Start knowing.
A daily newsletter distilling the past two weeks of movement in a domain or two — delivered to your inbox while the index updates in the background.
Each dot marks the weighted maturity of practices within a domain — hover for a brief summary, click for more detail
AI-powered search and retrieval-augmented generation across internal documentation and enterprise systems. Includes cross-system search federation and context-aware answer generation; distinct from domain-specific RAG which targets specialised corpora rather than general enterprise knowledge.
Enterprise search and RAG is a proven practice with mature tooling, documented ROI, and broad adoption -- yet one where execution discipline, not technology, now determines success or failure. The pattern combines keyword and semantic retrieval with LLM-powered answer generation over proprietary data, grounding generative AI in internal knowledge rather than public corpora. Hybrid retrieval (vector plus BM25) settled as the production standard after vector-only approaches proved unreliable for exact matches, structured data, and multi-hop reasoning. The technology works. The harder problem is everything around it: chunking strategy, document quality, governance, cost control, and evaluation frameworks. Forty-five percent of enterprise AI deployments now incorporate RAG, and adopters report strong economics, but failure rates remain high among organisations that treat it as a plug-and-play capability rather than an operational discipline. The practice's defining tension is this gap between technological maturity and organisational readiness -- a gap that has stalled further tier advancement despite a market exceeding $2B and infrastructure that continues to improve.
The infrastructure layer is production-grade and still improving. Elasticsearch 9.3 shipped bfloat16 vector compression (halving storage) and GPU-accelerated indexing with 12x throughput gains. Azure AI Search added agentic retrieval with expanded knowledge sources including OneLake and SharePoint. Vector database adoption surged 377% year-over-year. These are not early-adopter tools -- they are GA platform features embedded in mainstream enterprise stacks.
Adoption metrics tell a confident story on the surface: 92% of RAG adopters report ROI within 12 months, averaging 3.2x return. But the denominator matters. Of roughly 1,000 enterprises that attempted RAG deployments through 2025, only about 200 succeeded, with 51% of enterprise AI failures being RAG implementations. The pattern that emerges is not technology risk but operational neglect -- 70% of deployments lack systematic evaluation frameworks, 30-40% of infrastructure budgets are wasted on poorly observed pipelines, and 87% of enterprises measure answer quality while ignoring data freshness and pipeline reliability.
Document quality remains the most underestimated barrier. Standard chunking destroys the logical structure of technical documents -- tables, cross-references, embedded images -- producing hallucinations even when retrieval is technically correct. Recent production analysis identifies five specific failure modes: irrelevant retrieval from poor ranking, partial answers split across multiple chunks, outdated answers from stale knowledge bases, answer refusal when retrieval fails, and hallucinated sources detached from actual documents. The emerging discipline of RAGOps attempts to address this by treating retrieval pipelines as production systems requiring monitoring, governance, and lifecycle management rather than one-time integrations. Evaluation frameworks (six-layer maturity models evaluating corpus quality, retrieval accuracy, groundedness, task success, latency/cost, and escalation design) are now standard in mature deployments. For organisations willing to invest in that operational discipline, enterprise RAG delivers 12-18% precision improvement through hybrid search, 69% error reduction from contextual compression, and 75% accuracy gains on complex regulatory documents via agentic reasoning. For those expecting turnkey results, the failure rate remains punishing.
— Azure AI Search April 2026 updates: GA semantic ranker on free tiers, agentic retrieval with reasoning control, document sensitivity labels, advancing platform maturity.
— 70-80% of large enterprises have production RAG; enterprise AI spending exceeds $300B in 2026 with 40%+ on generative AI, confirming mainstream adoption.
— Agentic RAG market projects $3.8B→$165B (2024-2034); named deployments: Morgan Stanley (financial research), PwC (tax/compliance), ServiceNow (task automation).
— RAGAS established as de facto evaluation standard; AWS, Microsoft, Databricks, Moody's running 5M+ monthly evaluations, advancing measurement infrastructure.
— Uber deployed OpenSearch for semantic search on 1.5B items, evaluated multiple platforms, solving ingestion and performance bottlenecks at scale.
— Gartner study: 52% of enterprise AI hallucinate on ungoverned RAG vs near-zero on governed data; IBM: 72% of AI failures from inadequate context, not models.
— Critical analysis: 50-90% of LLM responses lack full support; 57% of citations unfaithful (post-hoc rationalization); documents citation faithfulness gap.
— Detailed benchmark of vector databases (Pinecone, Weaviate, Qdrant, Milvus) with latency, recall, and cost metrics for enterprise RAG.