The AI landscape doesn't move in one direction — it lurches. Some techniques leap from experiment to table stakes in a single quarter; others stall against regulatory walls, technical ceilings, or organisational inertia that no amount of hype can dislodge. Knowing which is which is the hard part. The State of Play cuts through the noise with a rigorously maintained index of AI techniques across every major business domain — classified by maturity, evidenced by real-world adoption, and updated daily so you always know where you stand relative to the field. Stop guessing. Start knowing.
A daily newsletter distilling the past two weeks of movement in a domain or two — delivered to your inbox while the index updates in the background.
Each dot marks the weighted maturity of practices within a domain — hover for a brief summary, click for more detail
AI that retrieves relevant information from a single search query and synthesises a coherent answer with source attribution. Includes search-augmented generation and cited responses; distinct from deep research which conducts multi-step autonomous investigation.
Single-query research retrieval has crossed into production at forward-leaning organisations, but a persistent reliability gap keeps it from becoming standard infrastructure. The practice combines information retrieval with generative AI: a system executes one or a few searches, retrieves relevant sources, and synthesises a cited answer in a single pass. Perplexity, You.com, ChatGPT with browsing, and enterprise RAG deployments all embody this pattern. Adoption is real and growing fast -- Perplexity alone has surpassed 50 million monthly active users, and two-thirds of B2B buyers report using AI search tools. Yet the same deployments surface a stubborn accuracy ceiling: hallucinations, misattributed citations, and query-specific failures that aggregate metrics routinely mask. Architectural advances such as hybrid retrieval and adaptive depth are narrowing the gap, not closing it. The defining tension at leading-edge is exactly this: organisations are deploying at scale while knowingly accepting systematic factuality risks that no shipping product has resolved.
The vendor ecosystem is broad and scaling. Perplexity has reached 50 million monthly active users and ships production APIs (Agent API, Embeddings API) to general availability; You.com processes over one billion API queries per month for enterprise customers including Alibaba and DuckDuckGo; Databricks has launched Instructed Retrieval, a hybrid deterministic-probabilistic search yielding 35-50% recall gains; and Google Cloud now offers a production RAG platform on Vertex AI with hybrid search and re-rankers. Perplexity's acquisition of Carbon signals a push into enterprise grounding across internal data sources, complemented by a 200-seat law enforcement deployment that extends adoption into the public sector.
Enterprise deployments confirm productivity value. The Cleveland Cavaliers use Perplexity across 15+ teams, reporting 10+ hours saved per employee per week. Databricks documents 5,000 working hours of monthly savings. B2B buyer research finds AI search shortens research cycles by 34%. Yet a wide gap separates usage from bottom-line impact: 71% of organisations use generative AI regularly, but only 17% attribute more than 5% of earnings to it.
Reliability remains the binding constraint, with April 2026 evidence hardening the gap between adoption and performance. Large-scale accuracy testing (NP Digital, BBC/EBU) shows Perplexity at 49.3% accuracy with 12.2% outright error rate; 51% of AI answers contain significant issues, and 13% fabricate attributed quotes entirely. The CRAG benchmark from Meta/HKUST documents the capability ceiling: advanced LLMs achieve ≤34% accuracy, basic RAG improves to 44%, and state-of-the-art solutions reach 63% without hallucination—across just 4,409 question-answer pairs. A PRISMA review of 128 RAG studies documents persistent data quality failures across pipeline stages. A Nature-published Bixonimania experiment shows systems elaborated false statistics and contaminated peer review through citation laundering; practitioners report Charlotin database tracked 1,200+ hallucination cases by early 2026 (growing 5-6 daily), with Perplexity showing 37% citation error rate. Practitioner analysis finds that standard evaluation metrics mask query-specific catastrophes--a system can score well on average while hallucinating on the queries that matter most. Adaptive retrieval techniques show promise, with production implementations reporting 40-60% latency improvements, but these are engineering workarounds for a deeper problem: single-pass systems treat every query identically, under-serving complex questions while over-processing simple ones. No published breakthrough has closed this gap.
— Comparative analysis of retrieval and citation behavior across three major single-query research platforms; documents platform-specific ranking architectures and source selection biases.
— Stanford AI Index analysis documenting structural hallucination failure: knowledge-belief distinction collapse when users assert false premises; models fail at 86% rate under this condition.
— Comprehensive market analysis: ChatGPT 900M weekly users, 17% of all digital queries; conversion impact shows AI-referred visitors convert 23x higher than organic; cites are new ranking signal.
— Aggregated adoption metrics from primary sources (OpenAI, Adobe, Gartner); comprehensive evidence of mainstream single-query AI research adoption, search behavior shifts, and user quality expectations.
— Critical negative signal from Tow Center for Digital Journalism study. Shows real failure rates in production single-query research tools: Perplexity 37% wrong, ChatGPT 40% wrong, overall 60% inaccuracy.
— Market analysis of AI search platforms (ChatGPT, Gemini, Perplexity); describes market shift from keyword search to AI-synthesized answers with citations, directly addressing single-query research transition.
— Original rigorous benchmark: 5,000 prompts across frontier models measuring hallucination on factual recall, citation accuracy, and code reference; citation accuracy worst at 12.4% average.
— Measures AI search referral adoption at 0.9% of web traffic (5x YoY growth), names platforms (ChatGPT, Perplexity, Gemini, Claude); projects 3-5% share by end 2027 with steeper growth curve than organic.