Perly Consulting │ Beck Eco

The State of Play

A living index of AI adoption across industries — where established practice meets the bleeding edge
UPDATED DAILY

The AI landscape doesn't move in one direction — it lurches. Some techniques leap from experiment to table stakes in a single quarter; others stall against regulatory walls, technical ceilings, or organisational inertia that no amount of hype can dislodge. Knowing which is which is the hard part. The State of Play cuts through the noise with a rigorously maintained index of AI techniques across every major business domain — classified by maturity, evidenced by real-world adoption, and updated daily so you always know where you stand relative to the field. Stop guessing. Start knowing.

The Daily Dispatch

A daily newsletter distilling the past two weeks of movement in a domain or two — delivered to your inbox while the index updates in the background.

AI Maturity by Domain

Each dot marks the weighted maturity of practices within a domain — hover for a brief summary, click for more detail

DOMAIN
BLEEDING EDGEESTABLISHED

Code search & codebase Q&A

LEADING EDGE

TRAJECTORY

Stalled

AI-powered semantic search and question answering across large codebases, going beyond keyword matching. Includes tools that answer questions about architecture, dependencies, and usage patterns; distinct from documentation generation which produces static artefacts.

OVERVIEW

Code search and codebase Q&A has reached inflection: the practice is now table-stakes infrastructure for enterprise AI development, yet fundamental architectural limitations persist. Semantic code search—asking natural language questions about architecture, dependencies, and usage patterns—addresses a real bottleneck (developers spend ~15% of time on discovery), and deployments at Qualtrics, Coinbase, Booking.com, Altisource, and clinical programming teams demonstrate production value. GitHub and Sourcegraph reached GA milestones in Q1 2026: Copilot's semantic indexing achieved seconds-fast retrieval; Cody shifted enterprise-only (July 2025). Yet vendors are moving beyond RAG architectures precisely because semantic-only retrieval has hit limits. Retrieval precision drops sharply with corpus scale (87% degradation at 50K+ documents); retrieval consistency varies >50% across prompt formulations; context window constraints cause accuracy degradation on large codebases (50% on >10K LOC). The tier-defining tension has inverted: maturity is now about moving beyond semantic search toward hybrid keyword-semantic-structural approaches. Adoption and trust remain misaligned: 52% of developers use leading tools but 96% distrust AI output, and embedding drift silently degrades relevance in production systems. The practice remains leading-edge but constrained by architectural fragility and unresolved reliability barriers.

CURRENT LANDSCAPE

GitHub and Sourcegraph control enterprise deployments, with Q1 2026 and April 2026 milestones confirming market consolidation. Copilot's semantic code search reached GA (March 2026, sub-second indexing); Sourcegraph shipped Smart hover summaries (April 2026) grounding Q&A in precise code intelligence rather than embeddings alone. Cody moved enterprise-only ($19-$49/seat, July 2025), signaling market bifurcation. Real-world deployments demonstrate measurable ROI: Qualtrics' 1,000-developer rollout reduced IDE navigation 28% and code Q&A time 25%; Altisource modernized 350K lines with 25% productivity gain and 54% vulnerability reduction; clinical programming team achieved 4/5 satisfaction with 791 semantic queries over 6 weeks on 300K+ corpus. Scale Labs' 2026 benchmark (124 production tasks) revealed 30% frontier capability ceiling in architecture, root-cause, and onboarding analysis.

Adoption breadth masks critical limitations. Market surveys show 52% developer adoption of leading tools (Claude Code, Cursor) but 96% distrust AI output; code review time now exceeds writing time. Technical assessments document scaling failures: context window constraints cause 50% accuracy degradation on codebases >10K LOC; embedding drift silently degrades relevance in production without visible error signals; text-based search fails on complex inheritance and templating. Wikimedia Foundation deployed semantic search at scale (1.1M snippets, 83K files) but vendors simultaneously pivot toward hybrid strategies (keyword-embedding-AST, language servers) recognizing semantic-only approaches hit hard limits. Third-party tools remain niche; duopoly consolidation driven by adoption barriers—secret leakage (6.4% for Copilot users, 40% above baseline), vulnerability generation, and vendor lock-in concerns restraining broader ecosystem. The practice achieved mainstream integration but remains constrained by unresolved reliability and architectural boundaries.

TIER HISTORY

ResearchJan-2023 → Jan-2023
Bleeding EdgeJan-2023 → Jan-2024
Leading EdgeJan-2024 → present

EVIDENCE (92)

Deep Search quantitative code analysisProduct Launches

— Sourcegraph Deep Search ships programmatic aggregations for quantitative code analysis: counting, ranking, grouping across repository searches in single turn, extending code search beyond retrieval into analytics.

— GitHub ships semantic code search (all workspaces), grep-style cross-repo queries (githubTextSearch), and /chronicle chat history Q&A feature; semantic search expansion to all workspaces removes GitHub-only constraint.

— CoREB benchmark reveals code search as specialized retrieval domain: code-specialised embeddings dominate code-to-code by 2×, yet short keyword queries collapse all models to near-zero nDCG@10, identifying fundamental code search challenges.

— Technical analysis: semantic search destroys document ontology through fixed-size chunking, failing on hierarchical structures. Code is inherently hierarchical (package/class/method/block); ontological approach outperforms embeddings on structure-dependent queries.

— Productivity analysis across 15,000+ placements: unfamiliar codebase navigation shows -19% slowdown, revealing codebase Q&A and search immaturity as a productivity barrier and adoption constraint.

— Market shift detected: 'Most teams looking for a Sourcegraph alternative have moved past code search as the core problem. They want a context layer for autonomous development.' Code search matured from problem to table-stakes infrastructure component.

— Benchmarks show semantic indexing delivers 62× fewer tokens, 84% fewer agent steps vs grep. Five competing tools shipping production code search (Cursor, Zilliz, sverklo, SocratiCode, VS Code); ecosystem maturation signals code search as commodity.

— Q1 2026 multi-survey analysis: Claude Code dominant (70% net like, 46% 'most loved') with 75% adoption among small startups; excels at multi-file editing and entire codebase understanding, signaling market preference.

HISTORY

  • 2023-H1: Code search and codebase Q&A entered production with Sourcegraph Cody GA and bloop launch. Stack Overflow's 2023 survey showed 44–70% adoption intent among developers, but only 3% reported high confidence in accuracy, indicating real-world use paired with maturity concerns around validation and trust.
  • 2023-H2: Core platforms upgraded semantic search infrastructure. GitHub Copilot Chat reached GPT-4 with code referencing public beta; VS Code's @workspace added knowledge graphs and local indexing for context retrieval. Adoption metrics surged (84% of developers using AI tools as search engines). Academic research achieved 0.7795 MRR on CodeSearchNet, validating technical maturity. Real deployments demonstrated practical value in code migration and refactoring. Limitations remained: deep debugging, complex error analysis, and non-mainstream framework understanding all flagged as weak areas by users.
  • 2024-Q1: Enterprise adoption accelerated with Cody Enterprise GA (Qualtrics, Leidos deployments). Foundational infrastructure matured: Voyage AI released voyage-code-2 with 14.52% recall improvement on code retrieval. Research community focused on standardization: InfiBench and CodeQueries datasets launched to benchmark code Q&A across 100+ models. Parallel concerns emerged: GitClear research documented code churn increase and maintainability risks from AI-assisted coding, raising questions about long-term codebase health.
  • 2024-Q2: Large-scale remote repository context handling advanced: Cody Enterprise demonstrated capability for 300,000+ repository deployments and monorepos exceeding 90GB. Ecosystem tools matured: bloop reached v0.6.5 with conversational search and code studio features (9,486 GitHub stars). Evaluation infrastructure expanded with CoSQA+ benchmark (92% quality verification). Adoption surged to 76% of surveyed developers using AI coding tools, but 38% reported frequent inaccuracies, signaling trust remains a limiting factor for production-critical code understanding tasks.
  • 2024-Q3: Adoption reached 97% of developers using AI coding tools (GitHub survey, 2,000+ respondents), but critical trust gap emerged—developer reliance significantly lagged behind awareness. Sourcegraph Cody expanded free tier with Claude 3.5 Sonnet, Mixtral, Gemini 1.5 support and lifted query limits; enterprise model selection entered early access. Gartner forecasted 30% GenAI project abandonment by end of 2025. Ecosystem consolidation accelerated: Bloop ceased development and repository archived, leaving GitHub/Sourcegraph duopoly as primary code Q&A platforms. The practice solidified as mainstream but constrained by accuracy concerns and narrow value capture outside large enterprise deployments.
  • 2024-Q4: Platform maturation accelerated with GitHub releasing experimental semantic search relevance sorting in Copilot Chat (October 2024) and Sourcegraph Cody reaching GA for enterprise model selection across Bedrock, Azure OpenAI, and Vertex AI (November 2024). Research advances demonstrated capability gains: RAG-powered LLM agents achieved 78.2% CodeSearchNet success; SCAM 2024 presented REINFOREST, improving cross-language code search by 44.7%. Cortex survey identified context gathering as the leading productivity blocker (26% weekly unproductive work), validating core pain point. Ecosystem consolidation continued: Bloop's technical degradation (initialization failures, account bugs in Q4 2024) preceded January 2025 archival. The practice remained concentrated in GitHub/Sourcegraph duopoly, with business model sustainability constraining broader vendor ecosystem growth.
  • 2025-Q1: Platform optimization accelerated: GitHub Copilot's semantic code search indexing reached GA (March 2025) with indexing time reduced from ~5 minutes to seconds, removing latency barriers to codebase-aware context retrieval. Sourcegraph deployed Analytics infrastructure for enterprise ROI measurement of Code Search and Cody deployments. Developer adoption remained near-universal (98% using AI coding tools weekly for 'explain this' and debugging workflows) but trust remained low—Stack Overflow's 2025 survey of 65,000 developers confirmed 84% adoption while 46% actively distrust AI accuracy, indicating the practice had achieved ubiquity without resolving core reliability challenges. The GitHub/Sourcegraph duopoly consolidated control while smaller vendors struggled with sustainability, leaving code search and Q&A tightly integrated into mainstream platforms but geographically concentrated in vendor ecosystems.
  • 2025-Q2: Enterprise deployments validated production value: Qualtrics reported 28% reduction in IDE navigation for code understanding and 25% faster code Q&A via Cody (1,000+ developer rollout), demonstrating measurable ROI at scale. GitHub roadmapped Copilot Analytics dashboards (Q3 2025) signaling advanced enterprise metrics. Yet reliability concerns deepened: InfoWorld documented 59% of engineers reporting frequent AI code errors and 67% spending more time debugging AI outputs; Greptile analysis revealed fundamental embedding gap in semantic code search (12% performance drop when searching raw code vs. natural language summaries). Third-party ecosystem matured incrementally—CodeCompanion.AI integrated voyage-code-3 embeddings—but GitHub/Sourcegraph duopoly remained unchallenged. Academic research (FORGE 2025) advanced retrieval techniques via semantic graphs, but adoption remained constrained by accuracy and trust barriers, not technical capability.
  • 2025-Q3: Adoption surged to 80% of 49,000-developer Stack Overflow survey, but trust collapsed to 29%—revealing a widening paradox in code search and Q&A maturity. GitHub and Sourcegraph tightened duopoly: Sourcegraph shipped enterprise analytics for ROI measurement while GitHub advanced semantic search feature expansion. Developers reported 45% spending more time debugging AI-generated "almost-right" code than writing manually, inverting productivity gains. Research advances (structural code search via natural language queries, 55-70% precision/recall on 400-query benchmark) provided technical progress, but ecosystem stagnation persisted—third-party tools like CodeCompanion.AI matured incrementally while independent code search platforms (bloop) remained archived. The practice remained architecturally mature but functionally constrained by unresolved reliability barriers and vendor consolidation.
  • 2025-Q4: Research community pivoted toward addressing hallucinations in code Q&A systems: peer-reviewed advances (92% citation accuracy via hybrid retrieval, adaptive RAG for black-box models) demonstrated technical maturity in hallucination mitigation. Platform vendors accelerated feature delivery: GitHub shipped CLI semantic search with natural language Q&A; Sourcegraph maintained analytics infrastructure. Third-party analysis documented ecosystem consolidation with Cody and Copilot dominating enterprise deployments (Coinbase, Booking.com, Qualtrics). Despite technical progress and widespread adoption, the practice remained trapped in reliability-trust paradox—improvements in retrieval accuracy did not yet translate to user confidence at scale. By year-end 2025, code search and codebase Q&A had achieved mainstream integration but remained limited by unresolved accuracy validation mechanisms and vendor lock-in.
  • 2026-Jan: Ecosystem continued consolidation with Sourcegraph Cody expanding enterprise security certifications and deployment flexibility (self-hosted, cloud, hybrid options). Third-party semantic code search tools matured: llm-tldr and similar tools demonstrated sub-100ms query latency and 16-language support, signaling infrastructure commoditization. Practitioner analysis reaffirmed semantic code search ROI: developers spend 15% of time on code discovery, semantic retrieval improves LLM success by 20% vs. keyword search. Duopoly control (GitHub/Sourcegraph) remained unchallenged; independent tools showed specialization rather than competition. The practice remained architecturally mature with persistent adoption-trust misalignment.
  • 2026-Feb: Platform vendors published enterprise comparisons positioning semantic code search as mainstream: Cody recognized in Gartner Magic Quadrant (Sep 2025 Visionary) with 1M-token RAG contexts and multi-repo retrieval; deployment analysis showed 30% reduction in manual code exploration but highlighted vendor lock-in risks. Meanwhile, leading vendors (GitHub Copilot, Cody) began architectural shifts away from pure RAG toward hybrid strategies (keyword matching, AST analysis, agent-based retrieval) to address semantic drift and structure-awareness limitations. February 2026 marked an inflection: code search and Q&A had achieved ubiquitous enterprise adoption but leading voices questioned whether RAG alone was the right architectural direction. The practice remained entrenched in GitHub/Sourcegraph duopoly with measurable deployment value alongside unresolved accuracy barriers.
  • 2026-Mar: Platform vendors released Q1 2026 feature milestones confirming semantic code search as table-stakes: GitHub Copilot shipped semantic code search GA (March 17) with automatic meaning-based retrieval; Sourcegraph Cody enterprise-only repositioning (discontinued free/pro July 2025) confirmed market bifurcation. Independent evaluations documented retrieval improvements (semantic search: 90% relevant context vs 30% keyword) and deployment ROI (clinical programming, enterprise monorepos). Yet simultaneously, peer-reviewed research surfaced critical limitations: semantic collapse at scale (87% precision drop on 50K+ corpuses), RAG-induced inconsistency (>50% variance across prompts), retrieval latency overhead (300-400ms). Market positioning reflected the tension: industry analysis positioned code search as essential infrastructure while vendor roadmaps signaled moves toward hybrid keyword-semantic-structural approaches. The practice achieved universal enterprise adoption but remained constrained by architectural limits that single-modality semantic search cannot overcome.
  • 2026-Apr: Semantic code search matured as a differentiator signal while capability benchmarks revealed hard ceilings. GitHub shipped semantic code search GA in Copilot for VS Code (v1.111–v1.115) with an auto-managed index requiring no configuration; Sourcegraph shipped Smart hover summaries (GA) using precise code intelligence to ground Q&A outputs in actual symbol usage rather than embeddings alone. Scale Labs' SWE Atlas benchmark (124 tasks across 11 production repos) revealed a 30% frontier capability ceiling for AI agents on architecture, root-cause, and onboarding comprehension tasks — quantifying the gap between search retrieval and genuine codebase understanding. A 55,000-developer survey found 52% adoption of Claude Code and Cursor combined, but 96% distrust of AI output, with code review time now exceeding writing time. Production deployment evidence widened: Altisource modernized 350K lines of legacy Java with Amazon Q Developer achieving a 25% productivity gain and 54% vulnerability reduction in 4 months versus 9–12 months prior. NxCode documented Copilot's 8K context window causing 50% accuracy degradation on codebases over 10K LOC; Wikimedia Foundation deployed semantic code search at scale (1.1M snippets, 83K files, 2,400+ repos) confirming production viability of meaning-based retrieval. Embedding drift emerged as a systemic reliability concern: semantic search relevance silently degrades in production without triggering alerts. JetBrains AI Pulse (10K+ developers) found Claude Code at 18% adoption and 91% CSAT — signalling market broadening beyond the GitHub/Sourcegraph duopoly, though the trust-adoption gap remained unresolved.
  • 2026-May: Platform capability expansion accelerated in early May with Sourcegraph Deep Search adding programmatic aggregations for quantitative code analysis (counting, ranking, grouping across repository searches in single turn, extending code search beyond retrieval into analytics). GitHub Copilot April releases (v1.116–v1.119) expanded semantic search to all workspaces (removing GitHub-only constraint), added githubTextSearch for grep-style cross-repo queries, and introduced experimental /chronicle feature for chat history Q&A. Market analysis revealed ecosystem maturation: semantic codebase indexing delivered 62× fewer tokens and 84% fewer agent steps vs grep-based search across five competing tools (Cursor, Zilliz, sverklo, SocratiCode, VS Code); Q1 2026 developer surveys showed Claude Code dominance (70% net like, 46% 'most loved') with 75% adoption among small startups for multi-file editing and codebase understanding. Yet critical limitations persisted: CoREB research paper identified code search as specialized domain where code-specialised embeddings dominate code-to-code tasks by 2× yet fail on short keyword queries (near-zero nDCG@10); hierarchical data analysis showed semantic search destroys ontology through fixed-size chunking (inherent weakness for code's class/method/block structure); embedding fine-tuning for precision degrades broad retrieval 40%, creating architectural tradeoff preventing single-solution deployment. Productivity analysis documented negative signal: unfamiliar codebase navigation showed -19% slowdown, revealing codebase Q&A immaturity as adoption barrier. Market positioning shifted: vendors recognized code search had matured from standalone problem to table-stakes infrastructure component, with customer demand shifting toward context layers for autonomous development rather than code search features alone. The practice remained leading-edge but fundamentally constrained by single-modality semantic approaches hitting architectural limits.

TOOLS