Perly Consulting │ Beck Eco

The State of Play

A living index of AI adoption across industries — where established practice meets the bleeding edge
UPDATED DAILY

The AI landscape doesn't move in one direction — it lurches. Some techniques leap from experiment to table stakes in a single quarter; others stall against regulatory walls, technical ceilings, or organisational inertia that no amount of hype can dislodge. Knowing which is which is the hard part. The State of Play cuts through the noise with a rigorously maintained index of AI techniques across every major business domain — classified by maturity, evidenced by real-world adoption, and updated daily so you always know where you stand relative to the field. Stop guessing. Start knowing.

The Daily Dispatch

A daily newsletter distilling the past two weeks of movement in a domain or two — delivered to your inbox while the index updates in the background.

AI Maturity by Domain

Each dot marks the weighted maturity of practices within a domain — hover for a brief summary, click for more detail

DOMAIN
BLEEDING EDGEESTABLISHED

Model interpretability & explainability

GOOD PRACTICE

TRAJECTORY

Advancing

Techniques for understanding and explaining how AI models reach decisions, supporting transparency and accountability. Includes SHAP, LIME, and attention visualisation; distinct from model documentation which records metadata rather than explaining decision mechanisms.

OVERVIEW

Model interpretability and explainability is proven governance infrastructure -- the operational question is no longer whether to adopt it, but how to deploy it effectively within domain-specific and regulatory constraints. The practice centres on techniques (SHAP, LIME, attention visualisation, and newer mechanistic approaches) that make AI decision-making transparent enough for oversight, audit, and compliance. A mature vendor ecosystem, GA tooling from multiple cloud platforms, and enforceable regulatory mandates across jurisdictions confirm that explainability has crossed the threshold from forward-leaning initiative to expected organisational capability. The defining tension has not disappeared, though; it has sharpened. Post-hoc explanation methods remain the workhorse of production deployments, yet years of empirical research document their instability, disagreement across techniques, and vulnerability to misinterpretation by non-technical stakeholders. Inherently interpretable models offer stronger guarantees but narrower applicability. The field's maturity is visible less in triumphant adoption metrics than in its hardened pragmatism: organisations that succeed treat explainability as a design-time commitment requiring human oversight, domain expertise, and realistic expectations about what explanations can deliver.

CURRENT LANDSCAPE

Regulation is the dominant adoption driver, with enforcement now active across jurisdictions. EU AI Act enforcement began in January 2026 for general-purpose AI models, with high-risk system requirements following in August; financial regulators (OSFI Guideline E-23, CFPB, BaFin, FINMA) have made explainability a compliance requirement, with DACH banking regulators explicitly mandating SHAP or LIME implementation before production deployment. A survey of 600 global data leaders found that seven in ten have adopted generative AI, but 75% acknowledge governance and literacy gaps -- demand for explainability tooling outpaces organisational readiness to use it well. The vendor ecosystem reflects maturity: IBM watsonx.governance, Azure ML, Palantir AIP Control Tower, ServiceNow AI Control Tower, and DataRobot all offer production-grade explainability workflows with SHAP as standard functionality; the Thales-IBM governance solution brings EU AI Act compliance capabilities to market. Named deployments confirm real enterprise traction: JPMorgan Chase uses SHAP to explain credit card approvals with per-decision feature attribution; Wells Fargo processes 400,000 mortgage applications annually with SHAP-generated adverse action explanations; BNP Paribas uses SHAP for AML alert justification, reducing false positive investigation time by 35%. Yet deployment remains challenging; Azure ML's XAI dashboard still lacks multi-level classification support, and 60% of AI-based credit decisions reviewed by the CFPB lack adequate explainable reasoning. Mechanistic interpretability has begun crossing from research into production: Google DeepMind added activation probes to live Gemini deployments for misuse detection, demonstrating production-scale adoption of circuit-level analysis for governance. Meanwhile, real-world deployment evidence expands: wastewater treatment facilities across Wisconsin (3.7 to 250 MGD capacity) validate SHAP and LIME for predicting effluent quality across operational scales; clinical deployments of SHAP-based explainability in preterm infant care achieve 92.2% AUC with web-based decision support enabling practitioner adoption. Yet these advances coexist with critical limitations documented in Q1 2026: mechanistic interpretability research reveals a 53-percentage-point knowledge-action gap (models with 98.2% internal feature detection but only 45.1% error correction capability), questioning whether internal understanding translates to actionable fixes; design science studies show XAI explanations produce no significant improvement in user understanding, trust, or usability across test populations (n=344); sparse autoencoders, the canonical mechanistic interpretability tool, underperform linear probes while remaining unused in real engineering workflows. The field's maturity is visible not in triumphalism but in pragmatism: 42% of companies abandoned AI initiatives in 2025 due to compliance gaps; retrofitting explainability costs 2-3x more than building it in from design; and practitioners consistently report that XAI outputs get "lost in translation" with non-technical stakeholders. The XAI market projects 18% CAGR growth ($11.28B in 2025 to $57.9B by 2035), driven by regulation and trust requirements, yet <5% of frontier model computations are actually understood despite $75-150M annual investment in mechanistic interpretability research.

TIER HISTORY

ResearchJan-2018 → Jan-2019
Bleeding EdgeJan-2019 → Jan-2023
Leading EdgeJan-2023 → Jan-2026
Good PracticeJan-2026 → present

EVIDENCE (115)

— Critical practitioner assessment: explainability techniques (feature importance, local explanations, attention, chain-of-thought) show technical limitations and deployment risks in high-stakes domains; candid analysis of what explanations can and cannot deliver.

— Market adoption evidence: XAI grew to $11.74B (2026) from $9.73B (2025), 20.6% CAGR; EU AI Act August 2026 deadline with €35M penalties driving enterprise adoption; only 20% of orgs seeing AI ROI—governance/explainability the missing link.

— Materials science deployment: SHAP-guided feature selection in self-driving laboratories achieving 33% reduction in experimental effort; demonstrates operational efficiency gains from interpretability in production automation.

— Survey of 950 banking executives: 50% report governance/compliance barriers limiting AI performance; only 18% confident in audit readiness; explainability positioned as critical missing link for scaled deployment.

— Manufacturing case study: explainable ML (including SHAP) optimizing 3D-printed composites with quantified outcomes (14.6% tensile strength, 9.2% thermal conductivity improvements, R²=0.937).

— Peer-reviewed clinical deployment: SHAP-interpreted random forest for stroke-associated pneumonia prediction at Huizhou Central People's Hospital (290-patient cohort); demonstrates production integration of interpretability in high-stakes medical decision support.

— SHAP applied to Alzheimer's diagnosis and prognosis using 53,318 participants from NACC-UDS; high accuracy diagnostic models with empirical performance metrics; evidence of clinical-scale SHAP adoption in healthcare ML.

— Enterprise adoption barrier data: 30% cite lack of explainability as top barrier to AI trust; 28% cite model transparency; 46% of planned AI investments stalled due to trust concerns.

HISTORY

  • 2018: Interpretability and explainable AI emerge as a research and early-vendor initiative in response to governance concerns. IBM launches AI OpenScale with built-in explanation and bias detection. Academic research consolidates XAI theory and competing approaches; critical voices question post-hoc explanation viability. Open-source tools (LIME, SHAP) gain traction in data science communities.

  • 2019: Field maturity accelerates with comprehensive XAI taxonomies, practitioner adoption studies, and regulatory compliance drivers. Poland's GDPR implementation mandates explanations in credit decisions, creating enforceable organizational demand. Critical self-assessment reveals validation gaps (< 1% of papers use human subjects). Domain-specific deployments emerge in agriculture and medical imaging. Vendor support consolidates around interpretability as governance capability.

  • 2020: Vendor operationalization accelerates with IBM Watson OpenScale shipping production-ready explainability features. Academic research advances with formal computational complexity theory comparing interpretability across model types and practitioner studies revealing real-world adoption gaps. Inherently interpretable models (Neural-Backed Decision Trees) achieve near-neural-network accuracy, challenging traditional trade-offs. Critical reassessment of popular methods (SHAP, LIME) documents limitations around collinearity and model-dependency; regulatory pressure remains primary adoption driver in finance.

  • 2021: Vendor adoption accelerates with Microsoft and IBM shipping production interpretability features with named enterprise deployments (SAS, EY). Critical gap emerges: real-world evaluations show LIME/SHAP often disagree and fraud analysts achieve lower accuracy with explanations, challenging utility claims. Practitioner demand grows (84% value transparency) but field splits between post-hoc explanations (practical but unreliable) and inherently interpretable models (stronger guarantees but less scalable). Regulatory drivers remain strong; critical voices argue XAI insufficient without structured oversight.

  • 2022-H1: Vendor interpretability features mature across cloud platforms (Azure, IBM, AWS), with integrated tools for fairness assessment and model monitoring. Research reveals method limitations: LIME and SHAP show different bias-variance trade-offs by data density, and produce conflicting feature rankings, challenging practitioner trust. Critical journalism highlights vendor overpromises and unreliability of popular methods. Positive signals from healthcare deployments: physicians use interactive tools (GAM Changer) to debug models for pneumonia/sepsis prediction. Consensus remains that interpretability is necessary but insufficient without human oversight.

  • 2022-H2: Empirical research continues validating XAI limitations across domains—LIME/SHAP show inconsistent performance in bug prediction (software engineering), ICD-10 classification (healthcare), and insurance actuarial models, with domain experts rating SHAP superior to LIME in medical contexts. Critical finding from Nature Communications: public survey shows users prioritize accuracy over interpretability in decision-making trade-offs, signaling adoption friction. Enterprise survey reveals employees recognize XAI as important for governance but identify organizational barriers to deployment. Interpretability established as necessary but insufficient governance practice; real-world adoption remains constrained by method reliability and user preferences for performance.

  • 2023-H1: Vendor interpretability features consolidate around cloud platforms with IBM watsonx.governance GA and Azure/AWS maintaining production tooling. Critical research reveals methodological concerns: May 2023 paper documents SHAP/LIME sensitivity to model choice and collinearity; February 2023 arXiv paper challenges XAI paradigm alignment with human decision-making. Positive domain-specific signals persist: healthcare deployments achieve high accuracy with interactive XAI tools; CMU SEI publishes research-to-practice XAI framework with case studies. Pragmatic consensus solidifies: interpretability necessary for governance but insufficient without domain expertise and human oversight.

  • 2023-H2: IBM watsonx.governance achieves GA in December 2023, cementing vendor platform integration of explainability features. Critical research deepens with comprehensive survey (Which LIME should I trust) documenting LIME's foundational limitations in stability and applicability. Healthcare evidence strengthens: Swedish study of 15,612 hospital admissions shows explainable models match deep learning on clinical outcomes while providing actionable insights. Materials science research challenges assumed interpretability-performance trade-off, finding linear models competitive in extrapolation. Technical advances in neural-based interpretability (CNAM) demonstrate practical progress. Debate emerges over domain-specificity of explainability requirements, moving toward contextual rather than universal adoption.

  • 2024-Q1: Research consolidates evidence of XAI method limitations: LIME/SHAP disagreement documented in geophysics, CVPR workshops identify persistent gaps between academic methods and remote sensing practice. LLM interpretability emerges as new frontier with both opportunities (natural language explanations) and challenges (hallucinations). Industrial deployments expand to predictive maintenance using multiple XAI techniques (LIME, SHAP, PDP, ICE). CFPB regulatory guidance on algorithmic credit decisions creates new organizational demand for explainability in financial services, anchoring practice to compliance requirements.

  • 2024-Q2: Healthcare deployments achieve measurable outcomes—PLOS ONE case study documents 96.89% accuracy with SHAP/LIME in cancer detection. Theoretical critique deepens with position papers from leading institutions arguing current interpretability paradigms fail to ensure faithfulness. Qualitative research reveals critical gaps between technical transparency and contextual user needs in loan decision systems. CFPB compliance analysis highlights accuracy validation challenges in using XAI tools for regulatory adverse actions. Methodological framework (XAIE) proposes systematic tool selection, addressing fragmentation in XAI toolkit landscape. Consensus solidifies: interpretability essential for governance but requires domain-specific validation and alignment with user decision contexts.

  • 2024-Q3: Field matures with mixed signals on capability and limitations. Empirical evaluation across 68,500 model runs (20 datasets) demonstrates no strict performance-interpretability trade-off for tabular data, validating interpretable GAM models as viable alternatives to black-box approaches. Real-world deployments continue: financial services case study applies SHAP/LIME to stock prediction with measurable outperformance. However, critical concerns surface: IBM Research reveals XAI tools (LIME/SHAP) can be exploited for model extraction attacks, exposing security vulnerabilities in interpretable models; practitioner feedback from cybersecurity integration shows XAI outputs 'lost in translation' with non-technical users in production workflows; vendor opinion argues post-hoc explanations fundamentally insufficient due to spurious correlations and fragility. Consensus by end-Q3: interpretability remains essential for governance, but requires careful security assessment, domain-specific deployment, and realistic expectations about explanation quality and user uptake. Industry recognizes both advancing technical capabilities and persisting practical limitations.

  • 2024-Q4: Methodological progress accelerates with focus on accessibility and domain consolidation. MIT research introduces EXPLINGO, converting SHAP visualizations to natural language via LLMs, addressing persistent user comprehension gaps. Healthcare evidence consolidates: systematic review of 23 studies validates LIME/SHAP applications in Alzheimer's detection and clinical decision support. Financial services deployments continue with empirical validation: fraud detection research at XAI 2024 conference compares SHAP, LIME, ANCHORS, and DiCE methods; loan approval case study provides specific performance metrics showing SHAP's superior feature attribution depth (0.38s runtime) versus LIME's speed (0.15s). Vendor support remains stable: Azure Machine Learning, IBM watsonx.governance, and AWS maintain production-grade tooling. Consensus by end-Q4: interpretability established as foundational for governance and regulatory compliance, but success depends on domain-specific method selection, realistic expectations about explanation quality, user alignment, and security considerations in production workflows.

  • 2025-Q1: Vendor ecosystem consolidates with IBM watsonx.governance v2.1.2 adding Evaluation Studio and improved inventory features. Critical assessment surfaces: CSET analysis reveals inconsistent explainability definitions and evaluation gaps (correctness prioritized at 88% versus effectiveness testing at only 4%). Open-source ecosystem remains active: InterpretML library updates to v0.6.10 with ARM support. Academic integration deepens: CMU integrates critical perspectives (Cynthia Rudin's interpretable-model advocacy) and regulatory references (GDPR, ECOA) into ML production curricula. Practitioner analysis documents method reliability challenges: LIME shows 0.72 consistency in high dimensions but recovers to 0.97 with increased sampling; SHAP maintains stability across dimensions. Field maturity signals mixed: continued deployment in healthcare and finance alongside acknowledgment of burnout in interpretability education and persistent gaps between technical transparency and user comprehension. Consensus by end-Q1: interpretability non-negotiable for governance but real-world effectiveness requires standardized evaluation methodologies, domain-specific validation, and sustainable investment in both tools and field culture.

  • 2025-Q2: Vendor maturity continues with IBM watsonx GA product page highlighting explainable governance workflows (Vodafone case study showing 99% improvement in testing efficiency). Technical advancement persists: peer-reviewed research addresses LIME/SHAP limitations in specialized domains (spectroscopy), proposing grouped feature analysis to enhance reliability. Market growth signals strong: XAI market valued at USD 7.94B (2024), projected to reach USD 30.26B by 2032 (18.2% CAGR), with 83% of businesses incorporating AI and XAI as core strategy. Critical philosophical reassessment surfaces: opinion scholarship questions whether AI explainability paradigm is well-founded, proposing reframe toward epistemic soundness in development practices rather than post-hoc transparency. Practitioner analysis consolidates method limitations: LIME/SHAP face documented challenges in explanation stability (randomness-induced variance), faithfulness (surrogate models approximation gaps), and computational cost in production systems. Consensus by end-Q2: interpretability remains essential governance practice, but field increasingly recognizes that method reliability requires domain-specific tuning, philosophical clarity on explainability expectations, and realistic cost-benefit assessment of explanation versus model performance.

  • 2025-Q3: Clinical deployment deepens with longitudinal evidence: 18-month case study of cerebral palsy risk prediction system demonstrates clinicians trust explainability when it enables scrutiny against their own assessments, introducing "Evaluative Requirements" framework. Healthcare research consolidates around desiderata for XAI integration: peer-reviewed analysis identifies three escalating challenges (context-dependent explanations, genuine dialogue, social capabilities) and critiques current methods as too inflexible for clinical needs. Regulatory landscape firms with global financial regulators (OSFI Guideline E-23, FSI analysis) mandating explainability for high-risk AI systems; compliance becomes primary adoption driver in financial services. Methodological clarity emerges: industry analysis confirms accuracy-interpretability trade-off is context-dependent, with practical frameworks guiding model selection across finance, healthcare, and business domains. Field consensus by end-Q3: interpretability foundational for governance, but success requires domain-specific deployment, realistic expectations about explanation quality, and regulatory alignment—not technical transparency alone.

  • 2025-Q4: Ecosystem maturity and realistic limitations consolidate across vendor and research communities. Method validation expands to security domain: peer-reviewed study of SHAP and LIME in intrusion detection systems confirms 97.8% accuracy and explanation stability, extending evidence beyond traditional healthcare/finance contexts. Industry commentary emphasizes adoption barriers: regulatory compliance drives interpretability demand but practitioners highlight persistent gaps—human-in-the-loop remains essential as technical transparency alone proves insufficient. Vendor constraints surface: Azure ML's XAI dashboard lacks support for multi-level classification, revealing deployment gaps in major platforms. Consensus by end-Q4: interpretability indispensable for governance and regulatory compliance, but field maturity manifests as hardened pragmatism—success requires domain-specific method selection, realistic expectations about explanation quality, careful security assessment, and organizational investment in user comprehension. The period marks shift from overconfidence in XAI's explanatory power to disciplined deployment within domain-specific and regulatory constraints.

  • 2026-Jan: Regulatory enforcement and technical innovation reshape the interpretability landscape. EU AI Act enforcement launches (January 2026 for general-purpose AI, August 2026 for high-risk systems), catalyzing enterprise adoption across financial services, healthcare, and compliance workflows. LIME receives technical advancement with LIME-LLM, replacing random token masking with hypothesis-driven LLM-based perturbations for improved NLP explanation fidelity. Mechanistic interpretability emerges from research to production: Sparse Autoencoders enable circuit-level analysis and direct feature decomposition, with production case studies (Rakuten PII detection achieving 500x cost reduction vs. LLM alternatives). Enterprise deployments accelerate: e& and IBM deliver watsonx.governance proof-of-concept within eight weeks; multi-vendor ecosystem (IBM, Palantir, ServiceNow, Azure) integrates AI governance and explainability workflows at scale. Adoption metrics: 60% of large enterprises plan AI governance tool adoption; survey of 600 data leaders shows 7-in-10 GenAI adoption but 75% acknowledge governance/literacy gaps, revealing persistent human capital constraints. Compliance drivers remain dominant; survey data indicates regulatory mandates (not user demand) drive tool selection. Critical assessment: industry voices warn of "interpretability illusions" and black-box compliance failures remain common despite tool availability. Consensus by end-Q1: interpretability essential for regulatory compliance and governance, but real-world adoption constrained by method reliability limitations, organizational expertise gaps, and human-in-the-loop requirements beyond technical transparency.

  • 2026-Feb: Healthcare and vendor ecosystem continue deepening maturity with mixed outcomes. Chronic kidney disease and metabolic obesity prediction studies validate SHAP/LIME utility in clinical applications (88.4-99.1% AUC across datasets), confirming interpretability value in high-stakes medical domains. Geophysics research framework reveals SHAP/LIME can disagree on complex data, calling for causal foundations in interpretability. Vendor ecosystem maturity: Thales and IBM launch integrated AI governance solution with watsonx.governance for EU AI Act compliance. Critical findings surface: financial services adoptions reveal adoption barriers—42% of companies abandoned AI initiatives in 2025 due to compliance gaps, and retrofitting explainability costs 2-3x more than embedding from design. Regulatory scrutiny intensifies with shift from automation baselines to accountability-based transparency mandates. Consensus: interpretability remains central to governance, but growing evidence shows deployment requires domain-specific validation, human oversight integration, and realistic cost-benefit assessment.

  • 2026-Mar to Apr: Real-world deployment evidence consolidates across finance, healthcare, and critical infrastructure. JPMorgan Chase and Wells Fargo implement SHAP at scale (400,000 mortgage applications annually with per-decision explanations); BNP Paribas reduces AML false positive investigation time 35% via SHAP-based alert justification. Clinical deployments expand: preterm infant risk prediction achieves 92.2% AUC with web-based SHAP interface enabling practitioner adoption; wastewater treatment facilities across Wisconsin validate SHAP/LIME for operational prediction across facility scales (3.7-250 MGD). SHAP-guided feature selection in self-driving materials science laboratories achieved 33% reduction in experimental effort, extending interpretability into production automation. Regulatory codification accelerates: DACH banking regulators (BaFin, FINMA) explicitly mandate SHAP/LIME implementation before production for credit, AML, fraud models; the EU AI Act August 2026 deadline (with €35M penalties) is driving enterprise XAI adoption — market reached $11.74B in 2026 (20.6% CAGR from $9.73B in 2025), yet only 20% of orgs report AI ROI, with governance and explainability identified as the missing link. Yet critical research surfaces fundamental limitations: mechanistic interpretability study (400 clinical vignettes) reveals 53-percentage-point knowledge-action gap—models with 98.2% feature detection but only 45.1% error correction; design science research (n=344) shows XAI explanations produce NO significant improvement in user understanding, trust, or usability; sparse autoencoders (canonical mechanistic interpretability tool) underperform linear probes while remaining unused in real engineering workflows. Frontier model understanding remains <5% despite $75-150M annual mechanistic interpretability investment. Consensus by mid-Q2: interpretability matured from optional to mandatory for regulatory compliance, but deployment success depends on domain-specific validation, human oversight integration, and managing stakeholder expectations about explanation quality rather than assuming technical transparency alone enables effective governance.