The AI landscape doesn't move in one direction — it lurches. Some techniques leap from experiment to table stakes in a single quarter; others stall against regulatory walls, technical ceilings, or organisational inertia that no amount of hype can dislodge. Knowing which is which is the hard part. The State of Play cuts through the noise with a rigorously maintained index of AI techniques across every major business domain — classified by maturity, evidenced by real-world adoption, and updated daily so you always know where you stand relative to the field. Stop guessing. Start knowing.
A daily newsletter distilling the past two weeks of movement in a domain or two — delivered to your inbox while the index updates in the background.
Each dot marks the weighted maturity of practices within a domain — hover for a brief summary, click for more detail
AI that generates preliminary radiology reads autonomously, with radiologist confirmation for final diagnosis. Includes automated report generation and critical finding alerting; distinct from assisted detection which highlights findings for human interpretation rather than producing reports.
A handful of forward-leaning organisations have moved autonomous preliminary radiology reads from pilot to production, but the broader field has not followed. Systems that generate draft reports without radiologist input -- typically for high-volume screening of chest X-rays, mammography, and visa-processing workflows -- show genuine efficiency gains where deployed. Radiology Partners processes 40-50 million cases annually through RADPAIR, cutting report turnaround from 15-20 seconds to 2-5 seconds with a 12% error reduction. A peer-reviewed study of 100 complex cases found turnaround dropped from 6.1 to 3.43 minutes while accuracy and confidence scores improved significantly. These results are real, but they come from a vanguard. Most health systems have not adopted autonomous reads, and those that have tend to concentrate use on routine normal-case screening rather than complex diagnostic interpretation.
The gap between regulatory momentum and clinical readiness defines this practice's current tension. The Radiology Research Alliance published in March 2026 that human-AI collaboration produces strongest outcomes, signaling field consensus that full autonomy is not the primary path forward. Over 1,000 FDA-cleared AI devices exist, 80% in radiology, yet a scoping review of 67 LLM studies found 79% were single-centre proof-of-concepts with diagnostic accuracy ranging from 16% to 86%. Algorithms perform measurably worse in diverse patient populations, and fundamental questions around accountability -- who is liable when an autonomous read is wrong -- remain unresolved across jurisdictions. Workflow integration compounds the challenge: a Signify Research survey of 150 healthcare organisations rated seamless PACS/RIS integration a 9-10 out of 10 priority, and poor integration remains a deal-breaker for buyers. The technology works in controlled settings. Scaling it safely is the unsolved problem.
Deployment is concentrated among a few high-volume operators. Radiology Partners, the largest US radiology practice with 3,900+ radiologists, runs RADPAIR in production with documented 25% time savings per case. In India, autonomous TB screening at sites in Chhattisgarh increased case notifications by 80% with 162 confirmed cases. A systematic review of 11 studies concluded that autonomous chest X-ray triage is ready for clinical implementation, finding a 42.3% autonomous triage rate with 97.8% sensitivity across datasets exceeding 500,000 real-world cases. Danish mammography screening with 248,000+ images achieved a 48.8% workload reduction while maintaining cancer detection rates. These are meaningful deployments, not demos.
The commercial ecosystem is consolidating rapidly. RadNet acquired Gleamer in March 2026 for up to $270M, merging it into DeepHealth with 26 FDA-cleared devices and 2,700+ customer contracts across 50 countries. Qure.ai's qXR now covers 26 FDA indications, and newer entrants like Report Rad AI claim 60-95% faster reporting across CT, MRI, X-ray, and ultrasound. ARA Health Specialists (70+ physicians, 13 hospitals, 100k studies/month) deployed Rad AI Reporting achieving 20% reporting time reduction with 79% of radiologists showing efficiency gains. European momentum accelerates: United Imaging Intelligence unveiled CE-certified uAI Insight Image-to-Report agents generating structured preliminary reports from CT and brain MRI across healthcare sites in Poland, UK, Italy, Romania, and Bosnia. European radiologist adoption of AI tools has grown from 20% in 2018 to 48%, though this figure captures AI-assisted detection broadly, not autonomous reads specifically. LLMs are entering the toolchain: fine-tuned Llama-3-70B achieves F1 0.780 on error detection, outperforming GPT-4, and structured tasks like report simplification exceed 94% accuracy. Diagnostic performance remains unreliable, with published accuracy spanning 16% to 86%.
What separates leading-edge from good-practice here is the infrastructure and governance gap. A Brisbane hospital study catalogued 82 distinct barriers to sustained adoption, including performance inconsistency, medicolegal uncertainty, and weak communication between AI vendors and clinical teams. In response, Radiology Partners and Stanford's AIDE Lab have partnered to build validation and safety monitoring frameworks. The American College of Radiology has launched ARCH-AI (recognition for quality assurance best practices) and Assess-AI (national registry for post-deployment monitoring), signaling field consensus on governance maturity requirements. Yet these frameworks are nascent. Independent analysis distinguishes validated AI-assisted detection (supported by randomized trial evidence like MASAI across 100,000+ patients) from unvalidated autonomous-only reads that remain in pilot maturity. Multi-agent architectures for collaborative AI-assisted reporting show efficiency gains (7.8% average, 18.3% for complex cases) when radiologists review and modify AI proposals, but fully autonomous preliminary reads lack equivalent clinical trial validation. The FDA pipeline is prolific (1,039 cleared devices and counting), but regulatory clearance and clinical readiness remain different things.
— Regulatory ecosystem signal: radiology represents 75% of FDA-cleared clinical AI tools (1,357 total devices); FDA shifts toward lifecycle oversight for continuously-evolving autonomous AI systems.
— Peer-reviewed proof-of-concept (Insights Imaging) showing AI-assisted collaborative workflow (radiologists review and modify AI proposals) achieves 7.8% average efficiency gain with 18.3% improvement for complex cases without quality degradation.
— Critical analysis distinguishing validated AI-assisted detection (MASAI randomized trial, 100,000+ women) from unvalidated autonomous-only reads; documents evidence gap between regulatory approval and clinical trial validation for autonomous preliminary reads.
— Systematic pre-deployment assessment framework evaluating 13 AI models across 88,645 clinical exams; provides structured methodology for task-specific value prediction before autonomous implementation.
— ACR launches standardized governance framework (ARCH-AI recognition program, Assess-AI national monitoring registry) addressing maturity assessment and post-deployment performance benchmarking for radiology AI—signals field consensus on governance requirements.
— Rad AI Reporting (part of Rad AI Omni suite) deployed across 8 of 10 largest US private radiology practices; generates comprehensive autonomous impressions with 16–23% radiologist modification rate and 5% error correction rate.
— Gleamer's €230M acquisition by RadNet confirms autonomous draft reporting already deployed across Europe at scale (700+ customer contracts in 44 countries, €30M ARR expected 2026); positions merged DeepHealth as largest radiology AI provider worldwide.
— FDA Breakthrough Device Designation for Cognita CXR, a vision-language model generating autonomous preliminary chest X-ray reports with 16-65% detection improvement and 18% efficiency gain.
2023-H1: Emerging commercial deployment with multiple vendors at scale (3100+ sites, 10.7M scans processed by single vendor). Vendor landscape includes 179 CE-certified products by March 2023, with 67% having peer-reviewed evidence though evidence weighted toward diagnostic accuracy rather than clinical impact. Technical challenges documented: data imbalance in training sets, inadequate evaluation metrics, specific failure modes in non-standard imaging. Safety concerns noted including radiographer override rates and position-dependent accuracy degradation.
2023-H2: Substantial real-world deployment acceleration across multiple health systems and use cases. UK NHS trusts (Frimley, Greater Glasgow) deployed qXR for large-scale triage with 99.7% normal-case accuracy and 58% workload reduction potential. India-based TB screening program demonstrated 15.8% incremental yield improvement with autonomous systems. Teleradiology platforms (InHealth) integrated autonomous triage for critical finding alerting. Major commercial deployment at RSNA 2023: Microsoft/Nuance launched PowerScribe Smart Impression on platform used by 80% of radiologists. Adoption sentiment among clinicians remained positive (78% of Chinese residents surveyed support AI embrace), though replacement concerns persist (30% feared workforce reduction).
2024-Q1: Consolidation of commercial adoption and emergence of safety frameworks. Providence health system deployed PowerScribe autonomous reporting in largest US platform rollout, signaling mainstream health system adoption. Research momentum continues with IEEE review documenting methodological advances in automatic report generation. Emerging LLM approaches (ChatGPT/GPT-4) explored for autonomous reporting, though professional societies (ACR, CAR, ESR, RANZCR, RSNA) established formal evaluation and monitoring frameworks for autonomous AI tools, emphasizing rigorous safety assessment requirements alongside deployment.
2024-Q2: Major vendor expansion and critical deployment-readiness assessment. Microsoft Azure launches Radiology Insights preview service, signaling major cloud vendor entry into autonomous radiology tooling ecosystem. GPT-4 demonstrates capabilities matching radiologists on error detection (82.7% accuracy, lower cost and faster than human review). Independent academic assessment from Imperial College and Royal College of Radiologists documents widespread implementation barriers post-regulatory approval: reliability validation, accountability, trust, and safety governance gaps limit real-world adoption despite regulatory clearance. Asian Oceanian Society of Radiology formalizes regional adoption guidance. Open-source momentum continues with 100+ papers and tools curated in community repositories.
2024-Q3: Consolidation of clinical evaluation frameworks and LLM maturation. RSNA and MICCAI publish joint expert consensus on deployment barriers emphasizing trust, reproducibility, and accountability frameworks. LLM-based approaches demonstrate specific clinical capabilities: GPT-4 for synoptic report generation (F1 0.997), Claude-2 for RADS categorization (57% accuracy with prompting), and patient-friendly report generation (improving understanding scores). Radiologist adoption sentiment remains strong (75%+ engagement in AI practices in major markets). NHS Greater Glasgow & Clyde initiates prospective stepped-wedge clinical trial (RADICAL) for rigorous qXR evaluation across 24 months, signaling major health system commitment to evidence-based autonomous read validation. Systematic reviews synthesize efficiency gains (30-40% reporting time reduction) across diverse deployments, though net clinical outcome questions persist.
2024-Q4: Infrastructure consolidation and human factors maturation concerns. Largest US radiology practice (Radiology Partners, 3,900+ radiologists) partners with generative AI vendor for co-developed autonomous reporting, signaling shift toward strategic health system integration. Large-scale deployment validation continues: 1.3M chest X-rays processed across 33 UAE visa screening centers documented in peer-reviewed publication. Yet critical limitations emerge: multi-site prospective study documents over-reliance risk when AI provides local explanations (even incorrect ones), and practitioner surveys show mixed adoption sentiment with 100-radiologist cohort expressing concerns about reliability, job displacement, and ethical implications. JMIR viewpoint highlights persistent research-to-practice gap despite 190+ FDA-approved radiology AI devices, suggesting deployment maturity lags regulatory approvals.
2025-Q1: Broad hospital adoption and implementation gap evidence. Danish large-scale study demonstrates AI-driven mammography screening with 248,000+ images achieving 48.8% workload reduction while maintaining cancer detection, expanding evidence beyond chest X-ray modality. Pew Charitable Trusts survey (2022 data) documents 44% of US hospitals adopted AI imaging tools, yet critical implementation gaps persist: only 26% piloted before rollout, 34% lack comprehensive validation information, 31% lack monitoring protocols. Over 200 EU-approved radiology AI tools exist with limited real-world adoption, indicating continued research-to-practice gap. Patient acceptance studies show AI-simplified reports improve understanding but patients prefer physician delivery. Major health systems continue infrastructure investments despite unresolved questions about clinical effectiveness versus workflow efficiency gains.
2025-Q2: Continued deployment expansion and platform consolidation. UH Cleveland activated qXR for autonomous lung cancer detection on chest X-rays, demonstrating sustained adoption by major US health systems. Intelerad-RADPAIR partnership combines workflow orchestration with generative AI-driven reporting, signaling integration of autonomous reporting into mainstream clinical informatics platforms. Radiologist interviews reveal variability in AI monitoring practices and implementation approaches. Research continues on report generation advancements and faculty/trainee adoption sentiment.
2025-Q3: Vendor ecosystem consolidation and agentic AI emergence. RADPAIR-Fireworks partnership demonstrates production-scale autonomous reporting infrastructure at Radiology Partners (40-50M cases annually) with specific metrics: 15-20s reduced to 2-5s report turnaround, 25% time savings per case, 12% error reduction. RADPAIR expands geographic reach via AdvaHealth partnership into Asian healthcare markets. Commentary in Diagnostic and Interventional Radiology explores agentic AI for autonomous radiology workflows (RadGPT example for CT scanning). Academic research examines feasibility and barriers of autonomous CXR reporting in UK context, highlighting unresolved accountability framework gaps, regulatory challenges (IR(ME)R, GDPR), and need for post-market surveillance. RSNA covers autonomous report generation in neuroradiology using LLMs (GPT-4), discussing benefits and limitations including hallucinations and generalizability challenges.
2025-Q4: LLM performance maturation and validation gap emergence. Peer-reviewed study demonstrates efficiency gains in semi-automated AI reporting: 6.1-to-3.43 minute turnaround reduction on 100 complex cases with improved accuracy (3.81→4.65/5.0) and confidence (3.91→4.67/5.0, p<0.0001). Scoping review of 67 LLM studies shows strong structured-task performance (>94% accuracy in report simplification) but inconsistent diagnostic performance (16%-86%) with 79% single-center proof-of-concept designs. European adoption expands to 48% of radiologists using AI (up from 20% in 2018); 115+ FDA-approved algorithms by mid-2025. Agentic AI architectural shift: RADPAIR launches PAIRsdk developer framework with industry coalition (Fovia, Interlinx, deepc, Intelerad) planning open-source standards for voice-first autonomous workflows in 2026. Critical expert assessment documents widening validation gap: 100+ tools exhibited at RSNA 2025 lack rigorous clinical validation; algorithms perform significantly worse in diverse populations. Fundamental accountability and clinical outcome questions persist unresolved.
2026-Feb: Continued deployment expansion and regulatory acceleration. Autonomous TB screening in India demonstrates 80.21% increase in case notifications with 162 confirmed cases (44.63% positivity) at deployment sites in Chhattisgarh tribal population. FDA clearing accelerated with 56 new radiology AI devices (1,039 total devices, 80% of all FDA-authorized AI), signaling sustained regulatory momentum. Systematic review of 11 studies concludes autonomous chest X-ray triage systems ready for clinical implementation, with weighted average 42.3% autonomous triage rate and 97.8% sensitivity across real-world datasets (500K+ cases). Implementation barriers evidence: Brisbane hospital study identified 82 barriers and 33 enablers post-deployment using NASSS framework, with sustained adoption constrained by performance inconsistency, weak communication, and medicolegal uncertainty. Industry initiatives: Radiology Partners and Stanford AIDE Lab partnered to develop evidence-based validation and safety monitoring frameworks for autonomous AI tools. Practitioner perspective documents job displacement concerns and malpractice exposure constraints on adoption.
2026-Feb: LLM capability maturation and infrastructure framework development. Blinded evaluation studies compare LLM-generated to radiologist-written reports across clinical relevance and accuracy, advancing evidence on LLM autonomous report quality. Research on fine-tuned Llama-3-70B demonstrates F1 0.780 error detection capability, outperforming GPT-4 (0.683), signaling LLM performance gains in structured report tasks. Technical research on web-based automated chest X-ray report generation systems achieves BLEU-4 0.482 and ROUGE-L 0.718, advancing autonomous generation methodologies. Industry frameworks emerge: Signify Research survey of 150 healthcare organizations identifies seamless workflow integration as 9-10/10 priority and integration barriers as deal-breakers; SATMED and HealthManagement publish frameworks documenting transition from scattered pilots to core infrastructure, citing 1,000+ FDA-cleared AI devices (75% radiology), cloud-native platforms, and agentic AI orchestration requirements. Commercial tool proliferation: Qure.ai expands qXR-Detect to 26 FDA indications; Report Rad AI launches claiming 60-95% faster reporting with 10K+ cases. Regulatory moment: ABR evaluates cautious AI adoption for internal functions while assessing competency frameworks for clinical tool use. Infrastructure maturity signal: adoption frameworks systematize integration, governance, and sustainability challenges underlying "pilot purgatory" barriers.
2026-Mar: Market consolidation accelerated with RadNet acquiring Gleamer for up to $270M and merging it into DeepHealth (26 FDA-cleared devices, 2,700+ customer contracts across 50 countries), signalling that scale and device breadth are becoming competitive moats. Deployment evidence continued broadening: ARA Health (13 hospitals, 100k studies/month) achieved 20% reporting-time reduction with Rad AI across 79% of radiologists; United Imaging Intelligence unveiled CE-certified uAI Image-to-Report agents generating structured CT and brain MRI preliminary reports across five European countries; and a Ghana clinical study found autonomous AI TB screening achieved 91% accuracy versus 86% for radiologists in a resource-limited setting. The Radiology Research Alliance published multi-institutional consensus in March 2026 that human-AI collaboration produces stronger outcomes than full autonomy, reflecting field caution about replacing radiologist sign-off despite rising deployment velocity.
2026-May: Commercial adoption deepened with Rad AI Omni deployed in 8 of 10 largest U.S. private radiology practices, and the RadNet-Gleamer merger creating 700+ customer contracts with autonomous draft reporting as a core capability. ACR launched ARCH-AI and Assess-AI governance programs — formal quality assurance and post-deployment monitoring registries — signalling field consensus that governance infrastructure now defines the frontier rather than algorithm capability itself. Critical evidence clarified the autonomous-versus-assisted boundary: independent analysis documented the evidence gap between validated AI-assisted detection (MASAI RCT, 100,000+ patients) and autonomous-only reads lacking equivalent clinical trial validation, while FDA moved toward lifecycle oversight for continuously-evolving AI systems with radiology representing 75% of all 1,357 cleared devices.