The AI landscape doesn't move in one direction — it lurches. Some techniques leap from experiment to table stakes in a single quarter; others stall against regulatory walls, technical ceilings, or organisational inertia that no amount of hype can dislodge. Knowing which is which is the hard part. The State of Play cuts through the noise with a rigorously maintained index of AI techniques across every major business domain — classified by maturity, evidenced by real-world adoption, and updated daily so you always know where you stand relative to the field. Stop guessing. Start knowing.
A daily newsletter distilling the past two weeks of movement in a domain or two — delivered to your inbox while the index updates in the background.
Each dot marks the weighted maturity of practices within a domain — hover for a brief summary, click for more detail
AI tools that support individual accessibility needs including transcription, screen reading, alt-text generation, and voice control. Includes personal accessibility assistance and accommodation tools; distinct from accessibility auditing in product design which tests products rather than assisting individuals.
AI-powered accessibility support -- voice interfaces, real-time transcription, alt-text generation, speech recognition -- has matured into a proven practice with GA tooling from major platforms and specialised vendors alike. The question for most organisations is no longer whether these tools work, but how to deploy them reliably. Apple, Microsoft, Amazon, and Google embed accessibility features at the OS level; specialists like Speechmatics and Voiceitt run production workloads across healthcare, broadcasting, emergency services, and financial services. 84% of organisations now prioritise digital accessibility, and 78% use automated testing. The ecosystem is real.
Yet a persistent accuracy ceiling separates routine use from high-stakes dependability. Transcription degrades sharply with accented speech, overlapping voices, and domain-specific terminology -- and AI accessibility overlays detect only 30-40% of WCAG violations. Regulatory pressure is rising: DOJ mandates WCAG 2.1 AA compliance from April 2026, while ADA lawsuits increasingly target organisations relying on automated overlays alone. The practice is mature and accessible, but human-in-the-loop oversight remains essential wherever errors carry real consequences.
Production deployments continue delivering significant productivity gains while exposing persistent reliability and equity barriers. Oxford University Hospitals' clinical evaluation of ambient voice technology documented 88-90% clinician time savings balanced against 37.3% accuracy issues and 44.4% hallucinations in multi-speaker settings. UK ambulance services deployed voice AI across 100% of call intake with 40% time recovery; Speechmatics processes 90M+ multilingual calls monthly across healthcare, lending, and logistics. Healthcare-specific voice AI (Sunoh, Oracle Health Clinical AI) achieved 81,800+ clinical notes generated and 27% efficiency gains at named health systems. These deployments validate production-scale accessibility: the practice has crossed from research/bleeding-edge into mainstream enterprise and public-sector use.
Aging populations show dramatic adoption growth. Amazon Alexa Senior Living reached 400+ U.S. communities with 30% of adults 65+ now regularly using voice assistants (2x growth since 2024), enabling cognitive decline detection through conversation patterns. This 65+ adoption wave extends accessibility beyond traditional disability categories: voice becomes a natural interface when age-related motor and vision limitations emerge. Consumer accessibility tools (Be My Eyes for 340M+ visually impaired users, Voiceitt for atypical speech, Apple Live Speech for speech disabilities) continue maturing into deployment-ready options, enabling independence previously requiring caregiver assistance.
Yet critical production barriers remain unresolved. Speechmatics' founder analysis from March 2026 diagnoses the maturity shift: ASR accuracy is no longer the binding constraint — "the problem is everything else." LLM reliability and multi-speaker diarization now dominate production failures in voice AI pipelines. Real-world speech-to-text accuracy degrades sharply from benchmark claims: vendors advertise 95% on clean audio but deliver 70-80% in production (76% of voice AI builders cite accuracy as the most critical success factor). Whisper large-v3 matches human performance in controlled noise but exhibits a critical failure mode: confabulation rather than silence—the system invents plausible-sounding but false content when uncertain, creating accessibility risk in high-stakes contexts. Audio language models show even more severe reliability gaps: Audio Flamingo 3 exhibits 95.35% attack success rate when probed for hallucinated audio content, demonstrating that audio accessibility applications (captioning, audio description) cannot yet rely on LLM-generated output without human verification. W3C's April 2026 workshop on voice agents identified 8 unresolved standards gaps—including pronunciation markup, hallucination control, and accessibility in immersive contexts—as critical blockers for industry-wide accessibility deployment.
Demographic equity gaps persist despite decade-long vendor investment. Fine-tuning on multilingual data amplified Whisper v3 bias 5-6x in non-English languages; accent support remains unequally distributed across languages; age-related vocal changes (thinner cords, slower processing) degrade accuracy for 55% of adults over 50, yet this demographic gap has received minimal research attention. Recent benchmarking (April 2026) quantified these gaps at scale: global models deliver 3-4x worse accuracy for Indian languages (20-30% WER vs. 4-5% for US English), with code-switching and accent diversity creating multi-layer accessibility barriers for 400M+ Hindi-English bilinguals; researchers document that ASR bias extends across 5 demographic axes with pathological hallucinations on accented speech reaching 9.62% insertion rates. Regulatory momentum accelerates adoption: DOJ Title II WCAG 2.1 AA deadline (April 2026) and European Accessibility Act enforcement (June 2025+) are shifting accessibility from voluntary adoption to legal mandate. However, this creates a deployment paradox: organizations are adopting accessibility tools faster than they can validate reliability, with transcription-only automation risk creating false compliance while actual accessibility gaps persist.
— All India Institute of Medical Sciences (AIIMS) deployed AI-enabled smart glasses to 40 visually impaired people (32 children, 8 teachers) with real-time audio guidance, object recognition, obstacle detection, navigation; targeted distribution with training.
— Independent open benchmark reveals production-scale TTS failures on non-standard text (dates, currencies) across leading providers; silent failures undetectable in standard monitoring create accessibility risk.
— VolenScribe solved cost-accuracy tradeoff in real-world event accessibility (DevFest Ireland): 3.9% WER, 25 languages, enabling technical conference captions for Deaf developers at community rates.
— Kenya's Ministry of ICT launched national AI for Disability Project (May 2026) as multi-stakeholder initiative embedding accessibility in digital infrastructure from design stage; flagship government program.
— Peer-reviewed research on dysarthric speech recognition: frozen audio-language models ignore context, but LoRA fine-tuning achieves 52% WER reduction for speakers with speech disabilities.
— Justice AV Solutions deployed transcription and voice technology at scale in 10,000+ US courtrooms; Voice Lift amplification technology addresses hearing accessibility, real-time captions for Deaf participants.
— Pragmatic clinical trial testing AI-enabled vision screening at primary care clinics for underserved diabetic populations; reduces care access barriers by enabling early detection without specialist referral.
— Market growth to $5.61B (34.28% CAGR), TTS reached 4.8 MOS (human parity), but 52% of organizations encountered deepfakes and 45+ US states enacted legislation—quality mature but trust/regulatory barriers emerging.
2018: AI accessibility tools gained vendor commitment (Microsoft $25M program, Amazon Alexa studies) and academic visibility. Major platforms started offering accessibility features (real-time captioning, auto alt-text, voice interfaces), but limitations in non-standard speech recognition and organizational adoption barriers remain prominent concerns.
2019: Major vendors advanced OS-level accessibility features (Apple Voice Control in iOS 13/macOS Catalina, improved speech pattern detection). Specialized vendors secured funding (Voiceitt $1.5M), and research demonstrated both capability (SciA11y system at 12M+ papers) and persistent gaps (2.4% PDF compliance, insufficient automatic transcription accuracy). W3C identified unresolved accessibility challenges for voice agents including hearing disability gaps and emerging tech requirements.
2020: Voiceitt deployed direct Alexa integration for people with speech disabilities, including pilot deployment at Inglis House (wheelchair community); demonstrated real-world accessibility for motor and speech disabilities. Speechmatics tackled speech recognition accuracy gaps for regional accents and dialects. Educational research confirmed automatic transcription adoption in universities driven by resource constraints, with mixed accuracy results highlighting ongoing reliability challenges.
2021: Healthcare pilots validate voice accessibility (Alexa-based telehealth for heart failure patients); smart home deployments expand (Voiceitt-Alexa integration wins 2021 Speech Industry Award). Educational adoption accelerates (50% of universities use automated transcription, 92% of students report accessibility tools improve learning). Technical progress on bias reduction: Speechmatics reduces racial disparities in speech recognition accuracy. However, critical limitations surface: documented ASR racial bias (35 vs. 19 errors per 100 words across demographics), best-in-class automated accuracy at 84%, requiring ongoing human review for legal/medical/educational contexts.
2022-H1: Platform accessibility reaches mainstream deployment: Microsoft announces Windows 11 Live Captions and Voice Access (May); Apple advances web accessibility at WWDC with SSML and VoiceOver support (June). Production-scale sector deployments: ENCO uses Speechmatics for broadcast captioning; Webex reports 36% transcription accuracy gains. However, social inequity emerges as key barrier: Stanford study shows accented speakers and disadvantaged populations significantly underrepresented among users, with transcription adoption at only 41.3%. Organizational implementation gaps persist despite resources (Google accessibility failures documented); accuracy remains at 84% best-in-class, requiring human review for high-stakes contexts.
2022-H2: Vendor ecosystem expands: Speechmatics launches Real-Time SaaS and 50-language coverage (6.9% WER improvement for Latvian); Voiceitt raises $4.7M to scale atypical speech recognition with 24K+ recordings in Nuvoic pilot. However, critical limitations surface: American Foundation for the Blind documents that AI accessibility overlays fail to address 70%+ of WCAG violations; IT Pro interviews reveal transcription remains unreliable for professional work despite vendor investment. Academic research on dynamic visualization accessibility highlights structural gaps in access for visually impaired data analysts.
2023-H1: Vendor product innovation accelerates: Apple announces Live Speech, Personal Voice, and Point and Speak accessibility features; Speechmatics releases 22-35% accuracy improvements and translation expansion to 34 languages; NCI CaptionSentry achieves 99% usage growth powered by Speechmatics ASR. Enterprise partnerships scale accessibility to professional platforms: Cisco Webex-Voiceitt integration brings non-standard speech recognition to virtual meetings. However, maturity plateau emerges: academic study documents persistent ASR accuracy variation across vendors with streaming quality significantly degraded; vendor analysis confirms transcription barriers (accent bias, vocabulary gaps, 97% accuracy = 30 errors/1000 words); reliability remains insufficient for high-stakes contexts despite ecosystem growth.
2023-H2: Major vendors advance OS-level voice accessibility: Microsoft delivers Voice Access general availability in Windows 11 (Oct), enabling voice-only device control for mobility disabilities with offline functionality. Speechmatics expands multilingual reach, adding 14 languages (48 total) targeting 70% global population coverage within three years. Voiceitt launches Voiceitt 2 (Aug) through RAZ Mobility partnership, enabling spontaneous non-standard speech translation for dysarthria and speech disabilities. However, user-reported development gaps persist: Apple Voice Control shows feature stagnation with buggy vocabulary support and limited customization compared to Siri, signaling maintenance and investment challenges despite critical user reliance. Ecosystem matures toward inclusivity but faces persistent barriers: transcription accuracy limitations, feature development lag in visibility products, and organizational barriers continue constraining broader adoption.
2024-Q1: Vendor accessibility features expand with Microsoft Azure accessibility suite (Seeing AI, audio descriptions, Copilot), Speechmatics real-time transcription, and Voiceitt innovation award recognition. However, research synthesis reveals persistent implementation gaps: systematic review documents overemphasis on visual impairments, gaps in speech/hearing/motor support, and failure to meet accessibility standards. Independent testing shows critical limitations: ASR systems achieve only 50% accuracy on poor-quality audio, Whisper hallucinates in medical contexts, and AI remediation tools fail to address 70%+ of accessibility violations. Accessibility maturity plateaus where technology capability exceeds reliability and implementation quality.
2024-Q2: Accessibility support maturity reveals a deepening deployment paradox. Industry analysis notes AI tools (Be My AI for visual assistance, Deque's axe Assistant) scaling, yet real-world testing of accessibility overlays finds consistent failure modes: misdescription of images, triggering 4,500+ lawsuits in 2023; EU warnings against sole-AI compliance; evidence of user dissatisfaction. Transcription continues as primary barrier: independent assessment documents persistent failures (accuracy degraded with accents, background noise, technical vocabulary), with conclusion that human expertise remains essential for high-stakes contexts (legal, medical). Bias and accuracy concerns documented in speech recognition for disfluent speech populations; ASR systems show consistent accuracy penalties for non-standard speech patterns. The practice plateau deepens: solutions exist, deployment occurs, but reliability, accuracy, and organizational adoption barriers remain binding constraints on broader accessibility transformation.
2024-Q3: Voice and transcription accessibility mature with specific deployment wins offsetting persistent limitations. Voiceitt reports >90% accuracy in pilot with deaf/hard-of-hearing users after training (8% WER), advancing atypical speech recognition toward production reliability; AI-Media's LEXI 3.0 captioning surpasses human quality metrics in live video at reduced cost, signaling mainstream adoption. Speechmatics expands accessibility scope with Flow API (voice interactions for 40M blind, 250M visually impaired users). However, critical barriers persist: independent emergency medicine study finds all four ASR engines fail in medical contexts (medication transcription F1=0.577), validating that accuracy gaps remain binding constraint; real-world user reports document Apple Voice Control bugs rendering tool unusable for mobility-impaired users; healthcare providers and transcription experts conclude human review remains essential for high-stakes medical/legal contexts. The deepening paradox: vendors report production-ready deployments with quality gains, yet independent testing and user reports consistently reveal reliability barriers and accuracy failures that prevent adoption in high-stakes contexts. Accessibility support tools advance in specialized use cases (broadcast, atypical speech) but reliability ceiling remains constraining factor for broader organizational transformation.
2024-Q4: Accessibility support enters mature deployment phase with deepening ecosystem consolidation and persistent accuracy limitations. Speechmatics Flow API achieves general availability with explicit accessibility positioning for 40M blind, 250M visually impaired, and 7% with dexterity issues, signaling vendor confidence in voice-first interaction design. Market data shows strong maturity: 79% of organizations now use AI tools for accessibility tasks (alt-text, code generation); global TTS market projects $9.3B by 2030 (13.4% CAGR), indicating sustained economic investment. However, critical quality barriers persist: AP investigation documents Whisper transcription hallucinations in healthcare settings (1% fabrications, 38% with harmful potential), confirming earlier findings that high-stakes contexts remain unreliable. Voiceitt continues specialized deployment for non-standard speech at CES 2024 with 10+ years operational history. Paradox deepens: vendors deploy production systems with maturity signals (market growth, organizational adoption, feature expansion), yet independent testing and deployment evidence consistently reveal accuracy and reliability barriers constraining broader transformation. Accessibility support reaches mainstream platform parity but accuracy ceiling remains binding constraint on advancement.
2025-Q1: Accessibility support achieves production-scale voice deployments with persistent accuracy limitations constraining organizational adoption. Callers deploys Speechmatics ASR at production scale handling 90M+ multilingual calls across healthcare, lending, logistics, and gaming; Screen Systems delivers live broadcast captioning via Speechmatics ASR for deaf/hard-of-hearing accessibility. Nonprofit adoption accelerates: 96% of nonprofits understand AI capability; 25% adopt AI transcription tools (Otter.ai) for meeting accessibility. Market maturity deepens: voice typing adoption reaches 89% in healthcare (94% of hospitals use speech recognition), 76% in legal, 71% in education; K-12 schools report 43% usage among students with IEPs; global voice interaction user base reaches 4.2B with 23.7% CAGR. However, critical limitations persist: independent research finds transcription accuracy typically 80% in production contexts with significant penalties for accented speech (AAVE accuracy 35% worse than standard English); healthcare, legal, and finance deployments hit accuracy floors requiring human review. User experience gaps continue: disabled users report Apple Voice Control development stagnation and limited customization for non-standard speech patterns. Paradox sustains: vendors deploy at scale with market validation (adoption metrics, organizational deployment, specialized use cases), yet accuracy and reliability barriers remain binding constraints on universal accessibility transformation.
2025-Q2: Enterprise accessibility adoption accelerates with organizational leadership expansion and high-stakes deployment wins. 84% of organizations prioritize accessibility with 80% maintaining dedicated leadership; 40% plan AI adoption; automated accessibility market grows 23% annually to $1.2B. Critical deployment validation: UK emergency services deploy Voice AI across 100% of ambulance calls with 40% clinician time savings, proving production-grade accessibility in high-stakes healthcare. However, accuracy barriers persist: AI tools detect only 30-40% of WCAG violations; legal, medical, and finance domains remain constrained by transcription failures with specialized jargon (e.g., legal terminology misheard, medical terms confused). Voiceitt expands geographic reach with Australian award recognition and 36,000 NDIS-eligible users gaining access. The paradox sustains: enterprise adoption signals strengthen and high-stakes deployments validate maturity, yet accuracy and domain-specific reliability barriers remain binding constraints on universal accessibility.
2025-Q3: Accessibility support ecosystem advances with multilingual expansion and practitioner adoption research. Speechmatics launches bilingual voice models (Mandarin-English, Malay-English, Tamil-English) with 60%+ accuracy gains for code-switching, extending accessibility reach to non-English populations. Enterprise adoption continues: 47% of companies deployed voice AI in 2024 with 30-40% cost reduction in support operations. AFB launches survey on AI adoption across disabled and non-disabled populations, signaling research attention to real-world accessibility tool experience. Practical deployments documented: Be My AI for blind image description, NaviLens for transit navigation (NYC subways, Denver Airport), live captioning, Project Euphonia for atypical speech. However, accuracy barriers remain persistent: transcription accuracy in real-world conditions (95-99.5% in ideal settings) degrades with noise and accents; 70.3% of veterinarians distrust AI transcription; high-stakes domains (legal, medical, finance) remain constrained by reliability concerns. The deepening paradox sustains: multilingual innovation and enterprise adoption metrics indicate market maturity, yet real-world accuracy degradation and practitioner distrust signal that accessibility transformation remains incomplete.
2025-Q4: Platform accessibility advances with vendor feature expansion and escalating legal risks from AI-driven accessibility overlays. Microsoft launches Voice Live API for real-time voice interactions with avatar and emotional intelligence support; Google advances Android accessibility with Expressive Captions, Gemini in TalkBack, dark theme automation, and Voice Access updates. Speechmatics expands real-time transcription with medical terminology improvements and new language models. However, critical barriers intensify: transcription accuracy drops below 80% in real-world conditions (below 60% with noise/accented speech), creating acute liability for legal/medical/finance contexts; 22.6% of 2025 ADA lawsuits (456 of 2,024 filings) targeted websites with AI accessibility overlays, indicating escalating compliance risk from premature automation; AI tools detect only 30-40% of WCAG violations with false positives. The paradox sharpens: platform feature expansion and vendor deployment acceleration signal maturity, yet transcription accuracy floors and escalating lawsuit liability reveal that AI-driven accessibility overlays create legal and UX risks, requiring sustained human oversight and domain expertise.
2026-Jan: Voice AI deployments accelerate with production healthcare scale and regulatory compliance deadlines. Speechmatics reports 30M clinician minutes returned to healthcare via voice AI with specialist medical models achieving 70% error reduction; 9/10 top Norwegian banks deployed voice AI; real-time usage grew 4x year-on-year. AssemblyAI survey of 455+ builders shows 87.5% actively building voice agents with market projected to grow from $2.4B (2024) to $47.5B by 2034. Alt-text automation matures: Climate Policy Review nonprofit achieved 99.2% WCAG Level AA compliance for 11,832 images via local AI models with human-in-the-loop validation. DOJ Title II WCAG 2.1 AA compliance mandate effective April 2026, establishing regulatory requirement for alt-text, captions, keyboard navigation. However, critical limitations persist: AI transcription accuracy remains 95-98% ideal but drops sharply with overlapping speakers and accents; AI tools detect only 30-40% of WCAG violations; human expertise remains essential for high-stakes contexts. Balance sheet: production deployments accelerate and regulatory mandate drives adoption, yet accuracy floors and limited contextual understanding require sustained human-in-the-loop approach.
2026-Feb: Accessibility support shows strong productivity gains alongside emerging accuracy and operational challenges. NHS independent evaluation of ambient voice pilots documents 88-90% clinician time savings but also 37.3% accuracy issues and 44.4% hallucinations in complex multi-voice clinical settings. Independent research (Ada Lovelace Institute) on social work transcription confirms pattern: major time savings paired with clear hallucinations and dialect mishandling, highlighting "speed vs. scrutiny" tensions. Vendor ecosystem maturity deepens: Speechmatics-Boost.ai partnership positioned for regulated industries (9/10 Norwegian banks, 118 municipalities) claiming critical infrastructure status. W3C Editor's Draft on AI accessibility advances standards development. However, reliability concerns intensify: Seamly.ai benchmarking reveals Speechmatics English accuracy regression (WER 69%→77%), and operational analysis documents production failures driven by infrastructure seams (latency, state management, concurrent load) beyond model quality. Balance sheet: deployment momentum accelerates with documented time-savings and organizational adoption, yet accuracy regression signals, hallucination evidence in multi-voice contexts, and operational failure patterns sustain binding constraints on broader accessibility transformation.
2026-Mar: Research exposed demographic blind spots in voice AI accessibility: UC Berkeley documented that 55% of adults over 50 use voice AI but 64% report technology is not designed for them, while Whisper v3 analysis confirmed 5-6x bias amplification in non-English languages through synthetic fine-tuning — two underserved populations previously outside mainstream vendor focus. Simultaneously, the consumer accessibility tool ecosystem was documented at scale: Be My Eyes, Voiceitt, Apple Live Speech, and Lookout collectively serve 340M+ visually impaired users, with the healthcare synthetic voice market projected to reach $3.2B cumulative spend by 2030; AI transcription paired with human review is now established as a standard compliance workflow for the 430M people with hearing loss.
2026-Apr: Production accuracy gaps confirmed at scale: real-world speech-to-text delivers 70-80% accuracy against vendor claims of 95% on clean benchmarks, with 76% of voice AI builders citing accuracy as the most critical success factor. Audio LLM reliability emerged as a new front: Audio Flamingo 3 exhibits a 95.35% hallucination attack success rate, undermining captioning and audio description applications that cannot yet rely on LLM-generated output without human verification. Multilingual equity failures intensified: new benchmarks show no single provider achieves accessibility across languages, global models deliver 3-4x worse accuracy for Indian languages, and ASR bias extends across 5 demographic axes with hallucinations reaching 9.62% insertion rates on accented speech. On the positive side, Adobe Premiere 2026 shipped on-device STT across 55+ languages with near-cloud accuracy, Flockler AI Alt Text reached GA aligned with the April 2026 ADA WCAG 2.1 AA deadline, and a Cornell CHI '26 study with 20 blind/low-vision users quantified the state of play at 56.6% accuracy for dependent queries with 22.2% hallucination rate — useful but not yet reliable without review.
2026-May: Real-world accessibility deployments expand while accuracy and trust barriers remain binding. AIIMS distributed AI-enabled smart glasses to 40 visually impaired individuals (32 children, 8 teachers) with object recognition, navigation, and emergency support; DevFest Ireland deployed VolenScribe (3.9% WER) enabling technical event captioning for Deaf developers. Kenya launched national AI for Disability Project as multi-stakeholder initiative embedding accessibility from design stage; UK survey of 1,032 disabled adults revealed 40% prioritize co-design with disabled people, but 38% remain skeptical of AI's ability to help them—signaling user demand paired with persistent trust gap. TTS market grew to $5.61B (34.28% CAGR) with cloud models achieving 4.8 MOS (human parity), yet 52% of organizations encountered deepfakes with 45+ US states enacting legislation—quality has matured but regulatory and trust barriers emerging. Critical limitation: Async's open benchmark revealed production-scale TTS failures on non-standard text (dates, currencies) with "silent failures" undetectable in standard monitoring. Audio-language models for dysarthric speech recognition improved with LoRA fine-tuning (52% WER reduction), but frozen models continue ignoring clinical context. Balance sheet: May 2026 shows ecosystem maturity with consumer product deployments, national-scale programs, and market growth, yet silent quality failures in production and persistent demographic distrust sustain constraints on broader adoption.