The AI landscape doesn't move in one direction — it lurches. Some techniques leap from experiment to table stakes in a single quarter; others stall against regulatory walls, technical ceilings, or organisational inertia that no amount of hype can dislodge. Knowing which is which is the hard part. The State of Play cuts through the noise with a rigorously maintained index of AI techniques across every major business domain — classified by maturity, evidenced by real-world adoption, and updated daily so you always know where you stand relative to the field. Stop guessing. Start knowing.
A daily newsletter distilling the past two weeks of movement in a domain or two — delivered to your inbox while the index updates in the background.
Each dot marks the weighted maturity of practices within a domain — hover for a brief summary, click for more detail
Frameworks for managing AI-related liability, insurance, and indemnification across deployment contexts. Includes AI-specific insurance products and liability allocation; distinct from general risk management which covers organisational rather than AI-specific liability.
AI insurance and liability frameworks sit in the gap between surging AI deployment and the insurance industry's ability to price, underwrite, and cover the risks that deployment creates. A handful of specialised products now exist -- Munich Re's aiSure platform, Relm Insurance's dedicated AI liability lines, and Chaucer/Armilla AI's Vanguard AI product among them -- but these are early market entrants, not a mature ecosystem. Traditional policies in cyber, Tech E&O, and commercial general liability remain structurally misaligned with AI-specific exposures, and several major insurers have moved to explicitly exclude AI risks rather than attempt to cover them.
The liability question is equally unsettled. The EU's 2024 Product Liability Directive now extends strict liability to software and AI systems, but the companion AI Liability Directive remains unfinished. In the US, regulatory authority is fragmenting across states -- Texas, Colorado, California, and New York each impose distinct obligations -- while federal proposals like the AI LEAD Act lack clear enactment paths. Litigation is filling some of the vacuum, with cases like Mobley v. Workday establishing early precedent on AI vendor accountability. The practice remains experimental: products are shipping, regulations are forming, but standardised frameworks for allocating liability across the AI value chain do not yet exist.
The defining tension in AI insurance is now crystallizing into active market bifurcation. On one side, specialised products continue to launch: Mosaic Insurance and Munich Re's parametric AI performance insurance (EUR/USD 15M coverage), Chaucer and Armilla AI's Vanguard AI (separating AI liability from cyber/E&O with $25M+ limits), Munich Re HSB's SMB-focused AI Liability Insurance (March 2026, addressing 74% SMB AI adoption). On the other side, traditional carriers are systematically withdrawing. In April 2026, Berkshire Hathaway, Chubb, and Travelers won regulatory approval in 80% of states to add explicit AI exclusions to commercial general liability policies. Verisk's ISO Form CG 40 47 01 26 (launched January 2026) now appears in 82% of global commercial policies, creating standardised, industry-wide AI exclusions. A Gallagher survey of 1,200+ global businesses revealed the coverage-deployment gap: 63% have operationalised AI, but only about half of AI-related claims receive any insurance coverage.
The carrier pullback is driven by unpriced liability exposure. Verisk, AIG, Great American, and WR Berkley have filed AI exclusions across commercial general liability, directors & officers, errors & omissions, and employment practices liability lines. Underwriting has bifurcated: companies with documented AI governance (bias testing, human oversight, vendor assessment, prompt controls) receive workable coverage terms; companies without documented governance face exclusions, sublimits, or denial. Real deployment failures now validate carrier caution. A financial services firm that replaced its 12-person QA team with an AI testing system suffered a $6M loss in a single day when the AI generated a faulty discount code that reset prices to zero. A logistics firm's AI agent gradually drifted in freight rate negotiations, accumulating $2.1M in suboptimal contracts over 11 weeks; the claim was denied as "gradual degradation" not covered by the policy's security-breach trigger, and the company lacked audit logs to demonstrate governance. These are not hypotheticals—they are deployment stage failures reshaping underwriting discipline in real time.
Litigation is accelerating faster than insurance markets can price. Gallagher Re and MIT documented 978% growth in GenAI-related lawsuits from 2021-2025, with cumulative filings exceeding 700; Gartner forecasts 2,000+ 'death by AI' claims by end of 2026. Federal courts are now ordering disclosure. The Estate of Lokken v. UnitedHealth case reached discovery phase in March 2026, with judges ordering UnitedHealth to disclose internal records on its nH Predict algorithm—demonstrating that courts now treat AI-assisted insurance coverage decisions as subject to the same transparency and accountability standards as any professional service. Coverage disputes themselves are becoming litigation. LION Specialty documented three unresolved E&O and cyber coverage issues: (1) Lokken v. UnitedHealth created a professional/product boundary question with no settled answer (90% reversal rate on appealed denials); (2) deepfake wire fraud cases (Arup lost $25M) show jurisdictional splits on whether impersonation falls under cyber, crime, or social engineering coverage; (3) model poisoning (adversarial training-data manipulation of fraud detection systems) has no coverage trigger in standard policy forms. Surgical AI provides a concrete high-stakes example. The FDA cleared 221 AI medical devices in 2023, 97% via 510(k) expedited pathway without new clinical trials. An Acclarent AI-enabled sinus surgery device integrated in 2021 went from 7 pre-AI malfunction reports to 100+ post-AI reports with 100+ adverse events, including 10 injuries. Two patients suffered strokes from carotid artery damage when the AI misidentified instrument location inside patients' heads; surgeon records showed the AI provided no warning of proximity to critical anatomy. Insurance coverage remains opaque: does GL cover the clinic, does product liability cover the device manufacturer, does medical malpractice cover the surgeon, or does each party's AI exclusion endorse create a liability gap the patient must absorb? These are not edge cases—they are the deployment reality defining market readiness.
Regulatory frameworks are fragmenting. Texas's TRAIGA took effect in January 2026; Colorado's AI Act follows in June. Kansas legislative briefing (March 2026) documented that 84% of US health insurers use AI for utilization management and 71% for prior authorization, concentrating AI-driven coverage decisions across a fragmented regulatory landscape. The EU's Product Liability Directive (effective December 2024) presumes defectiveness when AI systems fail to comply with AI Act requirements, shifting evidentiary burdens to manufacturers. But federal liability authority remains fragmented: the White House proposes preemption, the Senate's TRUMP AMERICA AI Act proposes product liability, Colorado proposes relative fault allocation. Across the US, 45 states have introduced 1,561 AI bills. This regulatory patchwork leaves underwriters without consistent rules for pricing or allocating risk across jurisdictions. Lockton Re's analysis captures the structural problem: commercial lines categories—CGL, cyber, E&O—were not designed for AI exposures and remain fundamentally misaligned with them. The gap between litigation velocity, regulatory fragmentation, carrier pullback, deployment failures, and the absence of settled liability frameworks defines this market's bleeding-edge status in April 2026.
— Major carriers executing AI exclusion strategy with 80% state regulatory approval across commercial policies (employee discrimination, IP violations, autonomous damage); exclusions in effect early 2026, brokerages flagging deployment coverage gaps.
— Technology law firm analysis of January 2026 inflection: Verisk CG 40 47/48 endorsements, major carriers adding AI exclusions, underwriting now requires documented AI governance, bias-testing, human oversight, and vendor assessment for coverage approval.
— Structural insurance gap analysis: all four AI vendors cap liability at 12 months of fees; Verisk/ISO exclusions effective Jan 2026 allow carriers to drop coverage; 137% YoY GenAI lawsuit growth creates new uninsurable liability class flowing to deployers.
— Gartner forecasts 2,000+ 'death by AI' claims by end 2026; juries hold software to higher standard than human judgment; insurance implications extend across health, life, and all industries deploying AI-driven decision systems.
— Detailed mapping of Jan 2026 market bifurcation: ISO Form CG 40 47 01 26 in 82% of policies; W.R. Berkley absolute AI exclusion; four specific coverage gaps emerging (CGL, D&O, E&O, EPLI) from endorsement-schedule changes without explicit notification.
— Detailed structural analysis of ISO exclusions (CG 40 47/48/35 08) with four documented coverage gaps including surgical AI (100+ adverse events post-integration vs 7 pre-AI, strokes from carotid misidentification) and healthcare/warehouse robotics silence.
— Comprehensive market mapping of AI insurance (embedded/endorsed/standalone products) with detailed negative signal: logistics firm's AI agent drift caused $2.1M uninsured loss; policy excluded gradual degradation without security-breach trigger and lacked behavioral audit logs.
— ISO standardized three AI exclusion codes (CG 40 47, CG 40 48, CG 35 08) effective Jan 2026; carrier bifurcation between governed vs. autonomous AI; case study: financial services firm's AI testing replacement caused $6M loss in single day (faulty discount code).