The AI landscape doesn't move in one direction — it lurches. Some techniques leap from experiment to table stakes in a single quarter; others stall against regulatory walls, technical ceilings, or organisational inertia that no amount of hype can dislodge. Knowing which is which is the hard part. The State of Play cuts through the noise with a rigorously maintained index of AI techniques across every major business domain — classified by maturity, evidenced by real-world adoption, and updated daily so you always know where you stand relative to the field. Stop guessing. Start knowing.
A daily newsletter distilling the past two weeks of movement in a domain or two — delivered to your inbox while the index updates in the background.
Each dot marks the weighted maturity of practices within a domain — hover for a brief summary, click for more detail
AI that provides detailed developmental feedback on student work, going beyond grades to guide improvement. Includes specific improvement suggestions and learning pathway recommendations; distinct from automated grading which scores rather than develops.
AI-generated formative feedback works well enough to deploy -- but not well enough to trust on its own. That tension defines the practice's leading-edge status. Forward-leaning districts and vendor platforms have moved from pilots to GA products, proving that LLMs can produce structured, actionable feedback on student work at a speed no human team can match. The value proposition is real: teachers reclaim hours, students get faster turnaround, and institutions can scale feedback across large cohorts. Yet the empirical record consistently shows that AI feedback remains inferior to human feedback on nuance, tone calibration, and adaptive support for struggling learners. Students, meanwhile, tend to overestimate AI feedback quality -- a source-credibility bias that compounds the accuracy problem. Production reliability adds another layer of risk; repeated model-drift and sycophancy incidents have forced rollbacks in deployed systems. The result is a practice that functions as a "teacher-amplifier" -- AI drafts feedback, humans validate it -- rather than an autonomous replacement. Most institutions have not yet adopted this approach, and those that have maintain mandatory human review. The question facing the field is no longer whether AI can generate feedback, but whether the quality and consistency gaps can close fast enough to justify the integration cost.
A growing cohort of vendor platforms and early-adopter institutions are operationalizing formative feedback systems at scale. Formative's Luna AI assistant, generally available since August 2025, has reached broad distribution across 90% of US school districts with 6+ billion student responses processed. Instructure's Canvas LMS released IgniteAI (April 2026), integrating rubric generation and feedback drafting into its core grading workflow -- evidence of ecosystem maturity as major LMS vendors embed formative feedback tools natively. Microsoft Teams Assignments ships AI Feedback Suggestions with explicit responsible-deployment guidelines. LearnWise reports 84% student preference for AI-generated feedback (40,000+ student sample) with deployment across Canvas, Moodle, Brightspace, and D2L. Wichita Public Schools (47,000+ students) and UK institutions piloting through the Jisc AI Assessment program demonstrate formative assessment as the primary deployed use case. These deployments maintain human review as mandatory workflow -- teachers review and edit all AI suggestions before students see them -- confirming the "teacher-amplifier" model as the operational standard, not an interim step.
The empirical picture remains stubbornly mixed despite operational scaling. April 2026 research confirms positive cognitive effects: a large-scale Frontiers study (n=1,079) shows AI precision feedback significantly enhances thinking ability (p<0.001) with intrinsic value identification mediating 32% of learning gains. Systematic reviews on L2 writing (55 studies) and automated feedback in HE (10 studies) identify collaborative tool use, custom design, and pedagogical scaffolding as critical success factors -- suggesting tool capability alone is insufficient without institutional redesign. Yet deployment quality remains contingent on assessment infrastructure. A 654-student peer-review study found half of participants flagged AI feedback inaccuracies, with only 6% preferring AI feedback alone. Experimental evidence from Chinese high schools showed AI-enabled visual feedback improving achievement but also increasing test anxiety. Reliability gaps persist across domains: Washington State University's study found ChatGPT accuracy on scientific hypotheses only ~60% (barely better than random chance) with 73% consistency across identical prompts. An MIT practitioner documented five spurious ChatGPT suggestions for every useful correction on feedback tasks.
The critical finding from March 2026 research: assessment design determines whether AI feedback drives learning or merely accelerates autopilot answer-completion. Qualitative evidence reveals that students with visible future accountability -- in-person exams requiring genuine understanding -- use AI feedback for reasoning and self-testing; those without accountability use it on autopilot. A 50-scholar synthesis identifies scalability benefits but forewarns of student dependency and quality consistency barriers; OECD research documents the performance-learning paradox: students write better essays with AI feedback but retain 80% less content, attributed to "fast AI" eliminating productive cognitive friction. A systematic review of 83 automated feedback studies confirms the field remains immature with heterogeneous results and inconsistent implementation. April 2026 Stanford research documents systematic demographic bias: high-achieving and White students receive developmental feedback while ELL/Hispanic students receive grammar-focused feedback, and low-achieving students experience feedback withholding. The trend line remains stalled: adoption has plateaued around the teacher-amplifier model, with quality consistency, equity gaps, assessment design contingency, and ROI realization blocking the path to broader uptake. The field consensus is clear: formative feedback generation succeeds only when embedded in pedagogically sound assessment systems with human oversight, not as a standalone tool.
— William & Mary $300K GRI Accelerate grant-funded K-12 deployment of AI peer buddies that prompt reasoning and reflection rather than providing answers; focuses on critical thinking, equity, and teacher decision-making.
— Peer-reviewed white paper proposing theoretical reframing of sycophancy toward reflective responses; directly addresses feedback system design that acknowledges uncertainty and supports user autonomy.
— Meta-analysis of 72 studies showing AI teaching interventions yield significant positive effects (g_p=0.586) on effectiveness; clearly identifies boundary conditions and moderating factors enabling heterogeneous outcomes.
— Peer-reviewed research (Assessment & Evaluation in Higher Education, March 2026) with 10 principles for effective AI feedback in higher education; documents that students trust human feedback more and AI requires relational design.
— Real school district deploying adapted AI Assessment Scale framework across 20+ countries; teachers using framework to guide conversations about AI use, academic integrity, and demonstrations of learning.
— Stanford study (LAK best paper nominee, April 2026) documenting systematic demographic bias in AI writing feedback across 4 models; different tone and pedagogical expectations by student race, gender, achievement level.
— Expert analysis synthesizing research on feedback timing, specificity, and AI capability limits; emphasizes teacher judgment remains essential on creative and collaborative assessment despite AI routine assessment reliability.
— LMS-integrated product with 84% student preference for AI-generated rubric-aligned feedback; maintains teacher review and edit workflow; integrated across Canvas, Moodle, Brightspace, D2L demonstrating ecosystem maturity.