Stanford CS547 HCI Seminar | Winter 2026 | Does GenAI Work in Education?

| Podcasts | February 13, 2026 | 5.63 Thousand views | 56:54

TL;DR

This seminar argues that GenAI's effectiveness in education hinges on 'knowledge engineering'—the systematic mapping of expert cognitive processes—to ensure high-fidelity, personalized feedback. A randomized trial demonstrates that TAs using AI suggestions based on detailed reasoning rubrics produced significantly better student learning outcomes than human-only feedback.

📊 The Mixed Reality of GenAI in Education 2 insights

Widespread adoption meets uncertain outcomes

While 70-85% of college students use GenAI for homework and hundreds of EdTech startups have emerged, studies show AI use can reduce learning retention when removed and decrease brain activity during writing tasks.

Variable quality in AI tutoring

Research indicates 30-50% of AI-generated tutoring hints contain quality issues, though carefully designed systems have outperformed active learning in college physics and improved K-12 math tutoring when providing real-time suggestions.

🧠 Knowledge Engineering & Cognitive Fidelity 2 insights

Solving Bloom's Two Sigma Challenge

Cognitive tutors from the 1990s achieved two standard deviation gains over conventional instruction by encoding expert production rules, demonstrating that AI must model human reasoning processes rather than just generate answers.

Granular decomposition enables precision

Breaking skills into specific steps and anticipating precise misconceptions—such as distinguishing between common and least common denominators—allows AI to provide targeted feedback and deliberate practice instead of generic hints.

📝 Case Study: Feedback Writer Trial 2 insights

Significant improvement in writing quality

In a randomized trial of 360 Econ 101 students, AI-mediated feedback produced revised essays with an effect size of 0.5, equivalent to moving a student from the 50th to 70th percentile compared to human-only feedback.

High-fidelity rubrics drive success

The system used checklists representing expert reasoning steps—such as identifying specific decision-makers harmed by externalities—to generate concrete hints without revealing answers, with AI judges assessing quality at 85% expert accuracy.

Bottom Line

To make GenAI work in education, developers must engineer detailed cognitive models that externalize expert reasoning into high-fidelity rubrics, enabling targeted feedback on learning processes rather than just evaluating final outputs.

More from Stanford Online

View all
Stanford CS153 Frontier Systems | Andreas Blattmann from Black Forest Labs on Visual Intelligence
1:01:14
Stanford Online Stanford Online

Stanford CS153 Frontier Systems | Andreas Blattmann from Black Forest Labs on Visual Intelligence

Andreas Blattmann, co-founder of Black Forest Labs and co-creator of Stable Diffusion, argues that visual intelligence represents the critical next frontier for AI, requiring a fundamental shift from text-centric unimodal models to multimodal systems trained on 'natural representations' (video, audio, physics) to unlock true reasoning, robotics capabilities, and higher intelligence.

5 days ago · 9 points