Stanford CS25: Transformers United V6 I From Representation Learning to World Modeling

| Podcasts | April 22, 2026 | 6.28 Thousand views | 1:11:03

TL;DR

This lecture explores JEPA (Joint Embedding Predictive Architecture) as an energy-based framework for world modeling that operates in latent space rather than pixels, with Hazel Nam introducing Causal JEPA—a method using object-centric slot attention and aggressive masking to teach models physical object dynamics and interactions.

More from Stanford Online

View all
Stanford CS25: Transformers United V6 I Overview of Transformers
1:16:46
Stanford Online Stanford Online

Stanford CS25: Transformers United V6 I Overview of Transformers

Stanford's CS25 introductory lecture traces the evolution from hand-engineered features to Transformer architectures, explaining how self-attention mechanisms enable parallel processing and long-context modeling, while exploring how billion-parameter language models develop emergent reasoning capabilities through next-token prediction on internet-scale data.

1 day ago · 10 points
Stanford Robotics Seminar ENGR319 | Spring 2026 | Mechanical Intelligence in Locomotion
51:44
Stanford Online Stanford Online

Stanford Robotics Seminar ENGR319 | Spring 2026 | Mechanical Intelligence in Locomotion

This seminar introduces 'missile-scale' robotics (~1kg) as a critical gap between micro and macro robots, demonstrating that mechanical redundancy (morphological intelligence) enables reliable locomotion in unpredictable terrain without sensors by applying Shannon's information theory to legged locomotion, while biological gait-switching strategies can overcome inherent speed limitations.

1 day ago · 10 points