Stanford Robotics Seminar ENGR319 | Winter 2026 | Robot Motion Learning w/Physics-Based PDE Priors

| Podcasts | January 16, 2026 | 4.29 Thousand views | 53:02

TL;DR

This seminar introduces a robot motion planning framework that embeds physics-based PDE priors (specifically the Eikonal equation) into neural networks, achieving real-time inference in high-dimensional spaces while eliminating the need for weeks of expert data gathering required by conventional data-driven methods.

⚠️ The Three-Way Trade-off in Robot Planning 3 insights

Optimization methods trap themselves in local minima

While optimization-based planners offer high inference and training efficiency, they fail to adapt to complex terrains and high-DOF constraints because they easily get stuck in local minima.

Classical methods suffer from slow computation

Sampling-based approaches like RRT and search-based methods like A* require no training but have prohibitively slow inference times that prevent real-time application in high-dimensional spaces.

Data-driven approaches demand massive training costs

Imitation and reinforcement learning models adapt well to complexity and infer quickly, but require weeks of expert demonstration data and expensive GPU clusters to train, making transfer to new domains impractical.

📐 Physics-Based PDE Priors 3 insights

Eikonal PDE provides mathematical structure

The method uses the Eikonal PDE—which governs optimal wavefront propagation—as a prior rather than a physics simulator, where the solution represents a value function (travel time) whose gradient indicates optimal motion direction.

Gradient matching eliminates need for expert data

The neural network is trained via gradient matching: the network's gradients are compared against the inverse of obstacle distance constraints, requiring only randomly sampled configurations rather than expert trajectories.

Surpassing numerical solver limitations

Traditional numerical solvers like Fast Marching Method scale only to 3-4 dimensions; the neural approach aims to solve the same PDE for high-DOF robots (12-15 dimensions) where numerical methods fail.

🧠 Technical Solutions for High-Dimensional Scaling 2 insights

Metric learning handles multimodal solutions

To address the Eikonal PDE's multiple valid solutions (e.g., going left or right around an obstacle), the network uses a specialized structure with latent space encoding and max pooling to enforce geodesic distance properties like symmetry and triangle inequality.

Temporal difference learning controls gradients

To prevent error accumulation between consecutive trajectory points, the model applies Temporal Difference (TD) learning to enforce Bellman's principle of optimality, regulating gradients similarly to Q-learning.

Real-World Efficiency Gains 3 insights

Sub-100-millisecond inference in complex environments

The method achieves approximately 0.07 seconds planning time for 7-DOF robots and scales to 15-DOF systems navigating narrow passages, matching or exceeding the speed of Nvidia's EmpireNet while generalizing across 300 unseen environments.

Training in minutes versus weeks

Compared to baseline methods requiring several weeks of data gathering and one week of training on eight Tesla GPUs, this approach trains in under one hour on a single RTX 3090 after only 50 minutes of data collection.

Zero-shot transfer to new domains

Because the model requires no expert demonstrations and minimal training data, it transfers immediately to new environments without retraining, solving the deployment bottleneck faced by traditional imitation learning systems.

Bottom Line

By encoding the structure of the Eikonal PDE into neural networks through gradient matching and temporal difference learning, robots can achieve real-time motion planning in high-dimensional spaces with minutes of training on a single GPU, eliminating the costly expert data requirements of conventional data-driven approaches.

More from Stanford Online

View all
Stanford CS153 Frontier Systems | Andreas Blattmann from Black Forest Labs on Visual Intelligence
1:01:14
Stanford Online Stanford Online

Stanford CS153 Frontier Systems | Andreas Blattmann from Black Forest Labs on Visual Intelligence

Andreas Blattmann, co-founder of Black Forest Labs and co-creator of Stable Diffusion, argues that visual intelligence represents the critical next frontier for AI, requiring a fundamental shift from text-centric unimodal models to multimodal systems trained on 'natural representations' (video, audio, physics) to unlock true reasoning, robotics capabilities, and higher intelligence.

5 days ago · 9 points