Stanford Robotics Seminar ENGR319 | Winter 2026 | Resilient Autonomy

| Podcasts | January 23, 2026 | 4.51 Thousand views | 59:55

TL;DR

This seminar outlines research on resilient autonomy for robots in degraded environments like underground mines and caves, emphasizing unified perception models that process pixels once for multiple geometric tasks, multi-modal sensor fusion (thermal, IMU, radar), and novel representations enabling long-range semantic reasoning without depth.

🏗️ The Resilience Imperative 3 insights

Underground exploration demands full autonomy

Systems must operate with intermittent or no communication to base stations due to rock interference, requiring fully on-board mapping, planning, and object detection in caves, coal mines, and decommissioned nuclear plants.

Robustness over peak performance

Unlike current research optimizing for brittle high-performance metrics, resilient autonomy prioritizes robustness through redundancy and resourcefulness to handle dust, darkness, and featureless corridors without system failure.

Real-world validation in Pittsburgh

Testing occurred in diverse challenging locations including a former veterans hospital, limestone mines, and a coal mine cabin to ensure algorithms function across big open spaces, narrow passages, and urban structures.

🧠 Unified Perception with Map Anything 3 insights

Single encoder replaces task-specific models

Map Anything processes pixels once to perform multiple tasks—monocular depth, SFM, depth completion, and localization—avoiding GPU bottlenecks caused by running separate object detectors, SLAM, and VQA models simultaneously.

Factorized outputs enable arbitrary cameras

Unlike prior work (DUSt3R, VGGT) requiring redundant prediction heads, this architecture uses a single DPT head predicting scale and ray directions, allowing flexible inputs for calibrated or uncalibrated cameras and auxiliary data like poses.

Superior odometry with monocular input

Compared to off-the-shelf stereo visual-inertial odometry (ZX camera baseline), Map Anything achieves significantly less drift running at 15-16 Hz using only monocular images, and upcoming optimizations promise 7-10x speedups.

🔍 Multi-Modal Sensing for Adverse Conditions 3 insights

Thermal alignment with visual backbones

AnyThermal fine-tunes DINOv2 on thermal data to align thermal and visual representations, enabling existing visual pipelines (segmentation, place recognition) to work natively with longwave infrared for night and dust penetration.

Open-source synchronized dataset collection

The TartanRGBT dataset uses a custom open-source platform with hard-synchronized stereo thermal/visual sensors, an AGX, and a physical button (no typing in freezing field conditions) to address the scarcity of quality thermal training data.

Online IMU learning for vision degradation

When LiDAR or vision fails in featureless hallways or complete darkness, the system refines a pre-trained IMU model on-the-fly using available visual data, then relies solely on the IMU (which cannot be jammed) to maintain accurate odometry.

🗺️ Geometric Representations and Navigation 3 insights

Native spherical convolutions for fisheye lenses

Rather than projecting distorted fisheye images into multiple virtual pinholes, the system projects images into a canonical spherical space with specialized convolution and pooling, enabling efficient 360-degree perception for drones without edge distortion.

Unified Flow Matching for correspondence

UFM combines wide baseline matching and optical flow using co-visibility masks to handle large viewpoint changes or small motions, providing dense pixel correspondence critical for manipulation and localization tasks.

Ray Frustons for long-range semantic reasoning

This representation combines 3D semantic voxel maps with directional ray vectors at frontier points, allowing the robot to navigate toward distant semantic targets (like water towers) visible in images but beyond accurate depth sensing range.

Bottom Line

Achieving resilient autonomy requires shifting from brittle, handcrafted perception stacks to unified, multi-modal architectures that adaptively fuse visual, thermal, and inertial data while processing pixels once through factorized geometric representations.

More from Stanford Online

View all
Stanford CS221 | Autumn 2025 | Lecture 20: Fireside Chat, Conclusion
58:49
Stanford Online Stanford Online

Stanford CS221 | Autumn 2025 | Lecture 20: Fireside Chat, Conclusion

Percy Liang reflects on AI's transformation from academic curiosity to global infrastructure, debunking sci-fi misconceptions about capabilities while arguing that academia's role in long-term research and critical evaluation remains essential as the job market shifts away from traditional entry-level software engineering.

16 days ago · 7 points
Stanford CS221 | Autumn 2025 | Lecture 19: AI Supply Chains
1:14:36
Stanford Online Stanford Online

Stanford CS221 | Autumn 2025 | Lecture 19: AI Supply Chains

This lecture examines AI's economic impact through the lens of supply chains and organizational strategy, demonstrating why understanding compute monopolies, labor market shifts, and corporate decision-making is as critical as tracking algorithmic capabilities.

16 days ago · 7 points
Stanford CS221 | Autumn 2025 | Lecture 18: AI & Society
1:12:10
Stanford Online Stanford Online

Stanford CS221 | Autumn 2025 | Lecture 18: AI & Society

This lecture argues that AI developers bear unique ethical responsibility for societal outcomes, framing AI as a dual-use technology that requires active steering toward beneficial applications while preventing misuse and accidental harms through rigorous auditing and an ecosystem-aware approach.

16 days ago · 8 points
Stanford CS221 | Autumn 2025 | Lecture 17: Language Models
1:19:46
Stanford Online Stanford Online

Stanford CS221 | Autumn 2025 | Lecture 17: Language Models

This lecture introduces modern language models as industrial-scale systems requiring millions of dollars and trillions of tokens to train, explaining their fundamental operation as auto-regressive next-token predictors that encode language structure through massive statistical modeling.

16 days ago · 10 points