NVIDIA AI Podcast

NVIDIA AI Podcast

201 K subscribers

Welcome to the NVIDIA Developer YouTube Channel Subscribe to this channel for easy-to-follow “how-to” videos to learn about the latest technologies for developers from NVIDIA. Whether you’re a student, professional developer, or tech enthusiast, discover: 🧑‍💻 CUDA Programming: Parallel computing, debugging, and performance tips ✨ Agentic & Generative AI: Build intelligent agents and generative apps with AgentIQ, NeMo, and open-source tools 🤖 Robotics: Unlock smart automation and robotics solutions 📊 Data Science & Analytics: Accelerate data workflows with GPU-powered libraries like RAPIDS and popular tools 🛠️ And More: Deep learning, computer vision, simulation, high-performance computing, SDK tutorials, and expert guides Join a vibrant developer community, stay ahead with emerging tech, get real-world examples, and tips from NVIDIA engineers. Subscribe and start creating, optimizing, and deploying innovations with NVIDIA. 🙌

31 summaries available YouTube ← All channels
Videos Channels Newsletter
Apr 14 - Jetson AI Lab Research Group Call - Tensor RT Edge LLM on Jetson & Culture
51:38
NVIDIA AI Podcast NVIDIA AI Podcast

Apr 14 - Jetson AI Lab Research Group Call - Tensor RT Edge LLM on Jetson & Culture

NVIDIA researchers Lynn Chai and Luc introduce TensorRT Edge LLM, a purpose-built inference engine for deploying large language models on Jetson edge devices, showcasing NVFP4 quantization and speculative decoding techniques that achieve up to 7x faster prefill speeds and 500 tokens per second generation while previewing a simplified vLLM-style Python API coming soon.

5 days ago · 10 points
March 10 - Jetson AI Lab Research Group Call - Lightning talks
55:28
NVIDIA AI Podcast NVIDIA AI Podcast

March 10 - Jetson AI Lab Research Group Call - Lightning talks

This Jetson AI Lab Research Group call features lightning talks on open-source hardware for remote Jetson access, a real-time emotional AI engine for robots running entirely on Jetson Nano, and updates to the Jetson AI Lab model repository with new performance benchmarks and deployment guides.

5 days ago · 8 points
Feb 10 - Jetson AI Lab Research Group Call - Drones on Jetson & Isaac Lab on DGX Spark
57:34
NVIDIA AI Podcast NVIDIA AI Podcast

Feb 10 - Jetson AI Lab Research Group Call - Drones on Jetson & Isaac Lab on DGX Spark

Cameron Rose presents 'Operation Squirrel,' an autonomous drone project using Jetson Orin Nano for real-time target tracking and dynamic payload delivery. The system uses a modular C++ software stack with TensorRT-optimized YOLO and OSNet running at 21 FPS, communicating via UART with a flight controller to maintain following distance through velocity commands.

5 days ago · 9 points
Designing a Modular 6G System Using NVIDIA Aerial™ Framework
43:41
NVIDIA AI Podcast NVIDIA AI Podcast

Designing a Modular 6G System Using NVIDIA Aerial™ Framework

NVIDIA Aerial Framework eliminates the traditional bottleneck of manually converting 6G RAN research into production C++ code by automatically lowering Python, JAX, and PyTorch algorithms into real-time CUDA kernels with microsecond latency, enabling rapid over-the-air deployment cycles.

5 days ago · 7 points
From Theory to Practice—Prototyping 6G With the NVIDIA Sionna Research Kit
39:11
NVIDIA AI Podcast NVIDIA AI Podcast

From Theory to Practice—Prototyping 6G With the NVIDIA Sionna Research Kit

NVIDIA Research introduces the Sionna Research Kit, an open-source, $6,000-$8,000 platform running on DGX Spark that bridges simulation and reality by enabling real-time prototyping of AI-native 6G networks with neural receivers, digital twin channel emulation, and commercial 5G hardware integration.

5 days ago · 10 points
Building Towards Self-Driving Codebases with Long-Running, Asynchronous Agents
37:49
NVIDIA AI Podcast NVIDIA AI Podcast

Building Towards Self-Driving Codebases with Long-Running, Asynchronous Agents

Cursor co-founder Aman traces AI coding's evolution from autocomplete to synchronous agents, outlining the shift toward long-running async cloud agents that use multi-agent architectures to overcome context limits, and predicting a future of self-driving codebases with self-healing systems and minimal human intervention.

27 days ago · 9 points
Accelerate AI through Open Source Inference | NVIDIA GTC
48:21
NVIDIA AI Podcast NVIDIA AI Podcast

Accelerate AI through Open Source Inference | NVIDIA GTC

Industry leaders from NVIDIA, Hugging Face, Mistral AI, Black Forest Labs, and Lightricks discuss how open-source inference optimization—spanning quantization, latent compression, and Mixture of Experts architectures—is enabling both massive trillion-parameter models and efficient edge deployment while driving the shift toward sovereign AI and local data control.

28 days ago · 10 points