Full Workshop: Build Your Own Deep Research Agents - Louis-François Bouchard, Paul Iusztin, Samridhi
TL;DR
This workshop demonstrates how to build production-grade deep research agents by navigating the spectrum from rigid workflows to autonomous systems, emphasizing strategic context management and hybrid architectures to automate high-quality technical content creation while avoiding generic 'AI slop'.
🚫 The AI Content Problem 3 insights
Avoiding Generic AI Slop
Standard LLM outputs suffer from hallucinations, outdated information, and overused phrases like 'intricacies' and 'rapidly evolving' that provide shallow, meaningless value.
Automation Solution
The team built a deep research and writing agent system to automate technical course and article creation, replacing the expensive manual process of research plus human writing.
Meta-Learning Approach
They used this same agentic system to generate a comprehensive course teaching students exactly how to build the deep research system itself.
🎛️ The Autonomy Spectrum 3 insights
The Autonomy Slider
AI engineering exists on a spectrum from simple prompting to complex agentic systems, where increased autonomy reduces control and raises costs per task.
Workflows Defined
Workflows execute predetermined steps in fixed sequences—like a ticket handler that always classifies, routes, drafts, and validate—offering reliability without dynamic environmental reaction.
Agents Defined
True agents autonomously plan actions, dynamically select tools, and react to environmental changes, making them necessary only when tasks require branching logic like a CRM marketing content generator.
🧠 Context Management & Architecture 3 insights
Context Rot Threshold
LLM performance degrades significantly around 200,000 tokens due to the 'lost in the middle' training problem, occurring well before theoretical context window limits.
Delegation Pattern
Manage context budget by delegating specialized tasks to tools or sub-agents with isolated system prompts and validation logic rather than overloading a single agent's memory.
Multi-Agent Triggers
Deploy multi-agent systems only when facing specific constraints like 20+ tools, strict security or compliance requirements, or insufficient single-agent context windows.
🔍 Deep Research System Design 2 insights
Core Capabilities
Effective deep research agents combine planning, autonomous web navigation, reliable source citation, and feedback loops to perform thorough research without human micromanagement.
Hybrid Implementation
Production-grade AI products integrate workflows, single agents, and multi-agent systems rather than using one pattern exclusively, with deep research serving as a prime example of this unified approach.
Bottom Line
Start with the simplest workflow possible and only increase autonomy when tasks explicitly require dynamic decision-making or branching, while aggressively managing context windows through delegation to prevent performance degradation.
More from AI Engineer
View all
Building Generative Image & Video models at Scale - Sander Dieleman (Veo and Nano Banana)
Sander Dieleman from Google DeepMind explains the technical foundations of training large-scale generative image and video models like Veo, emphasizing that meticulous data curation and learned latent representations are as critical as the diffusion architecture itself. He details how diffusion models reverse a noise corruption process through iterative refinement rather than single-step prediction.
Harness Engineering: How to Build Software When Humans Steer, Agents Execute — Ryan Lopopolo, OpenAI
OpenAI engineer Ryan Lopopolo explains how AI agents have made code generation abundant and free, requiring a fundamental shift from writing implementations to designing systems, documentation, and guardrails that enable agents to execute software engineering tasks autonomously.
AI Didn’t Kill the Web, It Moved in! — Olivier Leplus (AWS) & Yohan Lasorsa (Microsoft)
AI has permeated every stage of the web development lifecycle, from coding agents that use customizable "skills" to browser-native AI debugging tools, fundamentally shifting how developers build, test, and optimize applications.
How METR measures Long Tasks and Experienced Open Source Dev Productivity - Joel Becker, METR
Joel Becker from METR argues that slowing compute growth would proportionally delay AI capabilities milestones measured by task time horizons, while presenting findings that experienced open-source developers showed minimal productivity gains from AI coding assistants like Cursor, challenging optimistic adoption curves.