Bret Taylor of Sierra on AI agents, outcome-based pricing, and the OpenAI board
TL;DR
Bret Taylor explores how AI agents are shifting from polished but forgetful tools to messy, context-rich systems that leverage markdown memory and code repository structures, predicting software engineering will evolve from writing code to crafting 'harnesses' of documentation while enterprises move beyond APIs toward agent-accessible infrastructure.
đź§ AI Memory and Agent Architecture 3 insights
Raw markdown memory outperforms polished forgetful apps
While mainstream consumer AI apps still present blank slates with no memory, experimental projects like OpenClaw achieve better continuity by writing messy notes to markdown files, creating a 'Memento'-style memory system that proves more effective than vector databases.
Code repositories provide ideal agent environments
Unlike general business tasks that scatter context across systems, code repositories concentrate textual context, test feedback, and formal change history in one place, creating an environment uniquely designed for robotic self-reflection and iteration.
File-based context enables true random access
Taylor argues that loading context from a directory of markdown files provides a more useful mix of structured and random access than vector databases, which require knowing what to search for upfront.
🛠️ The Future of Software Engineering 3 insights
Engineers must abandon emotional attachment to code
Taylor describes forcing himself to stop caring about the elegance of handwritten code, recognizing that future software engineering prioritizes correctness through AI agents rather than craftsmanship of raw artifacts.
Documentation becomes the primary engineering output
As AI agents generate implementation code, the durable value shifts to documentation artifacts capturing intention and customer problems, potentially forcing engineers to focus on the very task they historically avoided.
Harness engineering defines the new workflow
The emerging discipline involves building context-rich environments—combining documentation, tests, and rules—around agents rather than writing code directly, though the field is still defining its fundamental categories and tools.
🏢 Enterprise AI and Infrastructure 3 insights
Agent harnesses will supersede traditional APIs
Beyond REST endpoints, companies will need to provide comprehensive 'harnesses'—instruction manuals and context layers—that teach agents how to extract maximum value from business objects.
Existing infrastructure outperforms new protocols
In healthcare, Sierra's AI agents already communicate via 'English over PSTN' (phone calls) rather than waiting for perfect API integrations, demonstrating that AI can effectively leverage legacy rails like the telephone network.
Multi-agent MCP architectures show limitations
Taylor argues that stuffing context into subagents creates robotic orchestration layers, suggesting that true agency requires monolithic context sharing more akin to OpenClaw's markdown approach than modular MCP server architectures.
Bottom Line
Organizations should stop polishing consumer AI interfaces and instead build rich, file-based context systems and 'agent harnesses' that treat documentation as the primary output, while pragmatically leveraging existing infrastructure like phone calls and SSH rather than waiting for perfect API coverage.
More from Stripe
View all
The 20-year journey to fully autonomous cars with Dmitri Dolgov of Waymo
Waymo Co-CEO Dmitri Dolgov details the 20-year technical evolution from Google's self-driving moonshot to 500,000 weekly autonomous rides, explaining why full autonomy requires augmenting end-to-end AI with structured intermediate representations and a 'three teachers' training framework rather than relying solely on scaled-up vision models.
Creating prediction markets (and suing the CFTC) with Tarek Mansour and Luana Lopes Lara
Kalshi founders Tarek Mansour and Luana Lopes Lara recount their four-year battle to launch the first CFTC-regulated prediction market in the US, culminating in a lawsuit against their own regulator to offer election contracts, and why their 'permission-first' approach ultimately enabled $10+ billion monthly volumes.
Garrett Langley of Flock Safety on building technology to solve crime
Garrett Langley explains how Flock Safety grew from a neighborhood project solving car break-ins to a $500M ARR company serving 6,000+ cities by building solar-powered license plate cameras, AI search tools, and drones that help law enforcement clear over one million crimes annually through real-time data coordination.
Reiner Pope of MatX on accelerating AI with transformer-optimized chips
Former Google TPU architect Reiner Pope explains how Google's early research and custom silicon investments enabled its AI resurgence, while outlining MatX's strategy to build transformer-specific chips that solve the latency-throughput trade-off through a hybrid memory architecture, requiring $500M and massive supply chain scale to compete with incumbents.