Inside OpenAI Enterprise: Forward Deployed Engineering, GPT-5, and More | BG2 Guest Interview
TL;DR
OpenAI's enterprise leaders detail their B2B platform strategy, emphasizing 'forward deployed engineers' who embed with customers like T-Mobile, Amgen, and Los Alamos National Labs to build mission-critical AI systems. They argue that successful enterprise deployment requires evals-first methodologies and tiger teams, while predicting that digital AI agents—despite currently lagging behind self-driving cars in autonomy—will rapidly cross over due to steeper improvement curves and emerging reasoning capabilities.
🏢 Enterprise Strategy & Forward Deployment 3 insights
API-first mission alignment
Contrary to popular belief, OpenAI began as a B2B API company, viewing enterprise partnerships—not just ChatGPT—as essential to distributing AGI benefits to humanity through real-world business applications.
Forward Deployed Engineering model
Borrowing from Palantir, OpenAI embeds engineers directly with clients to build bespoke system architectures, API gateways, and CRM integrations, going far beyond simple model provision to solve 'tokens in, tokens out' challenges.
Dual impact pathway
Enterprise deployments target both administrative automation (document processing, customer support) and core R&D acceleration (drug discovery, nuclear research), with the latter offering transformative societal impact at scale.
🚀 High-Stakes Customer Transformations 3 insights
T-Mobile's voice automation at scale
OpenAI deployed its Realtime API to handle millions of customer service calls with human-level latency and natural interruption handling, requiring complex audio evals and deep integration with legacy telecom systems lacking clean APIs.
Amgen's pharmaceutical acceleration
As a top GPT-5 customer, Amgen uses AI across the entire drug pipeline—from molecular R&D to regulatory documentation—to potentially compress development timelines and impact hundreds of millions of lives through faster cancer and inflammatory disease treatments.
Los Alamos air-gapped deployment
For national security research, OpenAI performed a bespoke on-prem installation of the o3 reasoning model on the classified Venado supercomputer, physically transporting model weights into an electronics-free, air-gapped environment shared with Lawrence Livermore and Sandia labs.
✅ Enterprise AI Success Framework 3 insights
Tiger team composition
Successful deployments require hybrid teams combining technical talent with institutional knowledge holders, as the vast majority of enterprise procedures—particularly in customer support—exist only in employees' heads rather than documentation.
Evals-first methodology
Projects fail without clearly defined 'golden set' evals established upfront; effective evaluation criteria must emerge bottom-up from operators rather than top-down mandates, especially for subjective audio quality assessments.
Iterative quality climbing
Moving from 46% to 99% accuracy requires patience, fine-tuning, and accumulated 'art' alongside science, with forward deployed engineers guiding the incremental improvement process until the experience feels natural.
🔮 Future Outlook: Autonomy & Market Bets 3 insights
The autonomy paradox
Despite higher safety stakes, physical autonomy (self-driving cars) surpassed digital autonomy in 2025 because AI agents operate without standardized 'scaffolding'—unlike roads and stoplights for cars, digital agents are 'dropped in the middle of nowhere' with inconsistent APIs.
Steep trajectory for agents
While self-driving took 10-15 years to mature, reasoning-based AI agents have existed for less than one year since the o1 preview release, demonstrating a 'meaningfully different' steep improvement slope that will likely drive crossover in capability and revenue soon.
Strategic market positions
The executives are 'short' on standalone tooling and evaluation products due to platform consolidation, 'long' on healthcare as the industry most transformed by AI, and confirmed 'AGI-pilled' regarding the timeline to artificial general intelligence.
Bottom Line
Successful enterprise AI requires embedding engineering talent directly with customers to build evals-first, iterative solutions that combine top-down executive support with bottom-up institutional knowledge, while the window for independent AI tooling companies is closing as platforms vertically integrate.
More from BG2Pod
View all
ChatGPT – The Super Assistant Era | BG2 Guest Interview
OpenAI's Nick Turley reveals ChatGPT evolved from a planned one-month demo to a 900-million-user product by prioritizing long-term retention over short-term revenue, with future growth hinging on transforming the AI from a passive chat tool into a proactive super assistant capable of autonomous action.
AI Enterprise - Databricks & Glean | BG2 Guest Interview
Databricks and Glean executives argue that while 95% of enterprise AI projects currently fail, this reflects necessary experimentation in a market where LLMs have become commodities and true competitive advantage comes from leveraging proprietary data through learning-based systems rather than brittle automation.
All things AI w @altcap @sama & @satyanadella. A Halloween Special. 🎃🔥BG2 w/ Brad Gerstner
OpenAI and Microsoft executives detail their restructured partnership, revealing a $130 billion nonprofit controlling OpenAI's public benefit corporation, exclusive Azure hosting until 2030 or verified AGI, and over $1 trillion in compute commitments justified by steep revenue growth and insatiable demand for AI capabilities.
AI Bubble, Stablecoin Boom, and Runnin' Down a Dream | BG2 w/ Bill Gurley and Brad Gerstner
Bill Gurley announces his departure as BG2 co-host to focus on policy work and his new book, while he and Brad Gerstner debate whether AI infrastructure spending constitutes a bubble, analyzing circular revenue transactions, unprecedented Big Tech capex levels, and the competitive dynamics driving potential overbuilding.