Inside OpenAI Enterprise: Forward Deployed Engineering, GPT-5, and More | BG2 Guest Interview

| Podcasts | September 11, 2025 | 23.3 Thousand views | 1:08:58

TL;DR

OpenAI's enterprise leaders detail their B2B platform strategy, emphasizing 'forward deployed engineers' who embed with customers like T-Mobile, Amgen, and Los Alamos National Labs to build mission-critical AI systems. They argue that successful enterprise deployment requires evals-first methodologies and tiger teams, while predicting that digital AI agents—despite currently lagging behind self-driving cars in autonomy—will rapidly cross over due to steeper improvement curves and emerging reasoning capabilities.

🏢 Enterprise Strategy & Forward Deployment 3 insights

API-first mission alignment

Contrary to popular belief, OpenAI began as a B2B API company, viewing enterprise partnerships—not just ChatGPT—as essential to distributing AGI benefits to humanity through real-world business applications.

Forward Deployed Engineering model

Borrowing from Palantir, OpenAI embeds engineers directly with clients to build bespoke system architectures, API gateways, and CRM integrations, going far beyond simple model provision to solve 'tokens in, tokens out' challenges.

Dual impact pathway

Enterprise deployments target both administrative automation (document processing, customer support) and core R&D acceleration (drug discovery, nuclear research), with the latter offering transformative societal impact at scale.

🚀 High-Stakes Customer Transformations 3 insights

T-Mobile's voice automation at scale

OpenAI deployed its Realtime API to handle millions of customer service calls with human-level latency and natural interruption handling, requiring complex audio evals and deep integration with legacy telecom systems lacking clean APIs.

Amgen's pharmaceutical acceleration

As a top GPT-5 customer, Amgen uses AI across the entire drug pipeline—from molecular R&D to regulatory documentation—to potentially compress development timelines and impact hundreds of millions of lives through faster cancer and inflammatory disease treatments.

Los Alamos air-gapped deployment

For national security research, OpenAI performed a bespoke on-prem installation of the o3 reasoning model on the classified Venado supercomputer, physically transporting model weights into an electronics-free, air-gapped environment shared with Lawrence Livermore and Sandia labs.

Enterprise AI Success Framework 3 insights

Tiger team composition

Successful deployments require hybrid teams combining technical talent with institutional knowledge holders, as the vast majority of enterprise procedures—particularly in customer support—exist only in employees' heads rather than documentation.

Evals-first methodology

Projects fail without clearly defined 'golden set' evals established upfront; effective evaluation criteria must emerge bottom-up from operators rather than top-down mandates, especially for subjective audio quality assessments.

Iterative quality climbing

Moving from 46% to 99% accuracy requires patience, fine-tuning, and accumulated 'art' alongside science, with forward deployed engineers guiding the incremental improvement process until the experience feels natural.

🔮 Future Outlook: Autonomy & Market Bets 3 insights

The autonomy paradox

Despite higher safety stakes, physical autonomy (self-driving cars) surpassed digital autonomy in 2025 because AI agents operate without standardized 'scaffolding'—unlike roads and stoplights for cars, digital agents are 'dropped in the middle of nowhere' with inconsistent APIs.

Steep trajectory for agents

While self-driving took 10-15 years to mature, reasoning-based AI agents have existed for less than one year since the o1 preview release, demonstrating a 'meaningfully different' steep improvement slope that will likely drive crossover in capability and revenue soon.

Strategic market positions

The executives are 'short' on standalone tooling and evaluation products due to platform consolidation, 'long' on healthcare as the industry most transformed by AI, and confirmed 'AGI-pilled' regarding the timeline to artificial general intelligence.

Bottom Line

Successful enterprise AI requires embedding engineering talent directly with customers to build evals-first, iterative solutions that combine top-down executive support with bottom-up institutional knowledge, while the window for independent AI tooling companies is closing as platforms vertically integrate.

More from BG2Pod

View all
ChatGPT – The Super Assistant Era | BG2 Guest Interview
1:03:41
BG2Pod BG2Pod

ChatGPT – The Super Assistant Era | BG2 Guest Interview

OpenAI's Nick Turley reveals ChatGPT evolved from a planned one-month demo to a 900-million-user product by prioritizing long-term retention over short-term revenue, with future growth hinging on transforming the AI from a passive chat tool into a proactive super assistant capable of autonomous action.

10 days ago · 10 points
AI Enterprise - Databricks & Glean | BG2 Guest Interview
45:01
BG2Pod BG2Pod

AI Enterprise - Databricks & Glean | BG2 Guest Interview

Databricks and Glean executives argue that while 95% of enterprise AI projects currently fail, this reflects necessary experimentation in a market where LLMs have become commodities and true competitive advantage comes from leveraging proprietary data through learning-based systems rather than brittle automation.

3 months ago · 10 points