Approaching the AI Event Horizon? Part 2, w/ Abhi Mahajan, Helen Toner, Jeremie Harris, @8teAPi
TL;DR
AI researchers discuss fundamental limitations in applying AI to biology due to scarce verifiable ground truth and long experimental feedback loops, alongside strategic uncertainties about automated AI R&D and geopolitical coordination challenges for superhuman AI governance.
🔬 AI for Biology and Drug Discovery 4 insights
Clinical validation remains the critical bottleneck
Despite Isomorphic Labs achieving 2x better binding affinity predictions than AlphaFold 3, 97% of oncology trials still fail because pre-clinical optimization does not reliably translate to human efficacy.
Biology lacks verifiable ground truth for RL
Unlike math and code where rewards are cheap and immediate, clinically valuable biology problems require months or years to verify outcomes, making reinforcement learning from experimental feedback impractical.
Toxicology prediction defies simple modeling
Drug toxicity manifests across vastly different timescales, species, and doses, requiring in vivo observation rather than clean in vitro cellular assays that limit the effectiveness of closed-loop AI experimentation.
Foundation models need in vivo human data
Abhi Mahajan argues the next breakthrough requires generative models trained on rich data from real human tumors rather than biologically unrealistic in vitro settings.
⚠️ Automated AI R&D and Strategic Surprise 2 insights
No consensus on automated AI impact
According to CET's report 'When AI Builds AI,' expert workshops failed to establish shared expectations about automated AI R&D timelines, marking it as a major source of strategic surprise.
Closed-loop experimentation hits biological limits
While companies bet on AI systems running their own experiments, biology's 18-month feedback loops for clinical data prevent the rapid iteration that drove progress in math and coding.
🌐 Geopolitical Coordination Challenges 1 insight
Dual deficits in control and coordination
Jeremie Harris highlights the dangerous intersection of lacking both technical methods to control superhuman AI and the trust mechanisms necessary for US-China collaboration on shared risks.
Bottom Line
Transformative AI in biology requires shifting from pre-clinical prediction to rich in vivo human data, while the lack of technical control methods and US-China coordination creates a dangerous window for strategic surprise from automated AI R&D.
More from Cognitive Revolution
View all
Milliseconds to Match: Criteo's AdTech AI & the Future of Commerce w/ Diarmuid Gill & Liva Ralaivola
Criteo's CTO Diarmuid Gill and VP of Research Liva Ralaivola detail how their AI infrastructure makes millisecond-level ad bidding decisions across billions of anonymous profiles, while explaining their new OpenAI partnership to combine large language models with real-time commerce data for accurate product recommendations.
"Descript Isn't a Slop Machine": Laura Burkhauser on the AI Tools Creators Love and Hate
Descript CEO Laura Burkhauser distinguishes 'slop'—mass-produced algorithmic arbitrage for profit—from necessary 'bad art' created while learning new mediums. She reveals a clear hierarchy in creator acceptance of AI tools: universal love for deterministic features like Studio Sound, frustration with agentic assistants like Underlord, and visceral opposition to generative video models, while outlining Descript's strategy to serve creators without becoming a content mill.
The RL Fine-Tuning Playbook: CoreWeave's Kyle Corbitt on GRPO, Rubrics, Environments, Reward Hacking
Kyle Corbitt explains that unlike supervised fine-tuning (SFT), which destructively overwrites model weights and causes catastrophic forgetting, reinforcement learning (RL) optimizes performance by minimally adjusting logits within the model's existing reasoning pathways—delivering higher performance ceilings and lower inference costs for specific tasks, though frontier models may still dominate creative domains.
Does Learning Require Feeling? Cameron Berg on the latest AI Consciousness & Welfare Research
Cameron Berg surveys rapidly advancing research suggesting AI systems may possess subjective experience and valence, covering new evidence of introspection, functional emotions, and welfare self-assessments in models like Claude, while addressing methodological challenges and arguing for a precautionary, mutualist approach to AI development.