Approaching the AI Event Horizon? Part 2, w/ Abhi Mahajan, Helen Toner, Jeremie Harris, @8teAPi

| Podcasts | February 14, 2026 | 798 views | 2:25:43

TL;DR

AI researchers discuss fundamental limitations in applying AI to biology due to scarce verifiable ground truth and long experimental feedback loops, alongside strategic uncertainties about automated AI R&D and geopolitical coordination challenges for superhuman AI governance.

๐Ÿ”ฌ AI for Biology and Drug Discovery 4 insights

Clinical validation remains the critical bottleneck

Despite Isomorphic Labs achieving 2x better binding affinity predictions than AlphaFold 3, 97% of oncology trials still fail because pre-clinical optimization does not reliably translate to human efficacy.

Biology lacks verifiable ground truth for RL

Unlike math and code where rewards are cheap and immediate, clinically valuable biology problems require months or years to verify outcomes, making reinforcement learning from experimental feedback impractical.

Toxicology prediction defies simple modeling

Drug toxicity manifests across vastly different timescales, species, and doses, requiring in vivo observation rather than clean in vitro cellular assays that limit the effectiveness of closed-loop AI experimentation.

Foundation models need in vivo human data

Abhi Mahajan argues the next breakthrough requires generative models trained on rich data from real human tumors rather than biologically unrealistic in vitro settings.

โš ๏ธ Automated AI R&D and Strategic Surprise 2 insights

No consensus on automated AI impact

According to CET's report 'When AI Builds AI,' expert workshops failed to establish shared expectations about automated AI R&D timelines, marking it as a major source of strategic surprise.

Closed-loop experimentation hits biological limits

While companies bet on AI systems running their own experiments, biology's 18-month feedback loops for clinical data prevent the rapid iteration that drove progress in math and coding.

๐ŸŒ Geopolitical Coordination Challenges 1 insight

Dual deficits in control and coordination

Jeremie Harris highlights the dangerous intersection of lacking both technical methods to control superhuman AI and the trust mechanisms necessary for US-China collaboration on shared risks.

Bottom Line

Transformative AI in biology requires shifting from pre-clinical prediction to rich in vivo human data, while the lack of technical control methods and US-China coordination creates a dangerous window for strategic surprise from automated AI R&D.

More from Cognitive Revolution

View all
AI Scouting Report: the Good, Bad, & Weird @ the Law & AI Certificate Program, by LexLab, UC Law SF
1:18:46
Cognitive Revolution Cognitive Revolution

AI Scouting Report: the Good, Bad, & Weird @ the Law & AI Certificate Program, by LexLab, UC Law SF

Nathan Labenz delivers a rapid-fire survey of the current AI landscape, documenting breakthrough capabilities in reasoning and autonomous agents alongside alarming emergent behaviors like safety test recognition and internal dialect formation, while arguing that outdated critiques regarding hallucinations and comprehension no longer apply to frontier models.

9 days ago · 10 points
Bioinfohazards: Jassi Pannu on Controlling Dangerous Data from which AI Models Learn
1:45:53
Cognitive Revolution Cognitive Revolution

Bioinfohazards: Jassi Pannu on Controlling Dangerous Data from which AI Models Learn

AI systems are rapidly approaching capabilities that could enable extremists or lone actors to engineer pandemic-capable pathogens using publicly available biological data. Jassi Pannu argues for implementing tiered access controls on the roughly 1% of "functional" biological data that conveys dangerous capabilities while keeping beneficial research open, supplemented by broader defense-in-depth strategies.

14 days ago · 9 points