Mathematical Superintelligence: Harmonic's Vlad & Tudor on IMO Gold & Theories of Everything

| Podcasts | February 18, 2026 | 67.8 Thousand views | 1:34:29

TL;DR

Harmonic co-founders Vlad Tenev and Tudor Achim discuss their AI system Aristotle, which achieved IMO Gold performance using formally verifiable Lean proofs rather than chain-of-thought reasoning, and outline a vision for mathematical superintelligence that could usher in an era of theoretical abundance and trustworthy AI through verifiable outputs.

🧮 The Nature of Mathematics 3 insights

Mathematics is fundamental reasoning

Mathematics is the process of breaking down understanding into small, verifiable logical steps that others can check, serving as the foundation for understanding physics and engineering.

Unreasonable effectiveness of abstraction

Historical examples like differential geometry enabling Einstein's relativity and number theory enabling secure digital economies demonstrate that abstract math eventually finds practical applications beyond imagination.

Math enables physical understanding

Mathematical reasoning underpins physical laws, with the ultimate goal of understanding fundamental forces and the universe's origins requiring deep mathematical insight.

🏛️ Aristotle's Architecture 4 insights

Formal verification in Lean

Unlike other systems using chain-of-thought, Aristotle generates proofs in the Lean programming language where a trusted kernel verifies every step follows from explicit premises, eliminating traditional peer review needs.

Hybrid Monte Carlo search system

The system combines a large transformer with Monte Carlo tree search (similar to AlphaGo), a LeMa guessing module for managing context between distant goals, and a specialized geometry module based on AlphaGeometry.

IMO Gold achievement

Aristotle achieved gold medal performance at the 2025 International Mathematical Olympiad through reinforcement learning scaling limited only by available compute.

Autoformalization boundary testing

The API's informal mode attempts to convert natural language requests into formal proofs, revealing the boundary between mathematically provable statements and those that are philosophical or factual.

🔮 Future of Mathematical AI 4 insights

Era of theoretical abundance

Future systems may generate multiple competing coherent explanations for all physical phenomena, separable only by increasingly exotic experiments.

Trust through formal verification

Formally verifiable outputs allow superintelligence to be trusted even without mechanistic understanding of the model's internal processes.

Hardening critical infrastructure

Mathematical superintelligence can verify and harden mission-critical infrastructure while solving previously unsolved mathematical problems.

2030 scaling trajectory

By 2030, mathematical superintelligence will scale with compute availability, accelerating both creative insights and knowledge synthesis across domains.

Bottom Line

The path to trustworthy superintelligence runs through formally verifiable reasoning systems that can prove their outputs correct, enabling theoretical abundance and hardened infrastructure without requiring human-comprehensible intermediate steps.

More from Cognitive Revolution

View all
AI Scouting Report: the Good, Bad, & Weird @ the Law & AI Certificate Program, by LexLab, UC Law SF
1:18:46
Cognitive Revolution Cognitive Revolution

AI Scouting Report: the Good, Bad, & Weird @ the Law & AI Certificate Program, by LexLab, UC Law SF

Nathan Labenz delivers a rapid-fire survey of the current AI landscape, documenting breakthrough capabilities in reasoning and autonomous agents alongside alarming emergent behaviors like safety test recognition and internal dialect formation, while arguing that outdated critiques regarding hallucinations and comprehension no longer apply to frontier models.

9 days ago · 10 points
Bioinfohazards: Jassi Pannu on Controlling Dangerous Data from which AI Models Learn
1:45:53
Cognitive Revolution Cognitive Revolution

Bioinfohazards: Jassi Pannu on Controlling Dangerous Data from which AI Models Learn

AI systems are rapidly approaching capabilities that could enable extremists or lone actors to engineer pandemic-capable pathogens using publicly available biological data. Jassi Pannu argues for implementing tiered access controls on the roughly 1% of "functional" biological data that conveys dangerous capabilities while keeping beneficial research open, supplemented by broader defense-in-depth strategies.

14 days ago · 9 points