We're sunsetting PodQuest on 2025-07-28. Thank you for your support!
Export Podcast Subscriptions
cover of episode Three Red Lines We're About to Cross Toward AGI (Daniel Kokotajlo, Gary Marcus, Dan Hendrycks)

Three Red Lines We're About to Cross Toward AGI (Daniel Kokotajlo, Gary Marcus, Dan Hendrycks)

2025/6/24
logo of podcast Machine Learning Street Talk (MLST)

Machine Learning Street Talk (MLST)

AI Deep Dive AI Chapters Transcript
People
D
Dan Hendrycks
D
Daniel Kokotajlo
G
Gary Marcus
一位批评当前人工智能研究方向的认知科学家和名誉教授。
Topics
Daniel Kokotajlo: 我认为美国可以通过阻止中国自动化AI研发来削弱其发展超级智能的能力。自动化AI研发可能导致情报爆炸,使其他国家无法赶超。AI实验室领导者认为,即使存在风险,他们也应该尽快发展AI并赢得竞赛,因为他们更信任自己。OpenAI成立的初衷是制衡DeepMind,防止Demis滥用AGI权力。 Gary Marcus: 我们都希望人工智能对人类有益,并为此共同努力。我们都认为AGI有积极的可能性,并希望引导其朝着积极的方向发展。如果停留在LLM上,就不会有持久的优势。如果范式保持不变,我不认为任何人会获得持久的优势。自动化AI研究过程是具有破坏性的、可怕的和危险的。即使AGI或ASI需要一段时间才能实现,人们已经在推动我们的红线,且透明度正在倒退。 Dan Hendrycks: 原始AGI遵循指令的能力“相当合理”,但这仍然令人担忧,特别是在涉及武器控制等高风险领域。我主要考虑阻止超级智能,因为这更容易实施,在地缘政治上与现有激励措施相容。如果朝着AGI发展,这可能是一个坏主意,因为我们将无法阻止超级智能。人们应该有自主权在AI增加GDP所能提供的资源下选择不同的生活方式。美国更有动力对前沿技术保持透明,因为中国已经了解情况。对前沿技术的高度透明对于建立更可信的威慑非常有用。

Deep Dive

Chapters
The conversation begins by discussing the dangers of automated AI R&D, particularly the "rationalization" that if one country doesn't develop superintelligence, another will. The risks of recursive self-improvement and the concentration of power in the hands of a few are highlighted.
  • Automated AI R&D could lead to an intelligence explosion and a durable advantage for one nation.
  • AI lab leaders prioritize speed over safety, driven by a lack of trust in each other.
  • The potential for AI to be used for malicious purposes, such as guiding weapons, is a major concern.

Shownotes Transcript

What if the most powerful technology in human history is being built by people who openly admit they don't trust each other? In this explosive 2-hour debate, three AI experts pull back the curtain on the shocking psychology driving the race to Artificial General Intelligence—and why the people building it might be the biggest threat of all. Kokotajlo predicts AGI by 2028 based on compute scaling trends. Marcus argues we haven't solved basic cognitive problems from his 2001 research. The stakes? If Kokotajlo is right and Marcus is wrong about safety progress, humanity may have already lost control.

Sponsor messages:

========

Google Gemini: Google Gemini features Veo3, a state-of-the-art AI video generation model in the Gemini app. Sign up at https://gemini.google.com

Tufa AI Labs are hiring for ML Engineers and a Chief Scientist in Zurich/SF. They are top of the ARCv2 leaderboard!

https://tufalabs.ai/

========

Guest Powerhouse

Gary Marcus - Cognitive scientist, author of "Taming Silicon Valley," and AI's most prominent skeptic who's been warning about the same fundamental problems for 25 years (https://garymarcus.substack.com/)

Daniel Kokotajlo - Former OpenAI insider turned whistleblower who reveals the disturbing rationalizations of AI lab leaders in his viral "AI 2027" scenario (https://ai-2027.com/)

Dan Hendrycks - Director of the Center for AI Safety who created the benchmarks used to measure AI progress and argues we have only years, not decades, to prevent catastrophe (https://danhendrycks.com/)

Transcript:

http://app.rescript.info/public/share/tEcx4UkToi-2jwS1cN51CW70A4Eh6QulBRxDILoXOno

TOC:

Introduction: The AI Arms Race

00:00:04 - The Danger of Automated AI R&D

00:00:43 - The Rationalization: "If we don't, someone else will"

00:01:56 - Sponsor Reads (Tufa AI Labs & Google Gemini)

00:02:55 - Guest Introductions

The Philosophical Stakes

00:04:13 - What is the Positive Vision for AGI?

00:07:00 - The Abundance Scenario: Superintelligent Economy

00:09:06 - Differentiating AGI and Superintelligence (ASI)

00:11:41 - Sam Altman: "A Decade in a Month"

00:14:47 - Economic Inequality & The UBI Problem

Policy and Red Lines

00:17:13 - The Pause Letter: Stopping vs. Delaying AI

00:20:03 - Defining Three Concrete Red Lines for AI Development

00:25:24 - Racing Towards Red Lines & The Myth of "Durable Advantage"

00:31:15 - Transparency and Public Perception

00:35:16 - The Rationalization Cascade: Why AI Labs Race to "Win"

Forecasting AGI: Timelines and Methodologies

00:42:29 - The Case for Short Timelines (Median 2028)

00:47:00 - Scaling Limits: Compute, Data, and Money

00:49:36 - Forecasting Models: Bio-Anchors and Agentic Coding

00:53:15 - The 10^45 FLOP Thought Experiment

The Great Debate: Cognitive Gaps vs. Scaling

00:58:41 - Gary Marcus's Counterpoint: The Unsolved Problems of Cognition

01:00:46 - Current AI Can't Play Chess Reliably

01:08:23 - Can Tools and Neurosymbolic AI Fill the Gaps?

01:16:13 - The Multi-Dimensional Nature of Intelligence

01:24:26 - The Benchmark Debate: Data Contamination and Reliability

01:31:15 - The Superhuman Coder Milestone Debate

01:37:45 - The Driverless Car Analogy

The Alignment Problem

01:39:45 - Has Any Progress Been Made on Alignment?

01:42:43 - "Fairly Reasonably Scares the Sh*t Out of Me"

01:46:30 - Distinguishing Model vs. Process Alignment

Scenarios and Conclusions

01:49:26 - Gary's Alternative Scenario: The Neurosymbolic Shift

01:53:35 - Will AI Become Jeff Dean?

01:58:41 - Takeoff Speeds and Exceeding Human Intelligence

02:03:19 - Final Disagreements and Closing Remarks

REFS:

Gary Marcus (2001) - The Algebraic Mind

https://mitpress.mit.edu/9780262632683/the-algebraic-mind/

00:59:00

Gary Marcus & Ernest Davis (2019) - Rebooting AI

https://www.penguinrandomhouse.com/books/566677/rebooting-ai-by-gary-marcus-and-ernest-davis/

01:31:59

Gary Marcus (2024) - Taming SV

https://www.hachettebookgroup.com/titles/gary-marcus/taming-silicon-valley/9781541704091/

00:03:01