Resolutions: Flourish or Die! The Choice is Ours
Автор: Mindplex
Загружено: 2026-01-01
Просмотров: 3562
Описание:
In this intense and comprehensive debate, AGI pioneer Dr. Ben Goertzel and AI safety expert Dr. Roman Yampolskiy clash over the ultimate fate of humanity. The central question: Will Artificial General Intelligence lead to our extinction or our unprecedented flourishing?
Their conversation covers the core disagreements shaping our future: The Alignment Problem. Is it solvable, or is it fundamentally "bunk" due to humanity's conflicting and evolving values? They explore Strategic Choices: Is open-source or proprietary development safer? Can we even stop the AGI race, or is it an uncontrollable force? They highlight The Future of Intelligence: Visions of transhumanist co-evolution, merging with AI, and "worthy successors" are debated against warnings of value manipulation and loss of meaning.
Key Moments You Don’t Want to Miss!
0:00 - Flourish or Die! The Choice is Ours
4:00 - Introducing Dr. Ben Gertel & Dr. Roman Yampolskiy
6:50 - Roman Yampolskiy: First Impressions of Ben's Work & AI Risk
8:22 - Ben Goertzel Responds to "If You Build It, We All Die"
10:20 - Defining AI Alignment & The Core Disagreement
16:34 - The Ethics of Risk: 1% Chance vs. 8 Billion Lives
24:43 - Formal Verification & Its Limits for Safety
28:10 - Concerns About LLMs and Paths to AGI
30:03 - Open Source vs. Proprietary AI: Risks and Benefits
34:27 - Ben's Vision: Transhumanism, Co-evolution & The "End Game"
37:06 - Roman on Merging with Superintelligence & Value Contribution
38:54 - The Greatest Risk: Nasty Humans vs. Nasty AI
41:23 - Roman Defines AI Alignment & Why It's "Bunk"
43:18 - Compassion vs. Alignment & The Evolution of Values
49:02 - Tools vs. Agents: The Fuzzy Boundary
52:10 - The Dangerous Transition from AGI to Superintelligence
55:20 - Roman's Strongest Argument for High P(Doom)
57:14 - Can We Control or Understand a Superintelligence?
58:53 - The "Mind Space" Argument Revisited: Culture vs. Randomness
1:04:10 - Can We Stop the AGI Race? A Debate on Feasibility
1:08:07 - Would You Stop If The World Voted Against AGI?
1:13:08 - The "Psychopath with AGI" Scenario & Transition Risks
1:15:48 - Historical Bans (Cloning, Nukes) vs. AI's Economic Momentum
1:18:05 - Practical Steps: Global Safety Standards & Certification
1:24:26 - Analogy: Why Can't We Even Stop Deep Fakes?
1:27:33 - The Uncontrollable Nature of Global Tech Ecosystems
1:35:43 - Can We Enforce Safety Rules? Roman's "Safety Theater"
1:42:05 - Do We Deserve to Be Succeeded by a Superior Intelligence?
1:44:25 - Personal Biases, Religion & Simulation Hypothesis
1:46:33 - Flourish or Die? Can We Have Our Cake and Eat It Too?
1:50:34 - Practical Collaboration on AI Security
👇 Want more?
Bonus: Dr. Mihaela and her guests explore technical arguments on formal verification, the "mind space" fallacy, and the perils of the transition period from AGI to superintelligence. This is not a surface-level chat but a rigorous, hour-long dissection of our most critical technology
🎛️Join and follow us for more👇:
Website: https://magazine.mindplex.ai/
X (formerly Twitter): https://x.com/Mindplex_AI
Linkedin: / mindplex-ai
Telegram Group: https://t.me/mindplex_ai
Telegram Channel: https://t.me/mindplexann
Subreddit: / mindplex
Повторяем попытку...
Доступные форматы для скачивания:
Скачать видео
-
Информация по загрузке: