Episode 2251: Kristian Ronn on why, in the short term, we all might be dead
In the long run, Keynes famously quipped, we are all dead. But Swedish entrepreneur Kristian Ronn reverses Keynes to argue that in the short term we, as a species, might also be death. In his new book Darwinian Trap, Ronn argues that we're hardwired to prioritize immediate benefits over long-term consequences, creating existential risks like nuclear war and uncontrolled AI development. Ronn suggests we need better system design with proper incentives to overcome these tendencies. He proposes controlling critical parts of technology supply chains (like AI chips) to ensure responsible use, similar to nuclear nonproliferation treaties. Despite acknowledging all the obvious challenges of these kind of UN style regulatory initiatives, Ronn remains hopeful that rational thinking and well-designed systems can help humanity transcend its evolutionary limitations.
Here are the 5 KEEN ON take-aways from our conversation with Kristian Ronn:
* The "Darwinian Trap" refers to how humans and systems are hardwired for short-term thinking due to evolutionary forces, creating both personal and existential risks.
* "Offensive realism" in international politics drives nations to compete for resources and develop increasingly dangerous weapons, creating existential threats through arms races.
* AI poses significant existential risks, particularly as a technology multiplier that could enable more destructive weapons and engineered pandemics.
* System design with proper incentives is crucial for overcoming our evolutionary short-term thinking—we need to "change the rules of the game" rather than blame human nature.
* Strategic control of technology supply chains (like AI chips) could potentially create frameworks for responsible AI development, similar to nuclear nonproliferation treaties.
Kristian Rönn is the CEO and co-founder of Normative, a software tool for sustainability accounting. He has a background in mathematics, philosophy, computer science, and artificial intelligence. Before he started Normative, he worked at the University of Oxford’s Future of Humanity Institute on issues related to global catastrophic risks.
This is a public episode. If you'd like to discuss this with other subscribers or get access to bonus episodes, visit keenon.substack.com/subscribe
Kimmel-Kirk and the End of the Television Age: Why Free Speech Has Never Been Freer
A 107 Reasons to Dislike 107 Days: Kamala Harris Throws Everyone, Including Herself, Under the Bus
Gutted and Glutted: The Dire Economics of Podcasting in the AI Age
The Innovation Paradox Undermining the Digital Revolution: How Magical Technology Isn't Translating into Miraculous Economic Progress
Should Billionaires Be Banned? Why Extreme Wealth Might Be Incompatible with Democracy and the Survival of the Earth
Why Trump Might Be Right About Greenland: How a 57,000-Person Island Became Critical to 21st Century Geopolitics
The Unluckiest Generation: Confessions of a Millennial
Why Humans Have Such Big Brains (No, it's not Because of our Intelligence)
How Should Criminals be Punished? From Bentham's "Enlightened" Panopticon to the Universal Human Rights of Prisoners
Why Misogyny May Be America's Most Dangerous Ideology: The Role of the Manosphere in Political Assassinations and Mass Shootings
Rational Exuberance: Why $3 Trillion in AI Investment is Mathematical Certainty, not Madness
From Dodgers Top Draft Pick to Harvard Trained Middle Eastern Maven: Does the American Dream Still Exist?
We're Burning 500 Million Years of Earth's History in a Few Decades: So Stop Pretending Recycling Will Save the Planet