

Future of Life Institute Podcast
Future of Life Institute
The Future of Life Institute (FLI) is a nonprofit working to reduce global catastrophic and existential risk from powerful technologies. In particular, FLI focuses on risks from artificial intelligence (AI), biotechnology, nuclear weapons and climate change. The Institute's work is made up of three main strands: grantmaking for risk reduction, educational outreach, and advocacy within the United Nations, US government and European Union institutions. FLI has become one of the world's leading voices on the governance of AI having created one of the earliest and most influential sets of governance principles: the Asilomar AI Principles.
Episodes
Mentioned books

Jan 5, 2023 • 39min
Sean Ekins on the Dangers of AI Drug Discovery
On this special episode of the podcast, Emilia Javorsky interviews Sean Ekins about the dangers of AI drug discovery. They talk about how Sean discovered an extremely toxic chemical (VX) by reversing an AI drug discovery algorithm.
Timestamps:
00:00 Introduction
00:46 Sean’s professional journey
03:45 Can computational models replace animal models?
07:24 The risks of AI drug discovery
12:48 Should scientists disclose dangerous discoveries?
19:40 How should scientists handle dual-use technologies?
22:08 Should we open-source potentially dangerous discoveries?
26:20 How do we control autonomous drug creation?
31:36 Surprising chemical discoveries made by black-box AI systems
36:56 How could the dangers of AI drug discovery be mitigated?

Dec 29, 2022 • 50min
Anders Sandberg on the Value of the Future
Anders Sandberg joins the podcast to discuss various philosophical questions about the value of the future.
Learn more about Anders' work: https://www.fhi.ox.ac.uk
Timestamps:
00:00 Introduction
00:54 Humanity as an immature teenager
04:24 How should we respond to our values changing over time?
18:53 How quickly should we change our values?
24:58 Are there limits to what future morality could become?
29:45 Could the universe contain infinite value?
36:00 How do we balance weird philosophy with common sense?
41:36 Lightning round: mind uploading, aliens, interstellar travel, cryonics

Dec 22, 2022 • 1h 3min
Anders Sandberg on Grand Futures and the Limits of Physics
Anders Sandberg joins the podcast to discuss how big the future could be and what humanity could achieve at the limits of physics.
Learn more about Anders' work: https://www.fhi.ox.ac.uk
Timestamps:
00:00 Introduction
00:58 Does it make sense to write long books now?
06:53 Is it possible to understand all of science now?
10:44 What is exploratory engineering?
15:48 Will humanity develop a completed science?
21:18 How much of possible technology has humanity already invented?
25:22 Which sciences have made the most progress?
29:11 How materially wealthy could humanity become?
39:34 Does a grand futures depend on space travel?
49:16 Trade between proponents of different moral theories
53:13 How does physics limit our ethical options?
55:24 How much could our understanding of physics change?
1:02:30 The next episode

Dec 15, 2022 • 58min
Anders Sandberg on ChatGPT and the Future of AI
Anders Sandberg from The Future of Humanity Institute joins the podcast to discuss ChatGPT, large language models, and what he's learned about the risks and benefits of AI.
Timestamps:
00:00 Introduction
00:40 ChatGPT
06:33 Will AI continue to surprise us?
16:22 How do language models fail?
24:23 Language models trained on their own output
27:29 Can language models write college-level essays?
35:03 Do language models understand anything?
39:59 How will AI models improve in the future?
43:26 AI safety in light of recent AI progress
51:28 AIs should be uncertain about values

Dec 8, 2022 • 48min
Vincent Boulanin on Military Use of Artificial Intelligence
Vincent Boulanin joins the podcast to explain how modern militaries use AI, including in nuclear weapons systems.
Learn more about Vincent's work: https://sipri.org
Timestamps:
00:00 Introduction
00:45 Categorizing risks from AI and nuclear
07:40 AI being used by non-state actors
12:57 Combining AI with nuclear technology
15:13 A human should remain in the loop
25:05 Automation bias
29:58 Information requirements for nuclear launch decisions
35:22 Vincent's general conclusion about military machine learning
37:22 Specific policy measures for decreasing nuclear risk
Social Media Links:
➡️ WEBSITE: https://futureoflife.org
➡️ TWITTER: https://twitter.com/FLIxrisk
➡️ INSTAGRAM: https://www.instagram.com/futureoflifeinstitute/
➡️ META: https://www.facebook.com/futureoflifeinstitute
➡️ LINKEDIN: https://www.linkedin.com/company/future-of-life-institute/

Dec 1, 2022 • 45min
Vincent Boulanin on the Dangers of AI in Nuclear Weapons Systems
Vincent Boulanin joins the podcast to explain the dangers of incorporating artificial intelligence in nuclear weapons systems.
Learn more about Vincent's work: https://sipri.org
Timestamps:
00:00 Introduction
00:55 What is strategic stability?
02:45 How can AI be a positive factor in nuclear risk?
10:17 Remote sensing of nuclear submarines
19:50 Using AI in nuclear command and control
24:21 How does AI change the game theory of nuclear war?
30:49 How could AI cause an accidental nuclear escalation?
36:57 How could AI cause an inadvertent nuclear escalation?
43:08 What is the most important problem in AI nuclear risk?
44:39 The next episode

Nov 24, 2022 • 52min
Robin Hanson on Predicting the Future of Artificial Intelligence
Robin Hanson joins the podcast to discuss AI forecasting methods and metrics.
Timestamps:
00:00 Introduction
00:49 Robin's experience working with AI
06:04 Robin's views on AI development
10:41 Should we care about metrics for AI progress?
16:56 Is it useful to track AI progress?
22:02 When should we begin worrying about AI safety?
29:16 The history of AI development
39:52 AI progress that deviates from current trends
43:34 Is this AI boom different than past booms?
48:26 Different metrics for predicting AI

Nov 17, 2022 • 60min
Robin Hanson on Grabby Aliens and When Humanity Will Meet Them
Robin Hanson joins the podcast to explain his theory of grabby aliens and its implications for the future of humanity.
Learn more about the theory here: https://grabbyaliens.com
Timestamps:
00:00 Introduction
00:49 Why should we care about aliens?
05:58 Loud alien civilizations and quiet alien civilizations
08:16 Why would some alien civilizations be quiet?
14:50 The moving parts of the grabby aliens model
23:57 Why is humanity early in the universe?
28:46 Could't we just be alone in the universe?
33:15 When will humanity expand into space?
46:05 Will humanity be more advanced than the aliens we meet?
49:32 What if we discovered aliens tomorrow?
53:44 Should the way we think about aliens change our actions?
57:48 Can we reasonably theorize about aliens?
53:39 The next episode

Nov 10, 2022 • 45min
Ajeya Cotra on Thinking Clearly in a Rapidly Changing World
Ajeya Cotra joins us to talk about thinking clearly in a rapidly changing world.
Learn more about the work of Ajeya and her colleagues: https://www.openphilanthropy.org
Timestamps:
00:00 Introduction
00:44 The default versus the accelerating picture of the future
04:25 The role of AI in accelerating change
06:48 Extrapolating economic growth
08:53 How do we know whether the pace of change is accelerating?
15:07 How can we cope with a rapidly changing world?
18:50 How could the future be utopian?
22:03 Is accelerating technological progress immoral?
25:43 Should we imagine concrete future scenarios?
31:15 How should we act in an accelerating world?
34:41 How Ajeya could be wrong about the future
41:41 What if change accelerates very rapidly?

Nov 3, 2022 • 54min
Ajeya Cotra on how Artificial Intelligence Could Cause Catastrophe
Ajeya Cotra joins us to discuss how artificial intelligence could cause catastrophe.
Follow the work of Ajeya and her colleagues: https://www.openphilanthropy.org
Timestamps:
00:00 Introduction
00:53 AI safety research in general
02:04 Realistic scenarios for AI catastrophes
06:51 A dangerous AI model developed in the near future
09:10 Assumptions behind dangerous AI development
14:45 Can AIs learn long-term planning?
18:09 Can AIs understand human psychology?
22:32 Training an AI model with naive safety features
24:06 Can AIs be deceptive?
31:07 What happens after deploying an unsafe AI system?
44:03 What can we do to prevent an AI catastrophe?
53:58 The next episode


