AXRP - the AI X-risk Research Podcast
En podkast av Daniel Filan
58 Episoder
-
17 - Training for Very High Reliability with Daniel Ziegler
Publisert: 21.8.2022 -
16 - Preparing for Debate AI with Geoffrey Irving
Publisert: 1.7.2022 -
15 - Natural Abstractions with John Wentworth
Publisert: 23.5.2022 -
14 - Infra-Bayesian Physicalism with Vanessa Kosoy
Publisert: 5.4.2022 -
13 - First Principles of AGI Safety with Richard Ngo
Publisert: 31.3.2022 -
12 - AI Existential Risk with Paul Christiano
Publisert: 2.12.2021 -
11 - Attainable Utility and Power with Alex Turner
Publisert: 25.9.2021 -
10 - AI's Future and Impacts with Katja Grace
Publisert: 23.7.2021 -
9 - Finite Factored Sets with Scott Garrabrant
Publisert: 24.6.2021 -
8 - Assistance Games with Dylan Hadfield-Menell
Publisert: 8.6.2021 -
7.5 - Forecasting Transformative AI from Biological Anchors with Ajeya Cotra
Publisert: 28.5.2021 -
7 - Side Effects with Victoria Krakovna
Publisert: 14.5.2021 -
6 - Debate and Imitative Generalization with Beth Barnes
Publisert: 8.4.2021 -
5 - Infra-Bayesianism with Vanessa Kosoy
Publisert: 10.3.2021 -
4 - Risks from Learned Optimization with Evan Hubinger
Publisert: 17.2.2021 -
3 - Negotiable Reinforcement Learning with Andrew Critch
Publisert: 11.12.2020 -
2 - Learning Human Biases with Rohin Shah
Publisert: 11.12.2020 -
1 - Adversarial Policies with Adam Gleave
Publisert: 11.12.2020
AXRP (pronounced axe-urp) is the AI X-risk Research Podcast where I, Daniel Filan, have conversations with researchers about their papers. We discuss the paper, and hopefully get a sense of why it's been written and how it might reduce the risk of AI causing an existential catastrophe: that is, permanently and drastically curtailing humanity's future potential. You can visit the website and read transcripts at axrp.net.