LESSWRONG
LW

Wikitags

AXRP

Edited by Multicore, DanielFilan, et al. last updated 30th Dec 2024

AI X-Risk Research Podcast is a podcast hosted by Daniel Filan.

See also: ,

Interviews
Subscribe
1
Subscribe
1
Audio
Discussion0
Discussion0
Posts tagged AXRP
72AXRP Episode 31 - Singular Learning Theory with Daniel Murfet
Ω
DanielFilan
1y
Ω
4
69AXRP Episode 27 - AI Control with Buck Shlegeris and Ryan Greenblatt
Ω
DanielFilan
1y
Ω
10
55AXRP Episode 24 - Superalignment with Jan Leike
Ω
DanielFilan
2y
Ω
3
52AXRP Episode 22 - Shard Theory with Quintin Pope
Ω
DanielFilan
2y
Ω
11
45AXRP Episode 19 - Mechanistic Interpretability with Neel Nanda
Ω
DanielFilan
2y
Ω
0
43AXRP Episode 25 - Cooperative AI with Caspar Oesterheld
Ω
DanielFilan
2y
Ω
0
41AXRP Episode 39 - Evan Hubinger on Model Organisms of Misalignment
Ω
DanielFilan
7mo
Ω
0
34AXRP Episode 33 - RLHF Problems with Scott Emmons
Ω
DanielFilan
1y
Ω
0
34AXRP Episode 15 - Natural Abstractions with John Wentworth
Ω
DanielFilan
3y
Ω
1
34AXRP Episode 38.2 - Jesse Hoogland on Singular Learning Theory
Ω
DanielFilan
8mo
Ω
0
31AXRP Episode 45 - Samuel Albanie on DeepMind’s AGI Safety Approach
Ω
DanielFilan
9d
Ω
0
28AXRP Episode 41 - Lee Sharkey on Attribution-based Parameter Decomposition
Ω
DanielFilan
1mo
Ω
1
26AXRP Episode 40 - Jason Gross on Compact Proofs and Interpretability
Ω
DanielFilan
4mo
Ω
0
25AXRP Episode 14 - Infra-Bayesian Physicalism with Vanessa Kosoy
Ω
DanielFilan
3y
Ω
10
25AXRP Episode 13 - First Principles of AGI Safety with Richard Ngo
Ω
DanielFilan
3y
Ω
1
Load More (15/59)
Add Posts