LESSWRONG
LW

2346
Wikitags

AI Success Models

Edited by plex last updated 17th Nov 2021

AI Success Models are proposed paths to an existential win via aligned AI. They are (so far) high level overviews and won't contain all the details, but present at least a sketch of what a full solution might look like. They can be contrasted with threat models, which are stories about how AI might lead to major problems.

Subscribe
Discussion
2
Subscribe
Discussion
2
Posts tagged AI Success Models
220An overview of 11 proposals for building safe advanced AI
Ω
evhub
5y
Ω
37
128AI Safety "Success Stories"
Ω
Wei Dai
6y
Ω
27
114Conversation with Eliezer: What do you want the system to do?
Orpheus16
3y
38
112Four visions of Transformative AI success
Ω
Steven Byrnes
2y
Ω
22
95Towards Hodge-podge Alignment
Ω
Cleo Nardo
3y
Ω
30
85Success without dignity: a nearcasting story of avoiding catastrophe by luck
HoldenKarnofsky
3y
17
85Various Alignment Strategies (and how likely they are to work)
Logan Zoellner
3y
34
82AI Safety via Luck
Ω
Jozdien
3y
Ω
7
81A positive case for how we might succeed at prosaic AI alignment
Ω
evhub
4y
Ω
46
80An Open Agency Architecture for Safe Transformative AI
Ω
davidad
3y
Ω
22
64Possible miracles
Orpheus16, Thomas Larsen
3y
34
63Solving the whole AGI control problem, version 0.0001
Ω
Steven Byrnes
5y
Ω
7
60Conditioning Generative Models for Alignment
Ω
Jozdien
3y
Ω
8
59Gradient Descent on the Human Brain
Ω
Jozdien, gaspode
2y
Ω
5
58Interpretability’s Alignment-Solving Potential: Analysis of 7 Scenarios
Ω
Evan R. Murphy
3y
Ω
0
Load More (15/32)
Add Posts