LESSWRONG
LW

170
Wikitags

AI Alignment Intro Materials

Edited by Raemon, Ruby, plex, et al. last updated 30th Dec 2024

AI Alignment Intro Materials are posts that help someone get oriented and skill up. Distinct from AI Public Materials is that they are more "inward facing" than "outward facing", i.e. for people who are already sold AI risk is a problem and want to upskill.
 

Some basic intro resources include:

  • Stampy's AI Safety Info (extensive interactive FAQ)
  • Scott Alexander's Superintelligence FAQ
  • The MIRI Intelligence Explosion FAQ
  • The AGI Safety Fundamentals courses
  • Superintelligence (book)
Subscribe
Discussion
2
Subscribe
Discussion
2
Posts tagged AI Alignment Intro Materials
84The Alignment Problem from a Deep Learning Perspective (major rewrite)
Ω
SoerenMind, Richard_Ngo, LawrenceC
3y
Ω
8
141Superintelligence FAQ
Scott Alexander
9y
39
32"Corrigibility at some small length" by dath ilan
Ω
Christopher King
2y
Ω
3
42A newcomer’s guide to the technical AI safety field
Ω
zeshen
3y
Ω
3
239AI Control: Improving Safety Despite Intentional Subversion
Ω
Buck, Fabien Roger, ryan_greenblatt, Kshitij Sachan
2y
Ω
24
72Alignment Org Cheat Sheet
Orpheus16, Thomas Larsen
3y
8
69How to pursue a career in technical AI alignment
Charlie Rogers-Smith
3y
1
103A short course on AGI safety from the GDM Alignment team
Ω
Vika, Rohin Shah
7mo
Ω
2
83Wikipedia as an introduction to the alignment problem
Ω
SoerenMind
2y
Ω
10
83Outreach success: Intro to AI risk that has been successful
Michael Tontchev
2y
8
61My first year in AI alignment
Alex_Altair
3y
10
55A starter guide for evals
Ω
Marius Hobbhahn, Jérémy Scheurer, Mikita Balesni, rusheb, AlexMeinke
2y
Ω
2
36UC Berkeley course on LLMs and ML Safety
Ω
Dan H
1y
Ω
1
27Advice for Entering AI Safety Research
scasper
2y
2
26Talk: AI safety fieldbuilding at MATS
Ryan Kidd
1y
2
Load More (15/50)
Add Posts