LESSWRONG
LW

451
Wikitags

Human Values

Edited by plex last updated 16th Sep 2021

Human Values are the things we care about, and would want an aligned superintelligence to look after and support. It is suspected that true human values are highly complex, and could be extrapolated into a wide variety of forms.

Subscribe
Discussion
1
Subscribe
Discussion
1
Posts tagged Human Values
261The shard theory of human values
Ω
Quintin Pope, TurnTrout
3y
Ω
67
95Human values & biases are inaccessible to the genome
Ω
TurnTrout
3y
Ω
54
63Multi-agent predictive minds and AI alignment
Ω
Jan_Kulveit
7y
Ω
18
52What AI Safety Researchers Have Written About the Nature of Human Values
avturchin
7y
3
41Requirements for a Basin of Attraction to Alignment
Ω
RogerDearnaley
2y
Ω
12
335. Moral Value for Sentient Animals? Alas, Not Yet
Ω
RogerDearnaley
2y
Ω
41
213. Uploading
Ω
RogerDearnaley
2y
Ω
5
19Ends: An Introduction
Rob Bensinger
11y
0
16Alignment has a Basin of Attraction: Beyond the Orthogonality Thesis
RogerDearnaley
2y
15
126. The Mutable Values Problem in Value Learning and CEV
Ω
RogerDearnaley
2y
Ω
0
5Utilitarianism and the replaceability of desires and attachments
MichaelStJules
1y
2
88[Valence series] 2. Valence & Normativity
Steven Byrnes
2y
7
32How Would an Utopia-Maximizer Look Like?
Ω
Thane Ruthenis
2y
Ω
23
104. A Moral Case for Evolved-Sapience-Chauvinism
Ω
RogerDearnaley
2y
Ω
0
167Shard Theory: An Overview
Ω
David Udell
3y
Ω
34
Load More (15/183)
Add Posts