LESSWRONG
LW

Wikitags

Human Values

Edited by plex last updated 16th Sep 2021

Human Values are the things we care about, and would want an aligned superintelligence to look after and support. It is suspected that true human values are , and could be extrapolated into a wide variety of forms.

Subscribe
1
Subscribe
1
Discussion0
Discussion0
highly complex
Posts tagged Human Values
259The shard theory of human values
Ω
Quintin Pope, TurnTrout
3y
Ω
67
95Human values & biases are inaccessible to the genome
Ω
TurnTrout
3y
Ω
54
63Multi-agent predictive minds and AI alignment
Ω
Jan_Kulveit
7y
Ω
18
52What AI Safety Researchers Have Written About the Nature of Human Values
avturchin
7y
3
41Requirements for a Basin of Attraction to Alignment
Ω
RogerDearnaley
1y
Ω
12
335. Moral Value for Sentient Animals? Alas, Not Yet
Ω
RogerDearnaley
2y
Ω
41
213. Uploading
Ω
RogerDearnaley
2y
Ω
5
18Ends: An Introduction
Rob Bensinger
10y
0
16Alignment has a Basin of Attraction: Beyond the Orthogonality Thesis
RogerDearnaley
1y
15
126. The Mutable Values Problem in Value Learning and CEV
Ω
RogerDearnaley
2y
Ω
0
5Utilitarianism and the replaceability of desires and attachments
MichaelStJules
1y
2
88[Valence series] 2. Valence & Normativity
Steven Byrnes
2y
7
32How Would an Utopia-Maximizer Look Like?
Ω
Thane Ruthenis
2y
Ω
23
104. A Moral Case for Evolved-Sapience-Chauvinism
Ω
RogerDearnaley
2y
Ω
0
167Shard Theory: An Overview
Ω
David Udell
3y
Ω
34
Load More (15/178)
Add Posts