LESSWRONG
LW

Wikitags

Value alignment problem

Edited by Eliezer Yudkowsky, et al. last updated 2nd Feb 2017

Disambiguation: For the research subject that includes the entire edifice of how and why to produce good AIs, see AI alignment.

The 'value alignment problem' is to produce sufficiently advanced machine intelligences that want to do beneficial things and not do harmful things. The largest-looming subproblem is 'value identification' or 'value learning' (sometimes considered synonymous with value alignment) but this also includes subproblems like Corrigibility, that is, AI values such that it doesn't want to interfere with you correcting what you see as an error in its code.

Parents:
AI alignment
Children:
Total alignment
Preference framework
5
5
Discussion0
Discussion0