You are viewing revision 2.16.0, last edited by Ruby

Artificial Intelligence is the study of creating intelligence in algorithms. On LessWrong, the primary focus of AI discussion is to ensure that as humanity builds increasingly powerful AI systems, the outcome will be good. The central concern is that a powerful enough AI, if not designed and implemented with sufficient understanding, would optimize something unintended by its creators and pose an existential threat to the future of humanity. This is known as the AI alignment problem.

Common terms in this space are superintelligence, AI Alignment, AI Safety, Friendly AI, Transformative AI, human-level-intelligence, AI Governance, and Beneficial AI. This entry and the associated tag roughly encompass all of these topics: anything part of the broad cluster of understanding AI and its future impacts on our civilization deserves this tag.

See also General Intelligence.

Basic Alignment Theory

AIXI
Corrigibility
Decision Theory
Embedded Agency
Fixed Point Theorems
Goodhart's Law
Inner Alignment
Instrumental Convergence
Logical Induction
Mesa-Optimization
Myopia
Newcomb's Problem
Optimization
Orthogonality Thesis
Outer Alignment
Solomonoff Induction
Utility Functions

Engineering Alignment

AI Boxing (Containment)
Debate
Factored Cognition
Humans Consulting HCH
Impact Measures
Iterated Amplification
Value Learning

 

Strategy

AI Progress
AI Risk
AI Services (CAIS)
AI Takeoff
AI Timelines

Other

Centre for Human-Compatible AI
Future of Humanity Institute
GPT
Machine Intelligence Research Institute
OpenAI
Ought
Research Agendas