LESSWRONG
LW

Wikitags

Language Models (LLMs)

Edited by Yaakov T, plex, et al. last updated 30th Dec 2024

Language Models are computer programs made to estimate the likelihood of a piece of text. "Hello, how are you?" is likely. "Hello, fnarg horses" is unlikely.

Language models can answer questions by estimating the likelihood of possible question-and-answer pairs, selecting the most likely question-and-answer pair. "Q: How are You? A: Very well, thank you" is a likely question-and-answer pair. "Q: How are You? A: Correct horse battery staple" is an unlikely question-and-answer pair.

The language models most relevant to AI safety are language models based on "deep learning". Deep-learning-based language models can be "trained" to understand language better, by exposing them to text written by humans. There is a lot of human-written text on the internet, providing loads of training material.

Deep-learning-based language models are getting bigger and better trained. As the models become stronger, they get new skills. These skills include arithmetic, explaining jokes, programming, and solving math problems.

There is a potential risk of these models developing dangerous capabilities as they grow larger and better trained. What additional skills will they develop given a few years?

See also

  • GPT - A family of large language models created by OpenAI
Subscribe
2
Subscribe
2
Discussion2
Discussion2
Posts tagged Language Models (LLMs)
662Simulators
Ω
janus
3y
Ω
168
369How it feels to have your mind hacked by an AI
blaked
3y
222
265Alignment Implications of LLM Successes: a Debate in One Act
Ω
Zack_M_Davis
2y
Ω
56
137How LLMs are and are not myopic
Ω
janus
2y
Ω
16
105On the future of language models
owencb
2y
17
93Inverse Scaling Prize: Round 1 Winners
Ω
Ethan Perez, Ian McKenzie
3y
Ω
16
65How to Control an LLM's Behavior (why my P(DOOM) went down)
Ω
RogerDearnaley
2y
Ω
30
64A "Bitter Lesson" Approach to Aligning AGI and ASI
Ω
RogerDearnaley
1y
Ω
41
47Try training token-level probes
StefanHex
5mo
6
144Transformer Circuits
Ω
evhub
4y
Ω
4
48Goodbye, Shoggoth: The Stage, its Animatronics, & the Puppeteer – a New Metaphor
Ω
RogerDearnaley
2y
Ω
8
44Mlyyrczo
lsusr
3y
14
37Striking Implications for Learning Theory, Interpretability — and Safety?
RogerDearnaley
2y
4
35Motivating Alignment of LLM-Powered Agents: Easy for AGI, Hard for ASI?
RogerDearnaley
2y
4
20A Chinese Room Containing a Stack of Stochastic Parrots
RogerDearnaley
2y
3
Load More (15/726)
Add Posts