This website requires javascript to properly function. Consider activating javascript to get access to all site functionality.
LESSWRONG
Tags
LW
Login
Anthropic (org)
•
Applied to
Rishi Sunak mentions "existential threats" in talk with OpenAI, DeepMind, Anthropic CEOs
by
Baldassare Castiglione
13d
ago
•
Applied to
Request to AGI organizations: Share your views on pausing AI progress
by
Akash
2mo
ago
•
Applied to
Anthropic is further accelerating the Arms Race?
by
Ruby
2mo
ago
•
Applied to
Anthropic: Core Views on AI Safety: When, Why, What, and How
by
RobertM
3mo
ago
•
Applied to
Anthropic's Core Views on AI Safety
by
RobertM
3mo
ago
•
Applied to
Podcast Transcript: Daniela and Dario Amodei on Anthropic
by
Ruby
3mo
ago
•
Applied to
[Preprint] Pretraining Language Models with Human Preferences
by
Giulio
4mo
ago
•
Applied to
Paper: The Capacity for Moral Self-Correction in Large Language Models (Anthropic)
by
LawrenceC
4mo
ago
•
Applied to
My understanding of Anthropic strategy
by
Siddharth Hiregowdara
4mo
ago
•
Applied to
[Linkpost] Google invested $300M in Anthropic in late 2022
by
Akash
4mo
ago
•
Applied to
Concrete Reasons for Hope about AI
by
Andrea_Miotti
5mo
ago
•
Applied to
Why I'm joining Anthropic
by
Multicore
5mo
ago
•
Applied to
Will research in AI risk jinx it? Consequences of training AI on AI risk arguments
by
Yann Dubois
6mo
ago
•
Applied to
A challenge for AGI organizations, and a challenge for readers
by
Andrea_Miotti
6mo
ago
•
Applied to
The limited upside of interpretability
by
Peter S. Park
7mo
ago
•
Applied to
Transformer Circuits
by
Neel Nanda
8mo
ago
•
Applied to
Toy Models of Superposition
by
Ruby
9mo
ago
•
Applied to
Anthropic's SoLU (Softmax Linear Unit)
by
Joel Burget
1y
ago
•
Applied to
How do new models from OpenAI, DeepMind and Anthropic perform on TruthfulQA?
by
Owain_Evans
1y
ago