LESSWRONG
LW

AI RiskOpenAIAI
Frontpage

1

OpenAI: Our approach to AI safety

by Jacob G-W
5th Apr 2023
1 min read
1

1

This is a linkpost for https://openai.com/blog/our-approach-to-ai-safety
AI RiskOpenAIAI
Frontpage

1

OpenAI: Our approach to AI safety
2RobertM
New Comment
1 comment, sorted by
top scoring
Click to highlight new comments since: Today at 7:42 PM
[-]RobertM2y20

Mod note: I've fixed the formatting for the quoted excerpts.

Reply
Moderation Log
More from Jacob G-W
View more
Curated and popular this week
1Comments

I appreciate the post and it has some interesting points.

We work hard to prevent foreseeable risks before deployment, however, there is a limit to what we can learn in a lab. Despite extensive research and testing, we cannot predict all of the beneficial ways people will use our technology, nor all the ways people will abuse it. That’s why we believe that learning from real-world use is a critical component of creating and releasing increasingly safe AI systems over time.

[...]

Importantly, we also believe that improving AI safety and capabilities should go hand in hand. Our best safety work to date has come from working with our most capable models because they are better at following users’ instructions and easier to steer or “guide.”

[...]

While we waited over 6 months to deploy GPT-4 in order to better understand its capabilities, benefits, and risks, it may sometimes be necessary to take longer than that to improve AI systems' safety. Therefore, policymakers and AI providers will need to ensure that AI development and deployment is governed effectively at a global scale, so no one cuts corners to get ahead. This is a daunting challenge requiring both technical and institutional innovation, but it’s one that we are eager to contribute to.