LESSWRONG
LW

Jemal Young
89Ω46240
Message
Dialogue
Subscribe

I specialize in regulatory affairs for AI-enabled Software as a Medical Device, and I'm interested in the safe development of frontier AI.

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
2Jemal Young's Shortform
2mo
2
No wikitag contributions to display.
Jemal Young's Shortform
Jemal Young2mo4-1

Not saying AI models can't be moral patients, but 1) if the smartest models are probably going to be the most dangerous, and 2) if the smartest models are probably going to be the best at demonstrating moral patienthood, then 3) caring too much about model welfare is probably dangerous.

Reply
The Sun is big, but superintelligences will not spare Earth a little sunlight
Jemal Young9mo20

You only set aside occasional low-value fragments for national parks, mostly for your own pleasure and convenience, when it didn't cost too much?

Earth as a proportion of the solar system's planetary mass is probably comparable to national parks as a proportion of the Earth's land, if not lower.

Maybe I've misunderstood your point, but if it's that humanity's willingness to preserve a fraction of Earth for national parks is a reason for hopefulness that ASI may be willing to preserve an even smaller fraction of the solar system (namely, Earth) for humanity, I think this is addressed here:

it seems like for Our research purposes simulations would be just as good. In fact, far better, because We can optimize the hell out of them, running it on the equivalent of a few square kilometers of solar diameter

"research purposes" involving simulations can be a stand-in for any preference-oriented activity. Unless ASI would have a preference for letting us, in particular, do what we want with some fraction of available resources, no fraction of available resources would be better left in our hands than put to good use.

Reply
What's a better term now that "AGI" is too vague?
Answer by Jemal YoungMay 31, 202430

I think the kind of AI you have in mind would be able to:

continue learning after being trained

think in an open-ended way after an initial command or prompt

have an

discover and exploit signals that were previously unknown to it

accumulate knowledge

become a closed-loop system

The best term I've thought of for that kind of AI is Artificial Open Learning Agent.

Reply
How do top AI labs vet architecture/algorithm changes?
Jemal Young1y12

Thanks for this answer! Interesting. It sounds like the process may be less systematized than how I imagined it to be.

Reply
How do top AI labs vet architecture/algorithm changes?
Jemal Young1y20

Dwarkesh's interview with Sholto sounds well worth watching in full, but the segments you've highlighted and your analyses are very helpful on their own. Thanks for the time and thought you put into this comment!

Reply
Model Organisms of Misalignment: The Case for a New Pillar of Alignment Research
Jemal Young2y10

I like this post, and I think I get why the focus is on generative models.

What's an example of a model organism training setup involving some other kind of model?

Reply
Which possible AI systems are relatively safe?
Answer by Jemal YoungAug 23, 202330

Maybe relatively safe if:

  • Not too big
  • No self-improvement
  • No continual learning
  • Curated training data, no throwing everything into the cauldron
  • No access to raw data from the environment
  • Not curious or novelty-seeking
  • Not trying to maximize or minimize anything or push anything to the limit
  • Not capable enough for catastrophic misuse by humans
Reply
What are the best non-LW places to read on alignment progress?
Jemal Young2y10

Here are some resources I use to keep track of technical research that might be alignment-relevant:

  • Podcasts: Machine Learning Street Talk, The Robot Brains Podcast
  • Substacks: Davis Summarizes Papers, AK's Substack

How I gain value: These resources help me notice where my understanding breaks down i.e. what I might want to study, and they get thought-provoking research on my radar.

Reply
Think carefully before calling RL policies "agents"
Jemal Young2y30

I'm very glad to have read this post and "Reward is not the optimization target". I hope you continue to write "How not to think about [thing] posts", as they have me nailed. Strong upvote.

Reply
“Unintentional AI safety research”: Why not systematically mine AI technical research for safety purposes?
Jemal Young2y10

Thanks for pointing me to these tools!

Reply
Load More
ontological crisis
2Jemal Young's Shortform
2mo
2
4Safe Search is off: root causes of AI catastrophic risks
5mo
0
4Can efficiency-adjustable reporting thresholds close a loophole in Biden’s executive order on AI?
1y
1
3How do top AI labs vet architecture/algorithm changes?
Q
1y
Q
5
27“Unintentional AI safety research”: Why not systematically mine AI technical research for safety purposes?
2y
3
11How might we make better use of AI capabilities research for alignment purposes?
Q
3y
Q
4