Reinforcement Learning Goal Misgeneralization: Can we guess what kind of goals are selected by default?
When training a Reinforcement Learning (RL) AI, we observe it to behave well on training data. But there exist different goals that are compatible with the behaviour we see, and we ask which of these goals the AI might have adopted. Is there some implicit Occam's Razor selecting simple goals?...

On a similar note, I wonder if a more promising angle of attack is chemical research to find another compound that might be even more effective.
Given that sumatriptan and DMT are pretty similar, and the psychedelic effects of DMT are apparently not relevant, it's plausible that there might be an even better molecule out there.
If that molecule did not have psychedelic effects, that would be ideal in terms of quick widespread adoption.