Do AI agents need "ethics in weights"?
In this article, I’m trying to argue why outer alignment is preferable and where, in my opinion, the error lies. I also explain why ethics must be part of the task, not embedded in the weights. Perhaps I’m wrong. But I believe it is necessary to consider any ideas in...
Nov 4, 20251