I've laid out a concrete example of this at https://www.lesswrong.com/posts/FgXjuS4R9sRxbzE5w/medical-image-registration-the-obscure-field-where-deep , following the "optimization on a scaffold level" route. I found a real example of a misaligned inner objective outside of RL, which is cool
I believe that current architecture of neural networks supports mesa-optimization: generally speaking, searching across some vectors in order to select one of them, which will be most useful for producing an answer.
Three inner optimization ways are already possible, and most likely there will be new ones.
Let's suppose the neural network consists of layers, as common now - composition of matrix multiplication and activation function - and activation function .
This construction can be extended to select maximum out of options in k layers; possibly, even options.
I believe that inner optimization might exist in current neural networks, and that it can be used as evidence to approximate what future AIs can do at what levels of capability.