Conditioning Generative Models for Alignment — LessWrong