x
Subliminal Preference Transfer in LLMs: Do Models Learn More Than We Intend? — LessWrong