Steganography via internal activations is already possible in small language models — a potential first step toward persistent hidden reasoning. — LessWrong