Any research in "probe-tuning" of LLMs? — LessWrong