Inverting the Most Forbidden Technique: What happens when we train LLMs to lie detectably? — LessWrong