Fuzzing LLMs sometimes makes them reveal their secrets — LessWrong