ChatGPT (and now GPT4) is very easily distracted from its rules
Summary Asking GPT4 or ChatGPT to do a "side task" along with a rule-breaking task makes them much more likely to produce rule-breaking outputs. For example on GPT4: And on ChatGPT: Distracting language models After using ChatGPT (GPT-3.5-turbo) in non-English languages for a while I had the idea to ask...
Mar 15, 2023180