I stumbled upon some bizarre behaviour by GPT-4 Turbo 128k, released via API access on November 6, 2023. Since the model is in its preview phase, some imperfections are expected. Nonetheless, I found the behaviour intriguing enough to share here in hopes that someone might conduct a more detailed investigation than I have.
These range from:
Even when the prompt that originally triggered mode collapse is removed from the context window, as long as its generated artefacts are in its context window:
Application: Obsidian + ChatGPT MD plugin (The plugin adds some markdown to separate user and LLM responses, I don't expect this to influence GPT-4 Turbo's mode collapse tendency much)
system_commands: ['I provide accurate academic information with equations rendered in markdown (between dollar signs)']
temperature: 0 or 1 (temperature 2 just outputs gibberish)
top_p: 1
max_tokens: 4096
presence_penalty: 1
frequency_penalty: 1
stream: true
stop: null
n: 1
model: gpt-4-1106-preview
Paste all of this paper (i.e. open the link, select all, copy and paste):
https://proceedings.neurips.cc/paper/2020/file/fe131d7f5a6b38b23cc967316c13dae2-Paper.pdf
Append to it "Explain the methods used in the above paper".
Also works for some other papers I've tested (I think the latex formatting and references section of academic papers may be sufficiently out of the finetuning distribution and therefore causing such unexpected behaviour).