The OpenAI playground for GPT-3 is a terrible interface. Is there any great local (or web) app for exploring/learning with language models?
I'm thinking something that you add your API keys, and it has a pleasant interface for writing prompts, seeing results, and storing all the prompts and outputs for your reference. Maybe integrating with more general tools for thought. This is what I've found so far that attempts to let you...
Assuming this is verified, contrastive decoding (or something roughly analogous to it) seems like could be helpful to mitigate this? There are many variants, but one might be actually intentionally training both the luigi and waluigi, and sampling from the difference of those distributions for each token. One could also just do this at inference time perhaps, prepending a prompt that would collapse into the waluigi and choosing tokens that are the least likely to be from that distribution. (Simplification, but hopefully gets the point across)