This is a accompanying blog post to work done by Callum McDougall, Arthur Conmy and Cody Rushing as part of SERI MATS 2023. The work was mentored by Neel Nanda and Tom McGrath. You can find our full paper at https://arxiv.org/abs/2310.04625. We i) summarize our key results, ii) discuss limitations and future work and iii) list lessons learnt from the project
Key Results
Copy Suppression
In the paper, we define copy suppression as the following algorithm:
If components in earlier layers predict a certain token, and this token appears earlier in the context, the attention head suppresses it.
We show that attention head L10H7 in GPT2-Small (and to a lesser extent L11H10) both perform copy suppression across... (read 2220 more words →)
No, it seems highly unlikely. Considered from a purely commercial perspective - which I think is the right one when considering the incentives - they are terrible customers! Consider:
- They are close to a monopsony (as any one would want exclusivity), so the deal would have to be truly enormous to work.
- If the deal is enormous they have a huge incentive to cut us out, and the tech is very close to their core competencies.
- Whatever techniques end up being good are likely to be major modifications to training stack that would be hard to integrate, so the options for doing
... (read more)