Redundant Attention Heads in Large Language Models For In Context Learning — LessWrong