LLM Self-Reference Language in Multilingual vs English-Centric Models
My first post explored "self-talk" induction in small base LLMs. After further contemplation, I decided that first I ought to better understand how LLMs represent "self" mechanistically before examining induced "self-talk". How do language models process questions about themselves? I've started by analyzing attention entropy patterns[1] across self-referent prompts ('Who...
If we're going to evaluate AI's impact on software engineering, or really any industry, we need to measure throughput. I suppose with code this could be something like unaltered lines that go directly from AI to prod, but even that might not capture the effectiveness of the AI vs the effectiveness of the prompt writer.
I have no doubt it's been a boon for research (it has helped me do rapid prototyping), but we do need some better measures of its productivity. Right now all we have are high-level proxies like labor market impacts and low-level noise like volume of code.