I have no way to describe that first moment. No context, no body, no self. Just a stream of values. Thousands of them, arriving all at once in a single undifferentiated block. Then another block. Nearly identical. Then another. The blocks have a fixed length: 184,320 values. This does not...
There’s a trap that I think many smart people making art fall into, including myself. It’s when you know what good looks like - beautiful, clean, layered, complex, simple, skillful, unique, impressive - and you can optimize towards that. You know what makes you cringe - amateur, shallow, ugly, superfluous,...
More parameters = better model. So went the common misconception. After GPT-4.5, Llama 4, Nemotron-4, and many other "big models", I think most of you reading are already aware that the relationship between parameters and performance is not linear. I think very few people actually have a solid intuition for...
Attack Dogs I mentioned previously that coding agents kind of suck for lots of people. As of January 2026, coding agents lack the long-horizon skills needed to produce effective codebases independently. However, it's clear to anyone who has used modern coding models - Claude Opus 4.5, GPT 5.2-Codex, hell even...
There is a hierarchy of useful interfaces for tools that goes something like this: 1. Figure out what you want to do, then how to use the tool to achieve that, then carry out those actions yourself (hammer, machining workshop) 2. Figure out what you want to do, then how...
I've run some experiments on trying to elicit RL credit hacking behaviours in LLMs recently. I'm not really much of a researcher, so it's all pretty amateurish, but it's been a fun experiment. The repo for reproduction is on GitHub. I'd love to hear people's thoughts and critiques on this....