Nice little video - audio is Neel Nanda explaining what mechanistic interpretability is and why he does it, and it's illustrated by the illustrious Hamish Doodles. Excerpted from the AXRP episode.
(It's not technically animation I think, but I don't know what other single word to use for "pictures that move a bit and change")
Lots of alpha in AI research distillers learning motion-canvas/motion-canvas: Visualize Complex Ideas Programmatically (github.com) and making explainers.
Or even better, finetuning an LLM to automate writing the code!
just don't use an overly large model.
For those reading (I imagine Sheikh knows about these already), some videos from the creator of that library:
A single word for this would be an animatic, probably.
I kinda guess that most people don't know what that means.
Here's a dumb idea: if you have a misaligned AGI, can you keep it inside a box and have it teach you some things about alignment, perhaps through some creative lies?