Getting LLMs to be deterministic when scoring the quality of a text is hard. If you ask ChatGPT to evaluate the same poem multiple times, you’ll get inconsistent responses. I’ve been thinking about whether there are ways to make LLM grading more consistent. We can take a hint from specific...
The following essay is largely plagiarized from Here is New York, by EB White. It is a miracle that Online works at all. The whole arrangement is improbable. People tap a piece of glass and expect, correctly, that their words will depart their room, enter a tangle of copper and...
If you accept the core premises of Eliezer's book, then you believe that we're building systems we cannot control.[1] Much of the field of AI alignment pretends: * we can control increasingly powerful AI systems * LLMs are the key AI system for us to learn how to control These...
TLDR: Method Iteration is a LLM prompting technique that causes better responses to hard problems. Some researchers think that for AI to solve truly hard problems, we need bigger models, more data, or new architectures. I wonder if there's another way. The text you get from an LLM is downstream...
In the face of any hard problem—reversing climate change, curing cancer, or starting a great novel—modern LLMs can generate thousands of possible solutions relatively cheaply. Most solutions from most prompts are bad: they’re not new relative to the state of the art, not feasible, or not significant enough. But for...
American democracy currently operates far below its theoretical ideal. An ideal democracy precisely captures and represents the nuanced collective desires of its constituents, synthesizing diverse individual preferences into coherent, actionable policy. Today's system offers no direct path for citizens to express individual priorities. Instead, voters select candidates whose platforms only...
This essay suggests the possibility that a loving superintelligence outcompetes a selfish superintelligence. Then, it recommends actions for AI labs to increase the chance of this possibility. The reasoning below is inspired primarily by Eliezer Yudkowsky, Joscha Bach, Michael Levin, and Charles Darwin. Superintelligence (SI) is near. Superintelligence will evolve...