Critiquing Risks From Learned Optimization, and Avoiding Cached Theories
What I'm doing with this post and why I've been told that it's a major problem (this post, point 4), of alignment students just accepting the frames they're given without question. The usual advice (the intro of this) is to first do a bunch of background research and thinking on...
I ended up going in a completely different direction with this: I intend to test my OWN rationality, and I figure that if rationality is about WINNING, about being EFFECTIVE, then I ought to find direct measures of the things I want, and test myself in 6 months or so (timeframe dependent on the toughness/length of the task). This will, in other words, be a test of my ability to understand the territory insofar as that understanding makes me more effective at a given task.
The things in particular, a few subgoals of my personal life-optimization:
- artistic endeavors and life enjoyment: engaging in things like art or gamedev or other mediums while aiming for
... (read more)