Some of you have been trying to raise money for the Singularity Institute, and I have an idea that may help.
The idea is to hold public competitions on LessWrong with money going to charity. Agree to a game and an amount of money, then have each player designate a charity. After the game, each player gives the agreed upon amount to the charity designated by the winner.1 It’s a bit like celebrity Jeopardy.2
Play the game here on LessWrong or post a record of it.3 That will spread awareness of the charities and encourage others to emulate you.
The game can be as simple as a wager or something more involved:
- Notations exist for go and chess.
- The AI Box Experiment is essentially an arguing game.
- Play anything else you can agree on.4
We’ve already played few games on LessWrong, more on those in a moment.
In most ways the AI problem is enormously more demanding than the personal art of rationality, but in some ways it is actually easier. In the martial art of mind, we need to acquire the realtime procedural skill of pulling the right levers at the right time on a large, pre-existing thinking machine whose innards are not end-user-modifiable.
First, I have a confession to make: I don’t really care how much money gets donated to the Singularity Institute, nor am I trying to drum up money for some other cause. I mainly want you all playing games.
Not just playing them, of course. Playing them here in front of the rest of LessWrong and analyzing the moves in terms of the “personal art of rationality.”
We need more approaches to improvement. Even in Eliezer_Yudkowsky’s Bayesian Conspiracy fictional series/manifesto, many other schools of thought (called, for dramatic effect, “conspiracies”) were present. As I recall, the “Competitive Conspiracy” was mentioned frequently, but there are other reasons for choosing to start with games.
Games are fun, of course. They also deal with the “personal art” of getting familiar with your own brain, which I think has been underrepresented on LW. I do believe there are certain important things we can’t learn properly just by reading, arguing, and doing math (valuable as those techniques are). Games are an easy, intuitive first step to filling that gap.
We’ve already had a few games here. Warrigal held an Aumann’s agreement game competition. I created Pract and played it with wedrifid. Neither one has caught on as a LessWrong pastime, but the comments revealed that people here know many interesting games.
And there’s the donation hook. Some of you believe the world is at stake, so that’s a nice motivator.
Of course it’s not always the case that you have exactly one winner. I suggest that if there’s no winner each player give to their own charity, and if there are multiple winners each player splits up their donation evenly among the charities of the winners. ↩
Unless it contains some remarkable insight, I wouldn’t make it a top-level post. The comment area here or in one of the open threads would be good. ↩
I see mutual consent as an important element of games. ↩