Previously in sequence: Doublecrux is for building products.
To recap – you might want to doublecrux if either:
- You're building a product, metaphorical or literal, and you disagree about how to proceed.
- You want to make your beliefs more accurate, and you think a particular person you disagree with is likely to have useful information for you.
- You just... enjoy resolving disagreements in a way that mutually pursues truth for whatever reason.
Regardless, you might find yourself with the problem:
Doublecruxing takes a lot of time.
For a 'serious' disagreement, it frequently takes a least an hour, and often much longer. Habryka and I once took 12 hours over the course of 3 days to make any kind of progress on a particularly gnarly disagreement. And sometimes disagreements can persist for years despite significant mutual effort.
Now, doublecruxing is faster than many other forms of truth-aligned-disagreement resolution. I actually it's helpful to think of doublecrux as "the fastest way for two disagreeing-but-honest-people to converge locally towards the truth", and if someone came up with a faster method, I'd recommend deprecating doublecrux in favor it that. (Meanwhile, doublecrux is not guaranteed to be faster for 3+ people to converge but I still expect it to be faster for smallish groups with particularly confusing disagreements)
Regardless, multiple hours is a long time. Can we do better?
I think the answer is yes, and it basically comes in the form of:
- Practice finding your own cruxes
- Practice helping other people find their cruxes
- Develop metacognitive skills that make cruxfinding natural and intuitive
- Caching the results into a clearer belief-network
Those are all things you can do unilaterally. If you get buy-in from your colleagues, you might also try something like "develop culture that encourages people to do those four things, and help each other to do so."
I'd summarize all of that as "develop the skill and practice of keeping your beliefs cruxy."
By default, humans form beliefs for all kinds of reasons, without regard for how falsifiable they are. The result is a tangled, impenetrable web. Productive disagreement takes a long time because people are starting from the position of "impenetrable web."
If you make a habit of asking yourself "what observations would change my mind about this?", then you gain a few benefits.
First, your beliefs should (hopefully?) be more entangled with reality, period. You'll gain the skill of noticing how your beliefs should constrain your anticipations, and then if they fail to do so, you can maybe update your beliefs.
Second, if you've cultivated that skill, then during a doublecrux discussion, you'll have an easier time engaging with the core doublecrux loop. (So, a conversation that might have taken an hour takes 45 minutes – your conversation partner might still take a long time to figure out their cruxes, but maybe you can do your own much faster)
Third, once you gotten into this habit, this will help your beliefs form in a cleaner, more reality-entangled fashion in the first place. Instead of building an impenetrable morass, you'll be building a clear, legible network. (So, you might have all your cruxes full accessible from the beginning of the conversation, and then it's just a matter of stating them, and then helping your partner to do so)
[Note: I don't think you should optimize directly for your beliefs being legible. This is a recipe for burying illegible parts of your psyche and missing important information. But rather, if you try to actually understand your beliefs and what causes them, the legibility will come naturally as a side benefit]
Finally, if everyone around you is doing, this radically lowers the cost of productive-disagreement. Instead of taking an hour (or three days), as soon as you bump into an important disagreement you can quickly navigate through your respective belief networks, find the cruxes, and skip to the part where you actually Do Empiricism.
I think (and here we get into speculation), that this can result in a phase shift in how disagreement works, enabling much more powerful discourse systems than we currently have.
I think keeping beliefs cruxy is a good example of a practice that is both a valuable "Rabbit Strategy", as well as something worth Stag Hunting Together on. (In Rabbit/Stag parlance, a Rabbit strategy is something you can just do without relying on anyone else to help you, that will provide benefit to you. A Stag strategy is something with a larger payoff, but only actually works if you've coordinated with people to do so)
If you have an organization, community, or circle of friends where many people have practiced keeping-beliefs-cruxy, I predict you will find individuals benefiting, as well as creating a truthseeking culture more powerful than the sum of its parts.
Hopefully obvious addenda, operationalizing:
My crux for all this is that I think this is locally tractable.
My recommendation is not that you go out of your way to practice this all the time – rather, that you spend some time on-the-margin practicing the "look-for-cruxes" whenever you actually have a noteworthy disagreement.
If you practice crux-finding whenever you have such a disagreement, and find that it isn't locally useful, I probably wouldn't recommend continuing.
If 10 people tried this and 6 of them told me it didn't feel useful, even superficially, I'd probably change my mind significantly. If they tried it continuously for a year and there weren't at least 3 of them that could point to transparently useful outcomes, (such as disagreements taking much less time than they'd naively predict), I'd also significantly change my beliefs here.
This operationalization has some important caveats, such as "I don't necessarily expect this to work for people who haven't read the sequences or some equivalent. I also think it might require some skills that I haven't written up explanations for yet –which I hope to do soon as I continue this sequence."
Another issue I've run into, somewhat confused about the right way to think about:
Recently, jimrandomh and I were having a debate about the new shortform features. I was in favor of a quick, hacky solution that just automated the system we were already using (i.e. "shortform" just means "you create a post called 'shortform' and then write comments on it). I expected to spend maybe 3 days on it, and then move onto other things. I mostly wanted to create features for it that would also be useful for non-shortform things (i.e. I wanted people to be able to subscribe to shortform feeds, but I also wanted people to be able to subscribe to posts generally).
Jim wanted to treat shortform more as a first class feature, with special UI and care taken to make it seem like not a random hacked together thing.
The conversation ended up touching on many things including:
We ended up successfully coming to a compromise on how to move forward, but not really resolving any of those issues.
And, a few weeks later... I have moved closer to Jim's position. But mostly for reasons unrelated to anything we talked about. (I'd summarize the change as mostly "hmm, now that I see shortform in action and how people are relating to it, it seems useful to lean in harder towards making it good." I still think hacked-together-looking features are fine, but it seems more strategically valuable to lean into it, hacky or no. We've spent 4 weeks on it instead of 3 days and I think it makes sense to spend more. Partly the shift is about the relative-value of the "push shortform" strategy vs the "open questions" strategy (which is what I would have previously pushed)
Now, doublecrux is ideally supposed to result in an experiment you can run. "Actually build the feature and see how you feel" is sort of obvious as an experiment to run. But, also an expensive one.
I feel like I'm supposed to learn some sort of lesson from this but not sure which one. "Start building a feature and see how you feel" seems a pretty okay, practical strategy. But I'm dissatisfied with it for purposes of developing good disagreement-resolution skills. Sometimes there won't be a 'fairly reasonable experiment'.
I have some sense that in the moment, it was possible for me to do some kind of "Step backwards and think about what might actually change my mind about the original disagreement, *without* getting bogged down in the weeds of Jim's sub-cruxes."
Pedantic correction: We didn't spend 4 weeks on it; while 4 weeks did pass, there was a lot of other stuff going on during that interval.