Splendid post. I am generally pro CFAR being alive; I was also pleased to read that the new workshops will still be ~2/3rds the good content from the original workshops, not exclusively new and experimental stuff, which makes me more confident in encouraging people to go (i.e. that the floor on the experience will still be quite good). Many things in this post seem to me to accurately address pathologies in CFAR 1.0; here's to CFAR 2.0 having even more success in developing an art of human rationality than CFAR 1.0.
Ways to help CFAR or to connect to CFAR besides donating:
[...]
- Book our venue (or help a friend realize they’d enjoy booking the venue, if they would)
I'll also let people know that I had a great experience renting the CFAR venue in Bodega Bay. I took the Inkhaven residents there for a weekend off-site, and the participants rated it really highly. It was a great bonding experience to be in a single house altogether, we had nice daily walks to the ocean, and Jack & Sunny were lovely hosts with their two adorable kittens. Endorsed as a good getaway space for up to ~50 people.
(And we do still need money to be viable, because being a custodian of such a community requires staff time and money for food/lodging/staff flights/etc.)
As a minor issue, I think I'm failing to understand this parenthetical. I already believe that many good non-profits need donations to live, and cannot sustain themselves fully on sales and revenue. This seemed to read to me though that aCFAR is justifying itself as needing funds primarily because of trying to sustain a community. Slightly earlier you wrote about the alumni community, which you felt was originally quite generative, then became lower quality, and you'd like to do something to get life into one again. But I don't think you mean to imply that the alumni community is the sole purpose of donations. What did you mean here?
This seemed to read to me though that aCFAR is justifying itself as needing funds primarily because of trying to sustain a community. ... What did you mean here?
From my POV, the main reason I want aCFAR in existence is to sustain (and allow to better-develop) a particular angle on rationality practice.
Like, suppose there were a bunch of people doing a bunch of bits of neat music, and you sorta had the idea that a jazz-type thing was nascent. And then maybe if you create a really cool bar or something with the right set of musicians and fans and stuff invited at once (a center), jazz congeals and becomes a thing.
My feeling is there's a particular take on rationality that I want to help pop into existence. There's a bunch of rhymes between stuff different ones of us staff are doing, near also Friedrich Hayek's work and Christopher Alexander's. I think we're onto something, I want to see where it goes. And I think this thing develops best via practicing it together, in various ways -- aCFAR staff trying it on ourselves and each other and volunteers; folks who don't work for aCFAR but who are independently doing "rhyming" stuff coming and riffing and showing us there angles; the bits and pieces combining; etc.
This is the "community" (of creating / practicing/ refining) that I meant we are trying to cultivate. (Not our new alumni community, although it'll overlap it.)
Does that make sense?
Hi Anna, thanks for laying out your perspective so extensively! In the spirit of having real conversations, there's a high-level thought that's been bopping around my mind for a while that I wanted to surface here.
When I went to a CFAR workshop in 2018 (specifically a MIRI/CFAR workshop) I wasn't really aware of anyone trying to do the same thing that CFAR was doing, in terms of making people more "rational" (except for maybe the Leverage workshops).
Since then, I've moved towards believing that emotional blocks are the main reason why people make bad decisions, and that you can help them dramatically improve their lives via removing those blocks. I get the sense that aCFAR has also moved in this direction (perhaps not as much as I have).
But there are a lot of workshops that aim to do roughly this; an incomplete list includes:
(Then there's a larger ecosystem of workshops which focus on more niche emotional processing techniques, like bodywork, breathwork, shadow work, sexual healing, drug-enabled healing, etc. I don't want to use these as central examples but do want to note that they exist and seem to often be quite powerful.)
I'm curious how you orient towards these kinds of things (which I'll broadly gloss as the "neo-hippie" community). I can imagine you having a range of different orientations, like:
(I have some of my own speculations about what the best approach is, but this feels like a good place to get more data from you before laying out my perspective further.)
Hi Richard, thanks for the engagement! My short answer is that I'm betting there's magic in doing two things at once:
The "neo-hippie" community you reference seems to me to include some gold near #2 (although some gaps, too, linked IMO sometimes to a lack of sufficient #1), but seems far worse than LW, aCFAR, etc at #1.
I'll try to comment back later with why exactly I want to bet on this combination. (Still thinking/articulating.)
Why I'm betting on doing both of these things at once: from my POV they're basically one thing. And that one thing is: having a go at the world with everything we've got.
When it comes to figuring out what's true: I want to let my perceptions "bubble up" and to let bits of how a given perception does/doesn't reconcile with other perceptions also bubble up, and to actively chew on the puzzle-points and maybe actively grab info that might help me make sense of things. Where the purpose of all that is to get the best guess I can about what's going on in the world. This is how Gendlin's Focusing comes to be useful in scientific research: a scientist notices their inclinations about a scientific puzzle, notices what perceptions/intuitions/frames/etc bubble up as they stare at it, notices how their current intuitions don't quite feel right (Eliezer's "tiny note of discord" / "noticing confusion"), etc. Which helps in science according to both Gendlin and our own attempts at CFAR (eg, Edward Kmett got a small breakthrough on a thing he'd been stuck on for awhile when he tried the "Focusing for research" class).
IMO this stuff is bottom-up-ish, mixed with top-down replies to and knitting together of the bottom-up bits, and applies in the absence of any particular emotional blocks, and also with more difficulty in their presence.
When it comes to accomplishing cool stuff [the other half of being human, besides figuring out what's true], I think there's a fairly similar structure, where [energy / morale / traction / a calibrated "yes I can do something here, let me bet on that and have a go] bubbles up from different models + perceptions ("someone on the internet is wrong! and I can help!"). And I can do top-down-ish things to help many different bits of the bottom-up morale-pieces cohere into a single project that is "onto something". (And to help disband that project's reified bits after a given one turns empty, so I can go back to locating something of quality and again be onto something.)
I'm not sure how to say this well.
But the usefulness of mechanistic reasoning seems like one of our best clues about the universe, and I sort of can't imagine someone making a serious go at the above without including it.
Why I'm betting on doing both of these things at once: from my POV they're basically one thing. And that one thing is: having a go at the world with everything we've got.
Yeah, to be clear, this is totally what I'm doing as well, and I strongly empathize with this sentence.
A better phrasing would not have been for me to challenge doing them at once, but rather to challenge learning them at once. Let me try to be concrete here. Some people have a level of stoicism or repression that means that they've almost never cried. For those people, getting to a level of emotional release that allows crying is a pretty important step towards emotional health. After that point, it's pretty powerful for them to mix that skill with a simultaneous "orienting towards the world" move (as we both observed at LARC). But when I look through the classes listed in the CFAR handbook, almost all of them seem pretty structurally antithetical to "learning to cry", because they are trying to get participants to absorb a primarily-epistemic frame.
(Though to be clear, I'm not saying that CFAR should be optimizing for making people cry. Running workshops where a lot of people cry is a risky endeavor. Again, it's more about the meta-level question of how aCFAR orients to all the crazy powerful stuff lying around.)
When it comes to accomplishing cool stuff [the other half of being human, besides figuring out what's true]
Not super important but wanted to flag that this feels like a very impoverished description of being human. There's also, you know, having fun, loving others, relaxing, etc.
I'm a fan of both these things; I think the crux is whether people should do them at once.
One reason I'm skeptical of that is that the point of a lot of neo-hippie techniques is to get people out of their "mechanistic reasoning" frame, e.g. to get them to start saying things that they feel without first checking whether those things are true. And more generally to let more aspects of one's psyche "bubble up", partly by creating a space where it feels acceptable to relax one's top-down control, and partly via practices that help channel attention to what's going on somatically. Lately I've been increasingly excited about practices that don't even involve talking at all (e.g. body work, breath work), since that's an easy way to quieten down the mechanistic reasoning parts.
On the other hand, I do think that there's something high-variance about trying to separate #1 and #2 too much. I feel like I haven't adopted many wild beliefs as a result of doing #1 and #2 separately, but I know people who have. But there might also be significant selection effects at play.
Overall I think the thing I'm actually trying to gesture at is something like:
From my perspective, CFAR is operating in a space where there are crazy powerful techniques lying around all over the place, many of which seem to have somewhat harmful side effects as typically applied. In part these harmful side effects seem to be related to how it's hard for a community to push frontiers without "going off the deep end".
I can imagine a version of aCFAR which internalizes insights from the people at the frontier of emotional processing (like the "mechanistic reasoning often inhibits emotional processing" insight I outlined above) while deploying them more carefully. But I can also imagine a version of aCFAR which is... let's say, structurally blind to what actually works re emotional processing, because of various aspects of its self-image. For example, a lot of the things I mentioned in my previous comment as "niche emotional processing techniques" are things that it would be very hard for CFAR to pivot to focusing on! But there's a big difference between a CFAR which is thinking "are these things we can't do actually the most powerful things, and if so can we approximate them?" and a CFAR which is thinking "those things are basically outside our domain, which is more anchored to 'rationality'".
My main update from writing out the above explicitly, btw, was that the "CFAR" branding might actually be a big hinderance, not so much because of external reputation but because it locks you into mentally orienting to the community too similarly to how y'all did in 2012-2020.
My other big update: I think of the neo-hippie community as a massive sprawling Thing. But in several other cases where I've thought that, I've turned out to be wrong—actually there was a relatively small set of people actually doing novel stuff, and they were fairly socially proximate to me. So it's possible that when I point to "the neo-hippie community" I'm pointing to a thing which is much more fleshed out in my map of the world than in other people's (potentially including yours). If so we should chat privately about what specifically I'm talking about and why I think it's valuable.
EDIT: a third update: man, I really want someone to actually try to "solve psychology" by combining insights from neo-hippies with systematizing thinking. Maybe this shouldn't be CFAR's job—I guess it used to be Leverage's job, and that got kinda complicated—but I really want it to happen, and I don't think academia will manage it any time soon. Probably a bunch of my comment above was written from a perspective of hoping for that, which may end up tilting my advice in unproductive directions compared with advice that's trying to help you achieve the thing you are actually aiming for (whatever that happens to be).
It seems like in your model mechanistic reasoning is more of a top down thing? It’s not in mine fwiw and maybe this points to an interesting disagreement
Yepp, sounds like a thing we should chat about.
I’m using “top-down” and “bottom-up” pretty loosely, but some intuitions to help triangulate:
I just replied to a chunk of this. To respond to a couple other chunks:
that the point of a lot of neo-hippie techniques is to get people out of their "mechanistic reasoning" frame, e.g. to get them to start saying things that they feel without first checking whether those things are true. And more generally to let more aspects of one's psyche "bubble up",
IMO, it's indeed good (and part of the art of rationality) to notice what you're perceiving without first "checking whether [your perception] is true." And then it's good to check afterward. Terry Pratchett has some solid basic rationality in his delightful novel "The Wee Free Men," in which his protagonist succeeds using "first sight, and second thought" -- "first sight" to see what her eyes are actually reporting, without rounding to what she was expecting to see, and "second thought" to double check her thinking.
So I believe the neo-hippies are inline with traditional rationality on this point. (IMO, traditional rationality doesn't want "mechanistic reasoning," authority, peer review, your priors, or anything else gating full conscious access to your perceptions, nor the ability to discuss your perceptions.)
I realize you said "saying things they feel", whereas I said "notice what you perceive". But "feel" in neohippy circles usually refers to a perception or belief (e.g., "I feel this is risky"). If we discuss only cases where it refers to an emotion ("I feel scared"), well, for those too I'd like to allow them into full consciousness.
man, I really want someone to actually try to "solve psychology" by combining insights from neo-hippies with systematizing thinking.
It makes sense to me that you want that! But it's not what I want to do, nor what I want aCFAR to do. Maybe try Chris Lakin? I believe he's trying for something like this, and I cheer for his efforts.
I don't desire to try this myself, because... my current angle is not [seeing people as things to solve or help, using techniques and models]. It's more [seeing people as wise self-authored creatures to befriend and build things with, who admittedly have a bunch of ~trauma and parasitic memes and stuff, but who even so mostly need listening and support in their ~autonomous process of making sense of the world and doing things they care about].
In other words, I want something more "I-and-thou" with the people we're interacting with, compared to "solving psychology" which is more I-it.
Past-me thinks that's an odd thing for me to trust in, in light of [phenomenon: sometimes I or others thought we handed someone a tool, but actually we handed it to some not-so-aligned-meme that's squatting in them, and that now uses this tool to get more control.] My current take, though, is that while [accidentally up-powering parasitic memes, to ill effect] definitely happens, this can be mostly avoided by a culture of assisting a central self, who is also a point of pride/authorship, to come more fully online and to notice where they do and don't have distaste for what's up. (This is why I like Focusing but don't like IDC -- only Self should be dialog with parts.) (It's different from pastCFAR!agentiness, which aimed to empower a system 2 virtual self, rather than the Self Gendlin and IFS talk about).
Somehow this is bouncing off me. We should probably talk more directly about it. I'll quickly give two pointers to what feels like it's bouncing off:
So I believe the neo-hippies are inline with traditional rationality on this point.
This is enough of a stretch in my ontology that I think we're probably talking past each other.
I don't desire to try this myself, because... my current angle is not [seeing people as things to solve or help, using techniques and models]. It's more [seeing people as wise self-authored creatures to befriend and build things with...]
Yeah, I very much got the latter vibe from LARC. But the former vibe does seem pretty hard to avoid in context where there are designated teachers, and they're teaching students, and it's a time-limited container with no expectation of future interaction, and so on.
I also notice that things like meditation traditions are super into a kind of "everyone's a wise self-authored creature" vibe while also being super into a "here is the tradition of knowledge I'm in, which y'all have come to learn from and be helped by" vibe, and don't think of them as in conflict.
Hence there's something here where isn't quite clicking for me. But you shouldn't feel pressure to try to explain it in this format.
I agree this conversational format may not work, but I'm finding it helpful for now. I'm gonna keep saying piecemeal bits I care about. (In a more proper conversation, I'd take more care to try to paraphrase you first etc. But I'm liking you sharing the bits you're sharing.)
I agree one can participate in a tradition of deep know-how (even as a teacher) without being I-it about it. Good cultures are typically like this IMO. I was complaining about "solving psychology," not about "drawing on a tradition." Chapter three of CS Lewis's "The Abolition of Man" points to the difference, if you've read it or feel like reading it. A passage:
In older systems both the kind of man the teachers wished to produce and their motives for producing him were prescribed by the Tao -- a norm to which the teachers themselves were subject and from which they claimed no liberty to depart. They did not cut men to some pattern they had chosen.... They initiated the young neophyte into the mystery of humanity which overarched him and them alike. It was but old birds teaching new birds to fly. This will be changed. Values are now mere natural phenomena.... It is one more part of Nature which they have conquered.
(If you somehow feel like investing time in this subpoint, the place I'm coming from here will likely make more sense if you read "The abolition of man"; also happy to try by talking here.)
I would ideally like a tradition where pieces can be individually tested and thrown out, but where it's still a tradition, not an I-it.
...
I agree there are many known psychological interventions out there of large effect-size and unclear sign, most clearly recreational drugs, but also including e.g. AoA and bodywork. It is at this point a clearly bad idea to go "I'll go hunting for things with large effects on psychology of unknown sorts, try lots of them, gather people who can try lots of them in dialog, and call this good."
What's needed instead is way to tell what's good.
The rationality community's traditional answer to "is X psychological intervention good?" is "does it help you form more accurate beliefs" (epistemic rationality) and "does it help you hit your stated / verbally endorsed goals" (instrumental rationality). I still think this notion of "good" helped build stuff that was, in fact, good mostly. (But limited in e.g. not helping the person you mention who doesn't know how to cry, and in setting people up for burnout, and many similar gaps).
Our new stuff at aCFAR is mostly a small but significant dif from the traditional rationality definition. We keep "form true beliefs" as a goal. But in place of the "instrumental rationality" part: instead of working to aid "agency" and verbally endorsed goals (and the ego who verbally endorses those goals), we're working with:
These are all attempts to empower something like "be a cohesive self, who doesn't glide past tiny notes of discord but instead tunes into all the data coming in, and is in full contact with the outside world, including all known science."
(I expect to be more articulately coherent on this in about six months.)
ETA: I want one self encountering the outside world full-on. My objection to "solve psychology" as an angle on things, is that it divides me in two: a "psychologist" part, and a "have psychology done to me" part.
(I imagine you, Richard, agree that the goal is like this, but then think the route to this goal is more like "do psychology" / "have psychology done to you"?)
Do you want to try a LessWrong dialog? Are those still a thing?
ETA: Or, if they're not (or even if they are), maybe a quickly-written coauthored post where we try to exposit where we're each coming from, what we agree about already, where we seem to have different inclinations, etc?
EDIT: a third update: man, I really want someone to actually try to "solve psychology" by combining insights from neo-hippies with systematizing thinking. Maybe this shouldn't be CFAR's job—I guess it used to be Leverage's job, and that got kinda complicated—but I really want it to happen, and I don't think academia will manage it any time soon.
FWIW, in a way that's totally unrelated to what CFAR is up to, I'm looking to do something very close to this. I think subjective science is real and doable, and I'm pretty excited to make it happen. Step two might be to mimic the history of chemistry. But step one is more like a "literature review" of what actual science has actually been done, and which things (e.g. in psychology) are maybe kind of close or promising but need adjustment to hug the query properly. That includes things like looking at what Leverage has done.
There's a ton to draw from in the neo-hippies' stuff for sure, but nearly all of it is what I'd tag as "pre-scientific subjective engineering" (analogous to bloodletting in medicine: sometimes actually effective, but memetically optimized to appear effective and relevant within a pre-scientific context). I view them as a source of inspiration. Kind of like how alchemy had a ton of really interesting results that were what was to be investigated to create chemistry, but the frameworks of alchemy were almost perfectly useless for developing a science of chemical reactions.
To pick on a couple of the institutions you mentioned:
I see AoA as being mostly orthogonal to what CFAR is doing. I imagine AoA can help you become unstuck in relationships, be more honest in your dealings with others, be more comfortable with uncomfortable emotions, etc. I think this could help you become more rational in certain domains where self-knowledge/introspection is the blocker; like, if you're doing a pro/con list on whether you should break up with your partner then you're probably ignoring the bigger emotional fact that given you are making a pro/con list you should probably just break up (to give a clichéd example).
But I'm not sure it really teaches any techniques that force you to engage your system 2 and actually grapple with a thorny problem in a way that Murphyjitsu or doing a pre-mortem or sitting with a timer and writing for 5 minutes actually do.
Jhourney seems much more focused on accessing particular psychological states that feel good - I'm not sure I see it as improving your rationality or decision-making?
I also think these things can have harms in some other ways that probably don't apply to CFAR techniques - e.g. in the AoA Connection Course, you're asked to think of something that would be really "triggering" to your partner (scare quotes because I think they use the word as a synonym for 'upset' rather than as 'triggering some underlying trauma/mental health issue') and then deliver it to them. When I did this my course partner cried and I felt very bad about it and I don't really think there was any benefit to it - and made me update away from the idea that Joe Hudson is some kind of emotional savant.
Likewise Jhourney explicitly forbids certain people from taking their courses, namely people who have ever had a psychotic episode or are prescribed psychiatric medication (except SSRIs, from what I recall). So I would suspect that these other more 'vibes-based' courses/approaches have quite different risk profiles from going to a CFAR retreat (although perhaps that also has its own risks!).
I think your claim about AoA being orthogonal relies on rejecting my claim above that "emotional blocks are the main reason why people make bad decisions".
For example, AoA doesn't teach stuff like Murphyjitsu or pre-mortems, but on my model both of those are primarily useful as ways of sidestepping how painful and aversive the prospect of failure usually is. So I'd count AoA as trying to do the same thing, but more robustly, if it tried to address the underlying pain of / aversion to failure (which it sounds like it does).
Re Jhourney, I know they advertise themselves as being about accessing bliss states, but when I attended a month or two ago they actually seemed to place just as much (or more) emphasis on emotional processing more generally.
Congratulations to Anna and the team for cohering around a vision and set of experiments. I donated to the new CFAR; I hope you continue posting about what you learn through the upcoming workshops.
One {note? suggestion? “real spirit” discussion point?} - I feel like the framing of aCFAR was missing something important about the state of rationality today. Namely from this, the year 2026 onward, being more rational is unlikely to be a “human technique” only affair. It will look more like cyborgs and centaurs - humans using AI tools and agents in different configurations and ways to make better decisions.
I won't belabor how good the AIs have gotten, and instead will just note that they are effective aids for rationalist techniques:
I appreciate that it's a Center for Applied Rationality, and maybe this particular center doesn't need to think about the cyborg angle and can just focus on developing better models of “who-ness”. Maybe a different center should!
But it seems valuable to consider, to the extent you want to push forward the frontier of rationality. I suspect there's some connection between the moments when AI meaningfully aids my real thinking, the moments when I'm doing slop-ful fake thinking where the AI is aiding my delusions, and the concept you're defining as "who-ness." Who-ness seems adjacent to taste, which might matter a lot for steering AI fleets towards goodness and meaningful concepts. And it’s probably the case that the general rationality techniques you’re imparting and working on with attendees could be more effective with AI assistance.
I would like to thank you Anna for this write-up! I love the thoughtfulness and ideas on institutional design, relationship to donors and EA, and others. Much of this post was quite viscerally relaxing for me to read, in a "finally it feels possible to jointly understand X and Y and Z" kind of way, which I'm quite glad for.
If anyone is considering donating and wants to talk a little first, do please book my calendly. 15 minutes, easy! :) (Also happy to talk with people about things other than donating, though easiest for me if we do that in January, all else equal; can still schedule now)
I'm glad to hear some form of CFAR experimentation/iteration is in the offing. (For those who don't know my handle, I'm Leah, who was a CFAR instructor/developer for a year).
To try to summarize back, it sounds like this version of CFAR has a big focus on an exploratory mindset vs an instrumentalizing mindset (e.g. insert thoughtful, driven people into workshop, output x-risk-pilled possible researchers). While I was at CFAR, it was clear that some folks in the project were working there as their best applied effort to improve X-risk (this wasn't my reason for working there, so it did a lot less to shape my view of curriculum/participant recruitment and outcomes).
When I explained my job to people, I said I taught defensive driving for your brain (how to make the best use of your tools for reasoning + how to steer away from the circumstances where you'd be least able to reason well).
If I expanded more, and talked a little more about System 1 and System 2, I'd say that we taught about how to more explicitly understand/actively apply System 2 reasoning, and how to treat System 1 as a rich source of data that System 2 might be prone to skip over too quickly, because it isn't well formatted as a source of input System 2 parses well.
I'm curious how well this maps on to aCFAR's current understanding of itself. It seems like there's a bigger emphasis on a receptive posture to the world.
I'd say that part of how I saw success for participants was
-more comfort reasoning under uncertainty
-more of a sense of where you want to lean into risk (strong EV from trying out a low cost 1 in 10 chance more often vs "that didn't work") that is built up by prediction market games etc
-more ability to sit curiously with conflicting impulses/desires vs reasoning only verbally and then sticking to the "right" choice no matter how much you don't like it
I think of all of this as gaining an internal freedom to think and to act, and thus to take on more ambitious projects because you have a sense you'll be able to navigate turbulence.
Great to hear from you; thanks for popping in! I like the thing you're describing, re: your work at CFAR_2013, and that angle on rationality/CFAR more generally. I'd say it describes our current stuff... 5 out of 10 maybe? I agree there's more emphasis now on receptive posture to the world (and also on tuning into taste, and the data taste brings up). I'm not sure what else to say at this moment, but I'm really glad you're bringing your perspective in, and wanted to say that and say hi.
Cool to see some more energy here - I have benefitted greatly from being around the CFAR nexus.
My kibitz: wait, who is driving? Who is leading? Who is holding the torch? Who is parenting this baby org?
You did a good job of pointing to people involved, but I'd like a better sense of how the people involved are themselves relating to the new org. My guess is that you (Anna) are intentionally holding a decent amount of the aCFAR torch by authoring this post, but also intentionally not wearing fancy titles or making a very formal org structure to allow for some of the magic of surrendered leadership (and to avoid some of the pitfalls of more structured leadership).
Still, I tend to expect more fruit borne from orgs with 1-3 'torch-bearers' or 'parents' or 'leaders'. I'd be more excited about aCFAR if there were some explicit parents, or if there were none then making that more explicit. At the very least getting more clarity would help me figure out how I want to relate to the org.
(disclaimer: I skimmed but didn't read closely your posts, perhaps you already touched on this. Hopefully this is helpful to surface in this context - feel free to ignore this comment if better discussed in another context, I can figure this out on my own time.)
(ETA: I changed most of these words about 25 minutes after posting, hope that isn't annoying.)
I'm executive director; the buck stops with me. aCFAR is doing things my inside view says are good, and is not doing anything that seems dumb or useless or wasted to me; I am promising to keep my eyes on the ball.
Caveats/details:
Does this answer your question?
We have been getting lots of good ideas from lots of other staff, and these ideas have been riffing off of each other and growing a lot of the cool stuff we have, but this isn't via me being shy about my own views. (In the past I was uneasy with some aspect of leadership, but I'm less so now.) A majority of our good stuff is coming from others, via a process that makes sense to me, and that does not at all remind me of "group work" in high school or of a "nobody home"/"dodging the social awkwardness" feeling. (I hated "group work" in school, because I didn't know how to navigate about contexts where I thought I knew something and yet didn't want to offend.)
A more articulate attempt:
My current theory of "how do good things ever happen in the world?" is "via someone doing things they care about, that make sense to them." (This is a ~necessary but not sufficient condition, TBC).
I aim to set up CFAR so that:
a) All of it makes sense to me, and is something I care about. Since I'm executive director. (It's fine if the "making sense to me" and "caring about it" is on the process level, e.g. if I care about and see sense in an instructor getting to test their unit, rather than about the specific details they're testing).
b) Each person within CFAR is basically only doing things that they themselves care about, and that make sense to them.
I used to be all wiggly-feeling about leadership because I didn't realize we could do (a) and (b) at the same time, but we totally can. I mistakenly thought that if I was leading, it meant I was asking other people to be in a weird relationship with their own perceptions of what made sense. Also, I somehow know now how to share my perceptions and my causes out loud, in a more simple and straightforward way that lets everyone else see them at once and talk about them and remember them. (Not perfect here, but better than I was.) And that makes it easier to find ways to do (a) and (b) at once.
Thank you, that answers my question - and it makes me more excited about the project! I'm glad to hear it's been unfolding nicely thus far, and I'm feeling pride/respect for you taking on the more formal mantle.
(I didn't see the early version)
Would you be up for saying a little more about "I have benefitted greatly from being around the CFAR nexus", and/or about what you're hoping for?
Hi, Anna. The post was very interesting, and I'd be happy to do some 'kibitzing' (had to check the word's meaning, actually), even if I give low credence that anything I'll say might be of use.
I am a relatively late arriver to the world of Rationalism (from about 3 years ago or so), having gone through some of the usual suspects (Julia Galef's Scout Mindset and her podcasts; Chivers' book on Rationalism; HPMOR and some online posts from Scott Alexander; some CFAR materials online, and some youtube videos; not much Less Wrong or The Sequences though, until now). What attracted me to it was my quasi-religious attachment for truth-seeking and for avoiding self-deception, and my desire to become more rational in my thinking and decision-making. I am still reading and exploring. You might remember that I've been one of the assistants to the CFAR test sessions, and would be interested in the future in going a workshop if it doesn't entail having to go across the Atlantic (I live in rural, NW Spain).
Through my eyes, CFAR seems like a group of people who are engaged in the sort of 'right thinking' I aspire to (LW also; perhaps the distinction from the outside is that CFAR seems to be a more educational-oriented project in practice).
A lot of what I see in the test sessions I find confusing, but I don't think it's from any fault of yours. They aren't workshops, and I feel a lot of uncanny valley-ness in that I recognize most of the ideas and terms but haven't really interiorized their meaning. Also, the only correlate your practice brings to mind is something like talk-therapy which, while pretty common in the States, is really unusual over here (the only, and very few people I know who have engaged in something like it do it with psychologists and for medical reasons). This isn’t a criticism so much as a cultural translation problem: from where I sit, it’s not always clear how to distinguish applied rationality coaching from therapeutic modes of engagement, and that makes it harder to know what norms to bring.
I think it would help me a lot to actually go in detail over some of the material you have online (particularly, the CFAR Handbook). I think I just mostly lack the grammar of how these things are supposed to be done.
'Hope for something out loud': I'd hope for a chance in the not-too-remote future to take once of your workshops this side of the ocean. 'Try to speak to why you care rather than rounding to the nearest conceptual category': as I said, I care really personally about truth-finding, truth-seeking and being part of an expanding circle of people who share this frame of mind, which makes me see your efforts, whatever they end up producing, as merit-worthy. And that’s why I care enough to watch closely how this iteration of CFAR actually plays out.
We are indeed so considering, downstream of your comment! Thanks. I think we'll submit our project there in an hour or two. Do you have thoughts re: what we should put as our fundraising goal? I assume it should be a target for how much donation comes to us via Manifund, rather than overall?
I haven't participated in CFAR workshops, I'm working from your written posts and my experience in other communities, so take these comments with the appropriate grains of salt.
I read the shift you describe as being from getting people on board with a largely preset agenda towards a more collaborative frame (and that the results have been promising so far). I wonder if the underlying issue here may be format-intent alignment. If CFAR has specific rationality techniques it believes are valuable and wants to teach, there's an inherent directiveness to that goal. Workshop/discussion formats, however, signal exploration and reciprocity. When the intent is directional but the format signals collaboration, participants can experience a disorienting mismatch.
I'm curious whether CFAR has considered hybrid formats that are explicit about which parts are directional and which are truly collaborative, or if you are leaning entirely on shifting intentions to fit the existing format—even if that means letting conversations drift away from anything adjacent to rationality or x-risk. I don't think there's an inherently right answer here, but understanding where you are on the spectrum and being transparent about this choice could help participants set their expectations, or self-filter regarding whether CFAR is a good fit for them at all.
Seeing this as a spectrum rather than a binary seems important because it prevents participants from running into "invisible" restrictions. Mutual learning and meeting people where they are at are great, but it doesn't seem realistic to try to be all things to all people. It's therefore important to be ready to say "this is what we offer, these are our constraints, respond as you will" even if where you think it is best to draw those lines has a wide range of valid answers. Sometimes a person is incompatible with an organization and that doesn't have to be anyone's fault.
A second, arguably more complex and charged dynamic I am wondering how you intend to navigate is responsibility for managing participant capacity. Is CFAR explicitly working to build participants' ability to recognize misalignment, maintain their sense of agency while in seemingly asymmetric power dynamics, and advocate for their needs? Or is the plan for CFAR to create sufficiently careful environments that participants won't need those skills? Or to screen applicants for having the capacity for self-advocacy already? Your post makes it sound like you are primarily taking the second approach, which is fine, but I am wondering how you are assessing the trade-offs involved.
For transparency, I personally believe in building capacity for self-advocacy and agency, with community care supporting that development and acting as a fallback when capacity isn't present, but I recognize that different approaches work for different purposes/people. What I am pushing for here is an explicit articulation of the balance you are striking so that participants can make a fully informed choice as to whether they wish to join.
In any case, I appreciate the thoughtfulness of your post and willingness to share CFAR's evolution publicly.
We seven years ago said publicly that Valentine “would not be any staff or volunteer roles going forward, but remain[ed] a welcome member of the alumni community”, and so it seems well to be similarly public about my revised intent
The link seems to be wrong; I couldn't find anything on that page about Valentine.
The quoted text is (mostly) on that page, under the heading ESPR 2017, but it refers to an unnamed "staff member". (And the quote isn't quite verbatim, even aside from the word 'remain', so it won't work in full as a ctrl+F search term.)
my fellow instructor Jack Carroll said they liked the participants for the first time instead of feeling at war
I find it sad to read and did not expect that Jack always felt at war and did not like previous participants.
To be fair, I was exaggerating when I used the phrase at war. I basically like people. I've always, in many ways, liked the participants. And, in many ways, I have not liked them. Not due to anything about them, but a sort of.. invisible structure I felt? Like a pressure that would ask me to hold myself in a particular way. And so in the past, I was in the workshop as an avatar of the entity CFAR, and an avatar of Holding the Cool Skills, and that invited in a lot of masking which blocked genuine connection. I used to spend as much time hiding outside of classes as possible. But at this most recent workshop there was an immediately apparent and distinct quality that I felt of.. I'm not trying to sell anything. I don't need to be in any way. In the absence of pressure, then it was so easy to be around participants and like them — and some of this is personal growthy stuff, my social anxiety has gone down etc etc, but I believe there's a bigger change in how CFAR is inviting staff and inviting participants that made for an even more wholesome and friendly environment.
FWIW I think it's pretty common for organizers of an event to feel exhausted and beleaguered by the attendees. It's pretty tiring and they make tons of requests and sometimes they act kind of counter to the spirit of the event.
Maybe, but that does not seem to be what @jackjack meant, though I am not sure whether I understand what he means, and why it's different now.
This post is the main part of a sequence of year-end efforts to invite real conversation about CFAR, published to coincide with our fundraiser.
Introduction / What’s up with this post
My main aim with this post is to have a real conversation about aCFAR[1] that helps us be situated within a community that (after this conversation) knows us. My idea for how to do this is to show you guys a bunch of pieces of how we’re approaching things, in enough detail to let you kibitz.[2]
My secondary aim, which I also care about, is to see if some of you wish to donate, once you understand who we are and what we’re doing. (Some of you may wish to skip to the donation section.)
Since this post is aimed partly at letting you kibitz on our process, it’s long.[3] Compared to most fundraiser posts, it’s also a bit unusually structured. Please feel free to skip around, and to participate in the comment thread after reading only whatever (maybe tiny) pieces interest you.
I’d like CFAR to live in a community
I’d like CFAR to live in a community where:
In the past, CFAR didn't know how to live in a community in this way (partly because I was often in charge, and I didn’t know how to do it). But I think CFAR and I now have the ability to do this.
As an example of the gap: I used to be somehow trying to claim that we were running our organization in the best, most EA-efficient, or most rational way. As a result, whenever someone argued in public that some revised action would be better, I thought I had to either:
But now, it’s different. We are visibly a particular organization, believed in by particular people, with details. The premises we believe in together (aka our operational premises for what we CFAR staff are building) are separated out from our epistemics, and from claims about what’s objectively best.
Anyhow: requesting community membership of this sort for CFAR, and setting you guys up to have a free and full conversation about CFAR, is the main business of this post, and is the main thing I’m trying to ask of you, Dear Reader, if you are interested and able.
Kibitzing requests
Some kinds of kibitzing I’d particularly appreciate:
Introductions: Me, aCFAR… and you, Reader?
I’ll start the introductions.
I’ll also try introducing aCFAR as though it’s a particular person with a history:
If you’re up for introducing yourself (which I’d appreciate!) there are two good ways:
On to the post proper!
Workshops
Workshops have always been the heart of our work at aCFAR. We spend most of our staff time tinkering toward making the workshop good, staring at folks at the workshop to see if it is good, iterating, etc. It’s where our take on rationality comes to life, changes us, is changed by its encounters with some of you guys, and so on.
So – if you want to kibitz on our current generators – it may be where you can best see them.
For those just meeting us: A CFAR workshop is a 4.5 day retreat with about 25 varied guests, 12-ish staff and volunteers, and a bunch of hard work, rationality, and conversation. The workshop typically involves a bunch of classes on rationality techniques and lots of time to apply those techniques and work on solving actual problems in real life. We currently have our next workshop scheduled for January 21-25 in Austin, TX.
Workshops: favorite bits
Among my favorite positive indicators from our workshops:
1. People made friends at the workshops and in the alumni network.
Many workshop guests across our history have told me a CFAR workshop was the first time they’d managed to make friends in the decade-or-more since they finished college.
This wasn’t an accidental side-effect of the workshops; we tuned the workshops toward: (a) creating contexts where people could update deeply (which helps with making real friends) and (b) arranging small and interactive classes with pair work, providing a “names and faces” Anki deck, hosting lightning talks, etc. to make it easy to make new friends at the workshop.
This wasn’t a side-goal for us, separate from the main aim of “rationality training”; IMO there’s a deep connection between [conversations and friendships, of the sort that can make a person bigger, and can change them] and the actual gold near “rationality,” such that each of (true friendships, rationality) can activate the other.
2. People had conversations at the workshops that updated the real generators of their actions.
Many conversations in the default world involve people explaining why a reasonable person might believe or do as they are doing, without sharing (or often knowing) the causes of their choices. But at CFAR, the real causes of actions often were (and are) properly in the conversation.
Relatedly, people at workshops would become at least briefly able to consider changing things they’d taken for granted, such as career paths, ways of relating to other people, etc., and they’d do it in a context full of curiosity, where there was room for many different thoughts.
3. The workshop was visibly “alive” in that it felt organic, filled with zany details, etc.
If this CFAR is going well, we should have spare energy and perceptiveness and caring with which to make many side-details awesome. We did this well in the past; we seem to be doing it even better now.
For example, during Questing at our November workshop, we had CFAR instructors run short “interludes” during which people can breathe and reflect a moment in between 10-minute hero-and-sidekick problem-solving blocks. However, due to a minor scheduling mishap, CFAR instructor Preston ended up committed to be in two places at once. Preston solved his problem by setting up an “oracle” to run his section of inner simulator-inspired Questing interludes.
For another example, chef Jirasek created waves of life emanating from the kitchen in the form of music, food art, and sort of ostentatious interactions with the locals (e.g. replacing his whole wardrobe with stuff from some local thrift stores).
4. Truth-seeking, curiosity-eliciting, rationality-friendly context
The context at our workshops is friendly both to hearing peoples’ perspectives deeply and to being able to point out possibly-contrary evidence.
Workshops: iffy bits, and their current state
Although there’s much I love about our old workshops, I would not be able to run them now, although I could probably cheer for someone else doing it; there’s too much I was eventually unable to stomach for myself. In particular:
Power over / doing something “to” people (current status: looks solved)
I currently aim not to take pains to impact someone unless I can take equal pains to hear them (in the sense of letting them change me, in deep and unpredicted ways). This is part of a general precept that conscious processes (such as CFAR guests) should not be subservient to processes that can’t see them (such as a rock with “follow policy X” written on it, or a CFAR instructor who hasn’t much attention to spare for the guest’s observations).
My main complaint about our past workshops is that I, and much of ‘we’, did not always hit this standard (although we tried some, and some of our staff did hit it). It’s part of my current take on how to do epistemics in groups.
More details about this complaint of mine, for those interested:
1. Excessively narrow backchaining / insufficient interest in both the world, and our workshop guests
I was scared about AI risk, all the time. I was in an emergency. And while I did try at the workshops to drop all that for a bit and take an interest in the people in front of me, I was also at the workshops to “make progress” on the AI risk stuff.
So, my notion of which participants were the coolest (most worth paying attention to, inviting back, etc) was mostly:
(As opposed to, say, who had neat make-stuff skills or research patterns we didn’t have, that might broaden our horizons; I was too tired to really see or care about such.)
2. Nudging the CFAR alumni culture toward #1, so our community also became narrower
I, and other CFAR staff, weren’t the only ones who evaluated coolness a bit too narrowly, by my present taste. I think I and others in positions of community leadership often helped set this up in various ways.
(As a contrast point, the 2007-2011 OvercomingBias commenter and meetup community had broad and deep engagement without being a “school of thought” in the way the CFAR and LW rationalists later were, IMO.)
3. Trying to do something “to” our guests; priming our guests to want something done to them.
Many of our guests signed up for the workshop so that we could help make them more rational so that they could be better EAs (for example). And we wanted them there for much the same reason (sometimes; some of us).
4. Casting ourselves as having more epistemic authority or charisma than I currently think warranted.
Deeply related to #1, 2, and 3 above.
I’m relieved that our Nov 2025 workshop (and our prior, tiny pilot at Arbor Summer Camp) did not have these problems AFAICT. Things I saw in November, that I believe I’d see differently if we did still have these problems:
This is enormously relieving to me; uncertainty about whether we could change this thing was my main reason for hesitating to run CFAR workshops. We will of course still be keeping our eyes out.
More workshop iffy bits
While the “power over” thing was the iffy bit that bugged me the most, there are also other things we want or need to change about the workshop. You can see our whole workshop-related bugs-and-puzzles-and-todos list here.
More about the new workshop
If you’ve been to a CFAR workshop in the ~2015-2020 era, you should expect that current ones:
Further detail, if you want it, at More details on CFAR’s new workshops.
Larger contexts surrounding our workshops
In this section, I’d like to talk about the larger contexts (in people, or in time) that our workshops depend on and contribute to, as well as some solved and unsolved pieces about those larger contexts.
aCFAR’s instructors and curriculum developers
Our major change, here, is that all instructors and curriculum developers are now very part-time. (In 2012-2020, most workshop instruction and curriculum development work was done by full-time staff.)
There are two big reasons I made this change.
A pleasant bonus is that we get more mileage per donor dollar: a few hours/week of trying our units on volunteers and on each other is enough to keep CFAR in our shower thoughts as we go through the week (for me, and for many other instructors AFAICT), and the rest of our normal life seems then to give us useful insights too. (And we’re paid hourly, so a "lighter" schedule that still gets curriculum development flowing is a good deal for donors!)
aCFAR’s alumni community
Our workshop development process is stronger with a healthy alumni community in several ways:
Our alumni community was extremely fun and generative in CFAR’s early years, but gradually became less invested and lower trust over time, partly as a natural side-effect of passing years, and partly because we weren’t doing community all that well. We still have an alumni mailing list and it hosts some interesting discussions, but things there feel less active and exciting than they once were.
We like our alumni and think they’re cool! We’d like to figure out how to freshly kindle some of the energy that made the old CFAR alumni community as cool a place as it was.
My guess (not a promise) is that we should start a new alumni community with these features:
Lineage-crediting and gatekeeping
It is vital to accurately, publicly track where good things come from (lineage-crediting). At the same time, it is necessary not to let people into our events or alumni networks who we can’t deal with having there. This combination can be awkward.
As an example of this awkwardness: Michael Vassar taught me and many people a bunch about rationality when I joined the rationalist and AI safety scene in 2008, and he was also quite involved in me changing my mind about the stuff I mentioned changing my mind about in 2020. I can see traces of his ideas all over this post. My thoughts in this post, and the ideas in the newer parts of CFAR, were also greatly influenced by my good friends Anonymous and Anonymous.
And yet, for varied reasons, I wouldn’t feel good about having any of those three visit an intro CFAR workshop (although I might well invite Michael Vassar to an alumni reunion or similar event, where my tolerances are a bit broader; and I’d gladly have all three to a retreat run by a more bespoke CFAR spin-off called LARC/Bramble). I think this is not unusual bad luck; my best guess is many of those who “woke up” as kids in strange surroundings and who forged their own paths to being unusually conscious and agentic, dodged some of the “be rule-abiding” training that makes most middle class people easy for other middle class people to predict and be safe around. And the CFAR alumni network is a large, semi-institutional context designed to work okay for folks who are within the normal range on rule-abiding and who are used to getting to assume others are too, for good reason. (To be clear, I also learned a pile of rationality from many others, most notably Eliezer, who are reliably rule-abiding.)
This sort of “awkward” isn’t only costly because of wanting not to alienate my friends. It’s also costly because it’s confusing (to me, to them, and to workshop guests and onlookers). When rationality content is presented within a context that couldn’t have made that content and that doesn’t help tend the sources of that content, it’s harder to set up good feedback loops. (Cf. the Caring that Tends its own Sources).
But, here I am, anyhow, having decided that this is the best world I can manage, and trying to describe something of its workings in public.
My plan, roughly, is the obvious one:
Michael “Valentine” Smith
While we are on the topic of both gatekeeping and lineage-tracking: we are considering bringing CFAR co-founder Michael “Valentine” Smith back onto our workshop staff.
I’d like to note this publicly now, because:
If it matters, I and various others have worked closely with Valentine at LARC/Bramble (CFAR’s more bespoke spinoff organization) for the last two years, and I have found it comfortable, wholesome, and generative.[8]
The broader rationality community
The broader rationality community makes our work at aCFAR feasible (e.g. via donations, via sending us participants who are already rationality fans, via giving us good rationality stuff to draw on, and via good critiques). We are grateful to you guys. It’s important to me that we give back to you, somehow, in the long run. My main current theory as to how to give back is that we should write substantive blog posts as our theories-of-rationality congeal, and should make our process open so if we fail this time, it’ll be easier for interested parties to see what exactly went wrong (no opaque fizzles).
Flows of money, and what financial viability looks like within our new ethos
We do not yet have a demonstrated-to-work plan under which aCFAR (in our new incarnation) can be financially sustainable.
In 2012-2020, a large majority of our donations came from AI risk donors, who donated because CFAR recruited for MIRI (or to a lesser extent other AI safety efforts) or because they otherwise believed we would help with AI risk.
Also, in 2012-2020, a significant chunk of our workshop revenue came from EAs (both AI risk people and EAs more broadly) who had heard that CFAR workshops would somehow make them better EAs, and perhaps also that CFAR itself was an EA organization worth supporting. And so they balked less at the (then) $3.9k price tag because it was parsed as an EA expense.
Double also, in 2012-2020, we workshop instructors broadly tried to position ourselves as people who know things and can give that knowledge to you (and so are worth paying for those things).
My current attempt at CFAR branding lets go of all three of these angles on “you should give us money,” in favor of an ethos more like: “we (including you, dear workshop guest) are a community of people who love to geek out (in a hands-on way) about a common set of questions, such as:
Under this model, CFAR instructors differ from workshop guests in that we spent a bunch of time testing and refining particular classes (which we try to make into good springboards for doing hands-on geeking out of this sort in common, and so for jumpstarting guests’ ability to have rich conversations with each other, and to do rich, grounded noticing together, and to point out traction-creating things that are visibly true once pointed-to). But we try not to differ in perceived/requested epistemic status, or in “you should believe us”-flavored social cues.
Also, under the new model, our requests aren’t backed by a claimed long-run EA payoff: we are not saying “please consider sacrificing parts of your well-being to work at CFAR, or to attend CFAR or implement our taught habits, because it’ll help with AI risk somehow.” Instead we are saying “please come nearby if it interests you. And if you like what happens next, and what changes it seems to give you in the observable near- and medium-term, then maybe keep trying things with us for as long as this seems actually healthy / rewarding / to give good fruits to you and visible others in a simple, cards-on-the-table way.”
I expect our new model is more wholesome – I expect it’ll bring healthier feedback loops to our curriculum and culture, will form a healthier town square that is more fruitful and has fewer stuck beliefs and forcefully propagated illusions, and will be an easier context in which to keep us staff wanting to share most info in public, including evidence we’re wrong. But I don’t know if it’ll bring in enough revenue to keep us viable or not. (And we do still need money to be viable, because being a custodian of such a community requires staff time and money for food/lodging/staff flights/etc.)
If we can’t make a financial go of things under our new ethos, my plan is not to revert to our past ethos, it’s to fold – though my guess is we’ll make it.[9]
How our ethos fits together
In this section, you’ll find pieces of what motivates us and principles we intend to follow.
Is aCFAR aimed at getting AI not to kill everyone? If not, why are you (Anna) working on it?
We are not backchained from “help get the world into state X which’ll be better for AI,” nor from “help recruit people to AI safety work,” “help persuade people to take better AI policy actions,” or anything like that.
My (Anna’s) motivations do and don’t relate to AI safety; it’s complicated; I’ll publish a separate post going into detail here in about a day.
Principles
This is an attempt to make visible the principles that I, and to some extent CFAR, are trying to act on in our CFAR work. I, and we, might change our mind about these – these aren’t a promise – but I plan to review these every three months and to note publicly if I change my mind about any (and to note publicly if CFAR changes leadership to someone who may run on different principles).
We’ll start with some short-to-explain ones, then head into some long explanations that really should be their own blog posts.
Truth is crucial
This principle is one of the “things that go without saying” around LessWrong most of the time (and is shared with past-CFAR), but it’s precious.
Honor who-ness
Remember each person is a miracle, is way larger than our map of them, and is sustained by knowledge and patterns of their own making. Honor this. Allow ourselves to be changed deeply by the knowledge, patterns, character, etc. of anyone who we deeply change.
Stay able to pivot or shut down, without leaving anybody in the lurch
It’s easier to stand by principles if there’s a known and not-too-painful-or-commitment-breaking path for quitting within a few months (should we prove unable to stick by these principles while remaining solvent, say).
Serious conversation, done in hearty faith
This section is written by my colleague John Salvatier.
Serious conversations deal with the real issues at play and go beyond literary genre patterns. And serious conversations in hearty faith apply enough real human trying to get to real discovery about the topic.
Serious discussion of problems we really care about where the participants are fully engaged are kind of a miracle. For example, if you’re wondering whether to quit your job, a serious and hearty conversation about the question and about what matters to you in life can have a profound life effect.
At this CFAR, we are trying to have hearty faith with each other and with others to create the possibility of serious conversations. (And we are trying to do this without forcing, via repeatedly asking ourselves something like: “does it feel good to share my real cruxes right now, and to hear where [person] is coming from? If not, what sensible reasons might I have for not (bearing in mind that there’s lots of useful stuff in me that conscious-me didn’t build)?” We aren’t trying to impose hearty faith; we’re taking its presence as a thermometer of whether life is going well right here.)
Serious conversations are like science experiments. Their success is not measured on reaching a particular outcome, but on their revealing substantial things about the world that bring us into closer contact with the world.
The classic Eliezer/Robin AI Foom Debate is a good example of something that might look like a serious conversation but somehow isn’t a “conversation” in quite the sense we mean. A conversation would spend a bunch of time doing asymmetric things where one person is mainly trying to understand the other (for example passing their ITT). Instead, Eliezer and Robin each use each other as a foil to better articulate their own view. This might be serious research, or good exposition to an audience, but it isn’t the thing we have in mind.
Hearty faith is necessary for successful serious conversations when our maps (or theirs) have messy relevance to the world and our goals. Which they will when the topic is a life frontier or a world frontier.
Hearty faith is different than just good faith.
Bad faith is lying, fraud. An abandoning of our integrity.
Lousy faith however is when our intentions are like a thin stew instead of a hearty, many-flavored, full-bodied one. In “lousy faith” we are putting in effort to keep integrity on some dimensions, but not very many.
Hearty faith, by contrast, is when we act with attention to many sorts of integrity all at once (the more, the heartier, like a hearty stew).
Hearty faith is necessary for serious conversations with messy world maps to be successful because every such conversation has many relevant-but-illegible layers that are otherwise obscured and hearty faith allows legibilizing them. It allows the relevant-but-illegible conversation layers into the conversation on good terms.
The caring that tends its own sources
This is a phrase I made up, inspired by Eliezer’s The Lens that Sees its Own Flaws (which is one of my very favorite Eliezer posts, and conveys an idea that’s on my shortlist for “most inspiring insights ever”) and and also by conversations with my friends Evan McMullen and Anonymous.
I hope to eventually write a blog post about this principle that makes sense. But this is not that blog post, it is a placeholder.
So: we find ourselves alive, awake, caring. How did I, or you, reader, get to be like this? It’s a bit of a miracle. We can tell decent causal stories (mine involves my parents, their parents, the United States, a brief opening in Hungary’s border during a war, my mom’s careful crafting of endless ‘math games’ for me, my dad’s absorbing a useful secularism from the Soviet Union that he rightly hated… going further back we have the European Enlightenment, eons of biological evolution, and more). We can tell decent causal stories, and it’s worth bothering to tell them, and bothering to try to get it right; and at the end of the day “a miracle” is still a decent term for it – the processes that let us be here are something large, and worth marveling at, and contain deep generative “magic” that we don’t yet know how to build.
How to relate to this?
Concretely:
No large costs without a feedback loop grounded in earned knowledge and caring
This principle is an attempt to articulate the main thing I changed my mind about in 2020.
It now seems to me that when you’re running an organization, such as aCFAR or the neighborhood bakery, you’ll benefit if you:
Take an interest in what produces and sustains these resources. Be aware of the rough extent to which you do or don’t have reliable maps of what’s involved in producing and sustaining these sources, so you can maintain the needed amount of [respect / Chesterton’s fence / actively watching out for needed conditions you shouldn’t disrupt], without being unduly cautious about everything.
For example, I understand how to turn hot water and peppermint teabags into peppermint tea. (Thus, I can change up my water heating method, its temperature, etc without being surprised by the results.)
On the other hand, my friend sometimes likes to walk his dog with me. I’m pretty sure there’s detail to where he will/won’t take his dog, when he does/doesn’t feel like doing it, etc., and I’m pretty sure that detail helps maintain cool functionality, but I also know I don’t know how it all works. Thus, I know that if I try making many of these decisions for my friend, without consulting him, I might mess up some resource he’s used to counting on.
Take an interest in the specific “bridging structures” that let particular resources coexist.
For example, a coaster is a good “bridging structure” to keep my hot teacup from damaging my wooden table.
For a more complex structure, a bakery’s proprietor might be careful to keep their sidewalk shoveled, to greet neighboring business owners, etc. as part of a plan to allow the bakery and the downtown it’s in to avoid harming each other. This kind of bridging structure will need to regularly take in new info, since one probably can’t have an adequate static map of downtown as a whole.
Let each resource-flow and each bridging structure have a keeper who maintains both an inside view about what’s necessary for sustaining the resource flow and an inside view about how much “magic” isn’t yet in their map.
That keeper must be responsible for deploying these resources only in ways that make inside-view sense to them (e.g., if there’s a small experiment, the keeper should have felt hope in doing small experiments; if there’s a large deployment, the keeper should have felt conviction that large deployments of this sort bring fruit)
That keeper must also have enough eyes on the results of that deployment that they can update sensibly.
I’ll spell out what this means in the case of CFAR, and then explain why I care.
What this means in the case of aCFAR:
This CFAR makes use of three main resource flows:
We want all these resources used in ways where their keepers have grounded reason to think it’ll help with something they care about (and have feedback loops for checking).
Concretely, I’m aiming for:
Staff and volunteers have better lives (or not-worse lives) via our involvement with CFAR, including in the short- and medium-run
In CFAR of 2012-2020, many of us sacrificed for CFAR – we e.g. worked 60+ hrs/week, had distorted social patterns with folks in the rationality community, and otherwise paid (and sometimes caused) large costs. I’d like to arrange our culture so that people don’t do that this time around. I want us to each be simply, groundedly in favor of what we’re doing, without trusting in long-term unseen effects on the post-AGI future or anything else.
(Here and elsewhere, it’s fine if staff and volunteers sometimes try things that hurt us. The principle isn’t “no costs” or “no one made worse-off ever.” It’s rather “no key resource flows, ones that CFAR is reinforced by and grows around, that make people worse-off.” One-off “ouches” are part of how we locate what works, and are fine as long as we update away from them instead of learning to depend on them.)
Participants try aCFAR’s suggested habits based on their own inside views (not our charisma or claimed knowledge)
Some participants have historically shown up to the workshop expecting to be told what to do by people who know the answer. But I want us to resist this pressure, and to create a culture of “practice trusting your own judgment, and making many small experiments while seeing yourself as the author and experiment-iterator for your life and habits.”
Donors
I go into much more detail on this one in who I hope does and doesn’t consider donating.
Why this principle
I’m afraid that otherwise we’ll do a bunch of hard work, at large costs, that nets out to “harmful, on average, after considering opportunity costs.” I’m also afraid that all that work won’t even teach us much because, for most of it, there was no conscious human who individually thought it a good idea. (This is coming out of my 2012-2020 experiences.)
To spell out my thinking:
First: people often learn more by making their own mistakes than by “making other peoples’ mistakes.”
This is easiest to see if we consider a concrete context such as chess. If I play chess from my own inside view, I will repeatedly make moves that look like good ideas to me – and then my opponent will often show me how exactly my inside view was wrong by exploiting my errors. If I instead play chess by repeatedly trying moves my friend thinks are good, I’m likely to learn less, because my friend’s moves aren’t rooted in a detailed inside-view lodged in my head.
There are exceptions – maybe my friend has a Cool Chess Trick that I can understand once I try it, and that wouldn’t have occurred to me on my own – but these exceptions work when they’re somehow supporting an existing, intact flow of my own autonomous choice.
Second: I don’t want to build habits or culture (in our alumni) that’ll be easy for cult leaders or others to exploit.
If workshop guests practice deferring to us about what weird things to do with their minds – especially if they do so for extended periods, based on wispy claims about long-term payoffs, e.g. “this’ll help with AI risk somehow” – this risks setting some up to later try deferring to people running more obviously unhealthy cults. I speak from experience.
I also hope a culture of “remember the buck stops with you; check whether it is producing fruits you directly feel good about” may help with the rationalist community’s tendency to enable AI companies. But this is only a hope.
Third: I want good hygiene near CFAR and the rationalists / I don’t want to leave metaphorical rotting meat in our kitchen counter.
If you’ll pardon a metaphor: having living, healthy humans in a kitchen is mostly fine, hygiene-wise. Having a large slab of unrefrigerated meat sitting in the kitchen (no longer alive, and so no longer tied in with a living organism’s immune system), is a hygiene problem, especially after a while.
I suspect that if we have “living resource flows” across CFAR, the memes and habits and culture-bits that survive and spread here will mostly be good for us and others. I suspect by contrast that if we have ungrounded resource flows (ie, if we ignore this principle), we’ll risk breeding “parasitic memes” (or people) that are optimized to use up all the free energy in the system and that don’t tend to the conditions required for healthy life.
I mean it
If we can’t hit this principle (or the truer spirit behind it), my plan is to either figure out how to to hit it, or close CFAR.
(Although, here as elsewhere, I may revise my views; and I’ll update this post if I do; these principles are not permanent promises.)
Some principles you might assume we have that we don’t have:
Why we need your support / some cruxes for continuing this CFAR
There’s a sense in which we don’t need anybody. I could sit in my room, call myself an “applied rationality researcher,” and write things I called “rationality exercises” on paper or something.
But if we’re going to do something that’s not pretend, then we need people. And we need to find a way that there’s something in it for those people – a resource flow that gives back to them. (Otherwise, it’s still pretend.)
Why ask for donations?
We’re asking for donations because it takes money to run CFAR. If there are enthusiastic people out there who are willing and able to help fund us, that’ll both help a lot and seem wholesome. We aim to find a set of people who want the kind of practice we are building, and who want to build it, believe in its possibility, and try it together.
If nobody donates, we’ll likely continue; in extremity, we could e.g. sell our Bodega Bay venue, which would give us a few years’ operating expenses at our current, fairly minimalist budget. (That said, we love our venue and don’t want to sell it; more on that later.)
But if nobody donates and nobody cool wants to kibitz and all the people who try our workshop kinda want their time back and so on, of course we quit. Our main business in interacting with the community is to find a way to do cool stuff, via resources from some of you, in such a way that everyone’s glad. I suspect, but am not sure, that getting some donations from some of you is part of how to build the good, living center we are seeking.
Some disagree with us, and we’re doing this anyway
It is not the case that everyone who’s had much contact with past-CFAR believes resuming workshops is a good idea.
In particular:
There were also cheers: a sizable majority (at least of those I heard from) offered enthusiasm, well-wishes, “I’m glad there are again CFAR workshops where I can send my friends,” “I missed you guys,” etc. Former CFAR instructors Renshin (aka Lauren Lee) and Adam Scholl did this in the public comment thread. And I of course landed solidly at “yes, I want this enough that I’m willing to put in real effort.”
But I want to acknowledge that some disagree, for a few reasons:
Let me elaborate on (c): Back in 2014-2020, I would freak out whenever some serious thread of public conversation cast doubts on CFAR. I’d do this because I knew I needed CFAR staff’s morale, and I believed (accurately, I think) that many would lose their morale if even a small vocal minority said we were doing it wrong.
I believe our morale is somehow stabler now. (Perhaps partly because we factored aCFAR’s believing in’s out separately from our epistemics, and also because we’re a particular experiment we each want to do rather than a claim about the ‘objective best’).
I care about (c) for several reasons, but one is that I want good but imperfect institutions to exist in our present world, and to do this without suppressing news of their failures. Many of the previous decades’ institutions are gone from the world of 2025.[10] I think this is in significant part caused by the combination of:
Also, I put real effort into dismantling parts of my and CFAR’s positive reputation that I believed were false or ill-founded, and I did that partly because I didn’t think we could build something good near CFAR before that stuff was dismantled. Having completed that step (as I see it), I am eager to see what we can build on the new, partially razed ground.
Donations
Our finances
We currently have about $129k available for CFAR and its projects, which gives us about four months of runway.
To make it comfortably to the end of 2026, we think we need about $200k of additional donations (counting donations into this fundraiser, any SFF funding, and any other donations, but not counting workshop payments or venue rental revenue). We expect to probably get some money from SFF (probably in the form of matching funds, in about a week), and so are setting a “basic target” of $125k, and a “reach target” of $200k (as we can do more with more).
For more detail on that, see this breakdown:
General Costs
CFAR has ongoing general administrative costs – accounting, staff wages for administrative tasks, and so on. We think this will cost us about $72,000 for 2026. This is a very significant decrease from e.g. 2019, as CFAR is running with a smaller and leaner staff and no longer maintains office space.
Venue
We maintain an event venue in Bodega Bay, California, which we also rent out to other groups. This venue is both one of our primary expenses and also a source of revenue. Since 2020, the venue has been a significant net expense as we have run fewer programs there and not had many bookings. However, we now have venue caretakers who are sprucing the place up, figuring out what outside groups are looking for in a venue and how we can hit it, etc. We also expect to use our venue for more CFAR programs than we have been in the past few years.
For 2026, we estimate that we will likely have total venue costs of about $285,000. This is primarily mortgage payments, utilities, various maintenance/repair/”venue caretaking” work, and property taxes, although it also includes supplies for programs held at the venue. We also anticipate bringing in approximately $200,000 of revenue from outside bookings (after deducting cleaning fees), as well as using the venue for our own programs, hosting some staff meetings there, and so on. The savings from our own programs there are difficult to calculate but would likely be in the tens of thousands of dollars, perhaps $35,000 to $65,000 or so across 2026.
This means we anticipate the venue will on net cost us something like $20,000 to $50,000 for 2026, depending on how many programs we end up running there, how many outside workshops we hold, and what other costs we may incur. This is not ideal but we consider it a cost worth bearing for now, and in the long run we hope to run more programs there ourselves and bring in more outside bookings such that the venue ends up breaking even or being financially positive for CFAR.[11]
Workshops
Workshops are both a source of revenue and a significant cost for CFAR to run. Generally speaking, workshops gain or lose money based on how many staff members and participants are involved and how much financial aid we do or don’t offer to participants; a workshop with twenty-five participants paying full price would be profitable, while workshops with fewer participants and/or more financial aid may well lose money for CFAR on net. For instance, our November workshop ended up approximately -$28,400 on net.
In 2026, we currently anticipate running about four mainline workshops (one Jan 21-25 in Austin, TX and three yet to be announced). The workshop in Austin will incur venue costs that workshops held at our venue won’t. Insofar as the workshops otherwise have overall similar costs and revenues as the November workshop, we will probably be net minus ~$130,600 from workshops.[12]
We are excited to run these workshops even at a potential loss. In addition to being helpful to the participants, running workshops greatly aids our efforts to develop and refine an art of rationality. (In the long run, if our programs are any good, we should be able to fund the workshops more fully from those who attend, which will be better feedback loops, though we may want ongoing exceptions for students / folks without much money and for folks who are coming mostly to aid rationality development work.)
We also think that workshops benefit people beyond those who attend directly – some workshop attendees have gone on to teach others concepts like double crux and other CFAR techniques, and we think running workshops provides significant value for these “grandstudents”[13] as well.
In the past, CFAR has even offered some workshops for free – for instance, the four workshops we ran in the Czech Republic during autumn 2022 were entirely free to participants. However, the overall state of the funding environment was different when those programs were being planned and offering free mainline workshops currently seems imprudent.
Curriculum Development
In addition to the above costs, we also pay staff for general curriculum development outside of workshops – research into various aspects of rationality, work on new techniques, running test sessions where we try new material on volunteers, and so on. We project something like $25,000 in costs for this area in 2026, though this is somewhat speculative.
Aspirational
In addition to the core categories mentioned earlier, there are additionally various other projects that CFAR would like to be able to spend money on but currently is not.
For instance, in the past CFAR has supported “telos projects” – a program where CFAR provided funding for rationality-related projects that felt relevantly alive to people. In 2025, we had a few legacy projects in this area but are not soliciting new applications for telos funding; in a world where we had better funding we would like to reopen the program and use it to help new alumni run cool projects, including infrastructure for the new alumni community.
We would like to be able to pay me (Anna) to write various LessWrong posts about concepts CFAR has recently been working with, but are currently holding off on that. We would also like to (slowly, slightly) grow out our staff of curriculum developers and to modestly increase staff wages if we can.
Who I hope does, and doesn’t, consider donating
As mentioned earlier in this post, I’d like to build toward a world in which aCFAR’s donations come from, and with, the right kind of feedback loops.
I’m particularly cheerful (happy, relieved, joyful, grateful) about donations stemming from any of:
There’s something in here that you personally are rooting for, and you’re moved to root for it harder, with your dollars, so it can really be tried. (Like a home team or a city or a project in which you have pride and have/want membership)
The more dollars you deploy here, the more I hope you have some heart to spare to come along with your dollars, as “I care about this, and I’ll be kibitzing from the sidelines, and updating my total view of life based on how it goes, with enough context that my kibitzes and updates will make sense.” (The more of your dollars you deploy here, the easier we’ll try to make this “kibitzing from the sidelines” for you, if you’re willing.)
I’m particularly wary of donations stemming from:
To be clear, EA is an excellent way to donate; I’m glad some people donate this way; there’d be something seriously wrong with the world if nobody did this. But it’s not what this CFAR is doing. (More on this above.)
And in my opinion (and Michael Nielsen’s in this podcast with Ajeya Cotra, if you want a case at more length), there’d be something even more wrong with the world if most resource expenditure flowed via EA-like analysis.[15]
Another reason people used to sometimes donate, that IMO don’t apply to us today, and so would not be good reasons today:
And a couple other reasons to donate:
Ways to help CFAR or to connect to CFAR besides donating:
There are several good ways to help CFAR financially besides donating. You can:
There are also a pile of ways to help this CFAR and our mission non-financially. (Most of the resources we run on are non-financial, and are shared with us by hopeful rationality fans). Basically: kibitz with us here, or in a test session, or at a workshop. Attending a workshop helps even if you come on full scholarship a lot of the time, as having varied, cool participants makes our workshops more perspectives-rich and generative.)
For bonus points, maybe come to a workshop and then write up something substantial about it on LessWrong. (Scholarships are available for this purpose sometimes.)
Perks for donating
If you donate before Jan 31, you’ll also get, if you want:
Also, if there’s something in particular you’d like CFAR to be able to do, such as run workshops in a particular city or run an alumni event focusing on a particular component of rationality, and you’re considering a more substantial donation, please reach out (you can book a meeting via calendly, or email donate@rationality.org).
To the conversation!
Thank you for your curiosity about CFAR, and for reading (at least some of) this post! I hope you introduce yourself in the comments and that – if you end up donating (or kibitzing, or attending a workshop, or getting involved in us in whatever way) – it ends up part of a thing that’s actually good for you and the contexts you care about. And that you and we learn something together.
Yours in aspiring rationality,
Anna and aCFAR
'aCFAR' stands for “a Center For Applied Rationality.” We adopted the 'a' part recently, because calling ourselves 'the' Center for Applied Rationality seems obviously wrong. But feel free not to bother with the 'a' if it’s too annoying. I personally say 'a' when I feel like it.
One of the best ways to get to know someone is to team up on something concrete; kibitzing on a current CFAR stuck point is my suggestion for how to try a little of that between you and aCFAR.
Thanks to Davis Kingsley, John Salvatier, Paola Baca and Zvi Mowshowitz for writing help. (Particularly Davis Kingsley, who discussed practically every sentence, revised many, and made the whole thing far more readable.) Thanks to Jack Carroll for photos. Thanks to Zack Davis and Claude Code for creating the thermometer graphic up top. Remaining errors, wrong opinions, etc. are of course all mine.
My mom wanted to teach her kids math, so we could be smart. And I wanted… to be like her… which meant I also wanted to teach myself/others math! :) (Rather than, say, wanting to learn math.) Rationality education gives me an even better chance to see the gears of thinking/updating.
This overcounts a bit since this number is based on totaling the attendee count of many different programs and some people attended multiple programs, so the number of unique individuals who attended CFAR programs is lower than this.
EA spaces were receiving large influxes of new people at the time, and I hoped CFAR workshops could help the EA and rationality communities to assimilate the large waves of new people with less dilution of what made these spaces awesome. (Lightcone has mostly taken over the “develop and spread useful vocabulary, and acculturate newcomers” role in recent years, and has done it spectacularly IMO.)
Unlike some bands, we didn’t have substance abuse. But, like traveling bands, we traveled a lot to do high intensity soul-bearing stuff in a context where we were often exhausted but “the show must go on." I believe many of us, and many of our working relationships, got traveling-band-like scars. Also, we had ourselves a roster of potentially-kinda-invasive “CFAR techniques”; in hindsight some of our uses of these seem unwholesome to me. (I think these techniques are neat when used freely by an autonomous person, but are iffy at best when used to “help” a colleague stretch themselves harder for a project one is oneself invested in.)
There would still be many details to sort through. Eg, CFAR is aiming to be an unusually low-staff-charisma organization in which staff suggest exercises or whatever to participants in ways that’re unusually non-dizzying; Valentine’s native conversational style has a bit more charismatic oomph than we’re aiming for. But I love the idea of collaborating with Valentine on stuff about memes, PCK-seeking, what sorts of systematicity might allow decent epistemics, etc.. I also like the idea of having one more person who’s been around from the beginning, and has seen both CFAR’s early generativity and our failure modes, keeping an eye out.
We would also try to find other ways to make money, and tinker/brainstorm broadly.
For instance, mainstream media and academia both have much less credibility and notably less money, the ACLU lost most of its vitality, many of the big organizations in EA space from 2015ish have either ceased to do much public leadership there or ceased existing altogether, and I would guess the trends in Bowling Alone have continued although I have not checked.
It’s unlikely this would look like the venue generating more than its costs in direct booking revenue, but rather that the combination of booking revenue and cost savings for our own programs would exceed the costs of operating and maintaining the venue. Additionally we think the venue gives us a bunch of spirit and beauty, saves a bunch of staff time on logistics for each workshop we hold there, lets us support LARC and other groups we care about, and makes it easier for us to consider possible large expansions to our programs.
There’s a lot of variability in what workshops end up looking like and there’s some reason to believe later workshops may generate more revenue, but we’re using November here as the most obvious basis for comparison.
A term coined by Duncan meaning “students of our students” and which we continue to find useful in thinking about the impact of workshops and other programs.
Lighthaven, the LW website, and other Lightcone-enabled social contexts are truly remarkable, IMO – one of the last bastions of general-purpose grounded truthseeking conversation on the internet. Many of you feel most at home there, and so should be sending such donations only to Lightcone. But some should perhaps put some or all of their ‘I want to support contexts that support people like me, or that support conversations I’ll feel at home near’ budget toward CFAR. Personally, I'm donating $10k to Lightcone and putting soul and work into aCFAR, and this leaves me personally feeling happier and more a part of things than if I were to skip either.
Briefly: we humans are local creatures and we probably create better things, that contribute more to the long run, if we let ourselves have deep local interests and loyalties (to particular lines of research, to particular friendships and communities, to particular businesses or projects we are invested in) without trying to be always doing the thing that would be highest-impact for an detailless agent who happens to be us, and without trying to always be ready to change our plans and investments on a dime. I admit I’m caricaturing EA a bit, but I believe the point holds sans caricature; I would very much love to discuss this point at arbitrary length in the comment thread if you’re interested.