There's a vision here of what LessWrong could/should be, and what a rationalist community could/should be more generally. I want to push back against that vision, and offer a sketch of an alternative frame.
The post summarizes the vision I want to push back against as something like this:
What I really want from LessWrong is to make my own thinking better, moment to moment. To be embedded in a context that evokes clearer thinking, the way being in a library evokes whispers. To be embedded in a context that anti-evokes all those things my brain keeps trying to do, the way being in a church anti-evokes coarse language.
Now, I do think that's a great piece to have in a vision for the LessWrong or the rationalist community. But I don't think it's the central piece, at least not in my preferred vision.
What's missing? What is the central piece?
Fundamentally, the problem with this vision is that it isn't built for a high-dimensional world. In a high-dimensional world, the hard part of reaching an optimum isn't going-uphill-rather-than-downhill; it's figuring out which direction is best, out of millions of possible directions. Half the directions are marginally-good, half are marg...
I think I agree with the models here, and also want to add a complicating factor that I think impacts the relevance of this.
I think running a site like this in a fully consequentialist way is bad. When you're public and a seed of something, you want to have an easily-understandable interface with the world; you want it to be the case that other people who reason about you (of which there will be many, and who are crucial to your plan's success!) can easily reason about you. Something more like deontology or virtue ethics ("these are the rules I will follow" or "these are the virtues we will seek to embody") makes it much easier for other agents to reason about you.
And so the more that I as a mod (or the mod team in general) rely on our individual prudence or models or so on, the more difficult it becomes for users to predict what will happen, and that has costs. (I still think that it ultimately comes down to our prudence--the virtues that we're trying to embody do in fact conflict sometimes, and it's not obvious how to resolve those conflicts--but one of the things my prudence is considering are those legibility costs.)
And when we try to figure out what virtues we should embody on...
+1 to all this, and in particular I'm very strongly on board with rationality going beyond AI safety. I'm a big fan of LessWrong's current nominal mission to "accelerate intellectual progress", and when I'm thinking about making progress in a high-dimensional world, that's usually the kind of progress I'm thinking about. (... Which, in turn, is largely because intellectual/scientific/engineering progress seem to be the "directions" which matter most for everything else.)
I think your main point here is wrong.
Your analysis rests on a lot of assumptions:
1) It's possible to choose a basis which does a good job separating the slope from the level
2) Our perturbations are all small relative to the curvature of the terrain, such that we can model things as an n-dimensional plane
3) "Known" errors can be easily avoided, even in many dimensional space, such that the main remaining question is what the right answers are
4) Maintenance of higher standards doesn't help distinguish between better and worse directions.
5) Drama pushes in random directions, rather than directions selected for being important and easy to fuck up.
1) In a high dimensional space, almost all bases have the slope distributed among many basis vectors. If you can find a basis that has a basis vector pointing right down the gradient and the rest normal to it, that's great. If your bridge has one weak strut, fix it. However, there's no reason to suspect we can always or even usually do this. If you had to describe the direction of improvement from a rotting log to a nice cable stayed bridge, there's no way you could do it simply. You could name the direction "more better", but in order to act...
This is a great comment. There are some parts which I think are outright wrong (e.g. drama selects for most-important-disagreements), but for the most part it correctly identifies a bunch of shortcomings of the linear model from my comment.
I do think these shortcomings can generally be patched; the linear model is just one way to explain the core idea, and other models lead to the same place. The main idea is something like "in a high dimensional space, choosing the right places to explore is way more important than speed of exploration", and that generalizes well beyond linearity.
I'm not going to flesh that out more right now. This all deserves a better explanation than I'm currently ready to write.
Yeah, I anticipated that the "Drama is actually kinda important" bit would be somewhat controversial. I did qualify that it was selected "(if imperfectly)" :p
Most things are like "Do we buy our scratch paper from walmart or kinkos?", and there are few messes of people so bad that it'd make me want to say "Hey, I know you think what you're fighting about is important, but it's literally less important than where we buy our scratch paper, whether we name our log files .log or .txt, and literally any other random thing you can think of".
(Actually, now that I say this, I realize that it can fairly often look that way and that's why "bikeshedding" is a term. I think those are complicated by factors like "What they appear to be fighting about isn't really what they're fighting about", "Their goals aren't aligned with the goal you're measuring them relative to", and "The relevant metric isn't how well they can select on an absolute scale or relative to your ability, but relative to their own relatively meager abilities".)
In one extreme, you say "Look, you're fighting about this for a reason, it's clearly the most important thing, or at least top five, ignore anyone arguing otherwise".
In a...
If you had to describe the direction of improvement from a rotting log to a nice cable stayed bridge, there's no way you could do it simply. You could name the direction "more better", but in order to actually point at it or build a bridge, many many design choices will have to be made. In most real world problems, you need to look in many individual directions and decide whether it's an improvement or not and how far to go. Real world value is built on many "marginal" improvements.
This was an outstandingly useful mental image for me, and one I suspect I will incorporate into a lot of thoughts and explanations. Thanks.
EDIT: finished reading the rest of this, and it's tied (with Vaniver's) for my favorite comment on this post (at least as far as the object level is concerned; there are some really good comments about the discussions themselves).
Drama
I object to describing recent community discussions as "drama". Figuring out what happened within community organizations and holding them accountable is essential for us to have a functioning community. [I leave it unargued that we should have community.]
I agree that figuring out what happened and holding people/orgs accountable is important. That doesn't make the process (at least the process as it worked this time) not drama. I certainly don't think that the massive amount of attention the recent posts achieved can be attributed to thousands of people having a deeply-held passion for building effective organizations.
(I like the above and agree with most of it and am mulling and hope to be able to reply substantively, but in the meantime I wanted to highlight one little nitpick that might be more than a nitpick.)
Maybe finding and fixing organizational problems will lead to marginally more researcher time/effort on alignment, or maybe the drama itself will lead to a net loss of researcher attention to alignment. But these are both mechanisms of going marginally faster or marginally slower along the direction we're already pointed. In a high-dimensional world, that's not the sort of thing which matters much.
I think this leaves out a thing which is an important part of most people's values (mine included), which is that there's something bad about people being hurt, and there's something good about not hurting people, and that's relevant to a lot of people (me included) separate from questions of how it impacts progress on AI alignment. Like, on the alignment forum, I get subordinating people's pain/suffering/mistreatment to questions of mission progress (maybe), but I think that's not true of a more general place like LessWrong.
Put another way, I think there might be a gap between the importance you reflectively assign to the Drama, and the importance many others reflectively assign to it. A genuine values difference.
I do think that on LessWrong, even people's pain/suffering/mistreatment shouldn't trump questions of truth and accuracy, though. Shouldn't encourage us to abandon truth and accuracy.
Addendum to the quoted claim:
Maybe finding and fixing organizational problems will lead to marginally more researcher time/effort on alignment, or maybe the drama itself will lead to a net loss of researcher attention to alignment. But these are both mechanisms of going marginally faster or marginally slower along the direction we're already pointed. In a high-dimensional world, that's not the sort of thing which matters much.
... and it's also not the sort of thing which matters much for reduction of overall pain/suffering/mistreatment, even within the community. (Though it may be the sort of thing which matters a lot for public perceptions of pain/suffering/mistreatment.) This is a basic tenet of EA: the causes which elicit great public drama are not highly correlated with the causes which have lots of low-hanging fruit for improvement. Even within the rationalist community, our hardcoded lizard-brain drama instincts remain basically similar, and so I expect the same heuristic to apply: public drama is not a good predictor of the best ways to reduce pain/suffering/mistreatment within the community.
But that's a post-hoc explanation. My actual gut-level response to this comment was ...
Executive summary: I have no idea what you're talking about.
Standards are not really popular. Most people don't like them. Half the people here, I think, don't even see the problem that I'm trying to point at. Or they see it, but they don't see it as a problem.
I gather that you're upset about how the Leverage conversation went, and also Cancel Culture, so I assume your chief proposition is that LessWrong is canceling Geoff Anders; but you haven't actually made that case, just vaguely gestured with words.
I think that a certain kind of person is becoming less prevalent on LessWrong, and a certain other kind of person is becoming more prevalent, and while I have nothing against the other kind, I really thought LessWrong was for the first group.
What are the two kinds of persons? Really, I honestly do not know what you are claiming here. Repeat: I don't have even a foggy guess as to what your two kinds of person are. Am I "a certain other kind of person"? How can I know?
Distinguish feeling from fact.
This post has virtually no facts. It has short, punchy sentences with italics for emphasis. It is written like a hortatory sermon. Its primary tool is rhetoric. The f...
Yikes, despite Duncan's best attempts at disclaimers and clarity and ruling out what he doesn't mean, he apparently still didn't manage to communicate the thing he was gesturing at. That's unfortunate. (And also worries me whether I have understood him correctly.)
I will try to explain some of how I understand Duncan.
I have not read the first Leverage post and so cannot comment on those examples, but I have read jessicata's MIRI post.
and this still not having incorporated the extremely relevant context provided in this, and therefore still being misleading to anyone who doesn't get around to the comments, and the lack of concrete substantiation of the most radioactive parts of this, and so on and so forth.
As I understand it: This post criticized MIRI and CFAR by drawing parallels to Zoe Curzi's experience of Leverage. Having read the former but not the latter, the former seemed... not very substantive? Making vague parallels rather than object-level arguments? Merely mirroring the structure of the other post? In any case, there's a reason why the post sits at 61 karma with 171 votes and 925 comments, and that's not because it was considered uncontroversially true. Similarly, there's...
This comment was much longer in draft, but I've deleted the remainder because I don't want to seem "impatient" or "sneering". I'm just confused: You wrote all these words intending to convince people of something, but you don't specify what it is, and you don't use the tools we typically use to convince (facts, reliable sources, syllogistic reasoning, math, game theory...). Am I just not part of the intended audience? If so, who are they?
Thanks very much for taking the time to include this paragraph; it's doing precisely the good thing. It helps my brain not e.g. slide into a useless and unnecessary defensiveness or round you off to something you're not trying to convey.
I gather that you're upset about how the Leverage conversation went, and also Cancel Culture, so I assume your chief proposition is that LessWrong is canceling Geoff Anders; but you haven't actually made that case, just vaguely gestured with words.
That's not, in fact, my chief proposition. I do claim that something-like-the-mass-of-users is doing something-resembling-canceling-Leverage (such that e.g. if I were to propose porting over some specific piece of Leverage tech to LW or an EA org's internal cul...
I spent 15 minutes re-reading the thread underneath orthonormal's comment to try to put myself in your head. I think maybe I succeeded, so here goes, but from a person whose job involves persuading people, it's Not Optimal For Your Argument that I had to do this to engage with your model here, and it's potentially wasteful if I've failed at modeling you.
I read both of the comments discussed below, at the time I was following the original post and comments, but did not vote on either.
***
orthonormal P1: Anders seemed like a cult leader/wannabe based on my first impressions, and I willingly incurred social cost to communicate this to others
orthonormal P2 [which I inferred using the Principle of Charity]: Most of the time, people who immediately come across as cult leaders are trying to start a cult
Duncan P1: It's bad when LW upvotes comments with very thin epistemic rigor
Duncan P2: This comment has very thin epistemic rigor because it's based on a few brief conversations
Gloss: I don't necessarily agree with your P2. It's not robust, but nor is it thin; if true, it's one person's statement that, based on admittedly limited evidence, they had a high degree of confidence that...
Thank you for the effort! Strong upvoted.
Quick point to get out of the way: re: the comment that you thought would likely meet my standards, yes, it does; when I hovered over it I saw that I had already (weak) upvoted it.
Here's my attempt to rewrite orthonormal's first comment; what I would have said in orthonormal's shoes, if I were trying to say what I think orthonormal is trying to say.
...All right, here comes some subjective experience. I'm offering this up because it seems relevant, and it seems like we should be in wide-net data gathering mode.
I met Geoff Anders at our 2012 CFAR workshop, and my overwhelming impression was "this person wants to be a cult leader." This was based on [specific number of minutes] of conversation.
The impression stuck with me strongly enough that I felt like mentioning it maybe as many as [specific number] of times over the years since, in various conversations. I was motivated enough on this point that it actually somewhat drove a wedge between me and two increasingly-Leverage-enmeshed friends, in the mid-2010's.
I feel like this is important and relevant because it seems like yet again we're in a situation where a bunch of peo
Thanks, supposedlyfun, for pointing me to this thread.
I think it's important to distinguish my behavior in writing the comment (which was emotive rather than optimized - it would even have been in my own case's favor to point out that the 2012 workshop was a weeklong experiment with lots of unstructured time, rather than the weekend that CFAR later settled on, or to explain that his CoZE idea was to recruit teens to meddle with the other participants' CoZE) from the behavior of people upvoting the comment.
I expect that many of the upvotes were not of the form "this is a good comment on the meta level" so much as "SOMEBODY ELSE SAW THE THING ALL ALONG, I WORRIED IT WAS JUST ME".
Thanks for asking about the ITT.
I think that if I put a more measured version of myself back into that comment, it has one key difference from your version.
"Pay attention to me and people like me" is a status claim rather than a useful model.
I'd have said "pay attention to a person who incurred social costs by loudly predicting one later-confirmed bad actor, when they incur social costs by loudly predicting another".
(My denouncing of Geoff drove a wedge between me and several friends, including my then-best friend; my denouncing of the other one drove a wedge between me and my then-wife. Obviously those rifts had much to do with how I handled those relationships, but clearly it wasn't idle talk from me.)
Otherwise, I think the content of your ITT is about right.
(The emotional tone is off, even after translating from Duncan-speak to me-speak, but that may not be worth going into.)
For the record, I personally count myself 2 for 2.5 on precision. (I got a bad vibe from a third person, but didn't go around loudly making it known; and they've proven to be not a trustworthy person but not nearly as dangerous as I view the other two. I'll accordingly not name them.)
I'm going to take a stab at cruxing here.
Whether it's better for the LW community when comments explicitly state a reasonable amount of the epistemic hedging that they're doing.
Out of all the things you would have added to orthonormal's comment, the only one that I didn't read at the time as explicit or implicit in zir comment was, "Not as anything definitive, but if I do an honest scan over the past decade, I feel like I'm batting ... 3/5, maybe, with 2 more that are undecided, and the community consensus is doing more like 1/5". I agree it would be nice if people gave more information about their own calibration where available. I don't know whether it was available to orthonormal.
As for the rest, I'm sticking that at the end of this comment as a sort of appendix.
If I'm right about the crux, that is totally not in the set of Things That I Thought You Might Have Been Saying after reading the original post. Re-reading the original post now, I don't see how I could have figured out that this is what our actual disagreement was.
I notice that I am surprised that {the norm of how explicit a comment needs to be regarding its own epistemic standard} prompted you to write the original pos...
Understood: the comment-karma-disparity issue is, for you, a glaring example of a larger constellation.
Also understood: you and I have different preferences for explicitly stating underlying claims. I don't think your position is unreasonable, just that it will lead to much-longer comments possibly at the cost of clarity and engagement. Striking that balance is Hard.
I think we've drilled as far down as is productive on my concerns with the text of your post. I would like to see your follow-up post on the entire constellation, with the rigor customary here. You could definitely persuade me. I maybe was just not part of the target audience for your post.
It's bad that comments which are good along three different axes, and bad along none as far as I can see, are ranked way below comments that are much worse along those three axes and also have other flaws
I have an alternative and almost orthogonal interpretation for why the karma scores are the way they are.
Both in your orthonormal-Matt example, and now in this meta-example, the shorter original comments require less context to understand and got more upvotes, while the long meandering detail-oriented high-context responses were hardly even read by anyone.
This makes perfect sense to me - there's a maximum comment length after which I get a strong urge to just ignore / skim a comment (which I initially did with your response here; and I never took the time to read Matt's comments, though I also didn't vote on orthonormal's comment one way or another, nor vote in the jessicata post much at all), and I would be astonished if that only happened to me.
Also think about how people see these comments in the first place. Probably a significant chunk comes from people browsing the comment feed on the LW front page, and it makes perfect sense to scroll past a long sub-sub-sub-comment that mig...
First, some off the cuff impressions of matt's post (in the interest of data gathering):
In the initial thread I believe that I read the first paragraph of matt's comment, decided I would not get much out of it, and stopped reading without voting.
Upon revisiting the thread and reading matt's comment in full, I find it difficult to understand and do not believe I would be able to summarize or remember its main points now, about 15 minutes after the fact.
This seems somewhat interesting to test, so here is my summary from memory. After this I'll reread matt's post and compare what I thought it said upon first reading with what I think it says upon a second closer reading:
[person who met geoff] is making anecdotal claims about geoff's cult-leader-ish nature based on little data. People who have much more data are making contrary claims, so it is surprising that [person]'s post has so many upvotes. [commenter to person] is using deadpan in a particular way, which could mean multiple things depending on context but I lack that context. I believe that they are using it to communicate that geoff said so in a non-joking manner, but that is also hearsay.
Commentary before re-reading: I expect ...
Strong upvote for doing this process/experiment; this is outstanding and I separately appreciate the effort required.
Do other readers find my summary, particularly my more concise summary, an accurate portrayal of matt's comment? How would they react to that much more concise comment, as compared to matt's comment?
I find your summary at least within-bounds, i.e. not fully ruled out by the words on the page. I obviously had a different impression, but I don't think that it's invalid to hold the interpretations and hypotheses that you do.
I particularly like and want to upvote the fact that you're being clear and explicit about them being your interpretations and hypotheses; this is another LW-ish norm that is half-reliable and I would like to see fully reliable. Thanks for doing it.
To add one point:
When it comes to assessing whether a long comment or post is hard to read, quality and style of writing matters, too. SSC's Nonfiction Writing Advice endlessly hammers home the point of dividing text into increasingly smaller chunks, and e.g. here's one very long post by Eliezer that elicited multiple comments of the form "this was too boring to finish" (e.g. this one), some of which got alleviated merely by adding chapter breaks.
And since LW makes it trivial to add headings even to comments (e.g. I used headings here), I guess that's one more criterion for me to judge long comments by.
(One could even imagine the LW site nudging long comments towards including stuff like headings. One could imagine a good version of a prompt like this: "This comment / post is >3k chars but consists of only 3 paragraphs and uses no headings. Consider adding some level of hierarchy, e.g. via headings.")
All right, a more detailed response.
You're fighting fire with fire.
I am not fighting fire with fire. I request that you explicitly retract the assertion, given that it is both a) objectively false, and b) part of a class of utterances that are in general false far more often than they are true, and which tend to make it harder to think and see clearly in exactly the way I'm gesturing at with the OP.
Some statements that would not have been false:
"This seems to me like it's basically fighting fire with fire."
"I believe that, in practice, this ends up being fighting fire with fire."
"I'm having a hard time summing this up as anything other than 'fighting fire with fire.'"
...and I reiterate that those subtle differences make a substantial difference in people's general ability to do the collaborative truth-seeking thing, and are in many ways precisely what I'm arguing for above.
I clearly outline what I am identifying as "fire" in the above post. I have one list which is things brains do wrong, and another list which lays out some "don'ts" that roughly correspond to those problems.
I am violating none of those don'ts, and, in my post, exhibiting none of those wrongbrains. &nbs...
I would like to register that I think this is an excellent comment, and in fact caused me to downvote the grandparent where I would otherwise have neutral or upvoted. (This is not the sort of observation I would ordinarily feel the need to point out, but in this case it seemed rather appropriate to do so, given the context.)
Huh. Interesting.
I had literally the exact same experience before I read your comment dxu.
I imagine it's likely that Duncan could sort of burn out on being able to do this [1] since it's pretty thankless difficult cognitive work. [2]
But it's really insightful to watch. I do think he could potentially tune up [3] the diplomatic savvy a bit [4] since I think while his arguments are quite sound [5] I think he probably is sometimes making people feel a little bit stupid via his tone. [6]
Nevertheless, it's really fascinating to read and observe. I feel vaguely like I'm getting smarter.
###
Rigor for the hell of it [7]:
[1] Hedged hypothesis.
[2] Two-premise assertion with a slightly subjective basis, but I think a true one.
[3] Elaborated on a slightly different but related point further in my comment below to him with an example.
[4] Vague but I think acceptably so. To elaborate, I mean making one's ideas even when in disagreement with a person palatable to the person one is disagreeing with. Note: I'm aware it doesn't acknowledge the cost of doing so and running that filter. Note also: I think, with skill and practice, this can be done without sacrificing the content of the message. It is a...
Given that there is lots of "let's comment on what things about a comment are good and which things are bad" going on in this thread, I will make more explicit a thing that I would have usually left implicit:
My current sense is that this comment maybe was better to write than no comment, given the dynamics of the situation, but I think the outcome would have been better if you had waited to write your long comment. This comment felt like it kicked up the heat a bunch, and while I think that was better than just leaving things unresponded, my sense is the discussion overall would have gone better if you had just written your longer comment.
Jumping in here in what i hope is a prosocial way. I assert as hypothesis that the two of you currently disagree about what level of meta the conversation is/should-be at, and each feels that the other has an obligation to meet them at their level, and this has turned up the heat a lot.
maybe there is a more oblique angle then this currently heated one?
I'll have more to say on this in the future, but for now I just want to ramble about something.
I've been reading through some of the early General Semantics works. Partially to see if there are any gaps in my understanding they can fill, partially as a historical curiosity (how much of rationality did they have figured out, and what could they do with it?), partially because it might be good fodder for posts on LW (write a thousand posts to Rome).
And somehow a thing that keeps coming into my mind while reading them is the pre-rigorous/rigorous/post-rigorous split Terrence Tao talks about, where mathematicians start off just doing calculation and not really understanding proofs, and then they understand proofs through careful diligence, and then they intuitively understand proofs and discard many of the formalisms in their actions and speech.
Like, the early General Semantics writers pay careful attention to many things that I feel like I have intuitively incorporated; they're trying to be rigorous about scientific thinking (in the sense that they mean) in a way that I think I can be something closer to post-rigorous. Rather than this just being "I'm sloppier than they were", I think...
For me, what this resonates most clearly with is the interaction I just had with Ben Pace.
Ben was like "X"
And I was like "mostly yes to X, but also no to the implicature that I think surrounds X which is pretty bad."
And Ben was like "oh, definitely not that! Heck no! Thanks for pointing it out, but no, and also I think I'll change nothing in response to finding out that many people might draw that from X!"
And my response was basically "yeah, I don't think that Ben Pace on the ideal LessWrong should do anything different."
Because the thing Ben said was fine, and the implicature is easy to discard/get past. Like "almost certainly Ben Pace isn't trying to imply [that crap], I don't really need to feel defensive about it, I can just offhandedly say 'by the way, not [implication], and it's fine for Ben to not have ruled that out just like it's fine for Ben to not also actively rule out 'by the way, don't murder people' in every comment."
But that's because Ben and I have a high-trust, high-bandwidth thing going.
The more that LW as a whole is clean-and-trustworthy in the way that the Ben-Duncan line segment is clean-and-trustworthy, the less that those implications are f...
And maybe I should update in that direction, and just ignore a constant background shrieking.
I'm not sure about this; there is some value in teaching undergrads rigor, and you seem more motivated to than I am. And, like, I did like Logan's comment about rumor, and I think more people observing things like that sooner is better. I think my main hope with the grandparent was to check if you're thinking the rigor is the moon or the finger, or something.
Some strings of wisdom that seem related:
"You have to know the rules before you can break them."
There has to be some sense that you're riffing deliberately and not just wrong about the defaults
The ability to depart from The Standard Forms is dependent on both the level of trust and the number of bystanders who will get the wrong idea (see my and Critch's related posts, or my essay on the social motte-and-bailey).
"Level three players can't distinguish level two players from level four players."
This suggests to me a different idea on how to improve LessWrong: make an automated "basics of disagreement" test. This involves recognizing a couple of basic concepts like cruxes and common knowledge, and involves looking at some comment threads and correctly diagnosing "what's going on" in them (e.g. where are they talking past each other) and you have to notice a bunch of useful ways to intervene.
Then if you pass, your username on comments gets a little badge next to it, and your strong vote strength gets moved up to +4 (if you're not already there).
The idea is to make it clearer who is breaking the rules that they know, versus who is breaking the rules that they don't know.
This post makes me kind of uncomfortable and I feel like the locus is in... bad boundaries maybe? Maybe an orientation towards conflict, essentializing, and incentive design?
Here's an example where it jumped out at me:
Another metaphor is that of a garden.
You know what makes a garden?
Weeding.
Gardens aren't just about the thriving of the desired plants. They're also about the non-thriving of the non-desired plants.
And weeding is hard work, and it's boring, and it's tedious, and it's unsexy.
Here's another:
But gardens aren't just about the thriving of the desired plants. They're also about the non-thriving of the non-desired plants.
There's a difference between "there are many black ravens" and "we've successfully built an environment with no white ravens." There's a difference between "this place substantially rewards black ravens" and "this place does not reward white ravens; it imposes costs upon them."
Like... this is literally black and white thinking?
And why would a good and sane person ever want to impose costs on third parties ever except like in... revenge because we live in an anarchic horror world, or (better) as punishment after a wise and just proceed...
Like... this is literally black and white thinking?
Yes, because there is in fact a difference between "stuff that promotes individuals' and groups' ability to see and think clearly" and stuff that does not, and while we have nowhere near a solid grasp on it, we do know some of it really well at this point. There are some things that just do not belong in a subculture that's trying to figure out what's true.
Some things are, in fact, better than others, especially when you have a goal/value set/utility function.
And why would a good and sane person ever want to impose costs on third parties ever except like in... revenge because we live in an anarchic horror world, or (better) as punishment after a wise and just proceeding where rehabilitation would probably fail but deterrence might work?
Note the equivocation here between "I can't think of a reason why someone would want to impose costs on third parties except X" and "therefore there probably aren't non-X reasons." This is an example of the thing; it makes it harder to see the answer to the question "why?" Harder than it has to be, by making the "why" seem rhetorical, and as-if-there-couldn't-be-a-good-answer.
I'd like to impose some costs on statements like the above, because they tend to confuse things.
On epistemic grounds: The thing you should be objecting to in my mind is not the part where I said that "because I can't think of a reason for X, that implies that there might not be a reason for X".
(This isn't great reasoning, but it is the start of something coherent. (Also, it is an invitation to defend X coherently and directly. (A way you could have engaged with is by explaining why adversarial attacks on the non-desired weeds would be a good use of resources rather than just... like... living and letting live, and trying to learn from things you initially can't appreciate?)))
On human decency and normative grounds: The thing you should be objecting to is that I directly implied that you personally were might not be "sane and good" because your advice seemed to be violating ideas about conflict and economics that seem normative to me.
This accusation could also have an epistemic component (which would be an ad hominem) if I were saying "you are saying X and are not sane and good and therefore not-X". But I'm not saying this.
I'm saying that your proposed rules are bad because they request expensive actions for unclear benefits that seem likely to lead to unproductive ...
If you think I'm irrational, please enumerate the ways. Please be nuanced and detailed and unconfused. List 100 little flaws if you like.
I'm having a hard time doing this because your two comments are both full of things that seem to me to be doing exactly the fog-inducing, confusion-increasing thing. But I'm also reasonably confident that my menu of options looks like:
...
- Don't respond, and the-audience-as-a-whole, i.e. the-culture-of-LessWrong, will largely metabolize this as tacit admission that you were right, and I was unable to muster a defense because I don't have one that's grounded in truth
- Respond in brief, and the very culture that I'm saying currently isn't trying to be careful with its thinking and reasoning will round-off and strawman and project onto whatever I say. This seems even likelier than usual here in this subthread, given that your first comment does this all over the place and is getting pretty highly upvoted at this point.
- Respond at length, here but not elsewhere, and try to put more data and models out there to bridge the inferential gaps (this feels doomy/useless, though, because this is a site already full of essays detailing all of the things wrong with your comments)
- Respond at length to all such comments, even though it's easier to produce bullshit than to refute bullshit, meaning that I'm basically committing to put forth two hours of effort for every one that other people can throw at me, which is a recipe for exhaustion and demoralization and failure, and which is precisely why the OP was written. "People not d
I'd contend that a post can be "in good faith" in the sense of being a sincere attempt to communicate your actual beliefs and your actual reasons for them, while nonetheless containing harmful patterns such as logical fallacies, misleading rhetorical tricks, excessive verbosity, and low effort to understand your conversational partner. Accusing someone of perpetuating harmful dynamics doesn't necessarily imply bad faith.
In fact, I see this distinction as being central to the OP. Duncan talks about how his brain does bad things on autopilot when his focus slips, and he wants to be called on them so that he can get better at avoiding them.
I want to reinforce the norm of pointing out fucky dynamics when they occur...
Calling this subthread part of a fucky dynamic is begging the question a bit, I think.
If I post something that's wrong, I'll get a lot of replies pushing back. It'll be hard for me to write persuasive responses, since I'll have to work around the holes in my post and won't be able to engage the strongest counterarguments directly. I'll face the exact quadrilemma you quoted, and if I don't admit my mistake, it'll be unpleasant for me! But, there's nothing fucky happening: that's just how it goes when you're wrong in a place where lots of bored people can see.
When the replies are arrant, bad faith nonsense, it becomes fucky. But the structure is the same either way: if you were reading a thread you knew nothing about on an object level, you wouldn't be able to tell whether you were looking at a good dynamic or a bad one.
So, calling this "fucky" is calling JenniferRM's post "bullshit". Maybe that's your model of JenniferRM's post, in which case I guess I just wasted your time, sorry about that. If not, I hope this was a helpful refinement.
(My sense is that dxu is not referring to JenniferRM's post, so much as the broader dynamic of how disagreement and engagement unfold, and what incentives that creates.)
It's often useful to have possibly false things pointed out to keep them in mind as hypotheses or even raw material for new hypotheses. When these things are confidently asserted as obviously correct, or given irredeemably faulty justifications, that doesn't diminish their value in this respect, it just creates a separate problem.
A healthy framing for this activity is to explain theories without claiming their truth or relevance. Here, judging what's true acts as a "solution" for the problem, while understanding available theories of what might plausibly be true is the phase of discussing the problem. So when others do propose solutions, do claim what's true, a useful process is to ignore that aspect at first.
Only once there is saturation, and more claims don't help new hypotheses to become thinkable, only then this becomes counterproductive and possibly mostly manipulation of popular opinion.
I liked the effort put into this comment, and found it worth reading, but disagree with it very substantially. I also think I expect it to overall have bad consequences on the discussion, mostly via something like "illusion of transparency" and "trying to force the discussion to happen that you want to happen, and making it hard for people to come in with a different frame", but am not confident.
I think the first one is sad, and something I expect would be resolved after some more rounds of comments or conversations. I don't actually really know what to do about the second one, like, on a deeper level. I feel like "people wanting to have a different type of discussion than the OP wants to have" is a common problem on LW that causes people to have bad experiences, and I would like to fix it. I have some guesses for fixes, but none that seem super promising. I am also not totally confident it's a huge problem and worth focussing on at the margin.
In light of your recent post on trying to establish a set of norms and guidelines for LessWrong (I think you accidentally posted it before it was finished, since some chunks of it were still missing, but it seemed to elaborate on things you put forth in stag hunt), it seems worthwhile to revisit this comment you made about a month ago that I commented on. In my comment I focused on the heat of your comment, and how that heat could lead to misunderstandings. In that context, I was worried that a more incisive critique would be counterproductive. Among other things, it would be increasing the heat in a conversation that I believed to be too heated. The other worries were that I expected that you would interpret the critique as an attack that needed defending, I intuited that you were feeling bad and that taking a very critical lens to your words would worsen your mood, and that this comment is going to take me a bunch of work (Author's note: I've finished writing it. It took about 6 hours to compose, although that includes some breaks). In this comment, I'm going to provide that more incisive critique.
My goal is to engender a greater degree of empathy in you when you engage with comm...
I'm glad you took the time to respond here, and there is a lot I like about this comment. In particular, I appreciate this comment for:
There are, however, some points of disagreement I'd like to raise and some possible deleterious consequences I'd like to flag.
I share the concern raised by habryka about the illusion of transparency, which may be increasing your confidence that you are interpreting the intended meaning (and intended consequences) of Jennifer's words. I'll go into (possibly too much) detail on one very short example of what you've written and how it may involve some misreading of Jennifer's comment. You quote Jennifer:
Perhaps you could explain "epistemic hygiene" to me in mechanistic detail, and show how I'm messing it up?
and respond:
Again the trap; ...
I was also confused about what you meant by epistemic hygiene when finishing the essays. Elsewhere someone asked whether they were one of the ones doing the bad thing you we...
[Obvious disclaimer: I am not Duncan, my views are not necessarily his views, etc.]
It seems to me that your comment is [doing something like] rounding off Duncan's position to [something like] conflict theory, and contrasting it to the alternative of a mistake-oriented approach. This impression mostly comes from passages like the following:
You're sad about the world. I'm sad about it too. I think a major cause is too much poking. You're saying the cause is too little poking. So I poked you. Now what?
If we really need to start banning the weeds, for sure and for true... because no one can grow, and no one can be taught, and errors in rationality are terrible signs that a person is an intrinsically terrible defector... then I might propose that you be banned?
...And obviously this is inimical to your selfish interests. Obviously you would argue against it for this reason if you shared the core frame of "people can't grow, errors are defection, ban the defectors" because you would also think that you can't grow, and I can't grow, and if we're calling for each other's banning based on "essentializing pro-conflict social logic" because we both think the other is a "weed"... well..
Thank you for this great comment. I feel bad not engaging with Duncan directly, but maybe I can engage with your model of him? :-)
I agree that Duncan wouldn't agree with my restatement of what he might be saying.
What I attributed to him was a critical part (that I object to) of the entailment of the gestalt of his stance or frame or whatever. My hope was that his giant list of varying attributes of statements and conversational motivations could be condensed into a concept with a clean intensive definition other than a mushy conflation of "badness" and "irrational". For me these things are very very different and I'll say much more about this below.
One hope I had was that he would vigorously deny that he was advocating anything like what I mentioned by making clear that, say, he wasn't going to wander around (or have large groups of people wander around) saying "I don't like X produced by P and so let's impose costs (ie sanctions (ie punishments)) on P and on all X-like things, and if we do this search-and-punish move super hard, on literally every instance, then next time maybe we won't have to hunt rabbits, and we won't have to cringe and we won't have to feel angry at ever...
dxu:
my model of Duncan predicts that there are some people on LW whose presence here is motivated (at least significantly in part) by wanting to grow as a rationalist,
Jennifer:
I think that it is likely that neither Duncan nor I likely consider ourselves in the first category.
Duncan, in the OP, which Jennifer I guess skimmed:
What I really want from LessWrong is to make my own thinking better, moment to moment. To be embedded in a context that evokes clearer thinking, the way being in a library evokes whispers. To be embedded in a context that anti-evokes all those things my brain keeps trying to do, the way being in a church anti-evokes coarse language.
… modular sub-communities …
… a staging area and audition space for more specific and more demanding subcultures …
Here is a thng I wrote some years ago (this is a slightly cleaned up chat log, apologies for the roughness of exposition):
...There was an analogue to this in WoW as well, where, as I think I’ve mentioned, there often was such a thing as “within this raid guild, there are multiple raid groups, including some that are more ‘elite’/exclusive than the main one”; such groups usually did not use the EPGP or other allocation system of the main group, but had their own thing.
(I should note that such smaller, more elite/exclusive groups, typically skewed closer to “managed communism” than to “regulated capitalism” on the spectrum of loot systems, which I do not think is a coincidence.)
[name_redacted]: Fewer people, higher internal trust presumably.
“Higher internal trust” is true, but not where I’d locate the cause. I’d say “higher degree of sublimation of personal interest to group interest”.
[name_redacted]: Ah. … More dedicated?
Yes, and more willing to sacrifice for the good of the raid. Like, if you’re trying to maintain a raiding guild of 100 people, keep it functioning a
I quoted it rather than paraphrasing because I wanted to not put words in your mouth while you were in a potentially adversarial mood
If a person writes "I currently get A but what I really want is B"
...and then you selectively quote "I currently get A" as justification for summarizing them as being unlikely to want B...
...right after they've objected to you strawmanning and misrepresenting them left and right, and made it very clear to you that you are nowhere near passing their ITT...
...this is not "simplification."
Apologizing for "over-simplifying," under these circumstances, is a cop-out. The thing you are doing is not over-simplification. You are [not talking about simpler versions of me and my claim that abstract away some of the detail]. You are outright misrepresenting me, and in a way that's reeeaaalll hard to believe is not adversarial, at this point.
It is at best falling so far short of cooperative discourse as to not even qualify as a member of the set, and at worst deliberate disingenuousness.
If a person wholly misses you once, that's run-of-the-mill miscommunication.
If, after you point out all the ways they missed you, at length, they brush that off a...
There's a lot here, and I've put in a lot of work writing and rewriting. After failing for long enough to put things in a way that is both succinct and clear, I'm going to abandon hopes of the latter and go all in on the former. I'm going to use the minimal handles for the concepts I refer to, in a way similar to using LW jargon like "steelman" without the accompanying essays, in hopes that the terms are descriptive enough on their own. If this ends up being too opaque, I can explicate as needed later.
Here's an oversimplified model to play with:
GWS took an approach of offering proof of security and making fairly modest bids for both security and respect. As a result, the message was accepted, but it was fairly restrained in what it attempted to communicate. Fo...
And why would a good and sane person ever want to impose costs on third parties ever except like in... revenge because we live in an anarchic horror world, or (better) as punishment after a wise and just proceeding where rehabilitation would probably fail but deterrence might work?
This paragraph sounds to me like when you say "costs" you are actually thinking of "punishments", with an implication of moral wrongdoing. I'm uncertain that Duncan intended that implication (and if he did, I'd like to request that both of you use the more specific term).
If you continue to endorse the quoted paragraph as applied to "costs" that are not necessarily punishments, then I further contend that costs are useful in several scenarios where there is no implication of wrongdoing:
The SUPER obvious example is markets, where costs are used as a mechanism to control resource allocation.
Another common scenario is as a filter for seriousness. Suppose you are holding a contest where you will award a prize to the best foozle. If entry in the contest is free of charge, you might get a lot of entries from amateur foozle artists who know that they have negligible chance of winning...
Please don't make this place worse again by caring about points for reasons other than making comments occur in the right order on the page.
I wish this statement explaining what goal your advice is designed to optimize had appeared at the top of the advice, rather than the bottom.
My current world-model predicts that this is not what most people believe points are for, and that getting people to treat points in this way would require a high-effort coordinated push, probably involving radical changes to the UI to create cognitive distance from how points are used on other sites.
Specifically, I think the way most people actually use points is as a prosthetic for nonverbal politics; they are the digital equivalent of shooting someone a smile or a glower. Smiles/glowers, in turn, are a way of informing the speaker that they are gaining/losing social capital, and informing bystanders that they could potentially gain/lose social capital depending on which side of the issue they support.
My model says this is a low-level human instinctive social behavior, with the result that it is very easy to make people behave this way, but simultaneously very hard for most people to explain ...
Like... this is literally black and white thinking?
This is written in a way that seems to imply that if it is black and white thinking that would be bad. It also doesn't read as a question despite having a question mark.
People whos neurotype makes them default to black and white thinking can get really good when a concept does apply or doesn't apply. It has strengths and weaknesses. You are taking the attitude that it is widely known for its weaknesses. Demonstrating what is being glossed over or what kinds of things would be missed by it. I guess later in the post there are descipriton how other stances have it better but I think failure analysis on what is worse for the strongly dualistic view is sorely needed.
One possible source of uncomfortableness if if the typical mind assumptions doesn't hold because the interaction target mind is atypical.
Phrasings like "And why would a good and sane person ever [...]" seem to prepare to mark individuals for rejection. And again it has a question word but doesn't read like a question.
I am worried about a mechanic where people recognised as deviant are categorised as undesirable.
I do share worry about a distinction between "in" and "out" leads to a very large out group and actions that are as large as possible instead of proprotioned to neccesity or level of guarantee.
"Black and white thinking" is another name for a reasonably well defined cognitive tendency that often occurs in proximity to reasonably common mental problems.
Part of the reason "the fallacy of gray" is a thing that happens is that advice like that can be a useful and healthy thing for people who are genuinely not thinking in a great way.
Adding gray to the palette can be a helpful baby step in actual practice.
Then very very similar words to this helpful advice can also be used to "merely score debate points" on people who have a point about "X is good and Y is bad". This is where the "fallacy" occurs... but I don't think the fallacy would occur if it didn't have the "plausible cover" that arises from the helpful version.
A typical fallacy of gray says something like "everything is gray, therefore lets take no action and stop worrying about this stuff entirely".
One possible difference, that distinguishes "better gray" from "worse gray" is whether you're advocating for fewer than 2 or more than 2 categories.
Compare: "instead of two categories (black and white), how about more than two categories (black and white and gray), or maybe even five (pure black, dark gray, gray, ...
An important note, that I couldn't manage to fold into the essay proper:
I have direct knowledge of at least three people who would like to say positive things about their experience at Leverage Research, but feel they cannot. People who are fully on board with the fact that their experiences do not erase Zoe's, but whose own very different stories are relevant to understanding what actually went wrong, so that it can be fixed for those who suffered, and prevented in the future.
And at least these three people are not speaking up, because the current state of affairs is such that they feel they can't do so without committing social suicide.
This is a fact about them, not a fact about LessWrong or what would actually happen.
But it sure is damning that they feel that way, and that I can't exactly tell them that they're wrong.
Personally, I think it's clear that there was some kind of disease here, and I'd like to properly diagnose it, and that means making it possible to collect all of the relevant info, not an impoverished subset.
We were missing Zoe's subset. It's good that we have it now.
It's bad that we're still dealing with an incomplete dataset, though, and it's really bad that a lot of people seem to think that the dataset isn't meaningfully incomplete.
But it sure is damning that they feel that way, and that I can't exactly tell them that they're wrong.
You could have, though. You could have shown them the many highly-upvoted personal accounts from former Leverage staff and other Leverage-adjacent people. You could have pointed out that there aren't any positive personal Leverage accounts, any at all, that were downvoted on net. 0 and 1 are not probabilities, but the evidence here is extremely one-sided: the LW zeitgeist approves of positive personal accounts about Leverage. It won't ostracize you for posting them.
But my guess is that this fear isn't about Less Wrong the forum at all, it's about their and your real-world social scene. If that's true then it makes a lot more sense for them to be worried (or so I infer, I don't live in California). But it makes a lot less to bring to bring it up here, in a discussion about changing LW culture: getting rid of the posts and posters you disapprove of won't make them go away in real life. Talking about it here, as though it were an argument in any direction at all about LW standards, is just a non sequitur.
When it comes to the real-life consequences I think we're on the same page: I think it's plausible that they'd face consequences for speaking up and I don't think they're crazy to weigh it in their decision-making (I do note, for example, that none of the people who put their names on their positive Leverage accounts seem to live in California, except for the ones who still work there). I am not that attached to any of these beliefs since all my data is second- and third-hand, but within those limitations I agree.
But again, the things they're worried about are not happening on Less Wrong. Bringing up their plight here, in the context of curating Less Wrong, is not Lawful: it cannot help anybody think about Less Wrong, only hurt and distract. If they need help, we can't help them by changing Less Wrong; we have to change the people who are giving out party invites and job interviews.
"three people... would like to say positive things about their experience at Leverage Research, but feel they cannot":
Oof. I appreciate you mentioning that.
(And a special note of thanks, for being willing to put down a concrete number? It helps me try to weigh it appropriately, while not compromising anonymity.)
Navigating the fact that people seem to be scared of coming forward on every side of this, is hard. I would love advice on how to shape this thread better.
If you think of something I can do to make talking about all of {the good, the bad, the neutral, the ugly, and the complicated}, easier? I can't guarantee I'll agree to it, but I really do want to hear it.
Please feel free to reach out to me on LW, anytime in the next 2 months. Not just Duncan, anyone. Offer does expire at start of January, though.
I am especially interested in concrete suggestions that improve the Pareto Frontier of reporting, here. But I'm also pretty geared up to try to steelman any private rants that get sent my way, too.
(In this context? I have already been called all of "possessed, angry, jealous, and bad with secrets." I was willing to steelman the lot, because there is a hint of truth in each, alth...
Hm... I notice I'm maybe feeling some particular pressure to personally address this one?
Because I called out the deliberate concentration of force in the other direction that happened on an earlier copy of the BayAreaHuman thread.
I am not really recanting that? I still think something "off" happened there.
But I could stand up and give a more balanced deposition.
To be clear? I do think BAH's tone was a tad aggressive. And I think there were other people in the thread, who were more aggressive than that. I think Leverage Basic Facts EA had an even more aggressive comment thread.
I also think each of the concrete factual claims BAH made, did appear to check out with at least one corner of Leverage, according to my own account-collecting (although not always at the same time).
(I also think a few of the LBFEA's wildest claims, were probably true. Exclusion of the Leverage website from Wayback Machine is definitely true*. The Slack channels characterizing each Pareto attendee as a potential recruit, seems... probably true?)
There were a lot of corners of Leverage, though. Several of them were walled off from the corners BAH talked about, or were not very near to it.
For what it's worth, ...
LW is likely currently on something like a Pareto frontier of several values, where it is difficult to promote one value better without sacrificing others. I think that this is true, and also think that this is probably what OP believes.
The above post renders one axis of that frontier particularly emotionally salient, then expresses willingness to sacrifice other axes for it.
I appreciate that the post explicitly points out that is willing to sacrifice these other axes. It nevertheless skims a little bit over what precisely might be sacrificed.
Let's name some things that might be sacrificed:
(1) LW is a place newcomers to rationality can come to ask questions, make posts, and participate in discussion, hopefully without enormous barrier to entry. Trivial inconveniences to this can have outsized effects.
(2) LW is a kind of bulletin board and coordination center for things of general interest to an actual historical communities. Trivial inconveniences to sharing such information can once again have an outsized effect.
(3) LW is a place to just generally post things of interest, including fiction, showerthoughts, and so on, to the kind of person who is interested in rationality, AI, ...
I meant a relative Pareto frontier, vis-a-vis the LW team's knowledge and resources. I think your posts on how to expand the frontier are absolutely great, and I think they (might) add to the available area within the frontier.
"If you want to suggest that OP is part of a "genre of rhetoric": make the case that it is, name it explicitly."
I mean, most of OP is about evoking emotion about community standards; deliberately evoking emotions is a standard part of rhetoric. (I don't know what genre -- ethos if you want to invoke Aristotle -- but I don't think it particularly matters.) OP explicitly says that he would like LW to be smaller -- i.e., sacrifice other values, for the value he's just evoked emotion about. I take this to just be a description of how the essay works, not a pejorative imputation of motives.
I could definitely have done better, and I too went through several drafts, and the one I posted was probably posted because I was tired of editing rather than because it was best. I have removed the sentences in the above that seem most pejorative.
Strong upvote. Thank you for writing this, it articulates the problems better than I had them in my head and enhances my focus. This deserves a longer reply, but I'm not sure if I'll get to write it today, so I'll respond with my initial thoughts.
What I really want from LessWrong is to make my own thinking better, moment to moment. To be embedded in a context that evokes clearer thinking, the way being in a library evokes whispers. To be embedded in a context that anti-evokes all those things my brain keeps trying to do, the way being in a church anti-evokes coarse language.
I want this too.
In the big, important conversations, the ones with big stakes, the ones where emotions run high—
I don’t think LessWrong, as a community, does very well in those conversations at all.
Regarding the three threads you list, I, others involved in managing of LessWrong, and leading community figures who've spoken to me are all dissatisfied with how those conversations went and believe it calls for changes in LessWrong.
Solutions I am planning or considering:
Regarding the three threads you list, I, others involved in managing of LessWrong, and leading community figures who've spoken to me are all dissatisfied with how those conversations went and believe it calls for changes in LessWrong.
I'm deeply surprised by this. If there is a consensus among the LW managers and community figures, could one of them write a post about it laying out what was dissatisfactory and what changes they feel need to be made, or at least the result they want from the changes? I know you're a highly conscientious person with too much on zir hands already, so please don't take this upon yourself.
I am also surprised by this! I think this sentence is kind of true, and am dissatisfied with the threads, but I don't feel like my take is particularly well-summarized with the above language, at least in the context of this post (like, I feel like this sentence implies a particular type of agreement with the OP that I don't think summarizes my current position very well, though I am also not totally confident I disagree with the OP).
I am in favor of experimenting more with some karma stuff, and have been encouraging people to work on that within the Lightcone team. I think there is lots of stuff we could do better, and definitely comparing us to some ideal that I have in my head, I think things definitely aren't going remotely as well as I would like them to, but I do feel like the word "dissatisfied" seems kind of wrong. I think there are individual comments that seem bad, but overall I think the conversations have been quite good, and I am mildly positively surprised by how well they have been going.
While I am not technically a "New User" in the context of the age of my account, I comment very infrequently, and I've never made a forum-level post.
I would rate my own rationality skills and knowledge at slightly above the average person but below the average active LessWrong member. While I am aware that I possess many habits and biases that reduce the quality of my written content, I have the sincere goal of becoming a better rationalist.
There are times when I am unsure whether an argument or claim that seems incorrect is flawed or if it is my reasoning that is flawed. In such cases, it seems intuitive to write a critical comment which explicitly states what I perceive to be faulty about that claim or argument and what thought processes have led to this perception. In the case that these criticisms are valid, then the discussion of the subject is improved and those who read the comment will benefit. If the criticisms are not valid, then I may be corrected by a response that points out where my reasoning went wrong, helping me avoid making such errors in the future.
Amateur rationalists like myself are probably going to make mistakes when it comes to criticism of other people's written content, even when we strive to follow community guidelines. My concern with your suggestions is that these changes may discourage users like me from creating flawed posts and comments that help us grow as rationalists.
I think there's a real danger of that, in practice.
But I've had lots of experience with "my style of moderation/my standards" being actively good for people taking their first steps toward this brand of rationalism; lots of people have explicitly reached out to me to say that e.g. my FB wall allowed them to do just those sorts of first, flawed steps.
A big part of this is "if the standards are more generally held, then there's more room for each individual bend-of-the-rules." I personally can spend more spoons responding positively and cooperatively to [a well-intentioned newcomer who's still figuring out the norms of the garden] if I'm not also feeling like it's pretty important for me to go put out fires elsewhere.
I think that smart people can hack LW norms and propagandize / pointscore / accumulate power with relative ease. I think this post is pretty much an example of that:
- a lot of time is spent gesturing / sermoning about the importance of fighting biases etc. with no particularly informative or novel content (it is after all intended to "remind people of why they care".). I personally find it difficult to engage critically with this kind of high volume and low density.
- ultimately the intent seems to be an effort to coordinate power against types of posters that Duncan doesn't like
I just don't see how most of this post is supposed to help me be more rational. The droning on makes it harder to engage as an adversary, than if the post were just "here are my terrible ideas", but it does so in an arational way.
I bring this up in part because Duncan seems to be advocating that his adherence to LW norms means he can't just propagandize etc.
If you read the OP and do not choose to let your brain project all over it, what you see is, straightforwardly, a mass of claims about how I feel, how I think, what I believe, and what I think should be the case.
...I explicitly underscore that I think
If I'm reading you correctly, it sounds like there's actually multiple disagreements you have here--a disagreement with Duncan, but also a disagreement with the current norms of LW.
My impression is primarily informed by these bits here:
I think that smart people can hack LW norms and propagandize / pointscore / accumulate power with relative ease. [...]
If people here really think you can't propagandize or bad-faith accumulate points/power while adhering to LW norms, well, I think that's bad for rationality.
Could you say more about this? In particular, assuming my reading is accurate, I'm interested in knowing (1) ways in which you think the existing norms are inadequate to the task of preventing bad-faith point-scoring, (2) whether you think it's possible to patch those issues by introducing better norms. (Incidentally, if your answer to (2) is "yes", then it's actually possible your position and Duncan's are less incompatible than it might first appear, since you might just have different solutions in mind for the same problem.)
(Of course, it's also possible you think LW's norms are irreparable, which is a fact that--if true--would still be worth drawing attention to, even if it is somewhat sad. Possibly this is all you were trying to do in the parent comment, in which case I could be reading more into what you said than is there. If that's the case, though, I'd still like to see it confirmed.)
Maybe it is good to clarify: I'm not really convinced that LW norms are particularly conducive to bad faith or psychopathic behavior. Maybe there are some patches to apply. But mostly I am concerned about naivety. LW norms aren't enough to make truth win and bullies / predators lose. If people think they are, that alone is a problem independent of possible improvements.
since you might just have different solutions in mind for the same problem.
I think that Duncan is concerned about prejudicial mobs being too effective and I am concerned about systematically preventing information about abuse from surfacing. To some extent I do just see this as a conflict based on interests -- Duncan is concerned about the threat of being mobbed and advocating tradeoffs accordingly, I'm concerned about being abused / my friends being abused and advocating tradeoffs accordingly. But to me it doesn't seem like LW is particularly afflicted by prejudicial mobs and is nonzero afflicted by abuse.
I don't think Duncan acknowledges the presence of tradeoffs here but IMO there absolutely have to be tradeoffs. To me the generally upvoted and accepted responses to jessicata's post are making a tradeo...
I like this highlighting of the tradeoffs, and have upvoted it. But:
But to me it doesn't seem like LW is particularly afflicted by prejudicial mobs and is nonzero afflicted by abuse.
... I think this is easier to say when one has never been the target of a prejudicial mob on LessWrong, and/or when one agrees with the mob and therefore doesn't think of it as prejudicial.
I've been the target of prejudicial mobbing on LessWrong. Direct experience. And yes, it impacted work and funding and life and friendships outside of the site.
propagandize / pointscore / accumulate power with relative ease
There's a way in which this is correct denotatively, even though the connotation is something I disagree with. Like, I am in fact arguing for increasing a status differential between some behaviors that I think are more appropriate for LW and others that I think are less appropriate. I'm trying at least to be up front about what those behaviors are, so that people can disagree. e.g. if you think that it's actually not a big deal to distinguish between observation and inference, because people already do a good job teasing those apart.
But yes: I wouldn't use the "power" frame, but there's a way in which, in a dance studio, there's "power" to be had in conforming to the activity of dance and dance instruction, and "less power" in the hands of people not doing those things.
an effort to coordinate power against types of posters that Duncan doesn't like
I don't think this is the case; I want to coordinate power against a class of actions. I am agnostic as to who is taking those actions, and even specifically called out that if there are people who are above the local law we should be candid about that ...
This is an essay about the current state of the LessWrong community, and the broader EA/rationalist/longtermist communities that it overlaps and bridges, inspired mostly by the dynamics around these three posts. The concepts and claims laid out in Concentration of Force, which was originally written as part one of this essay, are important context for the thoughts below.
Summary/thesis, mostly cribbed from user anon03's comment below: In many high-importance and high-emotion discussions on LessWrong, the comments and vote distribution seem very soldier-mindset instead of scout-mindset, and the overall soundness and carefulness of reasoning and discourse seems to me to be much lower than baseline, which already felt a smidge too low. This seems to indicate a failure of the LW community further up the chain (i.e. is a result of a problem, not the problem itself) and I think we should put forth real effort to fix it, and I think the most likely target is something like a more-consistent embrace and enforcement of some very basic rationality discourse norms.
I claim that something has gone a little bit wrong.
And as readers of many of my/other/essays/know, I claim that things going a little bit wrong is often actually quite a big problem.
I am not alone in thinking that the small scale matters. Tiny mental flinches, itty bitty little incentives, things thrown ever so slightly off course (and then never brought back). That small things often have outsized or cumulative effects is a popular view, either explicitly stated or discernible as an underlying assumption in the writings of Eliezer Yudkowsky, Nate Soares, Logan Brienne Strohl, Scott Alexander, Anna Salamon, and Andrew Critch, just to name a few.
Yet I nevertheless feel that I encounter resistance of various forms when attempting to point at small things as if they are important. Resistance rather than cooperative disagreement—impatience, dismissal, often condescension or sneering, sometimes projection and strawmanning.
This is absolutely at least in part due to my own clumsiness and confusion. A better version of me, more skilled at communication and empathy and bridging inferential gaps, would undoubtedly run into these problems less. Would better be able to recruit people's general enthusiasm for even rather dull and tedious and unsexy work, on that split-second level.
But it seems to me that I can't locate the problem entirely within myself. That there's something out there that's Actually Broken, and that it fights back, at least a little bit, when I try to point at it and fix it.
Here's to taking another shot at it.
Below is a non-exhaustive list of things which my brain will tend to do, if I don't put forth strategic effort to stop it:
I actually tend not to do these things. I do them fairly rarely, and ever more rarely as time goes on and I improve my cognitive software and add to my catalogue of mental checks and balances and discover more of my brain's loopholes and close them up one by one.
But it's work. It's hard work. It takes up a nontrivial portion of my available spoons every single day.
And more—it requires me to leave value on the table. To not-win fights that I could have won, had I been more willing to kick below the belt. There's a reason human brains slide toward those shortcuts, and it's because those shortcuts tend to work.
But the cost—
My brain does not understand the costs, most of which are distant and abstract. My brain was not evolved to understand, on an intuitive, reflexive level, things like:
All it sees is a chance to win.
My brain is a phoenix. It sees ways to win the immediate, local confrontation, and does not understand what it would be sacrificing to secure that victory.
I spend a lot of time around people who are not as smart as me.
(This is a rude statement, but it's one I'm willing to spend the points to make. It's right that we generally dock people points for rudeness; rudeness tracks a real and important set of things and our heuristics for dealing with it are actually pretty decent. I hereby acknowledge and accept the regrettable cost of my action.)
I spend a lot of time around people who are not as smart as me, and I also spend a lot of time around people who are as smart as me (or smarter), but who are not as conscientious, and I also spend a lot of time around people who are as smart or smarter and as conscientious or conscientiouser, but who do not have my particular pseudo-autistic special interest and have therefore not spent the better part of the past two decades enthusiastically gathering observations and spinning up models of what happens when you collide a bunch of monkey brains under various conditions.
(Repetition is a hell of a drug.)
All of which is to say that I spend a decent chunk of the time being the guy in the room who is most aware of the fuckery swirling around me, and therefore the guy who is most bothered by it. It's like being a native French speaker and dropping in on a high school French class in a South Carolina public school, or like being someone who just learned how to tell good kerning from bad keming. I spend a lot of time wincing, and I spend a lot of time not being able to fix The Thing That's Happening because the inferential gaps are so large that I'd have to lay down an hour's worth of context just to give the other people the capacity to notice that something is going sideways.
(Note: often, what it feels like from the inside when you are incapable of parsing some particular distinction is that the other person has a baffling and nonsensical preference between two things that are essentially indistinguishable. To someone with colorblindness, there's just no difference between those two shades. Sometimes, when you think someone is making a mountain out of a molehill, they are in fact making a mountain out of a molehill. But sometimes there's a mountain there, and it's kind of wild that you can't see it. It's wise to keep this possibility in mind.)
I don't like the fact that my brain undermines my ability to see and think clearly, if I lose focus for a minute.
I don't like the fact that my brain undermines other people's ability to see and think clearly, if I lose focus for a minute.
I don't like the fact that, much of the time, I'm all on my own to maintain focus, and keep my eye on these problems, and notice them and nip them in the bud.
I'd really like it if I were embedded in a supportive ecosystem. If there were clear, immediate, and reliable incentives for doing it right, and clear, immediate, and reliable disincentives for doing it wrong. If there were actual norms (as opposed to nominal ones, norms-in-name-only) that gave me hints and guidance and encouragement. If there were dozens or even hundreds of people around, such that I could be confident that, when I lose focus for a minute, someone else will catch me.
Catch me, and set me straight.
Because I want to be set straight.
Because I actually care about what's real, and what's true, and what's justified, and what's rational, even though my brain is only kinda-sorta halfway on board, and keeps thinking that the right thing to do is Win.
Sometimes, when people catch me, I wince, and sometimes, I get grumpy, because I'm working with a pretty crappy OS, here. But I try to get past the wince as quickly as possible, and I try to say "thank you," and I try to make it clear that I mean it, because honestly, the people that catch me are on my side. They are helping me live up to a value that I hold in my own heart, even though I don't always succeed in embodying it.
I like it when people save me from the mistakes I listed above. I genuinely like it, even if sometimes it takes my brain a moment to catch up.
I've got a handful of metaphors that are trying to triangulate something important.
One of them is "herd immunity." In particular, those nifty side-by-side time lapses that show the progression of virulent illness in populations with different rates of vaccination or immunity. The way that the badness will spread and spread and spread when only half the population is inoculated, but fizzle almost instantly when 90+% is.
If it's safe to assume that most people's brains are throwing up the bad stuff at least as often as mine does, then it seems to matter a lot how infect-able the people around you are. How quickly their immune systems kick in, before the falsehoods take root and replicate and spread.
And speaking of immune systems, another metaphor is "epistemic hygiene." There's a reason that phrase exists. It exists because washing your hands and wearing a mask and using disinfectant and coughing into your elbow makes a difference. Cleaner people get sick less, and propagate sickness less, and cleanliness is made up of a bunch of tiny, pre-emptive actions.
(There was a decent chance that there was going to be someone in the comments using the fact that this essay contains a Rick & Morty quote to delegitimize me and the point that I'm making, but then I wrote this sentence and that became a harder trick to pull off. Not impossible, though.)
Another metaphor is that of a garden.
You know what makes a garden?
Weeding.
Gardens aren't just about the thriving of the desired plants. They're also about the non-thriving of the non-desired plants.
And weeding is hard work, and it's boring, and it's tedious, and it's unsexy.
What I'm getting out of LessWrong these days is readership. It's a great place to come and share my thoughts, and have them be seen by people—smart and perceptive people, for the most part, who will take those thoughts seriously, and supply me with new thoughts in return, many of which I honestly wouldn't have ever come to on my own.
That's valuable.
But it's not what I really want from LessWrong.
What I really want from LessWrong is to make my own thinking better, moment to moment. To be embedded in a context that evokes clearer thinking, the way being in a library evokes whispers. To be embedded in a context that anti-evokes all those things my brain keeps trying to do, the way being in a church anti-evokes coarse language.
I'd like an environment that takes seriously the fact that the little things matter, and that understands that standards and principles that are only enforced 90% of the time aren't actually enforced.
I think LessWrong actually does a pretty good job of toeing the rationality line, and following its own advice, if you take the sum total of all of its conversations.
But if you look at the conversations that matter—the times when a dose of discipline is most sorely needed, and when its absence will do the most damage—
In the big, important conversations, the ones with big stakes, the ones where emotions run high—
I don't think LessWrong, as a community, does very well in those conversations at all. When the going gets tough, the number of people who are steadfastly unwilling to let their brains do the things, and steadfastly insistent that others not get away with it either feels like it dwindles to almost nothing, and as a result, the entirely predictable thing happens: people start using symmetric weapons, and they work.
(I set aside a few minutes to go grab some examples—not an exhaustive search, just a quick skim. There's the total vote count on this comment compared to these two, and the fact that it took nearly three weeks for a comment like this one to appear, and the fact that this is in negative territory, and this comment chain which I discussed in detail in another recent post, and this and its child being positive while this and this hover around zero, and this still not having incorporated the extremely relevant context provided in this, and therefore still being misleading to anyone who doesn't get around to the comments, and the lack of concrete substantiation of the most radioactive parts of this, and so on and so forth.)
To be clear: there are also many examples of the thing going well. If you count up from nothing, and just note all the places where LessWrong handled these conversations better than genpop, there are many! More, even, than what I'm highlighting as the bad stuff.
But gardens aren't just about the thriving of the desired plants. They're also about the non-thriving of the non-desired plants.
There's a difference between "there are many black ravens" and "we've successfully built an environment with no white ravens." There's a difference between "this place substantially rewards black ravens" and "this place does not reward white ravens; it imposes costs upon them." It should be possible—no, it should be easy to have a conversation about whether the incidence of white ravens has been sufficiently reduced, separate from the question of the total incidence of black ravens, and to debate what the ratio of white ravens to black ravens needs to be, and how long a white raven should hang out before being chased away, and what it would cost to do things differently, and whether that's worth it, and I notice that this very sentence is becoming pretty defensive, and is emerging in response to past experiences, and a strong expectation that my attempt at nuance and specificity is likely to fail, because the culture does not sufficiently disincentivize projection and strawmanning and misrepresentation, and so attempts-to-be-clear cannot simply be offhand but must be preemptively fortified and made proof against adversarial interpretation and geez, this kind of sucks, no?
In Concentration of Force, which was originally part one of this essay, I mention the process of evaporative cooling, and I want to ask: who is being evaporatively cooled out of LessWrong these days, and is that the feedback loop we want to set up?
I think it isn't. I think that a certain kind of person—
(one who buys that it's important to stick to the rationality 101 basics even when it's inconvenient, and that even a small percentage of slips along this axis is a pretty big deal)
—is becoming less prevalent on LessWrong, and a certain other kind of person—
(one who doesn't buy the claim that consistency-on-the-small-stuff matters a lot, and/or thinks that there are other higher goals that supersede approximately-never-letting-the-standards-slip)
—is becoming more prevalent, and while I have nothing against the latter in general, I really thought LessWrong was for the former.
Here's my vision of LessWrong:
LessWrong should be a place where rationality has reliable concentration of force.
Where rhetorical trickery does not work. Where supposition does not get mistaken for fact. Where people's words are treated as if they mean what they say, and if there seems to be another layer of implication or inference, that is immediately surfaced and made explicit so the hypothesis can be checked, rather than the assumption run with. Where we are both capable of distinguishing, and careful to distinguish, our interpretations from our observations, and our plausible hypotheses from our justified conclusions. Where we hold each other to that standard, and receive others holding us to that standard as prosocial and cooperative, because we want help holding the line. Where bad commentary is not highly upvoted just because our monkey brains are cheering, and good commentary is not downvoted or ignored just because our monkey brains boo or are bored.
Perhaps most importantly, where none of the above is left on the level of "c'mon, we all know." Where bad stuff doesn't go unmentioned because it's just assumed that everyone knows it's bad. That just results in newcomers not knowing the deal, and ultimately means the standards erode over time.
(A standard people are hesitant or embarrassed or tentative about supporting, or that isn't seen as cool or sophisticated to underline, is not one that endures for very long.)
I spend a decent chunk of my time doing stuff like upvoting comments that are mostly good, but noting in reply to them specific places in which I think they were bad or confused or norm-violating. I do this so that I don't accidentally create a social motte-and-bailey, and erode the median user's ability to tell good from bad.
This is effortful work. I wish more people pitched in, more of the time, the way this user did here and here and here and here.
In my opinion, the archetype of the Most Dangerous Comment is something like this one:
This is a bad comment (in context, given what it's replying to). It's the kind of thing my brain produces, when I lose focus for a minute.
But it sounds good. It makes you Feel Like You're On The Right Team as you read it, so long as you're willing to overlook the textbook strawmanning it does, of the comment it's replying to.
It's a Trojan horse. It's just such good-thoughts-wrapped-in-bad-forms that people give a pass to, which has the net effect of normalizing bad forms.
It's when the people we agree with are doing it wrong that we are most in need of standards, firmly held.
(I have a few theories about why people are abandoning or dismissing or undermining the standards, in each of a few categories. Some people, I think, believe that it's okay to take up banned weapons as long as the person you're striking at is in the outgroup. Some people seem to think that suffering provides a justification for otherwise unacceptable behavior. Some people seem to think that you can skip steps as long as you're obviously a good guy, and others seem to think that nuance and detail are themselves signs of some kind of anti-epistemic persuadery. These hypotheses do not exhaust the space of possibility.)
It is an almost trivial claim that there are not enough reasonable people in the world. There literally never will be, from the position of a group that's pushing for sanity—if the quality of thought and discourse in the general population suddenly rose to match the best of LessWrong, the best of the LessWrongers would immediately set their sights on the next high-water mark, because this sure ain't enough.
What that means is that, out there in the broader society, rationality will approximately always lose the local confrontations. Battles must be chosen with great care, and the forces of reason meticulously prepared—there will be occasional moments of serendipity when things go well, and the rare hero that successfully speaks a word of sanity and escapes unscathed, but for the most part those victories won't come by accident.
Here, though—here, within the walls of the garden—
A part of me wants to ask "what's the garden for, if not that? What precisely are the walls trying to keep out?"
In my post on moderating LessWrong, I set forth the following principle:
I no longer think that's sufficient. There aren't enough moderators for reliable concentration of force. I think it's the case that LessWrongers trying to adhere to the site's principles are often alone—and furthermore, they have no real reason, given the current state of affairs, to expect not to be alone.
Sometimes, people show up. Often, if you look on a timescale of days or weeks. But not always. Not quickly. Not reliably.
(And damage done in the meantime is rarely fully repaired. If someone has broadcast falsehoods for a week and they've been strongly upvoted, it's not enough to just say "Oops, sorry, I was wrong." That comes nowhere close to fixing what they broke.)
Looking at the commentary in the threads of the last month—looking at the upvotes and the downvotes—looking at what was said, and by whom, and when—
It's not promising.
It's not promising in the sense that the people in the parking lot consider it their responsibility to stand in defense of the parent who left their kids in the car.
That they do so reliably, and enthusiastically. That they show up in force. That they show up in force because they expect to be backed up, the way that people in a city expect to be backed up if they push back against someone shouting racist slurs. That they consider themselves obligated to push back, rather than considering it not-their-problem.
It is a defining characteristic of stag hunts that when everybody actually buys in, the payoff is pretty huge.
It is also a defining characteristic of stag hunts that when critical mass fails to cohere, those who chose stag get burned, and feel cheated, and lose big.
This post is nowhere close to being a sufficient coordination mechanism to cohere a stag hunt. No one should change their behavior in response to this alone.
But it's a call for a stag hunt.
The elephant in the room, which deserves its own full section but I wasn't able to pull it together:
Standards are not really popular. Most people don't like them. Or rather, most people like them in the abstract, but chafe when they get in the way, and it's pretty rare for someone to not think that their personal exception to the standard is more justified than others' violations. Half the people here, I think, don't even see the problem that I'm trying to point at. Or they see it, but they don't see it as a problem.
I think a good chunk of LW's current membership would leave or go quiet if we actually succeeded at ratcheting the standards up.
I don't think that's a bad thing. I'd like to be surrounded by people who are actually trying. And if LW isn't going to be that place, and it knows that it isn't, I'd like to know that, so I can go off and found it (or just give up).
Terrible Ideas
... because I don't have better ones, yet.
The target is "rationality has reliable concentration of force."
The current assessment is "rationality does not have reliable concentration of force."
The vector, then, is things which either increase the number of people showing up in true rationalist style, relative to those who are not, or things which increase the power of people adhering to rationalist norms, relative to those who are not.
More of the good thing, and/or less of the bad thing.
Here are some terrible ideas for moving in that direction—for either increasing or empowering the people who are interested in the idea of a rationality subculture, and decreasing or depowering those who are just here to cargo cult a little.
These are all terrible ideas.
These are all
terrible
ideas.
I'm going to say it a third time, because LessWrong is not yet a place where I can rely on my reputation for saying what I actually mean and then expect to be treated as if I meant the thing that I actually said: I recognize that these are terrible ideas.
But you have to start somewhere, if you're going to get anywhere, and I would like LessWrong to get somewhere other than where it was over the past month. To be the sort of place where doing the depressingly usual human thing doesn't pay off. Where it's more costly to do it wrong than to do it right.
Clearly, it's not going to "just happen." Clearly, we need something to riff off of.
The guiding light in front of all of those terrible ideas—the thing that each of them is a clumsy and doomed attempt to reach for—is making the thing that makes LessWrong different be "LessWrong is a place where rationality has reliable concentration of force."
Where rationality is the-thing-that-has-local-superiority-in-most-conflicts. Where the people wielding good discourse norms and good reasoning norms always outnumber the people who aren't—or, if they can't outnumber them, we at least equip them well enough that they always outgun them.
Not some crazy high-tower thing. Just the basics, consistently done.
Until users doing the above (and similar) consistently win against users who aren't, LessWrong is going to miss out on a thing that clearly a lot of us kind of want, and kind of think might actually matter to some other pretty important goals.
Maybe that's fine. Maybe all we really need is the low-effort rough draft. Maybe the 80/20 is actually the right balance. In fact, we're honestly well past the 80/20—LessWrong is at least an 85/37 by this point.
But it's not actually doing the thing, and as far as I can tell it's not really trying to do the thing, either—not on the level of "approximately every individual feels called to put forth a little extra effort, and approximately every individual feels some personal stake when they see the standards being degraded."
Instead, as a collective, we've got one foot on the gas and the other on the brake, and that probably isn't the best strategy for any worthwhile goal. One way or another, I think we should actually make up our minds, here, and either go out and hunt stag, or split up and catch rabbits.
Author's note: this essay is not as good as I wished it would be. In particular, it's falling somewhat short of the very standard it's pulling for, in a way that I silver-line as "reaching downward across the inferential gap" but which is actually just the result of me not having the spoons to do this or this kind of analysis on each of a dozen different examples. Having spent the past six days improving it, "as good as I wish it would be" is starting to look like an asymptote, so I chose now over never.