1256

LESSWRONG
LW

1255
HumorTerminology / Jargon (meta)Community
Frontpage

55

Shit Rationalists Say?

by Eneasz
25th Jan 2012
1 min read
119

55

55

Shit Rationalists Say?
93Alicorn
75Alicorn
69Alicorn
56Alicorn
14arundelo
30Viliam_Bur
67Alejandro1
47TheOtherDave
24fubarobfusco
13Multiheaded
7thomblake
3Multiheaded
7fubarobfusco
21Randaly
11Cthulhoo
1DanielH
1AlexanderRM
8Multiheaded
20thomblake
13Eliezer Yudkowsky
5thomblake
4wedrifid
-1loup-vaillant
8dbaupp
-1[anonymous]
1Bound_up
13Raemon
3gwern
9arundelo
9CronoDAS
3Alicorn
7Karmakaiser
13Karmakaiser
5curiousepic
3Alicorn
4Raemon
4curiousepic
1TimS
8steven0461
4[anonymous]
0wedrifid
4Zack_M_Davis
0wedrifid
64Raemon
30Raemon
58lessdazed
33Karmakaiser
20Alicorn
11lessdazed
7komponisto
5lessdazed
1[anonymous]
1XFrequentist
1Psychosmurf
55daenerys
7Nornagest
5Raemon
49RolfAndreassen
40Jack
37Solvent
2Armok_GoB
39arundelo
38RomeoStevens
2MugaSofer
0RomeoStevens
2LoganStrohl
0RomeoStevens
0Decius
-1MugaSofer
38fubarobfusco
33Raemon
33Raemon
7fubarobfusco
16Raemon
16Eneasz
1FiftyTwo
2MBlume
32mindspillage
6Joshua Hobbes
22Xachariah
3Joshua Hobbes
10Shmi
30NancyLebovitz
37TheOtherDave
29Anubhav
9MixedNuts
1Anubhav
1A1987dM
28MBlume
26Normal_Anomaly
2[anonymous]
1Multiheaded
26Anubhav
26[anonymous]
20EchoingHorror
20Curiouskid
19Prismattic
8NancyLebovitz
1DavidAgain
17Karmakaiser
11pedanterrific
2Karmakaiser
16Alejandro1
14FiftyTwo
8D_Alex
13CharlesR
7Raemon
24MBlume
5CharlesR
13Raemon
10maia
8Rubix
8thomblake
8Normal_Anomaly
6Solvent
10pedanterrific
2Karmakaiser
-1[anonymous]
-1[anonymous]
New Comment
119 comments, sorted by
top scoring
Click to highlight new comments since: Today at 5:16 PM
Some comments are truncated due to high volume. (⌘F to expand all)Change truncation settings
[-]Alicorn14y930

The "confidence interval" line should have a percentage ("What's your 95% confidence interval?").

"You make a compelling case for infanticide."

"Can you link me to that study?"

"I think I'm going to solve psychology." ("I think I'm going to solve metaethics." "I think I'm going to solve Friendliness.")

"My elephant wants a brownie."

"Is that your true rejection?"

"I wanna be an upload!"

"Does that beat Reedspacer's Lower Bound?"

"Let's not throw all our money at the Society for Rare Diseases in Cute Puppies."

"I have akrasia."

"I'm cryocrastinating."

"Do that and you'll wind up with the universe tiled in paperclips."

"So after we take over the world..."

"I want to optimize for fungibility here."

"This looks like a collective action problem."

"We can dissolve this question." ("That's a dissolved question.")

"My model of you likes this."

"Have you read Goedel, Escher, Bach?"

"What do the statistics say about cases in this reference class?"

Reply
[-]Alicorn14y750

"We need whiteboards."

"I'm trying paleo."

"I might write rationalist fanfiction of that."

"That's just an applause light." ("That's just a semantic stopsign." "That's just the teacher's password.")

"POLITICS IS THE MINDKILLER"

"If keeping my current job has higher expected utility than founding a startup, I wish to believe that keeping my current job has higher expected utility than founding a startup..."

"I think he's just being metacontrarian."

"Arguments are soldiers!"

"Not every change is an improvement, but every improvement is a change."

"There are no ontologically basic mental entities!"

"I'm an aspiring rationalist."

"Fun Theory!"

"The map is not the territory."

"Let's beware evaporative cooling, here."

"It's a sunk cost! Abandon it!"

"ERROR: POSTULATION OF GROUP SELECTION DETECTED"

"If you measure it and reward the measurement going up, you'll get what you measure, not what you want."

"Azahoth!"

"Death is bad."

Reply
[-]Alicorn14y690

This is too much fuuuuuuuun

"She's just signaling virtue."

"Money is the unit of caring."

"One-box!"

"Beliefs should constrain anticipations."

"Existential risk..."

"I'll cooperate if and only if the other person will cooperate if and only if I cooperate."

"I'm going to update on that."

"Tsuyoku naritai!"

"My utility function includes a term for the fulfillment of your utility function."

"Yeah, it's objective, but it's subjectively objective."

"I am a thousand shards of desire."

"Whoa, there's an inferential gap here that one of us is failing to bridge."

"My coherent extrapolated volition says..."

"Humans aren't agents." ("I'm trying to be more agenty." "Humans don't really have goals.")

"Wait, wait, this is turning into an argument about definitions."

"Look, just rejecting religion and astrology doesn't make someone rational."

"No, no, you shouldn't implement Really Extreme Altruism. Unless the alternative is doing it without, anyway..."

"I'll be the Gatekeeper, you be the AI."

"That's Near, this is Far."

"Don't fall into bottom-line thinking like that."

Reply
[-]Alicorn14y560

I think I'm done. If I think of any more I'll add them to this comment instead of making a new one.

"How do you operationalize that?"

"'Snow is white' is true if and only if snow is white."

"If I may generalize from one example here..."

"I'm suffering from halo effect."

"Warning: Dark Arts."

"Okay, but in the Least Convenient Possible World..."

"We want to raise the sanity waterline."

"You've fallen prey to the illusion of transparency."

"Bought some warm fuzzies today."

"What does the outside view say?"

"So the idea is that we make all scientific knowledge a sacred and closely guarded secret, so it will be treated with the reverence it deserves!"

"How could you test that belief?"

Reply
[-]arundelo14y140

RATIONALISTS SAY ALL THE THINGS!

Reply
[-]Viliam_Bur14y300

RATIONALISTS SAY ALL THE THINGS!

Solomonoff prior gives you 50%, that's pretty cool! :D

I hope someone will use Alicorn's (and other) quotes to make a good Eliza-bot. This could be an interesting AI challenge -- write a bot that will get positive karma on LW! If there are more bots, the bot with highest nonzero karma wins.

Reply
[-]Alejandro114y670

As a start, I copied all Alicorn's lines into a Markov text synthesizer . Some of the best results were:

"Whoa, there's an improvement, but it's subjectively objective."

"Okay, but every change is turning into bottom-line thinking like a collective action problem."

"If keeping my current job has higher expected utility than founding a brownie."

"I think he's just the unit of you shouldn't implement Really Extreme Altruism. Unless the teacher's password."

"If I wish to optimize for Rare Diseases in paperclips."

"My elephant wants a term for infanticide."

Reply
[-]TheOtherDave14y470

I burst out laughing while reading this, so of course my officemates wanted to know what was so funny.

I cannot remember the last time the gulf of inferential distances was so very very wide.

Reply
[-]fubarobfusco14y240

Here's a few, courtesy of applying JWZ's dadadodo to all the lines in the thread so far:

What does the best textbook on corrupted hardware. Dark Arts; Escher, Bach?

How could you credibly pre commit to see you as a compartmentalized belief?

I'm trying to be a cult.

Have super powers.

You've fallen prey to be condescending.

My current job has higher expected utility than you imagine; but in the sanity waterline.

Everyone is Far.

No idea is reliable? Have a lot of caring.

Conceptspace is the future.

Mysteriousness is a cult.

I'm going to be with the AI. I know the universe future.

Look, just generalize from the territory.

Everyone is bigger than you in Cute Puppies.

Emacs' M-x dissociated-press yields babble, but with some interesting words in it: "knowledgeneralize", "metacontrammer", "contrationalist", "choosequences", "the universal priory", "statististimate", "fanfused", "condescendista", "frobability", "dissolomonoff", "optimprovement", "estimagine", "cooperaterline", "pattern matchology". The only sensible sentence it's come up with is "I'm running on condescending".

Reply
[-]Multiheaded14y130

No idea is reliable? Have a lot of caring.

I visualized that being said simultaneously with the middle-finger gesture.

Reply
7thomblake14y
I seem to remember someone's already made a Bayesian Priory pun, but if not then it should happen prominently. EDIT: here
3Multiheaded14y
Wrong. Electronic old men are.
7fubarobfusco14y
To give an idea of what these look like raw, here's a paragraph of dadadodo: And here's a similar-sized chunk of M-x dissociated-press: Of these, I rather like: The blended-words effect seems to give M-x dissociated-press a sort of Finnegans Wake atmosphere which dadadodo doesn't have.
[-]Randaly14y210

From another generator:

"I'm going to solve metaethics." "I'm going, you're going to found the Society for infanticide."

""Snow is white" is failing to solve psychology."

"Wait, wait, "this is white" is a more technical explanation?"

"My utility function includes a semantic stopsign."

"If keeping my current job has little XML tags on it that say the Least Convenient Possible World...""

"Sure, I'd take over the sanity waterline."

"I'll be the symbol with ice cream trees."

"So after we take over the alternative universe that is the Least Convenient Possible World..."

"I want to tile the sanity waterline with the unit of a thing."

Reply
[-]Cthulhoo14y110

Not totally IT, but I tried it on Eliezer's "The 5-Second Level". Highlits include:

I won't socially kill you

Hope to reflect on consequentialist grounds

Say, what a vanilla ice cream, and not-indignation, and from green?

Associate to persuade anyone of how you were making the dreadful personal habit displays itself in a concrete example.

Rather you can't bear the 5-second level?

To develop methods of teaching rationality skills, you need more practice to get lost in verbal mazes; we will tend to have our feet on the other person.

Be sufficiently averse to the fire department and see if that suggests anything.

Reply
1DanielH12y
I do believe it suggests libertarianism. But I can't be sure, as I can't simply "be sufficiently averse" any more than I can force myself to believe something. Still, that one seems to be a fairly reasonable sentence. If I were to learn only that one of these had been used in an LW article (by coincidence, not by a direct causal link), I would guess it was either that one or "I won't socially kill you".
1AlexanderRM10y
I would be amazed if Scott Alexander has not used "I won't socially kill you" at some point. Certainly he's used some phrase along the line of "people who won't socially kill me". ...and in fact, I checked and the original article has basically the meaning I would have expected: "knowing that even if you make a mistake, it won't socially kill you.". That particular phrase was pretty much lifted, just with the object changed.
8Multiheaded14y
If we had signatures on LW, this would be mine.
[-]thomblake14y200

I hope someone will use Alicorn's (and other) quotes to make a good Eliza-bot.

Surely you mean Eliezer-bot.

Reply
[-]Eliezer Yudkowsky13y130

Should it be made, it will of course be known as Elieza.

But in any case I think you need to keep in mind that a blank map does not correspond to a blank territory.

Reply
5thomblake13y
I initially read the parent in a straightforward way, but then I noticed it is also a meta-joke.
4wedrifid13y
Usually. It could.
-1loup-vaillant13y
What is your prior? (For Eliezer being empty.)
8dbaupp14y
Hopefully they'd keep improving.
-1[anonymous]13y
1Bound_up8y
"My utility function includes a term for the fulfillment of your utility function." Awww.... :)
[-]Raemon14y130

Imagining this in my head has sold me on this being a good idea. Or at least a mildly amusing idea that will have relatively minor negative externalities. (I'm reminded of Eliezer Facts)

Reply
3gwern14y
I'm reminded of http://lesswrong.com/lw/21a/free_copy_of_feynmans_autobiography_for_best/
9arundelo13y
Gabe Newell (of Valve Software) wrote the following in an email to Yanis Varoufakis (an economist): Edit: And that reminds me of the Reamde character Richard Forthrast giving Zula Forthrast a job at his video game company because of her geology expertise.
9CronoDAS14y
I like this one. Mind if I actually use it?
3Alicorn14y
Go for it. I say it to recommend things to people. (Mostly one person.)
7Karmakaiser14y
I'm still laughing when I think about this one.
[-]Karmakaiser13y130

Update: Still laughing and using it in conversations.

Reply
5curiousepic14y
Sad as it is, this has potential to be effective outreach to Reddit, et al. Unless you'd like to do it yourself, or someone gives a good objection within a few days, I'll be posting it in one or more subreddits, perphaps including the GEB readthrough I'm participating in.
3Alicorn14y
I don't use Reddit. If there's interest in turning this into a video, I'm willing to film myself speaking some of my lines, but fear composing an entire video (ideally with several speakers) would take video editing skills and resources I don't have.
4Raemon14y
I actually want to film this, except I still think it has at least a 25% chance of turning out to be a horrible idea.
4curiousepic14y
I came into this thread with a negative set point because I see the "Shit X says" meme as thoroughly without value, being merely collections of stereotypes for no other purpose than to collect them. The OP confirmed this, and because my comment sorting had currently been on New, I scrolled through some of the comments, almost all of which continued to confirm this. Then I resorted to Top and saw your post, and my mind was immediately tickled. Some of these are genuinely funny, and in fact has value as a collection of LW memes and short rationality quotes.
1TimS14y
Could someone explain this reference?
8steven046114y
It's a metaphor used in Jonathan Haidt's book The Happiness Hypothesis: the rider is the conscious or deliberative mind and the elephant is everything underneath.
4[anonymous]14y
More to the point, the analogy is used in one of Luke's posts.
0wedrifid13y
Is there an original source for this one?
4Zack_M_Davis13y
Context.
0wedrifid13y
Thanks Zack. I had a feeling I'd seen it before but couldn't recall the details.
[-]Raemon14y640

"You should read the Sequences."
"It's not like Gandhi has little XML tags on that say 'moral' "
"It's not like natural selection put little XML tags there that say 'purpose'."
"Sure, I'd take a pill that made me bisexual."
"There's this really great fanfiction you should read."
"Oh man I wanna tile the universe with ice cream trees."
"I don't want to tile the universe with bananas and palm trees."

[Sequence that will be incredibly funny in context but is a terrible idea please for goodness' sake (literally) nobody film it]

"No, we're not a cult."
"We're not a cult or anything."
"It's not like organizations have little XML tags that say 'cult' and 'not cult'"
"Any organization with a good cause can have cult attractors"
"Look I'd explain it but there's a lot of inferential distance and I don't want to be condescending."
"We are not a cult. We are not a cult."

Reply
[-]Raemon14y300

"You have no idea how BIG mindspace is."
"You have no idea how BIG optimization process space is."
"You have no idea how BIG thingspace is."

Reply
[-]lessdazed14y580

"Did you just generalize from fictional evidence?"

"You're a one-boxer, right?" (Said with no context.)

"You'd choose specks, right?" (Said with no context.)

"Mysteriousness is not a property of a thing."

"You're running on corrupted hardware."

"Replace the symbol with the substance."

"Could you regenerate that knowledge?"

"Consider a group you feel prejudiced against, frequentists for example."

"But what's the best textbook on that subject?"

"Is that a compartmentalized belief?

"I notice I am confused."

"Of course I have super-powers. Everyone does."

"Beliefs are properly probabilistic."

"Is that your confidence level inside or outside the argument?"

"Did you credibly pre-commit to that rule?"

"That's just what it feels like from the inside."

"Conceptspace is bigger than you imagine."

"No you don't believe you believe that."

"No, money is the unit of caring."

"If that doesn't work out for you, you can still make six figures as a programmer."

"Purpose is not an inherent property."

"You think intr... (read more)

Reply
[-]Karmakaiser14y330

"Yeah, I read all the papers cited in lukeprog's latest article."

A bunch of links almost no one clicks. It's like the Anti-TVTropes.

Reply
[-]Alicorn14y200

"I notice I am confused."

I cannot believe I missed this one.

Reply
[-]lessdazed14y110

"We played reference class tennis."

"Those are just more available to you, not actually more likely."

"Are you more an aspiring rationalist, 'aspiring rationalist,' 'aspiring' rationalist, or aspiring 'rationalist'?"

"The invisible is implied here."

"Is that a disjunctive or conjunctive event?"

"It seemed hard until I hacked away at the edges."

"You didn't time yourself thinking about it before proposing solutions?"

"I have something to protect."

"Someone should type a transcript of that."

"I don't know if that's still an open problem, I've been following the HPMOR thread instead of that one." (Said to a Philosophy professor about a philosophical problem.)

"Is there a more technical explanation?

"Argument screens off authority."

"Go ahead and try to 'other optimize' me."

"That's one of my ugh fields."

"That's not a property, it's a dangling variable."

"ADBOC."

Shit and Bullshit Rationalists Don't Say:

"Gwern hasn't summarized any research on that."

"Did Yvain even edit that before posting?"

"What are his/her credentials?&qu... (read more)

Reply
7komponisto14y
Wrong list.
5lessdazed14y
Brain fart. This is relevant.
1[anonymous]13y
To be fair, this is true if you interpret "absence of evidence" as meaning "absence of evidence in either direction".
1XFrequentist14y
The last section is amazing!
1Psychosmurf14y
Oh man, had me laughing for a good while with this one. Nice job! ^_^
[-]daenerys14y550

If a "Shit Rationalists Say" thread would result in net positive utility, I want to believe that a "Shit Rationalists Say" thread would result in net positive utility.

If a "Shit Rationalists Say" thread would not result in net positive utility, I want to believe that a "Shit Rationalists Say" thread would not result in net positive utility.

Let me not become attached to beliefs I may not want.

Reply
7Nornagest14y
Well played.
5Raemon14y
This sums up my thinking.
[-]RolfAndreassen14y490

"Suppose we were all playing Prisoner's Dilemma with clones of ourselves..."

I heard this said at the Ohio meetup on Sunday; Yvain commented that, of all the meetups he'd been to, ours took the longest to reach that point.

Reply
[-]Jack14y400

"Oops."

Reply
[-]Solvent14y370

That's optimistic.

Reply
2Armok_GoB14y
moon explodes "Oops"
[-]arundelo14y390

Here's one I say a lot: "Everyone is too young to die."

Reply
[-]RomeoStevens14y380

I've thought a "shit your brain says" might be a good way of compactly presenting some cognitive biases.

Reply
2MugaSofer13y
Did anything come of this excellent idea?
0RomeoStevens13y
No, didn't realize this got so upvoted. The fad is a bit past. I still want cartoons on the level of RSA Animate for the sequences.
2LoganStrohl11y
I think this should still happen.
0RomeoStevens11y
RSA animate sequences or "shit your brain says"? Shit your brain says could probably be pulled off fairly easily in a couple days. I should talk to Alton about it. The animations are a longer term project that are still on my list of projects to fund/work on if MealSquares gets bigger.
0Decius11y
Concur.
-1MugaSofer13y
Shame. It's a good idea.
[-]fubarobfusco14y380

"Foo is not about bar."
"What odds are you offering on that?"
"Taboo 'monkey'."
"If you're not getting bad result, you're spending too many resources on avoiding it."
"Rationalists should win."
"Sorry, I'm running on corrupted hardware."

Reply
[-]Raemon14y330

Shit Rationalists Say:

'The thread should really be called "Shit LWers Say.'

Reply
[-]Raemon14y330

"So you know the My Little Pony episode about bayesian updating?"

Reply
7fubarobfusco14y
"'Feeling Pinkie Keen' comes so close!"
[-]Raemon14y160

I got into an argument recently over whether that episode was good or terrible. (I believe it is a good episode, specifically because of the broader context of the show. Fellow rationalist in question watched that episode FIRST, with no context, which is just about the worst place to do so)

Regardless, I have a heuristic about how much I am allowed to argue about My Little Pony before I should feel bad about myself, and... I felt a little bad that day.

(That day was today).

Reply
[-]Eneasz14y160

I did really hate the whole "Sometimes you just have to believe!" message. It felt almost like someone had thought MLP was becoming TOO rational and wanted to throw a bone to the Believer parents. :/

Reply
1FiftyTwo12y
Alternative interpretation is believing something you have strong evidence of that contradicts your pre-existing theories.
2MBlume14y
My brother saw the first of the cutie mark episodes first, he was unimpressed.
[-]mindspillage14y320

One huge category of utterances remains unrepresented:

"Ooh, is there a new Methods of Rationality chapter up yet?"

"I can't believe there's no new chapter yet."

"Have you read Methods of Rationality? You have to read it, OMG."

Reply
6Joshua Hobbes14y
Now that Eliezer is writing at a decent pace again I'm just desperate for progress updates.
[-]Xachariah14y220

Your response just prompted me to check. I wonder if this is how the pigeons in Skinner's Box felt.

Reply
3Joshua Hobbes14y
The worst part is that unlike with actual updates we might get them at any time. So I check multiple times per day just to see if he's going to tell us how many words he wrote last night.
[-]Shmi14y100

No one said that in a long time.

Reply
[-]NancyLebovitz14y300

Cached thoughts are the mindkiller.

Reply
[-]TheOtherDave14y370

I always think that.

Reply
[-]Anubhav14y290

Took every quote posted in this thread so far and Markov-ised it. (Depth of 1, i.e., each word only depends on the word before it. A depth of 2 basically regurgitates the input in this case.)

Here are some of the results:

"How do you regenerate that beat Reedspacer's Lower Bound?"

"It's a more karma. I interpreted your map, not a group you feel prejudiced against, frequentists for infanticide."

"Look, just generalize from halo effect."

"Okay, but in this is rational under them to make six figures as a cult."

"Beliefs are just an inferential distance and my belief is the territory."

"That's just a survey should do you measure, not throw all the scientific knowledge a dissolved question."

"Any organization with the teacher's password."

"That's just rejecting religion and closely guarded secret, so it but there's no idea how BIG thingspace is."

"Foo is the karma system more likely."

"You make six figures as a thousand shards of inferential distance and I cooperate."

"Conceptspace is white."

"Go ahead and only if and it's objective, but I think he's just what the Nazis did!"

"... (read more)

Reply
9MixedNuts14y
Frequentists for infanticide! (I hope nobody takes this seriously. I mean, ew. Who'd support such horrible statistics?)
1Anubhav14y
Most of those quotes aren't even in the post anymore. :( (Cut it down drastically because the wall of text was too high.)
1A1987dM14y
Awesome.
[-]MBlume14y280

"I know, it's pasta, and it's terrible for me, but at least I poured butter all over it."

Reply
[-]Normal_Anomaly14y260

"How do you know that?"

"Have I interpreted your position correctly?"

"That's uncomputable."

"That AI would destroy the world."

"That's a fact about your map, not about the territory."

"That's a duplicate quote."

"What does this have to with rationality?"

"That word doesn't carve reality at its joints."

"I want to do an AI Box experiment."

"Apply the reversal test! You're suggesting we kill all the old people."

"I wish I could self-modify to . . ."

"I'd say something politically incorrect about race, but it would start a flame war."

"I'd say something politically incorrect about gender, but it would start a flame war."

"I'd say something politically incorrect about my ability to say politically incorrect things, but it would start a flame war."

"Karma is being used to suppress dissent."

"Your position is incoherent unless you're a vegetarian."

"Rationality isn't that great."

"We should do a survey on this."

"That survey should have been bigger. We should do another survey."

"Way too much of the scientific lit... (read more)

Reply
2[anonymous]14y
Some of us read Moldbug and Foseti so I'm not too sure... ;)
1Multiheaded13y
(For the record, following your second link is what originally gave me the intuition that "fascist technocracy" might be a real problem, and that it's worth investigating. Before, I was mostly like: "Oh, nothing to worry about, it's just Mencius and he says creepy shit, but he's nice enough really.")
[-]Anubhav14y260

The thread should really be called "Shit LWers Say."

We're not the only group of people calling ourselves "Rationalists", nor are we the most well-known of these groups (not by a long shot).

Reply
[-][anonymous]14y260

I'm guilty of all of these:

"Cached thought!"

"That's a wrong question."

"Have you read the Sequences?"

"According to Solomonoff Induction/the universal prior..."

"Stop prepending "rational" to post titles!"

"I know the forbidden idea, and it's not that bad."

"That's just a status/signalling game."

"There's a signalling hypothesis by Robin Hanson..."

"TDT/UDT says..."

Reply
[-]EchoingHorror14y200

"I want to get my microexpressions analyzed so I can know what I'm thinking."

Reply
[-]Curiouskid14y200

Consider this a study guide for newbies who want to measure how much of LW they understand.

Reply
[-]Prismattic14y190

Post title should be "Shit LWers say". Not all Bayesians sound like regulars on this website.

Reply
8NancyLebovitz14y
What's your prior that a Bayesian is also an LWer?
1DavidAgain14y
And not all rationalists are Bayesians, come to that.
[-]Karmakaiser14y170

Rationalist PUA:

"What's you number? [...] No not that. Your Erdos number."

Rationalist insults:

"...Frequentist."

While twirling a paperclip: "I do not love you, nor do I hate you." (well, more of a threat.)

"I bet (73%) that you're a really consistent person. The sort of person whose decisions are final. Like say, in Monty Hall."

"Name three.."

Reply
[-]pedanterrific14y110

While twirling a paperclip: "I do not love you, nor do I hate you."

More of a threat, surely?

Reply
2Karmakaiser14y
I could go into a long semantic argument putting threat as some subclass in insult but you're right. Oops.
[-]Alejandro114y160

Some line in the video (maybe the "corrupted hardware" one) should be said by someone with a desk full of piles of papers (connoting unfinished, urgent work) on topics like FAI, and a computer screen with open windows on TvTropes, Fanfiction.net, and a LW post on akraisa.

Reply
[-]FiftyTwo14y140

"Source?"

Reply
8D_Alex14y
Nah, that's 4chan.
[-]CharlesR14y130

"What Shock Level are you?"

Reply
7Raemon14y
This is the first one that stumps me.
[-]MBlume14y240

That's because it's old -- more of a Shit SL4ers say

Reply
5CharlesR14y
The question came up at the West LA LW Meetup. Only two people knew what it meant.
[-]Raemon14y130

I think I like this conceptually (I would get warm fuzzies if done right), except I'm trying to detach (in my own mind as well as others), the word "rationalist" from "the Less Wrong memeplex." Which is probably a pipe dream, and such a video would not contribute (much) to the issue one way or another, and "Shit Less Wrong folks say" doesn't have the same ring to it.

Reply
[-]maia13y100

"You're going to [insert life plan here]? Why don't you just go work on Wall Street and donate your money to AMF?"

Reply
[-]Rubix14y80

"This, modulo that."

"It's not obvious that..."

"...is the obvious low-hanging fruit."

"HPJEV is my fantasy jerkass boyfriend."

"Eat this rock salt."

"Eat this potassium salt."

"Have you played Mage: The Ascention?"

"Can you be a CEV machine and order food for me?"

"Scumbag aliefs!"

"Scumbag subconscious!"

"Scumbag physics!"

"I want to want to do that."

Reply
[-]thomblake14y80

Wow, most of that looks really weird, and putting it all together like that seems to associate 'rationalist' with 'weird object-level beliefs'.

Reply
8Normal_Anomaly13y
There's a selection bias. We also say plenty of normal things, but they aren't unique to us, so they don't get put here.
[-]Solvent14y60

From RationalWiki:

How to spot a LessWrongian in your sceptical discussion

In the outside world, the ugly manifests itself as LessWrong acolytes, minds freshly blown, metastasising to other sites, bringing the Good News for Modern Rationalists, without clearing their local jargon cache. Sceptical discussion spaces will often have people show up labelling themselves "rationalist" and being as irritating as teenage nerds who've just discovered Ayn Rand. Take one drink for each of:

"As rationalists, we should ..."

"As a rationalist, you

... (read more)
Reply
[-]pedanterrific14y100

minds freshly blown

Band name alert.

Reply
[-]Karmakaiser14y20

"Rational Approach..."

Reply
[-][anonymous]13y-10

"Stop dissecting my hypothetical..."

"Do you operate under Crockers rules?"

"That is a question about quality of life, not about death badness!"

[This comment is no longer endorsed by its author]Reply
[-][anonymous]14y-10

"It's not obvious that..."

"...is the obvious low-hanging fruit."

"HPJEV is my fantasy jerkass boyfriend."

"This, modulo that."

[EDIT: Not actually Michael Vassar.]

[This comment is no longer endorsed by its author]Reply
Moderation Log
More from Eneasz
View more
Curated and popular this week
119Comments
HumorTerminology / Jargon (meta)Community
Frontpage
Mentioned in
100Random LW-parodying Statement Generator
82Son of Shit Rationalists Say
42Stop talking about p(doom)
41Cards Against Rationality
16Shit rationalists say - 2018

I assume everyone has run across at least one of the "Shit X's Say" format of videos? Such as Shit Skeptics Say. When done right it totally triggers the in-group warm-fuzzies. (Not to be confused with the nearly-identically formatted "Shit X's Say to Y's" which is mainly a way for Y's to complain about X's).

What sort of things do Rationalists often say that triggers this sort of in-group recognition which could be popped into a short video? A few I can think of...

You should sign up for cryonics. I want to see you in the future.

…intelligence explosion…

What’s your confidence interval?

You know what they say: one man’s Modus Ponens is another man’s Modus Tollens

This may sound a bit crazy right now, but hear me out…

What are your priors?

When the singularity comes that won’t be a problem anymore.

I like to think I’d do that, but I don’t fully trust myself. I am running on corrupted hardware after all.

I want to be with you, and I don’t foresee that changing in the near future.

…Bayesian statistics…

So Omega appears in front of you…

What would you say the probability of that event is, if your beliefs are true?

 

Others?