You are viewing revision 1.67.0, last edited by Ruby

This page contains summaries of LessWrong posts published in 2009.

Free to Optimize

It may be better to create a world that operates by better rules, that you can understand, so that you can optimize your own future, than to create a world that includes some sort of deity that can be prayed to. The human reluctance to have their future controlled by an outside source is a nontrivial part of morality.

The Uses of Fun (Theory)

Fun Theory is important for replying to critics of human progress; for inspiring people to keep working on human progress; for refuting religious arguments that the world could possibly have been benevolently designed; for showing that religious Heavens show the signature of the same human biases that torpedo other attempts at Utopia; and for appreciating the great complexity of our values and of a life worth living, which requires a correspondingly strong effort of AI design to create AIs that can play good roles in a good future.

Growing Up is Hard

Each part of the human brain is optimized for behaving correctly, assuming that the rest of the brain is operating exactly as expected. Change one part, and the rest of your brain may not work as well. Increasing a human's intelligence is not a trivial problem.

Changing Emotions

Creating new emotions seems like a desirable aspect of many parts of Fun Theory, but this is not to be trivially postulated. It's the sort of thing best done with superintelligent help, and slowly and conservatively even then. We can illustrate these difficulties by trying to translate the short English phrase "change sex" into a cognitive transformation of extraordinary complexity and many hidden subproblems.

Rationality Quotes 21

Emotional Involvement

Since the events in video games have no actual long-term consequences, playing a video game is not likely to be nearly as emotionally involving as much less dramatic events in real life. The supposed Utopia of playing lots of cool video games forever, is life as a series of disconnected episodes with no lasting consequences. Our current emotions are bound to activities that were subgoals of reproduction in the ancestral environment - but we now pursue these activities as independent goals regardless of whether they lead to reproduction.

Rationality Quotes 22

Serious Stories

Stories and lives are optimized according to rather different criteria. Advice on how to write fiction will tell you that "stories are about people's pain" and "every scene must end in disaster". I once assumed that it was not possible to write any story about a successful Singularity because the inhabitants would not be in any pain; but something about the final conclusion that the post-Singularity world would contain no stories worth telling seemed alarming. Stories in which nothing ever goes wrong, are painful to read; would a life of endless success have the same painful quality? If so, should we simply eliminate that revulsion via neural rewiring? Pleasure probably does retain its meaning in the absence of pain to contrast it; they are different neural systems. The present world has an imbalance between pain and pleasure; it is much easier to produce severe pain than correspondingly intense pleasure. One path would be to address the imbalance and create a world with more pleasures, and free of the more grindingly destructive and pointless sorts of pain. Another approach would be to eliminate pain entirely. I feel like I prefer the former approach, but I don't know if it can last in the long run.

Rationality Quotes 23

Continuous Improvement

Humans seem to be on a hedonic treadmill; over time, we adjust to any improvements in our environment - after a month, the new sports car no longer seems quite as wonderful. This aspect of our evolved psychology is not surprising: it is a rare organism in a rare environment whose optimal reproductive strategy is to rest with a smile on its face, feeling happy with what it already has. To entirely delete the hedonic treadmill seems perilously close to tampering with Boredom itself. Is there enough fun in the universe for a transhuman to jog off the treadmill - improve their life continuously, leaping to ever-higher hedonic levels before adjusting to the previous one? Can ever-higher levels of pleasure be created by the simple increase of ever-larger floating-point numbers in a digital pleasure center, or would that fail to have the full subjective quality of happiness? If we continue to bind our pleasures to novel challenges, can we find higher levels of pleasure fast enough, without cheating? The rate at which value can increase as more bits are added, and the rate at which value must increase for eudaimonia, together determine the lifespan of a mind. If minds must use exponentially more resources over time in order to lead a eudaimonic existence, their subjective lifespan is measured in mere millennia even if they can draw on galaxy-sized resources.

Eutopia is Scary

If a citizen of the Past were dropped into the Present world, they would be pleasantly surprised along at least some dimensions; they would also be horrified, disgusted, and frightened. This is not because our world has gone wrong, but because it has gone right. A true Future gone right would, realistically, be shocking to us along at least some dimensions. This may help explain why most literary Utopias fail; as George Orwell observed, "they are chiefly concerned with avoiding fuss". Heavens are meant to sound like good news; political utopias are meant to show how neatly their underlying ideas work. Utopia is reassuring, unsurprising, and dull. Eutopia would be scary. (Of course the vast majority of scary things are not Eutopian, just entropic.) Try to imagine a genuinely better world in which you would be out of place - not a world that would make you smugly satisfied at how well all your current ideas had worked. This proved to be a very important exercise when I tried it; it made me realize that all my old proposals had been optimized to sound safe and reassuring.

Building Weirdtopia

Utopia and Dystopia both confirm the moral sensibilities you started with; whether the world is a libertarian utopia of government non-interference, or a hellish dystopia of government intrusion and regulation, either way you get to say "Guess I was right all along." To break out of this mold, write down the Utopia, and the Dystopia, and then try to write down the Weirdtopia - an arguably-better world that zogs instead of zigging or zagging. (Judging from the comments, this exercise seems to have mostly failed.)

She has joined the Conspiracy

Justified Expectation of Pleasant Surprises

A pleasant surprise probably has a greater hedonic impact than being told about the same positive event long in advance - hearing about the positive event is good news in the moment of first hearing, but you don't have the gift actually in hand. Then you have to wait, perhaps for a long time, possibly comparing the expected pleasure of the future to the lesser pleasure of the present. This argues that if you have a choice between a world in which the same pleasant events occur, but in the first world you are told about them long in advance, and in the second world they are kept secret until they occur, you would prefer to live in the second world. The importance of hope is widely appreciated - people who do not expect their lives to improve in the future are less likely to be happy in the present - but the importance of vague hope may be understated.

Seduced by Imagination

Vagueness usually has a poor name in rationality, but the Future is something about which, in fact, we do not possess strong reliable specific information. Vague (but justified!) hopes may also be hedonically better. But a more important caution for today's world is that highly specific pleasant scenarios can exert a dangerous power over human minds - suck out our emotional energy, make us forget what we don't know, and cause our mere actual lives to pale by comparison. (This post is not about Fun Theory proper, but it contains an important warning about how not to use Fun Theory.)

Getting Nearer

How should rationalists use their near and far modes of thinking? And how should knowing about near versus far modes influence how we present the things we believe to other people?

In Praise of Boredom

"Boredom" is an immensely subtle and important aspect of human values, nowhere near as straightforward as it sounds to a human. We don't want to get bored with breathing or with thinking. We do want to get bored with playing the same level of the same video game over and over. We don't want changing the shade of the pixels in the game to make it stop counting as "the same game". We want a steady stream of novelty, rather than spending most of our time playing the best video game level so far discovered (over and over) and occasionally trying out a different video game level as a new candidate for "best". These considerations would not arise in most utility functions in expected utility maximizers.

Sympathetic Minds

Mirror neurons are neurons that fire both when performing an action oneself, and watching someone else perform the same action - for example, a neuron that fires when you raise your hand or watch someone else raise theirs. We predictively model other minds by putting ourselves in their shoes, which is empathy. But some of our desire to help relatives and friends, or be concerned with the feelings of allies, is expressed as sympathy, feeling what (we believe) they feel. Like "boredom", the human form of sympathy would not be expected to arise in an arbitrary expected-utility-maximizing AI. Most such agents would regard any agents in its environment as a special case of complex systems to be modeled or optimized; it would not feel what they feel.

Interpersonal Entanglement

Our sympathy with other minds makes our interpersonal relationships one of the most complex aspects of human existence. Romance, in particular, is more complicated than being nice to friends and kin, negotiating with allies, or outsmarting enemies - it contains aspects of all three. Replacing human romance with anything simpler or easier would decrease the peak complexity of the human species - a major step in the wrong direction, it seems to me. This is my problem with proposals to give people perfect, nonsentient sexual/romantic partners, which I usually refer to as "catgirls" ("catboys"). The human species does have a statistical sex problem: evolution has not optimized the average man to make the average woman happy or vice versa. But there are less sad ways to solve this problem than both genders giving up on each other and retreating to catgirls/catboys.

Failed Utopia #4-2

A fictional short story illustrating some of the ideas in Interpersonal Entanglement above. (Many commenters seemed to like this story, and some said that the ideas were easier to understand in this form.)

Investing for the Long Slump

What should you do if you think that the world's economy is going to stay bad for a very long time? How could such a scenario happen?

Higher Purpose

Having a Purpose in Life consistently shows up as something that increases stated well-being. Of course, the problem with trying to pick out "a Purpose in Life" in order to make yourself happier, is that this doesn't take you outside yourself; it's still all about you. To find purpose, you need to turn your eyes outward to look at the world and find things there that you care about - rather than obsessing about the wonderful spiritual benefits you're getting from helping others. In today's world, most of the highest-priority legitimate Causes consist of large groups of people in extreme jeopardy: Aging threatens the old, starvation threatens the poor, extinction risks threaten humanity as a whole. If the future goes right, many and perhaps all such problems will be solved - depleting the stream of victims to be helped. Will the future therefore consist of self-obsessed individuals, with nothing to take them outside themselves? I suggest, though, that even if there were no large groups of people in extreme jeopardy, we would still, looking around, find things outside ourselves that we cared about - friends, family; truth, freedom... Nonetheless, if the Future goes sufficiently well, there will come a time when you could search the whole of civilization, and never find a single person so much in need of help, as dozens you now pass on the street. If you do want to save someone from death, or help a great many people, then act now; your opportunity may not last, one way or another.

Rationality Quotes 24

The Fun Theory Sequence

Describes some of the many complex considerations that determine what sort of happiness we most prefer to have - given that many of us would decline to just have an electrode planted in our pleasure centers.

BHTV: Yudkowsky / Wilkinson

31 Laws of Fun

A brief summary of principles for writing fiction set in a eutopia.

OB Status Update

Rationality Quotes 25

Value is Fragile

An interesting universe, that would be incomprehensible to the universe today, is what the future looks like if things go right. There are a lot of things that humans value that if you did everything else right, when building an AI, but left out that one thing, the future would wind up looking dull, flat, pointless, or empty. Any Future not shaped by a goal system with detailed reliable inheritance from human morals and metamorals, will contain almost nothing of worth.

Three Worlds Collide (0/8)

The Baby-Eating Aliens (1/8)

Future explorers discover an alien civilization, and learns something unpleasant about their civilization.

War and/or Peace (2/8)

The true prisoner's dilemma against aliens. The conference struggles to decide the appropriate course of action.

The Super Happy People (3/8)

Humanity encounters new aliens that see the existence of pain amongst humans as morally unacceptable.

Interlude with the Confessor (4/8)

Akon talks things over with the Confessor, and receives a history lesson.

Three Worlds Decide (5/8)

The Superhappies propose a compromise.

Normal Ending: Last Tears (6/8)

Humanity accepts the Superhappies' bargain.

True Ending: Sacrificial Fire (7/8)

The Impossible Possible World tries to save humanity.

Epilogue: Atonement (8/8)

The last moments aboard the Impossible Possible World.

The Thing That I Protect

The cause that drives Yudkowsky isn't Friendly AI, and it isn't even specifically about preserving human values. It's simply about a future that's a lot better than the present.

...And Say No More Of It

In the previous couple of months, Overcoming Bias had focused too much on singularity related issues and not enough on rationality. A two month moratorium on the topic of the singularity/intelligence explosion is imposed.

(Moral) Truth in Fiction?

It is possible to convey moral ideas in a clearer way through fiction than through abstract argument. Stories may also help us get closer to thinking about moral issues in near mode. Don't discount moral arguments just because they're written as fiction.

Informers and Persuaders

A purely hypothetical scenario about a world containing some authors trying to persuade people of a particular theory, and some authors simply trying to share valuable information.

Cynicism in Ev-Psych (and Econ?)

Evolutionary Psychology and Microeconomics seem to develop different types of cynical theories, and are cynical about different things.

The Evolutionary-Cognitive Boundary

It's worth drawing a sharp boundary between ideas about evolutionary reasons for behavior, and cognitive reasons for behavior.

An Especially Elegant Evpsych Experiment

An experiment comparing expected parental grief at the death of a child at different ages, to the reproductive success rate of children at that age in a hunter gatherer tribe.

Rationality Quotes 26

An African Folktale

A story that seems to point to some major cultural differences.

Cynical About Cynicism

Much of cynicism seems to be about signaling sophistication, rather than sharing uncommon, true, and important insights.

Good Idealistic Books are Rare

Much of our culture is the official view, not the idealistic view.

Against Maturity

Dividing the world up into "childish" and "mature" is not a useful way to think.

Pretending to be Wise

Trying to signal wisdom or maturity by taking a neutral position is very seldom the right course of action.

Wise Pretensions v.0

An earlier post, on the same topic as yesterday's post.

Rationality Quotes 27

Fairness vs. Goodness

An experiment in which two unprepared subjects play an asymmetric version of the Prisoner's Dilemma. Is the best outcome the one where each player gets as many points as possible, or the one in which each player gets about the same number of points?

On Not Having an Advance Abyssal Plan

Don't say that you'll figure out a solution to the worst case scenario if the worst case scenario happens. Plan it out in advance.

About Less Wrong

Formative Youth

People underestimate the extent to which their own beliefs and attitudes are influenced by their experiences as a child.

Tell Your Rationalist Origin Story

Markets are Anti-Inductive

The standard theory of efficient markets says that exploitable regularities in the past, shouldn't be exploitable in the future. If everybody knows that "stocks have always gone up", then there's no reason to sell them.

Issues, Bugs, and Requested Features

The Most Important Thing You Learned

The Most Frequently Useful Thing

That You'd Tell All Your Friends

Test Your Rationality

You should try hard and often to test your rationality, but how can you do that?

Unteachable Excellence

If it were possible to teach people reliably how to become exceptional, then it would no longer be exceptional.

The Costs of Rationality

Teaching the Unteachable

There are many things we do that we can't easily understand how we do them. Teaching them is therefore a challenge.

No, Really, I've Deceived Myself

Some people who have fallen into self-deception haven't actually deceived themselves. Some of them simply believe that they have deceived themselves, but have not actually done this.

The ethic of hand-washing and community epistemic practice

Belief in Self-Deception

Deceiving yourself is harder than it seems. What looks like a successively adopted false belief may actually be just a belief in false belief.

Rationality and Positive Psychology

Posting now enabled

Kinnaird's truels

Recommended Rationalist Resources

Information cascades

Is it rational to take psilocybin?

Does blind review slow down science?

Formalization is a rationality technique

Slow down a little... maybe?

Checklists

The Golem

Simultaneously Right and Wrong

Moore's Paradox

People often mistake reasons for endorsing a proposition for reasons to believe that proposition.

It's the Same Five Dollars!

Lies and Secrets

The Mystery of the Haunted Rationalist

Don't Believe You'll Self-Deceive

It may be wise to tell yourself that you will not be able to successfully deceive yourself, because by telling yourself this, you may make it true.

The Wrath of Kahneman

The Mistake Script

LessWrong anti-kibitzer (hides comment authors and vote counts)

You May Already Be A Sinner

Striving to Accept

Trying extra hard to believe something seems like Dark Side Epistemology, but what about trying extra hard to accept something that you know is true.

Software tools for community truth-seeking

Wanted: Python open source volunteers

Selective processes bring tag-alongs (but not always!)

Adversarial System Hats

Beginning at the Beginning

The Apologist and the Revolutionary

Raising the Sanity Waterline

Behind every particular failure of social rationality is a larger and more general failure of social rationality; even if all religious content were deleted tomorrow from all human minds, the larger failures that permit religion would still be present. Religion may serve the function of an asphyxiated canary in a coal mine - getting rid of the canary doesn't get rid of the gas. Even a complete social victory for atheism would only be the beginning of the real work of rationalists. What could you teach people without ever explicitly mentioning religion, that would raise their general epistemic waterline to the point that religion went underwater?

So you say you're an altruist...

A Sense That More Is Possible

The art of human rationality may have not been much developed because its practitioners lack a sense that vastly more is possible. The level of expertise that most rationalists strive to develop is not on a par with the skills of a professional mathematician - more like that of a strong casual amateur. Self-proclaimed "rationalists" don't seem to get huge amounts of personal mileage out of their craft, and no one sees a problem with this. Yet rationalists get less systematic training in a less systematic context than a first-dan black belt gets in hitting people.

Talking Snakes: A Cautionary Tale

Boxxy and Reagan

Dialectical Bootstrapping

Is Santa Real?

Epistemic Viciousness

An essay by Gillian Russell on "Epistemic Viciousness in the Martial Arts" generalizes amazingly to possible and actual problems with building a community around rationality. Most notably the extreme dangers associated with "data poverty" - the difficulty of testing the skills in the real world. But also such factors as the sacredness of the dojo, the investment in teachings long-practiced, the difficulty of book learning that leads into the need to trust a teacher, deference to historical masters, and above all, living in data poverty while continuing to act as if the luxury of trust is possible.

On the Care and Feeding of Young Rationalists

The Least Convenient Possible World

Closet survey #1

Soulless morality

The Skeptic's Trilemma

Schools Proliferating Without Evidence

The branching schools of "psychotherapy", another domain in which experimental verification was weak (nonexistent, actually), show that an aspiring craft lives or dies by the degree to which it can be tested in the real world. In the absence of that testing, one becomes prestigious by inventing yet another school and having students, rather than excelling at any visible performance criterion. The field of hedonic psychology (happiness studies) began, to some extent, with the realization that you could measure happiness - that there was a family of measures that by golly did validate well against each other. The act of creating a new measurement creates new science; if it's a good measurement, you get good science.

Really Extreme Altruism

Storm by Tim Minchin

3 Levels of Rationality Verification

How far the craft of rationality can be taken, depends largely on what methods can be invented for verifying it. Tests seem usefully stratifiable into reputational, experimental, and organizational. A "reputational" test is some real-world problem that tests the ability of a teacher or a school (like running a hedge fund, say) - "keeping it real", but without being able to break down exactly what was responsible for success. An "experimental" test is one that can be run on each of a hundred students (such as a well-validated survey). An "organizational" test is one that can be used to preserve the integrity of organizations by validating individuals or small groups, even in the face of strong incentives to game the test. The strength of solution invented at each level will determine how far the craft of rationality can go in the real world.

The Tragedy of the Anticommons

Are You a Solar Deity?

In What Ways Have You Become Stronger?

Brainstorming verification tests, asking along what dimensions you think you've improved due to "rationality".

Taboo "rationality," please.

Science vs. art

What Do We Mean By "Rationality"?

When we talk about rationality, we're generally talking about either epistemic rationality (systematic methods of finding out the truth) or instrumental rationality (systematic methods of making the world more like we would like it to be). We can discuss these in the forms of probability theory and decision theory, but this doesn't fully cover the difficulty of being rational as a human. There is a lot more to rationality than just the formal theories.

Comments for "Rationality"

The "Spot the Fakes" Test

On Juvenile Fiction

Rational Me or We?

Dead Aid

Tarski Statements as Rationalist Exercise

The Pascal's Wager Fallacy Fallacy

People hear about a gamble involving a big payoff, and dismiss it as a form of Pascal's Wager. But the size of the payoff is not the flaw in Pascal's Wager. Just because an option has a very large potential payoff does not mean that the probability of getting that payoff is small, or that there are other possibilities that will cancel with it.

Never Leave Your Room

Rationalist Storybooks: A Challenge

A corpus of our community's knowledge

Little Johny Bayesian

How to Not Lose an Argument

Counterfactual Mugging

Rationalist Fiction

What works of fiction are out there that show characters who have acquired their skills at rationality through practice, and who we can watch in the act of employing those skills?

Rationalist Poetry Fans, Unite!

Precommitting to paying Omega.

Why Our Kind Can't Cooperate

The atheist/libertarian/technophile/sf-fan/early-adopter/programmer/etc crowd, aka "the nonconformist cluster", seems to be stunningly bad at coordinating group projects. There are a number of reasons for this, but one of them is that people are as reluctant to speak agreement out loud, as they are eager to voice disagreements - the exact opposite of the situation that obtains in more cohesive and powerful communities. This is not rational either! It is dangerous to be half a rationalist (in general), and this also applies to teaching only disagreement but not agreement, or only lonely defiance but not coordination. The pseudo-rationalist taboo against expressing strong feelings probably doesn't help either.

Just a reminder: Scientists are, technically, people.

Support That Sounds Like Dissent

Tolerate Tolerance

One of the likely characteristics of someone who sets out to be a "rationalist" is a lower-than-usual tolerance for flawed thinking. This makes it very important to tolerate other people's tolerance - to avoid rejecting them because they tolerate people you wouldn't - since otherwise we must all have exactly the same standards of tolerance in order to work together, which is unlikely. Even if someone has a nice word to say about complete lunatics and crackpots - so long as they don't literally believe the same ideas themselves - try to be nice to them? Intolerance of tolerance corresponds to punishment of non-punishers, a very dangerous game-theoretic idiom that can lock completely arbitrary systems in place even when they benefit no one at all.

Mind Control and Me

Individual Rationality Is a Matter of Life and Death

The Power of Positivist Thinking

Don't Revere The Bearer Of Good Info

You're Calling *Who* A Cult Leader?

Paul Graham gets exactly the same accusations about "cults" and "echo chambers" and "coteries" that I do, in exactly the same tone - e.g. comparing the long hours worked by Y Combinator startup founders to the sleep-deprivation tactic used in cults, or claiming that founders were asked to move to the Bay Area startup hub as a cult tactic of separation from friends and family. This is bizarre, considering our relative surface risk factors. It just seems to be a failure mode of the nonconformist community in general. By far the most cultish-looking behavior on Hacker News is people trying to show off how willing they are to disagree with Paul Graham, which, I can personally testify, feels really bizarre when you're the target. Admiring someone shouldn't be so scary - I don't hold back so much when praising e.g. Douglas Hofstadter; in this world there are people who have pulled off awesome feats and it is okay to admire them highly.

Cached Selves

Eliezer Yudkowsky Facts

When Truth Isn't Enough

BHTV: Yudkowsky & Adam Frank on "religious experience"

I'm confused. Could someone help?

Playing Video Games In Shuffle Mode

Book: Psychiatry and the Human Condition

Thoughts on status signals

Bogus Pipeline, Bona Fide Pipeline

On Things that are Awesome

Seven thoughts: I can list more than one thing that is awesome; when I think of "Douglas Hofstadter" I am really thinking of his all-time greatest work; the greatest work is not the person; when we imagine other people we are imagining their output, so the real Douglas Hofstadter is the source of "Douglas Hofstadter"; I most strongly get the sensation of awesomeness when I see someone outdoing me overwhelmingly, at some task I've actually tried; we tend to admire unique detailed awesome things and overlook common nondetailed awesome things; religion and its bastard child "spirituality" tends to make us overlook human awesomeness.

Hyakujo's Fox

Terrorism is not about Terror

The Implicit Association Test

Contests vs. Real World Problems

The Sacred Mundane

There are a lot of bad habits of thought that have developed to defend religious and spiritual experience. They aren't worth saving, even if we discard the original lie. Let's just admit that we were wrong, and enjoy the universe that's actually here.

Extreme updating: The devil is in the missing details

Spock's Dirty Little Secret

The Good Bayesian

Fight Biases, or Route Around Them?

Why *I* fail to act rationally

Open Thread: March 2009

Two Blegs

Your Price for Joining

The game-theoretical puzzle of the Ultimatum game has its reflection in a real-world dilemma: How much do you demand that an existing group adjust toward you, before you will adjust toward it? Our hunter-gatherer instincts will be tuned to groups of 40 with very minimal administrative demands and equal participation, meaning that we underestimate the inertia of larger and more specialized groups and demand too much before joining them. In other groups this resistance can be overcome by affective death spirals and conformity, but rationalists think themselves too good for this - with the result that people in the nonconformist cluster often set their joining prices way way way too high, like an 50-way split with each player demanding 20% of the money. Nonconformists need to move in the direction of joining groups more easily, even in the face of annoyances and apparent unresponsiveness. If an issue isn't worth personally fixing by however much effort it takes, it's not worth a refusal to contribute.

Sleeping Beauty gets counterfactually mugged

The Mind Is Not Designed For Thinking

Crowley on Religious Experience

Can Humanism Match Religion's Output?

Anyone with a simple and obvious charitable project - responding with food and shelter to a tidal wave in Thailand, say - would be better off by far pleading with the Pope to mobilize the Catholics, rather than with Richard Dawkins to mobilize the atheists. For so long as this is true, any increase in atheism at the expense of Catholicism will be something of a hollow victory, regardless of all other benefits. Can no rationalist match the motivation that comes from the irrational fear of Hell? Or does the real story have more to do with the motivating power of physically meeting others who share your cause, and group norms of participating?

On Seeking a Shortening of the Way

Altruist Coordination -- Central Station

Less Wrong Facebook Page

The Hidden Origins of Ideas

Defense Against The Dark Arts: Case Study #1

Church vs. Taskforce

Churches serve a role of providing community - but they aren't explicitly optimized for this, because their nominal role is different. If we desire community without church, can we go one better in the course of deleting religion? There's a great deal of work to be done in the world; rationalist communities might potentially organize themselves around good causes, while explicitly optimizing for community.

When It's Not Right to be Rational

The Zombie Preacher of Somerset

Hygienic Anecdotes

Rationality: Common Interest of Many Causes

Many causes benefit particularly from the spread of rationality - because it takes a little more rationality than usual to see their case, as a supporter, or even just a supportive bystander. Not just the obvious causes like atheism, but things like marijuana legalization. In the case of my own work this effect was strong enough that after years of bogging down I threw up my hands and explicitly recursed on creating rationalists. If such causes can come to terms with not individually capturing all the rationalists they create, then they can mutually benefit from mutual effort on creating rationalists. This cooperation may require learning to shut up about disagreements between such causes, and not fight over priorities, except in specialized venues clearly marked.

Ask LW: What questions to test in our rationality questionnaire?

Requesting suggestions for an actual survey to be run.

Bay area OB/LW meetup, today, Sunday, March 29, at 5pm

Akrasia, hyperbolic discounting, and picoeconomics

Deliberate and spontaneous creativity

Most Rationalists Are Elsewhere

Framing Effects in Anthropology

Kling, Probability, and Economics

Helpless Individuals

When you consider that our grouping instincts are optimized for 50-person hunter-gatherer bands where everyone knows everyone else, it begins to seem miraculous that modern-day large institutions survive at all. And in fact, the vast majority of large modern-day institutions simply fail to exist in the first place. This is why funding of Science is largely through money thrown at Science rather than donations from individuals - research isn't a good emotional fit for the rare problems that individuals can manage to coordinate on. In fact very few things are, which is why e.g. 200 million adult Americans have such tremendous trouble supervising the 535 members of Congress. Modern humanity manages to put forth very little in the way of coordinated individual effort to serve our collective individual interests.

The Benefits of Rationality?

Money: The Unit of Caring

Omohundro's resource balance principle implies that the inside of any approximately rational system has a common currency of expected utilons. In our world, this common currency is called "money" and it is the unit of how much society cares about something - a brutal yet obvious point. Many people, seeing a good cause, would prefer to help it by donating a few volunteer hours. But this avoids the tremendous gains of comparative advantage, professional specialization, and economies of scale - the reason we're not still in caves, the only way anything ever gets done in this world, the tools grownups use when anyone really cares. Donating hours worked within a professional specialty and paying-customer priority, whether directly, or by donating the money earned to hire other professional specialists, is far more effective than volunteering unskilled hours.

Building Communities vs. Being Rational

Degrees of Radical Honesty

Introducing CADIE

Purchase Fuzzies and Utilons Separately

Wealthy philanthropists typically make the mistake of trying to purchase warm fuzzy feelings, status among friends, and actual utilitarian gains, simultaneously; this results in vague pushes along all three dimensions and a mediocre final result. It should be far more effective to spend some money/effort on buying altruistic fuzzies at maximum optimized efficiency (e.g. by helping people in person and seeing the results in person), buying status at maximum efficiency (e.g. by donating to something sexy that you can brag about, regardless of effectiveness), and spending most of your money on expected utilons (chosen through sheer cold-blooded shut-up-and-multiply calculation, without worrying about status or fuzzies).

Proverbs and Cached Judgments: the Rolling Stone

You don't need Kant

Accuracy Versus Winning

Wrong Tomorrow

Selecting Rationalist Groups

Trying to breed e.g. egg-laying chickens by individual selection can produce odd side effects on the farm level, since a more dominant hen can produce more egg mass at the expense of other hens. Group selection is nearly impossible in Nature, but easy to impose in the laboratory, and group-selecting hens produced substantial increases in efficiency. Though most of my essays are about individual rationality - and indeed, Traditional Rationality also praises the lone heretic more than evil Authority - the real effectiveness of "rationalists" may end up determined by their performance in groups.

Aumann voting; or, How to vote when you're ignorant

"Robot scientists can think for themselves"

Where are we?

The Brooklyn Society For Ethical Culture

Open Thread: April 2009

Rationality is Systematized Winning

The idea behind the statement "Rationalists should win" is not that rationality will make you invincible. It means that if someone who isn't behaving according to your idea of rationality is outcompeting you, predictably and consistently, you should consider that you're not the one being rational.

Another Call to End Aid to Africa

First London Rationalist Meeting upcoming

On dollars, utility, and crack cocaine

Incremental Progress and the Valley

The optimality theorems for probability theory and decision theory, are for perfect probability theory and decision theory. There is no theorem that incremental changes toward the ideal, starting from a flawed initial form, must yield incremental progress at each step along the way. Since perfection is unattainable, why dare to try for improvement? But my limited experience with specialized applications suggests that given enough progress, one can achieve huge improvements over baseline - it just takes a lot of progress to get there.

The First London Rationalist Meetup

Why Support the Underdog?

Off-Topic Discussion Thread: April 2009

Voting etiquette

Formalizing Newcomb's

Supporting the underdog is explained by Hanson's Near/Far distinction

Real-Life Anthropic Weirdness

Extremely rare events can create bizarre circumstances in which people may not be able to effectively communicate about improbability.

Rationalist Wiki

Rationality Toughness Tests

Heuristic is not a bad word

Rationalists should beware rationalism

Newcomb's Problem standard positions

Average utilitarianism must be correct?

Rationalist wiki, redux

What do fellow rationalists think about Mensa?

Extenuating Circumstances

You can excuse other people's shortcomings on the basis of extenuating circumstances, but you shouldn't do that with yourself.

On Comments, Voting, and Karma - Part I

Newcomb's Problem vs. One-Shot Prisoner's Dilemma

What isn't the wiki for?

Eternal Sunshine of the Rational Mind

Of Lies and Black Swan Blowups

Whining-Based Communities

Many communities feed emotional needs by offering their members someone or something to blame for failure - say, those looters who don't approve of your excellence. You can easily imagine some group of "rationalists" congratulating themselves on how reasonable they were, while blaming the surrounding unreasonable society for keeping them down. But this is not how real rationality works - there's no assumption that other agents are rational. We all face unfair tests (and yes, they are unfair to different degrees for different people); and how well you do with your unfair tests, is the test of your existence. Rationality is there to help you win anyway, not to provide a self-handicapping excuse for losing. There are no first-person extenuating circumstances. There is absolutely no point in going down the road of mutual bitterness and consolation, about anything, ever.

Help, help, I'm being oppressed!

Zero-based karma coming through

E-Prime

Mandatory Secret Identities

This post was not well-received, but the point was to suggest that a student must at some point leave the dojo and test their skills in the real world. The aspiration of an excellent student should not consist primarily of founding their own dojo and having their own students.

Rationality, Cryonics and Pascal's Wager

Less Wrong IRC Meetup

"Stuck In The Middle With Bruce"

Extreme Rationality: It's Not That Great

"Playing to Win"

The term "playing to win" comes from Sirlin's book and can be described as using every means necessary to win as long as those means are legal within the structure of the game being played.

Secret Identities vs. Groupthink

Silver Chairs, Paternalism, and Akrasia

Extreme Rationality: It Could Be Great

The uniquely awful example of theism

Beware of Other-Optimizing

Aspiring rationalists often vastly overestimate their own ability to optimize other people's lives. They read nineteen webpages offering productivity advice that doesn't work for them... and then encounter the twentieth page, or invent a new method themselves, and wow, it really works - they've discovered the true method. Actually, they've just discovered the one method in twenty that works for them, and their confident advice is no better than randomly selecting one of the twenty blog posts. Other-Optimizing is exceptionally dangerous when you have power over the other person - for then you'll just believe that they aren't trying hard enough.

How theism works

That Crisis thing seems pretty useful

Spay or Neuter Your Irrationalities

The Unfinished Mystery of the Shangri-La Diet

An intriguing dietary theory which appears to allow some people to lose substantial amounts of weight, but doesn't appear to work at all for others.

Akrasia and Shangri-La

The Shangri-La diet works amazingly well for some people, but completely fails for others, for no known reason. Since the diet has a metabolic rationale and is not supposed to require willpower, its failure in my and other cases is unambigiously mysterious. If it required a component of willpower, then I and others might be tempted to blame myself for not having willpower. The art of combating akrasia (willpower failure) has the same sort of mysteries and is in the same primitive state; we don't know the deeper rule that explains why a trick works for one person but not another.

Maybe Theism Is OK

Metauncertainty

Is masochism necessary?

Missed Distinctions

Toxic Truth

Too much feedback can be a bad thing

Twelve Virtues booklet printing?

How Much Thought

Awful Austrians

Sunk Cost Fallacy

It's okay to be (at least a little) irrational

Marketing rationalism

Bystander Apathy

The bystander effect is when groups of people are less likely to take action than an individual. There are a few explanations for why this might be the case.

Persuasiveness vs Soundness

Declare your signaling and hidden agendas

GroupThink, Theism ... and the Wiki

Collective Apathy and the Internet

The causes of bystander apathy are even worse on the Internet. There may be an opportunity here for a startup to deliberately try to avert bystander apathy in online group coordination.

Tell it to someone who doesn't care

Bayesians vs. Barbarians

Suppose that a country of rationalists is attacked by a country of Evil Barbarians who know nothing of probability theory or decision theory. There's a certain concept of "rationality" which says that the rationalists inevitably lose, because the Barbarians believe in a heavenly afterlife if they die in battle, while the rationalists would all individually prefer to stay out of harm's way. So the rationalist civilization is doomed; it is too elegant and civilized to fight the savage Barbarians... And then there's the idea that rationalists should be able to (a) solve group coordination problems, (b) care a lot about other people and (c) win...

Actions and Words: Akrasia and the Fruit of Self-Knowledge

Mechanics without wrenches

I Changed My Mind Today - Canned Laughter

Of Gender and Rationality

Analysis of the gender imbalance that appears in "rationalist" communities, suggesting nine possible causes of the effect, and possible corresponding solutions.

Welcome to Less Wrong!

Instrumental Rationality is a Chimera

Practical rationality questionnaire

My Way

I sometimes think of myself as being like the protagonist in a classic SF labyrinth story, wandering further and further into some alien artifact, trying to radio back a description of what I'm seeing, so that I can be followed. But what I'm finding is not just the Way, the thing that lies at the center of the labyrinth; it is also my Way, the path that I would take to come closer to the center, from whatever place I started out. And yet there is still a common thing we are all trying to find. We should be aware that others' shortest paths may not be the same as our own, but this is not the same as giving up the ability to judge or to share.

The Art of Critical Decision Making

The Trouble With "Good"

While we're on the subject of meta-ethics...

Chomsky on reason and science

Anti-rationality quotes

Two-Tier Rationalism

My main problem with utilitarianism

Just for fun - let's play a game.

Rationality Quotes - April 2009

The Epistemic Prisoner's Dilemma

How a pathological procrastinor can lose weight (Anti-akrasia)

Atheist or Agnostic?

Great Books of Failure

Weekly Wiki Workshop and suggested articles

The True Epistemic Prisoner's Dilemma

Spreading the word?

The ideas you're not ready to post

Evangelical Rationality

The Sin of Underconfidence

When subjects know about a bias or are warned about a bias, overcorrection is not unheard of as an experimental result. That's what makes a lot of cognitive subtasks so troublesome - you know you're biased but you're not sure how much, and if you keep tweaking you may overcorrect. The danger of underconfidence (overcorrecting for overconfidence) is that you pass up opportunities on which you could have been successful; not challenging difficult enough problems; losing forward momentum and adopting defensive postures; refusing to put the hypothesis of your inability to the test; losing enough hope of triumph to try hard enough to win. You should ask yourself "Does this way of thinking make me stronger, or weaker?"

Masochism vs. Self-defeat

Well-Kept Gardens Die By Pacifism

Good online communities die primarily by refusing to defend themselves, and so it has been since the days of Eternal September. Anyone acculturated by academia knows that censorship is a very grave sin... in their walled gardens where it costs thousands and thousands of dollars to enter. A community with internal politics will treat any attempt to impose moderation as a coup attempt (since internal politics seem of far greater import than invading barbarians). In rationalist communities this is probably an instance of underconfidence - mildly competent moderators are probably quite trustworthy to wield the banhammer. On Less Wrong, the community is the moderator (via karma) and you will need to trust yourselves enough to wield the power and keep the garden clear.

UC Santa Barbara Rationalists Unite - Saturday, 6PM

LessWrong Boo Vote (Stochastic Downvoting)

Proposal: Use the Wiki for Concepts

Escaping Your Past

Go Forth and Create the Art!

I've developed primarily the art of epistemic rationality, in particular, the arts required for advanced cognitive reductionism... arts like distinguishing fake explanations from real ones and avoiding affective death spirals. There is much else that needs developing to create a craft of rationality - fighting akrasia; coordinating groups; teaching, training, verification, and becoming a proper experimental science; developing better introductory literature... And yet it seems to me that there is a beginning barrier to surpass before you can start creating high-quality craft of rationality, having to do with virtually everyone who tries to think lofty thoughts going instantly astray, or indeed even realizing that a craft of rationality exists and that you ought to be studying cognitive science literature to create it. It's my hope that my writings, as partial as they are, will serve to surpass this initial barrier. The rest I leave to you.

Fix it and tell us what you did

This Didn't Have To Happen

Just a bit of humor...

What's in a name? That which we call a rationalist...

Rational Groups Kick Ass

Instrumental vs. Epistemic -- A Bardic Perspective

Programmatic Prediction markets

Cached Procrastination

Practical Advice Backed By Deep Theories

Knowledge of this heuristic might be useful in fighting akrasia.

(alternate summary:)

Practical advice is genuinely much, much more useful when it's backed up by concrete experimental results, causal models that are actually true, or valid math that is validly interpreted. (Listed in increasing order of difficulty.) Stripping out the theories and giving the mere advice alone wouldn't have nearly the same impact or even the same message; and oddly enough, translating experiments and math into practical advice seems to be a rare niche activity relative to academia. If there's a distinctive LW style, this is it.

"Self-pretending" is not as useful as we think

Where's Your Sense of Mystery?

Less Meta

The fact that this final series was on the craft and the community seems to have delivered a push in something of the wrong direction, (a) steering toward conversation about conversation and (b) making present accomplishment pale in the light of grander dreams. Time to go back to practical advice and deep theories, then.

SIAI call for skilled volunteers and potential interns

The Craft and the Community

Excuse me, would you like to take a survey?

Should we be biased?

Theism, Wednesday, and Not Being Adopted

The End (of Sequences)

Final Words

The conclusion of the Beisutsukai series.

Bayesian Cabaret

Verbal Overshadowing and The Art of Rationality

How Not to be Stupid: Starting Up

How Not to be Stupid: Know What You Want, What You Really Really Want

Epistemic vs. Instrumental Rationality: Approximations

What is control theory, and why do you need to know about it?

Re-formalizing PD

Generalizing From One Example

Generalization From One Example is a tendency to pay too much attention to the few anecdotal pieces of evidence you experienced, and model some general phenomenon based on them. This is a special case of availability bias, and the way in which the mistake unfolds is closely related to the correspondence bias and the hindsight bias.

Wednesday depends on us.

How to come up with verbal probabilities

Fighting Akrasia: Incentivising Action

Fire and Motion

Fiction of interest

How Not to be Stupid: Adorable Maybes

Rationalistic Losing

Rationalist Role in the Information Age

Conventions and Confusing Continuity Conundrums

Open Thread: May 2009

Second London Rationalist Meeting upcoming - Sunday 14:00

TED Talks for Less Wrong

The mind-killer

What I Tell You Three Times Is True

Return of the Survey

Essay-Question Poll: Dietary Choices

Allais Hack -- Transform Your Decisions!

Without models

Bead Jar Guesses

Applied scenario about forming priors.

Special Status Needs Special Support

How David Beats Goliath

How to use "philosophical majoritarianism"

Off Topic Thread: May 2009

Introduction Thread: May 2009

Consider Representative Data Sets

No Universal Probability Space

Wiki.lesswrong.com Is Live

Hardened Problems Make Brittle Models

Beware Trivial Inconveniences

On the Fence? Major in CS

Rationality is winning - or is it?

The First Koan: Drinking the Hot Iron Ball

Epistemic vs. Instrumental Rationality: Case of the Leaky Agent

Replaying History

Framing Consciousness

A Request for Open Problems

How Not to be Stupid: Brewing a Nice Cup of Utilitea

Step Back

You Are A Brain

No One Knows Stuff

Willpower Hax #487: Execute by Default

Rationality in the Media: Don't (New Yorker, May 2009)

Survey Results

A Parable On Obsolete Ideologies

"Open-Mindedness" - the video

Religion, Mystery, and Warm, Soft Fuzzies

Cheerios: An "Untested New Drug"

Essay-Question Poll: Voting

Outward Change Drives Inward Change

Share Your Anti-Akrasia Tricks

Wanting to Want

"What Is Wrong With Our Thoughts"

Bad reasons for a rationalist to lose

Supernatural Math

Rationality quotes - May 2009

Positive Bias Test (C++ program)

Catchy Fallacy Name Fallacy (and Supporting Disagreement)

Inhibition and the Mind

Least Signaling Activities?

Brute-force Music Composition

Changing accepted public opinion and Skynet

Homogeneity vs. heterogeneity (or, What kind of sex is most moral?)

Saturation, Distillation, Improvisation: A Story About Procedural Knowledge And Cookies

This Failing Earth

The Wire versus Evolutionary Psychology

Dissenting Views

Eric Drexler on Learning About Everything

Anime Explains the Epimenides Paradox

Do Fandoms Need Awfulness?

Can we create a function that provably predicts the optimization power of intelligences?

Link: The Case for Working With Your Hands

Image vs. Impact: Can public commitment be counterproductive for achievement?

A social norm against unjustified opinions?

Taking Occam Seriously

The Onion Goes Inside The Biased Mind

The Frontal Syndrome

Open Thread: June 2009

Concrete vs Contextual values

Bioconservative and biomoderate singularitarian positions

Would You Slap Your Father? Article Linkage and Discussion

With whom shall I diavlog?

Mate selection for the men here

Third London Rationalist Meeting

Post Your Utility Function

Probability distributions and writing style

My concerns about the term 'rationalist'

Honesty: Beyond Internal Truth

Macroeconomics, The Lucas Critique, Microfoundations, and Modeling in General

indexical uncertainty and the Axiom of Independence

London Rationalist Meetups bikeshed painting thread

The Aumann's agreement theorem game (guess 2/3 of the average)

Expected futility for humans

You can't believe in Bayes

Less wrong economic policy

The Terrible, Horrible, No Good, Very Bad Truth About Morality and What To Do About It

Let's reimplement EURISKO!

If it looks like utility maximizer and quacks like utility maximizer...

Typical Mind and Politics

Why safety is not safe

Rationality Quotes - June 2009

Readiness Heuristics

The two meanings of mathematical terms

The Laws of Magic

Intelligence enhancement as existential risk mitigation

Rationalists lose when others choose

Ask LessWrong: Human cognitive enhancement now?

Don't Count Your Chickens...

Applied Picoeconomics

Representative democracy awesomeness hypothesis

The Physiology of Willpower

Time to See If We Can Apply Anything We Have Learned

Cascio in The Atlantic, more on cognitive enhancement as existential risk mitigation

ESR's comments on some EY:OB/LW posts

Nonparametric Ethics

Shane Legg on prospect theory and computational finance

The Domain of Your Utility Function

The Monty Maul Problem

Guilt by Association

Lie to me?

Richard Dawkins TV - Baloney Detection Kit video

Coming Out

The Great Brain is Located Externally

People don't actually remember much of what they know, they only remember how to find it, and the fact that there is something to find. Thus, it's important to know about what's known in various domains, even without knowing the content.

Controlling your inner control circuits

What's In A Name?

Atheism = Untheism + Antitheism

Book Review: Complications

Open Thread: July 2009

Fourth London Rationalist Meeting?

Rationality Quotes - July 2009

Harnessing Your Biases

Avoiding Failure: Fallacy Finding

Not Technically Lying

The enemy within

Media bias

Can chess be a game of luck?

The Dangers of Partial Knowledge of the Way: Failing in School

An interesting speed dating study

Can self-help be bad for you?

Causality does not imply correlation

Recommended reading for new rationalists

Formalized math: dream vs reality

Causation as Bias (sort of)

Debate: Is short term planning in humans due to a short life or due to bias?

Jul 12 Bay Area meetup - Hanson, Vassar, Yudkowsky

Our society lacks good self-preservation mechanisms

Good Quality Heuristics

How likely is a failure of nuclear deterrence?

The Strangest Thing An AI Could Tell You

"Sex Is Always Well Worth Its Two-Fold Cost"

The Dirt on Depression

Fair Division of Black-Hole Negentropy: an Introduction to Cooperative Game Theory

Absolute denial for atheists

Causes of disagreements

The Popularization Bias

Zwicky's Trifecta of Illusions

Are You Anosognosic?

Article upvoting

Sayeth the Girl

Timeless Decision Theory: Problems I Can't Solve

An Akrasia Anecdote

Being saner about gender and rationality

Are you crazy?

Counterfactual Mugging v. Subjective Probability

Creating The Simple Math of Everything

Joint Distributions and the Slow Spread of Good Ideas

Chomsky, Sports Talk Radio, Media Bias, and Me

Outside Analysis and Blind Spots

Shut Up And Guess

Of Exclusionary Speech and Gender Politics

Missing the Trees for the Forest

Deciding on our rationality focus

Fairness and Geometry

It's all in your head-land

An observation on cryocrastination

The Price of Integrity

Are calibration and rational decisions mutually exclusive? (Part one)

The Nature of Offense

People are offended by grabs for status.

AndrewH's observation and opportunity costs

Are calibration and rational decisions mutually exclusive? (Part two)

Celebrate Trivial Impetuses

Freaky Fairness

Link: Interview with Vladimir Vapnik

Five Stages of Idolatry

Bayesian Flame

The Second Best

Bayesian Utility: Representing Preference by Probability Measures

The Trolley Problem in popular culture: Torchwood Series 3

Thomas C. Schelling's "Strategy of Conflict"

Information cascades in scientific practice

The Obesity Myth

The Hero With A Thousand Chances

Pract: A Guessing and Testing Game

An Alternative Approach to AI Cooperation

Open Thread: August 2009

Pain

Suffering

Why You're Stuck in a Narrative

Unspeakable Morality

The Difficulties of Potential People and Decision Making

Wits and Wagers

The usefulness of correlations

She Blinded Me With Science

The Machine Learning Personality Test

A Normative Rule for Decision-Changing Metrics

Recommended reading: George Orwell on knowledge from authority

Rationality Quotes - August 2009

Why Real Men Wear Pink

The Objective Bayesian Programme

LW/OB Rationality Quotes - August 2009

Exterminating life is rational

Robin Hanson's lists of Overcoming Bias Posts

Fighting Akrasia: Finding the Source

A note on hypotheticals

Dreams with Damaged Priors

Would Your Real Preferences Please Stand Up?

Calibration fail

Guess Again

Misleading the witness

Utilons vs. Hedons

Deleting paradoxes with fuzzy logic

Sense, Denotation and Semantics

Towards a New Decision Theory

Fighting Akrasia: Survey Design Help Request

Minds that make optimal use of small amounts of sensory data

Bloggingheads: Yudkowsky and Aaronson talk about AI and Many-worlds

Oh my God! It's full of Nash equilibria!

Happiness is a Heuristic

Experiential Pica

Friendlier AI through politics

Singularity Summit 2009 (quick post)

Scott Aaronson's "On Self-Delusion and Bounded Rationality"

Ingredients of Timeless Decision Theory

You have just been Counterfactually Mugged!

Evolved Bayesians will be biased

How inevitable was modern human civilization - data

Timeless Decision Theory and Meta-Circular Decision Theory

ESR's New Take on Qualia

The Journal of (Failed) Replication Studies

Working Mantras

Decision theory: An outline of some upcoming posts

How does an infovore manage information overload?

Confusion about Newcomb is confusion about counterfactuals

Mathematical simplicity bias and exponential functions

A Rationalist's Bookshelf: The Mind's I (Douglas Hofstadter and Daniel Dennett, 1981)

Pittsburgh Meetup: Survey of Interest

Paper: Testing ecological models

The Twin Webs of Knowledge

Don't be Pathologically Mugged!

Some counterevidence for human sociobiology

Cookies vs Existential Risk

Argument Maps Improve Critical Thinking

Great post on Reddit about accepting atheism

Optimal Strategies for Reducing Existential Risk

Open Thread: September 2009

Rationality Quotes - September 2009

LW/OB Quotes - Fall 2009

Knowing What You Know

Decision theory: Why we need to reduce "could", "would", "should"

The Featherless Biped

The Sword of Good

Torture vs. Dust vs. the Presumptuous Philosopher: Anthropic Reasoning in UDT

Notes on utility function experiment

Counterfactual Mugging and Logical Uncertainty

Bay Area OBLW Meetup Sep 12

Decision theory: Why Pearl helps reduce "could" and "would", but still leaves us with at least three alternatives

Forcing Anthropics: Boltzmann Brains

Why I'm Staying On Bloggingheads.tv

An idea: Sticking Point Learning

FHI postdoc at Oxford

Outlawing Anthropics: An Updateless Dilemma

Let Them Debate College Students

Pittsburgh Meetup: Saturday 9/12, 6:30PM, CMU

The Lifespan Dilemma

Formalizing informal logic

Timeless Identity Crisis

The New Nostradamus

Formalizing reflective inconsistency

Beware of WEIRD psychological samples

The Absent-Minded Driver

What is the Singularity Summit?

Sociosexual Orientation Inventory, or failing to perform basic sanity checks

Quantum Russian Roulette

MWI, weird quantum experiments and future-directed continuity of conscious experience

Minneapolis Meetup: Survey of interest

Hypothetical Paradoxes

Reason as memetic immune disorder

How to use SMILE to solve Bayes Nets

The Finale of the Ultimate Meta Mega Crossover

Ethics as a black box function

Avoiding doomsday: a "proof" of the self-indication assumption

Anthropic reasoning and correlated decision making

Boredom vs. Scope Insensitivity

Minneapolis Meetup, This Saturday (26th) at 3:00 PM, University of Minnesota

The utility curve of the human population

Solutions to Political Problems As Counterfactuals

Non-Malthusian Scenarios

Correlated decision making: a complete theory

The Scylla of Error and the Charybdis of Paralysis

The Anthropic Trilemma

Your Most Valuable Skill

Privileging the Hypothesis

Why Many-Worlds Is Not The Rationally Favored Interpretation

Intuitive differences: when to agree to disagree

NY-area OB/LW meetup Saturday 10/3 7 PM

Regular NYC Meetups

Open Thread: October 2009

Why Don't We Apply What We Know About Twins to Everybody Else?

Are you a Democrat singletonian, or a Republican singletonian?

Scott Aaronson on Born Probabilities

'oy, girls on lw, want to get together some time?'

When Willpower Attacks

Dying Outside

Don't Think Too Hard.

The Presumptuous Philosopher's Presumptuous Friend

The First Step is to Admit That You Have a Problem

Let them eat cake: Interpersonal Problems vs Tasks

New Haven/Yale Less Wrong Meetup: 5 pm, Monday October 12

Boston Area Less Wrong Meetup: 2 pm Sunday October 11th

LW Meetup Google Calendar

I'm Not Saying People Are Stupid

How to get that Friendly Singularity: a minority view

The Argument from Witness Testimony

Link: PRISMs, Gom Jabbars, and Consciousness (Peter Watts)

What Program Are You?

Do the 'unlucky' systematically underestimate high-variance strategies?

Anticipation vs. Faith: At What Cost Rationality?

The power of information?

Quantifying ethicality of human actions

BHTV: Eliezer Yudkowsky and Andrew Gelman

We're in danger. I must tell the others...

PredictionBook.com - Track your calibration

The Shadow Question

Information theory and FOOM

Waterloo, ON, Canada Meetup: 6pm Sun Oct 18 '09!

How to think like a quantum monadologist

Localized theories and conditional complexity

Applying Double Standards to "Divisive" Ideas

Near and far skills

Shortness is now a treatable condition

Lore Sjoberg's Life-Hacking FAQK

Why the beliefs/values dichotomy?

Biking Beyond Madness (link)

Rationality Quotes: October 2009

The continued misuse of the Prisoner's Dilemma

Better thinking through experiential games

Extreme risks: when not to use expected utility

Pound of Feathers, Pound of Gold

Arrow's Theorem is a Lie

The Value of Nature and Old Books

Circular Altruism vs. Personal Preference

Computer bugs and evolution

Doing your good deed for the day

Expected utility without the independence axiom

Post retracted: If you follow expected utility, expect to be money-pumped

A Less Wrong Q&A with Eliezer (Step 1: The Proposition)

David Deutsch: A new way to explain explanation

Less Wrong / Overcoming Bias meet-up groups

Our House, My Rules

Open Thread: November 2009

Re-understanding Robin Hanson's "Pre-Rationality"

Rolf Nelson's "The Rational Entrepreneur"

Money pumping: the axiomatic approach

Light Arts

News: Improbable Coincidence Slows LHC Repairs

Bay area LW meet-up

All hail the Lisbon Treaty! Or is that "hate"? Or just "huh"?

Hamster in Tutu Shuts Down Large Hadron Collider

The Danger of Stories

Practical rationality in surveys

Reflections on Pre-Rationality

Rationality advice from Terry Tao

Restraint Bias

What makes you YOU? For non-deists only.

Less Wrong Q&A with Eliezer Yudkowsky: Ask Your Questions

Test Your Calibration!

Anti-Akrasia Technique: Structured Procrastination

Boston meetup Nov 15 (and others)

Consequences of arbitrage: expected cash

Auckland meet up Saturday Nov 28th

The Academic Epistemology Cross Section: Who Cares More About Status?

BHTV: Yudkowsky / Robert Greene

Why (and why not) Bayesian Updating?

Efficient prestige hypothesis

A Less Wrong singularity article?

Request For Article: Many-Worlds Quantum Computing

The One That Isn't There

Calibration for continuous quantities

Friedman on Utility

Rational lies

In conclusion: in the land beyond money pumps lie extreme events

How to test your mental performance at the moment?

Agree, Retort, or Ignore? A Post From the Future

Contrarianism and reference class forecasting

Getting Feedback by Restricting Content

Rooting Hard for Overpriced M&Ms

A Nightmare for Eliezer

Rationality Quotes November 2009

Morality and International Humanitarian Law

Action vs. inaction

The Moral Status of Independent Identical Copies

Call for new SIAI Visiting Fellows, on a rolling basis

Open Thread: December 2009

The Difference Between Utility and Utility

11 core rationalist skills

Help Roko become a better rationalist!

Intuitive supergoal uncertainty

Frequentist Statistics are Frequently Subjective

Arbitrage of prediction markets

Parapsychology: the control group for science

Science - Idealistic Versus Signaling

You Be the Jury: Survey on a Current Event

Probability Space & Aumann Agreement

What Are Probabilities, Anyway?

The persuasive power of false confessions

A question of rationality

The Amanda Knox Test: How an Hour on the Internet Beats a Year in the Courtroom

Against picking up pennies

Previous Post Revised

Man-with-a-hammer syndrome

Rebasing Ethics

Getting Over Dust Theory

Philadelphia LessWrong Meetup, December 16th

An account of what I believe to be inconsistent behavior on the part of our editor

December 2009 Meta Thread

Reacting to Inadequate Data

The Contrarian Status Catch-22

Any sufficiently advanced wisdom is indistinguishable from bullshit

Fundamentally Flawed, or Fast and Frugal?

Sufficiently Advanced Sanity

Mandating Information Disclosure vs. Banning Deceptive Contract Terms

If reason told you to jump off a cliff, would you do it?

The Correct Contrarian Cluster

Karma Changes

lessmeta

The 9/11 Meta-Truther Conspiracy Theory

Two Truths and a Lie

On the Power of Intelligence and Rationality

Are these cognitive biases, biases?

Positive-affect-day-Schelling-point-mas Meetup

Playing the Meta-game

A Master-Slave Model of Human Preferences

That other kind of status

Singularity Institute $100K Challenge Grant / 2009 Donations Reminder

Boksops -- Ancient Superintelligence?

New Year’s Resolutions Thread

New Year's Predictions Thread

End of 2009 articles