"I mean... if an external objective morality tells you to kill babies, why should you even listen?"
This is an incredibly dangerous argument. Consider this : "I mean... if some moral argument, whatever the source, tells me to prefer 50 years of torture to any number of dust specks, why should I even listen?"
And we have seen many who literally made this argument.
Maybe they are right.
People have been demonstrably willing to make everyone live at a lower standard of living rather than let a tiny minority grow obscenely rich and everyone else be moderately well off. In other words we seem to be willing to pay a price for equality. Why wouldn't this work in the other direction? Maybe we prefer to induce more suffering overall if this prevents a tiny minority suffering obscenely.
Too many people seem to think perfectly equally weighed altruism (everyone who shares the mystical designation of "person" has a equal weight and after that you just do calculus to maximize overall "goodness") that sometimes hides under the word "utilitarianism" on this forum, is anything but another grand moral principle that claims to, but fails, to really compactly represent our shards of desire. If you wouldn't be comfortable building an AI to follow that rule and only that rule, why are so many people keen on solving all their personal moral dilemmas with it?
You do realize that valuing equality in itself to any extent at all is always (because of opportunity cost at least) an example of this:
Are you sure?
If you take a concave function, such as a log, of the net happiness of each individual, and maximize the sum, you'd always prefer equality to inequality when net happiness is held constant, and you'd always prefer a higher minimum happiness regardless of inequality.
Personally, I don't know what morality is, or what's the 'inherently the right thing to do'. For me, the situation is simple.
If I hurt someone, my mirror neurons will hurt me. If I hurt someone's baby, I'll experience the pain I inflicted upon the baby, plus the pain of the parents, plus the pain of everyone who heard about this story and felt the pain thanks, in turn, to their mirror neurons.
And I'll re-experience all this pain in the future, every time I remember the episode -- unless I invent some way to desensitize myself to this memory.
I'm a meat mach...
I'm pretty sure you're doing it wrong here.
"What if the structure of the universe says to do something horrible? What would you have wished for the external objective morality to be instead?" Horrible? Wish? That's certainly not according to objective morality, since we've just read the tablet. It's just according to our intuitions. I have an intuition that says "Pain is bad". If the stone tablet says "Pain in good", I'm not going to rebel against it, I'm going to call my intuition wrong, like "Killing is good", &qu...
Vladimir, if there was a pill that would make the function of the mirror neurons go away, in other words, a pill that would make you able to hurt people without feeling remorse or anguish, would you take it?
@IL: Would I modify my own source code if I were able to? In this particular case, no, I wouldn't take the pill.
I don't believe in the existence of morals, which is to say there is no "right" or "wrong" in the universe. However, I'll still do actions that most people would rate "moral". The reasons I do this are found in my brain architecture, and are not simple. Also, I don't care about utilitarianism. One can probably find some extremely complex utility function that describes my actions, which makes everybody on earth a utilitarianist, but I don't consciously make utility calculations. On the other hand, if morality is defined as "the way people make decisions", then of course everybody is moral and morality exists.
It is the courage of a theist who goes against what they believe to be the Will of God, choosing eternal damnation and defying even morality in order to rescue a slave, or speak out against hell, or kill a murderer...I once read in some book about members of the Inquisition who thought that their actions - like torture and murder - might preclude them from going to heaven. But these people where so selflessly moral that they gave up their own place in heaven for saving the souls of the witches... great, isn't it?
If you believe that there is any kind of stone tablet in the fabric of the universe, in the nature of reality, in the structure of logic - anywhere you care to put it - then what if you get a chance to read that stone tablet, and it turns out to say "Pain Is Good"? What then?
Well, Eliezer, since I can't say it as eloquently as you:
"Embrace reality. Hug it tight."
"It is always best to think of reality as perfectly normal. Since the beginning, not one unusual thing has ever happened."
If we find that Stone Tablet, we adjust our model accordingly.
Eliezer: "Go ahead. Indulge your fantasy. Would you want the stone tablet to say people should die of old age, or that people should live as long as they wanted? If you could write the stone tablet yourself, what would it say?"
Excellent way of putting it... I would certainly want the option of living as long as I liked. (Though I find it worth noting that when I was depressed, I found the idea of needing to choose when to end program abhorrent, since I figured I could go several billion years in agony before making such a choice... Many people...
Vladimir, why not? From reading your comment, it seems like the only reason you don't hurt other people is because you will get hurt by it, so if you would take the pill, you would be able to hurt other people. Have I got it wrong? Is this really the only reason you don't hurt people?
The nice thing with believing in no objective morality is that you needn't to solve such poorly intelligible questions. I hope Eliezer is trying to demonstrate the absurdity of believing in objective morality, if so, then good luck!
"I mean... if an external objective morality tells you to kill babies, why should you even listen?" - this is perhaps a dangerous question, but still I like it. Why should you do what you should do? Or put differently, what is the meaning of "should"?
If everything I do and believe is a consequence of the structure of the universe, then what does it mean to say my morality is/isn't built into the structure of the universe? What's the distinction? As far as I'm concerned, I am (part of) the structure of the universe.
Also, regarding the previous post, what does it mean to say that nothing is right? It's like if you said, "Imagine if I proved to you that nothing is actually yellow. How would you proceed?" It's a bizarre question because yellowness is something that is in the mind anyway. There is simply no fact of the matter as to whether yellowness exists or not.
Eliezer, Your post is entirely consistent with what I said to Robin in my comments on "Morality Is Overrated": Morality is a means, not an end.
"...if there was a pill that would make the function of the mirror neurons go away, in other words, a pill that would make you able to hurt people without feeling remorse or anguish, would you take it?"
The mirror neurons also help you learn from watching other humans. They help you intuit the feelings of others which makes social prediction possible. They help communication. They also allow you to share in the joy and pleasure of others...e.g., a young child playing in a park.
I would like more control over how my mind functions. At times it would...
See: Good, Evil, Morality, and Ethics: "What would it mean to want to be moral (to do the moral thing) purely for the sake of morality itself, rather than for the sake of something else? What could this possibly mean to a scientific materialistic atheist? What is this abstract, independent, pure morality? Where does it come from? How can we know it? I think we must conclude that morality is a means, not an end in itself."
I think we must conclude that morality is a means, not an end in itself.
Morality is commonly thought of neither as a means nor as an end, but as a constraint. This view is potentially liberating, because the conception of morality as a means to an end implies the idea that any two possible actions can be compared to see which is the best means to the end and therefore which is the most moral. To choose the less moral of the two choices is, on this conception, the very definition of immoral. Thus on this conception, our lives are in principle mapped out for...
This is horrible, this is non-rational. You are telling us to trust our feelings, after this blog has shown us that our feelings think it's just as good to rescue ten men as a million? What is your command to "shut up and multiply", but an off switch for my morality that replaces it with math?
If it were inherently right to kill babies, I would hope I had the moral courage to do the right thing.
This is horrible, this is non-rational. You are telling us to trust our feelings, after this blog has shown us that our feelings think it's just as good to rescue ten men as a million? What is your command to "shut up and multiply", but an off switch for my morality that replaces it with math?I just wish Eliezer would take his own advice. But for some reason he seems quite unwilling to show us the mathematical demonstration of the validity of his opinions, and instead of doing the math he persists with talking.
"Maybe you should just do that?"
Heck, hell with physics too. Let's just make up all human knowledge. If we're going to invent the prescriptive, why not the descriptive too?
Eliezer,
Every time I think you're about to say something terribly naive, you surprise me. It looks like trying to design an AI morality is a good way to rid oneself of anthropomorphic notions of objective morality, and to try and see where to go from there.
Although I have to say the potshot at Nietzsche misses the mark; his philosophy is not a resignation to meaninglessness, but an investigation of how to go on and live a human or better-than-human life once the moral void has been recognized. I can't really explicate or defend him in such a short remark...
@IL: Of course, "I just feel that hurting living things is bad" sums the inner perspective quite well, but this isn't really an answer to the question why exactly hurting living things feels bad for me, and why I wouldn't take the pill that shuts down my mirror neurons.
By taking the pill, I create a people-hurter, a thing-that-hurts-people, which is undoubtedly a bad thing to do judging from the before-the-pill POV. It's not that different from pressing a button that says "pressing this button will result in a random person being hurt or killed every day for 40 years since this moment".
Doesn't the use of the word 'how' in the question "If "yes", how inherently right would it have to be, for how many babies?" presuppose that the person answering the question believes that the 'inherent rightness' of an act is measurable on some kind of graduated scale? If that's the case, wouldn't assigning a particular 'inherent rightness' to an act be, by definition, the result of a several calculations?
What I mean is, if you've 'finished' calculating, and have determined that killing the babies is a morally justifiable (and/or nece...
"Would you kill babies if it was inherently the right thing to do? Yes [] No []"
-->
"Imagine that you wake up one morning and your left arm has been replaced by a blue tentacle. The blue tentacle obeys your motor commands - you can use it to pick up glasses, drive a car, etc. ... How would I explain the event of my left arm being replaced by a blue tentacle? The answer is that I wouldn't. It isn't going to happen. "
If morality was objective and it said we should kill babies, we'd have, and likely want to do it. Appears it isn't objective, though, and that we just don't feel that way. Another question ?
What I'm wondering, in other words is this: Is our reluctance to carry out an act that we may have judged to be morally justifiable a symptom that the decision-making software we think we're running is not the software we're actually running?
I admit that I own no great familiarity with the works of Nietzsche - I've read only one or two things and that turned me off the rest - so I've edited the main article accordingly.
"Torture is a relative morality, as such, when a subculture like an intelligence agency tortures a terrorist, then it is allowed and it is moral. Any moral 'critique' of the torture is tantamount to a universal moralist rule: Torture is universally bad."
Torture is universally bad, with the exception of imperatives which are heirarchally superior.
"On the other hand, if morality is defined as "the way people make decisions", then of course everybody is moral and morality exists."
It's more like "the way people ought to make ...
I'm with Manon and Nominull: if, somehow, I actually believed such a Tablet existed, I hope I would overwrite my own moral intuitions with it, even if it meant killing babies. Not that I believe the Tablet is any more likely or coherent than fundamental apples - why should I listen, indeed? - although my volition extrapolating to something inhuman is.
The idea that there is no right and wrong is simply laughable.
The idea that our culturally inculcated senses of right and wrong have no objective basis is about as shocking to me as the idea that fashion has no objective basis. Oh no! However will we determine whether hemlines should be high or low next season? The topic itself has no interest for me, and even if it did, the idea simply wouldn't have anything to do with any of my opinions on it.
The sounds of words usually have no objective connection to the things they describe, either. The words are basically arbitrary. Oh, existential horror!
I mean, really - to be upset about these sorts of ideas, you have to be almost terminally naive.
For me, these questions create a tangle of conflicts between the real and the hypothetical. This is my best attempt to untangle, so far. First, if there were a tablet that could actually somehow be shown to reveal objective morality, I suspect that I might never have had any qualms about committing atrocities in the first place, since I would be steeped in a culture that unanimously approved. We already see this in the real world, merely as a result of controversial tablets that only some agree on! If you mean, what if I suddenly discovered the tablet just...
Caledonian: 1) Why is it laughable? 2) If hemlines mattered to you as badly as a moral dilemma, would you still hold this view?
Or, you have to want more justification than is really necessary or possible, which is quite understandable when it comes to fundamental values.
What is inclusive genetic fitness? Is it the same as inclusive fitness as defined on wikipedia?
What if you build a super-intelligent AI and you are convinced that it is Friendly, and it tells you to do something like this? Go kill such-and-such a baby, and you will massively increase the future happiness of the human race. You argue and ask if there isn't some other way to do it, and the FAI explains that every other alternative will involve much greater human suffering. Killing a baby is relatively humane, as newborn babies have only limited consciousness, and their experiences are not remembered anyway. You will kill the baby instantly and painles...
Hal Finney-
I probably wouldn't have argued that much with the AI... I've done things I've personally found more morally questionable since I didn't have quite as good a reason to believe I was right about the outcome... Moral luck, I was.
Hal: as an amoralist, I wouldn't do it. If there is not enough time to explain to me why it is necessary and convince me that it is necessary, no deal. Even if I thought it probably would substantially increase the future happiness of humanity, I still wouldn't do it without a complete explanation. Not because I think there is a moral fabric to the universe that says killing babies is wrong, but because I am hardwired to have an extremely strong aversion to like killing babies. Even if I actually was convinced that it would increase happiness, I still migh...
Hal: I wouldn't do it, nor do I think I'd want to live in a world governed thusly. My reasoning is that it violates individual liberty and self-possession. It seems to imply that individuals are somehow the "eminent domain", as it were, of society. I reject that. I say that nobody has the right to spend the baby's life. Granted, this is more of a political stance than a moral one. I can't claim that there's an objective reason to value individual rights so highly, but it is a fact that I do. I know you said the baby wouldn't suffer, but this question still put me in mind of the idea that pain and happiness may not be the same currency. It may not be valid to try to offer suffering as a payment for happiness.
Andy: "I can't claim that there's an objective reason to value individual rights so highly, but it is a fact that I do."
Hal: "You argue and ask if there isn't some other way to do it, and the FAI explains that every other alternative will involve much greater human suffering."
These things seem grossly disproportionate. Do you really believe utility(individual rights of one person)>>>utility(end great human suffering)
Andy- A man who is on the brink the death has a key to a safe deposit box in which there is an asthma inhaler. ...
Reading this thread has been fascinating. I'm perhaps naive & simplistic in my thinking but here are some of my thoughts.
There's no particular need to renew the torture and dust specks debate, so I'll just point out that GBM, Nominull, Ian C., and Manon de Gaillande have all made similar points: if you say, "if there is an external objective morality that says you should kill babies, why should you listen?" the question is the same as "if you should kill babies, why should you do it?"
Yes, and if 2 and 2 make 5, why should I admit it?
It isn't in fact true that I should kill babies, just as 2 and 2 don't make 5. But if I found out that 2 and 2 do make 5, of...
Laura: Yes, I absolutely steal the key. Given the context of the original question, I had in mind the right to life, in particular. I didn't make this distinction until you asked this question. I happen not to think that the right to property is anything like as valuable as the right to life. (By "right" I mean nothing more than ground rules that society has "agreed" on.) Again, I have a problem with acting as though an individual's life is the eminent domain of society. As in Shirley Jackson's "The Lottery," the picture looks...
Andy- I agree with your skepticism. I was taking for granted that the AI in the scenario was correct in its calculation, since I am 'convinced that it is friendly' but yes, I would need to be pretty fucking sure it really was both friendly and able to perform such calculations before I would kill anyone at its command.
What's 'objective' about morality doesn't take the form of moral commandments aka 'the 10 commandments', nor does it take the form of an optimization function that produces the commandments either.
There's a thrid possibility, one you've over-looked, that is, in fact, the objective compoenent of morality: namely purely abstract archetypes or moral ideals (ie beauty, freedom, virtue). These objective platonic abstractions are not in the form of commandments, and they're not optimization functions either. The objective component of morality built into the universe doesn't tell me to do anything. It's just a lot of abstract archetypes.
Assuming that we evolved in the moral climate that you are constructing I would guess that we would readily kill babies. Now of course, in the example you give there is an inherent limit to the number of babies that can be killed and still have sufficient life left over to be around to respond to your questions.
The spectrum of responses and moralities I've seen on display here (and elsewhere) are artifacts of our being and culture. Many of the behavioral tendencies that we ascribe as being "moral" have both an innate ("instinctual" fo...
Hmm... This whole baby-killing example is making me think...
Knecht: "Even if I thought it probably would substantially increase the future happiness of humanity, I still wouldn't do it without a complete explanation. Not because I think there is a moral fabric to the universe that says killing babies is wrong, but because I am hardwired to have an extremely strong aversion to like killing babies."
This does seem like what a true amoralist might say... yet, what if the idea of having forgone the opportunity to substantially increase the future hap...
The Greeks really did get it all right.No, they were simply less wrong than most on a limited number of memorable topics.
Laura ABJ: To expand on the text you quoted, I think that killing babies is ugly, and therefore would not do it without sufficient reason, which I don't think the scenario provides. The ugliness of killing babies doesn't need a moral explanation, and the moral explanation just builds on (and adds nothing but a more convenient way of speaking about) the foundation of aversion, no matter how it's dressed up and made to look like something else.
The idea is not compelling to me and so would not haunt me forever, because like I said, I'm not yet convinced that ...
I realize that just because I am fairly confident I wouldn't suffer terribly from killing the baby if my knowledge was fairly complete, I can't say that for all people. People's utility functions differ, as do their biological and learned aversions to certain types of violence. The cognitive dissonance created by being presented with such a situation might be too great for some, causing them to break down psychologically and rationalize their way out of the decision any way they could. What if we upped the stakes and took it from some anonymous baby pai...
Hal Finney:
Why doesn't the AI do it verself? Even if it's boxed (and why would it be, if I'm convinced it's an FAI?), at the intelligence it'd need to make the stated prediction with any degree of confidence, I'd expect it to be able to take over my mind quickly. If what it claims is correct, it shouldn't have any qualms about doing that (taking over one human's body for a few minutes is a small price to pay for the utility involved).
If this happened in practice I'd be confused as heck, and the alleged FAI being honest about its intentions would be prett...
By gum, I'm amazed that fifty comments have gone by and nobody's mentioned future toddler chopper Vox Day. Sure, it's nearly a year and a half old, but if anyone had doubt that there are apparently functioning humans out there who would tick the second box and fill in "until my arm got tired".
The Euthyphro hypothetical does remind me a bit of the Ticking Time Bomb--a thoroughly unrealistic situation designed to cause the quiz-taker to draw a conclusion about more realistic situations that they wouldn't have come to otherwise.
If there were no such thing as green, what color would green things be? MU.
Kierkegaard talked about all this in Fear and Trembling a long time ago. Was Abraham sacrificing Isaac immoral? If you call morality the universal of that time, then you have to suppose it was. But in the story we know that God told Abraham to do it.
Anyone who defies the universal in favor of their own experience of God, the universal has no way to judge. No argument on the basis of the universal morality can call Kierkegaard's knight of faith back from the quest.
"And if an ext...
Would you kill babies if it was inherently the right thing to do?
If it were inherently the right thing to do, then I wouldn't be here. Someone would have killed me when I was a baby.
There is no morality, it is a fiction to be discarded alongside god and rights. What informs are actions is the trifecta of self-interest, emotion, and social expectation. Our upbringing and later education shapes which of these is given more weight when we make our decisions.
There simply is no moral property to an action or consequence. There is no natural property that is moral. There is no discoverable law or property that can inform an ought. Our "ethical intuitions" are simply emotional responses. No one can say "killing is wrong&...
I don't get it. If killing babies was inherently good, I would kill them, sure. It's not like killing babies is inherently bad.
Or did you think that I thought so?
I understand that in many usual contexts killing babies would seem bad to me, because I was given instructions to take care of babies (generally) by evolution, only because having these instructions made it more likely for me to exist and have those instructions. So what? Is existing and having instructions inherently good?
In general, the Socratic questions in this sequence don't seem to work for me.. is this because I'm not answering in a way I was expected to?
If I AM utility function maximizer and I proved that killing baby reliably maximizes it, then sure I'll kill.
But I am not. My poorly defined unclosurable morality meter will break in and demand revising that nice and consistent utility function I've based my decision on. And so moral agonising begins.
Answer: I don't know, and it will be a painful work to decide, weighting all pro and cons, building and checking new utility functions, rewriting moral itself...
So... the correct answer is to dissolve the question, yes?
I like to think of this as being extreme artificiality. Humans have always attempted to either ignore or go against certain natural elements in order to flourish. It was never this fundamental, though. Logic has, at best, managed to straighten us out and make things better for us. And at worst, it reaches conclusions that are of no practical consequence. If it ever told us that killing babies is good, we would of course have to check all the consequences of what it would mean to ignore this logic. If we get lucky, it’s a logic that doesn’t really extend ve...
If it were revealed to me that, say, the Aztecs were right, their gods are real, and the One True Religion, then I believe it would be my duty to defy their will, and reject their plan for mankind. Power does not grant moral authority, even if it is the power that was used to make the world as it is.
Would I be brave enough to do it in practice? I have no idea, but I think it helps that I'm thinking about it beforehand.
What would you have wished for the external objective morality to be instead? What's the best news you could have gotten, reading that stone tablet?
That's an awesome question. I'm going to have to steal that one.
I find it funny that many of the people here who where pretty much freaked out by the idea of "objective morality built into the fabric of the universe" not really mattering for humans, yet when it comes to mythology don't have a problem criticizing Abraham for being willing to sacrifice his son because God told him too.
Leon Kass (of the President's Council on Bioethics) is glad to murder people so long as it's "natural", for example. He wouldn't pull out a gun and shoot you, but he wants you to die of old age and he'd be happy to pass legislation to ensure it.
Does anyone have sources to support this conclusion about Kass's views? I tracked down a transcript of an interview he gave that was cited on a longevity website, but it doesn't support that characterization at all. He does express concerns about greatly increased lifespans, but makes clear that he see...
Go ahead. Indulge your fantasy. Would you want the stone tablet to say people should die of old age, or that people should live as long as they wanted? If you could write the stone tablet yourself, what would it say?
I'm reminded of one of Bill Watterson's Calvin and Hobbes strips:
...Calvin: I'm at peace with the world. I'm completely serene.
Hobbes: Why is that?
Calvin: I've discovered my purpose in life. I know why I was put here and why everything exists.
Hobbes: Oh really?
Calvin: Yes. I am here so everyone can do what I want.
Hobbes: (rolling eyes) It'
What if the structure of the universe says to do something horrible?
If the "structure of universe" is something mathematical (e.g. the prime number theorem) then it's meaningless to ask "what if the structure of the universe says X" unless it truly says X. Assuming it says something different from what it really says immediately leads to a logical contradiction which allows deducing anything at all
If you could write the stone tablet yourself, what would it say?
You're suggesting that we should trust our moral intuition instead of ...
Responding to old post:
In 1966, the Israeli psychologist Georges Tamarin presented, to 1,066 schoolchildren ages 8-14, the Biblical story of Joshua's battle in Jericho:
If you ask a question to schoolchildren, you have to take into consideration that children are supposed to obey authority figures. And not only because the authority figures have power, but because children don't know and can't comprehend many important things about the world, and that makes it a good idea for children to put little weight on their own conclusions and a lot of weight on...
Followup to: What Would You Do Without Morality?, Something to Protect
Once, discussing "horrible job interview questions" to ask candidates for a Friendly AI project, I suggested the following:
Yesterday I asked, "What would you do without morality?" There were numerous objections to the question, as well there should have been. Nonetheless there is more than one kind of person who can benefit from being asked this question. Let's say someone gravely declares, of some moral dilemma—say, a young man in Vichy France who must choose between caring for his mother and fighting for the Resistance—that there is no moral answer; both options are wrong and blamable; whoever faces the dilemma has had poor moral luck. Fine, let's suppose this is the case: then when you cannot be innocent, justified, or praiseworthy, what will you choose anyway?
Many interesting answers were given to my question, "What would you do without morality?". But one kind of answer was notable by its absence:
No one said, "I would ask what kind of behavior pattern was likely to maximize my inclusive genetic fitness, and execute that." Some misguided folk, not understanding evolutionary psychology, think that this must logically be the sum of morality. But if there is no morality, there's no reason to do such a thing—if it's not "moral", why bother?
You can probably see yourself pulling children off train tracks, even if it were not justified. But maximizing inclusive genetic fitness? If this isn't moral, why bother? Who does it help? It wouldn't even be much fun, all those egg or sperm donations.
And this is something you could say of most philosophies that have morality as a great light in the sky that shines from outside people. (To paraphrase Terry Pratchett.) If you believe that the meaning of life is to play non-zero-sum games because this is a trend built into the very universe itself...
Well, you might want to follow the corresponding ritual of reasoning about "the global trend of the universe" and implementing the result, so long as you believe it to be moral. But if you suppose that the light is switched off, so that the global trends of the universe are no longer moral, then why bother caring about "the global trend of the universe" in your decisions? If it's not right, that is.
Whereas if there were a child stuck on the train tracks, you'd probably drag the kid off even if there were no moral justification for doing so.
In 1966, the Israeli psychologist Georges Tamarin presented, to 1,066 schoolchildren ages 8-14, the Biblical story of Joshua's battle in Jericho:
After being presented with the Joshua story, the children were asked:
66% of the children approved, 8% partially disapproved, and 26% totally disapproved of Joshua's actions.
A control group of 168 children was presented with an isomorphic story about "General Lin" and a "Chinese Kingdom 3,000 years ago". 7% of this group approved, 18% partially disapproved, and 75% completely disapproved of General Lin.
"What a horrible thing it is, teaching religion to children," you say, "giving them an off-switch for their morality that can be flipped just by saying the word 'God'." Indeed one of the saddest aspects of the whole religious fiasco is just how little it takes to flip people's moral off-switches. As Hobbes once said, "I don't know what's worse, the fact that everyone's got a price, or the fact that their price is so low." You can give people a book, and tell them God wrote it, and that's enough to switch off their moralities; God doesn't even have to tell them in person.
But are you sure you don't have a similar off-switch yourself? They flip so easily—you might not even notice it happening.
Leon Kass (of the President's Council on Bioethics) is glad to murder people so long as it's "natural", for example. He wouldn't pull out a gun and shoot you, but he wants you to die of old age and he'd be happy to pass legislation to ensure it.
And one of the non-obvious possibilities for such an off-switch, is "morality".
If you do happen to think that there is a source of morality beyond human beings... and I hear from quite a lot of people who are happy to rhapsodize on how Their-Favorite-Morality is built into the very fabric of the universe... then what if that morality tells you to kill people?
If you believe that there is any kind of stone tablet in the fabric of the universe, in the nature of reality, in the structure of logic—anywhere you care to put it—then what if you get a chance to read that stone tablet, and it turns out to say "Pain Is Good"? What then?
Maybe you should hope that morality isn't written into the structure of the universe. What if the structure of the universe says to do something horrible?
And if an external objective morality does say that the universe should occupy some horrifying state... let's not even ask what you're going to do about that. No, instead I ask: What would you have wished for the external objective morality to be instead? What's the best news you could have gotten, reading that stone tablet?
Go ahead. Indulge your fantasy. Would you want the stone tablet to say people should die of old age, or that people should live as long as they wanted? If you could write the stone tablet yourself, what would it say?
Maybe you should just do that?
I mean... if an external objective morality tells you to kill people, why should you even listen?
There is a courage that goes beyond even an atheist sacrificing their life and their hope of immortality. It is the courage of a theist who goes against what they believe to be the Will of God, choosing eternal damnation and defying even morality in order to rescue a slave, or speak out against hell, or kill a murderer... You don't get a chance to reveal that virtue without making fundamental mistakes about how the universe works, so it is not something to which a rationalist should aspire. But it warms my heart that humans are capable of it.
I have previously spoken of how, to achieve rationality, it is necessary to have some purpose so desperately important to you as to be more important than "rationality", so that you will not choose "rationality" over success.
To learn the Way, you must be able to unlearn the Way; so you must be able to give up the Way; so there must be something dearer to you than the Way. This is so in questions of truth, and in questions of strategy, and also in questions of morality.
The "moral void" of which this post is titled, is not the terrifying abyss of utter meaningless. Which for a bottomless pit is surprisingly shallow; what are you supposed to do about it besides wearing black makeup?
No. The void I'm talking about is a virtue which is nameless.
Part of The Metaethics Sequence
Next post: "Created Already In Motion"
Previous post: "What Would You Do Without Morality?"