Today's post, Failure By Affective Analogy was originally published on 18 November 2008. A summary (taken from the LW wiki):

 

Making analogies to things that have positive or negative connotations is an even better way to make sure you fail.


Discuss the post here (rather than in the comments to the original post).

This post is part of the Rerunning the Sequences series, where we'll be going through Eliezer Yudkowsky's old posts in order so that people who are interested can (re-)read and discuss them. The previous post was Failure By Analogy, and you can use the sequence_reruns tag or rss feed to follow the rest of the series.

Sequence reruns are a community-driven effort. You can participate by re-reading the sequence post, discussing it here, posting the next day's sequence reruns post, or summarizing forthcoming articles on the wiki. Go here for more details, or to have meta discussions about the Rerunning the Sequences series.

New Comment
1 comment, sorted by Click to highlight new comments since: Today at 5:42 AM

For this much I do know - if I don't know that my AI design will work, it won't.

Well, that's obviously not so - the design isn't changed when you complete your proof. I suppose this should be taken to mean, "If I don't know that my AI design will work, the chances of its working are low enough to be negligible"?