LESSWRONG
LW

69
Daniel Kokotajlo
30170Ω395611034880
Message
Dialogue
Subscribe

Was a philosophy PhD student, left to work at AI Impacts, then Center on Long-Term Risk, then OpenAI. Quit OpenAI due to losing confidence that it would behave responsibly around the time of AGI. Now executive director of the AI Futures Project. I subscribe to Crocker's Rules and am especially interested to hear unsolicited constructive criticism. http://sl4.org/crocker.html

Some of my favorite memes:


(by Rob Wiblin)

Comic. Megan & Cueball show White Hat a graph of a line going up, not yet at, but heading towards, a threshold labelled "BAD". White Hat: "So things will be bad?" Megan: "Unless someone stops it." White Hat: "Will someone do that?" Megan: "We don't know, that's why we're showing you." White Hat: "Well, let me know if that happens!" Megan: "Based on this conversation, it already has."
(xkcd)

My EA Journey, depicted on the whiteboard at CLR:

(h/t Scott Alexander)


 
Alex Blechman @AlexBlechman Sci-Fi Author: In my book I invented the Torment Nexus as a cautionary tale Tech Company: At long last, we have created the Torment Nexus from classic sci-fi novel Don't Create The Torment Nexus 5:49 PM Nov 8, 2021. Twitter Web App

Sequences

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
No wikitag contributions to display.
Agency: What it is and why it matters
AI Timelines
Takeoff and Takeover in the Past and Future
The Takeoff Speeds Model Predicts We May Be Entering Crunch Time
Daniel Kokotajlo4d50

For my part I just somehow never saw this. Dunno why. Maybe I was busy and didn't check LW for a while idk.

Reply
The Takeoff Speeds Model Predicts We May Be Entering Crunch Time
Daniel Kokotajlo4d42

Wow somehow this article never came to my attention, sorry I missed it, thanks Eli Lifland for sending it to me!

Reply
I Read Red Heart and I Heart It
Daniel Kokotajlo4d52

OK, thanks!

FWIW I do think that superintelligent AI, when it eventually exists, will be more analogous to a superhuman god doing research beyond the comprehension of mere humans, than to a difficult enemy that could in theory be beaten by the main character. Like, yes, in theory it could be beaten, in the same sense that in theory I could beat Magnus Carlsen in chess or in theory Cuba could invade and conquer the USA.

Reply
I Read Red Heart and I Heart It
Daniel Kokotajlo4d70

I've bought it and plan to read it, you are the second person to recommend it to me recently.

Curious about how you think it's more realistic than AI 2027? AI 2027 doesn't really feature any superpersuasion, much less intelligence-as-mind-control. It does have bioweapons but they aren't at all important to the plot. As for boiling the oceans... I mean I think that'll happen in the next decade or two unless the powers that be decide to regulate economic growth to prevent it, which they might or might not.

Reply
GradientDissenter's Shortform
Daniel Kokotajlo5d92

Huh. Seems super wrong to me fwiw. How would you rank AIFP on the list? 

Reply
The Main Sources of AI Risk?
Daniel Kokotajlo5d20

No, alas. :(  iirc I had a half-finished redraft and then ended up prioritizing other things.

Reply
GradientDissenter's Shortform
Daniel Kokotajlo6d120

For context, how would you rank the AI safety community w.r.t. integrity and honor, compared to the following groups:

1. AGI companies
2. Mainstream political parties (the organizations, not the voters, so e.g. the politicians and their staff)
3. Mainstream political movements e.g. neoliberalism, wokism, china hawks, BLM, 
4. A typical university department
5. Elite opinion formers (e.g. the kind of people whose Substacks and op-eds are widely read and highly influential in DC, silicon valley, etc.)
6. A typical startup
7. A typical large bloated bureaucracy or corporation
8. A typical religion e.g. christianity, islam, etc.
9. The US military


 

Reply5
Ontology for AI Cults and Cyborg Egregores
Daniel Kokotajlo6dΩ362

This ontology allows clearer and more nuanced understanding of what's going on and dispels some confusions. 


The ontology seems good to me, but what confusions is it dispelling? I'm out of the loop here.

Reply
Myopia Mythology
Daniel Kokotajlo6dΩ8115

IIUC the core of this post is the following:

There are three forces/reasons pushing very long-sighted ambitious agents/computations to make use of very short-sighted, unambitious agents/computations:


1. Schelling points for acausal coordination

2. Epistemology works best if you just myopically focus on answering correctly whatever question is in front of you, rather than e.g. trying to optimize your long-run average correctness or something.

3. Short-sighted, unambitious agents/computations are less of a threat, more easily controlled. 

I'll ignore 1 for now. For 2 and 3, I think I understand what you are saying on a vibes/intuitive level, but I don't trust my vibes/intuition enough. I'd like to see 2 and 3 spelled out and justified more rigorously. IIRC there's an academic philosophy literature on 2, but I'm afraid I don't remember it much. Are you familiar with it? As for 3, here's a counterpoint: Yes, more longsighted agents/computations carry with them risk of various kinds of Evil. But, the super myopic ones also have their drawbacks. (Example: Sometimes you'd rather have a bureaucracy staffed with somewhat agentic people who can make exceptions to the rules when needed, than a bureaucracy staffed with apathetic rule-followers.) You haven't really argued that the cost-benefit analysis systematically favors delegating to myopic agents/computations.

Reply
Load More
10Daniel Kokotajlo's Shortform
Ω
6y
Ω
928
120Vitalik's Response to AI 2027
4mo
53
177My pitch for the AI Village
5mo
32
99METR's Observations of Reward Hacking in Recent Frontier Models
5mo
9
140Training AGI in Secret would be Unsafe and Unethical
7mo
15
669AI 2027: What Superintelligence Looks Like
Ω
7mo
Ω
222
183OpenAI: Detecting misbehavior in frontier reasoning models
Ω
8mo
Ω
26
88What goals will AIs have? A list of hypotheses
Ω
8mo
Ω
20
36Extended analogy between humans, corporations, and AIs.
Ω
9mo
Ω
2
161Why Don't We Just... Shoggoth+Face+Paraphraser?
Ω
1y
Ω
58
70Self-Awareness: Taxonomy and eval suite proposal
Ω
2y
Ω
2
Load More