LESSWRONG
LW

Gurkenglas
2516Ω8213126842
Message
Dialogue
Subscribe

I operate by Crocker's rules.

I try to not make people regret telling me things. So in particular:
- I expect to be safe to ask if your post would give AI labs dangerous ideas.
- If you worry I'll produce such posts, I'll try to keep your worry from making them more likely even if I disagree. Not thinking there will be easier if you don't spell it out in the initial contact.

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
Comparison of decision theories (with a focus on logical-counterfactual decision theories)
Gurkenglas6d20
UDT(s)=argmaxfn∑i=1U(Oi)⋅P(┌UDT:=f:s↦a┐□→Oi)

The left hand side of the equation has type action (Hintze page 4: "An agent’s decision procedure takes sense data and outputs an action."), but the right hand side has type policy, right?

Reply
Comparison of decision theories (with a focus on logical-counterfactual decision theories)
Gurkenglas7d20

(FDT(P,x))(x)

Should this be FDT(P,x)? As is this looks to me like the second (x) introduces x into scope, and the first x is an out-of-scope usage.

Reply
Critiques of FDT Often Stem From Confusion About Newcomblike Problems
Gurkenglas10d30

Let me try again:

Does the note say that I was predicted to choose the right box regardless of what notes I am shown, and therefore the left box contains a bomb? Then the predictor is malfunctioning and I should pick the right box.

Does the note say that I was predicted to choose the right box when told that the left box contains a bomb, and therefore the left box contains a bomb? Then I should pick the left box, to shape what I am predicted to do when given that note.

Reply
Critiques of FDT Often Stem From Confusion About Newcomblike Problems
Gurkenglas10d20

You'll also need to update the content of the note and the predictor's decision process to take into account that the agent may see a note. In particular, the predictor needs to decide whether to show a note in the simulation, and may need to run multiple simulations.

Reply
With enough knowledge, any conscious agent acts morally
Gurkenglas10dΩ131

Let's sharpen A6. Consider this stamp collector construction: It sends and receives internet data, it has a magically accurate model of reality, it calculates how many stamps would result from each sequence of outputs, and then it outputs the one that results in the most stamps.

By definition it knows everything about reality, including any facts about what is morally correct, and that stamps are not particularly morally important. It knows how to self-modify, and how many stamps any such self-modification will result in.

I'd like to hear how this construction fares as we feed it through your proof. I think it gums up the section "Rejecting nihilistic alternatives". I think that section assumes the conclusion: You expect it to choose its biases on the basis of what is moral, instead of on the basis of its current biases.

Reply
Arjun Panickssery's Shortform
Gurkenglas10d40

The analogous argument would be:

If I have no way to do something, then it's nonsensical to say that I should avoid doing that thing. For example, if you say that I should have avoided arriving to an appointment on time and I say that it would be impossible because you only told me about it an hour ago and it's 1000 miles away, then it would be nonsensical for you to say that I should have avoided arriving in time anyway. This is equivalent to saying that if I should avoid doing something, then I can do it.

Reply1
Arjun Panickssery's Shortform
Gurkenglas11d20

I think that makes as much sense as "Whatever ought to be done can actually be done". Do you have some argument that makes sense of one but not the other?

Reply
Arjun Panickssery's Shortform
Gurkenglas11d20

By analogous reasoning, if determinism is true, then whatever ought not to be done also actually is done.

Reply
Tomás B.'s Shortform
Gurkenglas19d40

or an audio interface to a camera

Reply
Tech Tree for Secure Multipolar AI
Gurkenglas19d20

I plugged your middle paragraph into the provided AI because that's its point. Here's the response:

Currently, no technical or governance scheme can reliably guarantee that all private entities are prevented from developing or running AGI outside official oversight. Even strong international agreements or hardware controls can be circumvented by determined actors, especially as required compute drops with research progress. Without ubiquitous surveillance or global control over compute, models, and researchers, a determined group could realistically “go rogue,” meaning any system that depends on absolute prevention is vulnerable to secret efforts that might reach AGI/ASI first, potentially unleashing unaligned or unsafe systems beyond collective control.

sounds kinda sycophantic, e.g. you only need global control over one of the three.

Reply
Load More
5Gurkenglas's Shortform
6y
30
Reflective category theory
3y
(+100)
Reflective category theory
3y
(+193/-111)
Reflective category theory
3y
(+11/-13)
Reflective category theory
3y
(+344/-78)
Reflective category theory
3y
(+5)
83I'm offering free math consultations!
8mo
7
24A Brief Theology of D&D
3y
2
65Would you like me to debug your math?
4y
16
22Domain Theory and the Prisoner's Dilemma: FairBot
4y
5
7Changing the AI race payoff matrix
5y
2
68Using GPT-N to Solve Interpretability of Neural Networks: A Research Agenda
Ω
5y
Ω
11
43Mapping Out Alignment
Ω
5y
Ω
0
18What are some good public contribution opportunities? (100$ bounty)
Q
5y
Q
1
5Gurkenglas's Shortform
6y
30
41Implications of GPT-2
7y
28
Load More