LESSWRONG
LW

496
aelwood
1530
Message
Dialogue
Subscribe

Physicist and AI researcher with a passion for philosophy. Interested in consciousness, models of reality, meditation, and how we should choose to live and act. Writing over at https://pursuingreality.com

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
No wikitag contributions to display.
A stubborn unbeliever finally gets the depth of the AI alignment problem
aelwood3y2-1

This is a great comment, but you don't need to worry that I'll be indoctrinated! 

I was actually using that terminology a bit tongue in cheek, as I perceive exactly what you say about the religious fervour of some AI alignment proponents. I think the general attitude and vibe of Yudkowsky etc is one of the main reasons I was suspicious about their arguments for AI takeoff in the first place.

Reply
A stubborn unbeliever finally gets the depth of the AI alignment problem
aelwood3y20

I actually agree that it's likely an AGI will at least start thinking in a way kind of similar to a human, but that in the end this will still be very difficult to align. I actually really recommend that you checkout Understand by Ted Chiang, which basically plays out the exact scenario you mentioned -- a normal guy gets super human intelligence and chaos ensues. 

Reply
A stubborn unbeliever finally gets the depth of the AI alignment problem
aelwood3y10

Thanks for the comment, I'll read some more on the distinction of inner and outer alignment, that sounds interesting.

I don't think you would need to get anywhere near perfect simulation in order to begin to have extremely good predictive power over the world. We're already seeing this in graphics and physics modeling.

I think this is a good point, although these are cases where lots of data is available. So I guess any case in which you don't have the data ready would still have more difficulties. Off the top of my head I don't know how limiting this would be in practice, but it should be in lots of cases.

Reply
-5The second bitter lesson — there’s a fundamental problem with aligning distributed AI
8mo
0
-12Utilitarianism is the only option
3y
7
17A stubborn unbeliever finally gets the depth of the AI alignment problem
3y
8
4A hierarchy of truth
5y
0
1The ways of knowing
5y
0