LESSWRONG
LW

Frontpage

15

[ Question ]

Does there exist an AGI-level parameter setting for modern DRL architectures?

by TurnTrout
9th Feb 2020
1 min read
A
2
3

15

Frontpage

15

Does there exist an AGI-level parameter setting for modern DRL architectures?
3Gurkenglas
3Steven Byrnes
2Steven Byrnes
New Answer
New Comment

2 Answers sorted by
top scoring

Gurkenglas

Feb 09, 2020*

30

Yes. Modelspace is huge and we're only exploring a smidgen. The busy beaver sequence hints at how much you can do with a small number of parts and exponential luck. I think feeding a random number generator into a compiler could theoretically have spawned an AGI in the eighties. Given a memory tape, transformers (and much simpler architectures) are Turing-complete. Even if all my reasoning is wrong, can't the model just be hardcoded to output instructions on how to write an AGI?

Add Comment
[-]Steven Byrnes6y30

Very clever! Yes I agree with you that there is a parameter setting for modern DRL architectures for an agent that has an "instinct" to walk over to the nearest computer, and write and execute code that turns on a real-deal superintelligent AGI. Or for a program that manually steps through the execution steps of an AGI Turing machine. I guess I interpreted the question to say that that kind of thing doesn't count. :-P

Reply

Steven Byrnes

Feb 09, 2020

20

Jumping out on a limb—and I might change my mind next week—but I would say "no", if using current popular mainstream DRL techniques, because these lack (1) foresight (i.e., running a generative model to predict the result of different possible courses of action, and choosing on the basis of the results), and (2) analysis-by-synthesis (processing inputs by continually running searches through a space of generative models to find the model that best matches that input). I think humans do both, and without both (among other requirements), I picture systems as sorta more like "operating on instinct" rather than "intelligent".

So (in my mind), your question would be "can we get 'robustly human+ capabilities' from a system operating on instinct?" and the answer is "Obviously yes when restricted to any finite set of tasks in any finite set of situations", e.g. AlphaStar. With enough parameters, the set of tasks and situations could get awfully high, and maybe that counts as "robustly human+"—just as a large enough Giant Lookup Table might count as "robustly human+". But my hunch is that systems with foresight and analysis-by-synthesis will be "robustly human+" earlier than any systems that operate on instinct.

Add Comment
Rendering 0/1 comments, sorted by
top scoring
(show more)
Click to highlight new comments since: Today at 1:10 PM
Moderation Log
More from TurnTrout
View more
Curated and popular this week
A
2
0

Suppose the architecture includes memory (in the form of a recurrent state) and will act as the policy network for an observation-based RL agent. Evaluating the agent from a reasonable initial state, would you guess that there exists a model with robustly human+ capabilities for current architectures?

How many parameters would it take before you estimate there's a fifty-fifty chance of such a parameter setting existing? 1 billion? 1 trillion? More?