Apparently with reflection technique (answer-critique-improve) GTP4 capable of giving much better answers. But that implies it should be capable of doing essentially Alpha Go Zero type of learning! It can't do complete self play from zero as there is no ground truth for it to learn from, but that basically burns all hopes of having further progress bottlenecked by data. Also, while still severely lacking, it constitutes limited self-improvement capability. Not only we have unclear distinction of GTP4 being an AGI or not, but it also have some slight self improvement capabilities! We really are boiling frog, aren't we?

New to LessWrong?

New Comment
3 comments, sorted by Click to highlight new comments since: Today at 9:55 PM

I don't see the "recursive" part. If you ask GPT4 to critique and improve its previous answer, will you keep getting increasingly better results if you do this repeatedly? Will you keep getting better results in the future as a result of having this done in the past?

Shortly, can you do something like "GPT, how to conquer the world? critique and improve your plan 1000 times" and get a superior plan you can easily follow?

I am not AI researcher. (I have a feeling you may have mistaken me for one?)

I don't expect much gain after first iteration of reflection (don't know if it was attempted). When calling it recursive I was referring to Alpha Go Zero style of distillation and amplification: We have model producing Q->A, reflect on A to get A' and update model in direction Q->A'. We got to the state where A' is better than A, before if tried result probably would be distilled stupidity instead of distilled intelligence. 

Such process, in my opinion, is significant step in direction of "capable of building new, better AI" worth explicitly noticing before we take such capability for granted.

Will you keep getting better results in the future as a result of having this done in the past?

Only if you apply "distillation and amplification" part, and I hope if you go too hard in absence of some kind of reality anchoring it may go off the rails and result in distilled weirdness. And hopefully you need bigger model anyway.

[-]awg1y20

Agreed that this seems like the first step in something big. Methods will probably be discovered to operationalize this in a more end-to-end manner given a single prompt/problem, potentially along with compression techniques to distill and analyze many different iterations of a solution before finding and amplifying the right one to the user.

Incidentally, I think this is a pretty good video on these recent developments: