The report presents little evidence of successful implementation of the model because it is more an outline of the theory than a rigorous scientific testing of the model. It is not peer reviewed.
Are you... on the right forum? How did you even show up here? Why would you think that somehow the people here will appreciate an appeal to academic authority of all things? This blogpost itself is not peer reviewed!
I did note that the report is not peer reviewed, but I still have no problem treating it as a credible report. Whether it's peer reviewed or not (it doesn't matter), it still does not contain what the AI 2027 authors claim/imply it contains (i.e. evidence of AI successes in "Self-improvement for general intelligence")
I'll have my full review of AI 2027 finished in a few weeks. It's clear if you look at the links they provide to "evidence" that they actually have no evidence. They consistently misrepresent research (peer reviewed or not). I showed up here because I also reviewed Yudkowsky and Soares' book and wanted to talk to the people here about that, but my review wasn't approved for posting. You can find it in the link at the end of the article above.
What’s their evidence that any such extrapolation is warranted?
You click on the provided link to supporting evidence and you are taken to a 2017 report titled “Supervising strong learners by amplifying weak experts”.
The link to Supervising strong learners by amplifying weak experts is in the following sentence:
Self-improvement for general intelligence had seen minor successes before.
This is the very first sentence in the Iterated distillation and amplification (IDA) collapsible section, and is clearly not being offered as evidence that is meant to justify extrapolating to the very last sentence in that section:
Now, the models have become sufficiently good at verifying more subjective things (e.g. the quality of a work product), allowing the use of IDA to improve the model at many tasks.
The rest of your post has a lot of other objections that seem invalid or confused, like attempting to use the lack of the paper's peer review as meaningful evidence about whether a technique like that might generalize or not, but I don't think it's worth getting into them because the entire argument is premised on a misunderstanding of what evidence is being offered for what purpose.
True, I didn't include my review of the other article they reference in that last sentence (https://arxiv.org/pdf/2210.11610). That article also contains no evidence to support their claims. I'll have my full review of AI 2027 finished in a few weeks.
Even granting what you said about the gap between the first and last sentence, the sentence in which the article is referenced is: "Self-improvement for general intelligence had seen minor successes before."
The report referenced clearly has nothing to do with "general intelligence": they test the model on 5 narrow algorithmic tasks. And they explicitly and repeatedly say that they have provided no evidence for the model's applicability in wider tasks.
The AI 2027 authors reference the report apparently as evidence of AI's successes in self-improvement in "general intelligence". The report contains no such evidence. So the report is misrepresented by the AI 2027 authors
If you are going to predict that that gap will be bridged—as AI 2027’s authors predict—you would need to explain how it will be bridged and present evidence.
Not really, it's a forecast, it's supposed to be inherently handwavy.
It's actually a very good science -- the autiors are formulating a hypothesis which is perfectly verifiable -- just wait for 1 more year!
I'll have my full review of AI 2027 finished in a few week. I've looked at every link they provide and found that none of the relevant articles linked provide evidence to support the claim that computers will be intelligent
I got –17 karma for this post! I understand people disagree, but it is a good faith post. Now my account is restricted because of the negative karma
LessWrong is not a forum in which posting in good faith is sufficient to be welcomed! Think of it as a professional community. Just because you are writing a physics paper in good faith doesn't mean it will be well-received by the physics community as a contribution. Similarly here, I think you are missing a large number of prerequisites that are assumed to be understood by participants on LW.
I would recommend checking out the New User's Guide to LessWrong .
By Oscar Davies
The authors of AI 2027 repeatedly misrepresent the scientific reports they reference as evidence to support their arguments.
I’m writing a comprehensive review of AI 2027 (to be posted soon), but in the meantime I want to briefly present an example of the misrepresentations:
In a section in AI 2027 on “Iterated distillation and amplification (IDA)” (there are no page numbers so to find a passage, use control + F search), the authors write:
“Self-improvement for general intelligence had seen minor successes before. But in early 2027, it’s seeing huge returns [with] IDA”
(I’ll outline IDA below—but understanding the details is not very important for our purposes). A few sentences later, they go on:
“Early versions of IDA have been working for many years on easily verifiable tasks, like math and coding problems that have a clear answer …
Now [they’re predicting what things will be like in 2027 here], the models have become sufficiently good at verifying more subjective things (e.g. the quality of a work product), allowing the use of IDA to improve the model at many tasks.”
Here the authors take success in “math and coding problems that have a clear answer” and extrapolate to future “models [that] have become sufficiently good at verifying more subjective things”—i.e. models that are capable in many areas beyond narrow math and coding problems (such wider capacities are essential to the predicted “superintelligent” computers). What’s their evidence that any such extrapolation is warranted?
You click on the provided link to supporting evidence and you are taken to a 2017 report titled “Supervising strong learners by amplifying weak experts”.
In that report the authors outline their attempt to get AIs to imitate the operations a human carries out as the human breaks down (they call it “decomposing”) a difficult problem into smaller, easier subproblems, and then combines the answers to those subproblems to find a solution to the larger problem. (That’s more or less what is meant by “IDA”—again, don’t worry if it’s not completely clear, it’s not important here.)
The report presents little evidence of successful implementation of the model because it is more an outline of the theory than a rigorous scientific testing of the model. It is not peer reviewed.
The authors do test the model on, as they put it, “5 toy algorithmic tasks”—i.e. math tasks with very narrow goals and methods. They report some success.
One of the five test “algorithmic tasks” in the paper, for example, is: “Given a directed graph with 64 vertices and 128 edges, find the distance from s to t.”
It’s no secret that computers can do some math. The possibility of machine intelligence revolves around the question whether computers’ success at math and similarly clear-cut tasks could be extrapolated to wider less-clear-cut problems.
The authors explicitly say that their report provides no evidence that their theory (called “Iterated Amplification”) is useful outside the bounds of narrow math (or “algorithmic”) tasks. Here’s a few passages from the article making that point:
“Having successfully applied Iterated Amplification to synthetic algorithmic problems, the natural question is whether it can actually be applied to complex real-world tasks that are “beyond human scale.” We leave a convincing demonstration to future work.”
“In our experiments questions can be algorithmically decomposed into subquestions, and we replace the human with a hand-coded algorithm. These experiments don’t shed any light on whether humans can decompose interesting real world tasks [they say the model needs humans to decompose tasks], nor on whether it would be feasible to learn messy real world decompositions.”
“Removing these simplifications is a task for future work, which will ultimately test the hypothesis that Iterated Amplification can be usefully applied to complex real-world tasks for which no other training strategy is available.”
Repeatedly, the authors state that their paper does not touch on whether their theory can be applied outside of “synthetic algorithmic problems”. I might point out the article was written eight years ago, and the theory still has found little success outside that narrow domain.
This lack of real-world application (the paper explicitly does not even attempt such application), along with the authors’ repeated reminder that their paper does not address “real world” tasks, does not deter the authors of AI 2027 from referencing it as alleged evidence to support their argument that computers will be superintelligent and take over the world.
The AI 2027 authors’ link to the study appears in the sentence “Self-improvement for general intelligence had seen minor successes before.” I don’t see how anyone could construe the referenced report as having to do with “general intelligence”. To repeat, the authors of the report explicitly say that they do not address wider (“general”) problems, but rather focus on narrow math tasks. Nor does the report provide very promising proof of “self-improvement” on a level approaching anything like intelligence.
In quotes we saw above, the AI 2027 authors imply that AI’s relative success in “math and coding problems that have a clear answer” is evidence that soon they will “become sufficiently good at verifying more subjective things (e.g. the quality of a work product), allowing … the model [to improve] at many tasks”—i.e. the model will succeed in domains beyond narrow “math and coding problems”.
Success in such wider domains is the crux of the problem of inventing intelligent computers. If you are going to predict that that gap will be bridged—as AI 2027’s authors predict—you would need to explain how it will be bridged and present evidence.
No such evidence or explanation appears in AI 2027. In fact, the authors of the paper referenced as evidence in the passage we’re reviewing explicitly state the opposite—that the report does not contain proof their model is applicable for “more subjective things”, but has only been used in narrow math tasks.
As this is just one minor example, it might seem as if I am nitpicking. I am getting too deep into the details by reviewing the specific claims in materials referenced in AI 2027. But the details are what is important. The referenced materials are supposed to be the evidentiary basis for the claims. Without evidence, the claims are empty.
And perhaps you wonder whether the authors of AI 2027 slipped up in this one case and in fact they present more substantial evidence elsewhere. They do not. In every relevant instance they make basic mistakes like the one I’ve reviewed in this article.
These sorts of misrepresentations and unfounded extrapolations are repeated throughout AI 2027. I will soon publish my full review demonstrating that the authors have no evidence to support their claims, and repeatedly reference reports that do not contain anything close to evidence to warrant their predictions.
.
.
.
Take a look at my other articles:
On the errors in AI inventor Geoffrey Hinton’s claims