I think it's unlikely we get there in the foreseeable future, with the current paradigms
It would be nice if you could define “foreseeable future”. 3 years? 10 years? 30? 100? 1000? What?
And I’m not sure why “with the current paradigms” is in that sentence. The post you’re responding to is “Ten arguments that AI is an existential risk”, not “Ten arguments that Multimodal Large Language Models are an existential risk”, right?
If your assumption is that “the current paradigms” will remain the current paradigms for the “foreseeable future”, then you should say that, and explain why you think so. It seems to me that the paradigm in AI has had quite a bit of change in the last 6 years (i.e. since 2018, before GPT-2, i.e. a time when few had heard of LLMs), and has had complete wrenching change in the last 20 years (i.e. since 2004, many years before AlexNet, and a time when deep learning as a whole was still an obscure backwater, if I understand correctly). So by the same token, it’s plausible that the field of AI might have quite a bit of change in the next 6 years, and complete wrenching change in the next 20 years, right?
why would something that you admit that is temporary (for now) matter in an exponential curve? It's like saying it's ok to go out for drinks on March 5 2020. Ok sure but 10 days after it wasn't. The argument must stand for a very long period of time or it's better not said. And that is the best argument for why we should be cautious, because a) we don't know for sure and b) things change extremely fast.
Because you could make the same argument could be made earlier in the "exponential curve". I don't think we should have paused AI (or more broadly CS) in the 50's, and I don't think we should do it now.
but you are comparing epochs before and after Turing test passed. Isnt' that relevant? The Turing test unanimously was/is an inflection point and arguably most experts think we have already passed it in 2023.
- Thus most decisions will probably be allocated to AI systems
- If AI systems make most decisions, humans will lose control of the future
- If humans have no control of the future, the future will probably be bad for humans
- Sure - at some point in the future, maybe.
- Maybe, maybe not. Humans tend to have a bit of an ego when it comes to letting a filthy machine make decisions for them. But I'll bite the bullet.
- There's several levels on which I disagree here. Firstly, we're assuming that "humans" have control of the future in the first place. It's hard to assign coherent agency to humanity as a whole, it's more of a weird mess of conflicting incentives, and nobody really controls it. Secondly, if those AI systems are designed in the right way, the might just become the tools for humanity to sorta steer the future the way we want it.
I agree with your framing here that systems made up of rules + humans + various technological infrastructure are the actual things that control the future. But I think the key is that the systems themselves would begin to favour more non-human decision making because of incentive structures.
Eg, corporate entities have a profit incentive to have the most efficient decision maker in charge of the company, and maybe that includes a CEO but the board might insist on the use of an AI assistant for that CEO, and if the CEO makes a decision that goes against the AI and it turns out to be wrong shareholders in that company will come to trust the AI system more and more of the time. They don't necessarily care about the ego of the CEO they just care about the outcomes, within the competitive market.
In this way, more and more decision making gets turned over to non-human systems because of the competitive structures which are very difficult to escape from. As this transition continues it becomes very hard to control the unseen externalities from these decisions.
I suppose this doesn't seem too catastrophic in its fundamental form, but I think the outcomes of playing it forward essentially seem to be a significant potential for harm from these externalities, without much of a mechanism for recourse.
This is a polemic to the ten arguments post. I'm not a regular LW poster, but I'm an AI researcher and mild-AI-worrier.
I believe that AI progress, and the risks associated with it, is one of the most important things to figure out as humanity in the current year. And yet, in most discussions about x-risk, I find myself unaligned with either side.
My overall thesis about AI x-risk is that it's absolutely real, but also far enough into the future that at this moment, we should simply continue progress on both capabilities and safety. I'm not trying to argue that sufficiently powerful AI could never pose an x-risk, this belief seems rather silly.
Disclaimers:
- Humans will build AI systems that are 'agents', i.e. they will autonomously pursue goals
- Humans won’t figure out how to make systems with goals that are compatible with human welfare and realizing human values
- Such systems will be built or selected to be highly competent, and so gain the power to achieve their goals
- Thus the future will be primarily controlled by AIs, who will direct it in ways that are at odds with long-run human welfare or the realization of human values
Overall this is the core x-risk argument that I completely agree with - but I think it's unlikely we get there in the foreseeable future, with the current paradigms.
- Human dominance over other animal species is primarily due to humans having superior cognitive and coordination abilities
- Therefore if another 'species' appears with abilities superior to those of humans, that species will become dominant over humans in the same way
- AI will essentially be a 'species' with superior abilities to humans
- Therefore AI will dominate humans
Not a fan of this argument. Might be effective as an intuition pump if someone can't even conceive of how a powerful AI could lead to x-risk, but I don't take it too seriously.
- AI systems will become much more competent than humans at decision-making
- Thus most decisions will probably be allocated to AI systems
- If AI systems make most decisions, humans will lose control of the future
- If humans have no control of the future, the future will probably be bad for humans
This largely sounds like a rehash of the previous argument. AI will become more powerful, we can't control it, we're screwed. The argument has a different coat of paint, so my response is different, but ultimately the point is that an AI will take over the world with us as an under-species.
- Advances in AI will produce very rapid changes, in available AI technology, other technologies, and society
- Faster changes reduce the ability for humans to exert meaningful control over events, because they need time to make non-random choices
- The pace of relevant events could become so fast as to allow for negligible relevant human choice
- If humans are not ongoingly involved in choosing the future, the future is likely to be bad by human lights
In one sense, this argument is obviously true - if we get an AI that's superintelligent, super-quickly, and misaligned, then we're probably screwed because we won't react in time. But it's a spectrum, and the real x-risk is only on the extreme end of the spectrum.
- People who broadly agree on good outcomes within the current world may, given much more power, choose outcomes that others would consider catastrophic
- AI may empower some humans or human groups to bring about futures closer to what they would choose
- From 1, that may be catastrophic according to the values of most other humans
Here we go, full agreement. This really is an issue with any sufficiently powerful technology. If only one person/country had nukes, we'd probably be worse off than in the current multipolar situation. Can the same multipolar approach help in the specific case of AI? Maybe - that's why I tend to favor open-source approaches, at least as of 2024 with the current state of capabilities. So far, for other technologies, we're somehow handling things through governance, so we should keep doing this with AI - and everything else.
- There appear to be non-AI technologies that would pose a risk to humanity if developed
- AI will markedly increase the speed of development of harmful non-AI technologies
- AI will markedly increase the breadth of access to harmful non-AI technologies
- Therefore AI development poses an existential risk to humanity
Most technologies have good and bad uses. AI will be a force multiplier, but if we can't handle Nukes 2.0 obtained via AI, we probably can't handle Nukes 2.0 obtained via good ol' human effort. This is fundamentally an argument against technological progress in general, which could be a significantly larger argument. My overall stance is that technological progress is generally good.
- So far, humans have developed technology largely through understanding relevant mechanisms
- AI systems developed in 2024 are created via repeatedly modifying random systems in the direction of desired behaviors, rather than being manually built, so the mechanisms the systems themselves ultimately use are not understood by human developers
- Systems whose mechanisms are not understood are more likely to produce undesired consequences than well-understood systems
- If such systems are powerful, then the scale of undesired consequences may be catastrophic
I was always a little bit anti-interpretability. Sure, it's better if a model is more interpretable than less interpretable, but at the same time, we don't need it to be fully interpretable to be powerful and aligned.
The core argument here seems to be that if the black boxes remain forever pitch black, and we multiply their potential side effects by a gazillion (in the limit of a "powerful" AI), then the consequences will be terrible. Which... sure, I guess. If it actually remains entirely inscrutable, and it becomes super powerful, then bad outcomes are more likely. But not by much in my opinion.
- Competition can produce outcomes undesirable to all parties, through selection pressure for the success of any behavior that survives well, or through high stakes situations where well-meaning actors' best strategies are risky to all (as with nuclear weapons in the 20th Century)
- AI will increase the intensity of relevant competitions
This feels like a fairly generic "force multiplier" argument. AI, just like any technology, will amplify everything that humans do. So if you take any human-caused risk, you can amplify it in the "powerful AI" limit to infinity and get an x-risk.
This goes back to technological progress in general. The same argument can be made for electricity, so while I agree in principle that it's a risk, it's not an extraordinary risk.
- AI development will have very large impacts, relative to the scale of human society
- Large impacts generally raise the chance of large risks
Once again, take AI as a mysterious multiplier for everything that we do. Take something bad that we (may) do, multiply it by AI, and you get an x-risk.
- The people best placed to judge the extent of existential risk from AI are AI researchers, forecasting experts, experts on AI risk, relevant social scientists, and some others
- Median members of these groups frequently put substantial credence (e.g. 0.4% to 5%) on human extinction or similar disempowerment from AI
I liked that post. It's a coherent summary of the main AI x-risks that I can address. I largely agree with them in principle, but I'm still not convinced that any threat is imminent. Most discussions that I tried to have in the past usually started from step zero ("Let me explain why AI could even be a risk"), which is just boring and unproductive. Perhaps this will lead to something beyond that.