Assuming I have an IQ of 145, there are ~11 million people on Earth smarter (have a higher IQ) than me, but almost none of them, including e.g. Terence Tao, are trying to do something about AI x-risk even at this late date. Updating on this has to move one away from HIA directionally, right, versus the prior 10 years ago?
10 years ago you could say that those ~11m have just never thought about AI, but today the conclusion seemingly has to be that strategic competence is surprisingly little correlated with, or not much scaled by, intelligence, which if true would mean that HIA wouldn't do much for the key bottleneck of humanity's strategic incompetence[1], but could easily make things worse by creating more and harder strategic problems.
Yeah, so I think this is probably my most salient crux at this point: what does the "strategic competence landscape" look like after significant HIA has occurred?
(To nitpick a little, if your 10m estimate is assuming N(100,15) globally, it's going to be a large overestimate, because that's normed from the richest healthiest countries. The mean global IQ is of course lower, which by tail effects drops the size of the extremes rapidly. And since there's such a mixture of countries and populations, N(90,15) is a sketch estimate too, and if you try to do it per-country, you run into issues like whether your China estimate is really trustworthy due to the danger of selective testing, especially testing just of Shanghai/Beijing, and most of these biases will tend to bias your estimate upwards.)
Oops, thought I could trust "reasoning" AI (Gemini 3 Pro) for such a simple seeming question. Had it redo the estimate taking your comment into account, and it came up with 1m assuming N(90,15) globally, which still felt wrong, so I had it redo the estimate using country-level data, and it ended up with 7.5m with 6.1m in East Asia, 1.1m in the West, and .3m in RoW. This assumed N(105,15) for East Asia (so not quite using country-level data), which Opus and GPT point out might be an overestimate due to China being a bit lower than this. Had them redo the EA estimate using country-level data and they came up with 4.5m and 5.5m for EA (using N(103,15) and N(104,15) for China) respectively.
This is actually a significant update for my mental world model, as I didn't previously realize that China had more than half of the world's population of IQ>145 people.
what does the "strategic competence landscape" look like after significant HIA has occurred?
It's a good question, thanks. I'm still thinking these things through, but my guess at where I'll end up is something like:
Strategic competence is in a class with several other things. The class is something like: A kind of competence, which humanity needs, and doesn't have, and which you could hope that some people would exert if they could, but apparently it's rare. For example, competence at solving deep philosophical problems; at deep / full-spectrum cognitive empathy for others; at organizing large groups (e.g. shepherding group epistemics); at Wisdom. For things in this class, you need some combination of high cognitive capacity and high values of other unknown traits. If you have the unknown traits, then more IQ still helps a lot with what you can achieve. Also, IQ can funge against those traits, but only with enough of the right memes and at a fairly poor rate. (For example, you can do cognitive empathy well just by working really hard at it and being smart, without innate talents / attunements / whatever, but you have to work really hard and be really smart.)
I think this will imply a high value of HIA. It also implies a high value on figuring out how to influence other traits (e.g. through good parenting, or though reprogenetics though that has other fraughtness). It also implies a high value on generally contributing to a good memetic / group epistemic / group agentic environment in which HIA kids could grow up.
(I don't update very much on anecdotes from people about their supposed IQ and talents; I don't feel I know how to evaluate it. IQ measurements such as the SAT do have noise; are you not the smartest person in a room of 1000 actually-random people? How heavily selected was your crypto research environment?)
My first two thoughts are:
1) HIA is very clearly going to be a lot slower than the development of ASI, so it doesn't solve the problem of controlling artificial intelligence much smarter than us.
2) On the other hand, Homo sapiens is only a quarter-million years old. We're clearly still very little past the minimum intelligence threshold for having technology (let along a sophisticated technological society). To use a software analogy, if we were a computer language, we'd be barely Turing complete, and still a Turing tarpit. So even an increase of only 20 or 30 IQ points on average might well make a significant difference in how well we coped with difficult problems such as AI alignment.
HIA is very clearly going to be a lot slower than the development of ASI
(FWIW, I don't think that's right. I think there's quite substantial chance we have time for reprogenetics; see https://www.lesswrong.com/posts/sTDfraZab47KiRMmT/views-on-when-agi-comes-and-on-strategy-to-reduce and https://www.lesswrong.com/posts/5tqFT3bcTekvico4d/do-confident-short-timelines-make-sense . Also, in theory, some of the methods listed here (https://www.lesswrong.com/posts/jTiSWHKAtnyA723LE/overview-of-strong-human-intelligence-amplification-methods) could be much faster, e.g. brain implants or signaling molecules. Pushing up HIA timelines still helps: https://tsvibt.blogspot.com/2022/08/the-benefit-of-intervening-sooner.html )
This is a timelines question. For the record, my current AGI timelines are 5–20 years from now, with ASI 1–10 years after that — a little longer than some people here on LW, and definitely longer than Sam Altman or Dario Amodei are discussing. So at the long ends of both of those ranges, that does give time for one generation of HIA recipients to reach young adulthood after an education received while augmented — but (especially if the HIA techniques used need to be applied before birth), very little before that time for the HIA techniques to actually be developed.
I'm not aware of rapid progress in HIA. People tend to be very cautious about what experiments you're allowed to do on the brains of young children, so I'm very dubious that we're going to see sudden advances in HIA: even once experimentation starts, at a minimum I'd expect it to take at least one lightly modified generation reaching adulthood with little-or-no problems (that we didn't by then know how to fix) and measurable improvements, and then a second generation of significantly more modified individuals to reach even 20–30 IQ point improvements. So my HIA timelines are ~~50 years, longer than even the long end of my ASI timelines. But I really haven't looked into it, so that's basically just a wild-assed-guess.
So, I retract "very clearly", and would substitute "quite likely".
FWIW, I think that in some sort of hypothetical involving a bunch more resources (something like $200 million -- $1 billion, maybe), you could plausibly technically get to strong reprogenetic HIA within 5 or 10 years. This would go through IVG plus either iterated CRISPR, iterated recombinant selection, and/or chromosome selection. (Then you'd have to wait for the kids to grow up, and as you say, uptake would be slow at first and would see regulatory obstacles.)
I suspect you'd need to spend those resources in some jurisdiction advanced enough to be a good place to do biomedical research yet small enough that it would be willing to take a hands-off attitude with an investment of that size. Places like Singapore or Costa Rica come to mind.
(I mentally checked again, and I still don't feel like posting it, IDK why. It was written under a lot of time pressure during Inkhaven. I intend to at some point post my fuller / more fully explained reasoning.)
As a very general argument, we might expect HIA that targets IQ or IQ-like traits specifically to be bad because it's imbalanced. Specifically, it makes people who are more capable but not necessarily wiser, to the extent that wisdom is orthogonal to IQ. Since we're in a regime where the unwise competent pursuit of technology is an existential risk, this implies HIA would be bad.
this, in my opinion is the strongest argument against HIA
IQ test are the most widely used intelligence measurement tool and would likely be used as a target for the first widely available forms of HIA
mass producing tech wizards with wis 10 and int 20 is a technoptimists wet dream and a rationalists nightmare
what would be ideal is to somehow raise humanities collective scores on some hypothetical wisdom analog to the IQ test, lets call it a WQ test
on one hand wisdom seems to be more teachable then other forms of intelligence
on the other hand it seems to be harder to boost through nootropics or reprogenetics then IQ
and on a third hand, we don't have a WQ test, not only is it harder to measure but collecting enough data with it to work out a form of HIA using it would take way too long
Thanks.
I am interested in at some point thinking about how to measure "WQ" (see some speculations on what wisdom is here: https://www.lesswrong.com/posts/fzKfzXWEBaENJXDGP/what-is-wisdom-1 ).
I do also think you could increase wisdom with better memes plus some intelligence, which upvotes making wise memes.
Crosspost from my blog.
Context
Previously, in "HIA and X-risk part 1: Why it helps", I laid out the reasons I think human intelligence amplification would decrease existential risk from AGI. Here I'll give all the reasons I can currently think of that HIA might plausibly increase AGI X-risk.
Questions for the reader
Caveats
The world is very complicated and chaotic and I can't plausibly predict even important questions like "what actual effect would such and such have". I can't even plausibly resolve much uncertainty, and the world is full of agents who will adaptively do surprising things. So the actual search procedure is something like: What is a way, or reason to think, that HIA might increase AGI X-risk, that could plausibly hypothetically convince me that HIA is bad to do? This is mostly a breadth-first search, with a bit of deeper thinking.
In particular, many of the reasons listed below, as they are presented, are, according to my actual beliefs, not true or misrepresented or misemphasized. However, that said, this is an attempt at True Doubt, which partly succeeded; some of the reasons listed do give me some real pause.
This is a similar project as "Potential perils of germline genomic engineering". As in that case, keep in mind that part of the reason for this exploration is not just to answer "Should we do this, yes or no?", but also to answer "As we do this, how do we do this in a beneficial way?". See "To be sharpened by true criticisms" in "Genomic emancipation".
What is HIA?
I'll generally leave it fairly undefined what intelligence is, and what human intelligence amplification is. See "Overview of strong human intelligence amplification methods" for some concrete methods that might be used to implement HIA; those methods suggest (various different) specific functional meanings of intelligence and HIA.
Because we are being imprecise:
Vague definitions of intelligence and HIA
Vaguely speaking, by HIA I mean any method for increasing a living human's intelligence, or for making some future people who in expectation have a higher intelligence than they would have otherwise had by default. Generally, we're discussing strong HIA, meaning that the increase in intelligence is large—imagine 30 or 50 IQ points, so going from average to genius or genius to world-class genius.
Vaguely speaking, by intelligence I mean someone's ability to solve problems that are bottlenecked on cognition (as opposed to physical strength or stamina, or financial resources, etc.). A priori, this could include the whole range of cognitive problem-solving. So, we include stereotypically IQ-style cognitive problems, like math or engineering. But, we also include for example the ability and inclination towards political charisma, wisdom, good judgement, philosophical ability, learning, questioning and attending to something steadily, creativity, good performance under stress, empathy, contributing well to teams, memory, taste, and speed.
On the other hand, we do not include other cognitive traits, such as kindness, agreeableness, emotional valence, emotional regulation, determination, conscientiousness, and so on. These are important traits in general, and it might be good to also give people the tools to influence themselves on those traits (though that might also be fraught due to coercion risks). But this article is focused more narrowly on intelligence, rather than all cognitive traits.
In practice, intelligence refers to whatever we can reasonably easily measure. If a trait is hard to measure, it's hard to increase. (This is indeed a cause for concern, in that only increasing traits that are easily measured could be distortive somehow; the claim under discussion is whether HIA is good even under this restriction.) More specifically, intelligence refers to IQ, because IQ is fairly easy to measure. IQ is far from capturing everything about someone's ability to solve problems that are bottlenecked on cognition. But in this article we take it for granted that IQ is a significant factor in those abilities, and we presume that IQ can be increased.
HIA as a general access good
One dimension we will fix is distribution: We will assume that HIA comes in an open access way. In other words, defenses of HIA can't say "we'll only give HIA to the people who are morally good, and therefore there will be a bunch more brainpower directed in a morally good way". That's because a restricted access implementation of HIA seems largely infeasible and also morally and ethically very fraught.
I don't think it's an absolute principle that if you come up with an effective HIA method you have to immediately share it with everyone. But I do think there's a strong moral weight towards doing so; and there's separately a politico-ethical weight (meaning roughly "it's not the sort of thing you should do as a member in good standing of society, even if it's moral, because it would justifiably cause a lot of conflict"). Because of the politico-ethical weight especially, in many scenarios it seems logistically infeasible to do very much selection of who gets access.
This ethical weight towards general access is strongly increased in the case of reprogenetics. Reprogenetics is inherently a multi-use technology, and is already being used by polygenic embryo screening companies to enable parents to decrease disease risks in their future children. This means that society has a very strong justified interest in reprogenetics being equal-access (in the medium-term, once the initial expensive development stages have been completed). Since reprogenetics is likely to be the most feasible HIA method (see "Overview of strong human intelligence amplification methods"), open access seems like a reasonable mainline assumption.
Finally, open-access HIA might be harder to defend as helpful for decreasing AGI existential risk, compared to some sort of hypothetical restricted-access HIA. So, defending the claim that even open-access HIA decreases X-risk is a stricter test; if passed, it should provide stronger evidence that HIA is good to pursue.
HIA and reprogenetics
Since reprogenetics is likely to be the most feasible strong HIA method, it's hard to discuss HIA in general completely separately from reprogenetics. The type of HIA available, the timing of its advent, what other traits can be influenced, and how society will react are all potentially heavily affected if the method is reprogenetics specifically.
Still, as much as possible, this article aims to discuss the impact of HIA in general, factoring out impacts from any specific HIA methods. For thoughts on the downside risks of reprogenetics, see "Potential perils of germline genomic engineering".
AGI X-risk
Background assumptions
This exploration assumes:
Red vs. Blue AGI capabilities research
I want to introduce a piece of vague terminology to help with discussing the strategic landscape. Very vaguely speaking, there's a spectrum of AGI capabilities from "Red" (near-term, big training runs, lots of attention) to "Blue" (blue-sky research). It's of course far from actually one-dimensional, and some entries in the below table are quite debatable (i.e. maybe the two entries should be swapped). Still, I want to use this one dimension as a rough-and-ready way to divide the space up by one degree. To give more flavor of the dimension:
A plausible (AFAIK) first-approximation model is that at any given time, Red research is the most likely to set off an intelligence explosion. Red research takes existing ideas that have already been somewhat proven, and then cranks them up to 11 to see what happens. On the other hand, Blue research is most likely to contribute to getting to AGI in the longer run.
Red research is easier to regulate than Blue research. That's because Red research requires big piles of resources, and is generally more visible (PR, products, large salaries, brand recognition). In particular, the physical needs of a large datacenter (energy, heat, chips) can be detected and regulated. Blue research can be carried out with consumer computers and via intellectual discourse, and it uses more specialized theoretical ideas, so it is harder to detect or even define.
An ontology of effects of interventions on world processes
In general, with some strategic intervention, the question arises: What processes in the world does this intervention speed up / support, and what processes does the intervention slow down / disrupt?
To a rough first approximation, the intervention is good if and only if the expected net change in all the speeds of the affected processes is a good change. So, we can get a rough guess for the value of an intervention by making guesses at how it affects each separate world process. Then, an argument that HIA is bad takes the form "This process is bad and is especially accelerated by HIA" or "This process is good and is especially decelerated by HIA", or "Process X is worse than process Y and process X is accelerated by HIA more than process Y is accelerated".
The next subsection will make some general remarks about the meaning of "acceleration". The following two subsections will give a list of categories of ways that HIA could affect the speed of some process. (They don't try to present a comprehensive ontology; I just think dividing up the space somewhat, even a bit arbitrarily, is helpful because it makes it easier to think in terms of specifics while also searching broadly through much of the whole space.)
The meaning of "acceleration"
To get some kind of handle on the menagerie of plausible effects of HIA, I'll give a list of categories of ways that HIA could affect the speed of some process. These will be phrased as which processes "are accelerated" by HIA. This is vague, for convenience, but some notes to clarify a bit:
Effects of HIA on a single process
Effects of HIA involving multiple processes
Processes
This is a list indicating some of the processes relevant to AGI X-risk:
Some plausible bad effects of HIA on processes
The following subsections list reasons to think that HIA would speed up / support risky processes more than it speeds up / supports derisky processes.
Speeding up Blue research
HIA would (a priori, in expectation) speed up all research. If progress on some research problem is more bottlenecked on very difficult ideas (compared to e.g. money, legwork, regulatory approval, etc.), then it will tend to be sped up by HIA more than another research problem that's less bottlenecked on ideas. Therefore, at a guess, HIA would directly speed up Blue research relatively more than many other kinds of research (including Red research).
Smart people might tend to be most interested in endeavors that are individualistic, technical, ambitious, computer-y, and puzzle-y. So they'd tend to be drawn to AGI research. This effect might be currently added to by society's tendency to not naturally offer ideal social and economic niches for very smart people.
As a basic note, we observe people already being directed to Blue research, so by default we expect that to continue.
Blue research might also have some second-order self-acceleration effects. E.g. there would be intellectual network effects, and maybe some self-improvement effects via better credit assignment and resource allocation internal to the field. These effects might be relatively weak because Blue research is fairly diffuse, but still substantive. On the other hand, there might be significant "coordination overhang": there could be a threshold effect, where with some difficult new ideas, a large number of small siloed Blue research groups could coordinate with each other. Since there's far more absolute Blue research than alignment research, there's more such overhang for Blue research.
Blue research is especially bad, because:
Speeding up Red research
Relative to Blue research, Red research is less bottlenecked on very difficult ideas, so it gets less of a relative direct speedup.
However, Red is likely to have strong indirect acceleration. Because of money and status incentives, Red research attracts people. Red research is likely to attract excess big piles of resources to AGI capabilities. It will probably continue attracting investment as it gets applied to more sectors of the economy, and it gets applied more as it progresses more. It also gains social cachet.
As with Blue research, we observe people being directed to Red research. This observation is even more indicative of trends for Red research in particular, because Red research has upticked a lot recently. That means people are still being directed to Red research even in a memetic environment that already includes a lot of warnings about AGI X-risk. In particular, this suggests that kids who benefit from HIA, growing up in a memetic environment with X-risk warnings but also a very prominent money incentive to do AGI research, might tend to work on Red research.
Red researchers might be especially prone and able to take agentic, conflictual stances towards efforts to avert AGI X-risk. That's because they are more concentrated, have more resources at hand, and tend to be more anti-social and greedy. For example:
Less speeding up legal and social regulation
People (the public at large; policymakers) could socially and legally push against AGI research. They'd first have to be convinced to do so. That process may be less bottlenecked on ideas, compared to AGI research. Instead it may be more bottlenecked on, for example:
Nonlinear / race-condition regulatory escape
In general, processes that regulate AGI research are in some conflict with AGI researchers. The results of this conflict could be quite nonlinear, with a soft threshold effect where the ability of AGI researchers to carry on dangerous research could overpower the ability of regulators to prevent it.
Similar things happen with tax evasion and with regulation of pirating media.
Since AGI research is likely to be accelerated relatively more than regulation of AGI research, HIA would increase the likelihood of regulatory escape.
Alignment loses the race anyway
Even if HIA speeds up alignment, the plan of making an aligned pivotal AI still probably requires making AGI-potent capabilities advances. So, a fortiori, aligned pivotal AI would still probably lose the race against (unaligned omnicidal) AGI. So, the current trajectory is bad, and HIA doesn't change that.
What HIA does do, is speed up that trajectory. So even if alignment and capabilities research got the same speedup from HIA, the overall effect would not benefit the chances of alignment beating AGI.
Intrinsic regulatory escape
HIA people, especially extremely smart ones, would in general be out of distribution. That could be because of selection effects, the specific form of HIA, or just because of the high intelligence itself.
Because HIA people are out of distribution, society would tend to be less good at regulating them in general, e.g.:
Disrupting regulatory systems
In general, HIA could cause conflict. Conflict could destabilize systems. If systems are destabilized, they might be less able to regulate in general. Therefore, HIA could make it easier for AGI capabilities research to evade regulation. Examples:
In general, many aspects of the current state of affairs will be somewhat at equilibrium, and in particular will be somewhat adapted to the current state of affairs. To the extent that the current state of affairs includes some ability to regulate dangerous technologies, that ability would be disrupted by fast shifts that move out of the regime of adaptation. [H/t so and so for this point.] Further, this adaptation would tend to be poor at benefiting from HIA acceleration, so it would tend to fall even further behind, leading to even more escape.
Note that this argument is a response to the reversal test, because it argues that the status quo is best.
Social values favor following local incentives
Generally, given society's current set of values (that it instills in people), long-term altruistic payoffs aren't incentivized. So in general, processes that only have long-term altruistic payoffs will receive less benefit from HIA. In particular, alignment research, the decision to stop doing AGI research, and the decision to regulate AGI research, are not incentivized.
That is a first order effect, where rewards and punishments don't directly incentivize long-term thinking. As a second order effect, besides the direct effect, there's an indirect effect where the fact that society is like this further breaks reasonable faith someone might have in society being good long-term. Since long-term society is a stag hunt game, this further disincentivizes long-term thinking; long-term thinking is partly incentivized because others are directly incentivized to do long-term thinking, but if they aren't then that incentive is gone. E.g. if there's a lot of fraud and injustice, that diminishes your expectation that being honest and just will pay off, because others won't collaborate with you on your honest and just endeavors. This directly interferes with good endeavors. It also might indirectly interfere with good endeavors by more generally distorting people's values. That happens because the general environment of bad incentives makes there be less expectation of a good long-term future in general, which makes people care less about, for example, omnicide. So AGI X-risk would seem less bad. In that mindset, the thrill and money from AGI research would be more tempting on net.
Less speeding up change towards better values
In general, for humanity to respond better to AGI is to some extent a question of values, broadly construed to include wisdom, sanity, calmness, patience, coherence, goodness, long-term thinking, altruism, empathy. Policymakers and the public would have to care about long-term global outcomes rather than short-term ones; AGI researchers would have to care about not harming others more than a small chance of large personal gain, and would have to have hope in the future without AGI.
Rather than being bottlenecked on ideas, value change may be relatively more bottlenecked on e.g.:
Alignment harnesses added brainpower much less effectively than capabilities research does
In addition to just being more difficult, the conceptual structure of the problem of AGI alignment has some more specific disfavorable properties compared to capabilities, which are salient in this context. Alignment progress is less parallelizable, cascading, tractionful, and purely technical than capabilities. In more detail:
There's generally much less traction in alignment research.
Ideas in capabilities more easily cascade into more ideas.
Capabilities is more parallelizable.
Alignment depends more on cognitive traits that are less IQ-correlated than raw technical problem-solving.
For these reasons, alignment harnesses the gains from HIA much less effectively than capabilities research does.
A related point / another way to say this is that alignment benefits the most from HIA that makes there be more extremely-smart people, but does not benefit differentially from HIA that makes there be more somewhat-smart people, whereas capabilities research does benefit from more somewhat-smart people.
HIA people may tend to be transgressive
In general, there are several reasons to be cautious about using HIA. Therefore, people who make use of HIA technology might tend to be especially transgressive, i.e. ignoring reasons to not use it. (Cf. "Transgression" in "Potential perils of germline genomic engineering".) Also, extremely smart people might tend to be transgressive in some ways. Being transgressive might correlate with other traits that lead to HIA people having transgression-related traits. Those traits would tend to make HIA people do more bad things, including pursuing AGI research.
Examples:
Other arguments
This section gives other arguments that HIA increases AGI X-risk.
Concentration of power
If there's a large early cohort of people who benefit a lot from HIA, they might form a somewhat cohesive community. This could have bad effects, including exacerbating some of the dynamics mentioned above. For example:
HIA is unpredictable and therefore risky
Generally, we don't understand the tails of cognitive performance, so we don't understand what HIA would be like. If there's some strong tendency in HIA people, that tendency would have a large effect in a world with HIA. Most changes are bad, so a priori large effects are bad. Since HIA is unpredictable, we don't have a good reason to expect it to have good effects.
More capable but not wiser
As a very general argument, we might expect HIA that targets IQ or IQ-like traits specifically to be bad because it's imbalanced. Specifically, it makes people who are more capable but not necessarily wiser, to the extent that wisdom is orthogonal to IQ. Since we're in a regime where the unwise competent pursuit of technology is an existential risk, this implies HIA would be bad.
As an analogy, consider a 3-year-old. Suppose they suddenly gained the strength of an adult and the self-control of an adult. Let us ignore, for the sake of the hypothetical, all the probable badness that would entail for the child's experience and development, and just ask, what are the direct kinetic consequences of that change? It wouldn't be so bad: they have the self-control of an adult so they won't do too much harm. But what if the child suddenly gained the strength of an adult, but not the self-control? This disproportionate change in abilities would be disastrous.
Acknowledgements
Thanks to many people for conversations about this, especially RK, DB, MS, VP, SE, TY.