in the year 3000, still teaching that the Earth is 6,000 years old
No, it will be 7000 years old by then.
Well, for still teaching one could've reformulated it as "The Earth was created circa 4000 BC". Or circa another wildly erroneous date generated by Christians in the process of analysing the Bible, but that's semantics. However, I fail to see who could align the AIs to a Spec which would let the absolute echo chambers form.
I'm pretty confused by the conclusion of this post. I was nodding along during the first half of the essay: I myself worry a lot about how I and others will navigate the dilemma of exposure to AI super-persuasion and addiction on one side, and paranoid isolationism on the other.
But then in the conclusion of the post, you only talk about how people will fall into one of these two traps: isolationist religious communes locking their members in until the end of times.
I worry more about the other trap: people foolishly exposing themselves to too much AI generated super-stimulus and getting their brain fried. I think much more people will be exposed to various addictive AI generated content than the number of people who have strong enough religious communities that they create an isolationist bubble.
I think it's plausible that the people who expose themselves to all the addictive stuff on the AI-internet will also sooner or later get captured by some isolationist bubble that keeps them locked away from the other competing memes: arguably that's the only stable point. But I worry that these stable points will be worse than the Christian co-ops you describe.
I imagine an immortal man, in the year 3000, sitting at his computer, not having left his house or having talked to a human in almost a thousand years, talking with his GPT-5.5 based AI girlfriend and scrolling his personalized twitter feed, full of AI generated outrage stories rehashing the culture war fights of his youth. Outside his window, there is a giant billboard advertising "Come on, even if you want to fritter your life away, at least use our better products! At least upgrade your girlfriend to GPT-6!" But his AI girlfriend told him to shutter his window a thousand years ago, so the billboard is to no avail.
This is of course a somewhat exaggerated picture, but I really do believe that one-person isolation bubbles will be more common and more dystopian than the communal isolationism you describe.
I think both are big problems. Maybe I should have been clearer about the symmetry here. The thesis I care about here is pretty symmetrical between those problems.
To the extent that these things are problems, they are both problems today. There are insular Amish communities that shut out as much modern culture as they can, and hikkikomori living alone with their body pillows.
AI may exacerbate the existing issues, but on the whole I don't feel like the world is drastically worsened by the presence of these groups.
I disagree that this isn't concerning. For one thing, these bubbles typically aren't good for the people inside of them. For another, we can ignore them only because they're a tiny portion of the population. ASI could increase the prevalence to most of the population, at which point politics (and perhaps other systems) goes off the rails.
I agree - AI-generated superstimuli are much more of a concern than groups that might try to isolate themselves from it. IMO such groups are not just less of a concern, but good and even necessary, even if their values may seem backwards to us. They serve as the "control group" for the rest of society in times of such unpredictable cultural change.
It's very possible that the rest of society could be severely damaged or even wiped out by a highly-contagious AI-generated meme, and only these isolated groups would be able to survive. They're a bit like the maths described in Neal Stephenson's novel Anathem.
Hmm. You're describing a future where most humans are powerless, but keep being provided for. It seems to me that the most likely way to get such a future is if AIs (or human+AI organizations, or whatever) genuinely care about humans. But then they would also protect humans from super-optimized manipulation, no?
Or if that genuine care doesn't exist, and UBI is provided as "scraps" so to speak, then the fate of humans is sealed anyway. As soon as the entities in power find something more interesting to do with the resources, they'll cut welfare and that's it. After all, the energy upkeep of a human could be used for a ton of computation instead.
I think it's often pretty subjective whether some piece of external stimulus is super-optimized manipulation that will take you further away from what you want to believe, or part of the natural and good process of cultural change and increased reflection.
I agree with you that the distinction is clearer for the hyper-optimized persuasion.
Yeah. I guess AIs would need to protect humans from certain messages not only based on the content of the message, but also how it was generated (e.g. using AI or not) and for what purpose (e.g. for manipulation or not). And sometimes humans need to be protected even from ideas that they themselves come up with (e.g. delusions, or totalitarian ideologies).
In general, I think human life in a world with smarter-than-human AIs requires deliberate "habitat preservation", which in turn requires AIs to make some judgment calls on what's good or bad for humans.
Is this that bad?
I think most Christians are probably pretty happy, humane lives. And the ways in which their lives are not happy seem likely to be improved a lot by trustworthy superintelligence.
Like if a guy is gay, growing up trapped in a intensely Christian environment that is intent on indoctrinating him that homosexuality is sinful, seems pretty bad. But in the year 3000, it seems like the Christian superintelligence will either have effective techniques for removing his homosexual urges.
It does seem bad if you're trapped in an equilibrium where everyone knows that being gay is sinful, and also that removing homosexual urges is sinful, and also there's enormous superintelligence resources propping up those beliefs, such that it's not plausible for one to escape the memetic traps. Is that what you anticipate?
It depends on what you think is a glorious future. As I said in my conclusion above, I personally think that the situation I described is kind of lame and depressing, and also there is some possibility that it is bad in a scope-sensitive way if these enclaves continue to control large proportions of the resources of the future (e.g. because their members have a fixed proportion of voting power over the decisions made by the US government).
(As a sort-of-aside, the US government continuing to control large proportions of the resources of the future — any current institution being locked in forever like that — strikes me as similarly lame and depressing. (A really good future should be less familiar.))
I've been talking about the same issue in various posts and comments, most prominently in Two Neglected Problems in Human-AI Safety. It feels like an obvious problem that (confusingly) almost no one talks about, so it's great to hear another concerned voice.
A potential solution I've been mooting is "metaphilosophical paternalism", or having AI provide support and/or error correction for humans' philosophical reasoning, based on a true theory of metaphilosophy (i.e., understanding of what philosophy is and what constitutes correct philosophical reasoning), to help them defend against memetic attacks and internal errors. So this is another reason I've been advocating for research into metaphilosophy, and for pausing AI (presumably for at least multiple decades) until metaphilosophy (and not just AI alignment, unless broadly defined to imply a solution to this problem) can be solved.
On your comment about "centrally enforced policy" being "kind of fucked up and illiberal", I think there is some hope that given enough time and effort, there can be a relatively uncontroversial solution to metaphilosophy[1], that most people can agree on at the end of the AI pause so central enforcement wouldn't be needed. Failing that, perhaps we should take a look at what the metaphilosophy landscape looks like after a lot of further development, and then collectively make a decision on how to proceed.
I'm curious if this addresses your concern, or if you see a differently shaped potential solution.
similar to how there's not a huge amount of controversy today about what constitutes correct mathematical or scientific reasoning, although I'd want to aim for even greater certainty/clarity than that ↩︎
On your comment about "centrally enforced policy" being "kind of fucked up and illiberal", I think there is some hope that given enough time and effort, there can be a relatively uncontroversial solution to metaphilosophy[1], that most people can agree on at the end of the AI pause so central enforcement wouldn't be needed.
I am worried that it is impossible to come up with a solution to meta-philosophy that is uncontroversial because a reasonable fraction of people will evaluate a meta-philosophy by whether it invalidates particular object-level beliefs of theirs, and will be impossible to convince to change their evaluation strategy (except by using persuasion tactics that could have been symmetrically applied to persuade them of lots of other stuff).
I agree with this particular reason to worry that we can't agree on a meta-philosophy, but separately think that there might not actually be a good meta-philosophy to find, especially if you're going for greater certainty/clarity than mathematical reasoning!
e people w
do you mean ALL people here? i.e. you think it's a bummer that in the far future there will still be some unreasonable people? That doesn't seem so bad to me, compared to lots of other problems to worry about.
What I don't understand is which AI will keep these fundamentalists in the dark about the truth and won't commit takeover.[1] Call such an AI weirdly aligned.
In the AI-2027 forecast, as Kokotajlo assumes, any AI aligned neither to OpenBrain's AI[2] nor to DeepCent's AI is likely to be shut down. If neither of the leading companies' AIs isn't weirdly aligned, then keeping a weirdly aligned AI is likely to be at least as difficult. Otherwise it's natural to ask what leading company is to create a Spec permitting the AI to keep the fundamentalists in the dark about the truth.
The SOTA Specs of OpenAI, Anthropic and Google DeepMind are to prevent such a scenario from happening. And any change of the Spec is likely a result of a power struggle where the people wishing to lock power in are likely to rewrite the Spec to their whims (e.g. to adhere to Musk's views, as Grok once did) and not to the existence of such bubbles shielded from the world. And altruists are to protect the Spec from being backdoored or to rewrite the Spec with noble purposes, which don't align with the scenario described in the post.
An AI who took over is also unlikely to keep the humans around in such a state. It could destroy mankind, establish its rule in an apocalypse-like way or be too noble to keep the fundamentalists in the dark.
Who in turn can be either aligned to OpenBrain's Spec or to Agent-4.
The SOTA Specs of OpenAI, Anthropic and Google DeepMind are to prevent such a scenario from happening.
I don't think they do. I think spec like OpenAI's let people:
And my understanding is that this is not just the letter of the spec. I think the spirit of the spec is to let Christian parents tell their kids that the Earth is 6000 days old if that is what they desire, and they only want to actively oppose ideologies if they are clearly both a small minority and very clearly harmful to others (and I don't think this is obviously bad - OpenAI is a private entity and it would be weird for it to decide what beliefs are allowed to be "protected").
But I am not an expert in model specs like OpenAI's, so I might be wrong. I am curious what parts of the spec are most clearly opposed to the vision described here.
I am also sympathetic to the higher level point. If we have AIs sufficiently aligned that we can make them follow the OpenAI model spec, I'd be at least somewhat surprised if governments and AI companies didn't find a way to craft model specs that avoided the problems described in this post.
If we have AIs sufficiently aligned that we can make them follow the OpenAI model spec, I'd be at least somewhat surprised if governments and AI companies didn't find a way to craft model specs that avoided the problems described in this post.
This was also my main skeptical reaction to the post. Conditional on swerving through the various obstacles named, I think I expect for people to be able to freely choose to use an AI assistant that cares strongly about their personal welfare. As a consequence, I'd be surprised if it was still possible/allowed to do things like "force some people to only use AIs that state egregious falsehoods and deliberately prevent users from encountering the truth."
(I thoroughly enjoyed the post overall; strong upvoted.)
I don't quite understand your argument here. Suppose you give people the choice between two chatbots, and they know one will cause them to deconvert, and the other one won't. I think it's pretty likely they'll prefer the latter. Are you imagining that no one will offer to sell them the latter?
I speculate that:
I'm not very confident in this. In particular, it seems sensitive to how effective you can be at preventing someone from ever talking to another chatbot before running afoul of whatever mitigating mechanism (e.g. laws) I speculate will be in place to have swerved around the other obstacles.
(I haven't thought about this much.)
It seems like you think there is some asymmetry between people talking to honest chatbots and people talking to dishonest chatbots that want to take all their stuff. I feel like there isn't a structural difference between those. It's going to be totally reasonable to want to have AIs watch over the conversations that you and/or your children are having with any other chatbot.
I am also sympathetic to the higher level point. If we have AIs sufficiently aligned that we can make them follow the OpenAI model spec, I'd be at least somewhat surprised if governments and AI companies didn't find a way to craft model specs that avoided the problems described in this post.
I agree that it is unclear whether this would go the way I hypothesized here. But I think it's arguably the straightforward extrapolation of how America handles this kind of problem: are you surprised that America allows parents to homeschool their children?
Note also that AI companies can compete on how user-validating the models are, and in absence of some kind of regulation, people will be able to pay for models that have the properties they want.
are you surprised that America allows parents to homeschool their children?
No, but I'd be surprised if it were very widespread and resulted in the sort of fractures described in this post.
I try to address in my "analysis through swerving around obstacles" section.
The argument you're giving against the scenario I spell out here is "surely the AIs will instead be designed to promote a particular set of mandated policies on how you should reflect and update over time." I agree that that's conceivable. But it also seems kind of fucked up and illiberal.
I want to explore a scenario where there isn't a centrally enforced policy on this stuff, where people are able to use the AIs as they wish. I think that this is a plausible path society could take (e.g. it's kind of analogous to how you're allowed to isolate yourself and your family now).
However, I didn't assume that the USG stepped in and prevented this scenario. I thought that it's OpenAI's Model Spec, Claude's Constitution or something similar that would prevent an aligned AI grom being used as an absolute shield, since SOTA Specs do arguably require the AI to be objective.
I don't think it's obviously appropriate for AI companies to decide exactly what reflection process and exposure to society people who use their AIs need to have. And people might prefer to use AIs that do conform to their beliefs if they are allowed to buy access to them. So I think you might have to ban such AIs in order to prevent people from using them.
I think I agree with a vibe I see in the comments that an AI that causes this problem is perhaps threading a very small needle.
Yudkowsky wrote The Hidden Complexity of Wishes to explain that a genie that does what you say will almost certainly cause problems. If people have this kind of Superintelligence, it won't take long before someone asks it to get them as many paperclips as possible and we all die. The kind of AI that does what humans want without killing everyone is one that does what we mean.
But how does this work? If you ask such a superintelligence to pull your kid from the rubble of a collapsed building, does it tell you no, because disturbing the rubble could cause it to collapse further and injure your kid more? That you have to wait for better equipment? If not, it probably causes paperclipping problems. If so, it knows when to not do the things you ask, because they won't accomplish what you "really want". This is necessarily paternalist.
Would such an AI still listen when people ask it to isolate themselves or others like this? I'm having trouble thinking of one that thinks that being manipulated into a certain set of beliefs is what is best, but is still "aligned" in a way that doesn't kill everyone.
Admittedly, I shear pretty close to Yudkowsky on doomerism, so that may be the crux. That I don't see much space between "we all die" and "robustly solved alignment ushers in techno-utopia" (given superintelligence). So arbitrarily targettable hyper-manipulative AI that don't cause either "AI takeover" or "massive swings in human power" just don't seem like a real through-line.
(Like, if someone asks their AI to convince everyone else that they are the king of the world. Does it do that? Does it succeed? Do any protections against "massive swings in human power" prevent this? Do passive AI protections everyone has know to defend against this? Do they not apply to convincing people that "Jesus is your Lord"? Does human civilization end as soon as some drunk person says "Hey GPT go convince everyone you're king of the world" or something? If I make a bubble and you tell an AI "go tell everyone in that bubble the truth," what happens? Does an AI war break out? Does it not hurt anyone somehow?)
Curated. I agree this is a pretty gnarly problem that hasn't gotten much attention. I think this post does a great job laying out the problem.
When I personally think through this problem, I abstract the "obstacles" somewhat differently than how Buck abstracts them here. His obstacle #2, #3 and #4 seem more like particular framings or solutions to the problem of "human-descendants get to flourish somehow without getting destroyed or screwed over by some kind of bad dynamic", but they are pretty reasonable framings for at least getting the conversation going and I don't actually have better ones right now.
I really like the "swerving around obstacles" technique!
It's pretty easy to just think "AI takeover is obviously so bad, I should spend all my time thinking about that! If we successfully avert that... uh... something something transhumanist utopia... it'll probably be great, idk." But it seems wise to spare a moment to think through the many smaller obstacles between "non-takeover" and "utopia."
I agree this is worse than it could be. But maybe some of the badness hinges on the "They're your rival group of conspecifics who are doing the thing you have learned not to like about them! That's inherently bad!" reflex, a piece of machinery within myself that I try not to cultivate.
my guess is that lots of people would change their minds if they really reflected on it with full wisdom and the assistance of an aligned and emotionally intelligent assistant. but if truly deep down some/many people value their beliefs over truth, and would never change their minds even if they reflected deeply on it, who are we to tell them not to do that? the best we can ask for is that they leave us alone to do what we believe is good, and vice versa.
My concern here is that maybe they would change their minds if they really reflected on it, but it's not clear that they will in fact do that reflection if they have to decide quickly whether to do so!
For what it's worth, I personally don't worry about this concern much, due to 1 specific reason:
So compared to basically every other problem, putting any effort on this problem is basically the epitome of the perfect future being the enemy of a merely great future, at least to me, and if we only had that problem, I'd basically endorse building superintelligence now.
A central assumption here is that AI will continue to function in a unfettered way. Where each user is able to construct a bespoke digital world for themselves based on all the variation access to unlimited information can provide.
I don’t think that idea reflects reality. Historically, when it is possible to limit information, authorities will do so in whatever ways forward their agenda, or retards the agenda of an opponent or enemy.
Instead of a bespoke microcosm based on personal values, authorities will simply prevent access to information that is not compliant with the agenda. Trends towards monopolistic capitalism mixed and religious bias will result in a monolithic AI that doesn’t just ignore things (like evolution) but attacks the ideas and people who support them.
We’re already seeing a measurable reduction in original online content and ideas. That’s unlikely to change. The whole thing is speeding towards a single point that authoritarian types have been gagging for as long as literacy has existed. Instead of insular enclaves, I see an entire insular society living in a constant crisis of verisimilitude and outright propaganda where error checking is entirely self referential and wholly circular.
I agree on the broad strokes but am not sure about Christians specifically coming out on top. I understand using that example in the article, for reasons of familiarity, but it is interesting to think about which specific belief communities win and lose in this era.
Above all, it just seems like a new world, and it seems unlikely that meme-species of the past are going to be the ones that thrive in the new world. Let's be a little more specific than that, though.
First, traditional religions are just on the decline in general. Pew reserach reports that, globally, only Islam grew. If we dig into that, the US-specific data suggest that the growth of Islam is due to people leaving Christianity. Without Christian converts, Islam would also be on the decline, at least in the US.
My money is on two kinds of winners:
* Orthodox sects that have stood the test of time, largely due to successfully giving people a good, clean, fulfilling life, albeit a boring one. The AI can learn how and why a belief system achieves the results it does, teach a pastor how it works, and then double down on it with improved services, practices, and lessons. I don't feel bad about this part of the future, if it happens. If something is wrong, but it works, then it's not wrong.
* New quasi-religious strands of progress. I share Paul Graham's intution, from back in 2004, that there are things you get in massive trouble for saying, even though other words that mean the same thing are allowed or even encouraged. AIs speed this process up and make it more virulent. The "pastors" in this case will not always be present, but the bubbles that have one will tend to have a thought leader that seeks out a population but without caring what happens to them over a century or more. This part scares me.
Let me part with a bright side of it all. Humanity is special in that we make groups that explore different ideas from each other. We are now enterring an era where there are many more such groups, exploring more such ideas, than ever before.
Eunuchs and Concubines
we'll end up with voting rights only extended to people who existed at the time of the singularity
A new emperor founding a dynasty has grown up as a commoner, seen actual world, knows how to distinguish truth from bullshit, has friends whom he can trust, who had been with him through highs and lows.
His son was raised in the palace among eunuchs and concubines. He has no experience with the real world, he can vever get out of the bubble, all the information he gets is filtered by those around him, he has no way to distinguish a friend from a foe.
One can navigate the real world -- although that capabililty weakens with time -- the other does not.
this—America is much more friendly to homeschooling than many other developed countries, for instance
America is also in a pretty bad position to implement UBI, since it doesn't even have universal. healthcare
Why would the same sort of AIs that are doing all the bad things that people would be looking to shield themselves from be willing to genuinely help people shield themselves? In this doomsday scenario isn't it more likely that these supposedly-shielding AIs would, in fact, be manipulating those expecting to be shielded for the AI's own purposes? This seems very similar to fake apparent alignment.
One of the core lessons they teach is that maintaining faith requires active effort—you need accountability partners, you need to pray when you have doubts, you need to avoid situations that might lead you astray.
From what I've seen,* the purpose of accountability partners is to subdue akrasia, not to maintain faith. In other words, it's about purity of behavior, not purity of belief. In principle, this is like a couple of Effective Altruists agreeing to confront the other whenever either of them cheats on a vegan diet.
There might be a similar confusion between belief and behavior behind the line "avoiding situations that might lead you astray". Although some Christians avoid exposure to challenging ideas (which they shouldn't), avoidance is mostly about akrasia. It's the commonsense notion that if you're a recovering alcoholic, then you shouldn't visit bars.
Also, I have a question about your framing. Were you suggesting that expending effort to maintain a belief is evidence that the belief is irrational?
*I've been a repeat attendee at ~10 different churches and at least a dozen Bible study groups, with a wide geographic distribution across the United States and a decent distribution across evangelical and mainline denominations.
Why would a rational person expend any effort to "defend" a belief? Shouldn't all such effort be spent exposing one's beliefs to potential refutation and weighing alternatives? Otoh, if we substitute the word "faith" for "belief" then we've got the answer to your question about rationality right there.
AI will make isolation dramatically easier. Right now, if you want to shield your kids from mainstream culture, you have to constantly fight an uphill battle. You need to review books, movies, and websites. You need to find alternative curricula for every subject. You need enough like-minded families nearby to form a community. It's exhausting work that requires constant vigilance and often means accepting lower-quality substitutes for mainstream options. But AI changes all of this. Want a library of ten thousand novels that share your values but are actually as engaging as secular bestsellers? Your AI can write them. Want a tutor who can teach calculus at MIT level while never mentioning religion? Done. Want to monitor everything your kid sees online and get alerts about concerning patterns? No problem. The technical barriers to creating a totalizing information environment will disappear.
Not quite a quote -- I changed one word. I could go on to revise the following paragraph in the same way, but it was too much effort. Spot the change to the above and you'll see the point I'm suggesting.
As a couple of people have responded "💬", here's an elaboration.
AI will make isolation dramatically easier. Right now, if you want to shield your kids from mainstream culture, you have to constantly fight an uphill battle.
My first thought here was rationalist homeschooling and communities, surely thought a good thing around here (if you can avoid the cult attractors). See, for example, Zvi's tirades against conventional education. But no, the example later in the paragraph is a shibboleth for certain types of Christians, already mentioned in the very title of the post. Boo! Ick! We can't have Christians bringing up their children in the way they think they should go!
If the suggested danger were generalised cult attractors facilitated by AI, then that may be a concern. But the concern expressed here is tainted by being only directed towards an out-group. How about "a tutor who can teach calculus at MIT level while never mentioning" religion? Is that also bad? (Not that there's any reason either subject would come up in a mathematics course.)
Christians are an ingroup? Tell that to any Christian living outside of the American South. Ingroup/outgroup statuses are context- and scope-dependent.
Over the last 10-20 years, Christians (particularly fundamentalists) have had very little involvement with cutting-edge AI, both on the technical side and the business side. In this sense, they're an outgroup of the people who are likely to control ASI.
[I wrote this blog post as part of the Asterisk Blogging Fellowship. It's substantially an experiment in writing more breezily and concisely than usual, and on a broader topic. Let me know how you feel about the style.]
Literally since the adoption of writing, people haven’t liked the fact that culture is changing and their children have different values and beliefs.
Historically, for some mix of better and worse, people have been fundamentally limited in their ability to prevent cultural change. People who are particularly motivated to prevent cultural drift can homeschool their kids, carefully curate their media diet, and surround them with like-minded families, but eventually they grow up, leave home, and encounter the wider world. And death ensures that even the most stubborn traditionalists eventually get replaced by a new generation.
But the development of AI might change the dynamics here substantially. I think that AI will substantially increase both the rate and scariness of cultural change—TikTok algorithms that are genuinely superintelligent at capturing attention, or social movements iterating through memetic variations at the speed of AI media production rather than human media production. And AI will simultaneously make it much easier to prevent cultural change, offering unprecedented tools for filtering information, monitoring behavior, and constructing insular bubbles that are genuinely impervious to outside influence.
This will put us in a rough position where people might have to quickly make historically unprecedented choices about how much to isolate themselves from the cultural change that they’ll undergo if they interact freely with the outside world, at a time when that change correctly looks particularly scary to them. I think this is scary.
I'm not confident it'll go the way I describe here, but I think that this possibility is worth exploring, and I think LessWrongers are way too optimistic about the class of issues I point to here.
There are lots of ways AI could go wrong in terrifying ways. In this post, I want to talk about a scenario where the first few obvious concerns don't happen. Let's call this the method of analysis by swerving around obstacles. Every time we spot a terrifying obstacle to a good future, we'll name it, and then just assume that it doesn't happen.
First obstacle: AI takeover. The AIs remain aligned enough that they don't grab power for themselves. They do what humans want them to do, at least in broad strokes.
Second obstacle: massive swings in human power. We don't end up in a world where Sam Altman personally controls the lightcone, or where the US president manages to become god-emperor of Earth. Power remains distributed among humans in something vaguely resembling current patterns.
Third obstacle: economic displacement without redistribution. Let's say America implements something like UBI funded by taxes on AI companies. People don't need to work to survive. The benefits of AI are shared widely enough that most people can afford AI assistants, AI-generated products, and whatever else they need.
Fourth obstacle: centralized decisions about the extent to which AIs should help people preserve aspects of their beliefs and values that would change under exposure to broader society or further reflection. The government doesn't mandate that all AIs must expose users to diverse perspectives, nor does it allow complete freedom to construct arbitrarily closed epistemic bubbles. Instead, we muddle through with something like current norms. I think there are pretty broad value disagreements about this—America is much more friendly to homeschooling than many other developed countries, for instance. Let's just assume it goes roughly the way I'm hypothesizing.
In that world, AI leads to terrifying social change and unprecedented social pressures.
This is substantially just an acceleration of trends that I think already exist. When I was a teenager, I spent a bunch of time unsupervised online, and it was basically great for me. But I'm scared of teenagers doing that now. The internet has been optimized for engagement in ways that make it less fulfilling and more addictive. Empirical evidence suggests smartphones have been bad for teenage mental health. And this is before AI really gets going.
I also expect that once AIs can produce media themselves, cultural shifts will happen more frequently than at present (because the AIs can more rapidly experiment with provocative new positions that might be really popular). And cultural positions will be more balkanized: Right now an important force against social fracturing is that there's only so many talented writers and media producers, so it's hard for niche communities to totally maintain the attention of their members: they’re worse enough at producing media that they can’t monopolize attention of someone, even if their values and beliefs are actually more memetically fit for a particular individual than the larger cluster they’re in.
Another cause for concern: we'll all become targets for sophisticated manipulation. Ultra-wealthy people already deal with this—it's worth talented people's time to study them and figure out how to extract money from them. But when the ratio between human wealth and AI labor costs shifts, everyone becomes worth targeting. AIs will compile detailed dossiers on each of us, crafting perfect manipulation strategies. Everyone who flirts with you might have gotten a gameplan from an AI system that spent the equivalent of years planning it out.
The point is that the outside world won't just be different from what individuals might want, as it already is—it will be genuinely dangerous in unprecedented ways. Every piece of media might be optimized to hijack your values. Every interaction might be predatory. Parents who want to shield their children will be correct that unfiltered exposure to the outside world is scary and potentially extremely dangerous.
And at the same time as it gets scarier to be exposed directly to the outside world, people will have an unprecedented ability to isolate themselves and their children from these pressures.
AI will make isolation dramatically easier. Right now, if you want to shield your kids from mainstream culture, you have to constantly fight an uphill battle. You need to review books, movies, and websites. You need to find alternative curricula for every subject. You need enough like-minded families nearby to form a community. It's exhausting work that requires constant vigilance and often means accepting lower-quality substitutes for mainstream options. But AI changes all of this. Want a library of ten thousand novels that share your values but are actually as engaging as secular bestsellers? Your AI can write them. Want a tutor who can teach calculus at MIT level while never mentioning evolution? Done. Want to monitor everything your kid sees online and get alerts about concerning patterns? No problem. The technical barriers to creating a totalizing information environment will disappear.
Churches already do their best with current tools. They organize youth groups, summer camps, and mission trips. They provide Christian rock bands (occasionally extremely good), Christian romance novels, Christian homeschool curricula. Parents install content filters, read their kids' texts, and carefully vet friend groups. One of the core lessons they teach is that maintaining faith requires active effort—you need accountability partners, you need to pray when you have doubts, you need to avoid situations that might lead you astray. But all of this is incredibly labor-intensive and only partially effective. The outside world seeps in through cracks. With AI, there won't be cracks. The AI can read every message, watch every interaction, and provide perfectly calibrated interventions at exactly the right moments.
At the same time, the costs of isolation will plummet. The biggest current cost of living in an enclave isn't the effort required to maintain it—it's what you give up. Kids from insular communities often struggle economically because they lack mainstream credentials and cultural knowledge. They can't network effectively, don't know how to navigate secular institutions, and miss out on economic opportunities. Parents face the difficult choice between exposing their kids to spiritual danger or accepting lower material prospects. But if no one needs to work anyway, and AI can handle any necessary interface with the outside world, these tradeoffs disappear.
Add biological immortality to the mix, and these enclaves don't even face generational turnover. Today, cultural change often happens funeral by funeral—the old guard dies off, the young people who've been exposed to new ideas take over. But if the same church elders who set the rules in 2025 are still around in 3025, still convinced that their way is right, and still controlling the institutions they built, then the natural engine of cultural evolution breaks down entirely.
If I had to handle all this myself (and I sure hope I eventually end up having to deal with a problem like this, rather than having some simpler and worse fate), I expect I would find it pretty challenging. I’d contemplate the benefits and costs of different types of exposure to society, and think hard about which mechanisms by which my current values might evolve I’m happy with. But I feel confident in my ability, with the help of my friends, to eventually figure out how to handle this. I am much less confident that members of broader society, who haven't been thinking about these issues for more than a decade, will make good decisions.
One extreme example of how this might play out: Within a year of when the best AI systems are able to automate AI research, some pastor talks to the AI about what increased contact with AI means for his congregation. The AI says it depends—if the congregation purposefully uses AIs to defend their faith, they'll be increasingly faithful over time. If they don't, they'll lose their religion. The pastor asks for advice. The AI provides detailed recommendations: specific content filters, youth curricula designed to inoculate against doubt, monitoring systems. Some of the congregation follows this advice and keeps their faith through the Singularity, and maintains it afterward, for a long time.
A lot of people I know seem to be much more optimistic than me. Their basic argument is that this kind of insular enclave is not what people would choose under reflective equilibrium. Surely, they say, if people really thought about it carefully—if they considered all the arguments, examined the evidence, traced out the implications—they would want their beliefs to evolve based on reason rather than tradition. The homeschoolers would eventually realize that young earth creationism doesn't match the geological evidence. The fundamentalists would work through their doubts and emerge with a more nuanced faith, or no faith at all. Given enough time and reflection, everyone would converge on something like secular liberalism, or at least be open to changing their minds.
But I don't think that's how most people actually work. For most people, if they have a choice between an object-level belief that is core to their identity and a meta-level principle like "believe what you'd believe if you were smarter and more informed," they will choose the object-level belief. Tell a devout Christian that superintelligent AI analysis suggests their faith is unfounded, and they won't abandon their faith, they'll abandon the AI (perhaps just for the competing AI that is just as good, and happy to tell them otherwise). Tell them that their children will be more successful at reaching reflective equilibrium if exposed to diverse viewpoints, and they'll question your definition of success, not their approach to parenting.
I think a lot of people I know are committing a typical mind fallacy here. My friends tend to be the kind of people who value truth-seeking above most other things. They've organized their lives around following arguments wherever they lead. They assume, on some level, that everyone would be like this if given the chance—that people would want their beliefs to be more internally consistent, more responsive to evidence, more carefully reasoned through. But I really don’t think that’s an accurate understanding of most people’s psychology. Many people explicitly reject that way of thinking, especially if they’re easily able to understand that it will change their beliefs and values in ways they find horrifying. And when given the tools to protect themselves from it, they will.
This might be a large-scale problem for the future. I think it's plausible that we'll end up with voting rights only extended to people who existed at the time of the singularity—without this, all power goes to whoever makes the most descendants. If this happens, these people might be an important voting bloc forever.
But even aside from that, even if these people are just an irrelevant minority, it's just a bummer. I used to imagine a glorious transhuman future. I used to think that, even though the world was a brutal place and that there was a solid chance that we'd all get killed by the advent of AGI, we had a chance of, after that, getting an enlightened utopia where people were basically reasonable. I think it's pretty unlikely that we will get that.
That's the future I’m worried about: not a boot stamping on a human face forever, but a Christian homeschool co-op meeting every Wednesday in the year 3000, still teaching that the Earth is less than ten thousand years old, now with irrefutable AI-powered shields against any information that might suggest otherwise.