Sorted by New

Wiki Contributions



Hmmm I guess I don't really use the terms 'investing' and 'trading' interchangeably.


Humans are the most dangerous entity on earth and my only fear with ai is that it ends up being too human.

Answer by J30

For arbitrary time horizons nothing is 'safe', but that just means our economy shifts to a new model. It doesn't mean the outcome is bad for humans. I don't know if it makes sense to worry about which part of the ship will become submerged first because everyone will rush for the other parts and those jobs will be too competitive. It might be better to worry about how to pressure the political system to take proactive action to rearchitect our economy. Ubi and/or a shorter workweek are inevitable and the sooner we sort out how to implement that the better.

for the sake of understanding the roadmap for the autopocalypse, I think you can consider these factors:

The most obvious: can the work be entirely computer-based? A corollary to this is whether a unskilled human with the assistance of an ai can replace the skilled human (e.g. Healthcare roles requiring knowledge work and physical work)

Regulatory environment. Even if it were possible for software to replace a human worker licensing requirements and other laws may protect human workers for awhile beyond that point.

Eventually machines will be able to transcend software to perform every physical task a human can now perform. And that mostly won't be anthropomorphic robots but more automation of machines that currently are operated by humans, like cars and drones. The anthropomorphic robots will appear (on the job) the furthest into the future.

But again everyone will be racing away from these jobs (and toward the remaining 'safe' ones) at the same time. Financial investment may be the safest source of income. Owning your own business may possibly benefit from the autopocalypse but at the extreme you will just be an investor in a company run by machines.


So the best advice is probably to build an investment portfolio (and the knowledge to do that well). If you own the companies it doesn't matter who the workers are.


Ben Shapiro wanted to say that he likes this quote


Thanks! Just read some summaries of parfit. Do you know any literature that addresses this issue within the context of a) impacts to other species, or b) using artificial minds as the additional population? I assume the total utilitarianism theory assumes arbitrarily growing physical space for populations to expand into and would not apply to finite spaces or resources (I think I recall bostrom addressing that).

Reading up on parfit also made me realize that deep utopia really has prerequisites and you were right that it's probably more readily understood by those with philosophy background. I didn't really understand what he was saying about utilitarianism until just reading about parfit.


This is a major theme in Star Trek: The Next Generation, where they refer to it as the Prime Directive. It always bothered me when they violated the Prime Directive and intervened because it seemed like it was an act of moral imperialism. But I guess that's just my morals (an objection to moral imperialism) conflicting with theirs.

A human monoculture seems bad for many reasons analogous to the ones that make an agricultural monoculture bad, though. Cultural diversity and heterogeneity should make our species more innovative and more robust to potential future threats. A culturally heterogeneous world would also seem harder for a central entity to gain control of. Isn't this largely why the British, Spanish, and Roman empires declined?


I've only skimmed it but so far I'm surprised bostrom didn't discuss a possible future where ai 'agents' act as both economic producers and consumers. Human population growth would seem to be bad in a world where ai can accommodate human decline (i.e. Protecting modern economies from the loss of consumers and producers), since finite resources will be a pie that gets divided either into smaller slices or larger ones depending on the number of humans around to allocate them to. And larger slices would seem to increase average well-being. Maybe he addressed but I missed it in my skim.