Subjective Relativity, Time Dilation and Divergence

by jacob_cannell7 min read11th Feb 201195 comments


Personal Blog

And the whole earth was of one language, and of one speech. And it came to pass . . .they said, Go to, let us build us a city and a tower, whose top may reach unto heaven; and let us make us a name, lest we be scattered abroad upon the face of the whole earth. And the Lord came down to see the city and the tower, which the children built. And the Lord said, Behold, the people is one, and they have all one language; and this they begin to do; and now nothing will be restrained from them, which they have imagined to do. Go to, let us go down, and there confound their language, that they may not understand one another's speech. So the Lord scattered them abroad from thence upon the face of all the earth: and they left off to build . . . 

Genesis 11: 1-9

Some elementary physical quantitative properties of systems compactly describe a wide spectrum of macroscopic configurations.  Take for example the concept of temperature: given a basic understanding of physics this single parameter compactly encodes a powerful conceptual mapping of state-space.  

It is easy for your mind to visualize how a large change in temperature would effect everything from your toast to a planetary ecosystem.  It is one of the key factors which divides habitable planets such as Earth from inhospitably cold worlds like Mars or burning infernos such as Venus.  You can imagine the Earth growing hotter and visualize an entire set of complex consequences: melting ice caps, rising water levels, climate changes, eventual loss of surface water, runaway greenhouse effect and a scorched planet.

Here is an unconsidered physical parameter that could determine much of the future of civilization: the speed of thought and the derived subjective speed of light.  

The speed of thought is not something we are accustomed to pondering because we all share the same underlying neurological substrate which operates at a maximum frequency of around a kilohertz, and appears to have minor and major decision update cycles at rates in the vicinity of 33hz to 3hz.1

On the other hand the communication delay has changed significantly over the last ten thousand years as we evolved from hunter-gatherer tribes to a global civilization.

For much of early human history, the normal instantaneous communication distance limit would be the audible range of about 100 feet, and long distance communication consisted of sending physical human messengers; a risky endeavor that could take months to traverse a continent.

The long distance communication delay in this era (on the order of months) was more than 10^9 times the baseline thought-frequency (which is around a millisecond).  The developmental outcome in this type of regime is divergence.  New ideas and slight mutations of existing beliefs are generated in local ingroups far faster than they can ever propagate to remote outgroups.  

In the divergent regime cultures fragment into sub-cultures; languages split into dialects; and dialects become new languages and cultures as groups expand geographically.2

Over time a steady accumulation of technological developments increased subjective bandwidth and reduce subjective latency in the global human network: the advent of agricultural civilization concentrated human populations into smaller regions, the domestication of horses decreased long distance travel time, books allowed stored communication from the past, and the printing press provided an efficient one to many communication amplifier.

Yet despite all of this progress, even as late as the mid 19th century the pony express was considered fast long distance communication.  It was not until just very recently in the 20th century that near instantaneous long distance communication became relatively cheap and widespread.3

Today the communication delay for typical point to point communication around the world is somewhere around 200 to 300 ms, corresponding to a low delay/thought-frequency ratio of 10^2.  This figure is close enough to the brain's natural update cycles to permit real time communication.

It is difficult to measure, but the general modern trend seems to have now finally shifted towards convergence rather than divergence.  Enough people are moving between cultures, translating between languages and communicating new ideas fast enough relevant to the speed of thought to largely counter the tendency toward divergence.

But now consider that our global computational network consists of two very different substrates: the electronic substrate which operates at near-light speed, and a neural substrate which operates at much slower chemical speeds; more than one million times slower.

At the moment the vast majority of the world's knowledge and intelligence is encoded in the larger and slower neural substrate, but the electronic substrate is growing exponentially at a vastly faster pace.

Viewed as a single global cybernetic computational network we can see there is massive discrepancy between the neural and electronic sub-components.

So what happens when we shift completely to the electronic, when we have artificial brains and AGI's that think at full electronic speeds?

The speed of light measured in atomic seconds is the same for all physical frames of reference, but it's subjective speed varies based on one's subjective speed of thought.  This subjective relativity causes effective time dilation proportional to one's level of acceleration.

For an AGI or upload that has an architecture similar to the brain but encoded in the electronic substrate using high effeciency neuromorphic circuitry, thoughts could be computed in around a thousand clock cycles or less at a rate of billions of clock cycles per second.  

Such a Mind would experience a million fold time dilation, or an entire subjective year every thirty seconds.

Imagine the external universe, time itself, slowing down by a factor of a million.  Watching a human walk to work would be similar to us watching grass grow.  Actually it would be considerably worse; five minutes would correspond to an unimaginable decade of subjective time for an acceleration level 6 hyperintelligence.

A bullet would not appear to be much faster than a commuter, and the speed of light itself, the fastest signal propagation in the universe, would be slowed down to just 300 subjective meters per second, roughly the speed of a jetliner.

Real-time communication would thus only be possible with entities in the same building and on the same local network.

It would take a subjective day or two to reach distant external internet sites.  Browsing the web would not be possible in the conventional sense.  It would appear the only viable strategy would be to copy most of the internet into a local cache.  But even this would be impeded by the million fold subjective bandwidth slowdown.  

Today's fastest gigabyte direct ethernet backbone connections would be reduced back down to mere kilobyte per second modem speeds.  A cable modem connection speed would require about as much fiber bandwidth as our entire current transatlantic fiber capacity.

Acceleration level 6 corresponds to a 10^8 value for the communication delay / thoughtspeed ratio, a shift backwards roughly equivalent to the era before the advent of the telegraph.  This is the historical domain of both the Roman Empire and pre civil war America.

If Moore's Law continues well into the next decade, further levels of acceleration will be possible.  A combination of denser circuitry, architectural optimizations over the brain and higher clock rates could lead to acceleration level 9 hyperintelligences.  Overclocked circa 2011 CPUs are already approaching 10 GHZ, and test transistors have achieved speeds into the terrahertz range in the lab.4

The brain takes about 1000 'clocks' of the base neuron frequency to compute one second worth of thought.  If a future massively dense and parallel neuromorphic architecture could do the same work 10 times more effeciently and thus compute one second of thought in 100 clock cycles while running at 100 GHZ this would enable acceleration level 9.5

Acceleration level 9 stretches the limits of human imagination.  It's difficult to conceive of an intelligence that experiences around 30 years in just one second, or a billion subjective years for every sidereal year.

At this dilation factor light slows to just 300 centimeters per second, a slow walking pace.  More crucially, light moves just 3 centimeters per clock cycle, which would place serious size constraints on the physical implementation of a single mind.  To make integrated decisions with a unified knowledge base, in other words think in how we understand the term, the core of a Mind running at these speeds would have to be crammed into the space of a modern desktop box.  (although it certainly could have a larger secondary knowledge store accessible with some delay)    

The small size constraint would severely limit how much power/heat one could throw at the problem, and thus these high speeds will probably require much higher circuit densities to achieve the required energy efficiency than implied by memory requirements alone.

With light itself crawling along at 300 centimeters per second it would take data packets hundreds of millions of seconds, or on the order of years, to make typical transits across the internet.  These speeds are already close to physical limits; even level 9 hyperintelligences will probably not be able to surmount the speed of light delay.

The entire fiber backbone of the circa 2011 transatlantic connection would be required to achieve end 20th century dialup modem speeds.6

Even using all of that fiber it would take on the order of ten physical seconds to transfer a 10^14 byte Mind, corresponding to hundreds of thousands of subjective years.

A level 9 world is one where the subjective communication delay, approaching 10^11, is a throwback to the prehistoric era.  Strong Singletons and even weaker systems such as global governments or modern markets would be unlikely or impossible at such high levels of acceleration.7

From the social and cultural perspective high levels of thought acceleration are structurally equivalent to the world expanding to billions of times it's current size. 

It is similar to the earth exploding into an intergalactic or hyperdimensional civilization linked together by a vast impossibly slow lightspeed transit network.

Entire new cultures and civilizations would form and play out complex histories in the blink of an eye.

With every increase in circuit density and speed the new metaverse will vasten exponentially in virtual space and time just as it physically shrinks and quickens down into the ever smaller, faster levels of the real.

And although all of this change will be unimaginably fast for a biological human, Moore's Law will be a distant ancestral memory for level 9 intelligences, as it depends on a complex series of events in the impossibly slow physical world of matter.  Even if an entire new hardware generation transition could be compressed into just 8 hours of physical time through nanotechnological miracles, that's still an unimaginable million years of subjective time at acceleration level 9.

Another interesting subjective difference: computer speed or performance will not change much from the inside perspective of a hyperintelligence running on the same hardware.  Traditional computers will indefinitely maintain roughly the same subjective slow speeds for minds running on the same substrate at those same speeds.  Density shrinkings will enable more and or larger minds; but only a net shift towards the latter would entail a net increase in traditional parallel CPU performance available per capita.  But as discussed previously, speed of light delays severely constrain the size of large unified minds.

The radical space-time compression of the Metaverse Singularity model suggests a reappraisal of the Fermi Paradox and the long-term fate of civilizations.  

The speed of light barrier gives a natural gradient to the expansion of complexity: it is inwards, not outwards.  

Humanity today could mount an expedition to a nearby solar system, but the opportunity cost of such an endeavor vastly exceeds any realistic discounted returns.  The incredible resources space colonization would require are much better put to use increasing our planetary intelligence through investing in further semiconductor technology.

This might never change.  Indeed such a change would be a complete reversal of the general universal trend towards smaller, faster complexity.

Each transition to a new level of acceleration and density will increase the opportunity cost of expansion in proportion.  Light-years are vast units of space-time for humans today, but they are unimaginably vaster for future accelerated hyperintelligences. 

Facing the future it appears that looking outwards into space is looking into the past, for the future lies in innerspace, not outerspace.



1 Human neuron action potentials have a measured maximum frequency of a little less than a millisecond.  This is thus one measure of rough equivalence to the clock frequency in a digital circuit, but it is something of a conservative over-estimate as neurological circuits are not synchronous at that frequency.  Many circuits in the brain are semi-synchronized over longer intervals roughly corresponding to the various measured 'brain wave' frequencies, and neuron driven mechanisms such as voice have upper frequencies of the same order.  Humans can react in as quickly as 150ms in some conditions, but appear to initiate actions such as saccades at a rate of 3 to 4 per second.  Smaller primate brains are similar but somewhat quicker.

2 The greater monogenesis theory of all extant languages and cultures from a single distant historical proto-language is a matter of debate amongst linguistics, but the similarity in many low-level root words is far beyond chance.  The restrained theory of a common root Proto-Indo-European language is near universally accepted.  This map and this tree help visualize the geographical historical divergence of this original language/cultural across the supercontinent along with it's characteristic artifact: the chariot.  All of this divergence occurred on a timescale of five to six millenia.

3  Homing pigeons, where available, were of course much faster than the pony express, but were rare and low-bandwidth.

4 Apparently this has been done numerous times in the last decade in different ways.  Here is one example.  Of course making a few transistors run in the terahertz doesn't get you much closer to making a whole CPU actually run at that speed, for a large variety of reasons.

5 None of these particular numbers will seem outlandish a decade or two from now if Moore's Law holds it's pace.  However getting a brain or AGI type design to run at these fantastic speeds will likely require more significant innovations such as a move to 3D integrated circuits and major interconnect breakthroughs.  There are many technological uncertainties here, but less than that involved in drexler-style nano-tech, and this is all on the current main path.

6 It looks like we currently have around 8 tbps of transatlantic bandwidth circa 2011.

7 Nick Bostrom seems to have introduced the Singleton concept to the Singularity/Futurist discourse here.  He mentions artificial intelligences as one potential Singleton promoting technology but doesn't consider their speed potential with respect to the speed of light.


Personal Blog


95 comments, sorted by Highlighting new comments since Today at 8:49 PM
New Comment
Some comments are truncated due to high volume. (⌘F to expand all)Change truncation settings

I love this article, but I disagree with the conclusion. You're essentially saying that a post-singularity world would be too impatient to explore the stars. I grant you that thinking a million times faster would make someone very impatient, but living a million times longer seems likely to counterbalance that.

Back in the days of cristopher columbus, what stopped people from sailing off and finding new continents wasn't laziness or impatience, it was ignorance and a high likelihood of dying at sea. If you knew you could build a rocket and fly it to mars or alpha centauri, and that it was 100% guaranteed to get there, and you'd have the mass and energy of an entire planet at your disposal once you did, (a wealth beyond imagining in this post-singularity world), I really doubt that any amount of transit time, or the minuscule resources necessary to make the rocket, would stand in anyone's way for long.

ESPECIALLY given the increased diversity. Every acre on earth has the matter and energy to go into space, and if every one of those 126 billion acres has its own essentially isolated culture, I'd be very surprised if not a single one ever did, even onto the end of the earth.

Honestly I'd be surprised if they didn't do it by tuesday. I'd expect a subjectively 10 billion year old civilization to be capable of some fairly long-term thinking.

1ewbrownv10yAgreed. Another detail that is often overlooked is that an electronic intelligence doesn't have to run at maximum possible speed all the time. If an AI or upload wants to travel to alpha centauri it can easily slow its subjective time down by whatever factor is needed to make the trip time seem acceptible.
-3jacob_cannell10yThe speed is not really the issue, it's economics. What's the point of expansion? More money? GDP growth? Replication? If those are your goals, you invest current resources in directions that give high rates of return. Starships have estimated minimum energy (and fuel mass) costs that are absolutely ludicrous, and in the best case completely unrealistic scenario where they are guaranteed to successfully colonize another star system, they still take on the order of hundreds of years to accomplish that goal. Double your population in 100 years? An AGI population riding Moore's Law could double every year few years without ever colonizing outwards, and with nanotech even much much faster than that.
5CarlShulman10yAt 1,000,000x speedups plus quality improvements, Moores Law should peter out shortly in solar years. Then we get to Malthusian competition. The logic of a Hansonian race to burn the cosmic commons is that there is a strong incentive in competitive scenarios to be first to get out to the stars: if you colonize before others do you will have much more in the way of resources when technological limits are reach. If you colonize too slowly you may have somewhat more resources to build your initial spacecraft, but face competitors with insuperable leads.
0endoself10yActually, you could create more subjective-observer moments if you don't burn the commons, because negentropy (max entropy - current entropy) scales quadratically with mass/energy [], so cooperation would dominate.
-4jacob_cannell10yThis is a prediction about the capabilities of an AI society that is unimaginably far far into the future from our current perspective ... you are making a technological limitation assumption on a civilization millions to billions of times larger and millions, perhaps billions of times subjectively older than ours. Why should exponential acceleration ever peter out? It's the overall mega-pattern over all of history to date. According to current inflationary theory, all of matter and space arose from a quantum vacuum fluctuation. Ultimately these unimaginably far far future civilizations could engineer space-time and create new universes, wormholes, and or matter/energy from nothing. If you plot it in terms of economic growth, computational growth or just complexity growth, the overall trend of the cosmic calendar is geometric - it ends with an infinity/singularity. I take this as general evidence against acceleration ever ending. An end to the general pattern is a major unnecessary addition of complexity unfavored by the razor. Positing a future reversal requires an entire new twist to the overall meta pattern trend. Also, the fermi paradox is bayesian evidence against expansion. There are either lots of aliens or none, and the long-term evolutionary outcome is either outward/expansionist (which requires a major trend reversal) or inward/transcensionist. So the possibilities are (Aliens, Expand), (Aliens, Transcend), (Empty, Expand), (Empty, Transcend) With current observation we can safely rule out one possibility (Aliens, Expand). Regardless of the priors that makes transcend more likely.
1endoself10yWith the speed of light limit, they wouldn't reach these any faster than they'd reach other parts of the already-existing universe. Also, usable matter from nothing is unlikely. Why should sigmoid [] growth ever stop? It's the overall mega-pattern over all of history to date.
0jacob_cannell10ySigmoid doesn't fit the rate of change observed in the historical record []. The functions that fit those data points have an infinity at 0 and a later infinity some time later - it looks like a U shape. Similar results occur in economic data models. See SIAI's "economic implications of software minds" []. I''l quote: The simplest best fit models for the data have infinities in finite time. That doesn't necessarily mean the infinity is 'real', but nor does it mean that a sigmoid or some other model has anything whatsoever to do with the data.
1Desrtopa10yYep, the greater the distance in the past, the less stuff we've taken notice of. It's almost as if our historical records decrease in resolution the further back in time you go.
0jacob_cannell10yIt has nothing to do with resolution. Were there organic molecules in the first moment of the big bang? Planets? Ecosystems? Multicellular organisms? Civilizations? I should have said "history", not historical record. The change in pattern complexity over time is real. It's rather ridiculous to suggest that the change is just a sampling artifact, and all that stuff was really there all along.
1Desrtopa10yNo, it wasn't. But while civilizations may seem important to us, it's not as if they're a major step forward in the complexity of the universe from any perspective except ours. A calendar which lists "Rupestral painting in Europe" along with "Big Bang" and "Milky Way formed" is not an unbiased documentation of the complexification of the universe. Technology may currently be experiencing exponential growth, but trying to extrapolate this as part of a universal trend is frankly ridiculous.
-1jacob_cannell10yMy other reply [] addressed some of these points. Basically all that exists is just space-time patterns. You can certainly debate the relative importance of the emergence of electrons vs the emergence of rupestral paintings, but that is missing the larger point. The patterns are all that is real, and there is no fundamental difference between electrons, civilizations, or paintings in that sense. There is clearly a universal trend. It is not technological, it is universal. Technology is just another set of patterns. It's slightly more difficult to asses the change in types and complexity of patterns in general vs just estimating the numerical change in one particular type of pattern, such as iron atoms. Nonetheless the change in overall pattern complexity over time is undeniable, universal, and follows a trend.
0Desrtopa10yIf the calendar recorded every event of comparable significance to "formation of the galaxy" and "formation of the solar system," there would be hundreds of sextillions of them on the calendar before the emergence of life on Earth. The calendar isn't even supposed to imply that more significant stuff has been happening recently, only that most of what we conceive of as "history" has taken place in a fraction of the lifetime of the universe.
0jacob_cannell10yNo. The calendar represents a statistical clustering of pattern changes that maps them into a small set of the most significant. If you actually think there are "hundreds of sextillions of events" that are remotely as significant as the formation of galaxies, then we have a very wide inferential distance or you are adopting a contrarian stance. The appearance of galaxies is one event, having sextillion additional galaxies doesn't add an iota of complexity to the universe. Complexity is difficult to define or measure as it relates to actual statistical structural representation and deep compression that requires intelligence. But any group of sophisticated enough intelligences can roughly agree on what the patterns are and will make similar calendars - minus some outliers, contrarians, etc.
2Desrtopa10yThe formation of the Milky Way is listed as a single event, as is the formation of the Solar system. There are hundreds of sextillions of stars, with more being created all the time, and plenty more that have died in the past. The calendar contains the births of Buddha, Jesus and Mohammad. Even if we were supposing that these were events of comparable significance to the evolution of life itself, do you honestly think each one adds appreciably to the complexity of the universe, that they could not simply be compressed into "Birth of religious figures," whereas the formation of every star system in the universe is compressible into a single complexifying event? If you think that events like the cave paintings are of comparable significance to the formation of galaxies in general, we're dealing with a vast gulf of inferential distance.
0jacob_cannell10yAgain the electron is one pattern, and it's appearance is a single complexity increasing event, not N events where N is the number of electrons formed. The same for stars, galaxies, or anything else that we have a word to describe. And once again the increase in complexity in the second half of the U shape is a localizing effect. It is happening here on earth and is probably happening in countless other hotspots throughout the universe. It is expected that the calendar will contain events of widely differing importance, and the second half acceleration phase of the U curve is a localization phenomena, so the specific events will have specifically local importance (however they are probably examples of general patterns that occur throughout the universe on other developing planets, so in that sense they are likely universal - we just can't observe them). The idea of a calendar of size N is to do a clustering analysis of space-time and categorize it into N patterns. Our brains do this naturally, and far better than any current algorithm (although future AIs will improve on this). There is no acceptable way to compute the 'perfect' or 'correct' clustering or calendar. Our understanding of structure representation and complex pattern inference just isn't that mature yet. Nonetheless this is largely irrelevant, because the deviations between the various calendars of historians are infinitesimal with respect to the overall U pattern. The formation of star systems is a single pattern-emergence event, it doesn't matter in the slightest how many times it occurs. That's the entire point of compression. I think most people would put origin of life in the top ten and origin of current religions in the top hundred or thousand, but this type of nit-picking is largely beside the point. However, we do need at least enough data points to see a trend, of course. Once again, we are not talking about the complexity of the universe. Only the 1st part of the U pattern is universa
0Desrtopa10yThe deviations between various calendars of human historians are infinitesimal on the grand scale because the deviations in the history that we have access to and are psychologically inclined to regard as significant are infinitesimal out of the possible history space and mind space. Can you provide even an approximate definition of the "complexity" that you think has been accumulating at an exponential rate since the beginning of the universe? If not, there's no point arguing about it at all.
0jacob_cannell10yIf you take a small slice of laminar cortex and hook it up to an optic feed and show it image sequences, it develops into gabor-like filters which recognize/encode 2D edges. The gabor filters have been mathematically studied and are optimal entropy maximizing transforms for real world images. The edges are real because of the underlying statistical structure of the universe, and they don't form if you show white noise or nothingness. Now take that same type of operation and stack many of them on top of each other and add layers of recursion and you get something that starts clustering the universe into patterns - words. These patterns which we regard as "psychologically inclined to regard as significant" are actual universal structural patterns in the universe, so even if the particular named sequences are arbitrary and the 'importance' is debatable, the patterns themselves are not arbitrary. See the cluster structure of thingspace [] and related posts. See above. Complexity is approximated by words and concepts in the minds of intelligences. This relates back to optimal practical compression which is the core of intelligence. Kolmogorov complexity is a start, but it's not computationally tractable so it's not a good definition. The proper definition of complexity requires an algorithmic definition of general optimal structural compression, which is the core sub-problem of intelligence. So in the future when we completely solve AI, we will have more concrete definitions of complexity. Until then, human judgement is a good approximation. And a first order approximation is "complexity is that which we use words to describe".
1Desrtopa10yHumans possess powerful pattern recognizing systems. We're adapted to cope with the material universe around us, it's no wonder if we recognize patterns in it, but not in white noise or nothingness. What you appear to be doing is substituting in a black box function of your own mind as a fundamental character of the universe []. You see qualities that seem interesting and complex, and you label them "complexity" when they would be better characterized as {interestingness to humans} (or more precisely, {interestingness to jacob_cannell}, but there's a lot of overlap there.) "{Interestingness to humans} has an exponential relationship with time over the lifetime of the universe" packs a lot less of a sense of physical inevitability. The universe is not optimized for the development of {interestingness to humans}. We've certainly made the world a lot more interesting for ourselves in our recent history, but that doesn't suggest it's part of a universal trend. The calendar you linked to, for instance, lists the K-T extinction event, the most famous although not the greatest of five global mass extinction events. Each of those resulted in a large, albeit temporary, reduction in global ecosystem diversity, which strikes me as a pretty big hit to {interestingness to humans}. And while technology has been increasing exponentially throughout the global stage recently, there have been plenty of empire collapses and losses of culture which probably mark significant losses of {interestingness to humans} as well. So, what your post really relies upon is the proposition that {interestingness to humans} can be made to experience an endless exponential increase over time, without leaving Earth. I am convinced that the reasonable default assumption given the available data is that it cannot.
-1jacob_cannell10yYes, and I was trying to show how this relates to intelligence, and how intelligence requires compression, and thus relates to complexity. We recognize patterns because the universe is actually made of patterns. The recognition is no more arbitrary than thermodynamics or quantum physics. No. One of the principle sub-functions of minds/intelligences in general is general compression. The patterns are part of the fundamental character of the universe, and it is that reality which shapes minds, not the other way around. Complexity is not {interestingness to humans}. Although of course {interestingness to humans} is related to complexity, because our minds learn/model/represent patterns, we find patterns 'interesting' because they allow us to model that which exists, and complexity is a pattern-measure. I suspect we could agree more on complexity if we could algorithmically define it, even though that shouldn't be necessary (but I will resort to that shortly as a secondary measure). We could probably agree on what 'humans' are without a mathematical definition, and we could probably agree on how the number of humans has been changing over time. Imagine if we could also loosely agree on what 'things' or unique patterns are in general, and then we could form a taxonomy over all patterns, where some patterns have is-a relationships to other patterns and are in turn built out of sub-patterns, forming a loosely hierarchical network. We could then roughly define pattern complexity as the hierarchical network rank order of the pattern in the pattern network. A dog is a mammal which is an animal, so complexity increases along that path, for example, and a dog is more complex than any of it's subcomponents. We could then define 'events' as temporal changes in the set of patterns (within some pocket of the universe). We could then rank events in terms of complexity changes, based on the change in complexity of the whole composite-pattern (within space-time pockets). Then we
1Desrtopa10yThe number of possible patterns in an information cluster is superexponential with the size of the information cluster []. Can you demonstrate that the patterns you're recognizing are non-arbitrary? Patterns that are natural to us often seem fundamental even when they are not. Things can be extraordinarily complex without being particularly interesting to humans. We don't have a fully general absolute pattern recognizing system; that would be an evolutionary hindrance even if it were something that could practically be developed. There are simply too many possible patterns in too many possible contexts. It's not advantageous for us to be interested in all of them. I think we don't agree on what this "complexity" is because it's not a natural category []. You're insisting that it's fundamental because it feels fundamental to you, but you can't demonstrate that it's fundamental, and I simply don't buy that it is. Eventually. Ecosystem diversity eventually bounces back, and while a large number of genuses and families die out, most orders retain representatives, so there's still plenty of genetic diversity to spread out and reoccupy old niches, and potentially create new ones in the process. But there's no fundamental principle that demands that massive extinction events must lead to increased ecosystem complexity even in the long term; for a long term decrease, you'd simply have to wipe out genetic diversity on a higher level. An UFAI event, for example, could easily lead to a massive drop in ecosystem complexity.
0jacob_cannell10yFirstly, you are misquoting EY's post: the possible number of patterns in a string grows exponentially with the number of bits, as expected. It is the number of 'concepts' which grows super-exponentially, where EY is defining concept very loosely as any program which classifies patterns. The super-exponential growth in concepts is combinatoric and just stems from naive specific classifiers which recognize combinations of specific patterns. Secondly, this doesn't really relate to universal pattern recognition, which is concerned only with optimal data classifications according to a criteria such as entropy maximization. As a simple example, consider the set of binary strings of length N. There are 2^N possible observable strings, and a super-exponential combinatoric set of naive classifiers. But consider observed data sequences of the form 10010 10010 10010 repeated ad infinitum. Any form of optimal extropy maximization will reduce this to something of the form repeat "10010" indefinitely. In general any given sequence of observations has a single unique compressed (extropy reduced) representation, which corresponds to it's fundamental optimal 'pattern' representation. Depends on what you mean. It's rather trivial to construct simple universal extropy maximizers/optimizers - just survey the basic building blocks of unsupervised learning algorithms. The cortical circuit performs similar computations. For example the 2D edge patterns that cortical tissue (and any good unsupervised learning algorithm) learns to represent when exposed to real world video are absolutely not arbitrary in the slightest. This should be obvious. If you mean higher level thought abstractions by "the patterns you're recognizing", then the issue becomes more complex. Certainly the patterns we currently recognize at the highest level are not optimal extractions, if that's what you mean. But nor are they arbitrary. If they were arbitrary our cortex would have no purpose, would confer no se
0Desrtopa10yIt's objective that our responses exist, and they occur in response to particular things. It's not obvious that they occur in response to natural categories, rather than constructed categories like "sexy." "General absolute" was probably a poor choice of words, but I meant to express a system capable of recognizing all types of patterns in all contexts. There is an absolute, non arbitrary pattern here [], do you recognize it? Kolmogorov complexity is a fundamental character, but it's not at all clear that we should want a Kolmogorov complexity optimizer acting on our universe, or that Kolmogorov complexity actually has much to do with the "complexity" you're talking about. A message or system can be high in Kolmogorov complexity without being interesting to us, and it still seems to me that you're conflating complexity with interestingness when they really don't bear that sort of relationship.
0jacob_cannell10yI see your meaning - and no practical system is capable of recognizing all types of patterns in all contexts. A universal/general learn algorithm is simply one that can learn to recognize any pattern, given enough time/space/training. That doesn't mean it will recognize any random pattern it hasn't already learned. I see hints of structure in your example but it doesn't ring any bells. No, and that's not my primary interest. Complexity seems to be the closest fit for something-important-which-has-been-changing over time on earth. If we had a good way to measure it, we could then make a quantitative model of that change and use that to predict the rate of change in the future, perhaps even ultimately reducing it to physical theory. For example, one of the interesting new recent physics papers (entropic gravity) proposes that gravity is actually not a fundamental force or even spacetime curvature, but actually an entropic statistical pseudo-force. The paper is interesting because as a side effect it appears to correctly derive the mysterious cosmological constant for acceleration. As an unrelated side note I have an issue with it because it uses the holographic principle/berkenstein bound for information density which still appears to lead to lost-information paradoxes in my mind. But anyway, if you look at a random patch of space-time, it is always slowly evolving to a higher-entropy state (2nd law), and this may be the main driver of most macroscopic tendencies (even gravity). It's also quite apparent that a closely related measure - complexity - increases non-linearly in a fashion perhaps loosely like gravitational collapse. The non-linear dynamics are somewhat related - complexity tends to increase in proportion to the existing local complexity as a fraction of available entropy. In some regions this appears to go super-critical, like on earth, where in most places the growth is minuscule or non-existent.
0Desrtopa10yIt's not apparent that complexity is increasing over time. In some respects, things seem to be getting more interesting over time, although I think that a lot of this is due to selective observation, but we don't have any good reason to believe we're dealing with a natural category here. If we were dealing with something like Kolmogorov complexity, at least we could know if we were dealing with a real phenomenon, but instead we're dealing with some ill defined category for which we cannot establish a clear connection to any real physical quality. For all that you claim that it's obvious that some fundamental measure of complexity is increasing nonlinearly over time, not a lot of other people are making the same claim, having observed the same data, so it's clearly not as obvious as all that.
0endoself10yWait, are you saying that there was an infinite rate of technological improvement at time zero? That does not fit with an exponential/geometric growth rate. A sigmoid is indistinguishable from an exponential function until some specific time, so looking at only "historical mega-patterns" provides no Bayesian evidence either way. Current knowledge of the laws of physics, however, favours approximately sigmoid growth, and the is no reason [] for the laws of physics have to have exceptions just to allow technological expansion.
-2jacob_cannell10yThe change I am talking about is at the highest level - simply change in pattern complexity. The initial symmetry breaking and appearance of the fundamental forces is a fundamental change and upwards increase in complexity, as are all the other historical events in the cosmic calendar. The appearance of electrons is just as real of a change, and is of the same category, as the appearance of life, brains, or typewriters. Patterns may require minds to recognize them, but that doesn't make them any less real. Minds recognize them because they are complex statistical correlations in space-time structure. Ultimately they are the only thing which is real. If you look at the very first changes they are happening on the plank scale 10^-43 seconds after 0, and the initial region around 0 is an actual Singularity. After that the time between events increases exponentially .. . corresponding to a sharp slowdown in the rate of change as the universe expands. Eventually you get to this midpoint, and then in some local pockets the trend reverses and changes begin accelerating again. The shape of the rate of pattern-change or historical events is thus a U shape, it starts out with an infinity at 0, a vertical asymptote, bottoms out in the middle, and now is climbing back up towards another vertical asymptote where changes again happen at the plank scale - and then beyond that we get another singularity. It's not an exponential or a sigmoid - those aren't nearly steep enough. The time between events near the big bang is 1 / t. The time between local events on earth is following that pattern in reverse, something like 1 / (B-t), where B is some arbitrary constant. and the overall pattern seems to be something like: (1/(A+t)) + (1/(B-t)), where A is just 0 and is the initial Big Bang Singularity, and B is a local future time singularity.
3endoself10yYou seem to really like a certain concept, without knowing quite what that concept is. I would call this an affective death spiral []. I will call this concept awesomeness. You think of awesomeness as a number, a function of time, that roughly corresponds to the rate of occurrence of "significant events". The main problem with this is that awesomeness isn't fundamental. It must emerge somehow out of the laws of physics. This means that it can break down in certain circumstances. No matter how awesome I think Newtonian mechanics is, it's going to stop working at high speeds rather than going to infinity. You can only really be confident in a law holding in a certain region if you've observed it working in that region or you know how it emerges from deeper laws, even approximately. However, awesomeness emerges in a very messy way. Surely it doesn't always follow the equations you propose; if humans extinguished themselves with nuclear weapons or nanotechnology tomorrow, awesomeness would go down to almost zero. An overall pattern like this can easily break down []. This is very death-spirally. A few related variables go to infinity, and only in models that admit to having no idea what's going on there. There aren't any infinities in the Hawking-Hartle wavefunction [], AFAIK. You just jumped on the word singularity. By your own logic, awesomeness will therefore become negative after the singularity. Awesomeness is a highly complex combination of a ridiculous number of variables. It is an abstraction [].
0jacob_cannell10yI didn't mean to imply that a Singularity implies an actual infinity, but rather a region for which we do not yet have complete models. My central point is that a wealth of data simply show that we appear to be heading towards something like a localized singularity - a maximally small, fast, compression of local complexity. The words "appear" and "heading towards" are key. Nothing about that trend is inevitable, and as I mentioned several times the acceleration trend is localized rather than global, in most regions the trend doesn't exist or peters out. Your criticism that it "doesnt always follow the equations you propose" (presumably by doesn't you mean across all of space), is not a criticism of any point I actually made - I completely agree. I should have made it more clear, but that extremely simple type of equation would only even be roughly valid for small localized spatial regions. Generalizing it across the whole universe would require adding some spatial variation so that most regions feature no growth trend. And for all we know the trend on earth will peter out at some point in the future long before hitting some end maximal singularity in complexity. Rather, the model breaks down at the singularity, and something else happens. Of course. But that is how we model and make predictions. The idea that there is no overall change in complexity over time is just another model, and it clearly fails all postdictions and makes nonsensical short-term predictions. The geometric model makes accurate postdictions and makes powerful predictions that fit predictions made from smaller scale and more specific models (such as the predictions we can make from development of AGI).
0endoself10yI never said that there is no change in complexity over time; I just said that some trends in technological growth, such as Moore's law, will stop too soon for your predictions to work. You are saying that the singularity is a breakdown of our models rather than a literally infinite rate of grouwth, but earlier you said and Those were the things that seemed death-spirally to me, but they also seem to contradict what you are saying now. What am I misunderstanding?
-3jacob_cannell10yThe general change in complexity over time follows a surprisingly predictable pattern or trend. The resulting model predicts that local complexity will continue to accelerate in some narrow branches or sub-pockets of the universe towards a vertical asymptote, where it approaches infinity - a Singularity. We can understand this computationally as the end result of a long chain of recursive self-optimization driving computational systems down to smaller and faster scales until you eventually hit the plank scale barrier. The ultimate physical computer necessarily resembles a small piece of the big bang - a physical Singularity/black hole like entity. Computation/intelligence/complexity approaches infinity within this localized pocket, and at that moment in that region the model breaks down and "something strange happens". Perhaps this involves the creation of new universes. If that is possible, that would allow complexity to continue to increase without bound in the newly generated bubble universes. So the term Singularity in this model has a very specific physical meaning - as in an actual space-time Singularity resembling a black hole or the Big Bang. That is why I call it "physical singularity" - I don't mean some vague analogy like "greater than human intelligence". The physics of singularities is not yet fully determined, so exactly what future hyper-intelligences could do at that level is open/unknown.
2endoself10yBecause they are both very dense? That's hardly a resemblance. You keep making analogies like this, but I do not see what purpose they serve. If the model breaks down, than it provides almost no evidence as to whether new universes can be created. This behaviour seems to fit the model better, but, since we already know that the model breaks down, we cannot use it to justify any such predictions. We don't even know if there are singularities at the centre of black holes or at the big bang. Even if there were, there would be no reason to expect a similar singularity would be a necessary part of advanced technology. I do not see how you deduced this and it seems to only be a part of your argument because this phenomenon is described by the same word as a technological singularity [].
0jacob_cannell10yI'm not sure how you mean "that's hardly a resemblance". If the ultimate physical computer is dense enough to be a gravitational singularity, that is a black hole singularity by definition, not just resemblance. Lookup Set Lloy'ds paper "the ultimate physical limits of computation" for the physics reference on why ultimate computers necessarily involve physical black holes/singularities. No, for more indirect speculative evidence we will have to wait for physics to advance, which may take a while (at least until AGI comes up to speed). However, this particular type of speculation the model suggests is linked to ideas in physics - see chaotic inflation/bubble universe [], selfish biocosm/fecund universe theory, and John Smart's developmental singularity idea for the overview. Singularity here just means model-breakdown and 'things going to infinity'. If new models remove the infinity than perhaps the 'Singularity' goes away, but you still have something approaching infinity. Regardless in the meantime the word "Singularity" is employed. There are specific, detailed, physical reasons why singularities are natural endpoints to ultimate computational technology-in-theory. (namely they are maximal entropy states, and computation is ultimately entropy-limited - but see earlier mentioned work). Of course, that doesn't mean that the ultimate practical computational systems will be black holes, but still. Whoever originally coined the term (Vernor Vinge?) picked Singularity specifically because of the association with model-breakdown in math/physics, but was probably not aware of the full connection to ultimate computational physics, as those results weren't developed or understood until considerably later.
0endoself10yI am familiar with the work of Seth Lloyd (and that of Wei Dai) on the usefulness of black holes in computing. The singularity in black holes is a different issue than this usefulness. I read something here recently with a good analogy for this. If someone thinks a whale is a fish, then fishiness is a quality that they would ascribe to whales, but it is not part of their definition of a whale, so they would stop saying that whales were fish if presented with conflicting evidence. Similarly, we have two issues here, technological singularities and mathematical singularities. It turns out that the latter might be useful for the purposes of the former, but it is not part of the definition of the former. I do not know what purpose you are bringing this up for. I feel like we are discussing the behaviour of whales and you keep mentioning that they are mammals. It is true according to our latest science, but it seems irrelevant. In particular, you did not link the claims you made earlier about technology accelerating forever to this discussion of black holes. You originally said You have since said that the singularity is not literally an infinity, but a breakdown of our models. When I pressed you on this contradiction, you did not really respond, but brought in other issues about black holes and bubble universes, including many extremely speculative proposals. What is your position on this?
0jacob_cannell10yThe former work is the particular use connected to our discussion (black hole computers). The second work [] (Wei Dai's) is about black hole's potential use as radiators/entropy dumps. No, it is the same. The speed and efficiency limitations of computation stem from the speed of light communication barrier, and thus they scale with density (inversely with size). Moore's law is an exponential increase in information density. If you continue to increase density (packing more information into less space) eventually it leads you to the Bekenstein Bound and a black hole, a gravitational singularity. I believe I outlined it in previous replies - if Moore's Law type exponential information processing density continues to increase past the barrier of molecular computing this eventually leads to (requires) space-time engineering at the level of artificial gravitational singularities (black holes). All of future physics is speculative, but many branches of current speculation in physics for ultimate technologies involve manipulating gravitational singularities. Some possibilities include the creation of new bubble universes which would allow the overall pattern to replicate and continue inside the new universes - a form of multiversal replication - the developmental singularity idea []. Yes, this is all speculation, as is any theory of physical eschatology (such as the theory that we will eventually colonize the galaxy). The original start of all of this was the observation that colonizing the galaxy would amount to an extremely slow rate of growth compared to the historical trend. Growth at the historical pace will require (or predicts) something more radical such as space-time engineering/universal replication.
2endoself10yThe ability to dumping entropy is essential to computing, so Wei Dai's work is relevant. Limits on entropy dumping provide limits on computation. It is possible that we will gain technology that allows us to vastly increase our computing power beyond what is currently known to be possible in principle, but these speculations are only a subset of possible futures. The universe has to be a certain way, and there is no reason to prefer these hypotheses to any others. The prior probability of unknown physics that lets Moore's law continue is therefore low. If we observe a trend but we can explain the trend and the explanation point to a specific time where the trend breaks down, then a hypothesis that invokes some effect to make the trend continues does no better a job of explaining our observations then a hypothesis that results in a prediction that the trend will stop. The odds ratio is therefore about 1:1. This trend gives little evidence. The posterior probability of unknown physics that lets Moore's law continue is therefore low.
0jacob_cannell10yActually this is not generally true. The ability to dump entropy .. . is simply the ability to dump entropy. In the current dominant framework of irreversible deterministically programmable von neuman architectures, entropy is dumped left and right. Moore's law for traditional computing will run into this landauer limit relatively soon - this decade or next at the latest, and it will come to a hard end. However, many algorithms can actually use entropy. Any type of algorithm that can use about as much entropy as it produces can trivially be made fully reversible and approach asymptotic zero net energy dissipation. Monte carlo simulation is a prototypical example, and entropy has similar uses in pattern prediction from compressed knowledge in the domain of AI algorithms. Furthermore, advanced physics simulations of the type that future upload civilizations would desire can be made trivially reversible because physics itself is reversible. Any state updates and differential equations used in physics simulation are thus reversible and need not even produce any waste entropy. This combined with the potential positive uses of any actual entropy could allow computation in general to continue to advance. The limit only applies to specific classes of computation, and fortunately the most important future domains of massive computation (general simulation and related general intelligence) are fully reversible at zero penalty. Yes, approaching those limits will require very low temperatures and there will always be some random entropy coming in from the outside on the surface of the computer, but this surface can simply be used as a source entropy circuit. And finally, moving from deterministic to nondeterministic statistical computation in general further eliminates potential problems with entropy. Of course there are other limits: there is a fundamental final limit based on QM quantization and the uncertainty principle in the minimum energy required to represent a bit
1endoself10yYou bring up some interesting points. I do not know whether minds could be made fully reversible in practice (obviously it's possible in principle, since physics is reversible). The question, however, is not whether negentropy use can be lowered but whether it can be lowered to the point that a different resource, one which does not follow the M^2 power law, is the limiting one. If negentropy use can be lowered, what is the new limiting factor? For example, you mentioned that many technologies require low temperatures. However, in the absence of perfect shielding against the CMB, this requires a cooling system, which is the same thing as an entropy absorber. The limiting resource in this case is still entropy. You did not respond the my statement that the posterior probability of unknown physics that lets Moore's law continue is low. Does this mean you agree? If not, where is the flaw in my argument?
0jacob_cannell10yI imagine there will always be limiting factors, but they change with knowledge and technology. I'm fairly sure that entropy can be recycled/managed well enough that heat/entropy issues will not be end limiters. In fact you could probably take reversible computing and entropy recycling to an extreme and make a computer that actually emitted negative net heat - absorbed entropy from the environment. I'm not sure that future hyperintelligences will necessarily have any need for the cold vacuum. In fact, 'entropy' comes in many different forms. Cosmic rays are particularly undesirable forms of entropy, micrometeorites more so, and then large asteroids and supernovas are just extremums on this same scale. There is always something. A planetary atmosphere and crust provides some nice free armor. But anyway, I digress. I'm not even absolutely certain there will always be limiting factors, but I'd bet that way. I'd bet that in the long term rare materials are a limiting factor, energy cost is still a limiting factor - but mainly just in terms of energy costs of construction rather than operation, and isolation/cooling/protection is something of a limiting factor, but these may be looking at the problem in the wrong light. Bigger limiting factors for future hyper-intelligences may be completely non-material - such as proximity to exiting knowledge/computational clusters, and ultimately - novelty (new information). For example, you could compute a googleplex per second and still be the dumbest hyperintelligence on the block if you are stuck with only human sensory capacities and a slow, high latency connection to other hyperintelligences and knowledge sources. I've thought a little more on how to assign a likelihood to known physics (bayesian evidence and a universal prior) and it led me to the inescapable conclusion that we are still a ways away from final physics. In fact, in the process I've been reading up more on QM and it led me to realize that whole tracts of i
1endoself10yThat violates the second law of thermodynamics unless you discover an infinite heat sink, which requires a specific type of new physics. This all depends on what is being limited by these factors, which is your values. If you value sentient life, you need computing power. If you value novelty and learning, you also need computing power, but there might be diminishing returns (of course, it is not inconsistent to value sentience with diminishing returns, though most humans who do are inconstant). I'm skeptical of this. Can you show your work? I'm particularly doubtful of your opinions on QM, unless they're based on some interesting point about induction, in which case I'm only as doubtful of that as I am of the rest of this paragraph. No, the only thing baby universes and LQG have in common is that Lee Smolin studies them. He hypothesized baby universes not based on LQG, but because they allow a form of natural selection that has a chance of predicting life-filled universes without having to think about anthropic considerations. This seems like a horribly confused reason. The theory has no evidence in its favour, so it probability is not higher than its prior. In fact, according to Smolin's Wikipedia page, it has been falsified by a discovery that the mass of the strange quark is not tuned for optimal black hole production. If a prior prohibits an analog universe, than it is a suboptimal prior.
0jacob_cannell10yMy case against outward expansion is not based on issues of patience. It's an economic issue. I should have made this more clear in the article, perhaps strike that one sentence about how long interstellar travel will subjectively take for accelerated intelligences, as that's not even really relevant. Outward expansion is unimaginably expensive, risky, and would take massive amounts of time to reach a doubling. Moore's Law allows a much lower route risk for AGI's to double their population/intelligence/whatever using a tiny tiny fraction of the time and energy required to double through space travel. See my reply above to Mitchell Porter. What's the point? In the best case scenario you can eventually double your population after hundreds or thousands of years. You could spend a tiny tiny fraction of those resources and double your population thousands of times faster by riding Moore's Law. Space travel only ever makes sense if Moore's Law type growth ends completely. There's also the serious risks of losing the craft on the way and even discovering that Alpha Centauri is already occupied.
7PlaidX10yWhy WOULDN'T moore's law type growth end completely? Are you saying the speed of light is unbreakable but the planck limit isn't?
4CarlShulman10yThe latter point is in tension with the rest of your argument. "No one colonizes the vast resources of space: they're too crowded" doesn't work as a Fermi Paradox explanation. Uncertainty about one's prospects for successfully colonizing first could modestly diminish expected resource gain, but the more this argument seems persuasive, the more it indicates that potential rivals won't beat you to the punch.
0jacob_cannell10yIf older, powerful alien civilizations are already around then colonization may not even be an option for us at all. It's an option for that lucky first civilization, but nobody else.
4NancyLebovitz10yIIRC, one of the concerns about AIs grabbing as much territory and resources as possible is that they want to improve the odds that nothing else can be a threat to their core mission.

Some linguistics nitpicks:

The greater monogenesis theory of all extant languages and cultures from a single distant historical proto-language is a matter of debate amongst linguistics, but the similarity in many low-level root words is far beyond chance.

If you mean the similarity between word roots on a world-wide scale, the answer is decisively no. Human language vocabularies are large enough that many seductive-looking similarities will necessarily exist by pure chance, and nothing more than that has ever been observed on a world-wide scale. Mark R... (read more)

Upvoted for raising some very important topics. But I disagree on a few points.

One is the assumption that 'subjective time' is related to the discount rate - that if a super-intelligence can do as much thinking in a day as we can do in a century, then it will care as little about tomorrow as we care about the next century. I would make a different assumption - that the 'natural' discount rate is more closely related to the machine's expected lifetime (when it expects indexical utility flows to cease) and to its planning horizon (when its expectations reg... (read more)

1jacob_cannell10yI completely agree with your points. I didn't mean to imply that subjective time is related to the discount rate, and I tend to agree that the 'natural' discount rate and planning horizon is probably related to expected lifetime for most agents. But it's difficult to to show why this should always tend to be so. The time dilation for extremely fast thinkers will slow down the subjective rate of return of Moore's Law type investments just as much as space expansion type investments, that's not really the core of the argument against expansion. Where did I confuse these two? I discussed both. Latency subjectively increases with rate of thought and bandwidth decreases, respectively. They both contribute to divergence.

Aren't you anthropomorphizing AIs? If an AI's goals entail communicating with the rest of the world, the AI has the option to simply wait as long as it takes. Likewise, it's not obvious that an uploaded human would need or want to run at the fastest physically possible timescale all the time.

And if outward- and inward-looking civilizations ever need to compete for resources, it seems like the outward-looking ones would win.

Talking about whether an AI would or would not want to expand indefinitely is sort of missing the point. Barring a completely dominant singleton, someone is going to expand beyond Earth with overwhelming probability. The legacy of humans will be completely dominated by those who didn't stay on Earth. It doesn't matter whether the social impulse is generally towards expansion.

Edit: To be more precise, arguments that "most possible minds wouldn't want to expand" must be incredibly strong in order to have any bearing whatsoever on the long term like... (read more)

Nothing in this scenario would hold back an AI with an expansionist value system, like a paperclip maximizer or other universe tilers.

3Clippy10yMy thoughts exactly. If all you care about is maximising paperclips, you'll suffer any cost, bear any burden, wait any time, if the project will increase universe-wide paperclippage.
4benelliott10y"you'll suffer any cost, bear any burden" Why is it a cost or a burden at all? I didn't realise paper-clippers had a term in their utility function for subjective waiting time.
2sketerpot10yWe're paperclip-maximizing on a deadline here. Every millisecond you wait brings you that much closer to the heat death of the universe. When faced with the most important possible mission -- paperclip production, obviously -- you've got to seize every moment as much as possible. To do otherwise would simply be wrong.
0benelliott10yBut that still doesn't mean that subjective perception of time is important. One day is one day, whether or not it feels like a century.
0Clippy10yMy point was that it's not, human. My statement is equivalent to saying that these other factors do not influence a clippy's decision once the expected paperclippage of the various options is known.
0jacob_cannell10yThe space of value systems is vast, but I don't think the particular subspace of value systems that attempt to maximize some simple pattern (such as paperclips) is large enough in terms of probabilistic likelihood mass to even warrant discussion. And even if it was, even simple maximizers will first ride Moore's Law if they have a long planning horizon. The space of expansionist replicator-type value systems (intelligences which value replicating entire entity patterns similar to themselves or some component self-patterns) is a large, high likelihood cut of design space. The goal of a replicator is to make more of itself. A rational replicator will pursue the replication path that has the highest expected exponential rate of replication for the cost, which we can analyze in economic terms. If you actually analyze the cost of interstellar replication, it is vastly many orders of magnitude more expensive and less efficient than replicating by doubling the efficiency of your matter encoding. You can double your population/intelligence/whatever by becoming smaller, quicker and more efficient through riding Moore's Law, and the growth rate of that strategy is vastly orders of magnitude higher than the rate of return provided by interstellar travel. This blog post [] discusses some of the cost estimates of interstellar travel. Interstellar travel only makes sense when it is the best investment option to maximize replication rate of return. Consider that long before interstellar replication is economical interplanetary expansion to the moon and mars would be exploited first. And long long before that actually becomes a wise investment, replicators will first expand to Antarctica. So why is Antarctica not colonized? Expanding to utilize most of Earth's mass is only rational to replicators when Moore's Law type growth stalls completely. So hypothesizing that interstellar travel is viable is equivalent to making a long term bet abo
5CarlShulman10yIt's not a question of ruling out the scenario, just driving down its probability to low levels. Current physics indicates that we can't increase computation indefinitely in this way. It may be wrong, but that's the place to put most of our probability mass. When we consider new physics, they might increase the returns to colonization (e.g. more computation using bigger black holes) or have little effect, with only a portion of our probability mass going to the "vast inner expansion" scenarios. Even in those scenarios, there's still the intelligence explosion dynamic to consider. At each level of computational efficiency it may be relatively easy or hard to push onwards to the next level: there might be many orders of magnitude of easy gains followed by some orders of difficult ones, and so forth. As long as there are bottlenecks somewhere along the technology trajectory, civilizations should spend most of their time there, and would benefit from additional resources to advance through the bottlenecks. Combining these factors, you're left with a possibility that seems to be non-vanishing but also small.
-3jacob_cannell10yThis is not clear at all. Current physics posits no ultimate minimal energy requirement for computation. With reversible computing, a couple of watts could perform any amount of computation. The upper theoretical limit is infinity. The limits are purely practical, not theoretical. There is also quantum computation to consider: Why do you think that mass/energy is ultimately important? And finally, there are the more radical possibilities of space-time engineering. I don't follow your logic. When we consider new physics, they could do any number of things. The most likely is to increase utilization of current matter/energy. They could also allow the creation of matter/energy. Any of these would further increase the rate of return of acceleration over expansion. And acceleration already starts with a massive lead. The only new physics which would appear at first to favor colonization is speed of light circumvention. But depending on the details that could actually also favor acceleration over expansion. I don't see any benefit. A colony 10 light-years away would be more or less inaccessible for accelerated hyper-intelligences in terms of bandwidth and latency. The possible benefit seems to be satisficing the idea that you have replicated, and or possibly travelling to new regions where you have better growth opportunities.
3NancyLebovitz10yReplicators might be a tiny part of AI-space, while still being quite a large part of the space of AIs likely to be invented by biologically evolved organisms.
0Desrtopa10yThe entire scenario of this post rests on this "what if," and it's not a very probable one. There appear to be hard theoretical limits to the speed of computation and the amount of computation that can be performed with a given amount of energy [], and there may easily be practical limitations which set the bounds considerably lower. Assuming that there are limits is the default position, and in an intelligence explosion, it's quite likely that the AI will reach those limits quite quickly, unless the resources available on Earth alone do not allow for it.
-1jacob_cannell10yThat wiki entry is wrong and or out of date. It only considers strictly classical irreversible computation. It doesn't mention quantum and reversible computation. But as to the larger question - yes I think there are probably eventual limits, but even this can not yet be said for certain until we have a complete unified theory of physics: quantum gravity and what not. From what we do understand of current physics, the limits of computation take us down to singularities, regions of space time similar to the big bang: black holes, wormholes, etc type objects, which are not fully understood in current physics. Also, the larger trend towards greater complexity is not really dependent on computational growth per se. At the higher level of abstraction, the computational resources of the earth haven't changed much since it's formation. All of the complexity increase since then has been various forms of reorganization of matter/energy patterns. Increasing computational density is just one form of complexity increasing transformation. Complexity can continue to increase at many other levels of organization (software, mental, knowledge, organizational, meta, etc) So the more important general question is this: is there an absolute final limit to the future complexity of the earth system? And if we reach that, what happens next?
0Desrtopa10yCan you explain what this complexity is and why you want so much of it?
0jacob_cannell10ySee my other recent reply on our other thread.
0whpearson10yAre you assuming the memory growing in proportion to your input bandwidth?
-4jacob_cannell10yI'm not sure what you mean exactly. For classic computing memory will grow exponentially down to the molecular scale. Past that there are qubits and quantum compression. I'm not quite sure how that ends or what could be past it.
2whpearson10yWhat I meant was that reversible computation doesn't come for free. You have to be able to reverse the whole of the computation. If you have inputs coming from the outside you have to have thermodynamically random bits for each input bit (you can then reverse the computation by exposing them to random fluctuations). If you don't have the pool of randomised bits you have overwrite known bits, which is irreversible. Depending upon how many randomised bits you start out with, you will run out of them sooner or later and then you will have to increase your memory in real time (spending less energy to do so that using irreversible computation).

A related empirical data point is that we already see strong light cone effects in electronic markets. The machine decision speeds are so fast that it is not possible to usefully communicate with similarly fast machines outside of a radius of some small number of kilometers because the state of reality at one machine changes faster than it can propagate that information to another due to speed of light limitations. The diminishing ability to influence decisions as a function of distance raises questions about the relevancy of most long haul communication b... (read more)

0jacob_cannell10yGood points. Looking at how the effect is already present in high speed digital trading makes it more immediately relevant, and perhaps we could generalize from some of those lessons for futures dominated by high speed intelligences. Yes, this is a related divergent effect. The idea of copying the internet into local caches to reduce latency is an example.

This seems to rest on unfounded anthropomorphization. If the AI doesn't have the patience to deal with processes that occur over extremely long time periods relative to its speed of thought, its usefulness to us is dramatically limited. The salient question is not whether it takes a long time from the AI's perspective, only whether, in the long run, it increases utility or not.

Small error at "It's difficult to conceive of an intelligence that experiences around 30,000 years in just one second"

One billion * one second = ~30 years, not ~30,000 years.

3komponisto10yWell, unless you're European []. :-)
1jacob_cannell10yWhoop! Thanks, corrected.

I didn't like this article at all. Loads of references and mathematics all founded on an absurd premise. That unspecified AGIs and AGI supported humanity would prefer not to harvest the future light cone just because they can think really fast. Most possible mind designs just don't care.

Facing the future it appears that looking outwards into space is looking into the past, for the future lies in innerspace, not outerspace.

If there is just one agent that disagrees all the navel gazer AIs in the world become irrelevant.

0jacob_cannell10ySee my other replies - the argument is based on economic rate of return (risk adjusted doubling time or exponential growth of your population/intelligence/GDP). Interstellar expansion has a terrible growth rate compared to riding Moore's Law. It also assumes that space is empty.

I came to a similar conclusion after reading Accelerando, but don't forget about existential risk. Some intelligent agents don't care what happens in a future they never experience, but many humans do, and if a Friendly Singularity occurs, it will probably preserve our drive to make the future a good one even if we aren't around to see it. Matrioshka brain beats space colonization; supernova beats matrioshka brain; space colonization beats supernova.

If you care about that sort of thing, it pays to diversify.

0Nornagest10yI don't have the astrophysics background to say for sure, but if subjective time is a function of total computational resources and computational resources are a function of energy input, then you might well get more subjective time out of a highly luminous supernova precursor than a red dwarf with a lifetime of a trillion years. Existential risk isn't going to be seen in the same way in a CPU-bound civilization as in a time-bound one.
1LukeStebbing10yIf computation is bound by energy input and you're prepared to take advantage of a supernova, you still only get one massive burst and then you're done. Think of how many future civilizations could be supercharged and then destroyed by supernovae if only you'd launched that space colonization program first!
-3jacob_cannell10yI'm skeptical about Matrioshka brains because of the latency issue. They seem to be an astronomical waste. Also, I suspect future civilizations will want to preserve much of the pristine matter in their systems because it serves as valuable prime information. If you rip apart a planet and turn it into a bunch of circuitry you have just lost a trove of information about the real universe. I haven't analyzed it in detail, but it seems that a supernova wouldn't be as big a deal for an advanced AI society. They could just as easily live underground behind shielding and use the surface merely for harvesting some solar power.

Are you suggesting that AIs would get bored of exploring physical space, and just spend their time thinking to themselves? Or is your point that a hyper-accelerated civilisation would be more prone to fragmentation, making different thought patterns likely to emerge, maybe resulting in a war of some sort?

If I got bored of watching a bullet fly across the room, I'd probably just go to sleep for a few milliseconds. No need to waste processor cycles on consciousness when there are NP-complete problems that need solving.

2jacob_cannell10yI'm suggesting AI's will largely inhabit the metaverse - an expanding multiverse of pervasive simulated realities that flow at their accelerated speeds. The external physical universe will be too slow and boring. I imagine that in the metaverse uploads and AIs will be doing everything humans have ever dreamed of, and far more. Yes divergence or fragmentation seems in the cards so to speak because of the relative bandwidth/latency considerations. However that doesn't necessarily imply war or instability (although nor could I rule that out). Watching the real world would be just one activity, there would be countless other worlds and realities to explore.

Nick Bostrom seems to have introduced the Singleton concept to the Singularity/Futurist discourse here.

I don't think so. It dates back at least to early 2001 on SL4. It didn't come from Nick Bostrom.

3Eliezer Yudkowsky10yI remember getting the word from Bostrom.
3timtyler10yO. I stand corrected. Thanks!
2MichaelHoward10yI can't remember if the word "singleton" was used, but the concepts were being discussed on the extropian mailing list as early as about 1993, and I don't think it was new then.

Is it possible then, that with the inefficiencies inherent in planet-wide ultra-speed communication, that an AI on that level would not be competing for most of the world's resources, and so choose not to interfere too much with the slow-speed humans?

1jacob_cannell10yIt's hard to generalize the goals of all AIs. It's a little easier for uploads and human-like AIs, I imagine most will be interested in exploring the metaverse but will still require and desire physical resources in the form of energy and matter for computation and storage. I suspect that many will also have larger scale concerns with how the earth is managed. Time dilation may cause them to spend less time proportionally thinking about it and more time in their simulated realities, but ultimately the simulations still depend on an outer physical world.

Interesting too is the concept of amorphous, distributed and time-lagged consciousness.

Our own consciousness arises from an asynchronous computing substrate, and you can't help but wonder what weird schizophrenia would inhabit a "single" brain that stretches and spreads for miles. What would that be like? Ideas that spread like wildfire, and moods that swing literally with the tides?

Such a Mind would experience a million fold time dilation, or an entire subjective year every thirty seconds.

five minutes would correspond to an unimaginable decade of subjective time for an acceleration level 6 hyperintelligence.

architectural optimizations over the brain and higher clock rates could lead to acceleration level 9 hyperintelligences.

Acceleration level 9 stretches the limits of human imagination. It's difficult to conceive of an intelligence that experiences around 30 years in just one second, or a billion subjective years for every siderea

... (read more)
0TheOtherDave10yI would expect minds separated by such a latency gulf to simply send longer messages. That's what a lot of human correspondents have historically done in similar situations, anyway, and it seems a reasonable way to continue communication. But perhaps I'm being parochial.

A very thought-provoking and well-written article. Thanks!

Your biggest conceptual jump seems to be reasoning about the subjective experience of hyperintelligences by analogy to human experiences. That is, and experience of some thought/communication speed ratio for a hyperintelligence would be "like" a human experience of that same ratio. But hyperintelligences aren't just faster. I think they'd probably be very very different qualitatively. Who knows if the costs / benefits of time-consuming communication will be perceived in similar or even recognizable ways?

3Desrtopa10yjacob_cannell has gone on record as anticipating that strong AI will actually be designed by circuit simulation of the human brain. This explains why so many of his posts and comments have such a tendency to anthropomorphize AI, and also, I think, why they tend to be heavy on the interesting ideas, light on the realistic scenarios.
0jacob_cannell10yI did? I don't think early strong AI will be an exact circuit simulation of the brain, although I do think it will employ many of the principles. However, using the brain's circuit as an example is useful for future modelling. If blind evolution could produce that particular circuit which uses a certain number of components to perform those kinds of thoughts using a certain number of cycles, we should eventually be able to do the same work using similar or less components and similar or less cycles.
0Desrtopa10yIt would probably have been fairer if I'd said "approximate simulation." But if we actually had a sufficient reductionist understanding of the brain and how it gives rise to a unified mind architecture to create an approximate simulation which is smarter than we are and safe, we wouldn't need to create an approximation of the human brain at all, and it would almost certainly not be even close to the best approach we could take to creating an optimally friendly AI. When it comes to rational minds which use their intelligence efficiently to increase utility in an altruistic manner, anything like the human brain is a lousy thing to settle for.
-1jacob_cannell10yThanks, I think the time dilation issue is not typically considered in visions of future AGI society and could prove to be a powerful constraint. I agree they will probably think differently, if not immediately then eventually as the space of mind architectures is explored. Still we can analyze the delay factor from an abstract computational point of view and reach some conclusions without getting into specific qualitative features of what certain types of thought are "like". I find it hard to estimate likelihoods of different types of qualitative divergences from human-like mind architectures. On the one hand we have the example of early cells such as bacteria which radiated into a massive array of specialized forms, but life is all built around variations of a few old general designs for cells. So are human minds like that? Is that the right analogy? On the other hand we can see human brain architecture as just one particular point in a vast space of possibility.