zhukeepa

Wiki Contributions

Comments

zhukeepa115

People sometimes say that AGI will be like a second species; sometimes like electricity. The truth, we suspect, lies somewhere in between. Unless we have concepts which let us think clearly about that region between the two, we may have a difficult time preparing.

I just want to strongly endorse this remark made toward the end of the post. In my experience, the standard fears and narratives around AI doom invoke "second species" intuitions that I think stand on much shakier ground than is commonly acknowledged. (Things can still get pretty bad without a "second species", of course, but I think it's worth thinking clearly about what those alternatives look like, as well as how to think about them in the first place.)

zhukeepaΩ330

Thanks, Alex. Any connections between this and CTMU? (I'm in part trying to evaluate CTMU by looking at whether it has useful implications for an area that I'm relatively familiar with.)

No direct connections that I'm aware of (besides non-classical logics being generally helpful for understanding the sorts of claims the CTMU makes). 

Good question! Yeah, there's nothing fundamentally quantum about this effect. But if the simulator wants to focus on universes with 1 & 2 fixed (e.g. if they're trying to calculate the distribution of superintelligences across Tegmark IV), the PNRG (along with the initial conditions of the universe) seem like good places for a simulator to tweak things. 

It is not clear to me that this would result in a lower Kolmogorov complexity at all. Such an algorithm could of course use a pseudo-random number generator for the vast majority quantum events which do not affect p(ASI) (like the creation of CMB photons), but this is orthogonal to someone nudging the relevant quantum events towards ASI. For these relevant events, I am not sure that the description "just do whatever favors ASI" is actually shorter than just the sequence of events.

Hmm, I notice I may have been a bit unclear in my original post. When I'd said "pseudorandom", I wasn't referring to the use of a pseudo-random number generator instead of a true RNG. I was referring to the "transcript" of relevant quantum events only appearing random, without being "truly random", because of the way in which they were generated (which I'm thinking of as being better described as "sampled from a space parameterizing the possible ways the world could be, conditional on humanity building superintelligence" rather than "close to truly random, or generated by a pseudo-random RNG, except with nudges toward ASI".) 

I mean, if we are simulated by a Turing Machine (which is equivalent to quantum events having a low Kolmogorov complexity), then a TM which just implements the true laws of physics (and cheats with a PNRG, not like the inhabitants would ever notice) is surely simpler than one which tries to optimize towards some distant outcome state.


As an analogy, think about the Kolmogorov complexity of a transcript of a very long game of chess. If both opponents are following a simple algorithm of "determine the allowed moves, then use a PRNG to pick one of them", that should have a bound complexity. If both are chess AIs which want to win the game (i.e. optimize towards a certain state) and use a deterministic PRNG (lest we are incompressible), the size of your Turing Machine -- which /is/ the Kolmogorov complexity -- just explodes.

Wouldn't this also serve as an argument against malign consequentialists in the Solomonoff prior, that may make it a priori more likely for us to end up in a world with particular outcomes optimized in their favor? 

It is not clear to me that this would result in a lower Kolmogorov complexity at all. 

[...]

Look at me rambling about universe-simulating TMs. Enough, enough.

To be clear, it's also not clear to me that this would result in a lower K-complexity either. My main point is that (1) the null hypothesis of quantum events being independent of consciousness rests on assumptions (like assumptions about what the Solomonoff prior is like) that I think are actually pretty speculative, and that (2) there are speculative ways the Solomonoff prior could be in which our consciousness can influence quantum outcomes. 

My goal here is not to make a positive case for consciousness affecting quantum outcomes, as much as it is to question the assumptions behind the case against the world working that way. 

This. Physics runs on falsifiable predictions. If 'consciousness can affect quantum outcomes' is any more true than the classic 'there is an invisible dragon in my garage', then discovering that fact would seem easy from an experimentalist standpoint. Sources of quantum randomness (e.g. weak source+detector) are readily available, so any claimant who thinks they can predict or affect their outcomes could probably be tested initially for a few 100$. 

Yes, I'm also bearish on consciousness affecting quantum outcomes in ways that are as overt and measurable in the way you're gesturing at. The only thing I was arguing in this post is that the effect size of consciousness on quantum outcomes is maybe more than zero, as opposed to obviously exactly zero. I don't think of myself as having made any arguments that the effect size should be non-negligible, although I also don't think that possibility has been ruled out for non-neglible effect sizes lying somewhere between "completely indistinguishable from no influence at all" and "overt and measurable to the extent a proclaimed psychic could reproducibly affect quantum RNG outcomes". 

I'll take a stab at this. Suppose we had strong a priori reasons for thinking it's in our logical past that we'll have created a superintelligence of some sort. Let's suppose that some particular quantum outcome in the future can get chaotically amplified, so that in one Everett branch humanity never builds any superintelligence because of some sort of global catastrophe (say with 99% probability, according to the Born rule), and in some other Everett branch humanity builds some kind of superintelligence (say with 1% probability, according to the Born rule). Then we should expect to end up in the Everett branch in which humanity builds some kind of superintelligence with ~100% probability, despite the Born rule saying we only have a 1% chance of ending up there, because the "99%-likely" Everett branch was ruled out by our a priori reasoning. 

I'm not sure if this is the kind of concrete outcome that you're asking for. I imagine that, for the most part, the kind of universe I'm describing will still yield frequencies that converge on the Born probabilities, and for the most part appear indistinguishable from a universe in which quantum outcomes are "truly random". See my reply to Joel Burget for some more detail about how I think about this hypothesis. 

If we performed a trillion 50/50 quantum coin flips, and found a program with K-complexity far less than a trillion that could explain these outcomes, that would be an example of evidence in favor of this hypothesis. (I don't think it's very likely that we'll be able to find a positive result if we run that particular experiment; I'm naming it more to illustrate the kind of thing that would serve as evidence.) (EDIT: This would only serve as evidence against quantum outcomes being truly random. In order for it to serve as evidence in favor of quantum outcomes being impacted by consciousness, the low K-complexity program explaining these outcomes would need to route through the decisions of conscious beings somehow; it wouldn't work if the program were just printing out digits of pi in binary, for example.)

My inside view doesn't currently lead me to put much credence on this picture of reality actually being true. My inside view is more like "huh, I notice I have become way more uncertain about the a priori arguments about what kind of universe we live in -- especially the arguments that we live in a universe in which quantum outcomes are supposed to be 'truly random' -- so I will expand my hypothesis space for what kinds of universes we might be living in". 

Shortly after publishing this, I discovered something written by John Wheeler (whom Chris Langan cites) that feels thematically relevant. From Law Without Law

I was hoping people other than Jessica would share some specific curated insights they got. Syndiffeonesis is in fact a good insight.

I finally wrote one up! It ballooned into a whole LessWrong post. 

Load More