LESSWRONG
LW

2294
johnswentworth
59286Ω691737135380
Message
Dialogue
Subscribe

Sequences

Posts

Sorted by Top (Inflation Adjusted)

Wikitag Contributions

Comments

Sorted by
Newest
11johnswentworth's Shortform
Ω
6y
Ω
744
No wikitag contributions to display.
From Atoms To Agents
"Why Not Just..."
Basic Foundations for Agent Models
Framing Practicum
Gears Which Turn The World
Abstraction 2020
Gears of Aging
Model Comparison
Resampling Conserves Redundancy (Approximately)
johnswentworth9h70

(Update 3)

We're now pursuing two main threads here.

One thread is to simplify the counterexamples into something more intuitively-understandable, mainly hopes of getting an intuitive sense for whatever phenomenon is going on with the counterexamples. Then we'd build new theory specifically around that phenomenon.

The other thread is to go back to first principles and think about entirely different operationalizations of the things we're trying to do here, e.g. not using diagram DKL's as our core tool for approximation. The main hope there is that maybe DKL isn't really the right error metric for latents, but then we need to figure out a principled story which fully determines some other error metric.

Either way, we're now >80% that this is a fundamental and fatal flaw for a pretty big chunk of our theory.

Reply
Resampling Conserves Redundancy (Approximately)
johnswentworth2d90

(Update 0)

I'm starting by checking that there's actually a counterexample here. We also found some numerical counterexamples which were qualitatively similar (i.e. approximately-all of the weight was on one outcome), but thought it was just numerical error. Kudos for busting out the sympy and actually checking it.

Looking at the math on that third-order issue... note that the whole expansion is multiplied by P[X]. So even if δ[X]∼√P[X], P[X] itself will still go to zero for small δ, so P[X](δ[X]√P[X])3 will go to zero. So it's not obviously a fatal flaw, though at the very least some more careful accounting would be needed at that step to make sure everything converges.

Reply
Resampling Conserves Redundancy (Approximately)
johnswentworth2d102

I plan to spend today digging into this, and will leave updates under this comment as I check things.

Reply
Coherence of Caches and Agents
johnswentworth3d40

Yup! A Simple Toy Coherence Theorem walks through a toy version of that idea, and I do think it's a ripe area for someone to figure out more realistic theorems.

Reply
My Empathy Is Rarely Kind
johnswentworth3d20

It's not that conscious/reflective. Respect is an emotion; my standards for it are more on the instinctive level. Which is not to say that there aren't consistent standards there, but they're not something I have easy direct control over or ready introspective access to.

Reply
ISO: Name of Problem
johnswentworth9d31

Haven't thought about a name for this problem for a while, but I still don't have one.

Reply1
johnswentworth's Shortform
johnswentworth10d20

Notably that post has a section arguing against roughly the sort of thing I'm arguing for:

Making the definition of what constitutes a low level language dependent on laws of physics is removing it from the realm of mathematics and philosophy. It is not a property of the language any more, but a property shared by the language and physical reality.

My response would be: yes, what-constitutes-a-low-level-language is obviously contingent on our physics and even on our engineering, not just on the language. I wouldn't even expect aliens in our own universe to have low-level programming languages very similar to our own. Our low level languages today are extremely dependent on specific engineering choices made in the mid 20th century which are now very locked in by practice, but do not seem particularly fundamental or overdetermined, and would not be at all natural in universes with different physics or cultures with different hardware architecture. Aliens would look at our low-level languages and recognize them as low-level for our hardware, but not at all low-level for their hardware.

Analogously: choice of a good computing machine depends on the physics of one's universe. 

I do like the guy's style of argumentation a lot, though.

Reply
johnswentworth's Shortform
johnswentworth10d40

I think that's roughly correct, but it is useful...

'The best UTM is the one that figures out the right answer the fastest' is true, but not very useful.

Another way to frame it would be: after one has figured out the laws of physics, a good-for-these-laws-of-physics Turning machine is useful for various other things, including thermodynamics. 'The best UTM is the one that figures out the right answer the fastest' isn't very useful for figuring out physics in the first place, but most of the value of understanding physics comes after it's figured out (as we can see from regular practice today).

Also, we can make partial updates along the way. If e.g. we learn that physics is probably local but haven't understood all of it yet, then we know that we probably want a local machine for our theory. If we e.g. learn that physics is causally acyclic, then we probably don't want a machine with access to atomic unbounded fixed-point solvers. Etc.

Reply
Natural Latents: Latent Variables Stable Across Ontologies
johnswentworth10d20

I think you might have misread something? The graphical statement of theorem 2 does not say that if ΛA is determined by ΛB, then ΛA is a mediator; that would indeed be false in general. It says that:

  • If ΛB is a mediator and we have agreement on observables, then...
  • ... naturality of ΛA implies that ΛA is determined by ΛB.

In particular, the theorem says that under some conditions ΛA is determined by ΛB. Determination is in the conclusion, not the premises. On the flip side, ΛA being a mediator is in the premises, not the conclusion.

Reply
johnswentworth's Shortform
johnswentworth10d50

What I have in mind re:boundedness...

If we need to use a Turing machine which is roughly equivalent to physics, then a natural next step is to drop the assumption that the machine in question is Turing complete. Just pick some class of machines which can efficiently simulate our physics, and which can be efficiently implemented in our physics. And then, one might hope, the sort of algorithmic thermodynamic theory the paper presents can carry over to that class of machines.

Probably there are some additional requirements for the machines, like some kind of composability, but I don't know exactly what they are.

This would also likely result in a direct mapping between limits on the machines (like e.g. limited time or memory) and corresponding limits on the physical systems to which the theory applies for those machines.

The resulting theory would probably read more like classical thermo, where we're doing thought experiments involving fairly arbitrary machines subject to just a few constraints, and surprisingly general theorems pop out.

Reply
Load More
591Making Vaccine
5y
249
564Orienting Toward Wizard Power
5mo
146
491Being the (Pareto) Best in the World
6y
61
485How To Write Quickly While Maintaining Epistemic Rigor
4y
39
420You Are Not Measuring What You Think You Are Measuring
3y
45
376What Money Cannot Buy
6y
53
368Accounting For College Costs
3y
41
362How To Get Into Independent Research On Alignment/Agency
Ω
4y
Ω
38
361Generalized Hangriness: A Standard Rationalist Stance Toward Emotions
3mo
69
359The Case Against AI Control Research
9mo
84
Load More