LESSWRONG
LW

2363
Daniel C
168Ω86500
Message
Dialogue
Subscribe

Master's student in applied mathematics, funded by Center on Long-Term Risk to investigate the cheating problem in safe pareto-improvements. Former dovetail fellow with @Alex_Altair. 

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
No wikitag contributions to display.
johnswentworth's Shortform
Daniel C8d30

The current theory is based on classical hamiltonian mechanics, but I think the theorems apply whenever you have a markovian coarse-graining. Fermion doubling is a problem for spacetime discretization in the quantum case, so the coarse-graining might need to be different. (E.g. coarse-grain the entire hilbert space, which might have locality issues but probably not load-bearing for algorithmic thermodynamics)

On outside view, quantum reduces to classical (which admits markovian coarse-graining) in the correspondence limit, so there must be some coarse-graining that works

Reply2
johnswentworth's Shortform
Daniel C8d61

I also talked to Aram recently & he's optimistic that there's an algorithmic version of the generalized heat engine where the hot vs cold pool correspond to high vs low k-complexity strings. I'm quite interested in doing follow-up work on that

Reply
johnswentworth's Shortform
Daniel C8d76

The continuous state-space is coarse-grained into discrete cells where the dynamics are approximately markovian (the theory is currently classical) & the "laws of physics" probably refers to the stochastic matrix that specifies the transition probabilities of the discrete cells (otherwise we could probably deal with infinite precision through limit computability)

Reply2
Synthesizing Standalone World-Models, Part 1: Abstraction Hierarchies
Daniel C14d30

As in, take a set of variables X, then search for some set of its (non-overlapping?) subsets such that there's a nontrivial natural latent over it? Right, it's what we're doing here as well.

I think the subsets can actually be partially overlapping, for instance you may have a λ that’s approximately deterministic w.r.t {X1,X2} and {X2,X3} but not X2 alone, weak redundancy (approximately deterministic w.r.t {¯¯¯¯¯¯Xi}∀i) is also an example of redunds across overlapping subsets

Reply1
Research Agenda: Synthesizing Standalone World-Models
Daniel C19d30

Mm, this one's shaky. Cross-hypothesis abstractions don't seem to be a good idea, see here.


yea so I think the final theory of abstraction will have a weaker notion of equivalence espeically when we incorporate ontology shifts. E.g. we want to say that water is the same concept before and after we discover water is H2O, but the discovery obviously breaks predictive agreement (Indeed, the solomonoff version of natural latent is more robust to the agreement condition)

Also, you can totally add new information/abstraction that is not shared between your current and new hypothesis, & that seems consistent with the picture you described here (you can have separate ontologies but you try to capture the overlap as much as possible)

My guess is that there's something like a hierarchy of hypotheses, with specific high-level hypotheses corresponding to several lower-level more-detailed hypotheses, and what you're pointing at by "redundant information across a wide variety of hypotheses" is just an abstraction in a (single) high-level hypothesis which is then copied over into lower-level hypotheses. (E. g., the high-level hypothesis is the concept of a tree, the lower-level hypotheses are about how many trees are in this forest.)

yes I think that's the right picture

But we don't derive it by generating a bunch of low-level hypotheses and then abstracting over them, that'd lead to broken ontologies.

I agree that we don't do that practically as it'd be slower (instead we simply generate an abstraction & use future feedback to determine whether it's a robust one), but I think if you did generate a bunch of low-level hypotheses and look for redundant computation among them, then an adequate version of it would just recover the "high-level low-level hypotheses" picture you've described?

In particular, with cross-hypothesis abstraction we don't have to separately define what the variables are, so we can sidestep dataset-assembly entirely & perhaps simplify the shifting structures problem

Reply1
Synthesizing Standalone World-Models, Part 2: Shifting Structures
Daniel C21d40

Nice, I've gestured at similar things in this comment, conceptually the main thing you want to model is variables that control the relationships between other variables, the upshot is you can continue the recursion indefinitely: Once you have second order variables that control the relationships between other variables, you can then have variables that control the relationship among second order variables and so on.

Using function calls as an analogy: When you're executing a function that itself makes a lot of function calls, there are two main ways these function calls can be useful:

  1. The results of these function calls might be used to compute the final output
  2. The results of these function calls can tell you what other function calls would be useful to make (e.g. if you want to find the shape of a glider, the position tells you which cells to look at to determine that)

an adequate version of this should also be turing complete which means it can accomodate shifting structures, & function calls seem like a good way to represent hierarchies of abstractions

 

 

CSI in bayesian networks also deals with the idea that the causal structure between variables changes over time/depending on context (you're probably more interested in how relationships between levels of abstraction changes with context, but the two directions seem linked). I plan to explore the following variant at some point(not sure if it's already in the literature):

  • Suppose that there is a variable Y that "controls" the causal structure of X, we use the good-old KL approximation to represent the error conditional on a particular value of Y DKL(P(X|Y=y)∥ΠiP(Xi|XpaG(i),Y=y)) under a particular diagram G
  • You can imagine that the conditional distrbution initially approximately satisfies a diagram G1, but as you change the value of Y, the error for G1 goes up while the error for some other diagram G2 goes to 0
  • In particular, if Y is a continuous variable, and the conditional distribution P(X|Y=y) changes continuously with Y, then DKL(P(X|Y=y)∥ΠiP(Xi|XpaG(i),Y=y)) changes continuously with Y which is quite nice
  • So this is a formalism that deals with "context-dependent structure" in a way that plays well with continuity, and if you have discrete variables controlling the causal structure, you can use it to accommodate uncertainty over the discrete outcomes (that determine causal structure). 
Reply
Synthesizing Standalone World-Models, Part 1: Abstraction Hierarchies
Daniel C22d*50

But note that synergistic information can be defined by referring purely to the system we're examining, with no "external" target variable. If we have a set of variables X={x1,…,xn}, we can define the variable s such that I(X;s) is maximized under the constraint of ∀Xi∈(P(X)∖X):I(Xi;s)=0. (Where P(X)∖X is the set of all subsets of X except X itself.)


That's a nice formulation of synergistic information, it's independent with redundant info via the data-processing inequality 0=I(Xi;s)≥I(f(Xi);s) so somewhat promising that it can add up to total entropy. 

 

You might be interested in this comment if distinguishing betweeen synergistic and redundant information is not your main objective: You can simply define redunds over collections of subsets, such that e.g. "dogness" is a redund over every subset of atoms that allows you to conclude you're looking at a dog. In particular, the redundancy lattice approach seems simpler when the latent depends on not just synergistic but also redundant and unique information

 

One issue with PID worth mentioning is that they haven't figured out what measure to use for quantifying multivariate redundant information. It's the same problem we seem to have. But it's probably not a major issue in the setting we're working in (the well-abstracting universes).

Recent impossibility result seems to rule out general multivariate PID that guarantees non-negativity of all components, though partial entropy decomposition may be more tractable
 

  1. If there's a pair of qi, qk such that Xi⊂Xk, then qi necessarily contains all information in qk. Re-define qi, removing all information present in qk.

This seems similar to capturing unique information, where the constructive approach is probably harder in PID than PED. E.g. in BROJA it involves an optimization problem over distributions with some constraints on marginals, but it only estimates the magnitude of unique info, not an actual random variable that represents unique info

Reply
Research Agenda: Synthesizing Standalone World-Models
Daniel C23d*30

Nice post!

Some frames about abstractions & ontology shifts I had while thinking through similar problems (which you may have considered already):

  • The dual of "abstraction as redundant information across a wide variety of agents in the same environment" is "abstraction as redundant information/computation across a wide variety of hypotheses about the environment in an agent's world model" (E.g. a strawberry is a useful concept to model for many worlds that I might be in). I think this is a useful frame when thinking about "carving up" the world model into concepts, since a concept needs to remain invariant while the hypothesis keeps being updated
  • The semantics of a component in a world model is partly defined by its relationship with the rest of the components (e.g. move a neuron to a different location and its activation will have a different meaning), so if you want a component to have stable semantics over time, you want to put the "relational/indexical information" inside the component itself
  • In particular, this means that when an agent acquires new concepts, the existing concepts should be able to "specify" how it should relate to that new concept (e.g. learning about chemistry then using it to deduce macro-properties of strawberries from molecular composition)

happy to discuss more via PM as some of my ideas seem exfohazardous

Reply
Jemist's Shortform
Daniel C1mo10

Neat idea, I've thought about similar directions in the context of traders betting on traders in decision markets

A complication might be that a regular deductive process doesn't discount the "reward" of a proposition based on its complexity whereas your model does, so it might have a different notion of logical induction criterion. For instance, you could have an inductor that's exploitable but only for propositions with larger and larger complexities over time, such that with the complexity discounting the cash loss is still finite (but the regular LI loss would be infinite so it wouldn't satisfy regular LI criterion)

(Note that betting on "earlier propositions" already seems beneficial in regular LI since if you can receive payouts earlier you can use it to place larger bets earlier)

There's also some redundancy where each proposition can be encoded by many different turing machines, whereas a deductive process can guarantee uniqueness in its ordering & be more efficient that way

Are prices still determined using Brouwer’s fixed point theorem? Or do you have a more auction-based mechanism in mind?

Reply1
Alignment as uploading with more steps
Daniel C1mo32

Yes I agree

I think it's similar to CIRL except less reliant on the reward function & more reliant on the things we get to do once we solve ontology identification

Reply
Load More
16Sleeping Experts in the (reflective) Solomonoff Prior
Ω
2mo
Ω
0
29Towards building blocks of ontologies
8mo
0
11Can subjunctive dependence emerge from a simplicity prior?
Q
1y
Q
0
19Jonothan Gorard:The territory is isomorphic to an equivalence class of its maps
1y
18
23What program structures enable efficient induction?
1y
5
22My decomposition of the alignment problem
1y
22