Zero-math-just-words summary:
So the natural world seems to be full of this symmetry thing, which is great because I suspect otherwise we wouldn't have any hope of making a good map of it. Here's some math that makes constructive epistemic claims about a particularly simple example -- permutation symmetry -- that people have been generally confused about for 90 years. Then here's how that idea would testably interact with a thought experiment that you could almost pull off in a lab (BYO magic box).
Remember T is a program that's supposed to emit all the computable parts of D that are relevant for doing inference on A. That's P(A|D) = P(A|T(D)). (Naturally, the identity function on D does a perfectly fine job of preserving the relevant parts of D, so to be non-trivial you really want T to compress D in some way.) Here is some permutation of D (this doesn't compress D). T(D) = T((D)) just says that T doesn't care about the ordering of D. Then you combine these facts to resolve that inference on A also doesn't care about the ordering of D. That's P(A|D) = P(A|(D)).
Negative example:
= "this sensory input contains a cat"
D = "the pixels of the sensory input"
T = "AlexNet without the final softmax layer"
Observe destroys inference for . P(|D) P(|(D)). Classification of natural images isn't invariant to arbitrary reordering of pixels. AlexNet isn't invariant to reordering of pixels because gradient descent would rapidly exit that region of weight space. T(D) T((D)).
Perhaps there are some other (maybe approximate) symmetries () like reflections, translations, blurring, recoloring, rescaling. P(|D) P(|(D)). AlexNet might be in or near those regions of weight space. T(D) T((D)). This should also apply to your cortex or aliens or whatever has a notion of a cat!
I would be very surprised if this didn't connect to natural latents. It connects to universality. I don't think it connects directly to infrabayes but they might be compatible. It also connects to SLT, but for reasons that are beyond the scope of this post.
In the Ising model, the inability -- through local fluctuations, well below -- for the coarse-grained state to transition to/from is an example of superselection. Unfortunately, I can't find a good primer on the subject, but here's something: https://physics.stackexchange.com/a/56570
To be able to take your desired step, reasoning from local->global, you also have to make additional measurements to rule out non-uniform temperature and non-uniform magnetic fields.
Your notation is clear to me! It can be shown that:
Df(p⋆,q)=∫dxq(x)f⋄(f⋄#(⟨λ,T(x)⟩))
Even though f⋄ and f⋄# are both convex, their composition is not necessarily convex. Sorry, I don't have any clear counterexamples at the ready. (The Hessian determinants all vanish in the graphene/bit-string model.) It's likely that I've configured my numerical solvers badly for this example.
As far as binding to reality: