Claim: Given constraints on the entropy of a latent variable , the redundancy and mediation errors are optimized exactly when the sum of mutual information is maximized.
Mediation is one of the main conditions for naturality of latent variables. Intuitively, an (approximate) mediator captures (approximately) all the correlation between a collection of observables . We will prove that a latent variable is an (optimal) approximate mediator it maximizes the sum of mutual information given constraint on entropy
Intuition: To maximize the sum of mutual information , it is beneficial to include information that is shared among multiple in (as that would increase multiple mutual information term). A mediator is a latent variable that captures all the shared information among the s, so it will be selected for when maximizing .
Proof:
The correspondence with the redundancy condition is quite simple: The sum of the redundancy errors are , so if we have a constraint of the form , then we have , and the sum of redundancy errors are minimized exactly when is maximized and .
We've shown that given constraints on , both the mediation and the redundancy conditions are minimized exactly when the sum of mutual information is maximized, we can use this to simplify the search for natural latents, and while optimizing for this quantity there is no tradeoff between the redundancy and mediation errors.
However, note that mediation error increases as decreases (the mediation error for the empty latent is simply total correlation), while the redundancy error increases with (which is why we imposed for mediation but for redundancy). So the entropy of the latent is exactly the parameter that represents the tradeoff between the mediation and redundancy errors.
In summary, we can picture a pareto-frontier of latent variables with maximal and different entropies, by ramping up the entropy of along the pareto-frontier we gradually increase the redundancy errors while reducing the mediation errors, and these are the only parameters relevant for latent variables that are pareto-optimal w.r.t the naturality conditions.
(See this post for background about the stochastic deterministic natural latent conjecture)
We've shown that given fixed , both the redundancy and mediation errors of a latent are minimized when is maximized, while is exactly the parameter the determines the tradeoff between redundancy and mediation errors (among pareto-optimal latent). We'll discuss how this could open up new angles of attack for the stochastic deterministic natural latent conjecture.
Suppose that we have a stochastic natural latent that satisfies:
From our result, we know that to construct a deterministic natural latent , all we have to do is to determine the entropy and then select the latent that maximizes . The latter ensures that the latent is pareto-optimal w.r.t the mediation and determinism conditions, while the former selects a particular point on the pareto-frontier.
Now suppose that our stochastic natural latent has a particular amount of mutual information with the joint observables . If the stochastic natural latent was a deterministic function of the observables, then we would have:
(as that would imply )
So one heuristic for constructing a deterministic natural latent is to just set and maximize given the entropy constraint (so that hopefully captures all the mutual info between and ). We will show that if preserves the mutual information with each observable (i.e. ), then the mediation condition is conserved and the stochastic redundancy conditions implies the deterministic redundancy conditions
Note that the mediation error is
Since all and terms are fixed relative to , the mediation error is completely unchanged if we replace with a deterministic latent that satisfies and for each .
Note that using partial information decomposition[1], we can decompose the stochastic redundancy errors as the following:
where represents synergistic information of and w.r.t while represents unique information of w.r.t . Intuitively, represents the information that has about when we have access to , which should include unique information that we can only derive from but not , but also synergistic information that we can only derive when we have both and .
We also have:
Intuitively, this is because contains both the unique information about that you can only derive from but not , and also the redundant information that you can derive from either or . Note that since and , we have
Similarly, we have:
where
As a result, both and are bounded by . This means that if we can find a deterministic latent that conserves all the relevant mutual information terms , and , then we can bound the deterministic redundancy errors:
We've shown that a sufficient condition for mediation and redundancy to transfer from the stochastic to deterministic case is if the deterministic latent preserves the mutual information of the stochastic latent with both the joint observable as well as the individual observables and . Given this, the remaining task would be to prove that such a deterministic latent always exists, or that it can preserve the mutual information terms up to some small error. In particular, if existence is guaranteed, then a tractable way to find the deterministic latent given a stochastic latent is to just set and maximize
Note that PID depends on a choice of redundancy measure, but our proof holds for any choice that guarantees non-negativity of PID atoms
Previously, we've shown that given constraints on the entropy of a natural latent variable , the mediation and redundancy errors are minimized exactly when the sum of mutual information with observables is maximized. In addition, the entropy of the latent variable is exactly the parameter that represents the tradeoff between the mediation and redundancy condition. In particular, the mediation error can only reduce with while the redundancy errors can only increase with .
However, there may be regimes where changes in can reduce the mediation error without increasing the redundancy errors or vice versa. For instance:
If we define a maximum redund as a latent variable that satisfies the redundancy conditions and has the maximum entropy among redunds, then represents the regime where we can increase without increasing the redundancy errors, since increasing beyond would necessarily violate the redundancy condition given our assumption of maximum entropy
Similarly, define a minimum mediator as a mediator with minimal entropy (among mediators). Then represents the regime where we can reduce entropy without increasing the mediation error, since reducing below necessarily violates the mediation condition.
Combining these ideas, represents the regime where changing actually presents a tradeoff between the mediation and redundancy errors; the minimum mediator and maximum redund marks the boundaries for when weak pareto-improvements are possible.
In natural latents we care about the uniqueness of latent variables, which is why we have concepts like minimal mediators and maximal redunds:
Through a universal-property-flavored proof, we can show approximate isomorphism among any pair of minimal mediators : Since is a minimal mediator and is a mediator, approximately determines , and using a similar argument we conclude determines . The same reasoning also allows us to derive uniqueness of any pair of maximal redunds. Naturality occurs when the maximal redund converge with the minimal mediator.
However, note that the concepts of minimal mediators and maximal redunds are at least conceptually distinct from minimum mediators and maximum redunds. We shall therefore prove that these concepts are mathematically equivalent. This can be useful because it's much easier to find minimum mediators and maximum redunds computationally, but we ultimately care about the unqiueness property offered by minimal mediators and maximal redunds, proving an equivalence enables the former to have the uniqueness guarantees of the latter.
Proof:
In addition, we have so is also an approximate minimum mediator
Proof:
Similarly, suppose that is a maximal redund, then , which means and is also an approximate maximum redund.
Recall that (where is the minimum mediator and is the maximum redund) represents the regime where changing actually presents a tradeoff between the mediation and redundancy errors. Due to the equivalence we proved, we can also think of as the minimal mediator and as the maximal redund.
We also know that naturality occurs when the minimal mediator converges with the maximal redund (as a natural latent satisfies both mediation and redundancy, and mediator determines redund); we can picture this convergence as if we're shrinking the gap between and . In other words, naturality occurs exactly when the regime of tradeoff () between the redundancy and mediation error is small. If we have exact naturality , then pareto-improvements on the naturality conditions can always be made by nudging closer to .
Combining this with our previous result, we conclude that that maximizing represents strong pareto-improvements over the naturality conditions; and represents the regime where we can have weak pareto-improvements by nudging closer to the boundary of or ; whereas represents the regime of real tradeoffs between naturality conditions. An approximate natural latent exist exactly when the regime of real tradeoffs is small and we can pareto-improve towards naturality