Using Finite Factored Sets for Causal Representation Learning?
One of the primary motivations of Finite Factored Sets (shorthand: FFS) that initially caught my eye was, to quote Scott: > "Given a collection of variables and a joint probability distribution over those variables, Pearl can infer causal/temporal relationships between the variables.", the words "Given a collection of variables" are...
Based on playing around recently with a similar setup (but only toy examples), I'm actually surprised you get only 85%, as I've only observed NDE=0 when I freeze the entire reasoning_trace.
My just-so explanation for this was that whenever the reasoning trace includes the conclusion (that is, the bolded text in your example), then freezing the reasoning trace preserves the final conclusion. Put another way, the <recommendation> is ~deterministically determined by <reasoning>, which suggests a strong bias towards seeing low direct effects.
If this just-so story is true, it suggests that we might need a more granular mediator than the <entire reasoning_trace>, if possible