(Inspired by https://stats.stackexchange.com/questions/175153/approximating-pa-b-c-using-pa-b-pa-c-pb-c-and-pa-pb-pc)

Imagine you're working at an electronics manufacturing company. There's a type of component your company needs for an upcoming production process. Your company can either produce these components in-house or it can source them from a vendor. Your job is to determine whether or not to produce the component in-house. Your team has been budgeted $2,000,000 for procuring as many working components as possible. You won't know whether a part works until the downstream production process is up and running later in the year, but you need to make a decision on part procurement now. Let's say your company agrees to pay you a bonus next year proportional to how many parts turned out to be functional and your personal utility function is linear in this bonus and doesn't depend on anything else.

Using the in-house machine, your $2,000,000 budget will get you 1,000,000 parts which might or might not work. A part will work if and only if it has dimples on the bismuth layer (event ), it has an osmium layer between 40 and 60 microns thick (event ), and the misalignment between the polonium and tungsten layers is less than 20 microns (event ). Thus the expected number of working parts is . The only way to determine whether a part satisfies properties , and  is by doing destructive testing: you have to physically cut it up. Because of the way this cutting works, you can't measure  if you've already measured  and  on a part, you can't measure  if you've already measured  and  on a part, and you can't measure  if you've already measured  and  on a part. The only way to get data on , and  together would be to send a part downstream into production but the production process doesn't exist yet.

Through destructive testing, you've been able to determine that , and  (you've tested such a truly unfathomable number of parts that you've gotten the uncertainty on these numbers down arbitrarily small).[1] Otherwise, the machine you're using to create these parts is a black box. The company that sold you the machine has since gone defunct and you can't find anyone else with prior experience working with this machine.

On the other hand, for $2,000,000 the external vendor is willing to give you  parts for some specified . These parts have been used in many other production processes and are known to never, ever fail. 

If , you should probably go with the in-house approach. If , you should probably source externally. What is the value of  below which you stay in-house but above which you go with the vendor?[2]


I'm working on a project where I'm going to need to estimate probabilities of combinations of events like this with incomplete information. I asked some friends about my problem and they said it was ill-posed in the abstract, so I wanted to create a real-world thought experiment where you can't wave the problem away with "There is no proper answer." If you were in this position, you'd have to actually make a choice.

This problem reminds me of the discussion around the presumption of independence. I think that a good philosophical justification for assuming independence in the absence of evidence to the contrary should be able to be generalized to providing a method for finding the "most independent" assignment of probabilities in settings where we know that actual independence is out of the question.

  1. ^

    No pair of these events is independent. To see that there exist two (and by taking convex linear combinations, infinitely many) possible values of  from this provided information, note that there are two valid probability mass functions,  and , on truth assignments to  which differ on their value of :

    0.150.14
    0.10.11
    0.20.21
    0.10.09
    0.050.06
    0.10.09
    0.250.24
    0.050.06
  2. ^

    Technically there could be an interval of width  where you're indifferent.

New Comment
4 comments, sorted by Click to highlight new comments since: Today at 1:16 AM

You want the widget problem in chapter 14 of Jaynes' Probability Theory: The Logic Of Science; it is extremely similar to the problem you present. Long story short: express each of the known probabilities as expectations of indicator variables (e.g. ), then maximize entropy subject to the constraints given by those expectations. Jaynes covers a bunch of conceptual arguments for why that's a sensible procedure to follow.

To do better than that, the next step would be to look for any structure/pattern in the known probabilities and exploit those - e.g. if the known probabilities approximately factor over a certain Bayes net, then a natural guess is that the unknown probabilities will too, which may allow backing out of the unknown probabilities.

Wow, this is exactly what I was looking for! Thank you so much!

These are the probabilities of each state:

StateProbability

with  being the probability of all three parts of a component being fine. (Obviously, , because .)

This is not enough information to solve for x, of course, but note that . Note also that  and  (ie A is not strongly correlated or anti-correlated with  or ). However,  by quite a long way:  is fairly strongly anti-correlated with .

Now here's the estimation bit, I suppose: given that  holds, we'd probably expect a similar distribution of probabilities across values of  and , given that  is not (strongly) correlated with  or . So  etc. This resolves to 

StateProbability

This seems... not super unreasonable? At least, it appears slightly better than going for the most basic method, which is , so split the difference and say it's  or thereabouts.

The key assumption here is that "if  is pretty much uncorrelated with  and , it's probably uncorrelated with the conjunction . This is not strictly-always true as a matter of probability theory, but we're making assumptions on incomplete information based on a real-world scenario, so I'd say this skewing our guess by a factor of 10% from the most naive approach is probably helpful-on-net.

This means in expectation, we guess the in-house machine to produce  good widgets. I'd take that many from the Super Reliable Vendor if offered, but if they were offering less than that I'd roll the dice with the Worryingly Inconsistent In-House Machine. That is, I'm indifferent at .

If your main intent is just to make the right decision for the company, then my answer is buy those parts unless your main goal is doing math. You don't need statistics to elaborate that, common sense is enough.

Since the manufacturers already produces more on top of those components, you already have enough to risk and test after manufacturing. Once your product works, go for optimisations and consider producing parts in-house.