LESSWRONG
LW

FutarchyPrediction MarketsWorld Modeling
Curated

173

Futarchy's fundamental flaw

by dynomight
13th Jun 2025
Linkpost from dynomight.net
10 min read
48

173

FutarchyPrediction MarketsWorld Modeling
Curated

173

Futarchy's fundamental flaw
31niplav
8dynomight
24RobinHanson
5dynomight
2RobinHanson
4dynomight
3RobinHanson
3Christopher King
2dynomight
13cousin_it
5dynomight
3cousin_it
3dynomight
2cousin_it
1dynomight
7So8res
5dynomight
6So8res
3dynomight
7So8res
6Dweomite
2[comment deleted]
2mako yass
11Ruby
7Baram Sosis
7dynomight
5Daniel C
5dynomight
3Daniel C
2dynomight
1Daniel C
3dynomight
3Daniel C
2dynomight
1Daniel C
2dynomight
1Daniel C
2dynomight
1Daniel C
3Matrice Jacobine
2Mikhail Samin
2Clay S
1_nuchi
1Bunthut
1Kyle Pena
1rain8dome9
1Austin Welch
1Vin
-1deepthoughtlife
2[comment deleted]
1[comment deleted]
New Comment
48 comments, sorted by
top scoring
Click to highlight new comments since: Today at 7:29 PM
[-]niplav1mo310

Previous discussion of the same (similar?) point:

  • Futarchy implements evidential decision theory (Caspar Oesterheld, 2017)
  • Conditional prediction markets are evidential, not causal (philh, 2024)

Futarchy also has some other issues, see e.g.

  • Futarchy Fix (Abram Demski, 2017)
  • Issues with Futarchy (Lizka, 2021)
Reply3
[-]dynomight1mo83

The earliest I'm aware of is the 2015 post I linked at the end. (Though as far as I know, my impossibility theorem for alternate payoff strategies is novel.)

Reply
[-]RobinHanson1mo242

I respond here: https://www.overcomingbias.com/p/decision-conditional-prices-reflect

Reply
[-]dynomight1mo50

Do I correctly understand that you claim that under some plausible assumptions, the market will converge to P(A|do(B))? Can you state what those assumptions are? The challenge for me is that you go through a set of at least four possible sets of assumptions and give informal arguments for each. But I can't tell which of these sets of assumptions you believe is realistic, and under which of these you claim market prices will converge to p(A|do(B)). (Feel free to make simplifying assumptions like an infinite number of traders, no market fees, etc.)

Further, when you state that my result is wrong, that would seem to imply that no additional assumptions are needed. Yet all your arguments seem to rely on additional assumptions, which makes me question if my result really is wrong as stated, or rather that you prefer to add some additional assumptions.
 

Reply
[-]RobinHanson1mo20

Market estimates will converge to the most profitable P(X if A), the one that wins bets vs other versions. And that is the version you want to use when you make decisions.

Reply
[-]dynomight1mo40

Sorry to be persistent, but can you confirm that this means you do not claim markets in general converge to p(A|do(B))? That's my central claim, so when you state that I'm wrong, the obvious interpretation would be that that you believe that central claim is wrong. In your post, you don't identify what mistake supposedly exists, so I'd like to confirm if you're actually claiming to refute my central claim or if, rather, you're arguing that it doesn't matter.

Reply
[-]RobinHanson1mo30

You say that markets give evidential conditionals while decisions want causal conditionals. For this comment, I'm not taking a position on which conditional we want for decisions. I'm just saying that both trades and the decision advised should use the same conditional, but I'm not saying which one that is. 

Reply
[-]Christopher King1mo30

I'm a bit confused. What happens in a concrete example where CDT and EDT normally give a different answer?

For example, would a futarchy one-box (evidential decision theory) or two-box (casual decision theory)?

Reply
[-]dynomight1mo20

Thanks, I added a link to the main post on my blog as well.

Reply
[-]cousin_it1mo132

I'm a bit confused. Let's say we're deciding between actions X and Y. We set up a pair of conditional prediction markets to determine which would give us more utility. And we promise to follow the market's decision - nothing can influence our choice between X and Y, except the market's decision. In this situation, is there still a difference between conditional and causal probabilities?

Reply1
[-]dynomight1mo5-4

I think so. That's basically what I'm trying to argue in this section: https://www.lesswrong.com/posts/vqzarZEczxiFdLE39/futarchy-s-fundamental-flaw#Putting_markets_in_charge_doesn_t_work

(An alternative argument would be that by default P(A|B) != P(A|do(B)), so maybe the burden of proof should fall on whoever is claiming something special happens that makes them be the same in this case. It's possible that there are some assumptions under which this happens, but to the best of my knowledge, no one has specified what these assumptions would be, or made the argument that they're sufficient.)

Reply
[-]cousin_it1mo30

Can you explain that section a bit? By your argument, it makes sense to pay at least 50 cents for the contract. But if that argument was correct, then everyone else in the market would also apply it, and the price of the contract would be at least 50 cents with certainty. That seems weird.

Reply
[-]dynomight1mo3-1

Sure, but could you help me understand exactly where the argument is failing for you? (E.g. with which of the three thought experiments?)

I should emphasize, I'm only claiming that the price will be at least 50 cents in the scenario where the coin is laser-scanned, so you know the market only activates when the coin has at least a 50% probability of landing heads. In the scenario where market activation is based on market prices, I only claim that you will pay more than 40 cents (probably less than 10 cents more), because the the fact that the market price closed above 50 cents still provides some extra info on top of your initial guess about the coin.

Reply
[-]cousin_it1mo20

Yeah, I'm trying to think about the market scenario. Still can't tell if it works. If everyone in the market used the argument you describe, would the extra info actually be nonzero for everyone?

Reply
[-]dynomight1mo1-1

It's quite hard to argue against this in the abstract. My view would be that the burden of proof falls on the person who claims it works out. I've shown that in general people will bid at prices that don't reflect their true beliefs. It's possible there exists some set of assumptions where the market price converge to the true "market beliefs". But I'd like to see someone actually make that argument! It's very hard for me to prove that no such assumption+proof exist, because non-existence proofs are typically very hard. (See P=NP etc.)

That said, I do think the extra info would indeed almost always still be nonzero. People in general have different beliefs about the value of the coin. It's incredibly rare to start with a bunch of interrelated random variables and end up with something independent. If you were about to bet, and then I told you what the final market closing price would be, wouldn't you update your beliefs at least a little?

Reply
[-]So8res1mo72

the trick is that the argument stops working for conditions that start to look like they might trigger. so the argument doesn't disrupt the idea that conditional prediction markets put the highest price on the best choice, but it does disrupt the idea that the pricings for unlikely conditions are counterfactually accurate.

for intuition, suppose there's a conditional prediction market for medical treatments for cancer. one of the treatments is "cut off the left leg." if certain scans and tests come back just the right way (1% chance likely) then cutting off the left leg is the best possible treatment, but otherwise, it's a dumb idea. if that condition is trading as if it's very unlikely to be a good idea, you can buy it up at very low risk -- most likely, the contract is canceled and you get your money back. but on the off-chance that the scans and tests come back in just the right way, you make a killing.

however, this incentive only exists insofar as "cut off the left leg" is not at risk of winning (before the tests and scans come back). if you thought the leg was going to be cut off simply because everyone else was buying the price up, and that it wouldn't actually heal the cancer, you'd sell off your shares.

this argument implies that you can't trust conditional prediction markets with cardinal ranking. however, afaict you'd need other arguments to imply that you can't trust their choice of "best action." such arguments probably exist, but this one isn't sufficient for it. (off the top of my head, i'd consider the case where option A is better for society and option B is better for some malicious actor, and the malicious actor is rich enough to convince the market to take option B. my intuition without thinking further is that this should 'obviously' work if the malicious actor is rich enough, in a fashion that's disanalogous to prediction markets in that it's not solved automatically with time (because the counterfactuals never settle and so the more-correct traders can't drain the rich-manipulator's money), but i haven't actually thought about it.)

Reply
[-]dynomight1mo53

the trick is that the argument stops working for conditions that start to look like they might trigger.

 

Can you give an argument for this claim? You're stating that there's an error in my argument, but you don't really engage with the argument or explain where exactly you think the error is.

For example, can you tell me what's incorrect in my example of two coins where you think one has a 60% probability and the other 59%, yet you'd want to pay more for a contract on the 59% coin? https://www.lesswrong.com/posts/vqzarZEczxiFdLE39/futarchy-s-fundamental-flaw#No__order_is_not_preserved (If you believe something is incorrect there.)

Reply
[-]So8res1mo*61

short version: the analogy between a conditional prediction market and the laser-scanner-simulation setup only holds for bids that don't push the contract into execution. (similarly: i agree that, in conditional prediction markets, you sometimes wish to pay more for a contract that is less valuable in counterfactual expectation; but again, this happens only insofar as your bids do not cause the relevant condition to become true.)

longer version:

suppose there's a coin that you're pretty sure is biased such that it comes up heads 40% of the time, and a contract that pays out $1 if the market decides to toss the coin and it comes up heads, and suppose any money you pay for the contract gets refunded if the market decides not to toss the coin. suppose the market will toss the coin if the contracts are selling for more than 50¢.

your argument (as i understand it) correctly points out that it's worth buying the contract from 40¢ to 45¢, because conditional on the market deciding to toss the coin, probably the market figured out that the coin actually isn't biased away from heads (e.g. via their laser-scanner and simulator). and so either your 45¢ gets refunded or the contract is worth more than 45¢, either way you don't lose (aside from the opportunity cost of money). but note that this argument depends critically on the step "the contract going above 50¢ is evidence that the market has determined that the coin is biased towards heads." but that argument only holds insofar as the people bidding the contracts from (say) 49¢ to 51¢ have actually worked out the coin's real bias (e.g., have actually run a laser-scanner or whatever).

intuition pump: suppose you bid the coin straight from 40¢ to 51¢ yourself, by accident, while still believing that the coin was very likely 40% likely to come up heads. the market closes in 5 minutes. what should you do? surely the answer is not "reason that, because the market price is above 50¢, somebody must have figured out that the coin is actually biased towards heads;" that'd be madness. sell.

more generally, nobody should bid a conditional branch into the top position unless they personally believe that it's worth it in counterfactual expectation. (or in other words: the conditional stops meaning "somebody else determined this was worth it" when you're the one pushing it over the edge; so when it comes to the person pushing the contract into the execution zone, from their perspective, the conditional matches the counterfactual.)

Reply1
[-]dynomight1mo30

Again, if there's a mistake, it would be helpful if you could explain exactly what that mistake is. You're sort of stating that the conclusion is mistaken and then giving a parallel argument for a different conclusion. It would be great (for multiple reasons) if you could explain exactly where my argument fails.

It might be helpful to focus on this example, which is pretty self-contained:

Suppose there’s a conditional prediction market for two coins. After a week of bidding, the markets will close, whichever coin had contracts trading for more money will be flipped and $1 paid to contract-holders for head. The other market is cancelled.

Suppose you’re sure that coin A, has a bias of 60%. If you flip it lots of times, 60% of the flips will be heads. But you’re convinced coin B, is a trick coin. You think there’s a 59% chance it always lands heads, and a 41% chance it always lands tails. You’re just not sure which.

We want you to pay more for a contract for coin A, since that’s the coin you think is more likely to be heads (60% vs 59%). But if you like money, you’ll pay more for a contract on coin B. You’ll do that because other people might figure out if it’s an always-heads coin or an always-tails coin. If it’s always heads, great, they’ll bid up the market, it will activate, and you’ll make money. If it’s always tails, they’ll bid down the market, and you’ll get your money back.

You’ll pay more for coin B contracts, even though you think coin A is better in expectation. Order is not preserved. Things do not work out.

Are you claiming that this is mistaken, or rather that this is correct but it's not a problem? (Of course, if this example is not central to what you see as a mistake, it could be the wrong thing to focus on.)

I've seen one argument which seems related to the one you're making and I do agree with. Namely, right before the market closes the final bidder has an incentive to bid their true beliefs, provided they know they will be the final bidder. I certainly accept that this is true. If you know the final closing price, then Y is no longer a random variable, and you're essentially just bidding in a non-conditional prediction market. I don't think this is completely reassuring on its own, though, because there's a great deal of tension with the whole idea of having a market equilibrium that reflects collective beliefs. I think you might be able to generalize this into some kind of an argument that as you get closer to closing, there's less randomness in Y and so you have more of an incentive to be honest. But this worries me because it would appear to lead to weird dynamics where people wait until the last second to bid. Of course, this might be a totally different direction from what you're thinking.

Reply
[-]So8res1mo74

Are you claiming that this is mistaken, or rather that this is correct but it's not a problem?

mistaken.

But if you like money, you’ll pay more for a contract on coin B.

this is an invalid step. it's true in some cases but not others, depending on how the act of paying for a contract on coin B (with no additional knowledge of whether it's double-headed) affects the chance that the market tosses coin B.

Reply
[-]Dweomite23d60

you’ll pay more for a contract on coin B. You’ll do that because other people might figure out if it’s an always-heads coin or an always-tails coin. If it’s always heads, great, they’ll bid up the market, it will activate, and you’ll make money. If it’s always tails, they’ll bid down the market, and you’ll get your money back.

So8res seems to be arguing that this reasoning only holds if your own purchase decision can't affect the market (say, if you're making a private bet on the side and both you and your counter-party are sworn to Bayesian secrecy). If your own bet could possibly change which contract activates, then you need to worry that contract B activates because you bid more than your true belief on it, in which case you lose money in expectation.

(Easy proof:  Assume all market participants have precisely the same knowledge as you, and all follow your logic; what happens?)

I think dynomight's reasoning doesn't quite hold even when your own bet is causally isolated, because:

  1. In order for you to pay more than $.59, you need to believe that the market is at least correlated with reality; that it's more likely to execute contract B if contract B actually is more valuable. (This is a pretty weak assumption, but still an important one.)
  2. In order for you to pay more than $.60 (not merely $.59 + epsilon), you not only need to believe that the market is correlated with reality, you need a quantitative belief that the correlation has at least a certain strength (enough to outweigh $.01). It's not enough for it to be theoretically possible that someone has better info than you; it needs to be plausible at a certain quantitative threshold of plausibility.

You can sort-of eliminate assumption #2 if you rework the example so that your true beliefs about A and B are essentially tied, but if they're essentially tied then it doesn't pragmatically matter if we get the order wrong. Assumption #2 places a quantitative bound on how wrong they can be based on how plausible it is that the market outperforms your own judgment.

Reply
[+][comment deleted]1mo20
[-]mako yass1mo20

At some point I'm gonna argue that this is a natural dutch book on CDT. (FDT wouldn't fall for this)

Reply
[-]Ruby23d*110

Curated. The idea of using Futarchy and prediction markets to make decision markets was among the earliest ideas I recall learning when I found the LessWrong/Rationality cluster in 2012 (and they continue to feature in dath ilani fiction). It's valuable then to have an explainer for fundamental challenges with prediction markets. I suggest looking at the comments and references, as there's some debate here, but overall I'm glad to have this key topic explored critically.

Reply
[-]Baram Sosis1mo72

In your theorem, I don't see how you get that E1[f(x,Y,Z)∣Y≥c]=E2[f(x,Y,Z)∣Y≥c]. Just because the conditional expectation of Z is the same doesn't mean the conditional expectation of f(x,Y,Z) is the same (e.g. you could have two different distributions over Z with the same expected value conditional on Y≥c but different shapes, and then have f depend non-linearly on Z, or something similar with Y). It seems like you'd need some stronger assumptions on f or whatever to get this to work. Or am I misunderstanding something?

(Your overall point seems right, though)

Reply1
[-]dynomight1mo73

Ack, I think you're right. I think I need to replace the assumption that 𝔼₁[Z | Y≥c] = 𝔼₂[Z | Y≥c] with the assumption that ℙ₁[(Y,Z)|Y≥c] = ℙ₂[(Y,Z)|Y≥c] which will guarantee the equality you're pointing out is true for all f. This seems totally fine since it's just a construction, but it definitely looks like an error to me. I'll fix that, thank you!

Reply1
[-]Daniel C1mo52

Suppose you run a market where if you pay x and the final market price is y and z happens, then you get a payout of f(x,y,z) dollars. The payout function can be anything, subject only to the constraint that if the final market price is below some constant c, then bets are cancelled, i.e. f(x,y,z)=x for y < c.


But in futarchy the "threshold price" c wouldn't be constant, it would be the price of the market conditional on the scenario y<c.

IIUC the theorem is saying that you would be indifferent to whatever happens to Z if y<c, but that counterfactual would be estimated by another market (which estimates c) that activates when y<c and cancels when y>=c

Reply
[-]dynomight1mo50

I think that's right. (I guess technically it depends on what version of Futarchy you're trying to use. You could have a single market for a single coin that's flipped iff the final price is above some threshold.)

In general I agree this doesn't capture the complexity of how markets would get resolved. For example, the most common case would probably be that you have two markets for two coins and you resolve whichever one has a higher price. That doesn't fit into my assumptions.

I guess I was implicitly trying to argue that it doesn't work even with this extra simplifying assumption.

But I think you can also pretty easily generalize the proof? Suppose we change C to be a random variable (an arbitrary random variable, which I think can reflect pretty much any market design), and we change the payout function to be f(x,y,z,c) with the restriction that f(x,y,z,c)=x for y<c. Then I think the same proof strategy still works, just changing f(x,Y,Z) to f(x,Y,Z,C) everywhere?

Reply
[-]Daniel C1mo30

I think that's right. (I guess technically it depends on what version of Futarchy you're trying to use. You could have a single market for a single coin that's flipped iff the final price is above some threshold.)[...] That doesn't fit into my assumptions.

Yep agreed.

But I think you can also pretty easily generalize the proof? [...] just changing f(x,Y,Z) to f(x,Y,Z,C) everywhere?

The proof definitely shows that within a single market (e.g. conditional on y>=c), you would be indifferent to Z given the opposite counterfactual (y<c), but that's okay because (in the two market for two coin case) we have two markets (estimating y and c) and each of the market would respond to each of the two possible counterfactuals (y>=c or y<c).

So although I would be indifferent between P1 & P2 in the first market (conditional on y>=c) when they imply different distribution on Z conditional on y<c, I would not be indifferent between P1 & P2 in the second market (conditional on y<c), and that difference would be reflected in c, which affects whether y>=c or y<c (& therefore which decision will be chosen).

Reply
[-]dynomight1mo20

Can you help me understand what you're claiming? Is it fair to think of your argument as supporting one of the following conclusions?

1. There is a function f with f(x,y,z,c)=x for y<c such that E[f]=E[z] always.

2. Rational agents would have some beliefs about the conditional distribution of c, and there is some function f for which that property is true once you add that assumption.

3. That property isn't necessary, some other (weaker) property is all that's needed.

Reply
[-]Daniel C1mo10

I think I'm claiming 3, namely all we want from futarchy is for it to select the decision with the highest expected payout, and for that the property isn't necessary.

Ex: For the two coin two market case, the first market's price estimates the expected payout if we flip coin one (& similarly for the second market), & while neither market satisfies the property (E[f]=E[z] always), we would still select the decision that leads to the higher expected payout (as we select the higher price), and that's all that's needed

Reply
[-]dynomight1mo30

How do you feel about this example, which gives a setup where you have an incentive to bid more for a coin you think has a lower expected value?

Suppose there’s a conditional prediction market for two coins. After a week of bidding, the markets will close, whichever coin had contracts trading for more money will be flipped and $1 paid to contract-holders for head. The other market is cancelled.

Suppose you’re sure that coin A, has a bias of 60%. If you flip it lots of times, 60% of the flips will be heads. But you’re convinced coin B, is a trick coin. You think there’s a 59% chance it always lands heads, and a 41% chance it always lands tails. You’re just not sure which.

We want you to pay more for a contract for coin A, since that’s the coin you think is more likely to be heads (60% vs 59%). But if you like money, you’ll pay more for a contract on coin B. You’ll do that because other people might figure out if it’s an always-heads coin or an always-tails coin. If it’s always heads, great, they’ll bid up the market, it will activate, and you’ll make money. If it’s always tails, they’ll bid down the market, and you’ll get your money back.

You’ll pay more for coin B contracts, even though you think coin A is better in expectation. Order is not preserved. Things do not work out.

More generally, what's the argument that the market will always select the decision that leads to he higher expected payout?

Reply
[-]Daniel C1mo30

We want you to pay more for a contract for coin A, since that’s the coin you think is more likely to be heads (60% vs 59%). But if you like money, you’ll pay more for a contract on coin B. You’ll do that because other people might figure out if it’s an always-heads coin or an always-tails coin. If it’s always heads, great, they’ll bid up the market, it will activate, and you’ll make money. If it’s always tails, they’ll bid down the market, and you’ll get your money back.

 

Let's call "Bidding on B, hoping that other people will figure out if B is an always-head or always-tails coin" strategy X,  and call "Figure out if B is an always-head or always-tail myself & bid accordingly, or if I can't, bid on A because it's better in expectation" strategy Y.

 

If I believe that sufficient number of people in the market are using strategy Y, then it's beneficial for me to use strategy X, and insofar as my beliefs about the market are accurate, this is okay, because sufficient number of people using strategy Y means the market will actually figure out if B is always-head or always-tail, then bid accordingly. So the market selects the right decision, insofar as my beliefs about the market is correct (Note that I'm never incentivized to place a bid on B so large that it causes B to activate, since I don't actually know if B is always-head).

 

On the other hand, if I believe that the vast majority of people in the market are using strategy X instead of strategy Y, then it's no longer beneficial for me to use strategy X myself, I should instead use strategy Y because the market doesn't actually do the work of finding out if coin B is always-head for me. Other traders who have accurate beliefs about the market will switch to strategy Y as well, until there is a sufficient number of trader to push the market towards the right decision.

So insofar as people have accurate beliefs about the market, the market will end up selecting the right decision (either sufficient number of people use strategy Y, in which case it's robust for me to use strategy X, or not enough people are using strategy Y, in which case people are incentivized to switch to Y)

More generally, what's the argument that the market will always select the decision that leads to he higher expected payout?

"Always" might be too strong, but very informally:

Suppose that we have we have decision d1 d2, with outcome/payoff u & conditional market prices p1 (corresponds to d1) p2 (corresponds to d2)
 

if p1>E[u|d1], then traders are incentivized to sell & drive down p1. Similarly they will be incentivized to bid up p1 if p1<E[u|d1]. So p1 will tend toward E[u|d1]. We can argue similar for p2 tending towards E[u|d2]

Since we choose the decision with the higher price, and prices tend towards the expected payoff given that decision, the market end up choosing the decision that leads to the higher expected payoff.


 

Reply
[-]dynomight1mo20

Let's call "Bidding on B, hoping that other people will figure out if B is an always-head or always-tails coin" strategy X,  and call "Figure out if B is an always-head or always-tail myself & bid accordingly, or if I can't, bid on A because it's better in expectation" strategy Y.

 

Regarding this, I'll note that my logic is not that different traders are following different strategies. I assume that all traders are rational agents and will maximize their expected return given their beliefs. My intended setup is that you believe coin A and coin B could have the biases stated, but you also believe that if you were to aggregate your beliefs with the beliefs of other people, the result would be more accurate than your beliefs alone.

I think this feeds into my objection to this proof:

Suppose that we have we have decision d1 d2, with outcome/payoff u & conditional market prices p1 (corresponds to d1) p2 (corresponds to d2)
 

if p1>E[u|d1], then traders are incentivized to sell & drive down p1. Similarly they will be incentivized to bid up p1 if p1<E[u|d1]. So p1 will tend toward E[u|d1]. We can argue similar for p2 tending towards E[u|d2]

Since we choose the decision with the higher price, and prices tend towards the expected payoff given that decision, the market end up choosing the decision that leads to the higher expected payoff.

My main objection to this logic is that there doesn't seem to be any reflection of the idea that different traders will have different beliefs. (It's possible that the market does give causal estimates with that assumption, but it's definitely not an assumption I'd be willing to make, since I think the central purpose of prediction markets is to aggregate diverse beliefs.) All my logic is based on a setup where different traders have different beliefs.

So I don't think the condition "p1>E[u|d1]" really makes sense? I think a given trader will drive down that market iff their estimate of the utility conditioned on that market activating is higher than p1, i.e. if p1>E_i[u|d1, market 1 activates]. I'm claiming that for trader i, E_i[u|d1, market 1 activates] != E_i[u|d1], basically because the event that market 1 activates contains extra information, and this makes it unlikely that the market will converge to E[u|d1].

Reply
[-]Daniel C1mo10

My main objection to this logic is that there doesn't seem to be any reflection of the idea that different traders will have different beliefs.[...] All my logic is based on a setup where different traders have different beliefs.


 Over time, traders who have more accurate beliefs (& act rationally according to those beliefs) will accumulate more money in expectation (& vice versa), so in the limit we can think of futarchy as aggregating the beliefs of different traders weighted by how accurate their beliefs were in the past

So I don't think the condition "p1>E[u|d1]" really makes sense? [...]and this makes it unlikely that the market will converge to E[u|d1].

If I pay p1 for a contract in market 1, my expected payoff is:

(E[u|d1]−p1)P(d1)+0×P(d2) (since I get my money back if d2/market 2 is activated)

this is negative iff p1>E[u|d1] and positive iff p1<E[u|d1]

and if we commit to using futarchy to choose the decision, then d1 is chosen iff market 1 activates, so E_i[u|d1, market 1 activates] should equal E_i[u|d1]
 

Reply
[-]dynomight1mo20

If I pay p1 for a contract in market 1, my expected payoff is:

(E[u|d1]−p1)P(d1)+0×P(d2) (since I get my money back if d2/market 2 is activated)

this is negative iff p1>E[u|d1] and positive iff p1<E[u|d1]

 

This is incorrect. There are two errors here:

  1. The first expectation needs to be conditioned on the market activating. (That is not conditionally independent of u given d1 in general.)
  2. Different people have different beliefs, so the expectations are different for different traders. You can't write "E" without specifying for which trader.

I agree that if you assume u is conditionally independent of market activation given d1 and that all traders have the same beliefs then the result seems to hold. But those assumptions are basically always false.

Reply
[-]Daniel C1mo10

The first expectation needs to be conditioned on the market activating. (That is not conditionally independent of u given d1 in general.)

 

If we commit to using futarchy to choose decision, then market 1 activating will have exactly the same truth conditions as executing d1, so "market activating and d1" would be the exact same thing as "d1" itself (commiting to use futarchy to choose decision means we assign 0 probability to "first market activating & execute d2" or "Second market activating & execute d1")

Different people have different beliefs, so the expectations are different for different traders. You can't write "E" without specifying for which trader.

Yes, we can replace with E_i, and then argue that traders with accurate beliefs will accumulate more money over time, making market estimates more accurate in the limit
 


 

Reply
[-]dynomight1mo20

Yes, we can replace with E_i, and then argue that traders with accurate beliefs will accumulate more money over time, making market estimates more accurate in the limit

 

There's a chicken-and-egg problem here. You're assuming that markets are causal (meaning traders that are better at estimating causal probabilities) and then using that assumption to prove that markets are causal.

Reply
[-]Daniel C1mo10

There's a chicken-and-egg problem here[...] and then using that assumption to prove that markets are causal.

 That argument was more about accomodating "different traders with different beliefs", but here's an independent argument for market being causal:

When I cause a particular effect/outcome, that means I mediate the influence between the cause of my action and the effect/outcome of my action, the cause of my action is conditionally independent of the effect of my action given me

Futarchy is a similar case: There may be many causes that influence market prices, which in turn determines the decision chosen, & market prices mediate the influence between the cause of market prices (e.g. different traders' beliefs) and the decision chosen. Any information can only influence what decision will be chosen through influencing the market prices. This seems like what it means for market to be causal (In a bayesnet, the decision chosen will literally only have market prices as the parent, assuming we commit to using futarchy to choose decisions).

 

 


 

Reply
[-]Matrice Jacobine22d30

This seems related to the ? Especially @Scott Garrabrant's version, considering logical induction is based on prediction markets.

Reply
[-]Mikhail Samin17d20

I’ve discussed this with some people about a month ago. I don’t believe this is a real flaw. Prediction markets can be structured to explicitly implement CDT, and I argue they normally do that on their own.

Reply
[-]Clay S25d20

 

i'm trying to understand the point about the vitamin d example. if a futarchy market is set up to predict whether "increasing vitamin d consumption" will "increase average lifespan," wouldn't participants who believe that wealth (or another confounder) is the actual causal factor, and not vitamin d itself, be incentivized to bet against the vitamin d policy leading to increased lifespan?

the market's incentive structure seems designed to filter for beliefs about the causal efficacy of the proposed intervention, not merely correlations. if people believe wealth is the cause, they wouldn't expect a vitamin d policy to succeed in raising lifespans, and would bet accordingly. it appears there might be a slight confusion between correlation in observational data and the causal impact of a direct intervention as assessed by a prediction market.

Reply
[-]_nuchi16d10

I commented on the post as well, but repeating here — I think denominating conditional markets in shares of contracts of the event happening (rather than in $) is more natural and resolves some confusions. i.e. denominate a contract that pays out if A conditional on B in shares of contracts that pay out $1 if B.

Shares of (coin comes up heads, conditional on coin being flipped) will then be traded in currency denominated in shares of (coin is flipped).

If everyone knows the coin will be flipped only if the laser scan comes up estimating the bias at > 0.5, then it's no surprise if the market is trading at > 0.5, and that really is everyone's true belief — namely their true belief of the probability of the coin coming up heads conditional on the coin being flipped. (Not their unconditional belief of the probability of the coin coming up heads!)

If instead everyone knows the coin is flipped only if the market is trading at > 0.5, then (probably?) most people's estimate of the coin's bias won't be meaningfully different depending on whether they condition on the coin being flipped or not — it's not strong evidence of the coin's bias — and so I expect the price to remain at whatever the market's prior of the bias was.

But those are two different situations, and I don't think the first one provides useful intuition for the second one. Setting up the right market is still a skill issue — in your Elon firing example, one would learn more from it by announcing ahead of time that the firing decision itself will be determined only by the results of the market. Then one doesn't have to worry about "well maybe their conditional probability is really due to some outside factor...". (Of course making such an announcement might cause both results to trade downward, because the market thinks that's a dumb way to make firing decisions. That's why it's a skill issue. If it seems like a dumb way to make firing decisions, maybe don't make such an announcement.)

Reply
[-]Bunthut19d*10

Suppose b is the true bias of the coin (which the supercomputer will compute). Then your expected return in this game is

𝔼[max(b, 0.50)] = 0.50 + 𝔼[max(b-0.50, 0)]

No. That formula would imply that, if the coin is 30% for sure and you buy it for 0.3, you make 0.2 in expectation, which you don't, you make 0 regardless of what price you buy at.

Note that this kind of problem has also shown up in decision theory more generally. This is a good place to start. In particular, it seems like your problem can be fixed with epsilon exploration (if it doesn't do so automatically, as per Soares), both the EDT and CDT variant should work.

Reply
[-]Kyle Pena20d10

There are specific conditions under which conditional probability differs from causal intervention. Suppose we are comparing conditional probability P(Y|X=x) to intervention P(Y|do(X=x)). When there are no "backdoor paths" from Y to X - which loosely speaking, are indirect paths of influence - then these are equal.

Reply
[-]rain8dome922d10

The best use of prediction markets is for decision makers to prove their competence.  Or as Lizka suggests to advise voting public.

Reply
[-]Austin Welch22d10

Can we build simulations that capture real-world causal relationships and optimize via RL, similar to the AI Economist paper?

Reply
[-]Vin1mo*10

Is one of the methods of getting causal estimates that you're going to write about simply to give each possible decision a non-zero random chance of being selected? Here's what Hanson wrote about this approach:

Conditioning on random decisions should very effective, but seems expensive. However, it seems much less expensive to sometimes randomly not do a policy change that you were going to do, than to sometimes randomly do a policy change you were not going to do. So I suggest that a futarchy system for considering and adopting proposals randomly reject say 5% of the changes that it would otherwise have accepted. This should ensure good estimates conditional on not adopting proposals, leaving only the potential problem of a decision selection bias distorting estimates conditional on adopting some proposal.

https://open.substack.com/pub/overcomingbias/p/decision-selection-bias

Reply
[-]deepthoughtlife26d-1-10

I believe that the most important fundamental flaw for Futarchy isn't what is written here, but that the flaw this essay identifies is in fact sufficient to be described as an important fundamental flaw. I do not believe it is possible to patch out the most fundamental flaw. 

My thought is that the most fundamental flaw is that you are asking for how people think the question will be resolved, and not the thing that is being referenced, so I think that prediction markets in general should be seen as just a (slightly unusual) type of opinion polling where you can theoretically win money by guessing other people's opinion's in the future. It's very meta. Prediction markets are very heavily selected for a specific type of person, so it is also an extremely biased poll. I do not believe that the self-selection is notably for accuracy. The chance to win money is likely to lead to notably more rigor than many other forms of polling (but also attracts gamblers).

Futarchy is just an unusual patch on top of prediction markets, and doesn't fundamentally change what a prediction market is. Like noted in the essay, that question is: what is the world like if this happens? And not: what effect does doing this have?

Reply
[+][comment deleted]1mo20
[+][comment deleted]1mo10
Moderation Log
Curated and popular this week
48Comments
Deleted by cousin_it, 06/15/2025
Reason: Comment deleted by its author.
Deleted by cousin_it, 06/14/2025
Reason: Comment deleted by its author.
Deleted by Vin, 06/18/2025
Reason: i get it know - there is still selection bias
5-and-10 problem
Mentioned in
24No, Futarchy Doesn’t Have an EDT Flaw

Say you’re Robyn Denholm, chair of Tesla’s board. And say you’re thinking about firing Elon Musk. One way to make up your mind would be to have people bet on Tesla’s stock price six months from now in a market where all bets get cancelled unless Musk is fired. Also, run a second market where bets are cancelled unless Musk stays CEO. If people bet on higher stock prices in Musk-fired world, maybe you should fire him.

That’s basically Futarchy: Use conditional prediction markets to make decisions.

People often argue about fancy aspects of Futarchy. Are stock prices all you care about? Could Musk use his wealth to bias the market? What if Denholm makes different bets in the two markets, and then fires Musk (or not) to make sure she wins? Are human values and beliefs somehow inseparable?

My objection is more basic: It doesn’t work. You can’t use conditional predictions markets to make decisions like this, because conditional prediction markets reveal probabilistic relationships, not causal relationships. The whole concept is faulty.

There are solutions—ways to force markets to give you causal relationships. But those solutions are painful and I get the shakes when I see everyone acting like you can use prediction markets to conjure causal relationships from thin air, almost for free.

I wrote about this back in 2022, but my argument was kind of sprawling and it seems to have failed to convince approximately everyone. So thought I’d give it another try, with more aggression.

Conditional prediction markets are a thing

In prediction markets, people trade contracts that pay out if some event happens. There might be a market for “Dynomight comes out against aspartame by 2027” contracts that pay out $1 if that happens and $0 if it doesn’t. People often worry about things like market manipulation, liquidity, or herding. Those worries are fair but boring, so let’s ignore them. If a market settles at $0.04, let’s assume that means the “true probability” of the event is 4%.

(I pause here in recognition of those who need to yell about Borel spaces or von Mises axioms or Dutch book theorems or whatever. Get it all out. I value you.)

Right. Conditional prediction markets are the same, except they get cancelled unless some other event happens. For example, the “Dynomight comes out against aspartame by 2027” market might be conditional on “Dynomight de-pseudonymizes”. If you buy a contract for $0.12 then:

  • If Dynomight is still pseudonymous at the end of 2027, you’ll get your $0.12 back.
  • If Dynomight is non-pseudonymous, then you get $1 if Dynomight came out against aspartame and $0 if not.

Let’s again assume that if a conditional prediction market settles at $0.12, that means the “true” conditional probability is 12%.

A non-causal kind of thing

But hold on. If we assume that conditional prediction markets give flawless conditional probabilities, then what’s left to complain about?

Simple. Conditional probabilities are the wrong thing. If P(A|B)=0.9, that means that if you observe B, then there’s a 90% chance of A. That doesn’t mean anything about the chances of A if you do B.

In the context of statistics, everyone knows that correlation does not imply causation. That’s a basic law of science. But really, it’s just another way of saying that conditional probabilities are not what you need to make decisions. And that’s true no matter where the conditional probabilities come from.

For example, people with high vitamin D levels are only ~56% as likely to die in a given year as people with low vitamin D levels. Does that mean taking vitamin D halves your risk of death? No, because those people are also thinner, richer, less likely to be diabetic, less likely to smoke, more likely to exercise, etc. To make sure we’re seeing the effects of vitamin D itself, we run randomized trials. Those suggest it might reduce the risk of death a little. (I take it.)

Futarchy has the same flaw. Even if you think vitamin D does nothing, if there’s a prediction market for if some random person dies, you should pay much less if the market is conditioned on them having high vitamin D. But you should do that mostly because they’re more likely to be rich and thin and healthy, not because of vitamin D itself.

If you like math, conditional prediction markets give you P(A|B). But P(A|B) doesn’t tell you what will happen if you do B. That’s a completely different number with a different notation, namely P(A|do(B)). Generations of people have studied the relationship between P(A|B) and P(A|do(B)). We should pay attention to them.

This is not hypothetical

Say people bet for a lower Tesla stock price when you condition on Musk being fired. Does that mean they think that firing Musk would hurt the stock price? No, because there could be reverse causality—the stock price dropping might cause him to be fired.

You can try to fight this using the fact that things in the future can’t cause things in the past. That is, you can condition on Musk being fired next week and bet on the stock price six months from now. That surely helps, but you still face other problems.

Here’s another example of how lower prices in Musk-fired world may not indicate that firing Musk hurts the stock price. Suppose:

  1. You think Musk is a mildly crappy CEO. If he’s fired, he’ll be replaced with someone slightly better, which would slightly increase Tesla’s stock price.
  2. You’ve heard rumors that Robyn Denholm has recently decided that she hates Musk and wants to dedicate her life to destroying him. Or maybe not, who knows.

If Denholm fired Musk, that would suggest the rumors are true. So she might try to do other things to hurt him, such as trying to destroy Tesla to erase his wealth. So in this situation, Musk being fired leads to lower stock prices even though firing Musk itself would increase the stock price.

Or suppose you run prediction markets for the risk of nuclear war, conditional on Trump sending the US military to enforce a no-fly zone over Ukraine (or not). When betting in these markets, people would surely consider about the risk that direct combat between the US and Russian militaries could escalate into nuclear war.

That’s good (the considering), but people would also consider that no one really knows exactly what Trump is thinking. If he declared a no-fly zone, that would suggest that he’s feeling feisty and might do other things that could also lead to nuclear war. The markets wouldn’t reflect the causal impact of a no-fly zone alone, because conditional probabilities are not causal.

Putting markets in charge doesn’t work

So far nothing has worked. But what if we let the markets determine what action is taken? If we pre-commit that Musk will be fired (or not) based on market prices, you might hope that something nice happens and magically we get causal probabilities.

I’m pro-hope, but no such magical nice thing happens.

Thought experiment. Imagine there’s a bent coin that you guess has a 40% chance of landing heads. And suppose I offer to sell you a contract. If you buy it, we’ll flip the coin and you get $1 if it’s heads and $0 otherwise. Assume I’m not doing anything tricky like 3D printing weird-looking coins. If you want, assume I haven’t even seen the coin.

You’d pay something like $0.40 for that contract, right?

(Actually, knowing my readers, I’m pretty sure you’re all gleefully formulating other edge cases. But I’m also sure you see the point that I’m trying to make. If you need to put the $0.40 in escrow and have the coin-flip performed by a Cenobitic monk, that’s fine.)

Now imagine a variant of that thought experiment. It’s the same setup, except if you buy the contract, then I’ll have the coin laser-scanned and ask a supercomputer to simulate millions of coin flips. If more than half of those simulated flips are heads, the bet goes ahead. Otherwise, you get your money back.

Now you should pay at least $0.50 for the contract, even though you only think there’s a 40% chance the coin will land heads.

Why? This is a bit subtle, but you should pay more because you don’t know the true bias of the coin. Your mean estimate is 40%. But it could be 20%, or 60%. After the coin is laser-scanned, the bet only activates if there’s at least a 50% chance of heads. So the contract is worth at least $0.50, and strictly more as long as you think it’s possible the coin has a bias above 50%.

(Math for people who like math.)

Suppose b is the true bias of the coin (which the supercomputer will compute). Then your expected return in this game is

  𝔼[max(b, 0.50)] = 0.50 + 𝔼[max(b-0.50, 0)],

where the expectations reflect your beliefs over the true bias of the coin. Since 𝔼[max(b-0.50, 0)] is never less than zero, the contract is always worth at least $0.50. If you think there’s any chance the bias is above 50%, then the contract is worth strictly more than $0.50.

To connect to prediction markets, let’s do one last thought experiment, replacing the supercomputer with a market. If you buy the contract, then I’ll have lots of other people bid on similar contracts for a while. If the price settles above $0.50, your bet goes ahead. Otherwise, you get your money back.

You should still bid more than $0.40, even though you only think there’s a 40% chance the coin will land heads. Because the market acts like a (worse) laser-scanner plus supercomputer. Assuming prediction markets are good, the market is smarter than you, so it’s more likely to activate if the true bias of the coin is 60% rather than 20%. This changes your incentives, so you won’t bet your true beliefs.

No, order is not preserved

I hope you now agree that conditional prediction markets are non-causal, and choosing actions based on the market doesn’t magically make that problem go away.

But you still might have hope! Maybe the order is still preserved? Maybe you’ll at least always pay more for coins that have a higher probability of coming up heads? Maybe if you run a market with a bunch of coins, the best one will always earn the highest price? Maybe it all works out?

Nope. You can create examples where you'll pay more for a contract on a coin that you think has a lower probability.

Suppose there’s a conditional prediction market for two coins. After a week of bidding, the markets will close, whichever coin had contracts trading for more money will be flipped and $1 paid to contract-holders for head. The other market is cancelled.

Suppose you’re sure that coin A, has a bias of 60%. If you flip it lots of times, 60% of the flips will be heads. But you’re convinced coin B, is a trick coin. You think there’s a 59% chance it always lands heads, and a 41% chance it always lands tails. You’re just not sure which.

We want you to pay more for a contract for coin A, since that’s the coin you think is more likely to be heads (60% vs 59%). But if you like money, you’ll pay more for a contract on coin B. You’ll do that because other people might figure out if it’s an always-heads coin or an always-tails coin. If it’s always heads, great, they’ll bid up the market, it will activate, and you’ll make money. If it’s always tails, they’ll bid down the market, and you’ll get your money back.

You’ll pay more for coin B contracts, even though you think coin A is better in expectation. Order is not preserved. Things do not work out.

No, it’s not easily fixable

Naive conditional prediction markets aren’t causal. Using time doesn’t solve the problem. Having the market choose actions doesn’t solve the problem. But maybe there’s still hope? Maybe it’s possible to solve the problem by screwing around with the payouts?

Theorem. Nope. You can’t solve the problem by screwing around with the payouts. There does not exist a payout function that will make you always bid your true beliefs.

(Click here for a version of that theorem with math. Warning: Math.)

Suppose you run a market where if you pay x and the final market price is y and z happens, then you get a payout of f(x,y,z) dollars. The payout function can be anything, subject only to the constraint that if the final market price is below some constant c, then bets are cancelled, i.e. f(x,y,z)=x for y < c.

Now, take any two distributions ℙ₁ and ℙ₂. Assume that:

  • ℙ₁[Y<c] = ℙ₂[Y<c] > 0
  • ℙ₁[Y≥c] = ℙ₂[Y≥c]
  • 𝔼₁[Z | Y≥c] = 𝔼₂[Z | Y≥c] ℙ₁[(Y,Z) | Y≥c] = ℙ₂[(Y,Z) | Y≥c] (h/t Baram Sosis)
  • 𝔼₁[Z | Y<c] ≠ 𝔼₂[Z | Y<c]

Then the expected return under ℙ₁ and ℙ₂ is the same. That is,

𝔼₁[f(x,Y,Z)]
  = x ℙ₁[Y<c] + ℙ₁[Y≥c] 𝔼₁[f(x,Y,Z) | Y≥c]
  = x ℙ₂[Y<c] + ℙ₂[Y≥c] 𝔼₂[f(x,Y,Z) | Y≥c]
  = 𝔼₂[f(x,Y,Z)].

Thus, you would be willing to pay the same amount for a contract under both distributions.

Meanwhile, the difference in expected values is

𝔼₁[Z] - 𝔼₂[Z]
  = ℙ₁[Y<c] 𝔼₁[Z | Y<c] - ℙ₂[Y<c] 𝔼₂[Z | Y<c]
    + ℙ₁[Y≥c] 𝔼₁[Z | Y≥c] - ℙ₂[Y≥c] 𝔼₂[Z | Y≥c]
  = ℙ₁[Y<c] (𝔼₁[Z | Y<c] - 𝔼₂[Z | Y<c])
  ≠ 0.

The last line uses our assumptions that ℙ₁[Y<c] > 0 and 𝔼₁[Z | Y<c] ≠ 𝔼₂[Z | Y<c].

Thus, we have simultaneously that

𝔼₁[f(x,Y,Z)] = 𝔼₂[f(x,Y,Z)],

yet

𝔼₁[Z] ≠ 𝔼₂[Z].

This means that you should pay the same amount for a contract if you believe ℙ₁ or ℙ₂, even though these entail different beliefs about how likely Z is to happen. Since we haven’t assumed anything about the payout function f(x,y,z), this means that no working payout function can exist. This is bad.

It’s not that bad

Just because conditional prediction markets are non-causal does not mean they are worthless. On the contrary, I think we should do more of them! But they should be treated like observational statistics—just one piece of information to consider skeptically when you make decisions.

Also, while I think these issues are neglected, they’re not completely unrecognized. For example, in 2013, Robin Hanson pointed out that confounding variables can be a problem:

Also, advisory decision market prices can be seriously distorted when decision makers might know things that market speculators do not. In such cases, the fact that a certain decision is made can indicate hidden info held by decision makers. Market estimates of outcomes conditional on a decision then become estimates of outcomes given this hidden info, instead of estimates of the effect of the decision on outcomes.

This post from Anders_H in 2015 is the first I’m aware of that points out the problem in full generality.

Finally, the flaw can be fixed. In statistics, there’s a whole category of techniques to get causal estimates out of data. Many of these methods have analogies as alternative prediction market designs. I’ll talk about those next time. But here’s a preview: None are free.