On Metaculus, people have attempted to solve this problem by having moderators review each question before it goes live. The result has generally been that most vague questions require substantial re-editing before they can appear on the website.
Also, over time, users have become quite skilled at noticing when questions are likely to resolve ambiguously.
Even with this cultural difference between the sites, many questions still resolve ambiguously on Metaculus. It's just really, really hard to say exactly what you mean when predicting the future.
I think Metaculus's level of verbosity in resolution criteria is bad in that it makes questions longer to write and longer to understand (because it takes longer to read and because its more complex). Part of the goal of Manifold is to remove trivial inconveniences so that people actually forecast at all, and so that we get markets on literally everything.
I think the synthesis here is to have a subset of high quality markets (clear resolution criteria, clear meta-resolution norms) but still have a fat tail of medium-quality questions.
David here from Manifold.
Really cool to see you taking the time to present your thoughts in a way that is constructive whilst not holding back with your criticisms. A lot of people would have just gotten frustrated with their experience on the site and have just moved on, so thank you for this post! We definitely think about these things a lot as we design Manifold and this post could definitely have an impact on where we lean within tough design problems.
First off, let me briefly discuss why we are bullish on using a creator-based resolution mechanism.
Now that that's out of the way, let's discuss some of the specific criticisms you outlined. The following is more a reflection of my own thoughts and hasn't necessarily been discussed in depth with the team.
Thanks again for taking the time to try out our site (even if it didn't go as you expected) and for doing this write-up. Am happy to discuss things further. Maybe what you've outlined is a bigger problem than we've realised and we need to spend a lot more effort designing the site around it, but I'm not yet convinced. It is good to know though that it is turning some users away.
I agree that ambiguity is bad, and most Manifold markets are probably too imprecise and ambiguous. My usual style is trying to be fairly precise in the forecasting questions I write, and I definitely second your recommendations!
However, I want to point out that the problem isn't just ambiguity, but really complexity. The more you try to nail down the resolution criteria, the more likely it becomes for there to be a serious mismatch between the top-line summary of the question (the question that you are actually interested in answering) and the detailed decision tree used to resolve it.
Here's a great example with a real-money, CFTC-regulated prediction market "Will student loan debt be forgiven?" on Kalshi. The rules were as unambiguous as possible up-front, specifying what would count as "broad-based debt relief" and things like that - things that people would reasonably think are important details to check. Unfortunately, it turned out that there was a huge detail in the resolution criteria that the market missed - the resolution rules required that the forgiveness be done via Executive Order or law. And in this case, the debt relief was clearly done but it wasn't done by law or executive order. So lots of traders lost lots of real money, and the prediction market clearly failed at its job. Link to the market: https://kalshi.com/events/SDEBT/markets/SDEBT-23JAN01. (Note: the page has been updated after the fact to highlight this issue.)
This is to highlight the point that writing unambiguous resolution criteria that also match the top-line summary question is really really hard! It's not uncommon on Metaculus for there to be a fairly clear subjective answer that differs from the actual resolution because of some issue that most people agree is a mere technicality. So I think that trying to disambiguate is good, but being willing to step back from technicalities is also potentially good. There are advantages to both approaches. I agree with you that the average market on Manifold is way too underspecified and ambiguous, but if you require Metaculus levels of precision then many people wouldn't be willing to spend the effort to ask the questions at all, and that is also a loss.
Also, on the Petrov day market, let's suppose the question had been "What % of Petrov Day will elapse before someone uses the big red button to take down Less Wrong's frontpage for the rest of the day?" ("for the rest of the day" is the only change.) I would consider this reasonably unambiguous - if LW decides to bring the page back up because it was a "mistake" then it shouldn't resolve yet. But I suspect that people would have bet on that nearly the same as the actual question, and your hypothetical user who saw the site was down at 10% would also have been burned. It's still better to avoid the ambiguity, I agree, but the problem of traders being burned by details is still there even if you avoid ambiguity in the details.
This sort of thing happens in the financial markets too. I'm thinking of all the games that happen over credit-default-swaps, for example. It would be nice if we could magically reduce complexity to reduce the impact of these sorts of issues, but it's a risk market participants are taking on by trading in the markets, and I think the value the markets provide is still clearly worth it (or else people wouldn't be willing to trade in them)!
Traders have the option, where a question is ambiguous, of asking the resolver how they would resolve it in some hypothetical scenario. This is true on Manifold as on Metaculus. I find this is normally more profitable for me than trying to get inside the head of the resolver.
There is a separate issue of resolver reputation, where some resolvers have a history of being biased in favor of their own positions, or just getting wrong. Definitely a weakness of current Manifold.
This post has a lot of good advice that I agree with, thanks for writing it.
I find this is normally more profitable for me than trying to get inside the head of the resolver.
If you ask in the comments, then the information about how the reviewer is likely to answer becomes public knowledge. There's more profit to be made if you are able to correctly predict the behavior without it being public knowledge.
Good point. Sometimes I do both. First I bet based on my attempt to get inside the head of the resolver. Then I ask them a question. When they respond I bet further (in one direction or another) based on their answer. When the market catches up with the new information I can slowly exit the market and bet elsewhere.
You're right that someone else can move the market based on the response first, but the site gives me a small assist: I get a notification when someone replies to my question, whereas nobody (yet) is subscribed to every comment thread. Maybe that will change with more volume. Also, since I asked the question, I probably have already thought about how different answers should move this and other markets.
Someone who was better at psychology-prediction than me has some better strategies, especially if someone else is betting on the market who thinks they are good at psychology-prediction and is not. There's lots of profit to be made that way, but also lots of loss.
FYI the creator of the market decided to resolve the market to N/A.
This means everyone's mana is fully restored as if they never interacted with the market.
Strongly agree about the existence of the problem. It's something I've put a bit of thought into.
One thing I think could help, in some cases, would be to split the market definition into
And then specify the relationship between them. For example:
Question: How many reported covid cases will there be in the US on [DATE]?
Resolution method: Look at https://covid.cdc.gov/covid-data-tracker/ a week after [DATE] for the reported values for [DATE].
Resolution notes: "Whatever values are reported that day will be used unconditionally." or "If the values change within the following week, the updated values will be used instead." or "The resolution method is a guideline only; ultimately the intent of the question, as interpreted by the question author, overrides."
This will only solve a subset of ambiguous resolutions, but I think it would still be a big help to spell some of these things out more clearly.
Markets should probably have multiple outcomes (which in politically controversial scenarios should ideally also be evaluated by different people with different biases) and allow people to make correlated bets over the outcomes.
An off-hand attempt at market mechnanism to solve this. Have "resolution issue" predictors. Like "There will be a news paper article of lenght more than 20 cm written by a full time journalist about biological weapons AND the vague biological weapons question will resolve in the negative".
Maybe we should just let people bet on N/A similar to Augur (with some stronger norm of resolving N/A in ambiguous cases)
(Engineer at Manifold here.) I largely agree! Letting people make markets on anything means that many people will make poorly operationalized markets. Subjective resolution is good for bets on personal events among friends, which is an important use case, but it bad for questions with an audience bigger than that.
We need to do a better job of:
1. resolution reliability, like by adding a reputation system or letting creators delegate resolution to more objective individuals / courts.
2. helping users turn their vague uncertainties into objective questions - crucial but less straightforward.
3. surfacing higher quality content
For the "Will Russia use chemical or biological weapons in 2022?" question, the creator provided information about an ambiguous outcome, though it seems very subjective:
If, when the question closes, there is widespread reporting that Russia did the attacks and there is not much reported doubt, then I will resolve YES. If it seems ambiguous I will either resolve as N/A or at a percentage that I think is reasonable. (Eg. resolve at 33% if I think there’s a 2/3 chance that the attacks were false flag attacks.) This of course would also go the other way if there are supposed Ukrainian attacks that are widely believed to be caused by Russia.
I think there are several similar such markets - the one I was looking at was at https://manifold.markets/Gabrielle/will-russia-use-chemical-or-biologi-e790d5158f6a and lacks such a comment.
EDITED: Ah, you are correct and I am wrong, the text you posted is present, it's just down in the comments section rather than under the question itself. That does make this question less bad, though it's still a bit weird that the question had to wait for someone to ask the creator that (and, again, the ambiguity remains).
I'll update the doc with links to reduce confusion - did not do that originally out of a mix of not wanting to point too aggressively at people who wrote those questions and feeling lazy.
I really buy the argument Sinclair makes about reducing trivial inconveniences here. Let’s make a model.
Ambiguity has two main negative effects, according to me:
Let’s set 1. aside for now. 2. seems like a big deal for sure. But also big is the drag on creating prediction markets imposed by Metaculus-style process. The way to balance these two seems to me to be a question that hinges on what you want your impact to be. If you’re trying to make the world as good a place as possible, you might have quite a strong preference for there being plenty of markets that can be made with low overhead. If the experience for forecasters is bad enough, then you won’t get predictions on those questions, but my empirical belief is that Manifold is striking a better balance right now than Metaculus.
As for 1., again, we can answer it according to, what’s most useful for our goals, and again, I want to claim Manifold is doing well here.
tl;dr There are real tradeoffs here.
I like the post btw!
(Disclaimers: I work in the financial industry, though not in a way related to prediction markets. Anything I write here is my opinion and not that of my employer.)
SOMEBODY SET US UP THE BOMB
You've heard of this new 'prediction market' fad, so you decide to try it out. You make a Manifold account and look for interesting questions.
There's a question about the 2022 LessWrong Petrov Day celebration "What % of Petrov Day will elapse before someone uses the big red button to take down Less Wrong's frontpage?" You navigate to the LW homepage to find out more information...and the site is down! Someone has pressed the button only a couple hours into Petrov Day! If it just went down now, that would mean the market would resolve to 10%...and if it went down a while ago the market would resolve even lower...but the market is currently trading at ~20%.
You're new to the idea of prediction markets, but even you have heard that if a market seems to be doing something silly, you can profit from correcting it. You buy the market down to 10%, and sit happily waiting for your profits.
Huh. It seems like LW admins have put the homepage back up? Apparently there was some kind of bug letting unauthorized users press the button. That's a bit odd? Still, it doesn't seem like it should affect you? Someone did use the big red button to take down LW's homepage only a few hours into Petrov Day, so the market should....
Wait, what? The market is not resolving? Apparently whoever organized it decided that the webpage going down due to someone pressing the red button "didn't count" and is letting the market continue and waiting for a "real" takedown of the webpage.
You are now short a large investment at prices ranging from 10-20%, while the market has responded to this announcement by going back up to ~30%...and it's rising further as the webpage continues not going down.
You thank your lucky stars that this fiasco only involved play money, delete your Manifold account, and resolve never to support real-money prediction markets for any genuine issue.
CURRENT MANIFOLD QUESTIONS
The Petrov Day question above is what prompted me to write this. However, I don't think it's an isolated bad question. I think many other Manifold questions have larger levels of ambiguity. (The Petrov Day question at least only ended up ambiguous due to an unusual event).
A few examples from scrolling down my Manifold search page (apologies to the authors of these. I don't think you're unusually bad people, I'm just picking on you as examples of what I consider a common problem):
MARKET STRUCTURE
One of the most valuable things a market resolution process can do is be unambiguous. Once reality has happened, it should be easy for someone who looks at reality, and looks at the market, to determine how the market should resolve.
If this is not the case, traders in the market are not trading on their knowledge of reality - they are trading on their knowledge of the market resolution process. And market prices aren't informative about reality - they are informative about the resolution process.
(While I don't work in this area directly, I've heard that similar things happen in real markets in the field of distressed-debt investing and credit default swaps - investments in these areas apparently are frequently decided by court cases over the legality of various things, and so serious investing in this area requires serious legal support and careful reading of documents).
The Manifold questions above are related to the real-world issues they ask about. However, if I were trying to trade real money in them, by far the most important thing I would want to determine would be the political leanings/prior views of the authors:
These are real-world questions of their own in a sense. There is real-world research you could do to investigate them (e.g. you could look at the author's prior statements on Russia/Ukraine/AlphaZero). These questions are not what these markets claimed to be about...so what? These questions are what these markets are actually about.
If you put serious money on the Russian-chemical-weapons Manifold market and got hedge funds interested in it (do not do this), the question these hedge funds would actually be researching to decide between 5% and 10% would be 'how sympathetic has the author previously been to Russia/Ukraine'.
This is a problem.
ACTIONS
My recommended courses of action would be: