Great work with this newsletter, keep it up. It's by far the best forecasting digest I've seen!
The major friction for me is that some of the formatting makes it feel overwhelming. Maybe use bold headings instead of bullet points for each new entry? Not sure.
Thanks.
The major friction for me is that some of the formatting makes it feel overwhelming. Maybe use bold headings instead of bullet points for each new entry? Not sure.
Fair point; will consider.
Mod note: Fixed the broken formatting. Looks like you pasted some markdown into our WYSIWYG editor.
Great links, thanks.
The Augur launch has unfortunately been a complete catastrophe, as the high transaction costs of ETH right now make it so that simply making a trade costs about $30...I hope they manage to come up with some sort of solution.
Highlights
Social Science Prediction Platform launches.
Ioannidis and Taleb discuss optimal response to COVID-19.
Report tries to foresee the (potentially quite high) dividends of conflict prevention from 2020 to 2030.
Index
Highlights.
Prediction Markets & Forecasting Platforms.
New undertakings.
Negative Examples.
News & Hard to Categorize Content.
Long Content.
Sign up here, browse past newsletters here, or view it on the EA forum here.
Prediction Markets & Forecasting Platforms.
Ordered in subjective order of importance:
Metaculus continues hosting great discussion.
In particular, it has recently hosted some high-quality AI questions.
User @alexrjl, a moderator on the platform, offers on the EA forum to operationalize questions and post them on Metaculus, for free. This hasn't been picked up by the EA Forum algorithms, but the offer seems to me to be quite valuable. Some examples of things you might want to see operationalized and forecasted: the funding your organization will receive in 2020, whether any particularly key bills will become law, whether GiveWell will change their top charities, etc.
Foretell is a prediction market by the University of Georgetown's Center for Security and Emerging Technology, focused on questions relevant to technology-security policy, and on bringing those forecasts to policy-makers.
Some EAs, such as myself or a mysterious user named foretold, feature on the top spots of their (admittedly quite young) leaderboard.
I also have the opportunity to create a team on the site: if you have a proven track record and would be interested in joining such a team, get in touch before the 10th of August.
Replication Markets
published their first paper
had some difficulties with cheaters:
still have Round 10 open until the 3rd of August.
At the Good Judgement family, Good Judgement Analytics continues to provide its COVID-19 dashboard.
PredictIt & Election Betting Odds each give a 60%-ish to Biden.
See Limits of Current US Prediction Markets (PredictIt Case Study), on how spread, transaction fees, withdrawal fees, interest rate which one could otherwise be earning, taxes, and betting limits make it so that:
New undertakings
On the one hand, I could imagine this having an impact, and the enthusiasm of the founders is contagious. On the other hand, as a forecaster I don't feel enticed by the platform: they offer a $25 reward to grad students (which I am not), and don't spell it out for me why I would want to forecast on their platform as opposed to on all the other alternatives available to me, even accounting for altruistic impact.
Ought is a research lab building tools to delegate open-ended reasoning to AI & ML systems.
Since concluding their initial factored cognition experiments in 2019, they’ve been building tools to capture and automate the reasoning process in forecasting: Ergo, a library for integrating model-based and judgmental forecasting, and Elicit, a tool built on top of Ergo to help forecasters express and share distributions.
They’ve recently run small-scale tests exploring amplification and delegation of forecasting, such as: Amplify Rohin’s Prediction on AGI researchers & Safety Concerns, Amplified forecasting: What will Buck’s informed prediction of compute used in the largest ML training run before 2030 be?, and Delegate a Forecast.
In addition to studying factored cognition in the forecasting context, they are broadly interested in whether the EA community could benefit from better forecasting tools: they can be reached out to team@ought.org if you want to give them feedback or discuss their work.
The Pipeline Project is a project similar to Replication Markets, by some of the same authors, to find out whether people can predict whether a given study will replicate. They offer authorship in an appendix, as well as a chance to get a token monetary compensation.
USAID's Intelligent Forecasting: A Competition to Model Future Contraceptive Use. "First, we will award up to 25,000 USD in prizes to innovators who develop an intelligent forecasting model—using the data we provide and methods such as artificial intelligence (AI)—to predict the consumption of contraceptives over three months. If implemented, the model should improve the availability of contraceptives and family planning supplies at health service delivery sites throughout a nationwide healthcare system. Second, we will award a Field Implementation Grant of approximately 100,000 to 200,000 USD to customize and test a high-performing intelligent forecasting model in Côte d’Ivoire."
Omen is another cryptocurrency-based prediction market, which seems to use the same front-end (and probably back-end) as Corona Information Markets. It's unclear what their advantages with respect to Augur are.
Yngve Høiseth releases a prediction scorer, based on his previous work on Empiricast. In Python, but also available as a REST API
Negative Examples.
News & Hard to Categorize Content.
See also: How accurate are [US] agencies’ procurement forecasts? and Forecasting Inflation in a Data-Rich Environment: The Benefits of Machine Learning Methods (which finds random forests a hard to beat approach)
Bloomberg on the IMF's track record on forecasting (archive link, without a paywall).
I keep seeing evidence that Trump will lose reelection, but I don't know how seriously to take it, because I don't know how filtered it is.
For example, the The Economist's model forecasts 91% that Biden will win the upcoming USA elections. Should I update somewhat towards Biden winning after seeing it? What if I suspect that it's the most extreme model, and that it has come to my attention because of that fact? What if I suspect that it's the most extreme model which will predict a democratic win? What if there was another equally reputable model which predicts 91% for Trump, but which I never got to see because of information filter dynamics?
The the Primary Model confirmed my suspicions of filter dynamics. It "does not use presidential approval or the state of the economy as predictors. Instead it relies on the performance of the presidential nominees in primaries", and on how many terms the party has controlled the White House. The model has been developed by an otherwise unremarkable professor of political science at New York's Stony Brook University, and has done well in previous election cycles. It assigns 91% to Trump winning reelection.
Forecasting at Uber: An Introduction. Uber forecasts demand so that they know amongst other things, when and where to direct their vehicles. Because of the challenges to testing and comparing forecasting frameworks at scale, they developed their own software for this.
Forecasting Sales In These Uncertain Times.
Long Content.
Lukas Gloor on takeaways from Covid forecasting on Metaculus
Ambiguity aversion. "Better the devil you know than the devil you don't."
Expert Forecasting with and without Uncertainty Quantification and Weighting: What Do the Data Say?: "it’s better to combine expert uncertainties (e.g. 90% confidence intervals) than to combine their point forecasts, and it’s better still to combine expert uncertainties based on their past performance."
How to build your own weather forecasting model. Sailors realize that weather forecasting are often corrupted by different considerations (e.g., a reported 50% of rain doesn't happen 50% of the time), and search for better sources. One such source is the original, raw data used to generate weather forecasts: GRIB files (Gridded Information in Binary), which lack interpretation. But these have their own pitfalls, which sailors must learn to take into account. For example, GRIB files only take into account wind speed, not tidal acceleration, which can cause a significant increase in apparent wind.
I liked the following taxonomy of what distinct targets the agency the first speaker works for is aiming to hit with their forecasts:
as an input into the policy-making process,
as a transparent assessment of public finances
as a prediction of whether the government will meet whatever fiscal rules it has set itself,
as a baseline against which to judge the significance of further news,
as a challenge to other agencies "to keep the bastards honest".
The limitations were interesting as well:
they require us to produce a forecast that's conditioned on current government policy even if we and everyone else expect that policy to change that of course makes it hard to benchmark our performance against counterparts who are producing unconditional forecasts.
The forecasts have to be explainable; a black box model might be more accurate but be less useful.
they require detailed discussion of the individual forecast lines and clear diagnostics to explain changes from one forecast to the next precisely to reassure people that those changes aren't politically motivated or tainted - the forecast is as much about delivering transparency and accountability as about demonstrating predictive prowess
the forecast numbers really have to be accompanied by a comprehensible narrative of what is going on in the economy and the public finances and what impact policy will have - Parliament and the public needs to be able to engage with the forecast we couldn't justify our predictions simply with an appeal to a statistical black box and the Chancellor certainly couldn't justify significant policy positions that way.
Note to the future: All links are added automatically to the Internet Archive. In case of link rot, go here