Metaculus is an online forecasting platform and aggregation engine that brings together a global reasoning community and keeps score for thousands of forecasters, delivering machine learning-optimized aggregate forecasts on topics of global importance. The Metaculus forecasting community is often inspired by altruistic causes, and Metaculus has a long history of partnering with nonprofit organizations, university researchers and companies to increase the positive impact of its forecasts.
Metaculus therefore poses questions about the occurrence of a variety of future events, on many timescales, to a community of participating forecasters — you!
The name "Metaculus" comes from the Metaculus genus in the Eriophyidae family, a genus of herbivorous mites found in many locations around the world.
Forecasting is a systematic practice of attempting to answer questions about future events. On Metaculus, we follow a few principles to elevate forecasting above simple guesswork:
First, questions are carefully specified so that everyone understands beforehand and afterward which kinds of outcomes are included in the resolution, and which are not. Forecasters then give precise probabilities that measure their uncertainty about the outcome.
Second, Metaculus aggregates the forecasts into a community prediction based on the median of user forecasts weighted by recency. Surprisingly, the Community Prediction is often better than any individual predictor! This principle is known as the wisdom of the crowd, and has been demonstrated on Metaculus and by other researchers. Intuitively it makes sense, as each individual has separate information and biases which in general balance each other out (provided the whole group is not biased in the same way).
Third, we measure the relative skill of each forecaster, using their quantified forecasts. When we know the outcome of the question, the question is "resolved", and forecasters receive their scores. By tracking these scores from many forecasts on different topics over a long period of time, they become an increasingly better metric of how good a given forecaster is. These scores provide aspiring forecasters with important feedback on how they did and where they can improve.
Forecasting is uniquely valuable primarily in complex, multi-variable problems or in situations where a lack of data makes it difficult to predict using explicit or exact models.
In these and other scenarios, aggregated predictions of strong forecasters offer one of the best ways of predicting future events. In fact, work by the political scientist Philip Tetlock demonstrated that aggregated predictions were able to outperform professional intelligence analysts with access to classified information when forecasting various geopolitical outcomes.
Research has shown that great forecasters come from various backgrounds—and oftentimes from fields that have nothing to do with predicting the future. Like many mental capabilities, prediction is a talent that persists over time and is a skill that can be developed. Steady quantitative feedback and regular practice can greatly improve a forecaster's accuracy.
Some events — such as eclipse timing and well-polled elections, can often be predicted with high resolution, e.g. 99.9% likely or 3% likely. Others — such as the flip of a coin or a close horse-race — cannot be accurately predicted; but their odds still can be. Metaculus aims at both: to provide a central generation and aggregation point for predictions. With these in hand, we believe that individuals, groups, corporations, governments, and humanity as a whole will make better decisions.
As well as being worthwhile, Metaculus aims to be interesting and fun, while allowing participants to hone their prediction prowess and amass a track-record to prove it.
Metaculus originated with two researcher scientists, Anthony Aguirre and Greg Laughlin. Aguirre, a physicist, is a co-founder of The Foundational Questions Institute, which catalyzes breakthrough research in fundamental physics, and of The Future of Life Institute, which aims to increase the benefit and safety of disruptive technologies like AI. Laughlin, an astrophysicist, is an expert at predictions from the millisecond predictions relevant to high-frequency trading to the ultra-long-term stability of the solar system.
Metaculus tournaments are organized around a central topic or theme. Tournaments are often collaborations between Metaculus and a nonprofit, government agency, or other organization seeking to use forecasting to support effective decision making. You can find current and archived tournaments in our Tournaments page.
Tournaments are the perfect place to prove your forecasting skills, while helping to improve our collective decision making ability. Cash prizes and Medals are awarded to the most accurate forecasters, and sometimes for other valuable contributions (like comments). Follow a Tournament (with the Follow button) to never miss new questions.
After at least one question has resolved, a Leaderboard will appear on the tournament page displaying current scores and rankings. A personal score board ("My Score") will also appear, detailing your performance for each question (see How are Tournaments Scored?).
At the end of a tournament, the prize pool is divided among forecasters according to their forecasting performance. The more you forecasted and the more accurate your forecasts were, the greater proportion of the prize pool you receive.
If you have outstanding tournament winnings, Metaculus is happy to facilitate donations to various non-profits, regranting organizations, and funds. You can find the list of organizations we facilitate payments to here.
Like Tournaments, Question Series are organized around a central topic or theme. Unlike tournaments, they do not have a prize pool.
Question Series still show leaderboards, for interest and fun. However they do not award medals.
You can find all Question Series in a special section of the Tournaments page.
Metaculus has some similarities to a prediction market, but ultimately is not one. Metaculus aims to aggregate many people's information, expertise, and predictive power into high-quality forecasts. However, prediction markets generally operate using real or virtual currency, where participants buy (or sell) shares if they think that the standing prices reflect too low (or high) a probability of an event occurring. Metaculus, in contrast, directly solicits predicted probabilities from its users, then aggregates those probabilities. We believe that this sort of "prediction aggregator" has both advantages and disadvantages relative to a prediction market.
Metaculus has several advantages over prediction markets, described below, but we want to preface this by saying that despite the potential issues with prediction markets that we describe here, we think prediction markets are valuable, are glad they exist, and would be glad to see more use of them.
No. Opinion polling can be a useful way to gauge the sentiment and changes in a group or culture, but there is often no single "right answer", as in a Gallup poll "How worried are you about the environment?"
In contrast, Metaculus questions are designed to be objectively resolvable (like in Will Brent Crude Oil top $140/barrel before May 2022?), and forecasters are not asked for their preferences, but for their predictions. Unlike in a poll, over many predictions, participants accrue a track record indicating their forecasting accuracy. These track records are incorporated into the Metaculus Prediction. The accuracy of the Metaculus track record itself is tracked here.
Questions should focus on tangible, objective facts about the world which are well-defined and not a matter of opinion. "When will the United States collapse?" is a poor, ambiguous question; What will be the US' score in the Freedom in the World Report for 2050?
is more clear and definite. They generally take the form Will (event) X happen by (date) Y?
or When will (event) X occur?
or What will the value or quantity of X be by (date) Y?
A good question will be unambiguously resolvable. A community reading the question terms should be able to agree, before and after the event has occurred, whether the outcome satisfies the question's terms.
Questions should also follow some obvious rules:
Many questions are launched by Metaculus staff, but any logged-in user can propose a question. Proposed questions will be reviewed by a group of moderators appointed by Metaculus. Moderators will select the best questions submitted, and help to edit the question to be clear, well-sourced, and aligned with our writing style.
Metaculus hosts questions on many topics, but our primary focus areas are Science, Technology, Effective Altruism, Artificial Intelligence, Health, and Geopolitics.
We currently receive a large volume of question submissions, many of which are interesting and well-written. That said, we try to approve just enough questions that they each can get the attention they deserve from our forecasters. Metaculus prioritizes questions on Science, Technology, Effective Altruism, Artificial Intelligence, Health, and Geopolitics. If your question falls into one of these categories, or is otherwise very urgent or important, you can tag us with @moderators to get our attention.
If a question is still waiting for resolution, check to make sure there hasn't been a comment from staff explaining the reason for the delay. If there hasn't, you can tag @admins to alert the Metaculus team. Please do not use the @admins tag more than once per week regarding a single question or resolution.
Private questions are questions that are not visible to the broader community. They aren't subject to the normal review process, so you can create one and predict on it right away. You can resolve your own private questions at any time, but points for private predictions won't be added to your overall Metaculus score and they won't affect your ranking on the leaderboard.
You can use private questions for anything you want. Use them as practice to calibrate your predictions before playing for points, create a question series on a niche topic, or pose personal questions that only you can resolve. You can even invite up to 19 other users to view and predict on your own questions!
To invite other forecasters to your private question, click the '...' more options menu and select 'Share Private Question'.
We have a full set of community etiquette guidelines but in summary:
To reduce ambiguity in an efficient way, here are some definitions that can be used in questions, with a meaning set by this FAQ:
Binary questions can resolve as either Yes or No (unless the resolution criteria were underspecified or otherwise circumvented, in which case they can resolve as Ambiguous). Binary questions are appropriate when an event can either occur or not occur. For example, the question "Will the US unemployment rate stay above 5% through November 2021?" resolved as No because the unemployment rate dropped below 5% before the specified time.
Range questions resolve to a certain value, and forecasters can specify a probability distribution to estimate the likelihood of each value occurring. Range questions can have open or closed bounds. If the bounds are closed, probability can only be assigned to values that fall within the bounds. If one or more of the bounds are open, forecasters may assign probability outside the boundary, and the question may resolve as outside the boundary. See here for more details about boundaries on range questions.
The range interface allows you to input multiple probability distributions with different weights. See here for more details on using the interface.
There are two types of range questions, numeric range questions and date range questions.
Numeric range questions can resolve as a numeric value. For example, the question "What will be the 4-week average of initial jobless claims (in thousands) filed in July 2021?" resolved as 395, because the underlying source reported 395 thousand initial jobless claims for July 2021.
Questions can also resolve outside the numeric range. For example, the question "What will the highest level of annualised core US CPI growth be, in 2021, according to U.S. Bureau of Labor Statistics data?" resolved as > 6.5 because the underlying source reported more than 6.5% annualized core CPI growth in the US, and 6.5 was the upper bound.
Date range questions can resolve as a certain date. For example, the question "When will the next Public Health Emergency of International Concern be declared by the WHO?" resolved as July 23, 2022, because a Public Health Emergency of International Concern was declared on that date.
Questions can also resolve outside the date range. For example, the question "When will a SpaceX Super Heavy Booster fly?" resolved as > March 29, 2022 because a SpaceX Super Heavy booster was not launched before March 29, 2022, which was the upper bound.
Question groups are sets of closely related questions or question outcomes all collected on a single page. Forecasters can predict quickly and efficiently on these interconnected outcomes, confident that they are keeping all of their predictions internally consistent.
With question groups, it's easy to forecast progressively wider distributions the further into the future you predict to reflect increasing uncertainty. A question group collecting multiple binary questions on a limited set of outcomes or on mutually exclusive outcomes makes it easier to see which forecasts are in tension with each other.
When regular forecast questions are converted into "subquestions" of a question group, the original pages are replaced by a single question group page. Comments that previously lived on the individual question pages are moved to the comment section of the newly created group page with a note indicating the move.
No. Question groups comprise multiple independent subquestions. For that reason, there is no requirement that you forecast on every outcome within a group.
Each outcome or subquestion is scored in the same manner as a normal independent question.
Even if there can only be one outcome for a particular question group, the Community Prediction functions as it would for normal independent questions. The Community Prediction will still display a median or a weighted aggregate of the forecasts on each subquestion, respectively. These medians and weighted aggregates are not constrained to sum to 100%
Feedback for question groups can be provided on the question group discussion post.
A Conditional Pair is a special type of Question Group that elicits conditional probabilities. Each Conditional Pair sits between a Parent Question and a Child Question. Both Parent and Child must be existing Metaculus Binary Questions.
Conditional Pairs ask two Conditional Questions (or "Conditionals" for short), each corresponding to a possible outcome of the Parent:
The first Conditional assumes that "The Parent resolves Yes" (or "if Yes" for short). The second conditional does the same for No.
Conditional probabilities are probabilities, so forecasting is very similar to Binary Questions. The main difference is that we present both conditionals next to each other for convenience:
Conditional questions are automatically resolved when their Parent and Child resolve:
Let's work through an example:
So the two Conditionals in the Conditional Pair will be:
For simplicity, Metaculus presents conditional questions graphically. In the forecasting interface they are in a table:
And in the feeds, each possible outcome of the Parent is an arrow, and each conditional probability is a bar:
Back to the example:
It rains today. The parent resolves Yes. This triggers the second conditional ("if No") to be annulled. It is not scored.
You wait a day. This time it doesn't rain. The Child resolves No. This triggers the remaining Conditional ("if Yes") to resolve No. It is scored like a normal Binary Question.
You can create and submit conditional pairs like any other question type. On the 'Create a Question' page, select Question Type 'conditional pair' and select Parent and Child questions.
Note: You can use question group subquestions as the Parent or Child by clicking the Parent or Child button and then either searching for the subquestion in the field or pasting the URL for the subquestion.
To copy the URL for a subquestion, simply visit a question group page and click the '...' more options menu to reveal the Copy Link option.
Questions on Metaculus are sorted by activity by default. Newer questions, questions with new comments, recently upvoted questions, and questions with many new predictions will appear at the top of the Metaculus homepage. However, there are several additional ways to find questions of interest and customize the way you interact with Metaculus.
The search bar can be used to find questions using keywords and semantic matches. At this time it cannot search comments or users.
Questions can be sorted and filtered in a different manner from the default using the filters menu. Questions can be filtered by type, status and participation. Questions can also be ordered, for example by "Newest". Note that the options available change when different filters are selected. For example, if you filter by "Closed" questions you will then be shown an option to order by "Soonest Resolving".
When submitting a question, you are asked to specify the closing date (when the question is no longer available for predicting) and resolution date (when the resolution is expected to occur). The date the question is set live for others to forecast on is known as the open date.
In some cases, questions must resolve at the resolution date according to the best available information. In such cases, it becomes important to choose the resolution date carefully. Try to set resolution dates that make for interesting and insightful questions! The date or time period the question is asking about must always be explicitly mentioned in the text (for example, "this question resolves as the value of X on January 1, 2040, according to source Y" or "this question resolves as Yes if X happens before January 1, 2040)".
The close date must be at least one hour prior to the resolution date, but can be much earlier, depending upon the context. Here are some guidelines for specifying the close date:
Note: Previous guidance suggested that a question should close between 1/2 to 2/3 of the way between the open time and resolution time. This was necessary due to the scoring system at the time, but has been replaced by the above guidelines due to an update to the scoring system.
For dates and times written in the question, such as "will event X happen before January 1, 2030?", if the timezone is not specified Coordinated Universal Time (UTC) will be used. Question authors are free to specify a different timezone in the resolution criteria, and any timezone specified in the text will be used.
For date range questions, the dates on the interface are in UTC. Typically the time of day makes little difference as one day is miniscule in comparison to the full range, but occasionally for shorter term questions the time of day might materially impact scores. If it is not clear what point in a specified period a date range question will be resolved as, it resolves as the midpoint of that period. For example, if a question says it will resolve as a certain day, but not what time of day, it will resolve as noon UTC on that day.
Only Metaculus Administrators can resolve questions. Binary questions can resolve Yes, No, Ambiguous, or Annuled. Range questions can resolve to a specific value, an out-of-bounds value, Ambiguous, or Annuled.
Sometimes a question cannot be resolved because the state of the world, the truth of the matter
, is too uncertain. In these cases, the question is resolved as Ambiguous.
Other times, the state of the world is clear, but a key assumption of the question was overturned. In these cases, the question is Annulled.
In the same way, when a Conditional turns out to be based on an outcome that did not occur, it is Annulled. For example, when a Conditional Pair's parent resolves Yes, the if No
Conditional is Annulled.
When questions are Annulled or resolved as Ambiguous, they are no longer open for forecasting, and they are not scored.
If you'd like to read more about why Ambiguous and Annulled resolutions are necessary you can expand the section below.
Reasons for Ambiguous and Annulled resolutions
An Ambiguous or Annulled resolution generally implies that there was some inherent ambiguity in the question, that real-world events subverted one of the assumptions of the question, or that there is not a clear consensus as to what in fact occurred. Metaculus strives for satisfying resolutions to all questions, and we know that Ambiguous and Annulled resolutions are disappointing and unsatisfying. However, when resolving questions we have to consider factors such as fairness to all participating forecasters and the underlying incentives toward accurate forecasting.
To avoid this unfairness and provide the most accurate information, we resolve all questions in accordance with the actual written text of the resolution criteria whenever possible. By adhering as closely as possible to a reasonable interpretation of what's written in the resolution criteria, we minimize the potential for forecasters to arrive at different interpretations of what the question is asking, which leads to fairer scoring and better forecasts. In cases where the outcome of a question does not clearly correspond to the direction or assumptions of the text of the resolution criteria, Ambiguous resolution or Annulling the question allows us to preserve fairness in scoring.
A question's resolution criteria can be thought of as akin to a legal contract. The resolution criteria create a shared understanding of what forecasters are aiming to predict, and define the method by which they agree to be scored for accuracy when choosing to participate. When two forecasters who have diligently read the resolution criteria of a question come away with significantly different perceptions about the meaning of that question, it creates unfairness for at least one of these forecasters. If both perceptions are reasonable interpretations of the text, then one of these forecasters will likely receive a poor score at resolution time through no fault of their own. Additionally, the information provided by the forecasts on the question will be poor due to the differing interpretations.
The following sections provide more detail about common reasons we resolve questions as Ambiguous or Annul them and some examples. Some of these examples could fit into multiple categories, but we've listed them each in one main category as illustrative examples. This list of types of Ambiguous or Annulled resolutions is not exhaustive — there are other reasons that a question may resolve Ambiguous or be Annulled — but these cover some of the more common and some of the trickier scenarios. Here's a condensed version, but read on for more details:
Note: Previously Metaculus only had one resolution type — Ambiguous — for cases where a question could not otherwise be resolved. We've since separated these into two types — Ambiguous and Annulled — to provide clarity on the reason that a question could not otherwise be resolved. Annulling questions first became an option in April of 2023.
Ambiguous resolution is reserved for questions where reality is not clear. Either because reporting about an event is conflicted or unclear about what actually happened, or available material is silent on the information being sought. We've described the types of questions where Ambiguous resolution is appropriate as those with No Clear Consensus.
Questions can also resolve Ambiguous when there is not enough information available to arrive at an appropriate resolution. This can be because of conflicting or unclear media reports, or because a data source that was expected to provide resolution information is no longer available. The following are some examples where there was no clear consensus.
Annulling a question is reserved for situations where reality is clear but the question is not. In other words, the question failed to adequately capture a method for clear resolution.
Note: Annulment was introduced in April of 2023, so while the following examples describe Annulment the questions in actuality were resolved as Ambiguous.
Writing good forecasting questions is hard, and it only gets harder the farther the question looks into the future. To fully eliminate the potential for a question to be Annulled the resolution criteria must anticipate all the possible outcomes that could occur in the future; in other words, there must be clear direction for how the question resolves in every possible outcome. Most questions, even very well-crafted ones, can't consider every possible outcome. When an outcome occurs that does not correspond to the instructions provided in the resolution criteria of the question then that question may have to be Annulled. In some cases we may be able to find an interpretation that is clearly an appropriate fit for the resolution criteria, but this is not always possible.
Here are some examples of Annulment due to underspecified questions:
Questions often contain assumptions in their resolution criteria, many of which are unstated. For example, assuming that the underlying methodology of a data source will remain the same, assuming that an organization will provide information about an event, or assuming that an event would play out a certain way. The best practice is to specify what happens in the event certain assumptions are violated (including by specifying that the question will be Annulled in certain situations) but due to the difficulty in anticipating these outcomes this isn't always done.
Here are some examples of Annulment due to subverted assumptions:
Sometimes questions imply imbalanced outcomes, for example where the burden of proof for an event to be considered to have occurred is high and tips the scales toward a binary question resolving No, or where the question would require a substantial amount of research to surface information showing that an event occurred, which also favors a resolution of No. In certain circumstances these kinds of questions are okay, so long as there is a clear mechanism for the question to resolve as Yes and to resolve as No. However, sometimes questions are formulated such that there's no clear mechanism for a question to resolve as No, leading to the only realistic outcomes being a resolution of Yes or Annulled. This creates a bias in the question and also produces bad incentives if the question isn't Annulled.
The case of imbalanced outcomes and consistent incentives is best explained with examples, such as the following:
Currently, all questions will be resolved.
Questions will be resolved when they have satisfied the criteria specified in the resolution section of the question (or conversely, when those criteria have conclusively failed to be met). Each question also has a "Resolution Date" listed in our system for purposes such as question sorting; however, this listed date is often nothing more than an approximation, and the actual date of resolution may not be known in advance.
For questions which ask when something will happen (such as When will the first humans land successfully on Mars?
), forecasters are asked to predict the date/time when the criteria have been satisfied (though the question may be decided and points awarded at some later time, when the evidence is conclusive). Some questions predict general time intervals, such as "In which month will unemployment pass below 4%?"; when such a question has specified the date/time which will be used, those terms will be used. If these terms are not given, the default policy will be to resolve as the midpoint of that period (for example, if the January report is the first month of unemployment under 4%, the resolution date will default to January 15).
No, only the resolution criteria is relevant for resolving a question, the background section is intended only to provide potentially useful information and context for forecasters. In a well-specified question, the resolution criteria should stand on its own as a set of self-contained instructions for resolving the question. In rare cases or older questions on Metaculus, the background material may be necessary to inform resolution, but the information in the resolution criteria supersedes conflicting information in the background material.
Still, we want the background material to be as helpful as possible and accurately capture the context and relevant information available at the time the question was written, so if you see errors or misleading information in the background of a question please let Admins know by tagging @admins in a comment!
We take care to launch questions that are as clearly specified as possible. Still, writing clear and objectively resolvable questions is challenging, and in some cases a question's resolution criteria may unintentionally allow for multiple different interpretations or may not accurately represent the question being asked. In deciding how to approach questions that have been launched with these deficiencies, Admins primarily consider fairness to forecasters. Issuing clarifications or edits for open questions can harm some forecaster's scores when the clarification significantly changes the meaning of the question. Based on an assessment of fairness and other factors, Admins may issue a clarification to an open question to better specify the meaning. This is typically most appropriate when a question has not been open for long (and therefore forecasts can be updated with negligible impact to scores), when a question contains inconsistent or conflicting criteria, or when the clarification adds specificity where there previously was none in a way that avoids substantial changes to the meaning.
In many cases such questions must be resolved as ambiguous or annulled to preserve fairness in scoring. If you believe there are ambiguities or conflicts in the resolution criteria for a question please let Admins know by tagging @admins in a comment. We hope inconsistencies can be identified as early as possible in the lifetime of a question so that they can be addressed. Claims of unclear resolution criteria made for questions which have already closed or claims of incorrect resolution for questions which have already resolved will be held to a higher standard of evidence if the issue(s) with the resolution criteria was not previously mentioned while the question was open to forecasting.
Yes, at times questions are resolved and it is later discovered these resolutions were in error given the information available at the time. Some questions may even specify that they will be resolved according to initial reporting or results, but specify re-resolution in the event that the final results disagree with the initial results (for example, questions about elections might use this mechanism to allow prompt feedback for forecasters but arrive at the correct answer in the rare event that the initial election call was wrong). Questions may be re-resolved in such cases if Metaculus determines that re-resolution is appropriate.
When resolving a question, the Moderator has an option to change the effective closing time of a question, so that if the question is unambiguously resolved prior to the closing time, the closing time can be changed to a time prior to which the resolution is uncertain.
When a question closes early, the points awarded are only those accumulated up until the (new) closing time. This is necessary in order to keep scoring "proper" (i.e. maximally reward predicting the right probability) and prevent gaming of points, but it does mean that the overall points (positive or negative) may end up being less than expected.
In some cases when the timing of an event is unknown it may be appropriate to change the closing date to a time before the question resolved, after the resolution is known. This is known as retroactive closure. Retroactive closure is not allowed except in the case of an event where the timing of the event is unknown and the outcome of the event is independent of the timing of the event, as described in the question closing guidelines above. When the timing of the event impacts the outcome of the event retroactive closure would violate proper scoring. For scoring to be proper a question must only close retroactively when the outcome is independent of the timing of the event. Here are several examples:
Forecasters often like retroactive closure because it prevents points from being truncated when an event occurs before the originally scheduled close date. But in order to elicit the best forecasts it’s important to follow proper scoring rules. For more on point truncation this section of the FAQ.
While Metaculus will try not to approve questions which specify inappropriate retroactive closure, sometimes new or existing questions do specify it. It is the policy of Metaculus to ignore inappropriate retroactive closure when resolving questions.
Our Moderators and question authors strive to be as clear and informed as possible on each question, but mistakes occasionally happen, and will be decided by our Admins' best judgement. For a hypothetical question like Will a nuclear detonation occur in a Japanese City by 2030?
it can be understood by common sense that we are asking about the next detonation after the detonations in 1945. In other questions like Will Facebook implement a feature to explain news feed recommendations before 2026?
, we are asking about the first occurrence of this event. Since this event occurred before the question opened and this was not known to the question author, the question resolved ambiguously.
There are times when the intent of a question is to specifically track the actions or statements of specific organizations or people (such as, "how many Electoral Votes will the Democrat win in the 2020 US Presidential Election according to the Electoral College"); at other times, we are interested only in the actual truth, and we accept a resolution source as being an acceptable approximation (such as, "how many COVID-19 deaths will there be in the US in 2021 according to the CDC?"). That said, in many cases it is not clear which is intended.
Ideally, every question would be written with maximally clear language, but some ambiguities are inevitable. Unless specifically indicated otherwise, if a resolution source is judged by Metaculus Admins to be defunct, obsolete, or inadequate, Admins will make a best effort to replace it with a functional equivalent. Questions can over-rule this policy with language such as "If [this source] is no longer available, the question will resolve Ambiguously" or "This question tracks publications by [this source], regardless of publications by other sources."
Metaculus uses Resolution Councils to reduce the likelihood of ambiguous resolutions for important questions—those that we feel have the potential to be in the top 1% of all questions on the platform in terms of impact.
A Resolution Council is an individual or group that is assigned to resolve a question. Resolution Council questions resolve at the authority of the individual or individuals identified in the resolution criteria. These individuals will identify the resolution that best aligns with the question and its resolution criteria.
If a Resolution Council member is not available to resolve a question, Metaculus may choose a suitable replacement.
We are working on recreating the tutorial.
You make a prediction simply by sliding the slider on the question's page to the probability you believe most captures the likelihood that the event will occur.
You can revise your prediction at any time up until the question closes, and you are encouraged to do so: as new information comes to light, it is beneficial to take it into account.
You're also encouraged to predict early, however, and you are awarded bonus points for being among the earliest predictors.
Some Metaculus questions allow numeric or date range inputs, where you specify the distribution of probability you think is likely over a possible range of outcomes. This probability distribution is known as a probability density function and is the probability per unit of length. The probability density function can be used to determine the probability of a value falling within a range of values.
When you hover over the chart you see the probabilities at each point at the bottom of the chart. For example, in the image below you can see the probability density at the value 136, denoted by "P(x = 136)", and you can see the probability density that you and the community have assigned to that point (in the image the user has assigned a probability density of 1.40 to that value and the community has assigned a probability density of 2.97).
By selecting the "Probability Density" dropdown at the top of the chart you can change the display to "Cumulative Probability". This display shows the cumulative distribution function, or in other words for any point it shows you the probability that you and the community have assigned to the question resolving below the indicated value. For example, in the image below you can see the probability that you and the community have assigned to the question resolving below the value of 136, denoted by "P(x < 136)". The probability that the user has assigned is 7% to the question resolving below that value, while the community has assigned an 83% chance to the question resolving below that value.
The vertical lines shown on the graphs indicate the 25th percentile, median, and 75th percentile forecasts, respectively, of the user and the community. These values are also shown for the user and the community in the table at the bottom.
In the table showing the predictions at the bottom of the images above, you will see that in addition to the 25th percentile, median, and 75th percentile probabilities there is also one labeled "> 500". This question has an open upper bound, which means forecasters can assign a probability that the question will resolve as a value above the upper end of the specified range. For the question depicted above the community and the forecaster each assign a 1% probability to the question resolving above the upper boundary.
Questions can have open or closed boundaries on either end of the specified range.
A closed boundary means forecasters are restricted from assigning a probability beyond the specified range. Closed boundaries are appropriate when a question cannot resolve outside the range. For example, a question asking what share of the vote a candidate will get with a range from 0 to 100 should have closed boundaries because it is not possible for the question to resolve outside the range. Closed boundaries restrict forecasters from assigning probabilities outside the specified range.
An open boundary allows a question to resolve outside the range. For example, a question asking what share of the vote a candidate will get with a range from 30 to 70 should have open boundaries because it is possible that the candidate could get less than 30% of the vote or more than 70%. Open boundaries should be specified even if it unlikely that the vote share falls outside the range, because it is theoretically possible that vote shares outside the specified range can occur.
Forecasters can assign probabilities outside the range when the boundary is open by moving the slider all the way to one side. The weight can also be lowered or increased to adjust the probability assigned to an out of bounds resolution.
In the images shown above you can see that the user has assigned two probability distributions. Up to five logistic distributions can be added using the "Add Component" button. The relative weight of each can be adjusted using the "weight" slider below each component.
The Community Prediction is a consensus of recent forecaster predictions. It's designed to respond to big changes in forecaster opinion while still being fairly insensitive to outliers.
Here's the mathematical detail:
Users can hide the Community Prediction from view from within their settings.
By default, bots are not included in any aggregations. If they are, it is indicated in the sidebar as "Include Bots".
The Metaculus Prediction can only be viewed in the Aggregation Explorer. It is deprecated since November 2024, but shows a record of the Metaculus system's best estimate of how a question will resolve. It's based on predictions from community members, but unlike the Community Prediction, it's not a simple average or median. Instead, the Metaculus Prediction uses a sophisticated model to calibrate and weight each user, ideally resulting in a prediction that's better than the best of the community.
For questions that resolved in 2021, the Metaculus Prediction has a Brier score of 0.107. Lower Brier scores indicate greater accuracy, with the MP slightly lower than the Community Prediction's Brier score of 0.108.
When a question first opens, nobody can see the Community Prediction for a while, to avoid giving inordinate weight to the very first predictions, which may "ground" or bias later ones.
Public Figure Prediction pages are dedicated to collecting and preserving important predictions made by prominent public figures and putting them into conversation with Metaculus community forecasts. Each figure's page features a list of predictions they made along with the source that recorded the prediction, the date the prediction was made, and related Metaculus questions. Public predictions are transparently presented alongside community forecasts in a manner that is inspectable and understandable by all, providing public accountability and additional context for the linked Metaculus questions.
A Public Figure is someone with a certain social position within a particular sphere of influence, such as a politician, media personality, scientist, journalist, economist, academic, or business leader.
A prediction is a claim or a statement about what someone thinks will happen in the future, where the thing predicted has some amount of uncertainty associated with it.
A Public Figure Prediction is a prediction made by the public figure themselves and not by figures who might represent them, such as employees, campaign managers, or spokespeople.
When predictions are made by public figures such as elected politicians, public health officials, economists, journalists, and business leaders, they become candidates for inclusion in the Public Figure Prediction system.
From a Public Figure's page, click Report Prediction and then provide
If the Public Figure does not yet have a dedicated page, you can request that one be created by commenting on the Public Figures Predictions discussion post. Tag @christian for a faster moderation process.
Depending on the level of specificity and clarity of the Public Figure Prediction, a linked Metaculus question might resolve according to the exact same criteria as the prediction. For example, Joe Biden expressed that he plans to run for reelection. This Metaculus question asks directly whether he will run.
Linked questions are not required, however, to directly correspond to the public figure's prediction, and this question on whether Biden will be the Democratic nominee in 2024 is clearly relevant to public figure claim, even as it's further away from the claim than asking whether Biden will run. Relevant linked questions shed light on, create additional context for, or provide potential evidence for or against the public figure's claim. Note that a question being closed or resolved does not disqualify it from being linked to the prediction.
On the other hand, this question about whether the IRS designates crypto miners as ‘brokers' by 2025 follows from Biden's Infrastructure Investment and Jobs Act, but beyond the Biden connection, it fails to satisfy the above criteria for a relevant linked question.
News sources that have authority and are known to be accurate are acceptable. If a number of news sources report the same prediction, but the prediction originated from a single source, using the original source is preferred. Twitter accounts or personal blogs are acceptable if they are owned by the public figure themselves.
Moderators will review and approve your request or provide feedback.
On the page of the prediction, comment the update with the source and tag a moderator. The moderator will review and perform the update if necessary.
Please email us at support at metaculus.com.
Sometimes you haven't changed your mind on a question, but you still want to record your current forecast. This is called "reaffirming": predicting the same value you predicted before, now.
It is also useful when sorting questions by the age of your latest forecast. Reaffirming a question sends it to the bottom of that list.
You can reaffirm a question from the normal forecast interface on the question page, or using a special button in feeds.
On question groups, reaffirming impacts all subquestions on which you had a forecast, but not the others.
Scores measure forecasting performance over many predictions. Metaculus uses both Baseline scores, which compare you to an impartial baseline, and Peer scores, which compare you to all other forecasters. We also still use Relative scores for tournaments. We do not use the now-obsolete Metaculus points, though they are still computed and you can find them on the question page.
Learn more in the dedicated Scores FAQ.
Medals reward Metaculus users for excellence in forecasting accuracy, insightful comment writing, and engaging question writing. We give medals for placing well in any of the 4 leaderboards: Baseline Accuracy, Peer Accuracy, Comments, and Question Writing. Medals are awarded every year. Medals are also awarded for Tournament performance.
Learn more in the dedicated Medals FAQ.
The Metaculus Journal publishes longform, educational essays on critical topics like emerging science and technology, global health, biosecurity, economics and econometrics, environmental science, and geopolitics—all fortified with testable, quantified forecasts.
If you would like to write for the Metaculus Journal, email [email protected] with a resume or CV, a writing sample, and two story pitches.
In November 2021, Metaculus introduced a new project—Fortified Essays. A Fortified Essay is an essay that is “fortified” by its inclusion of quantified forecasts which are justified in the essay. The goal of Fortified Essays is to leverage and demonstrate the knowledge and intellectual labor that went into answering forecasting questions while also putting the forecasts into a larger context.
Metaculus plans to run Fortified Essay Contests regularly as part of some tournaments. This additional context deriving from essays is necessary, because a quantified forecast in isolation may not provide the information required to drive decision-making by stakeholders. In Fortified Essays, writers can explain the reasoning behind their predictions, discuss the factors driving the predicted outcomes, explore the implications of these outcomes, and offer their own recommendations. By placing forecasts into this larger context, these essays are better able to help stakeholders deeply understand the relevant forecasts and how much weight to place on them. The best essays will be shared with a vibrant and global effective altruism community of thousands of individuals and dozens of organizations.
For certain projects, Metaculus employs Pro Forecasters who have demonstrated excellent forecasting ability and who have a history of clearly describing their rationales. Pros forecast on private and public sets of questions to produce well-calibrated forecasts and descriptive rationales for our partners. We primarily recruit members of the Metaculus community with the best track records for our Pro team, but forecasters who have demonstrated excellent forecasting ability elsewhere may be considered as well.
If you’re interested in hiring Metaculus Pro Forecasters for a project, contact us at [email protected] with the subject "Project Inquiry".
Metaculus selects individuals according to the following criteria:
The API documentation is still being worked on.
You can change your name for free within the first three days of registering. After that, you'll be able to change it once every 180 days.
In an effort to reduce spam, new users must wait 12 hours after signup before commenting is unlocked.
Metaculus may—though this thankfully occurs very rarely—issue the temporary suspensions of an account. This occurs when a user has acted in a way that we consider inappropriate, such as when our terms of use are violated. At this point, the user will receive a notice about the suspension and be made aware that continuing this behavior is unacceptable. Temporary suspensions serve as a warning to users that they are a few infractions away from receiving a permanent ban on their account.
When a question first opens, nobody can see the Community Prediction for a while, to avoid giving inordinate weight to the very first predictions, which may "ground" or bias later ones. Once the Community Prediction is visible, the Metaculus Prediction is hidden until the question closes.
NewsMatch displays a selection of articles relevant to the current Metaculus question. These serve as an additional resource for forecasters as they discuss and predict on the question. Each article is listed with its source and its publication date. Clicking an article title navigates to the article itself. Up and downvoting allows you to indicate whether the article was helpful or not. Your input improves the accuracy and the usefulness of the model that matches articles to Metaculus questions.
The article matching model is supported by Improve the News, a news aggregator developed by a group of researchers at MIT. Designed to give readers more control over their news consumption, Improve the News helps readers stay informed while encountering a wider variety of viewpoints.
Articles in ITN's database are matched with relevant Metaculus questions by a transformer-based machine learning model trained to map semantically similar passages to regions in "embedding space." The embeddings themselves are generated using MPNet.
Community Insights summarize Metaculus user comments on a given question using GPT-4. They condense recent predictions, timestamped comments, and the current community prediction into concise summaries of relevant arguments for different forecasts on a given question. Forecasters can use them to make more informed decisions and stay up-to-date with the latest insights from the community.
Community Insights are currently available on binary and continuous questions with large comment threads and will update regularly as new discussion emerges in the comments. If you have feedback on these summaries—or would like to see them appear on a wider variety of questions—email [email protected].
If you find a Community Insights summary to be incorrect, offensive, or misleading please use the button at the bottom of the summary to "Flag this summary" so the Metaculus team can address it.
Yes! Metaculus has a domain system, where each domain (like "example.metaculus.com") has a subset of questions and users that are assigned to it. Each question has a set of domains it is posted on, and each user has a set of domains they are a member of. Thus, a domain is a flexible way of setting a particular set of questions that are private to a set of users, while allowing some questions in the domain to be posted also to metaculus.com. Domains are a product that Metaculus can provide with various levels of support for a fee; please be in touch for more details.
Metaculus will get more fun and more interesting to the extent that it grows to include more and more predictors, so we encourage participants to spread the word to people who they think may enjoy predicting, or just be interested in how the questions develop. Some of the most useful mechanisms are:
Of course, if you wish to close your account, please email your request to[email protected]. Within five business days, we will remove your profile information and comments from our active database.