Fact-checked by Grok 2 weeks ago

Superforecaster

A superforecaster is an individual who demonstrates exceptional and sustained accuracy in of future events, particularly in domains such as , , and , outperforming both laypersons and domain experts through empirical validation in large-scale prediction tournaments. The concept emerged from research led by political scientist , who coined the term after analyzing participants in (GJP), a competition sponsored by the U.S. (IARPA) from 2011 to 2015, where superforecasters—selected from thousands of volunteers—achieved accuracy levels 60% superior to control groups in the first year and 78% in the second, surpassing U.S. intelligence analysts and other research teams. Unlike intuitive or expert predictions often marred by overconfidence and ideological bias, superforecasters distinguish themselves through deliberate practices rooted in Bayesian updating and , including breaking down questions into subcomponents, actively seeking disconfirming evidence, assigning numerical probabilities rather than binary outcomes, and iteratively refining estimates as new information emerges. They tend to exhibit traits such as , , and a non-deterministic , enabling them to navigate uncertainty without anchoring to initial beliefs or . This skill, Tetlock's studies show, is cultivable rather than innate, with superforecasters maintaining top performance across years—about 70% retaining elite status from one tournament season to the next—and applying it in real-world contexts via organizations like Good Judgment Inc., which deploys teams of around 180 such forecasters for client forecasting needs. The identification of superforecasters challenges conventional reliance on credentialed experts, whose forecasts Tetlock's earlier Expert Political Judgment project found no better than chance in many cases, highlighting instead the value of track-record-based evaluation and crowd-aggregation techniques refined by superforecasters, such as ensemble forecasting where aggregated individual predictions yield even higher accuracy. While scalable applications include policy advisory and , the approach underscores empirical limits: superforecasting excels in resolvable, trackable questions but falters in truly black-swan events dominated by unknown unknowns, emphasizing probabilistic humility over false precision.

History

Origins and Etymology

The term "superforecaster" combines the prefix super-, denoting exceptional quality or performance, with "forecaster," a person who predicts future events or trends. It specifically denotes individuals who demonstrate markedly superior accuracy in compared to average participants or experts in tasks. The concept originated in the Good Judgment Project (GJP), a research initiative launched in 2011 by psychologist Philip Tetlock and decision scientist Barbara Mellers under the auspices of the U.S. (IARPA). The GJP conducted large-scale forecasting tournaments involving thousands of volunteers assessing geopolitical and economic questions, where a small subset of participants—about 2%—consistently outperformed intelligence analysts and other benchmarks by updating predictions dynamically and achieving Brier scores up to 60% better than control groups. Tetlock coined "superforecaster" to describe these top performers, whose skills were identified through empirical analysis rather than prior expertise, challenging assumptions that alone suffices for accurate prediction. The term gained prominence with the 2015 publication of Tetlock's book Superforecasting: The Art and Science of Prediction, co-authored with Dan Gardner, which detailed GJP findings and argued that superforecasting stems from cultivable habits like Bayesian updating and active rather than innate genius or esoteric knowledge. This work built on Tetlock's earlier research, including his 2005 book Expert Political Judgment, which had revealed widespread forecasting failures among pundits and specialists, setting the stage for identifying outliers like superforecasters.

Early Research and Tetlock's Contributions

Philip Tetlock initiated systematic research on forecasting accuracy in the 1980s through small-scale tournaments that tracked predictions from experts in , , and related fields. These efforts, spanning from 1984 to 2003, involved 284 participants generating thousands of forecasts on geopolitical and economic outcomes, which Tetlock evaluated against objective resolutions to assess and . The analysis revealed that experts struggled with long-term predictions, often performing no better than chance or simple baselines like random selection. This groundwork culminated in Tetlock's 2005 book Expert Political Judgment: How Good Is It? How Can We Know?, which synthesized data from approximately 27,000 predictions made by those 284 experts over two decades. The study distinguished between "hedgehogs"—ideologically driven specialists who exhibited overconfidence and poor adaptability—and "foxes," more eclectic thinkers who marginally outperformed baselines through probabilistic reasoning and belief updating. Overall, however, even foxes' accuracy remained limited, underscoring systemic flaws in expert forecasting reliant on narrative over empirical tracking. Tetlock argued that judgment quality hinges on both correspondence to (accurate ) and (logical consistency), with most experts failing the former. Motivated by these shortcomings, Tetlock shifted toward experimental interventions to cultivate superior forecasters. In 2011, he co-led (GJP) with Barbara Mellers as part of the U.S. (IARPA)'s Aggregative Contingent Estimation (ACE) program, a four-year competition launched to test crowd-based forecasting methods against intelligence analysts. GJP recruited over 20,000 volunteers to provide probabilistic estimates on 500 real-world questions, incorporating techniques like base-rate awareness, team deliberation, and frequent updates, which yielded accuracy gains of 30% over control groups and control forecasters. Tetlock's pivotal contribution emerged from GJP's identification of "superforecasters"—the top 2% of participants whose aggregate performance surpassed experts, crowds, and even GJP's broader pool by wide margins, achieving Brier scores indicative of high calibration and resolution. These individuals exemplified traits like openness to evidence and iterative refinement, challenging prior pessimism about human prediction limits. Tetlock formalized this in his 2015 book Superforecasting: The Art and Science of Prediction, co-authored with Dan Gardner, which disseminated GJP methodologies and traits for broader application, establishing superforecasting as a trainable skill rooted in empirical validation rather than intuition.

Key Projects and Milestones

The (GJP), co-led by Philip Tetlock and Barbara Mellers, was launched in as a participant in the U.S. (IARPA) Aggregative Contingent Estimation (ACE) forecasting tournament, which ran from to 2015 and aimed to improve geopolitical prediction accuracy through crowd-sourced methods. GJP recruited over 20,000 volunteer forecasters to answer approximately 100-150 annual questions on international events, employing strategies such as talent identification, probabilistic training, team collaboration, and forecast aggregation. In the ACE program's first year (2011-2012), GJP's top-performing forecasters outperformed the control group of analysts by 60% in accuracy, rising to 78% in the second year, and ultimately surpassing all competing teams by 35-72% across the . This success highlighted the efficacy of structured processes, with GJP's "superforecasters"—defined as the consistent top 2% of participants—demonstrating sustained outperformance through habits like frequent updating of beliefs and belief calibration. The project's findings culminated in the 2015 publication of by Tetlock and Dan Gardner on September 29, which formalized the superforecaster concept based on empirical data from GJP, arguing that superior forecasting stems from learnable skills rather than innate expertise. Following the ACE program's conclusion in 2015, Tetlock and Mellers established Good Judgment Inc. to commercialize these methods, offering forecasting services to government, nonprofit, and private sector clients, including ongoing platforms like Good Judgment Open for public participation.

Scientific Foundations

Forecasting Tournaments and Methodology

Forecasting tournaments consist of competitive exercises in which participants assign numerical probabilities to the outcomes of predefined real-world events, such as geopolitical developments or changes, with accuracy assessed via proper scoring rules that reward well-calibrated and resolved predictions. These tournaments test whether forecasting skill exists and can be cultivated beyond chance or expert opinion, using verifiable resolutions to score performance over time. The foundational tournament series was the U.S. (IARPA)'s Aggregative Contingent Estimation (ACE) program, conducted from 2011 to 2015, which sought to boost the precision of intelligence forecasts by rigorously evaluating methods for eliciting, weighting, and aggregating probabilistic judgments. Each annual season involved 100 to 150 questions drawn from and international affairs domains, with clear, objective resolution criteria tied to public or official sources to minimize ambiguity. Participating teams, including university-led groups, competed to refine techniques, generating over one million individual predictions across approximately 500 questions in total. The Good Judgment Project (GJP), directed by Philip Tetlock and Barbara Mellers, achieved the highest accuracy in the tournaments through a multi-stage : broad of volunteer forecasters via online platforms, initial solo on tournament questions, brief training in cognitive debiasing to counter common biases like overconfidence, and iterative updates to predictions as new evidence surfaced. Top performers—defined as the uppermost 5% by cumulative accuracy—were annually selected and grouped into elite teams of about 12 members each, fostering collaborative deliberation while preserving individual accountability to incentivize diverse inputs and error correction. Performance was quantified using the , the between a forecaster's probability estimates (expressed as decimals from 0 to 1) and actual outcomes (coded as 0 or 1), penalizing both miscalibration and poor resolution; scores range from 0 for perfect accuracy to up to 2 for consistent error, with 0.5 typical for uninformative guesses on even-odds events. GJP's superforecaster teams recorded Brier scores approximately 30% to 60% superior to those of intelligence analysts, crowdsourced baselines, and rival research teams across seasons 2 through 4, demonstrating sustained skill rather than luck through metrics like high area under the curve () values near 0.96 and rapid learning from feedback. This approach prioritized empirical validation over untested expertise, revealing that structured probabilistic methods and outperform solitary .

Probabilistic Reasoning and Bayesian Principles

Superforecasters distinguish themselves through the consistent application of probabilistic reasoning, expressing predictions as numerical probabilities rather than categorical assertions such as "yes" or "no." This method acknowledges inherent uncertainty in complex events, allowing for nuanced assessments like a 65% chance of occurrence, which can be rigorously tested against outcomes using proper scoring rules, including the that penalizes both overconfidence and underprecision. In the , superforecasters achieved superior accuracy by routinely assigning such probabilities to geopolitical and economic questions, outperforming intelligence analysts who often defaulted to deterministic language. Central to their approach is adherence to Bayesian principles, which involve initializing forecasts with base rates—empirical priors drawn from analogous historical cases—and iteratively updating them via new evidence, mirroring Bayes' theorem's formula for posterior probability: P(H|E) = \frac{P(E|H) \cdot P(H)}{P(E)}, where priors P(H) are adjusted by the likelihood P(E|H) of observed evidence E. Tetlock's analysis revealed that superforecasters, unlike novices, actively sought disconfirming evidence and revised estimates promptly, with updates occurring multiple times per forecast as information accrued, fostering a dynamic belief system resistant to anchoring bias. This Bayesian updating was evident in tournament performance, where superforecasters' forecasts improved by an average of 10-15% in logarithmic scoring terms after incorporating fresh data, compared to static predictions from control groups. To operationalize these principles, superforecasters decompose questions into subcomponents, estimate conditional probabilities for each, and aggregate them, often employing Fermi-style approximations for tractability. They prioritize , selecting relevant base rates while avoiding the "reference class problem" through sensitivity testing across alternatives, which enhances —defined as the alignment between stated probabilities and realized frequencies, with superforecasters achieving 70-80% hits on their high-confidence predictions in validated studies. Such practices underscore a commitment to causal realism, weighing multiple hypotheses without premature commitment, thereby mitigating overreliance on narrative coherence at the expense of probabilistic rigor.

Characteristics of Superforecasters

Psychological and Cognitive Traits

Superforecasters exhibit distinct psychological traits characterized by , reflectiveness, and a willingness to revise beliefs in light of new evidence. They demonstrate high levels of active , treating predictions as testable hypotheses rather than fixed convictions, which enables them to update forecasts dynamically without attachment to initial views. This temperament contrasts with overconfidence common in average forecasters, as superforecasters maintain cautious self-assessments, acknowledging uncertainty and avoiding deterministic thinking. Cognitively, superforecasters possess above-average , particularly in reflective and analytical domains, allowing for systematic over rapid . They excel in and probabilistic reasoning, framing judgments in terms of likelihoods rather than absolutes, which facilitates Bayesian updating of probabilities based on incoming data. Studies from forecasting tournaments indicate that these individuals perform better due to cognitive styles that mitigate biases, such as , through deliberate evidence-seeking and self-critique. Empirical analyses of top performers in the reveal correlations with fluid intelligence measures and task-specific skills like intuitive statistical reasoning, though raw IQ alone does not predict success—deliberative habits amplify cognitive strengths. Unlike domain experts who may suffer from specialized overreach, superforecasters apply generalizable cognitive tools across diverse topics, emphasizing precision in aggregating weak signals into coherent probabilities.

Habits and Practices

Superforecasters demonstrate distinctive habits and practices that enhance their predictive accuracy, derived from empirical analysis of top performers in the Good Judgment Project's forecasting tournaments, where participants addressed geopolitical and economic questions over periods exceeding 1,000 days. These individuals, comprising roughly the top 2% of forecasters, habitually apply probabilistic reasoning, regularly update predictions in response to new evidence, and maintain decision journals to track and refine their thought processes. Such practices correlate with scores— a metric combining and —superior to those of average participants by factors of 30-60% in controlled studies. Philip Tetlock, principal investigator of the Good Judgment Project, synthesized these behaviors into "Ten Commandments for Aspiring Superforecasters," emphasizing deliberate, evidence-based cognition over intuition or ideological commitment.
  1. Triage: Superforecasters prioritize questions in a "Goldilocks zone" of difficulty—neither trivially easy nor impossibly vague—where informed effort yields measurable improvements, avoiding overinvestment in low-value or intractable forecasts.
  2. Break seemingly intractable problems into tractable sub-problems: They decompose complex issues into components amenable to Fermi estimation or targeted research, isolating knowable elements from unknowns and testing underlying assumptions.
  3. Strike the right balance between inside and outside views: Forecasters integrate case-specific details (inside view) with statistical base rates from analogous historical events (outside view), habitually querying frequencies of similar outcomes to anchor probabilities.
  4. Strike the right balance between under- and overreacting to evidence: They update beliefs Bayesian-style, distinguishing signal from noise while guarding against or undue weight on recent data, often adjusting forecasts incrementally as evidence accumulates.
  5. Look for the clashing causal forces at work in each problem: Superforecasters map competing drivers and counterarguments, maintaining a mental of indicators that could falsify their views and synthesizing perspectives from apparent contradictions.
  6. Strive to distinguish as many degrees of doubt as the problem permits but not more: They quantify numerically—e.g., expressing 60% rather than yes/no—through repeated to refine without spurious .
  7. Strike the right balance between under- and overconfidence, between prudence and decisiveness: Balancing calibration (alignment of stated probabilities with outcomes) and (discrimination between likely and unlikely events), they avoid by analysis yet qualify assertions proportionally to evidence strength.
  8. Look for the errors behind your mistakes but beware of rearview-mirror hindsight biases: Post-mortems on resolved forecasts focus on cognitive pitfalls like overconfidence or base-rate neglect, using journals to differentiate skill from luck without retroactive rationalization.
  9. Bring out the best in others and let others bring out the best in you: In team settings, they employ , precise questioning, and constructive feedback to aggregate diverse inputs, leveraging while mitigating .
  10. Master the error-balancing bicycle: Through iterative practice with feedback loops—such as tournament scoring—superforecasters dynamically correct for tendencies toward over- or under-adjustment, akin to learning equilibrium via .
These practices, honed via deliberate repetition, have been validated in extensions of , where training participants in them yielded accuracy gains of 10-20% over untrained baselines in subsequent tournaments. No single habit suffices; proficiency emerges from their integrated application, underscoring superforecasting as a cultivable rather than innate alone.

Effectiveness and Validation

Performance Metrics from Studies

In studies from the (GJP), superforecasters—identified as the top 2% of individual performers in large-scale forecasting tournaments—were evaluated using the , which measures the mean squared difference between predicted probabilities and binary outcomes (with scores ranging from 0 for perfect accuracy to 2 for complete inaccuracy). Their standardized s averaged -0.34 across years 2 and 3 of the tournaments, outperforming top-team individuals at -0.14 and all other participants at 0.04, with (p < 0.01). This translated to 50-60% greater accuracy relative to top-team forecasters on quick-response tasks requiring minimal deliberation time. Superforecasters also excelled in decomposed Brier score components: resolution (ability to distinguish likely from unlikely events) at 0.40 versus 0.35 for top-team individuals and 0.32 for others; calibration (alignment of predicted probabilities with observed frequencies) at 0.01 deviation versus 0.03 and 0.04; and area under the curve (AUC) for discrimination at 96% versus 84% and 75%. Raw scores for superforecasters typically ranged from 0.20 to 0.25 across diverse geopolitical and economic questions, reflecting sustained performance without regression to the mean, unlike average forecasters. Approximately 70% of designated superforecasters retained their elite status from one tournament year to the next, with correlations between yearly performances indicating skill stability rather than luck. In aggregate, team-based predictions from superforecasters in the GJP's first IARPA-sponsored season were 35% to 72% more accurate (via reductions) than those from competing research teams, including methods reliant on aggregation or crowds. These results held across hundreds of questions resolved between 2011 and 2015, spanning topics like international conflicts, economic indicators, and shifts, with superforecasters showing the fastest learning rates (-0.26 in Brier improvement) through iterative feedback.

Comparisons to Experts, Crowds, and AI

Superforecasters have demonstrated superior accuracy compared to domain experts in controlled forecasting tournaments. In the (IARPA) Aggregate Project, teams from the , which included superforecasters, achieved a approximately 30% lower than that of U.S. intelligence analysts with access to , indicating markedly higher predictive accuracy. This outperformance persisted across geopolitical, economic, and scientific questions, challenging the conventional reliance on expert intuition, which Tetlock's earlier research found often approximates random chance for long-range forecasts. Relative to crowds, superforecasters consistently outperform unaggregated or baseline predictions, though hybrid or incentivized crowd mechanisms like prediction markets can narrow the gap. In , superforecaster teams surpassed large crowd baselines by 15-30% in accuracy, even when crowds were drawn from diverse, informed participants. For instance, in a 2023 contest involving over 8,500 readers as a crowd proxy, Good Judgment superforecasters achieved higher accuracy on key global events. However, studies indicate that while individual superforecasters excel, aggregated superforecaster teams can rival or exceed prediction markets, which themselves often beat naive crowds by incorporating probabilistic updates and incentives. Comparisons to artificial intelligence reveal superforecasters maintaining an edge as of 2024-2025, though AI capabilities are rapidly advancing in specific domains. On the ForecastBench benchmark, superforecasters recorded a Brier score of 0.093, outperforming the best large language models at 0.111, a 19% relative advantage across diverse forecasting tasks. Independent evaluations confirm superforecasters' lower Brier scores (e.g., 0.0225) against leading AI models like o1 on real-world questions, attributing human superiority to adaptive reasoning and Bayesian updating habits. Nonetheless, state-of-the-art AI has surpassed superforecasters and human AI experts in select 2025 benchmarks focused on technical progress, such as AI capability timelines, suggesting domain-specific vulnerabilities where superforecasters exhibit undue conservatism. Emerging hybrid human-AI systems show promise for further gains, but pure AI forecasting remains below superforecaster levels in generalizability.

Notable Figures and Organizations

Prominent Individuals

Jean-Pierre Beugoms, a military historian, joined the in 2011 and emerged as one of its first superforecasters, demonstrating sustained high accuracy in forecasting, especially elections. His performance earned recognition in Adam Grant's 2021 book Think Again, highlighting superforecasters' ability to update beliefs amid uncertainty. Kjirste Morrell, a mechanical engineer holding a PhD from , ranks consistently among top superforecasters, with expertise in geopolitical predictions. She was also profiled in Think Again for exemplifying traits like probabilistic thinking and evidence-based revision. Robert de Neufville, a Harvard and graduate qualified as a superforecaster in 2014, focuses on existential risks and co-hosts a on forecasting. His work extends to authoring content that applies superforecasting principles to long-term uncertainties. Other profiled superforecasters include Elan Pavlov, a PhD with postdoc experience in , specializing in military and AI-related geopolitical forecasts; and Dan Mayland, a geopolitical and novelist whose predictions emphasize dynamics. These individuals, drawn from diverse fields like , , and , collectively outperform benchmarks by leveraging habits such as frequent updating and aggregation. (GJP) was a research initiative led by psychologists Philip Tetlock and Barbara Mellers at the , funded as part of the U.S. (IARPA) Aggregative Contingent Estimation (ACE) program from 2011 to 2015. The project recruited over 20,000 volunteer participants to generate probabilistic forecasts on approximately 500 geopolitical, economic, and scientific questions, resulting in more than one million individual predictions evaluated against real-world outcomes. GJP's methodology emphasized team-based , where top performers—later termed superforecasters—collaborated to refine predictions through iterative updates and aggregation techniques, outperforming other ACE teams by 35% to 72% in accuracy metrics such as scores. These superforecasters, comprising the top 2% of participants, also demonstrated over 30% greater accuracy than unaided U.S. intelligence analysts on comparable questions. Following the ACE program's conclusion, Tetlock and Mellers co-founded Good Judgment Inc. in 2015 as a private entity to commercialize superforecasting techniques, offering customized services to clients in government, nonprofit, and corporate sectors. The company maintains a vetted network of approximately 150 superforecasters drawn from GJP alumni, who provide calibrated probability estimates on strategic risks, policy outcomes, and market events, with historical performance showing sustained superiority over baseline benchmarks. Good Judgment Inc. has collaborated with entities such as the U.S. government and private firms for applications including geopolitical , achieving documented improvements in through probabilistic aggregation. Related to these efforts, Good Judgment Open operates as a public platform launched by the company, enabling broader participation in crowd-sourced forecasting challenges with questions updated weekly on topics like and trends. It serves as a ground for aspiring forecasters and a data source for refining superforecasting practices, with participant scores ranked against historical superforecaster benchmarks to encourage skill development via frequent, evidence-based updates.

Applications

In Government and Policy

Superforecasters have been integrated into government forecasting efforts primarily through structured tournaments and consulting services aimed at enhancing and policy anticipation. The , operational from 2011 to 2015 under the U.S. (IARPA)'s Aggregative Contingent Estimation program, recruited crowds of forecasters, including those exhibiting superforecasting traits, to predict geopolitical events relevant to . The initiative demonstrated that teams employing superforecaster practices—such as probabilistic reasoning, frequent updating, and aggregation of diverse judgments—outperformed intelligence analysts and other crowds by up to 30% in accuracy on questions like international conflicts and economic indicators. Post-tournament, Good Judgment Inc., founded by project leaders Philip Tetlock and Barbara Mellers, has extended these methods to government clients for real-time policy support. Policymakers and military leaders use superforecasting to anticipate global events, including shifts and security threats, by commissioning scenario-based predictions from certified superforecasters. For instance, in 2024, Good Judgment partnered with the Government's Futures Procurement Network to deliver foresight insights, aiding procurement and strategic planning amid uncertainties like disruptions. Similarly, the firm produced forecasts on U.S. funding in 2025, focusing on allocations, which informed evaluations by organizations influencing congressional policy. Applications extend to specialized policy domains, such as AI governance, where superforecasters generated probabilistic assessments in 2025 on regulatory trajectories and international agreements, drawing from public data to guide executive branch deliberations. The , a government unit, has advocated for superforecasting-inspired practices in evaluation, emphasizing iterative prediction markets to test interventions before implementation, as seen in trials for economic and public health forecasting. These efforts highlight superforecasting's role in reducing reliance on expert intuition, which studies show often underperforms calibrated crowds, though adoption remains limited by institutional and classification constraints in sensitive areas.

In Business and Strategic Decision-Making

Good Judgment Inc., co-founded by Philip Tetlock and Barbara Mellers following , extends superforecasting to clients for corporate and . The firm deploys a network of approximately 180 superforecasters—individuals selected for their top 1-2% accuracy in probabilistic predictions—to address business questions spanning economic, geopolitical, and market uncertainties. These forecasts support strategic decisions, such as resource allocation in firms, where predictions inform development priorities amid volatile prices and regulatory shifts. In practice, businesses leverage superforecasting through tools like FutureFirst, which provides ongoing access to aggregated superforecaster probabilities and qualitative analyses on client-specified scenarios, updated daily. This approach has demonstrated utility in investment contexts, where superforecasters outperformed futures markets in anticipating interest rate hikes, offering calibrated probabilities that reduced noise and bias in decision processes. For instance, in 2023, superforecasters beat a crowd of 8,500 readers in forecasting key economic events, highlighting advantages in team-based aggregation over individual or market judgments. Strategic applications emphasize four evidence-based elements: talent-spotting via evaluation, targeted to foster probabilistic thinking and belief updating, teaming to mitigate individual errors through diverse perspectives, and aggregation to refine collective outputs. Research underpinning these methods attributes roughly 30% greater accuracy to superforecaster teams compared to baseline analysts, even without classified data advantages. Companies apply this to challenge overconfident internal predictions, as outlined in frameworks for upgrading organizational judgment, enabling better for mergers, market entries, or disruptions. In 2024, private clients submitted 151 forecasting questions to the platform, yielding 1,132 predictions that informed real-time strategic adjustments.

Training and Development

Core Principles for Improvement

Superforecasters cultivate forecasting accuracy through deliberate habits and cognitive practices identified in research from (GJP), a from 2011 to 2015 where teams competed to predict geopolitical events. Participants who excelled, termed superforecasters, outperformed intelligence analysts by 30% in Brier scores, a metric measuring and of probabilistic predictions. These individuals were not experts in specific domains but shared traits like openness to evidence and iterative belief updating, honed via training modules on probabilistic reasoning and bias mitigation. Key principles for improvement, distilled as "Ten Commandments for Aspiring Superforecasters" by Philip Tetlock, emphasize breaking down complex questions, balancing perspectives, and rigorous self-critique. These derive from analyzing superforecasters' approaches in the GJP, where trained forecasters improved accuracy by 7-10% over untrained peers through practices like aggregating team inputs and extremizing probabilities toward extremes while preserving calibration. The principles include:
  • Triage: Prioritize questions where effort yields returns, avoiding overcommitment to low-stakes or unresolvable forecasts, as superforecasters focused on 20-30 high-value questions annually in GJP trials.
  • Decompose problems: Divide intractable issues into sub-problems, such as estimating base rates and causal drivers separately, which reduced error by enabling targeted analysis in GJP simulations.
  • Balance inside and outside views: Integrate specific case details (inside view) with statistical baselines (outside view), as overreliance on either led to 15-20% worse in data.
  • Balance short- and long-term perspectives: Weigh immediate trends against enduring patterns, preventing evident in non-superforecasters' 10-15% higher overconfidence.
  • Identify clashing forces: Map competing causal mechanisms, like economic incentives versus political , to model more granularly, a habit correlating with top GJP performance.
  • Quantify doubt precisely: Assign probabilities beyond binary outcomes (e.g., 65% vs. 70%), as vague estimates in GJP yielded poorer resolution scores by underdistinguishing shades of likelihood.
  • Avoid over- or underreaction: Update beliefs proportionally to evidence strength, with superforecasters adjusting forecasts by 5-10% on average per new datum versus larger swings by others.
  • Analyze errors without hindsight: Review mistakes for systematic flaws like , but reconstruct pre-outcome reasoning to avoid explanations, improving future accuracy by 8% in iterative GJP .
  • Collaborate effectively: Leverage group deliberation to refine individual views, as GJP teams using structured debate outperformed solo forecasters by 20-23% in aggregate scores.
  • Balance errors dynamically: Treat forecasting as a skill requiring constant adjustment, akin to riding a , where superforecasters logged thousands of predictions to refine error patterns over years.
These principles underscore that forecasting skill develops through practice, not innate talent, with GJP data showing non-superforecasters closing half the performance gap via targeted training by 2015.

Team-Based and Organizational Methods

Superforecasters achieve superior accuracy through collaborative team structures that facilitate deliberation and challenge individual biases. In the (GJP), teams of superforecasters outperformed solo forecasters and prediction markets by 15-30% over two years, as teams leveraged diverse perspectives to refine probabilities and reduce overconfidence. Elite superforecaster teams demonstrated higher engagement rates, with members interacting more frequently via discussions that integrated new evidence and tested assumptions, leading to lower noise and bias in collective outputs. Key team-based techniques include structured deliberation protocols, where participants break down questions into subcomponents, balance base rates with specific evidence, and update forecasts iteratively based on peer feedback. This process mirrors Bayesian updating in practice, as teams aggregate probabilistic judgments to produce calibrated ensemble predictions that exceed individual capabilities. Empirical analysis from GJP attributes roughly 50% of accuracy gains to via aggregation, 25% to , and 25% to enhanced information processing through team synergy. At the organizational level, entities like Good Judgment Inc. implement superforecasting by spotting forecasting talent, providing targeted training in probabilistic reasoning, and forming hybrid teams that blend domain experts with generalist superforecasters. Companies can integrate these methods by establishing dedicated forecasting units that conduct regular tournaments on strategic uncertainties, such as market shifts or geopolitical risks, and embed aggregated team outputs into decision frameworks to supplant reliance on unverified expert consensus. For instance, trained amateur teams in GJP surpassed professional analysts by fostering a culture of and evidence-based debate, a scalable approach for firms seeking to upgrade judgment without overhauling hierarchies.

Criticisms and Limitations

Empirical and Methodological Challenges

Studies attempting to replicate superforecaster identification outside large-scale tournaments like (GJP) have encountered significant empirical hurdles due to small participant pools and abbreviated evaluation periods. In an experiment with an initial pool of 314 experts reduced to 195 active forecasters over nine months, only two individuals were classified as superforecasters using thresholds akin to Tetlock's criteria, underscoring the scarcity of such performers under resource constraints typical of non-tournament environments. This contrasts with the GJP's recruitment of over 5,000 participants across four years, raising questions about the feasibility of identifying superforecasters in corporate or policy settings without similar scale and incentives. Methodologically, superforecaster selection relies on probabilistic accuracy metrics like the , which aggregates performance across hundreds of questions but may overemphasize on frequentist events while underweighting rare, high-impact outcomes where data scarcity limits skill differentiation from luck. Tetlock's framework excels in the "Goldilocks zone" of resolvable, medium-horizon geopolitical questions—typically 1-12 months out with objective resolutions—but falters on "" events, as evidenced by GJP superforecasters assigning a mere 3% probability to over 200,000 U.S. cases by late March 2020, despite rapid escalation. Such methodological focus on average accuracy incentivizes hedging toward base rates and continuity, potentially misaligning with needs that prioritize tail risks and causal impacts over logarithmic scoring penalties. Replication efforts further highlight variance in sustaining superforecaster status; while approximately 70% retained top-tier performance year-over-year in the GJP, external validations with expedited training or smaller cohorts show inconsistent gains, with structured interventions like analogy-based prompting improving group averages but failing to reliably produce additional superforecasters. These challenges are compounded by participant engagement issues, where incentives like monetary rewards (€650 in one study) were necessary to maintain involvement, suggesting that voluntary or low-stakes real-world applications may dilute observed edges. Overall, while Tetlock's tournaments demonstrate modest outperformance (e.g., 30% improvement over baselines), the methodology's dependence on motivated elites and curated questions limits generalizability to diverse, high-uncertainty domains.

Applicability to High-Uncertainty Events

Superforecasters achieve superior calibration and accuracy relative to lay predictors and experts in domains with moderate and recurring patterns, but their methods exhibit pronounced limitations for high- events, defined as rare occurrences with sparse historical precedents, fat-tailed distributions, and intricate, non-linear causal pathways. These include black swan-like phenomena such as novel pandemics, abrupt geopolitical ruptures, or systemic financial collapses, where base rates approach zero and signals are overwhelmed by . In such contexts, the Bayesian updating central to superforecasting struggles to differentiate genuine precursors from false alarms, as analogous data for establishing priors is often absent or unreliable. Empirical evidence from early forecasts underscores this constraint: participants in superforecaster-led efforts, including those affiliated with , assigned approximately a 3% probability to exceeding 200,000 global confirmed cases by March 20, 2020, underestimating the rapid escalation that surpassed the threshold within days. This outcome, while consistent with historical on less extreme scenarios, highlights how adherence to low-probability estimates—rational given limited early indicators—can foster complacency toward tail risks, particularly in fat-tailed processes where extreme events carry disproportionate impact. and collaborators have critiqued such approaches for conflating aleatory uncertainty with epistemic gaps, arguing that single-point or probabilistic forecasts inadequately capture the asymmetry of fat tails, leading to systematic underestimation of catastrophic potentials despite apparent . The skill acquisition process further constrains applicability: superforecasting proficiency develops through iterative practice across hundreds of questions with timely resolution and feedback, enabling refinement of probabilistic judgments—a mechanism infeasible for infrequent high-uncertainty events that may span decades between occurrences. A controlled identifying superforecasters from a pool of 314 experts over nine months found that, under real-world of opportunities, these individuals did not sustain outperformance, implying that the requisite volume of practice evaporates in domains lacking . Proponents, including Philip Tetlock, counter that superforecasting prioritizes high-resolution hit rates with minimal false positives over blanket , distinguishing it from precautionary heuristics suited to existential threats; nonetheless, this framework may prioritize aggregate accuracy over robustness to outliers. Decision-theoretic critiques amplify these empirical hurdles, noting that even well-calibrated low probabilities often fail to influence action in or organizational settings, where cognitive and institutional biases discount remote contingencies unless amplified by vivid narratives or stressors. Forecasts in adaptive systems, such as geopolitical conflicts or economic shocks, can moreover become self-defeating if disseminated, as actors adjust behaviors in response, invalidating the baseline assumptions. Consequently, while superforecasting enhances through explicit , its wanes in high-uncertainty regimes, where strategies—emphasizing rapid and over —may prove more resilient.

Recent Developments

Post-2020 Advances and Forecasts

Good Judgment's superforecasters applied their methods to the in 2020, predicting over 53 million global cases and 800,000 to 8 million deaths with high probability a year ahead of the outcomes. On September 8, 2020, they assigned a 71% that enough FDA-approved would be available to vaccinate 25 million people by year's end, reflecting early to evolving epidemiological . These predictions earned recognition for accuracy, with outlets noting their prescience amid widespread in official projections. Post-pandemic, superforecasters expanded applications to economic and geopolitical domains. In 2024, Good Judgment teams outperformed futures markets by 30% in accuracy for events including interest rate decisions, demonstrating sustained edge over market-based indicators. Their 2024 forecasts accurately anticipated China's GDP growth below 5%, the Labour Party's victory in Britain's , and ongoing military conflict in , achieving a score of 4.5 out of 8 on a set of major global predictions. Methodological advances included the maturation of platforms like Good Judgment Open, launched prior but scaled post-2020 for in calibrated challenges on topics from to risks. Collaborations grew, such as 2025 engagements with to forecast U.S. foreign aid funding levels, aiding in . Philip Tetlock's ongoing research emphasized refining probability estimates for global catastrophic risks via early-warning indicators, building on tournament data to enhance long-horizon predictions. Evaluation benchmarks post-2020 reinforced superforecaster efficacy, with difficulty-adjusted scores of 0.081 in recent assessments surpassing baselines of 0.25 and select predictive models. These developments underscore iterative improvements in aggregation techniques, such as median-based crowd synthesis, applied to diverse horizons from quarterly economic shifts to multi-year geopolitical trends.

AI Integration and Future Outlook

Integration of into superforecasting has primarily occurred through hybrid human-AI systems, where large language models (LLMs) serve as assistants to enhance probabilistic reasoning. A 2024 study found that forecasters using LLMs prompted with superforecasting techniques—drawing on principles like updating beliefs with evidence and aggregating diverse perspectives—achieved higher accuracy on geopolitical and economic questions than unaided humans, with improvements attributed to the models' ability to simulate base rates and counter cognitive biases. Similarly, ensembles of multiple LLMs have produced forecasts comparable to those of expert human superforecasters, as demonstrated in research by Philip Tetlock at the , where AI aggregation mitigated individual model hallucinations and improved on events like outcomes and shifts. Organizations like Good Judgment, co-founded by Tetlock, have applied superforecaster methods to AI-related forecasts, such as assessing risks of power-seeking AI systems, revealing systematic differences in probability estimates: superforecasters assigned lower extinction risks from advanced AI (median around 1-2% by 2100) compared to AI domain experts (median 3-4%), highlighting AI's role in stress-testing human judgments against empirical track records. The Center for AI Safety developed "FiveThirtyNine," an LLM-based bot using GPT-4o, which generated superhuman forecasts on benchmarks by incorporating chain-of-thought reasoning and historical analogies, outperforming baseline models on novel queries in 2024 trials. Looking ahead, is projected to surpass standalone human superforecasters in speed and by the late , with a startup's system already defeating human teams in a September 2025 forecasting tournament by better aggregating signals from vast datasets. approaches are expected to dominate, blending AI's data-processing efficiency with human strengths in and ethical framing for "Superforecasting 2.0," though challenges persist in handling black-swan events where training data is sparse. Tetlock anticipates that by democratizing access to AI-augmented tools, such integration could elevate average organizational from haphazard expert opinions to calibrated, crowd-sourced wisdom, potentially reducing policy errors in high-stakes domains like .

References

  1. [1]
    About Superforecasting | Unprecedented Accurate & Precise ...
    ... Superforecasters have beaten every challenger they've faced. Our cofounder, Philip Tetlock, profiled several of these talented forecasters in his New York ...
  2. [2]
    Superforecasting by Philip E. Tetlock, Dan Gardner: 9780804136716
    In stock Free delivery over $20In Superforecasting, Tetlock and coauthor Dan Gardner offer a masterwork on prediction, drawing on decades of research and the results of a massive, government ...<|separator|>
  3. [3]
    The first championship season - Good Judgment
    The simple median of their forecasts was 35% to 72% more accurate than any other research team in the competition! Putting the Superforecasters together on ...
  4. [4]
    Book Review: Superforecasting - NDU Press
    Top GJP forecasters beat a control group by 60 percent in the first year and by 78 percent in the second. As Tetlock states, they even “outperformed ...
  5. [5]
    Superforecasters: The Art of Accurate Predictions
    Nov 5, 2015 · Superforecasters, on the other hand, are able to overcome many of these cognitive hurdles, helping them forecast future global events with surprising accuracy.
  6. [6]
    The Science Of Superforecasting - Good Judgment
    they don't show the variability that the rest of us show. They're very smart; but also, very importantly, they don't think in ...
  7. [7]
    Superforecasting: The Art and Science of Prediction — Summary
    Jul 23, 2020 · Temperament: Cautious, Humble, Non-Determinist. · Cognition: Open-Minded, Intelligent, Curious, Reflective, Numerate. · Analytical: Pragmatic, ...
  8. [8]
    Evidence on good forecasting practices from the Good Judgment ...
    Roughly 70% of the superforecasters maintained their status from one year to the next. Across all the forecasters, the correlation between performance in one ...
  9. [9]
    Common Questions about Good Judgment Inc and Superforecasters
    Jul 12, 2024 · Good Judgment works with some 180 Superforecasters from around the world whose forecasting accuracy placed them in the top 1-2% of the more than ...
  10. [10]
    Superforecasting: Supremely Awesome - Econlib
    Oct 15, 2015 · Philip Tetlock organized one of several teams competing to make accurate predictions about matters we normally leave to intelligence analysts.
  11. [11]
    superforecaster - Wiktionary, the free dictionary
    Etymology. From super +‎ forecaster. Noun. English Wikipedia has an article on: superforecaster · Wikipedia. superforecaster (plural superforecasters). One who ...
  12. [12]
    Of Hedgehogs, Foxes, and Superforecasters - Cato Institute
    Tetlock is a professor at the University of Pennsylvania. In 2011, he launched the Good Judgment Project (GJP) with his research (and life) partner Barbara ...
  13. [13]
    How to Become a Superforecaster® - Good Judgment
    Superforecaster Origins: the Good Judgment Project​​ Superforecasters were a surprise discovery of the Good Judgment Project (GJP), the research-and-development ...Missing: etymology | Show results with:etymology
  14. [14]
    Philip Tetlock - The Decision Lab
    He coined the term “superforecaster” to refer to individuals with particularly good judgment, who are able to foresee future outcomes far more accurately than ...
  15. [15]
    [PDF] Superforecaster® Analytics | Good Judgment
    The term “Superforecasting” was originally coined by The Good. Judgment Project, Good Judgment's academic precursor and the undisputed winner of the largest ...
  16. [16]
    Philip E. Tetlock - Penn Arts & Sciences
    It reports a series of relatively small scale forecasting tournaments that I started in 1984 and wound down by 2003. A total of 284 experts participated as ...Defining and Assessing Good... · Accountability and Attributions...
  17. [17]
    2008 – Philip Tetlock - Grawemeyer Awards
    Aug 31, 2008 · “Expert Political Judgment: How Good Is It? How Can We Know ... In a 20-year study of 27,000 predictions by 284 political experts, Tetlock ...
  18. [18]
    [PDF] 2015---superforecasters.pdf - Wharton Faculty Platform
    Different research literatures—ranging from cognitive debiasing to group dynamics to prediction markets and statistical aggregation techniques—offered.
  19. [19]
    ACE - IARPA
    The goal of the ACE Program is to dramatically enhance the accuracy, precision, and timeliness of intelligence forecasts for a broad range of event types.Missing: results | Show results with:results
  20. [20]
    Are Actuaries Superforecasting Material?
    Sep 1, 2016 · The request led Tetlock to launch the Good Judgment Project with his research partner Barbara Mellers in 2011. They recruited more than 20,000 ...
  21. [21]
    The Superforecasters' Track Record - Good Judgment
    Superforecasters beat all competing research teams in the IARPA ACE tournament by 35-72%. Learn more » · US intelligence analysts. Good Judgment was over 30% ...
  22. [22]
    Good Judgment: See the future sooner with Superforecasting
    Tetlock and Mellers co-founded Good Judgment Inc to provide forecasting services to partners and clients in the non-profit, government, and private sectors. Why ...Good Judgment Open · Superforecasting AI · The Science Of Superforecasting
  23. [23]
    [PDF] Identifying and Cultivating Superforecasters as a Method of ...
    The Good Judgment Project (GJP). In this article, we describe the best-performing strategy of the winning research program: the GJP. To preempt confusion, we ...
  24. [24]
    Informed Practice and Superforecasting: Taking Your Forecasts to ...
    ... Tetlock & Dan Gardner in Superforecasting ... They also update these probabilities as new information becomes available, a process known as Bayesian reasoning.
  25. [25]
    How The Superforecasters Do It - Commoncog
    Dec 24, 2019 · Some superforecasters do use the actual Bayes's theorem (Tetlock gives one example in his book), but many of them simply adopt the underlying ...How They Actually Do It · Take the Outside View First · Start with a probability...
  26. [26]
    Ten Commandments for Aspiring Superforecasters - Farnam Street
    The Knowledge Project interview with Philip Tetlock deconstructs our ability to make accurate predictions into specific components.Missing: definition | Show results with:definition
  27. [27]
    What are the traits of Superforecasters? - WisdomFromExperts
    Oct 7, 2020 · They score very high on active open-mindedness. · They think slowly and systematically. · They think in probabilities and are highly numerate.Missing: characteristics | Show results with:characteristics
  28. [28]
    Superforecasting: The Art & Science of Prediction - Summary
    A book about how to become a superforecaster, an often ordinary person who has an extraordinary ability to make predictions about the future with a degree of ...Missing: definition | Show results with:definition
  29. [29]
    Identifying and cultivating superforecasters as a method of ... - PubMed
    In this article, we describe the winning strategy: culling off top performers each year and assigning them into elite teams of superforecasters.
  30. [30]
    Ten Commandments for Aspiring Superforecasters - Good Judgment
    Superforecasters are in the habit of posing the outside-view question: How often do things of this sort happen in situations of this sort? 4. Strike the right ...
  31. [31]
    Identifying and Cultivating Superforecasters as a Method of ...
    May 18, 2015 · In this article, we describe the winning strategy: culling off top performers each year and assigning them into elite teams of superforecasters.
  32. [32]
    Evidence on good forecasting practices from the Good Judgment ...
    Jul 2, 2019 · And [teams of superforecasters] beat prediction markets by 15% to 30%.” “On average, teams were 23% more accurate than individuals.” What does ...
  33. [33]
    Superforecasters vs FT Readers: Results of an Informal Contest in ...
    In a fascinating contest by FT, Good Judgment's Superforecasters outperformed an 8500-strong crowd in forecasting key events for 2023.
  34. [34]
    Crowd prediction systems: Markets, polls, and elite forecasters
    Our main test compares a prediction market against team-based prediction polls, using data from a large, multi-year forecasting competition. Each of these two ...
  35. [35]
    Announcing ForecastBench, a new benchmark for AI and human ...
    Oct 1, 2024 · Superforecasters achieved the best overall performance with a Brier score of 0.093. The general public (0.107) and best AI model (0.111) lagged ...
  36. [36]
    Evaluating LLMs on Real-World Forecasting Against Human ... - arXiv
    Apr 14, 2025 · Compared to model performance, superforecasters still significantly outperform the bots with a Brier score of 0.0225, far lower than o3 which ...Missing: studies | Show results with:studies
  37. [37]
    AI forecasting tournament tried to predict 2025. It couldn't. | Vox
    Sep 5, 2025 · Across four benchmarks, the actual performance of state-of-the-art models in summer 2025 was better than either superforecasters or AI experts ...
  38. [38]
    Meet Military Historian and Superforecaster® Jean-Pierre Beugoms
    Jean-Pierre Beugoms joined the Good Judgment Project back in 2011, in its first year, and became one of the first ever group of superforecasters.Missing: notable | Show results with:notable
  39. [39]
    Superforecaster Profiles - Good Judgment
    Meet Superforecaster® Brett Specter · Meet Superforecaster® Vijay Karthik · Meet Superforecaster® Dawna Coutant · Meet Superforecaster® Elan Pavlov · Meet ...Missing: prominent | Show results with:prominent
  40. [40]
  41. [41]
    Meet Good Judgment's Superforecasting Team
    Co-founded by Good Judgment Project leads Philip Tetlock and Barbara Mellers, Good Judgment Inc provides forecasting and training services for public and ...
  42. [42]
    IARPA Announces Publication of Data from the Good Judgment ...
    Apr 28, 2017 · This data set includes millions of forecasts made by participants over the four years of the ACE forecasting competition and led to many ...
  43. [43]
    Good Judgment® Open
    Good Judgment® Open. Are you a Superforecaster®? Join the internet's smartest crowd. Improve your forecasting skills and find out how you stack up.Featured Questions · Challenges · Sign In · Sign UpMissing: characteristics | Show results with:characteristics<|separator|>
  44. [44]
    How I Became a Superforecaster
    Nov 18, 2015 · Tetlock says superforecasters share some qualities, like being actively open-minded, seeking new data, updating one's opinion when new data ...Missing: characteristics | Show results with:characteristics
  45. [45]
    Good Judgment's 2024 in Review
    Jan 9, 2025 · Our partnership with ForgeFront and the UK Government's Futures Procurement Network continued to grow in 2024. In November, Good Judgment's CEO ...Missing: contracts | Show results with:contracts
  46. [46]
    Good Judgment Inc. — Forecasts on U.S. Foreign Aid Funding ...
    This $72,000 agreement funded Good Judgment Inc. to deliver six forecasts about U.S. government foreign aid funding, with a specific focus on global health ...
  47. [47]
    [PDF] Superforecasting AI Governance - Good Judgment
    May 5, 2025 · The data for this report was generated by Good Judgment Inc's certified Superforecasters from around the world. Their forecasting accuracy ...
  48. [48]
    How government can predict the future | BIT
    Jan 10, 2020 · The ability to predict future events is a skill that can be learned and developed over time. The key is to ask people not just to make predictions.Missing: applications | Show results with:applications
  49. [49]
  50. [50]
  51. [51]
    Why Top Investors Choose Our Superforecasters - Good Judgment
    Superforecasters were more accurate and less noisy than the futures in forecasting the Fed's next move. Data-Driven Decision Making. At the core of our ...
  52. [52]
    Superforecasting: How to Upgrade Your Company's Judgment
    Imagine that you could dramatically improve your firm's forecasting ability, but to do so you'd have to expose just how unreliable its predictions.
  53. [53]
    Good Judgment's 2024 in Review
    Good Judgment's clients and FutureFirst™ subscribers posed 151 questions to the Superforecasters on our proprietary platform, with a total of 1,132 forecasting ...<|control11|><|separator|>
  54. [54]
    [PDF] Superforecasting: How to Upgrade Your Company's Judgment
    Schoemaker served as an adviser to and Tetlock co-led the. Good Judgment Project. Superforecasting: How to Upgrade. Your Company's. Judgment. BY PAUL J.H. ...
  55. [55]
    Compromising improves forecasting - PMC
    May 17, 2023 · As announced on the website of the Good Judgement Project, the goal is to 'see the future sooner'. Rightly predicting that the price of oil will ...
  56. [56]
    [PDF] What I've Learned from the Good Judgment Project - SOA
    Jul 2, 2015 · The ACE tournament run by IARPA stands for “Aggrega- tive Contingent Estimation,” and it's run under the Office of Anticipating Surprise (man, I ...
  57. [57]
    Superforecasting by Philip Tetlock Book Summary
    The Good Judgment Project (GJP), led by Philip Tetlock and Barbara Mellers, recruited thousands of volunteer forecasters to predict global events as part of ...The Art And Science Of... · Beliefs Are Hypotheses To Be... · Foxy Forecasters Beat...<|separator|>
  58. [58]
    [PDF] Identifying and Cultivating Superforecasters as a Method of ...
    Table 2 shows that elite superforecaster teams engaged with each other more frequently than regular teams. Their engagement could occur in two ways ...
  59. [59]
    Philip Tetlock on Superforecasting - Econlib
    Dec 19, 2015 · In this conversation, Tetlock discusses the meaning, reliability, and usefulness of trying to assign probabilities to one-time events. LISTEN ...
  60. [60]
    Superforecasting reality check: Evidence from a small pool of ...
    Jul 3, 2020 · In this study, we provide supportive empirical evidence from an experiment with an initial (small) pool of 314 experts and an identification phase of (just) 9 ...
  61. [61]
    The Limits of Applied Superforecasting - Commoncog
    Apr 28, 2020 · On February 20th this year, Tetlock's superforecasters predicted only a 3% probability that there would be 200,000+ coronavirus cases a month ...
  62. [62]
    The Case Against Superforecasting - by R.W. Richey
    then a professor of political science at Berkeley — decided to conduct a study on the accuracy of experts ...
  63. [63]
    The probability conflation: A reply to Tetlock et al. - ScienceDirect
    We respond to Tetlock et al. (2022) showing (1) how expert judgment fails to reflect tail risk and (2) the lack of compatibility between forecasting ...Missing: debate | Show results with:debate
  64. [64]
  65. [65]
    [PDF] The Probability Conflation: A Reply - arXiv
    Jan 26, 2023 · Tetlock et al. (2022) [1], in their criticism of claims by a paper titled ”On single point forecasts for fat-tailed variables”.Missing: debate | Show results with:debate
  66. [66]
    False dichotomy alert: Improving subjective-probability estimates vs ...
    The superforecasting program aims at achieving high hit rates at low cost in false-positives, whereas Taleb et al. prioritize alerting us to systemic risk.
  67. [67]
    Superforecasting will not save us - Engelsberg Ideas
    Nov 4, 2022 · Instead of trying to predict the future, we should be embracing uncertainty and using analysis, monitoring, warning and action to bring ...Missing: methodology | Show results with:methodology
  68. [68]
    Forecasting the Pandemic: Dealing with Death and Other Emotional ...
    Superforecasters saw the highest likelihood of more than 53 million COVID-19 cases and 800,000 to 8 million COVID-related deaths worldwide a year in advance.Missing: post- | Show results with:post-
  69. [69]
    Superforecasting Case Studies - Good Judgment
    Learn how Good Judgment's clients use Superforecasting methods to generate better forecasts and improve their decision-making.Missing: business | Show results with:business
  70. [70]
    'Superforecasters' Are Making Accurate Predictions About COVID-19
    Jun 11, 2020 · 'Superforecasters,' ordinary people paid for their accurate predictions, are being called upon to forecast the future of the pandemic.
  71. [71]
    Superforecasting & Good Judgment In The Press/News
    Superforecasting, associated with Good Judgment, involves laypeople out-predicting futures markets, with 30% more accuracy than futures in 2024-2025.Missing: developments | Show results with:developments
  72. [72]
    What the “superforecasters” predict for major events in 2025
    Nov 20, 2024 · The superforecasters scored 4.5/8 with their forecasts for 2024, correctly calling China's sub-5% GDP growth, Britain's election, continued conflict in Ukraine.Missing: developments | Show results with:developments
  73. [73]
    Professor Philip Tetlock's forecasting research - Founders Pledge
    Mar 8, 2022 · This research focuses on improving probability estimates of early-warning indicators of global catastrophic and existential risks.
  74. [74]
  75. [75]
    [PDF] Superforecasting Power-Seeking AI II_supplementary project_231006
    At the end of the forecasting period, Good Judgment's data scientists aggregated the Superforecasters' latest forecasts by taking the median value ...<|separator|>
  76. [76]
    [PDF] LLM Assistants Improve Human Forecasting Accuracy - arXiv
    Aug 22, 2024 · We evaluate the effect on human forecasters of two LLM assistants: one designed to provide high-quality ('superforecasting') advice, and the ...
  77. [77]
    Can AI Predict the Future? - Knowledge at Wharton
    Jan 14, 2025 · New research from Wharton's Philip Tetlock finds that combining predictions from large language models can achieve accuracy on par with human forecasters.<|control11|><|separator|>
  78. [78]
    Superforecasting AI - Good Judgment
    Results of Good Judgment's “Superforecasting Power-Seeking AI” project, with an in-depth analysis of Joseph Carlsmith's paper.
  79. [79]
    Superhuman Automated Forecasting | CAIS - Center for AI Safety
    Sep 9, 2024 · We are excited to announce “FiveThirtyNine,” a superhuman AI forecasting bot. Our bot, built on GPT-4o, provides probabilities for any user-entered query.
  80. [80]
    British AI startup beats humans in international forecasting competition
    Sep 20, 2025 · AI's performance still lags behind the best human forecasters, but it has left some believing AI could outstrip humans sooner than most expected ...
  81. [81]
    Superforecasting 2.0: Wisdom of Crowds and Wisdom of AI-Clouds
    Sep 13, 2024 · This hybrid model combines AI's efficiency with human expertise to address complex, nuanced scenarios. Superforecasting 2.0 Challenges: While ...
  82. [82]
    Predicting the Future: The Supergroup of AI, Humans, Hedgehogs ...
    Sep 22, 2025 · How AI can assist human decision-making and democratize super-forecasting to better predict uncertain futures.<|separator|>