Calibrated Confidence: Why Framing Decisions as Bets Exposes Sloppy Thinking

Vague confidence is unfalsifiable, which makes it useless. The moment you attach a number to your conviction — 70%, 55%, 90% — you've created something that reality can test.

8 min read · for the tool Bet Framing

A product lead says, “I’m confident we’ll hit our Q2 target.” The head of sales says, “I think the deal will close.” A founder tells investors, “We believe this market is ready.” Each statement sounds decisive. Each conveys conviction. And each is, from a decision-quality standpoint, nearly worthless — because none of them can ever be proven wrong in a way that generates learning.

If the Q2 target is missed, the product lead can say the confidence was always provisional. If the deal doesn’t close, the head of sales can point to an unforeseeable variable. If the market doesn’t materialise, the founder can adjust the timeline. Vague confidence is a form of insurance against accountability. It feels like commitment but functions as a hedge. And because it can’t be tested, it can’t teach you anything.

The research

Philip Tetlock, a psychologist at the University of Pennsylvania, spent twenty years studying the predictions of political and economic experts. The results, published in Expert Political Judgment in 2005, were sobering. The average expert performed barely better than chance — and worse than simple statistical algorithms — when making predictions about geopolitical events. But the failure wasn’t uniform. A subset of forecasters consistently outperformed the rest.

Tetlock and Dan Gardner explored this subset in Superforecasting (2015). The distinguishing characteristic of “superforecasters” wasn’t superior intelligence, deeper domain expertise, or access to better information. It was calibration — the alignment between their stated confidence and actual outcomes. When superforecasters said they were 80% confident, events occurred roughly 80% of the time. When they said 60%, events occurred roughly 60% of the time. Their probability estimates mapped onto reality.

The mechanism wasn’t mystical. Superforecasters simply did something that people often — including most experts — never do: they expressed their beliefs as specific probabilities, tracked their accuracy over time, and updated their estimates when new information arrived. The act of quantifying confidence forced granular thinking. Saying “I’m 70% confident” requires you to consider the 30% — the scenarios in which you’re wrong. Saying “I’m confident” does not.

Annie Duke, a former professional poker player and decision strategist, argued in Thinking in Bets (2018) that framing every decision as a bet fundamentally changes how you process it. A bet has three components: a specific prediction, a stated probability, and a consequence for being wrong. When you say “I bet £1,000 at 3:1 odds that this product will hit 10,000 users by June,” you’ve created something that is precise, time-bound, and falsifiable. You’ve also created something that forces you to reckon with the possibility of being wrong — which vague confidence conveniently avoids.

The mechanism

The cognitive shift from “I think this will work” to “I’m 70% confident this will deliver X by Y” is more profound than it appears. It engages what Gershon Keren described in a 1991 paper in Acta Psychologica as the calibration dimension of judgement — the correspondence between subjective probability and objective frequency.

People often are poorly calibrated. When they say they’re 90% confident, they’re right roughly 70-75% of the time. When they say 50-50, they’re often right 60-65% of the time. The gap between felt confidence and actual accuracy is consistent, measurable, and invisible without tracking. You can’t feel that you’re overconfident. Overconfidence feels exactly like well-founded confidence. The only way to detect it is to keep score.

Baruch Fischhoff and Ruth Beyth demonstrated a related phenomenon in a 1975 paper in Organizational Behavior and Human Decision Processes. They found that after learning outcomes, people consistently misremembered their prior predictions as having been more accurate than they actually were. This “hindsight bias” — the “I knew it all along” effect — means that without a written record, your brain retroactively adjusts your remembered confidence to match what happened. The overconfidence persists across decisions because the feedback loop that should correct it is corrupted by memory distortion.

Bet framing breaks this loop by externalising the prediction. A written probability estimate can’t be retroactively adjusted. When the outcome arrives, you have an unambiguous comparison: you said 70%, and it either happened or it didn’t. Across dozens of predictions, a pattern emerges — either your 70% estimates are coming true about 70% of the time, or they’re not. If they’re not, you’ve discovered something about your judgement that no amount of introspection could reveal.

A prediction without a probability is an opinion. A prediction with a probability is a hypothesis. Only one of these can teach you anything.

The practical implications

The percentage forces you to consider the counterfactual. When you write “I’m 75% confident this hire will work out,” you’ve implicitly acknowledged a 25% chance it won’t. That 25% creates space for the question: what would cause it to fail? What would I see if it were failing? This is the bridge between bet framing and tools like the pre-mortem — the probability estimate opens a cognitive door that vague confidence keeps shut.

The precision reveals the quality of your evidence. If you can’t distinguish between 60% and 80% confidence, that’s informative. It means your evidence base is thin — you don’t have enough data to differentiate between a marginal and a strong conviction. Tetlock found that superforecasters routinely used fine-grained estimates (67%, 73%), while poor forecasters clustered around round numbers (50%, 75%, 90%). The granularity isn’t false precision — it’s a reflection of how carefully the forecaster has interrogated their own reasoning.

Updating is where the real value lives. A bet isn’t a one-time event. As new information arrives, the probability should move. “I was at 70% last week; after Monday’s data, I’m now at 55%” is a sentence that tracks the evolution of your thinking in real time. This practice — Bayesian updating in everyday language — prevents the common failure mode of anchoring on an initial judgement and then defending it against all evidence. The bet frame makes updating feel natural rather than like an admission of error.

The bigger picture

Professional culture rewards conviction. The leader who says “I’m completely confident in this direction” is perceived as decisive. The one who says “I’m 65% confident, and here’s what would move me to 80%” is perceived as uncertain. This cultural norm produces a perverse incentive: leaders learn to express maximum confidence regardless of their actual epistemic state, because the social rewards for certainty outweigh the social rewards for accuracy.

The cost is invisible but enormous. Overconfident commitments consume resources that better-calibrated assessments would have allocated differently. Projects proceed at full investment when the actual probability of success warrants a smaller, staged approach. Teams rally behind directions that a more honest confidence estimate would have flagged for additional validation first.

Thinking in bets doesn’t make you less decisive. It makes you precisely as decisive as your evidence warrants. The superforecasters Tetlock studied weren’t paralysed by their probability estimates. They acted on 70% confidence when the situation called for it. The difference is that they knew it was 70% — and they planned for the 30%.

References

  1. Duke, A. (2018). Thinking in Bets: Making Smarter Decisions When You Don't Have All the Facts. Portfolio.
  2. Tetlock, P. E. (2005). Expert Political Judgment: How Good Is It? How Can We Know? Princeton University Press.
  3. Tetlock, P. E., & Gardner, D. (2015). Superforecasting: The Art and Science of Prediction. Crown.
  4. Fischhoff, B., & Beyth, R. (1975). I knew it would happen: Remembered probabilities of once-future things. Organizational Behavior and Human Decision Processes, 13(1), 1–16.
  5. Keren, G. (1991). Calibration and probability judgements: Conceptual and methodological issues. Acta Psychologica, 77(3), 217–273.