This simulator teaches executives how confidence, calibration, uncertainty, probability curves, and decision quality interact.
Exact answers are brutally unforgiving.
| Question | Your Estimate | Result |
|---|---|---|
| {{ q.text }} | {{ q.pointEstimate }} | {{ q.correct ? 'RIGHT' : 'WRONG' }} |
A probability curve lets you express where outcomes feel more or less likely across the range.
The edges are not impossible. They are near-zero probability.
The most likely value does not need to sit halfway between the lowest and highest values.
By manipulating these variables, we can model classic statistical archetypes that define how uncertainty behaves. Select an archetype below.
Profile: {{ archetypeDetails.setup }}
Business Impact: {{ archetypeDetails.lesson }}
The Board asks for an estimate to rewrite the core billing system. Historical data suggests the most likely duration is 200 days. The Board demands a precise delivery window.
Adjust the slider to see how giving them a "precise" (narrow) estimate mathematically destroys your probability of actually hitting the target.
Specific estimates are often mathematically identical to low-confidence estimates. When leaders demand an exact number, they are inherently forcing the team to provide an estimate with a near-zero probability of success.
Hubbard developed "The Equivalent Bet" to fix overconfidence. Imagine you can choose between two ways to win $10,000:
You win $10,000 ONLY IF the true answer falls inside the range you just built:
You spin a wheel that has a mathematically guaranteed 90% chance of landing on a win.
If you prefer to spin the wheel, your interval is too narrow. You aren't actually 90% confident. You must widen your bounds until you are completely indifferent between the two options.
Now that you know the Equivalent Bet, provide a calibrated range for these 5 new questions. You choose the boundaries, the most likely outcome, and your confidence level. Imagine spinning the wheel before every submission.
{{ calibrationScore.gap <= 10 ? "Excellent. Your ego is checked and you are properly accounting for uncertainty." : "Still uncalibrated. Your stated confidence does not match reality." }}
| Question | Your Range | Conf | True Answer | Result |
|---|---|---|---|---|
| {{ q.text }} | {{ q.low }} → {{ q.high }} | {{ q.conf }}% | {{ q.answer.toLocaleString() }} | {{ q.correct ? 'HIT' : 'MISS' }} |
Individuals are notoriously noisy and biased, but aggregated estimates often zero in on the truth. Provide a calibrated range for the following scenario, then ask 5 simulated "Experts" for theirs.
| Expert | Their 90% CI | Result |
|---|---|---|
| Expert {{ i + 1 }} | {{ e.low.toLocaleString() }} → {{ e.high.toLocaleString() }} | {{ (14500 >= e.low && 14500 <= e.high) ? 'HIT' : 'MISS' }} |
| THE AGGREGATE AVERAGE | {{ wisdomAvgLow.toLocaleString() }} → {{ wisdomAvgHigh.toLocaleString() }} | {{ (14500 >= wisdomAvgLow && 14500 <= wisdomAvgHigh) ? 'HIT' : 'MISS' }} |
Notice how individual experts missed the target entirely (high noise), but the mathematical average of their bounds perfectly captured the true answer. This is why collaborative estimation (like Planning Poker) is mathematically superior to asking a single Lead Architect.
Trivia questions break down ego, but estimation matters most in the real world. You are asked to estimate the duration of an upcoming Cloud Migration. Provide a calibrated range for the project's lead time in days.
Your Estimate: {{ state.businessScenario.low }} to {{ state.businessScenario.high }} days. You scored a {{ state.businessScenario.result }}.
The exact same overconfidence bias that makes us fail trivia questions will cause you to bankrupt a multi-million dollar IT project if you don't widen your bounds to respect uncertainty.
Estimate the Year 1 Revenue of a new AI product. Instead of guessing one giant range, break it down into 3 variables. Provide a 90% Confidence Interval for each.
You don't need exact numbers to forecast accurately. You just need calibrated bounds on the sub-components. By decomposing the problem, your uncertainty is mathematically contained.
Measurement is an investment. You have a $100,000 budget and must decide whether to launch a massive, highly uncertain project.
Information is only valuable if it reduces uncertainty enough to change a decision and justify its cost. By buying information, you transformed a gamble into a calculated, safe bet.
Pedagogically, 90% is the training standard to shock people out of false precision. Operationally, insisting on 90% confidence everywhere is dysfunctional. Confidence is a decision variable, not a virtue signal.
{{ confidenceTradeoffScenario.description }}
Waiting for 90% confidence in this scenario causes the delay and opportunity costs to vastly exceed the cost of failure. An honestly calibrated 80% forecast is vastly superior to a politically optimized 95% deterministic commitment.
The real question becomes: "When we say 80%, do outcomes actually land inside the range ~80% of the time?" That is calibration.
| 50% | Exploratory / speculative |
| 70% | Directional confidence |
| 80% | Actionable operational confidence |
| 90% | Strong planning confidence (Training Standard) |
| 95%+ | Extremely conservative / high-cost-of-failure |
A Brier score is a way to measure how good probabilistic predictions are. It rewards accurate confidence, honest uncertainty, and calibration, while penalizing unjustified certainty, overconfidence, and underconfidence.
Formula: (Predicted Probability - Actual Outcome)² (0 is perfect, 1 is worst)
You predicted {{ state.brierDemo.prediction }}% but it failed. This results in a terrible score ({{ brierDemoScore }}). This is why overconfidence is punished brutally.
You predicted {{ state.brierDemo.prediction }}% and it happened. Your score is {{ brierDemoScore }}. Very good. You were confident and correct.
You predicted {{ state.brierDemo.prediction }}% and it missed. Score: {{ brierDemoScore }}. Not amazing, but much better than fake certainty.
Brier scores force a shift from asking "Was I right?" to asking "Was my confidence appropriate?" That is a profound shift in executive thinking.
| 0.00 | Perfect |
| 0.05 - 0.10 | Exceptional |
| 0.10 - 0.20 | Strong |
| 0.20 - 0.30 | Moderate |
| 0.50+ | Terrible calibration |
Predicts 99%, but is wrong. Brier: 0.9801.
Looks decisive, but forecasts terribly.
Predicts 55%, and is correct. Brier: 0.2025.
Often calibrated, but timid.
Predicts 80% and is correct. Brier: 0.0400.
They are not always right, but they are probabilistically honest.
You achieved a {{ calibrationScore.hitRate.toFixed(0) }}% hit rate with a Brier Score of {{ calibrationScore.brierScore }} (0 is perfect, 0.25 is random guessing). This metric has been saved to your learner profile. Calibration is a trainable skill—you can watch this score improve over time!
"Strong forecasters are not people who know the future. They are people whose confidence behaves more like reality. Measurement is about uncertainty reduction, not absolute certainty."
"An estimate is a probability distribution. A target is a business desire. The moment a leader conflates a target with an estimate, they destroy the estimate and guarantee failure."