Dead Reckoning: How to Design Wargames That Punish Overconfident Intelligence Estimates
E. SokolovMost wargames hand players a sanitized intelligence picture and call it ground truth. Blue cell gets a map overlay. Red cell gets an order of battle. Everyone proceeds. What nobody stress-tests is what happens when the intelligence estimate itself is the failure point — not enemy action, not logistics, not weather. The estimate.
Photo by Matias Luge on Pexels.
This is a design problem. And it's fixable.
Why Intelligence Estimates Are the Invisible Assumption
When the U.S. wargamed the Normandy landings in 1944, FUSAG deception planning fed deliberately corrupted assessments into German command structures. The deception worked partly because German wargames had no native way to flag confidence intervals in their own intelligence — commanders received estimates as facts. After-action literature from OKW exercises in 1943 shows adjudicators treating agent reports as binary: true or false. No probabilistic middle ground existed in the game's adjudication rules.
The same failure mode appears in unclassified post-mortems of GLOBAL GUARDIAN exercises in the 1990s. Players used threat assessments almost as a given — a stable backdrop against which they moved pieces. When anomalies surfaced (unexpected force postures, ambiguous signals), the game had no mechanism to force players to update their estimates under time pressure. The anomaly got noted. Then play moved on.
That's not a player problem. That's a design problem.
The Dead Reckoning Mechanic
Navigators use dead reckoning when GPS fails: you start from a known position and project forward using speed, heading, and elapsed time. Error compounds with distance. The further you travel without a fix, the wider your cone of uncertainty grows.
Build that into your intelligence layer.
Here's the concrete approach. At game start, give each cell a baseline estimate with an attached confidence score — not a label like "high" or "low," but an actual number: 70%, 55%, 40%. Then introduce an estimate decay rate tied to elapsed game time and missing collection assets. Every turn without a confirming report, the confidence score drops by an amount the umpire sets in advance and keeps hidden from players.
When a player acts on an estimate, they must declare their assumed confidence level aloud before the umpire adjudicates the outcome. If their declared confidence exceeds the actual (hidden) score, the action carries a compounding penalty — not an automatic failure, but increased variance in the result. You're not punishing boldness. You're punishing unexamined boldness.
graph TD
A[Baseline Estimate Issued] --> B{Collection Assets Active?}
B -->|Yes| C(Confidence Maintained)
B -->|No| D[Confidence Decays per Turn]
C --> E{Player Declares Confidence}
D --> E
E --> F{Declared vs. Actual Score}
F -->|Overconfident| G[/Variance Penalty Applied/]
F -->|Calibrated or Under| H((Standard Adjudication))
The diagram is simple on purpose. The power is in running it. Once players realize their estimates are degrading in real time, they start asking for collection assets instead of assuming they already know enough. That behavioral shift is the whole point.
Calibration Sessions: Running the Mechanic Before the Main Game
Don't drop this cold into a large exercise. Pre-mortem it first.
Run a 45-minute calibration session before the main event. Give players ten historical intelligence estimates — real ones, drawn from declassified NIC assessments or academic case studies — and ask them to assign confidence scores. Then reveal the actual outcomes. Most groups discover they are systematically overconfident, often by 20-30 percentage points across the sample. That experience lands differently than a briefing slide telling them the same thing.
PETER SCHWARTZ's scenario planning work at Shell in the 1970s used a similar forcing function: before teams built scenarios, they had to enumerate their current assumptions and rank them by certainty. Surfacing the assumption was the intervention. The dead reckoning mechanic does the same thing mid-game, under pressure, when it actually hurts to be wrong.
What This Surfaces That Standard Games Miss
Tail events in real operations rarely announce themselves as tail events. They arrive disguised as confirmation of what the estimate already predicted — until suddenly they don't. The estimate wasn't wrong in an obvious way; it was confidently wrong in a way nobody checked.
Building confidence decay into your adjudication rules forces players to treat intelligence as a perishable asset rather than a fixed input. That's not a philosophical point. It's a playable mechanic. Run it twice with the same scenario and watch how differently teams behave when they know their picture is expiring.
The anomaly doesn't have to be exotic. Sometimes it just has to be the thing everyone assumed was confirmed.
Get Uncertainty Game in your inbox
New posts delivered directly. No spam.
No spam. Unsubscribe anytime.