Predicting a scoreless draw is a different kind of problem than forecasting a winner. A 0-0 result compresses everything valuable about a match—tactical caution, defensive structure, chance quality—into a single, rare outcome. This article walks through the statistical tools analysts use to estimate the probability of a goalless game, explains their strengths and weaknesses, and shows how modern data (like expected goals) changes the picture.
Why a specialized model for goalless draws matters
On the surface, a 0-0 outcome is just one possible scoreline among many. In practice it behaves differently: zeros happen more frequently than a naive independent-scoring model often predicts, and they carry outsized meaning for betting markets and tactical analysis. Because few events—goals—determine the result, small changes in team behavior or luck can swing a match from 0-0 to 1-0.
From an applied perspective, accurately estimating the probability of a 0-0 draw matters for pricing odds, designing in-play strategies, and assessing defensive value. Analysts and bettors want robust, well-calibrated probabilities, not just point estimates; getting the long tail of low-scoring outcomes right improves every downstream decision.
Poisson as the baseline: the math and a quick table
The workhorse starting point for soccer and many low-scoring sports is the Poisson distribution. If a team’s goal-scoring can be modeled as a Poisson process with mean λ, then the chance it scores zero goals in a match is e^(−λ). Under the simplest independence assumption, the match-level probability of 0-0 is e^(−λ_home − λ_away).
This baseline is elegant and easy to use, and it often performs surprisingly well for bulk predictions. But its simplicity hides two important issues: teams’ scoring processes are not always independent, and league data frequently shows more zeros than the pure Poisson predicts.
| Home λ | Away λ | P(0-0) = e^(−(λh+λa)) |
|---|---|---|
| 0.4 | 0.4 | e^(−0.8) ≈ 0.45 |
| 0.8 | 0.6 | e^(−1.4) ≈ 0.25 |
| 1.2 | 0.9 | e^(−2.1) ≈ 0.12 |
That table shows how sensitive the 0-0 probability is to average scoring rates. Low λ values—typical in defensive leagues or against heavy favorites who sit back—push the 0-0 probability materially higher.
Why independence fails: correlation, game state, and low-scoring bias
Real matches create dependencies. A red card, pitch conditions, or two ultra-conservative lineups can simultaneously depress both teams’ chances to score. Conversely, an aggressive opening can inflate both teams’ expected goals. A simple product of two Poisson probabilities misses these joint effects.
Statisticians address this with extensions: bivariate Poisson distributions introduce a covariance term that allows positive correlation between the two score processes. Alternatively, zero-inflated Poisson models add a separate probability that a match becomes “structurally” goalless (for example, heavy rain makes scoring extremely unlikely), producing more zeros than a pure Poisson would.
Dixon–Coles correction and the problem of low-score dependence
The Dixon–Coles adjustment, introduced in 1997, is widely used in applied football analytics to correct for systematic misestimation of very low-scoring outcomes. Instead of assuming independence at all score combinations, it multiplies the joint probability of specific low score pairs (like 0-0, 1-0, 0-1, 1-1) by a small correction factor that accounts for empirical deviations.
Practically, the method preserves the Poisson flavor for most scores but recognizes that defensive matches or certain tactical matchups alter the likelihood of zeros and ones. Many forecasting systems still use a Dixon–Coles-style penalty when historical data show a persistent excess of 0-0 results compared to the baseline.
Expected goals (xG) and richer inputs for λ estimation
Estimating the Poisson means λ_home and λ_away is the heart of probability modeling. Classic approaches estimate attack and defense strengths from historical goals, adjusting for home advantage. Modern analyses use expected goals (xG) as a much richer signal, because xG measures the quality of chances rather than outcome alone.
Replacing goals with xG in model training improves stability and responsiveness. A team that creates chances with low finishing quality may have a low observed goal rate but a higher xG; using xG helps distinguish bad finishing from genuine lack of chance creation, yielding better λ forecasts and hence more accurate 0-0 probabilities.
Fitting the model: practical steps and calibration
To deploy a reliable model you need three things: a good data feed (goals, xG, match context), a fitting strategy (maximum likelihood with time decay or Bayesian updating), and a calibration check. Maximum likelihood estimation of Poisson or bivariate Poisson parameters is standard; many practitioners add time decay weights so recent matches influence λ more than older ones.
Calibration matters: if your predicted P(0-0) = 0.30 on average, roughly 30% of matches with that prediction should be goalless. Reliability diagrams and Brier scores are useful diagnostics here. When calibration drifts, re-weight features, re-estimate home advantage, or consider a zero-inflation term to account for persistent deviations.
Simulation, betting markets, and how the market absorbs information
Once you have λ estimates and any dependence adjustment, simulations turn those into actionable probabilities across all scorelines. Sampling from a bivariate Poisson (or adjusted Poissons) thousands of times produces stable probability estimates for 0-0 and other outcomes while allowing in-play updating when live events occur.
Betting markets often disagree with model outputs, presenting opportunities or warnings. In my experience working with trading desks, small misspecifications in λ—or failure to account for lineup rotations and weather—explain most market divergences in 0-0 pricing. Markets also rapidly incorporate public signals like team news, so modelers who incorporate lineups and minutes-played improve accuracy.
Limitations, corner cases, and tactical confounders
No model eliminates uncertainty. Low counts imply high variance: even a well-calibrated model will be wrong frequently on any individual match. Red cards, late tactical shifts, or psychological events (a team playing for a single point) can create abrupt departures from predicted probabilities.
Other pitfalls include small-sample bias for newly promoted teams, overfitting to league-specific scoring patterns, and ignoring competition-level effects. Cup games, for example, may have different incentives and substitution patterns than league matches, changing the baseline frequencies of 0-0 draws.
Implementation example: a simple workflow
Here is a practical workflow I’ve used: (1) collect season-level goals and xG by team and venue; (2) estimate initial attack/defense parameters with a Poisson GLM, including home advantage; (3) add a Dixon–Coles correction for the low-score joint distribution; (4) validate with a rolling holdout and calibrate with isotonic regression if needed; (5) simulate 100,000 matches per fixture to produce final probabilities.
In a real project forecasting a mid-tier European league, switching from goals-only to xG reduced mean Brier score for 0-0 by roughly 5–8% and increased calibration in the low-probability bins. These were modest but meaningful gains for pricing small-margin markets like 0-0 or under 1.5 goals.
When to use simple vs. complex models
Keep it pragmatic. For large-scale forecasting across many leagues, a Poisson baseline with xG-based λs and a modest Dixon–Coles correction often hits a productive balance of accuracy and computational simplicity. Use bivariate Poisson or full hierarchical Bayesian models when you need precise joint probabilities for specific fixtures or want to capture covariance explicitly.
In fast-moving environments like in-play trading, simpler models that update quickly and incorporate lineup or weather changes are often preferable to heavier models that provide marginal accuracy improvements at the cost of latency.
Estimating the probability of a 0-0 result is a careful mix of statistical machinery and football judgment. Poisson-based models give a clear starting point; modern data like xG and sensible adjustments for dependence and low-score bias refine that baseline into something practical. With proper calibration and attention to match context, you can produce well-calibrated 0-0 probabilities that serve betting, coaching, and analytical needs alike.
Sources and experts
- Dixon, M. J., & Coles, S. G. (1997). Modelling association football scores and inefficiencies in the football betting market.
- StatsBomb: What is expected goals (xG)?
- FiveThirtyEight: How our club soccer predictions work (SPI and model explanations)
- Poisson distribution — overview and properties (Wikipedia)


