Loading...
Loading...
Hockey is the hardest major sport to predict. Here's why 61% is better than it sounds, how we measure it honestly, and where the model is strongest.
TL;DR
A coin flip gets you 50%. Vegas closing lines hit around 58-60%. The best public hockey models land in the 59-65% range. PuckCast's 60.90% was measured on 6,560 games the model never saw during training (5-fold walk-forward validation). Confidence grades let you filter for the strongest picks, where accuracy climbs well above that 61% average.
If you could predict NHL games at 70%, you'd be the best hockey forecaster who ever lived. That ceiling barely exists in hockey the way it does in other sports. Here's why.
The puck is small, fast, and bounces off sticks, skates, and bodies in ways that are genuinely unpredictable. A deflection off a shin pad can be the difference between a win and a loss.
Most NHL games produce 5-7 combined goals. With so few scoring events, a single lucky bounce has outsized impact on the final result. Compare that to basketball where 200+ points are scored.
A goalie having an off night can tank a team that dominates possession. Save percentage swings game to game in ways that are extremely difficult to forecast in advance.
The salary cap, draft lottery, and revenue sharing are designed to keep the league competitive. In any given NHL game, the underdog has a real shot. That is the point.
Context matters. 61% sounds modest until you see where it sits relative to other benchmarks.
Random guessing. Zero skill.
Naive baseline. Home ice advantage exists but it is not enough.
The market consensus. Backed by millions of dollars in action.
6,560 games across 5 temporal folds. 16 seasons of training data.
The realistic ceiling for hockey prediction. Beyond 65% is uncharted territory.
Any model can memorize the past. The real test is whether it works on games it has never seen. That is the difference between in-sample accuracy (testing on your training data) and walk-forward accuracy (testing on future data only).
PuckCast uses 5-fold temporal cross-validation. The model trains on earlier seasons and predicts later ones. It never peeks at the future. The 60.90% number comes entirely from these held-out predictions across 6,560 games.
Why this matters
Transparency is a core value here, so let's be direct: live performance does not always match backtest numbers. PuckCast's walk-forward accuracy is 60.90% across 6,560 historical games. Live accuracy during the 2025-26 season sits at 56.4% overall.
That gap is normal. Every model faces it. Live predictions deal with injuries breaking mid-game, lineup changes announced after the model runs, and the inherent variance of small sample sizes within a single season. The good news: recent performance has been strong, with the model hitting 71% over the last 30 days as of early March 2026. Performance fluctuates, and we track it publicly on the performance page.
60.9%
Walk-forward (6,560 games)
56.4%
2025-26 live (overall)
71%
Last 30 days (Mar 2026)
The 61% number is an average across all games. But the model knows when it has a strong read and when it is basically guessing. That is where confidence grades come in.
A
The model sees a clear edge (>20 pts). Strong team advantages, favorable matchups, and aligned underlying metrics. These games hit at 71.2%, well above the 61% average.
B
Solid lean (5-20 pts). The model has a meaningful edge but not a slam dunk. Hits at 60.6%, right around the overall average.
C
Marginal edge (0-5 pts). The model has a slight preference but the signal is weak. Hits at 55.9%, closer to a coin flip.
If you only follow A-grade picks, your effective accuracy is meaningfully higher than 61%. The tradeoff is volume: fewer games qualify. Confidence grades let you decide where on the accuracy-vs-volume spectrum you want to sit.
Most public hockey prediction models do not publish walk-forward accuracy numbers. The ones that do tend to report in the 55-63% range. Some claim higher, but without disclosure of their validation methodology, those numbers are difficult to verify.
PuckCast uses a logistic regression model (C=0.01) trained on 158 features across 16 NHL seasons. We tested ensemble approaches, but the simpler model generalized better. The edge comes from feature engineering, not model complexity.
Model specs
Yes. Hockey is the most random of the four major North American sports. Vegas closing lines sit around 58-60%, and the best public models top out in the 59-65% range. PuckCast's 60.90% walk-forward accuracy places it solidly in that top tier, especially because it was measured on 6,560 games the model had never seen during training.
Walk-forward validation trains the model only on past data and tests it on future games, mimicking real-world usage. This is the honest way to measure accuracy. In-sample accuracy (testing on the same data you trained on) is always inflated and misleading. PuckCast's 60.90% number comes from 5-fold temporal cross-validation across 6,560 games.
Three main reasons: goals are rare events (typically 5-7 per game combined), so a single lucky bounce can flip the outcome. The puck is small, fast, and deflects unpredictably off sticks, skates, and bodies. And goalie performance varies significantly game to game in ways that are difficult to model in advance.
Not all predictions are created equal. PuckCast assigns confidence grades (A through C) based on how much edge the model sees. A-grade games historically hit at 71.2%, well above the 61% average, while C games sit closer to a coin flip. By focusing on high-confidence picks, you can filter for the model's strongest signals.