/Kaggle Submission · March Machine Learning Mania 2026
We predicted the winner of every possible game in both the men's and women's NCAA tournaments — 132,133 matchups total — and submitted them to a real data science competition on Kaggle.
Imagine asking a computer: "If Duke played every other team in the tournament, how often would Duke win?" We did that for every single pair of teams — not just the games that will actually happen, but every possible matchup. Instead of guessing, our computer uses real stats from this season to calculate a win percentage for each game. A prediction of 75% means "if these teams played 100 times, we think Team A wins about 75 of them."
/Model Architecture
Instead of relying on one method, we asked four different "experts" for their prediction, then blended their answers. Think of it like asking a stats nerd, a basketball scout, the selection committee, and a momentum tracker for their picks — then averaging them based on how reliable each expert is. The stats nerd (KenPom) gets the most weight because they're right most often.
KenPom Efficiency
The Stats Nerd — looks at raw scoring efficiency
How many points a team scores vs allows per 100 possessions — the #1 predictor in college basketball
Bradley-Terry
The Scout — weighs quality of wins and losses
Strength ratings computed from who beat whom — beating a good team counts more than beating a bad one
Seed-Based
The Committee — uses the seed numbers
The committee's expert judgment turned into probabilities — a 1 seed beats a 16 seed 99% of the time
Conf Tourney
The Momentum Tracker — hot teams get a bump
Teams that just won their conference tournament get a small confidence boost
132,133
Total Predictions
76.0%
Avg Confidence
12,257
Above 90% Conf
11,015
True Toss-Ups
11,100
Extreme Upsets
/Round of 64
These are the actual games happening in the tournament. The longer the bar, the more confident we are.
We're very confident in this one — Duke wins 96 out of 100 times in our model
We genuinely don't know — this is pure March Madness — Ohio St wins 54 out of 100 times in our model
Should win, but don't be shocked if they don't — St John's wins 76 out of 100 times in our model
We genuinely don't know — this is pure March Madness — Kansas wins 53 out of 100 times in our model
We genuinely don't know — this is pure March Madness — South Florida wins 50 out of 100 times in our model
Barely leaning one way — could go either way — Michigan St wins 58 out of 100 times in our model
Should win, but don't be shocked if they don't — UCLA wins 76 out of 100 times in our model
Solid favorite — but upsets happen — Connecticut wins 89 out of 100 times in our model
We're very confident in this one — Florida wins 96 out of 100 times in our model
We genuinely don't know — this is pure March Madness — Iowa wins 52 out of 100 times in our model
We genuinely don't know — this is pure March Madness — Vanderbilt wins 52 out of 100 times in our model
Solid favorite — but upsets happen — Nebraska wins 81 out of 100 times in our model
We genuinely don't know — this is pure March Madness — VCU wins 54 out of 100 times in our model
We're very confident in this one — Illinois wins 93 out of 100 times in our model
Solid favorite — but upsets happen — St Mary's CA wins 80 out of 100 times in our model
We're very confident in this one — Houston wins 93 out of 100 times in our model
Should win, but don't be shocked if they don't — Michigan wins 72 out of 100 times in our model
Should win, but don't be shocked if they don't — St Louis wins 66 out of 100 times in our model
We genuinely don't know — this is pure March Madness — Akron wins 54 out of 100 times in our model
Barely leaning one way — could go either way — Alabama wins 56 out of 100 times in our model
We genuinely don't know — this is pure March Madness — Tennessee wins 51 out of 100 times in our model
Solid favorite — but upsets happen — Virginia wins 84 out of 100 times in our model
Barely leaning one way — could go either way — Santa Clara wins 62 out of 100 times in our model
Solid favorite — but upsets happen — Iowa St wins 88 out of 100 times in our model
We're very confident in this one — Arizona wins 93 out of 100 times in our model
Barely leaning one way — could go either way — Utah St wins 63 out of 100 times in our model
We genuinely don't know — this is pure March Madness — High Point wins 55 out of 100 times in our model
Should win, but don't be shocked if they don't — Arkansas wins 74 out of 100 times in our model
Should win, but don't be shocked if they don't — BYU wins 68 out of 100 times in our model
We're very confident in this one — Gonzaga wins 96 out of 100 times in our model
Should win, but don't be shocked if they don't — Miami FL wins 77 out of 100 times in our model
Solid favorite — but upsets happen — Purdue wins 89 out of 100 times in our model
/Model Alert
In March Madness, the higher seed (like a #1) is supposed to beat the lower seed (like a #16). But our computer found some matchups where the "underdog" is actually the better team statistically. The selection committee may have under-seeded them — our model disagrees with the committee's ranking. These are the games where we're predicting an upset.
/Elite Matchups
The four #1 seeds are supposed to be the best teams in the country. But even among the elite, our model has clear favorites. If Duke played Arizona, who wins? These predictions show which #1 seed our model thinks is actually the strongest — and by how much.
Duke
75% to win
Duke
70% to win
Duke
64% to win
Arizona
61% to win
Arizona
59% to win
Michigan
54% to win
/High Confidence
/Coin Flip
/Guide
When we say 75%, imagine the two teams playing 100 games. We think Team A wins about 75 of them. The higher the number, the more confident we are. Even a 90% pick loses 1 out of 10 times — that's why March Madness is so exciting.
The colored bars show how confident we are visually. A long teal bar = strong favorite. Two equal-length bars = coin flip. A pink bar on the underdog = we're predicting an upset. The longer the bar, the more confident the pick.
These are games where our computer disagrees with the "experts" (the selection committee). The committee ranked Team A higher, but our stats say Team B is actually better. This is where the fun is — will our data beat the experts?
Kaggle scores us based on how calibrated our predictions are. If we say 70%, roughly 70% of those games should actually go our way. Being confidently wrong gets punished hard — being cautious is safer but won't win the competition.