NBA front offices spend millions on scouts, interviews, workouts, medical evaluations, and film study. This model sees none of that — just college box scores. It still drafts better.
Comparing the model's top 10 college players against the actual top 10 draft picks each year from 2015 to 2022. NBA teams select from college, international, and G League pools — the model only sees NCAA stats.
For each draft from 2015 to 2022, the model ranks its top 10 college players. We compare the career cumulative VORP of those 10 players against the actual top 10 picks made by NBA teams. All VORP data from Basketball Reference.
| Year | Model Top 10 | GM Top 10 | Δ |
|---|---|---|---|
| 2015 | 77.8 | 71.4 | +6.4 |
| 2016 | 102.2 | 76.0 | +26.2 |
| 2017 | 93.9 | 62.4 | +31.5 |
| 2018 | 94.0 | 104.1 | −10.1 |
| 2019 | 51.3 | 33.0 | +18.3 |
| 2020 | 56.3 | 37.2 | +19.1 |
| 2021 | 44.6 | 49.6 | −5.0 |
| 2022 | 29.6 | 13.2 | +16.4 |
| Total | 549.7 | 446.9 | +102.8 |
+17% more value: 524.5 ÷ 446.9 = 1.174. The model's top-10 picks generated 17% more career VORP than the actual top-10 draft picks across 8 drafts.
18% bust rate: A bust is defined as negative career VORP. 14 of 80 model top-10 picks finished with career VORP below zero (18%), compared to 24 of 80 actual top-10 picks (30%).
28 players found: 28 players ranked in the model's top 10 were picked outside the actual top 10 and went on to produce positive career VORP.
The model and scout consensus boards capture different signals. Adding a small dose of consensus — just 16% — squeezes out a bit more value.
Scout consensus boards were compiled from Rookie Scale, NBA Mock Draft Database, Hoops Prospects, and cross-referenced contemporary ESPN/SI/NBA.com boards for 2015–2022. International and G League players were filtered out so the comparison is college-only.
Blended rank is computed as w × model_rank + (1−w) × consensus_rank for each college player. Players missing from one list are assigned a penalty rank of 40. The top 10 by blended rank are selected and their career cumulative VORP is summed.
Optimal weight was found by testing every integer percentage from 0–100. The best single weight is 84% model / 16% consensus (553.5 VORP), with a broad plateau from 24–93% model weight where performance stays within 5 VORP of the peak. The model carries the strategy — a small dose of scout signal adds marginal value by catching a few players the model underranks.
Ranked in the model's top 10 but selected outside the lottery by actual NBA teams.












The model's top 10 using only pre-draft college stats. VORP (Value Over Replacement Player) measures a player's total contribution compared to a baseline bench-level player — higher means more cumulative impact over their career so far.
Trained on Torvik advanced college stats (2010–2026) supplemented with international Kaggle datasets. 144 engineered features span shooting, creation, defense, trajectory, team context, and positional archetypes inspired by EuroLeague clustering research. Target is 3-year cumulative NBA VORP.
Walk-forward — for each draft year, the model trains only on prior years. No future data ever enters a prediction. Every configuration is evaluated against real career outcomes across all draft classes (2015–2022).
Continuous probabilistic position classification with post-prediction calibration. Point-forward detection prevents tall playmakers from being penalized as traditional bigs.
No access to film, interviews, medicals, or combine data. Can't evaluate character, coachability, or team fit — real edges that GMs have and the model doesn't. Only evaluates college players, meaning it competes against GM boards that also include international and G League prospects.
150–200 iterations per model. No human in the loop.
A custom hill-climbing algorithm runs the full model pipeline in a loop. Each iteration mutates the configuration — feature inclusion across 200+ candidates, feature weights, position calibration thresholds, regularization parameters, scoring cutoffs, position-specific adjustments, senior discount values — then trains an XGBoost model using walk-forward cross-validation and scores it against actual NBA career outcomes.
Configurations that improve the fitness score survive. Everything else is discarded. Each model converges in 150–200 autonomous iterations, arriving at a feature set and calibration parameters no human selected. Across nine model iterations, correlation improved from ~0.25 to 0.65+.