|
|
|
![]() |
|
|||||||
|
||||||||
![]() |
|
|
Thread Tools | Rate Thread | Display Modes |
|
|
|
#1
|
||||
|
||||
|
paper: Comparison of Statistical Prediction Models
Thread created to discuss this paper.
This paper provides a comparison of common statistical prediction methods in order to determine which methods have the most predictive power. To see each model's predictions for each match during the period 2008-2016, as well as each team's rating before each match during this period, go to its corresponding workbook. The "Data Summary and Methodology" workbook contains details on each model, a FAQ, a summary of predictive capabilities of each model, and a side-by-side comparison of each model for the year 2016. I am continuing on my journey of building a predictive model for the 2017 season. Here, I compared a bunch of different predictive methods to determine where my efforts will be best spent. The extremely short version is that, in order of most predictive to least predictive, we have: Calculated Contribution to Score (OPR) WM Elo Average Score Calculated Contribution to WM (CCWM) Average WM Calculated Contribution to Win Adjusted Winning Record I was surprised how predictive average score was, and generally how similar the "average" methods were with the "calculated contribution" methods. Moving forward, I am planning to continue development of WM Elo and Calculated Contribution to Score methods, and take some kind of weighted average of those two. |
|
#2
|
|||||
|
|||||
|
Re: paper: Comparison of Statistical Prediction Models
FYI, TBA has predictions for all 2016 official events, not just champs. They're just not publicly linked. Append /insights to the end of any event, like https://www.thebluealliance.com/event/2016casj/insights
|
|
#3
|
|||||
|
|||||
|
Re: paper: Comparison of Statistical Prediction Models
Nice work! Please share results, when you can, on your effort to improve predictions using a weighted combination of OPR and ELO.
For now, we can all continue to curse OPR as the worst way to rank teams, except for all the other ways. ![]() |
|
#4
|
|||||
|
|||||
|
Re: paper: Comparison of Statistical Prediction Models
I have a couple of questions after looking at it more deeply.
- Isn't best linear combination kind of cheating? It seems like you're picking weights after knowing the outcome. - Do you have any high-level takeaways for what you think will contribute to a better predictive model moving forward? (Assuming the 2017 game isn't super weird) Cool stuff overall, thanks for sharing! |
|
#5
|
||||
|
||||
|
Re: paper: Comparison of Statistical Prediction Models
Quote:
I wanted to know how much "better" the models could get by adding two or more models together. The result was that it has reasonable, but not extremely high value, so it is worth more investigation. I wanted to know how "different" each model's predictions were to each other. If two models both have reasonable predictive power, and their correct predictions are uncorrelated, taking a combination of the two will provide much better predictions than either could provide individually. It turned out that the good model's all made pretty similar predictions. But yes, it is cheating in that I knew the results as I was finding the best linear combination. In comparison, almost everything else I did was tuned using the period 2012-2014, so the 2016 predictions are actually true predictions for the other models. Quote:
My biggest takeaway though was best summarized by Richard Wallace above. |
|
#6
|
||||
|
||||
|
Re: paper: Comparison of Statistical Prediction Models
Quote:
https://www.chiefdelphi.com/forums/s....php?p=1483963 Also, take a look at Eugene's very interesting work here: https://github.com/the-blue-alliance...ment-210564302 |
|
#7
|
||||
|
||||
|
Re: paper: Comparison of Statistical Prediction Models
Quote:
Quote:
|
![]() |
| Thread Tools | |
| Display Modes | Rate This Thread |
|
|