|
|
|
![]() |
|
|||||||
|
||||||||
![]() |
| Thread Tools |
Rating:
|
Display Modes |
|
#61
|
||||
|
||||
|
Re: "standard error" of OPR values
Quote:
I could be wrong, but I doubt this is what Citrus Dad had in mind. Can we all agree that 0.1 is real-world meaningless? There is without a doubt far more variation in consistency of performance from team to team. Manual scouting data would surely confirm this. @ Citrus Dad: you wrote: Quote:
|
|
#62
|
|||||
|
|||||
|
Re: "standard error" of OPR values
Very interesting results. I wonder if you could run the same analysis on the 2015 Waterloo Regional. The reason I'm asking for that event in particular is because it had the ideal situation for OPR: high matches per team (13) and a small number of teams (30).
|
|
#63
|
||||
|
||||
|
Re: "standard error" of OPR values
Quote:
For example, this means that if a team had, say, an OPR of 50, that if they were in another identical tournament with the same matches and randomness in the match results, that the OPR computed from that tournament would probably be between 39 and 61 (if you're being picky, 68% of the time the score would lie in this range if the data is sufficiently normal or Gaussian). So picking a team for your alliance that has an OPR of 55 over a different team that has an OPR of 52 is silly. But picking a team that has an OPR of 80 over a team that has an OPR of 52 is probably a safe bet. ![]() In response to the latest post, this could be run on any other tournament for which the data is present. Ether made this particularly easy to do by providing the A match matrix and the vector of match results in nice csv files. BTW, the code is attached and scilab is free, so anybody can do this for whatever data they happen to have on hand. Last edited by wgardner : 18-05-2015 at 13:46. |
|
#64
|
||||
|
||||
|
Re: "standard error" of OPR values
Quote:
http://www.chiefdelphi.com/media/papers/3132 |
|
#65
|
||||
|
||||
|
Re: "standard error" of OPR values
Quote:
Here are the results for the Waterloo tournament: mpt = matches per team (so the last row is for the whole tournament and earlier rows are for the tournament through 4 matches per team, through 5, etc.) varM = variance of the match scores stdevM = standard deviation of the match scores varR and stdevR are the same for the match prediction residual so varR/varM is the fraction of the match variance that can't be predicted by the OPR linear prediction model. /sqrt(mpt) = the standard deviation of the OPRs we would have if we were simply averaging a teams match score to estimate their OPR, which is just stdevR/sqrt(mpt) StdErrO = the standard error of the OPRs using my complicated model derivation. stdevO = the standard deviation of the StdErrO values taken across all teams, which is big if some teams have more standard error on their OPR values than other teams do. Code:
mpt varM stdevM varR stdevR /sqrt(mpt) StdErrO stdevO 4 3912.31 62.55 206.90 14.38 7.19 12.22 1.60 5 4263.97 65.30 290.28 17.04 7.62 10.44 0.71 6 3818.40 61.79 346.49 18.61 7.60 9.44 0.43 7 3611.50 60.10 379.83 19.49 7.37 8.64 0.30 8 3617.25 60.14 429.42 20.72 7.33 8.28 0.17 9 3592.06 59.93 469.44 21.67 7.22 8.00 0.11 10 3623.44 60.20 539.33 23.22 7.34 8.01 0.10 11 3530.91 59.42 548.08 23.41 7.06 7.58 0.08 12 3440.36 58.65 578.65 24.06 6.94 7.38 0.07 13 3356.17 57.93 645.25 25.40 7.05 7.42 0.06 Code:
mpt varM stdevM varR stdevR /sqrt(mpt) StdErrO stdevO 4 1989.58 44.60 389.80 19.74 9.87 16.51 1.28 5 2000.09 44.72 714.81 26.74 11.96 16.31 0.57 6 2157.47 46.45 863.88 29.39 12.00 15.17 0.37 7 2225.99 47.18 916.16 30.27 11.44 13.64 0.29 8 2204.03 46.95 985.63 31.39 11.10 12.77 0.24 9 2235.14 47.28 1053.26 32.45 10.82 12.21 0.10 10 2209.46 47.00 1056.14 32.50 10.28 11.37 0.12 The OPR seems to do a much better job of predicting the match results in the Waterloo tournament (removing 80% of the match variance vs. 50% in Archmedes), and the standard deviation of the OPR estimates themselves is less (7.42 in Waterloo vs. 11.37 in Archimedes). Last edited by wgardner : 18-05-2015 at 20:01. |
|
#66
|
||||
|
||||
|
Re: "standard error" of OPR values
Quote:
To be honest setting up a pooled time series with this data would take me more time than I have at the moment. I've thought about it and maybe it will be a summer project (maybe my son Jake (themccannman) can do it!) Note that the 1 SD SE of 11.5 is the 68% confidence interval. For 10 or so observations, the 95% confidence interval is about 2 SD or about 23.0. The t-statistic is the relevant tool for finding the confidence interval metric. Last edited by Citrus Dad : 19-05-2015 at 15:09. Reason: Add note about confidence intervals |
|
#67
|
||||
|
||||
|
Re: "standard error" of OPR values
Quote:
From this and other prior statements, I had the very strong impression you were seeking a separate error estimate for each team's OPR. Such estimates would certainly not be virtually identical for every team! It would be very helpful if you would please provide more information about statistical software packages you know that provide "parameter standard errors". I couldn't find any that could provide such estimates for the multiple-regression model we are talking about for OPR computation using FRC-provided match score data. I suspect that's because it's simply not possible to get such estimates for that model and data. |
|
#68
|
||||
|
||||
|
Re: "standard error" of OPR values
Quote:
Note that this is computing the confidence of each OPR estimate for each team. This is different from trying to compute the variance of score contribution from match to match for each team, which is a very different (and also very interesting) question. I think it would be reasonable to hypothesize that the variance of score contribution for each team might vary from team to team, possibly substantially. For example, it might be interesting to know that team A scores 50 points +/- 10 points with 68% confidence but team B scores 50 points +/- 40 points with 68% confidence. At the very least, if you saw that one team had a particularly large score variance, it might make you investigate this robot and see what the underlying root cause was (maybe 50% of the time they have an awesome autonomous but 50% of the time it completely messes up, for example). Hmmm.... |
|
#69
|
||||
|
||||
|
Re: "standard error" of OPR values
Quote:
Manual scouting data would surely confirm this. Consider the following thought experiment. Team A gets actual scores of 40,40,40,40,40,40,40,40,40,40 in each of its 10 qual matches. Team B gets actual scores of 0,76,13,69,27,23,16,88,55,33 The simulation you described assigns virtually the same standard error to their OPR values. If what is being sought is a metric which is somehow correlated to the real-world trustworthiness of the OPR for each individual team (I thought that's what Citrus Dad was seeking), then the standard error coming out of the simulation is not that metric. My guess is that the 0.1 number is just measuring how well your random number generator is conforming to the sample distribution you requested. Last edited by Ether : 19-05-2015 at 22:05. |
|
#70
|
||||
|
||||
|
Re: "standard error" of OPR values
Quote:
Your model certainly might be valid, and my derivation explicitly does not deal with this case. The derivation is for a model where OPRs are computed, then multiple tournaments are generated using those OPRs and adding the same amount of noise to each match, and then seeing what the standard error of the resulting OPR estimates is across these multiple tournaments. If you know that the variances for each team's score contribution are different, then the model fails. For that matter, the least squares solution for computing the OPRs in the first place is also a failed model in this case. If you knew the variances of the teams' contributions, then you should use weighted-least-squares to get a better estimate of the OPRs. I wonder if some iterative approach might work: First compute OPRs assuming all teams have equal variance of contribution, then estimate the actual variances of contributions for each team, then recompute the OPRs via weighted-least-squares taking this into account, then repeat the variance estimates, etc., etc., etc. Would it converge? [Edit: 2nd part of post, added here a day later] http://en.wikipedia.org/wiki/Generalized_least_squares OPRs are computed with an ordinary-least-squares (OLS) analysis. If we knew ahead of time the variances we expected for each team's scoring contribution, we could use weighted-least-squares (WLS) to get a better estimate of the OPRs. The link also describes something like I was suggesting above, called "Feasible generalized least squares (FGLS)". In FGLS, you use OLS to get your initial OPRs, then estimate the variances, then compute WLS to improve the OPR estimate. It discusses iterating this approach also. But, the link also includes this comment: "For finite samples, FGLS may be even less efficient than OLS in some cases. Thus, while (FGLS) can be made feasible, it is not always wise to apply this method when the sample is small." If we have 254 match results and we're trying to estimate 76 OPRs and 76 OPRvariances (152 parameters total), we have a pretty small sample size. So this approach would probably suffer from too small of a sample size. Last edited by wgardner : 20-05-2015 at 07:20. |
|
#71
|
||||
|
||||
|
Re: "standard error" of OPR values
See also this link:
http://en.wikipedia.org/wiki/Heteroscedasticity "In statistics, a collection of random variables is heteroscedastic if there are sub-populations that have different variabilities from others. Here "variability" could be quantified by the variance or any other measure of statistical dispersion." And see particularly the "Consequences" section which says, "Heteroscedasticity does not cause ordinary least squares coefficient estimates to be biased, although it can cause ordinary least squares estimates of the variance (and, thus, standard errors) of the coefficients to be biased, possibly above or below the true or population variance. Thus, regression analysis using heteroscedastic data will still provide an unbiased estimate for the relationship between the predictor variable and the outcome, but standard errors and therefore inferences obtained from data analysis are suspect. Biased standard errors lead to biased inference..." |
|
#72
|
||||
|
||||
|
Re: "standard error" of OPR values
Quote:
|
|
#73
|
|||
|
|||
|
Re: "standard error" of OPR values
Quote:
|
|
#74
|
||||
|
||||
|
Re: "standard error" of OPR values
Quote:
|
|
#75
|
|||
|
|||
|
Re: "standard error" of OPR values
Quote:
Last edited by sur : 24-05-2015 at 17:48. |
![]() |
| Thread Tools | |
| Display Modes | Rate This Thread |
|
|