|
|
|
| Love is in the air compressor. |
![]() |
|
|||||||
|
||||||||
![]() |
| Thread Tools |
Rating:
|
Display Modes |
|
#91
|
|||||
|
|||||
|
Re: "standard error" of OPR values
Quote:
The elephant in the room here is that assumption that the alliance is equal to the sum of its members. For example, consider a 2015 (Recycle Rush) robot with a highly effective 2-can grab during autonomous, and the ability to build, score, cap and noodle one stack of six from the HP station, or cap five stacks of up to six totes during a match, or cap four stacks with noodles loaded over the wall. For argument's sake, it is essentially 100% proficient at these tasks, selecting which to do based on its alliance partners. I will also admit up front that the alliance match-ups are somewhat contrived, but none truly unrealistic. If I'd wanted to really stack the deck, I'd have assumed that the robot was the consummate RC specialist and had no tote manipulators at all.
The real point is that this variation is based on the alliance composition, not on "performance variation" of the robot in the same situation. I also left HP littering out, which would provide additional wrinkles. My takeaway on this thread is that it would be good and useful information to know the rms (root-mean-square) of the residuals for an OPR/DPR data set (tournament or season). This would provide some understanding as to how much difference really is a difference, and a clue as to when the statistics mean about as much as the scouting. On another slightly related matter, I have wondered why CCWM (Combined Contribution to Winning Margin) is calculated by combining separate calculations of OPR and DPR, rather than by solving a single matrix of winning margin. I suspect that the single calculation would prove to be more consistent for games with robot-based defense (not Recycle Rush); if a robot plays offense five matches and defense five matches, then both OPR and DPR would each have a lot of noise, whereas true CCWM should be a more consistent number. Last edited by GeeTwo : 12-07-2015 at 22:56. Reason: Several nitnoids |
|
#92
|
|||
|
|||
|
Re: "standard error" of OPR values
Quote:
Last edited by Oblarg : 13-07-2015 at 00:26. |
|
#93
|
||||
|
||||
|
Re: "standard error" of OPR values
Quote:
We just tried this about a day ago. Unfortunately, there isn't enough data in a typical tournament to get reliable estimates of the per-team offensive variation. With much larger tournament sizes, it does work OK, but it doesn't work when you only have about 5-10 matches played by each team. I'll send you some of our private messages where this is discussed and where the results are shown.Last edited by wgardner : 13-07-2015 at 06:32. |
|
#94
|
||||
|
||||
|
Re: "standard error" of OPR values
Quote:
|
|
#95
|
||||
|
||||
|
Re: "standard error" of OPR values
Quote:
The paper discusses MMSE-based estimation of the metrics (as opposed to the traditional least-squares method) which reduces the overfitting effects, does better at predicting previously unseen matches (as measured by the size of the squared prediction residual in "testing set" matches), and is better at predicting the actual underlying metric values in tournaments which are simulated using the actual metric models. |
|
#96
|
|||
|
|||
|
Re: "standard error" of OPR values
Quote:
Rather, it would be overfitting if the predictive power of the model (when tested against data not used to tune it) did not increase with the amount of data available to tune the parameters. I highly doubt that is the case here. Last edited by Oblarg : 13-07-2015 at 22:01. |
|
#97
|
|||||
|
|||||
|
Re: "standard error" of OPR values
Quote:
|
|
#98
|
||||
|
||||
|
Re: "standard error" of OPR values
Quote:
On the first sentence of that quote, I previously found that if I replaced the data from the 2014 casa tournament (which had the greatest number of matches per team of the tournaments I worked with) with completely random noise, the OPR could "predict" 26% of the variance and WMPR could "predict" 47% of it. So they're clearly describing the random noise in this case where a "properly fit" model would come closer to finding no relationship between the model parameters and the data, as should be the case when the data is purely random. On the second sentence, again for the 2014 casa tournament, the OPR calculation only has 4 data points per parameter and the WMPR only has 2, which again sounds like "having too many parameters relative to the number of observations" to me. BTW, I think the model is appropriate, so I view it more as a problem of having too few observations rather than too many parameters. And again, the casa tournament is one of the best cases. Most other tournaments have even fewer observations per parameter. So that's why I think it's overfitting. Your opinion may differ. No worries either way. ![]() This is also discussed a bit in the section on "Effects of Tournament Size" on my "Overview and Analysis of First Stats" paper. Last edited by wgardner : 14-07-2015 at 09:25. |
|
#99
|
|||
|
|||
|
Re: "standard error" of OPR values
Quote:
The problem here is that there are two separate things in that wikipedia article that are called "overfitting:" errors caused by fundamentally sound models with insufficient data, and errors caused by improperly revising the model specifically to fit the available training data (and thus causing a failure to generalize). If one is reasoning purely based on patterns seen in the data, then there is no difference between the two (since the only way to know that one's model fits the data would be through validation against those data). However, these aren't necessarily the same thing if one has an externally motivated model (and I believe OPR has reasonable, albeit clearly imperfect, motivation). We may be veering off-topic, though. |
![]() |
| Thread Tools | |
| Display Modes | Rate This Thread |
|
|