doclenz Posted September 8, 2004 Posted September 8, 2004 The Massey ratings. SDSU and NDSU posters were writing about how by using the Massey ratings they could play with so many and beat most of them. This weeks Massey ratings: UND-127 UNI -130 NDSu-138 Winona-160 St. Cloud-167 SDSU-178 According to this and they were legitimate last year....UND is the best in the area! Quote
Let'sGoHawks! Posted September 8, 2004 Posted September 8, 2004 I remember when this was an issue earlier this year...nice to see the Sioux are up there. Quote
Let'sGoHawks! Posted September 8, 2004 Posted September 8, 2004 I see Grand Valley is 73rd, this is the highest ranking of any non Division I-A team, and they are ahead of many I-A teams. Pretty impressive, but the season is young. Quote
IowaBison Posted September 8, 2004 Posted September 8, 2004 sounds good, call up gene and schedule a game Quote
bigmrg74 Posted September 8, 2004 Posted September 8, 2004 Oh God!! Not another Fan of the Massey!!! Even if it does make my team look very favorable, I still hate the thing. If any of you are on the D2 board and listen to Bulldog sing the praises of Massey, you would know what I'm talking about. Quote
Let'sGoHawks! Posted September 8, 2004 Posted September 8, 2004 I, along with most people, do not put much stock in computer rankings of any sort. Luckily, in Division II, along with all lower divisions, we don't have to worry about these rankings deciding champions. They provide an interesting conversation piece, but things are decided on the field like they are supposed to be. Quote
jimdahl Posted September 8, 2004 Posted September 8, 2004 I have the same problem with the Massey ratings now as I did last Spring -- the all-division matrix of game outcomes is way too sparse interdivisionally for meaningful analysis. Many of us Sioux fans start think about statistical ratings from a Bradley-Terry perspective: there are only about 20 interactions between the DII cluster of teams and the DIAA cluster of teams; those interactions are also overwhelmingly won by DIAA so you can't form a win-tie path from each team to each other. Massey attempts to work around that sparseness problem by not just analyzing wins/losses, but the score of each contest. UND was a team that, famously, last year played to the level of each opponent, yet they still managed to win; is a team that runs up the score really better? Massey does correct for that with what he calls a Bayesian Correction, which basically reduces the importance of score differential as more win-loss information becomes available. Right now, when win-loss information is sparse, UND is being rewarded massively for blowing out Mesa State. As more win-tie connections are made throughout the matrix, the scores will become less important and Massey's ratings will come to resemble Bradley-Terry, though there will always be the score differential kludge linking parts of the matrix that aren't naturally linked by a win-tie path. Quote
Bisonfan1234 Posted September 8, 2004 Posted September 8, 2004 I have the same problem with the Massey ratings now as I did last Spring -- the all-division matrix of game outcomes is way too sparse interdivisionally for meaningful analysis. Many of us Sioux fans start think about statistical ratings from a Bradley-Terry perspective: there are only about 20 interactions between the DII cluster of teams and the DIAA cluster of teams; those interactions are also overwhelmingly won by DIAA so you can't form a win-tie path from each team to each other. Massey attempts to work around that sparseness problem by not just analyzing wins/losses, but the score of each contest. UND was a team that, famously, last year played to the level of each opponent, yet they still managed to win; is a team that runs up the score really better? Massey does correct for that with what he calls a Bayesian Correction, which basically reduces the importance of score differential as more win-loss information becomes available. Right now, when win-loss information is sparse, UND is being rewarded massively for blowing out Mesa State. As more win-tie connections are made throughout the matrix, the scores will become less important and Massey's ratings will come to resemble Bradley-Terry, though there will always be the score differential kludge linking parts of the matrix that aren't naturally linked by a win-tie path. Statistics?! What kind of useless garbage is that? Why don't you go try and fit a line to a scatter plot? Quote
jimdahl Posted September 8, 2004 Posted September 8, 2004 Statistics?! What kind of useless garbage is that? Why don't you go try and fit a line to a scatter plot? Sorry, didn't mean to confuse the NDSU-educated. I don't share the opinion that all sports statistics, including Massey's ratings, are "useless garbage". Rather, I was just warning that we should bear in mind how ratings are formulated so we know what relative ratings actually mean. The one element of Massey's ratings that I forgot to mention was that it uses last season's results as a beginning baseline, so UND's rating is also being elevated for last season's performance. Quote
BisSioux Posted September 8, 2004 Posted September 8, 2004 Other noteworthy team's ratings: University of Mary 283 Montana Tech 363 Valpo 374 Quote
Bisonfan1234 Posted September 9, 2004 Posted September 9, 2004 Sorry, didn't mean to confuse the NDSU-educated. I don't share the opinion that all sports statistics, including Massey's ratings, are "useless garbage". Rather, I was just warning that we should bear in mind how ratings are formulated so we know what relative ratings actually mean. The one element of Massey's ratings that I forgot to mention was that it uses last season's results as a beginning baseline, so UND's rating is also being elevated for last season's performance. I was joking. Quote
bisonguy Posted September 9, 2004 Posted September 9, 2004 Massey's rankings, along with most other "rankings", are pretty much useless until five or six games into the season. Not enough data. Quote
airmail Posted September 9, 2004 Posted September 9, 2004 Other noteworthy team's ratings: University of Mary 283 Montana Tech 363 Valpo 374 You made my day. Thanks! Quote
IowaBison Posted September 9, 2004 Posted September 9, 2004 Here are some facts, not stats, not ratings NDSU DIAA Valpo DIAA UND D2 Mary NAIA Montana Tech NAIA (painful for us Bison Fans, but True) This is what most people see, then they look at the rankings, the conference standings, etc. Only nerds like me look at ratings. Jim, I really enjoyed the background on the Massey that you did. Quote
NDSU grad Posted September 9, 2004 Posted September 9, 2004 What I've never understood is how people say they use statistics to figure out these ratings. How can you gain any precision in these ratings when the games aren't replicated? Perhaps I know just enough about statistics to ask really stupid questions, I don't know. Quote
Bisonfan1234 Posted September 9, 2004 Posted September 9, 2004 Polls are for entertainment. The only way to know is to play the game. Quote
NDSU grad Posted September 9, 2004 Posted September 9, 2004 Polls are for entertainment. The only way to know is to play the game. Thanks for the wonderful insight . Massey's not a poll, it's a rating system, and I'm just curious as to how it's computed with any sort of precision. Quote
The Sicatoka Posted September 9, 2004 Posted September 9, 2004 It's not the system Massey uses, but it's the one that seems to have the best track record of predicting probable outcomes. http://siouxsports.com/hockey/rankings/help/krach.htm Quote
jimdahl Posted September 9, 2004 Posted September 9, 2004 What I've never understood is how people say they use statistics to figure out these ratings. How can you gain any precision in these ratings when the games aren't replicated? Perhaps I know just enough about statistics to ask really stupid questions, I don't know. Ignoring the sparseness problem for now, the link Sicatoka provided includes a brief description and links to more in-depth descriptions of the Bradley-Terry technique. It was developed independently in a few different instances to perform contest judging when a full round-robin isn't possible (e.g. dog show, or food competition). By including each entrant in a minimum number of pair comparisons and with no unconnected pockets in the comparison matrix, you can then use the technique to infer what would have happened had two uncompared items been compared. Of course, sporting events are performance-based so include an element of randomness that makes the technique non-perfect, but it's still a great place to start. The problem to which I think you were alluding is the sparseness issue I was describing earlier. We don't usually find the Bradley-Terry results to be predictive until each team has about 15 hockey games in the can (a good 25% of which are usually inter-conference, creating those crucial linkages). To compensate for the shortness of the football season and the desire for immediate ratings, Massey includes two additional techniques: starts with a base of last year's ratings (which diminishes in weight as real results come in), and uses scores to give more precision about team strength rather than just wins/losses (which also diminishes as more results come in, and which prevents those rankings from being part of the BCS because score-based ratings are considered sketchy by the NCAA). Quote
jimdahl Posted September 9, 2004 Posted September 9, 2004 For instance, in agronomy research I can use ANOVA to see how precise my results are (p values, LSD's, confidence intervals, standard errors, etc.)Â Is there any way to "fit" your predicted outcomes with actual outcomes? First, I'm not by any stretch a statistician -- I'm just interested in sports ratings and learned enough to implement the ratings on this site: (PWR, RPI, KRACH). However, analysis of variance strikes me as the wrong path because I tend to think of that as useful for comparing numerous factors between independent groups (though a Bison fan could use ANOVA to compare the power of the divisions to prove that they're really different!) Because Massey's technique uses scores rather outcomes, the model doesn't perfectly fit past results so it is derived via a maximum likelihood estimate, the best indiciation of error of which is the standard deviation (which Massey does provide). He says 68% of observed game results will fall within one deviation of the expected result. Quote
jimdahl Posted September 13, 2004 Posted September 13, 2004 Massey's football ratings look a lot better interdivisionally now (not surprising, since last weekend featured a lot more interactions between divisions, including a few DII wins over DIAA). The top teams in IAA are bubbling up above the top of II, as they undoubtedly should be. UND still hanging strong at 120, #2 in DII. Quote
Recommended Posts
Join the conversation
You can post now and register later. If you have an account, sign in now to post with your account.