5/14/2016 – In his first article studying the ratings disparities of juniors compared to adults, Ganesh Viswanath had studied the causal factors that might be responsible. In this second article, he posits, "I find that on average juniors are statistically underrated by up to 50 Elo rating points against adult players, and this bias increases with the age difference." Here is his statistical study.

An interesting phenomenon of the Elo rating system is that when examining game outcomes for a group of juniors, they typically overperform when they are lower rated. Using statistical methods based on game results for junior-junior, junior-adult and adult-adult pairings, I find that on average juniors are statistically underrated by up to 50 Elo rating points against adult players, and this bias increases with the age difference. Rating systems that have potential in solving the bias and noise elements are a momentum based system or a Glicko system.

Continuing on from my findings in my previous article(1), I have expanded my dataset to include the age of both players at the time their game is played using Chessbase’s Megabase 2016(2). This is useful as by identifying the opponents’ age I can now examine the game results of 3 groups, junior-junior, junior-adult and adult-adult pairings(3). I hypothesize that junior-adult pairings should show a systematic bias in that lower rated juniors are overperforming relative to the theoretical Elo formula. This is intuitive as junior players are often making large strides in their development and consequently their current rating lags their true rating by a significant margin. I compute the conditional expectation function (CEF) for each pairing group, which is the expected score conditional on different levels of the rating difference between the two players, and is from the perspective of the junior player for junior-adult pairings (Figure 1). Lower rated juniors perform significantly better than the Elo formula suggests, and this result is more acute for junior-adult pairings. In contrast, adult-adult pairings more closely correspond to the Elo prediction.

Figure 1: CEF for junior-junior, junior-adult and adult-adult pairings

To substantiate the age bias, I can filter my sample of junior-adult pairings into sub-samples of pairings with age differences greater than 10 and 20 years of age. The age bias increases with the age gap in the data (Figure 2), where the probability the junior wins increases by more with a larger age gap. Older players are typically past their peak and so any systematic overrated bias of older players will substatinate this effect.

There is a simple way to incorporate the empirical results to recalibrate the rating formula(4). To start off we need to distinguish between the noise aspect of the results and the systematic undervaluation or bias. The theoretical Elo formula is calibrated so that σ is standardized to 200 rating points for all players(5). This assumption is clearly false in light of the data. One way to make the slope of the Elo prediction flatter is to increase rating variance. In contrast, a bias term will shift the Elo prediction up. I can then calibrate the rating variance and bias parameters to match the CEF for the three groups of pairings.

Figure 2: CEF for Junior-adult pairings with different age gaps

By Elo’s assumption, a player’s rating is normally distributed with mean μ and standard deviation σ. Assuming juniors and adults are distinct groups, the junior’s rating is normally distributed , and the adult’s rating follows the distribution, . Conceptually, the bias and noise terms can be captured by imposing a higher variance parameter and a bias for the junior player. The expected score a junior scores against an adult is given by the following formula, where Φ is the cumulative distribution function of the normal distribution.

I have devised a two-step procedure to estimate the variance and bias parameters for juniors.

- Using the data for junior-junior pairings to estimate by non-linear least squares of equation
*F*=*Score*−Φ , where*Score*is a vector of junior-junior pairing results and*RD*is vector of rating differences for each game result. The same procedure is applied to estimate using adult-adult pairings. - Using the data for junior-adult pairings to estimate by non-linear least squares of equation,
*F*=*Score*−Φ , where*Score*is now a vector of junior-adult game outcomes, and and are point estimates from step 1.

The results of this exercise are =276, =238 and =48. This suggests that on average the true rating of juniors are 48 Elo rating points above their reported rating. The 95% confidence intervals(6) are tightly clustered around the point estimates, and is [275,277] for , [237,239] for and [48,50] for the bias parameter. Plotting the recalibrated Elo formula matches the junior-adult, junior-junior and adult-adult pairings much more closely, as seen in Figure 3.

Figure 3: Recalibrated Elo formula matches CEF for junior-adult and junior-junior pairings

A simplifying assumption of the estimation is treating juniors and adults as two distinct, homogenous groups, when in reality there is a lot of heterogeneity in and . For example, if we select the subsample of adult-junior pairings with age differences greater than 20, the bias increases to 85 rating points, and when looking at an age difference greater than 30 years of age, the bias increases to 105 points. Now that we know there is a systematic bias in junior ratings, what are some potential solutions?

There are examples of current rating systems that can deal with these rating discrepancies. To address bias, a momentum rating system which weights recent games of a player would be effective. For example, if there is a statistically significant difference between a player’s performance rating and their current rating after a threshold number of games, the expected score can be modified to take into account this performance rating instead of the player’s actual rating. An alternative rating system is the Glicko rating system devised by statistician Mark Glickman(7) and is currently used by the Australian Chess Federation. Using games played during a rating period his algorithm estimates a player volatility parameter that attempts to estimate each player’s individual σ to then calculate the expected score for rating calculations. The system was originally intended to address the issue of a player coming back to chess after a number of years, with their player variance increasing over time, however it can be generalized to situations where players can be underrated due to rapid improvement. Another advantage of this system is that rating changes do not have to be symmetric, as the K factor is now proportional to player variance. These solutions are by no means exhaustive and I believe there is more work to be done in terms of optimizing rating systems to address the empirical concerns in this article.

(1) See http://en.chessbase.com/post/Elo-rating-system-how-underrated-are-the-kids

(2) The only way I could do this was to create a database of players born in each year and record their set of game results. I then merge databases for each year to then identify the opponent’s age. The sample size for each group are 574,094 games for junior-adult pairings, 358,445 games for junior-junior pairings, and 317,113 games for adult-adult pairings.

(3) Juniors are defined under 18 years of age. I have only included subsample of players born in 1960 onwards in my analysis.

(4) I am aware there are numerous methods to use the data in Figures 1 and 2 to statistically estimate the age effect. For example, a different way to estimate the age bias is by logistic regressions with an age dummy as well as interacting age with the rating difference. My main concern is that it is hard to isolate the age bias from the noise in the data, and the rating formula you derive from a logistic regression deviates significantly from a more standard Elo formula.

(5) The Elo formula uses a logistic function and so slightly deviates from the Normal cumulative distribution function at high rating differences.

(6) Intervals constructed using non-parametric bootstrap with 1000 simulations.

(7) For more details, see http://www.glicko.net/glicko/glicko2.pdf

Ganesh Viswanath is from Perth, Australia and is currently studying a PhD in economics from the University of California, Berkeley. In his spare time he likes to play chess tournaments in the bay area (facing a lot of juniors!) and has more recently been trying to venture into chess statistics. |

Discussion and Feedback
Join the public discussion or submit your feedback to the editors

More...

43

More...

59

many will revert with the same concept that young ones are better than the old generation blah blah blah.

But even Kasparov with his small comeback (though you guys will consider Blitz as irrelevant) somewhat proved that he can compete, meaning ratings are really inflated

But even Kasparov with his small comeback (though you guys will consider Blitz as irrelevant) somewhat proved that he can compete, meaning ratings are really inflated

Karpov about elo system inflation : "I haven’t looked into the mathematical formulae for why it’s happening, but it seems to me there’s an issue – since Fischer had 2760 at his peak, and I got to 2730 or 2735, but when I was rated 2720 Korchnoi was second and he was 2670, so there was a 50-point gap. That indicates something, of course, as it does that Fischer, when he reached that peak, was dozens of points – even close to a hundred – above his rivals. That’s significant. But as for ratings having an absolute significance… well, now they’ve got to 2800. In my day I became World Champion when the best chess players had ratings at about 2600, 2700. After Fischer I was the first to reach 2700, but at that time 2650 was a great rating, while now 2650 – perhaps even… no, maybe you still make it into the Top 100 with 2650."

the entire rating system is inflated

1