Jeff Sonas on Man-Machine, Part II:
Are chess computers improving faster than grandmasters?
In Part I
we looked at the historical results of chess games played between the strongest
grandmasters and the strongest chess computers. The surprising conclusion was
this: top grandmasters and top chess computers are dead-even, and they have
been stuck that way for some time. Neither side has actually won a match from
the other in five years, and the last seven events between grandmasters rated
2700+ and chess computers have all been drawn. How long can this deadlock last?
You're probably thinking, "Okay, sure, they're even now, but isn't it
inevitable that computers will just get better and better and eventually leave
the humans behind?" Well, yes, computers will definitely get better, but
don't forget that human players are improving too! Human players are constantly
getting better at chess; it's just super-hard to measure this statistically,
because improving humans play against other improving humans. It is NOT inevitable
that computers will surpass humans.
Let's look at some numbers and graphs. At first glance, it seems like the top
computers must be gaining considerable ground on the top humans. For instance,
it is quite easy to calculate a performance rating for top computers, in their
games against humans with FIDE ratings. I did this for 742 games over the past
fifteen years, grouped into three-year spans (considering only computers who
were one of the top eight engines in the world at the time). Here are the results:

The overall performance rating of computers against humans has increased at
roughly 30 Elo points per year. This would suggest that however fast humans
are improving, the computers are improving even faster. Thirty Elo points a
year faster. However, if you split these same human opponents into two groups,
based on a rating cutoff point of 2550, an interesting pattern emerges. Computers
are not doing any better against 2550+ players right now, than they were a few
years ago.

Although computers have certainly been more successful in 2001-2003 than ever
before, that is only due to the fact that they are finally starting to dominate
the humans who aren't in the top 200 in the world. As you can see in that graph,
computers are doing no better today, against 2550+ opposition, than they were
three years ago. Let's look at this event-by-event. First, I'll show you the
results against the weaker group, over the past 5-6 years:

Up through mid-2001, a +4 score by a computer, in an event against rated humans,
was almost unheard-of. In the two or three years since then, every single computer
that played in a tournament against sub-2550 humans, has scored between +4 and
+7 against them. Seven straight events, and five different computers, but they
all scored at least +4.
On the other hand, you don't see the same level of improvement by computers
against the stronger players (in the 2550-2700 range). In fact, you don't see
ANY improvement by computers. Let's look at the events since 1998 where computers
faced opponents in that rating class:

Computers are becoming more and more dominant against everyone but the top
200 players in the world. That is leading to an overall performance rating for
computers that is getting higher and higher. However, the players in the top-200
are holding their ground even against the latest and greatest computers. Perhaps
that group will soon shrink down to only the top-100, or the top-50, but not
inevitably, and not irreversibly. As you can see from my previous graphics,
there is no sign that the top-200 players are losing ground at all against the
top computers.
The top 20 humans (the 2700+ crowd) are managing a long string of drawn matches
against computers, and the rest of the top-200 is averaging the same 35% to
40% score that they did a few years ago. So, amazing as it may seem, I don't
see any evidence that the top computers are suddenly going to take over the
chess world. Of course the top computers are improving, mostly through hardware
and software upgrades, but somehow the top humans are improving just as fast,
in other ways.
In Part I, I mentioned that there were two key questions to answer. Let's review
those questions and see what my answers have turned out to be:
Question #1: A decade ago, top grandmasters were undeniably
stronger than chess computers. There was a large gap in strength, roughly 300
Elo points. In chess terms, if a top grandmaster had played 100 games against
a top computer, the grandmaster would have won the match by a score of about
85-15 (roughly speaking). In the past ten years, computers have certainly reduced
the gap. How large is the gap right now, and who is ahead?
Answer #1: There is no measurable gap right now between top
computers and top-20 grandmasters. They have been deadlocked for about five
years. The remaining top-200 grandmasters, as a group, are slightly weaker than
the top computers.
Question #2: Who is improving faster, top grandmasters or
chess computers? What can we say about how the situation will be different in
one year, or ten years, or fifty years?
Answer #2: There is no measurable difference in how fast either
group is improving. There is nothing (yet) to suggest that one or the other
will suddenly pull ahead. It may well be that in ten years the top computers
and top grandmasters will still be deadlocked.
We are at a unique point in chess history, an unprecedented state of dynamic
balance. The top computers have caught up with the top grandmasters, and right
now I'm not convinced that the computers will zoom past the grandmasters. Everything
depends on whether computers or grandmasters can improve faster, starting today.
It may even be that the top humans can figure out how to improve their own anti-computer
play, to the point that they will pull ahead again. Perhaps Garry Kasparov can
lead the way once more.
There is much more to say about how computers and grandmasters are improving,
relative to each other. We'll compare and contrast the various ways that computers
and grandmasters are improving, and how we can try to measure those improvements,
starting with Part III next week.
 |
Jeff Sonas is a statistical chess analyst who has written
dozens of articles since 1999 for several chess websites. He has invented
a new rating system and used it to generate 150 years of historical chess
ratings for thousands of players. You can explore these ratings on his
Chessmetrics website.
Jeff is also Chief Architect for Ninaza, providing web-based medical software
for clinical trials.
Previous articles:
|