Check out Grant Acedrex, our featured variant for April, 2024.

Game Courier Ratings for Shogi

This file reads data on finished games and calculates Game Courier Ratings (GCR's) for each player. These will be most meaningful for single Chess variants, though they may be calculated across variants. This page is presently in development, and the method used is experimental. I may change the method in due time. How the method works is described below.

Game Filter: Log Filter: Group Filter:
Tournament Filter: Age Filter: Status Filter:
SELECT * FROM FinishedGames WHERE Rated='on' AND Game = 'Shogi'
Game Courier Ratings for Shogi
Accuracy:80.24%82.12%80.66%
NameUseridGCRPercent wonGCR1GCR2
Francis Fahystamandua180759.0/63 = 93.65%18201793
dax00dax0016209.0/9 = 100.00%16241616
Kevin Paceypanther15958.0/11 = 72.73%15831608
Vitya Makovmakov33315837.0/8 = 87.50%15841583
pheko Motaungcouriermabovini15736.0/7 = 85.71%15771569
Fergus Dunihofergus15606.0/9 = 66.67%15601561
Nicola Caridiniccar15513.0/3 = 100.00%15531548
shift2shiftshift2shift15372.0/2 = 100.00%15391534
Eric Greenwoodcavalier15342.0/2 = 100.00%15341534
Pericles Tesone de Souzaperitezz15332.0/2 = 100.00%15331533
Chuck Leegyw6t15313.0/4 = 75.00%15301532
Raymond Dlewel15215.0/10 = 50.00%15191523
Alexander Trotterqilin15181.0/1 = 100.00%15191518
Play Testerplaytester15181.0/1 = 100.00%15181518
Dougbughouse15181.0/1 = 100.00%15181518
Daniil Frolovflowermann15181.0/1 = 100.00%15181518
juan rodriguezrodriguez15171.0/1 = 100.00%15181517
John Smithultimatecoolster15171.0/1 = 100.00%15171518
Natalia Dolindowhitetiger15171.0/1 = 100.00%15151519
Oisín D.sxg15171.0/1 = 100.00%15161518
Nicholas Wolffmaeko15152.0/3 = 66.67%15131516
Jenard Cabilaomgawalangmagawa15142.0/5 = 40.00%15091520
Julien Coll Moratfacteurix15121.0/2 = 50.00%15131511
S Ssim15002.0/4 = 50.00%14951506
Bogot Bogotolbog14991.0/2 = 50.00%14981499
boukineboukine14981.0/2 = 50.00%14971500
ctzctz14985.0/10 = 50.00%14951501
kunkunkunkun14910.0/1 = 0.00%14941488
Hugo Mendes-Nuneshugo199514910.0/1 = 0.00%14941488
Fabner Cruz Gracilianofabner14910.0/1 = 0.00%14941487
Bob Brownbobhihih14900.0/1 = 0.00%14941486
wyatt wyattquimssarcasm14900.0/1 = 0.00%14951486
jesus babyboypokechamp14900.0/1 = 0.00%14951485
Sagi Gabaysagig7214900.0/1 = 0.00%14951484
Nicholas Archerchess_hunter14890.0/1 = 0.00%14951484
Hsa Saidh14890.0/1 = 0.00%14951483
Georg Spengleravunjahei14890.0/1 = 0.00%14961482
wdtrwdtr14890.0/1 = 0.00%14961481
Daniel Zachariasarx14841.0/3 = 33.33%14851483
Matias I.tsatziq14840.0/1 = 0.00%14861481
sixtysixty14840.0/2 = 0.00%14881479
Richard milnersesquipedalian14830.0/1 = 0.00%14841481
Éric Manálangedubble1914830.0/1 = 0.00%14841481
MichaÅ‚ Jarskihookz14830.0/1 = 0.00%14831483
Jose Canceljoche14830.0/1 = 0.00%14831482
btstwbtstw14830.0/1 = 0.00%14841481
Juan Pablo Schweitzer Kirsingerdefender14820.0/1 = 0.00%14831481
George Dukegwduke14820.0/1 = 0.00%14821481
Erlang Shenerlangshen14810.0/1 = 0.00%14811481
Sandra#Paul BRANDLYARDsandravers13067514810.0/1 = 0.00%14811481
voicantvoicant14810.0/1 = 0.00%14811481
wabbawabba14800.0/1 = 0.00%14781481
xeongreyxeongrey14771.0/4 = 25.00%14781477
Jeremy Goodjudgmentality14750.0/2 = 0.00%14771473
Jose Carrilloj_carrillo_vii14690.0/2 = 0.00%14691468
Greg Strongmageofmaple14671.0/4 = 25.00%14661468
Samuel de Souzasamsou14660.0/2 = 0.00%14661466
Jon Dannjon_dann14650.0/2 = 0.00%14661464
Matthew La Valleesherman10114640.0/2 = 0.00%14641463
Omnia Nihilosacredchao14631.0/5 = 20.00%14671459
Gary Giffordpenswift14590.0/3 = 0.00%14581459
Julianredpanda14550.0/3 = 0.00%14511459
Aurelian Floreacatugo14542.0/8 = 25.00%14551453
Arthur Yvrardtorendil14520.0/3 = 0.00%14501453
pallab basupallab14510.0/3 = 0.00%14531448
Erik Lerougeerik14500.0/3 = 0.00%14521449
darren paullramalam14182.0/26 = 7.69%13671469
Carlos Cetinasissa14170.0/9 = 0.00%14101424
wdtr2wdtr213930.0/9 = 0.00%13821404

Meaning

The ratings are estimates of relative playing strength. Given the ratings of two players, the difference between their ratings is used to estimate the percentage of games each may win against the other. A difference of zero estimates that each player should win half the games. A difference of 400 or more estimates that the higher rated player should win every game. Between these, the higher rated player is expected to win a percentage of games calculated by the formula (difference/8)+50. A rating means nothing on its own. It is meaningful only in comparison to another player whose rating is derived from the same set of data through the same set of calculations. So your rating here cannot be compared to someone's Elo rating.

Accuracy

Ratings are calculated through a self-correcting trial-and-error process that compares actual outcomes with expected outcomes, gradually changing the ratings to better reflect actual outcomes. With enough data, this process can approach accuracy to a high degree, but error remains an essential element of any trial-and-error process, and without enough data, its results will remain error-ridden. Unfortunately, Chess variants are not played enough to give it a large data set to work with. The data sets here are usually small, and that means the ratings will not be fully accurate.

One measure taken to eke out the most data from the small data sets that are available is to calculate ratings in a holistic manner that incorporates all results into the evaluation of each result. The first step of this is to go through pairs of players in a manner that doesn't concentrate all the games of one player in one stage of the process. This involves ordering the players in a zig-zagging manner that evenly distributes each player throughout the process of evaluating ratings. The second step is to reverse the order that pairs of players are evaluated in, recalculate all the ratings, and average the two sets of ratings. This allows the outcome of every game to affect the rating calculations for every pair of players. One consequence of this is that your rating is not a static figure. Games played by other people may influence your rating even if you have stopped playing. The upside to this is that ratings of inactive players should get more accurate as more games are played by other people.

Fairness

High ratings have to be earned by playing many games. They are not available through shortcuts. In a previous version of the rating system, I focused on accuracy more than fairness, which resulted in some players getting high ratings after playing only a few games. This new rating system curbs rating growth more, so that you have to win many games to get a high rating. One way it curbs rating growth is to base the amount it changes a rating on the number of games played between two players. The more games they play together, the more it approaches the maximum amount a rating may be changed after comparing two players. This maximum amount is equal to the percentage of difference between expectations and actual results times 400. So the amount ratings may change in one go is limited to a range of 0 to 400. The amount of change is further limited by the number of games each player has already played. The more past games a player has played, the more his rating is considered stable, making it less subject to change.

Algorithm

  1. Each finished public game matching the wildcard or list of games is read, with wins and draws being recorded into a table of pairwise wins. A win counts as 1 for the winner, and a draw counts as .5 for each player.
  2. All players get an initial rating of 1500.
  3. All players are sorted in order of decreasing number of games. Ties are broken first by number of games won, then by number of opponents. This determines the order in which pairs of players will have their ratings recalculated.
  4. Initialize the count of all player's past games to zero.
  5. Based on the ordering of players, go through all pairs of players in a zig-zagging order that spreads out the pairing of each player with each of his opponents. For each pair that have played games together, recalculate their ratings as described below:
    1. Add up the number of games played. If none, skip to the next pair of players.
    2. Identify the players as p1 and p2, and subtract p2's rating from p1's.
    3. Based on this score, calculate the percent of games p1 is expected to win.
    4. Subtract this percentage from the percentage of games p1 actually won. // This is the difference between actual outcome and predicted outcome. It may range from -100 to +100.
    5. Multiply this difference by 400 to get the maximum amount of change allowed.
    6. Where n is the number of games played together, multiply the maximum amount of change by (n)/(n+10).
    7. For each player, where p is the number of his past games, multiply this product by (1-(p/(p+800))).
    8. Add this amount to the rating for p1, and subtract it from the rating for p2. // If it is negative, p1 will lose points, and p2 will gain points.
    9. Update the count of each player's past games by adding the games they played together.
  6. Reinitialize all player's past games to zero.
  7. Repeat the same procedure in the reverse zig-zagging order, creating a new set of ratings.
  8. Average both sets of ratings into one set.


Written by Fergus Duniho
WWW Page Created: 6 January 2006