@thefrickouttaherelol said in #22:
You can compare chess.com and FIDE elo ratings with Lichess ratings and there is almost consistently across the board a +300 / +400 difference in rating. Lichess ratings are inflated in that way.
This difference between Lichess and chess com has nothing to do with the rating system used. Heck, both use Glicko.
It's also a little bit strange to call this 'inflation'. You might as well call chess com ratings 'deflated' and it would make equally little sense. Average ratings haven't actually changed on either site, and there is no standard for what is 'normal'.
I think the notion that player pools are incomparable is inherently flawed.
It's statistical comparisons only. You can't compare them in the way you can compare (say) meters to feet. For that they should be measuring the same thing, and they don't. (Heck, not even the rating categories are the same. 'Blitz' means something different here than it does on chess com. But that's really besides the point, something that makes the comparison even stranger.)
If you say that you are 6 foot tall, I could infer from that that you are 1.8288 meters tall, and I would be completely accurate. And that's because they are different measures of exactly the same thing: length.
If you say that your chess com rating is 1600, infering from that that your Lichess rating is 1950 is complete nonsense. That might be off as much as 200-300 points, which is really worse than guessing. Even if you disregard that you neglected to specify which rating category. (These diferences have NOTHING to do with the sites. They're differences between specific rating pools. There are huge differences between, say, Lichess blitz and Lichess rapid as well. By your reasoning, they should be the same across the board.)
And that's because, unlike what people believe, these ratings don't measure the same thing. The comparison is valid on average only. It makes no sense in individual cases.
That's the problem with statistics. People tend to apply them to individual cases, and that definitely is flawed. Heck, even scientists make this mistake way too often.
@thefrickouttaherelol said in #22:
> You can compare chess.com and FIDE elo ratings with Lichess ratings and there is almost consistently across the board a +300 / +400 difference in rating. Lichess ratings are inflated in that way.
This difference between Lichess and chess com has nothing to do with the rating system used. Heck, both use Glicko.
It's also a little bit strange to call this 'inflation'. You might as well call chess com ratings 'deflated' and it would make equally little sense. Average ratings haven't actually changed on either site, and there is no standard for what is 'normal'.
> I think the notion that player pools are incomparable is inherently flawed.
It's statistical comparisons only. You can't compare them in the way you can compare (say) meters to feet. For that they should be measuring the same thing, and they don't. (Heck, not even the rating categories are the same. 'Blitz' means something different here than it does on chess com. But that's really besides the point, something that makes the comparison even stranger.)
If you say that you are 6 foot tall, I could infer from that that you are 1.8288 meters tall, and I would be completely accurate. And that's because they are different measures of exactly the same thing: length.
If you say that your chess com rating is 1600, infering from that that your Lichess rating is 1950 is complete nonsense. That might be off as much as 200-300 points, which is really worse than guessing. Even if you disregard that you neglected to specify which rating category. (These diferences have NOTHING to do with the sites. They're differences between specific rating pools. There are huge differences between, say, Lichess blitz and Lichess rapid as well. By your reasoning, they should be the same across the board.)
And that's because, unlike what people believe, these ratings *don't* measure the same thing. The comparison is valid on average only. It makes no sense in individual cases.
That's the problem with statistics. People tend to apply them to individual cases, and *that* definitely is flawed. Heck, even scientists make this mistake way too often.