
Metacritic scores.
They’ve been thrown about quite a bit, usually to identify Game A as being superior to Game B, but is there any merit to these claims?
First off, there have been numerous criticisms of how metacritic collects and converts scores. I’m not going to get into that right now, but you can have a look at their system here - http://www.metacritic.com/a... just wanted to point it out to anyone interested.
What I want to point out today is this:
There is no scientific basis upon which you can use metacritic in its present state to scientifically and objectively compare ANY 2 games.
Metacritic scores for different games are usually collected from different websites. Even when they are collected from the same website, they are usually written by different staff members.
So that’s the problem. Can you use scores written by different people to compare games? The answer is yes, but it involves a bit of science and statistics.
Stop. Stop right there. Do not close tab due to anticipated boredom. I’ll keep it simple. I promise.
Here’s one example of how you could do it. There are probably better ways.
1. You identify the critics you want to use. Hopefully, Jim Sterling is not one of them.
2. You give each of these critics the same group of games to score.
3. Then, you rank the scores. Eg if they were given 10 games, rank 1 describes the best in the lot, rank 10 describes the worse.
4. Once you’ve done that, you test for correlation between ranks of reviewers. For example, you can use the Spearman’s Rank Order Correlation. This will tell you if there is agreement between reviewers and thus, if their ranks are comparable now and, arguably, in the future.
I say arguably because a reviewer can review in a predictable way for the test and then proceed to launch utter flamebait in the future, but you get the idea.
Whether or not this is the best way or even a valid way is not really the point I’m trying to make. My point is this:
Comparison - true, accurate, unbiased comparison, takes work and effort. Work which is not done by Metacritic.
At the end of the day, the Metascore is a good indicator of game quality - which is what it was designed for.
However, it cannot be used to OBJECTIVELY AND SCIENTIFICALLY argue which game is the best.

Microsoft announced its financial results for Q3 of fiscal year 2026, including an update on its gaming Xbox business and more.
Not looking good. Hopefully Asha Sharma is able to turn Phil’s disaster around.
To me it's still quite remarkable how they can cash-in 5.3bn in revenue in a single quarter, since their hardware is basically dead.

The charity event will be streamed live from Gamescom in August.

Thanks to the slip-up of an artist working on the title, we now have more evidence that a new Injustice game is in the works.
I use Metacritic all the time, but not for the average score. I use their collected list of reviewers to pick out a few high, mid, and low review scores and read those reviews.
But I do agree with your sentiments. The "average" on Metacritic is meaningless. It only became meaningful for a brief season when certain fanboys used Meta score to label certain games as "AAA" and any game that didn't get a 90/100 or above average on Meta was of course not "AAA" according to that standard. This practice quickly ceased once the PS3...er...a "certain" console began to dominate the Meta top-scoring charts.
For teh console WAAAAAAAAAAAAAAAAAARZZZZZZZZZZ ZZZZZ!!!!!!!!!!!!!
Absolutely nothing
Uh-huh
SAY IT AGAIN Y'ALL!!!!
It's good for fanboys.
Metacritic is only good for fanboy flame wars. "I'll raise my Uncharted 2 (96) against your Halo 3 (94), SUCK IT!"