twesterm said:
Most of those games you listed I'm sure you expect to get between a 9 and 10. A bad game (or even a game that just doesn't deserve that score but still good) isn't automatically an 8, it's just something else so that variation is meaningless. Lets just pretend that Awesome Exlclusive VII for the PS3 and Bodacious Exclusive IV for the 360 both have a 9 on metacritic. Edge reviews them and gives Awesome Exlclusive VII a 7.5 while giving Bodacious Exlclusive IV an 8.5 According to you, that means Edge is automatically biased against the PS3 and both should either get a 9 or an 8.5 or just the same score when in reality they thought one game was a 7.5 and the other was an 8.5. Do you see what I'm getting at? They're different games so they can get different scores. |
Just to be Devil's Advocate, while of course two seperate games can get different scores, if those scores are inconsistent with a general set of scores then it does raise questions about the scoring, particularly if you take the view that, opinions or no, they are opinions from supposed 'experts' on a common topic and should show a relative degree of consistency - i.e. either the lighting is great or its not, it can't be both. Either the AI is good or its not, it can't be both - and so on.
As an analogy, if a film get's consistently praised in 90% of reputable reviews for having a good script, and one review site says the script is bad, you can see it as just another opinion, or you can wonder as to why? Does that reviewer have weaker grasp of scriptwriting theory? Did they not like the film particularly and let this influence their view on the script? and so on. The evidence would definately imply that there is reason to be cautious of the single poor opinoin of the script vs the universal acclaim.
Rather than just moaning, which I agree with everyone is a pain, the post on Edge vs metacritic averages is interesting to me because it appears to be factual, based on real data, and it clearly appears to show that, on average, a 360 title will review closer to or above the metacritic average than a PS3 title, which certainly does imply a bias.
Of course, that bias could be accidental. I hate to sound elitist, but I don't really hold game reviewers on the whole as particularly 'expert' on game mechanics, etc. and it may be no more than the PS3 reviewer is simply being harsher than the 360 reviewer, who really likes the 360 titles and his enjoyment reduces the percieved impact of any flaws.
But clearly, there is a skew there - if the shown data is correct. I hate to actually pick up on this, as I normally really dislike the whole 'they're bought off by MS, etc' tinfoil hat nonsense. But this analysis does show a skew. Of course it could just be taste - that can't be discounted. The reviewer may have genuinely felt KZ2 was simply average (to take the obvious example) while ODST is pretty stellar.
What I'd really be interested to see is a similar analysis for all major review sites / sources to see if skews are fairly common and chance over time - i.e. it is probably just different tastes, etc. - or whether only a few sites have a noticable skew from others.
Try to be reasonable... its easier than you think...