dahuman said: For my sanity, I really hope they are not using Bulldozer based cores, for fuck sake, that shit sucks. |
Once you accept that none of the next generation consoles will have an Intel based Core i5/i7 due to cost and Intel's high profit margins, what's the next best gaming processor in the world for a console that is expected to last 6-8 years? Bulldozer/Vishera. Say it isn't so?
Also, you forgot another key word: Context.
On the PC gamers care a lot about CPU performance because the most popular gaming genres are MMOs & RTS/strategy games which are notoriously poorly multi-threaded and are very CPU limited style games. Furthermore, enthusiasts tend to spend $100-200 for small performance increases on the GPU side (like $100 extra to go from GTX670 to GTX680 or $200+ extra for 25% more performance to go from GTX660Ti to GTX680, or HD7950 to HD7970GE, etc.). For those users, especially if they overclock their CPUs, every little bit of performance and reduction in power at overclocked 4.5-5.0ghz states matters.
None of these things apply to consoles: (1) Console CPUs aren't overclocked (2) Consoles don't have RTS or MMOs (3) Consoles won't use high-end or dual-GPUs.
So why does context matter so much? Because when you put Bulldozer/Vishera (FX8150-8350) and pair it with a GTX670 in a wide variety of "console style games" games, the performance difference between it and Core i5/i7 is almost non-existent.
- Alan Wake, ARMA II: Operation Arrowhead, Assassin's Creed 3, Battlefield 3, Crysis 2, F1 2012, Far Cry 3, GTA IV, Hitman Absolution, Max Payne 3, Metro 2033, and Sleeping Dogs you end up with this:

http://pctuning.tyden.cz/hardware/graficke-karty/25994-vliv-procesoru-na-vykon-ve-hrach-od-phenomu-po-core-i7?start=16
At 1920x1080 with Anti-aliasing, when FX8120-8350 are paired with a GTX670, they are only 2% behind in performance! In the type of games that are on consoles (non-MMO/RTS games), the console will be primarily GPU limited. The context matters even more because it's almost a certainty that none of the next generation consoles will have a GPU as powerful as a GTX670. Essentially if you were to put an FX8000 CPU with any GPU slower than GTX670, the console will be 95% GPU limited for most of its life going forward assuming the games are running at 1920x1080 with some anti-aliasing that stresses the GPU. Most professionals reviews run CPU tests at useless resolutions like 1280x800 or 1680x1050 with no AA to show the differences in CPU speeds. This type of testing is absolutely meaningless if the consoles are targeting 1920x1080 with some AA, because the workload almost entirely shifts to the graphics sub-system. The slower the GPU is in the consoles, the more the workload will expose the GPU bottleneck.
FX8150 is fast enough to get you 64 fps minimum and 81 fps average in Battlefield 3 at 1920x1080 4xMSAA Ultra Quality settings with a GTX690, barely behind Core i5 2500K/2600K. Of course we know the consoles will never have anything like a GTX690 in them, which automatically means a GPU bottleneck with an FX8150.

Conclusion: FX8000 would not suck at all in a console -- that would actually be the 2nd best choice after Intel's CPUs i5/i7s, far superior to any processor made by IBM or any Core i3 from Intel; and it would handily trounce an 8-core Jaguar by miles. If anything, you should hope and pray that PS4 has an 8-core (4 module) Bulldozer/Vishera in it, not denounce it, because honestly that's the 2nd fastest CPU you can get once you step outside of Intel's CPU offerings. Also, if you look at FX8320's retail price of $169 and what Sony can possibly purchase this AMD processor for directly from AMD, then it's pretty obvious Intel's offerings are out of the question, unless you are willing to pay $600-700 for a PS4 with a Core i7. But again, that wouldn't even matter unless the console had a GPU more powerful than a GTX670...
I think console gamers need to reevaluate the context of this hardware more carefully. PS360 rendered most games at 1280x720 and later even at lower resolutions (Black Ops 2 at 880x720, Uncharted 3 at 896x504, God of War 3 at 1152x640). Because of such low resolutions with minimal or no AA filters, the console's CPU speed mattered a lot more. Once you move to 1920x1080 and add 2-4x AA, in probably 90-95% of cases the console's GPU willl become the bottleneck in non-MMO/non-RTS based games unless they pair an FX8000 with a GTX680 or faster. Whichever console has the best graphics will be dictated by its GPU setup, not CPU, assuming they don't put some slow anemic CPU aka Wii U in them.