By using this site, you agree to our Privacy Policy and our Terms of Use. Close
JRPGfan said:
Soundwave said:

I know it's not the be all end all, but wouldn't two Tegra X2's in unison constitute about 1.25 TFLOP of performance, and if we average out that Nvidia's floating point performance is generally 30% higher, that would put the two at 1.625 TFLOPS in AMD terms. 

Maybe if they added say 24-32MB of high speed eDRAM onto the SCD version of the TX2 ... would that change things? That would kinda cancel out the XB1's memory bandwidth advantage. 

Where would it be lacking vs a XB1 in that scenario? Surely two X2 units would boost things like the poly count and fillrate?

Thats useing PC numbers, with DX11.

On consoles the differnce will be smaller, as will it with a differnt used API, like DX12 or Vulkan.

The "avg" advantage of flops to flops will be alot smaller in a console. Id be surprised if it was more than 10-15%.

I think Pemalite is correct, in saying that 2 x tegra x2 = roughly Xbox One level of performances, maybe limited by memory bandwidth so running lesser resolution.

Direct3D or DX is a propietary API from MSoft, so neither Sony nor Nintendo will use it in their development tools. That leaves us with OpenGL/Vulkan.

Also, those theories with 2xTegras have, in my eyes, one big flaw: they are assuming a 100% scaling. That's something that won't happen.



Please excuse my bad English.

Former gaming PC: i5-4670k@stock (for now), 16Gb RAM 1600 MHz and a GTX 1070

Current gaming PC: R5-7600, 32GB RAM 6000MT/s (CL30) and a RX 9060XT 16GB

Steam / Live / NNID : jonxiquet    Add me if you want, but I'm a single player gamer.