By using this site, you agree to our Privacy Policy and our Terms of Use. Close
curl-6 said:
Chrkeller said:

But look at what developers are doing across the three main sectors of fidelity.

1) resolution impacts bandwidth.  is the S2 rendering games at 360p and look like the witcher 3 on the S1?  Nope.  Resolution rendering, in many cases is quite high.  

2) image quality impacts bandwidth. is the S2 rending games with rebuilt assets like Hogwarts on the S1?  Nope, image quality (especially textures are quite high).

3) fps impact bandwidth.  is the S2 running game at a reduced fps compared to current gene?  YES.  

Third party developers, thus far, are address the memory bandwidth bottleneck by dropping fps...  this is literally happening, it is a fact.

There are things other than bandwidth can limit framerate, from CPU load to pixel/polygon fillrate.

It will depend on the game of course; in the right circumstances (say, a game that pushes a ton of alpha transparencies at a high resolution) bandwidth could become the limiting factor, but I expect that in most ports of PS/Xbox games, it'll be more the CPU that necessitates 30fps.

Sure, but we are talking Elden, 7 Remake, Cyber, Madden, etc.  All 30 fps, I can't fathom the CPU being why.....  wait for it....  memory bandwidth is.  

Something like Snake Eater 3, yeah, the CPU is going to take hit.  But until stuff like Snake Eater, Wilds or Outlaws are out, it is hard to determine CPU clock speed impact. 

But we do have a spread of ps4 games....    you think the S2's CPU is a bottleneck with ps4 games?  I don't.  It is the memory bandwidth.  



i7-13700k

Vengeance 32 gb

RTX 4090 Ventus 3x E OC

Switch OLED