sc94597 said:
Consoles don't support v-sync, from what I recall. You'll have a lot of screen tearing if you run the game on a 30hz/60hz display at anything that isn't an even or half multiple of 60. Basically the 60hz means the screen refreshes 60 times/second. If you have an fps locked at 30 you'll have a frame for every other refresh. Or if your television supports 30hz, you'll have a frame for every refresh. Now if you have say a framerate of 40 fps, every so often you'll have an extra frame that occurs in between two others (rather than there being a frame every other refresh.) This leads to tearing of the image where your eye sees part of the previous and next frame. There are techniques that can accomodate for this in PC gaming, since they can't really lock framerates for all PC's. These are called v-sync, g-sync, etc. For consoles, however, if a game is 40 fps they'll have to accomodate for this tearing some other way, and in many cases they might not be able to without expending more resources which would bring down their framerate to 30fps anyway. So usually developers either lock their game at 30 fps or 60 fps. And if it is 30 fps they provide more graphical features. Also for PAL countries, the televisions have refresh rates of 50 hz and the games run at 25 fps, if I recall correctly. All displays tend to support 24hz, because that is the standard for television and movies. |
Consoles do support v-sync, where did you hear it doesnt? The reason why games arent locked to 45 fps are because on a 60hz monitor 45 frames refreshing 60 times will produce 2 of the same frames for every single frame displayed. This causes jutter looking and feeling worse than 30 fps.