Experienced Points

Experienced Points
The Great Framerate Debate

Shamus Young | 27 May 2014 19:00
Experienced Points - RSS 2.0
xbox one vs ps4

So the controversy over the past couple of months is that the PS4 outperforms the Xbox One in terms of frame rate and resolution. No, actually the Xbox One has a performance advantage. No, actually it doesn't, but it doesn't matter. But that hasn't stopped people from compiling massive lists comparing the performance of the two platforms.

This argument is actually four-dimensional. On one axis are the merits of PS4 vs. Xbox One with regards to price vs. performance and features. On another axis is the argument over whether framerate or resolution is more important. Then there's another axis where one extreme says that these numbers are critical and the other says they don't matter. Then finally we have the debate over the even-ness of the framerate and whether or not it trumps any of the other arguments. These debates form a non-Euclidian (and extremely hyperbolic) volume of debate-space and half the work in any given argument is in determining if any two participants are even operating in the same frame of reference.

Which is to say that the whole thing is a confusing angry mess. For the record, I don't have a horse in this race. I loved the PS2 and still think of it as the greatest console ever. I thought the PS3 was a technological misfire. I liked the Xbox 360, but once mine bricked I never had a strong desire to replace it. And now I've got such a massive backlog of PC games that I can't justify getting a now-gen console. Why pay hundreds of dollars to gain access to a small library of titles I don't have time for?

But while I don't have any first-hand experience with either platform, I do have a lot of experience analyzing, scrutinizing, and agonizing over frame rate. I can't settle the debate, but hopefully this will cut down on the ambient level of misinformation and misunderstanding we have going on here.

This debate isn't new. Way back in 1998, now-defunct graphics card manufacturer 3dfx Interactive was busy turning out hardware designed to render blazing fast, high-resolution images using 16 bit color. At the time there was a big debate on what was more important: Raw speed or color depth.

16bit vs 32bit

To illustrate the difference, I took this image and created a version to simulate the difference between 32- and 16-bit color depth (seen at left).

For the nitpickers: I did this by reducing the red, green, and blue channels to 4 bits each, assuming that in the context of a game the other 4 bits might be given over to the alpha channel. I also reduced the resolution, both to make the difference easier to see and also to better show off how big the pixels were back then. Of course, back then we were looking at cartoonish texture maps and not photographs of ladies cosplaying as Rainbow Brite, but in this same time period game developers were going mad with colored lighting and fog, which greatly exacerbated the problem. The point is: We can haggle over how accurate this image is, but for the purposes of this discussion this is good enough for showing the reader what the 16 bit vs 32 bit distinction might have looked like in the context of a late-90's PC game.

Most of the rest of the industry was moving to 32 bit color, which would produce something like the top image. 3dfx was doubling down on raw speed with 16 bit color. The game would (in theory) run faster and allow for higher resolutions, but would suffer from the color banding you see in the lower image.

People debated about which one was more important. Some people found the color banding to be incredibly ugly and distracting. While I could see the color loss if I tried, it never bothered me. Maybe it was because I spent so many years looking at 4-bit CGA graphics that 16 bits seemed "good enough" for me. In any case, it was important to a lot of gamers and developers. In the end 32 bit won out over 16 bit.

It's important to note that frame rate and refresh rate are two different things. I see the two terms being used interchangeably, and it always makes my eye twitch. Refresh rate is the interval at which your monitor or television will show changes. This is interval is fixed and completely unrelated to whatever the videogame might be doing. This gets confusing for some people, because the monitor refresh rate is usually expressed in hertz, while videogame speed is expressed in frames per second. In a perfect world these two terms would be synonymous, but because of the strange way that technology evolved most older displays showed 30fps and ran at 60hz. (Or if you lived in America, it ran at 29.97 frames a second, for reasons that seemed perfectly sane in 1941 but ultimately just drove everyone crazy for the next half-century.)

Back in the day, those old CRTs needed to refresh as often as possible. That Cathode Ray Tube generated images by blasting a fluorescent screen with electron guns, which sounds very Star Trek but is actually very low-tech by today's standards. That system produced an image that flickered very slightly. It needed to flicker as quickly as possible, because the faster it goes the smoother it seems to the eye. If it pulsed at only 30 frames a second, then watching television would feel like staring into a strobe light (which is actually the case) which can cause headaches, nausea, vomiting, and seizures. Refreshing 60 times a second was deemed fast enough to keep most of us from getting sick. (Although some sensitive people, like my daughter, can still get violently ill even at 60hz.)

Comments on