What do numbers like 100Hz, 200Hz, 600Hz really mean for the picture quality of your new HDTV?
Original video frames (images 1 and 2) at 50 frames per second (fps) aren't enough to fill modern LCDs with higher refresh rates. Duplicating the original frames is one method. Alternately, frames can be interpolated to fill the gaps. In this example, the TV's processor creates frame 1a from the difference between 1 and 2.
(Credit: Geoffrey Morrison/CNET)
With 100Hz, 200Hz, and even 600Hz, the refresh rate gets a lot of attention in the marketing of new HDTVs. What it is and how it works is interesting, but why it exists is even more so. And it can have a profound effect on the picture quality of your HDTV.
Curious? Let's start with the basics. Television is a series of images, shown rapidly enough that your brain sees it as motion. In Australia, our electricity runs at 50Hz, so it's only natural that our TVs run at the same rate (in the US, where a lot of online information comes from, 60Hz is common). This is largely a holdover from the cathode ray tube (CRT) display days, but our entire system is based on it, so there's no use changing it.
What this means is that modern HDTVs show 50 images per second (50Hz).
Upping the frame rate
A few years ago, LCDs hit the market with higher refresh rates. These started at 100Hz, though now you'll see 200Hz and beyond (for US TVs, it's 120Hz and 240Hz). In this case, higher is indeed better, but to understand why it's better, we have to discuss why it exists in the first place.
All LCDs have a problem with motion resolution. Which is to say, when there's an object in motion onscreen (or the whole image is moving), the image blurs, compared with when the object/scenery is stationary. In the early days of LCDs, this was predominantly because of the "response time", or how fast the pixels could change from light to dark. Response times on modern LCDs are quite good, and this isn't the big issue anymore.
The top is full-motion resolution. The bottom half is a representation of what motion blur looks like. Notice how the dolphin on the right is blurred, compared to the other three and the rest of the image.
(Credit: Geoffrey Morrison/CNET)
The issue is how your brain interprets motion. Because it's your brain, everyone is going to see motion resolution somewhat differently. Some people don't notice motion blur. Some people aren't bothered by it. Some (like me) notice it quite often and are bothered by it. Others, like my colleague David Katzmaier, know it exists, but don't notice it enough in normal program material to consider it a major factor in picture quality.
There are two primary ways to fool the brain into seeing better detail with LCDs: backlight flashing (also called backlight scanning) and frame insertion.
Backlight flashing is what it sounds like. The most basic version of backlight flashing is the backlight going dark in between video frames. This moment of darkness is much like how a film projector works: an image, then darkness, an image, then darkness, and so on. Done slowly, this can result in flicker. Done fast enough, and you don't notice it. A more advanced version, called backlight scanning, dims sections of the backlight in sequence with the video. In either case, the side effect is a loss of light output (sometimes significantly), because there are sections of time where the backlight is literally off (or close to it). Another type of backlight flashing is called black-frame insertion, which shows a black image in between the real frames, but doesn't actually manipulate the backlight.
With 100Hz and 200Hz displays, there's another option: frame insertion. This method, also called frame interpolation, actually creates entirely new frames of video to insert in between the "real" frames of video. With video sources, like live TV, sport and video games, there's very little downside to this method. You get excellent motion resolution, and you maintain the light output of the display. The image at the top of this article is an example of frame interpolation.
However, with film/24fps content (including movies and most scripted-TV shows), there's an issue. The interpolated frames smooth out the inherent juddery motion of 24fps content. On the surface, this may seem like a good thing, but the resulting ultra-smooth motion makes movies look like cheap soap operas. Fittingly, then, this is called the Soap Opera Effect. We, and many TV companies, call it "dejudder". Personally, I find motion-interpolated video annoying to watch. In some, it causes nausea, while others don't mind it, which I find rather shocking.
Most modern higher-refresh-rate TVs have one or both versions of this technology, and you can completely select which to use (if at all). In some cases, like the Cinema mode in Panasonic's WT50 series of LED LCDs, you're locked into the motion interpolation. The trade-off (usually) is that, if you don't use motion interpolation or backlight scanning, you don't get full motion resolution.
Nah, let's make it more confusing
With marketing being what it is, companies are now obfuscating their TVs' actual refresh rates. Samsung, LG, Sony, and the like, have stopped giving native refresh rates, instead, adopting bespoke motion-resolution ratings called "Clear Motion Rate", "TruMotion", "Motionflow XR", and so on.
In all cases, the companies use backlight scanning and/or extra processing to imply that their TVs have higher refresh rates than they actually do. So, for example, a Clear Motion Rate of 100 could be a 50Hz TV with a scanning backlight, or it could be a 100Hz LCD without a scanning backlight. The spec sheets for the televisions rarely, if ever, list the actual panel refresh rate.
Gary Merson wrote an excellent article on this at HDGuru, called "Beware of phony LCD HDTV refresh rates" (based on US numbers, but the basics still hold).
Because plasmas don't suffer from motion blur like LCDs do, they don't need higher refresh rates. The problem is that all plasma TV manufacturers also make LCDs. So you're not going to see a big marketing push from any of them, saying that, "No, no, buy our cheaper plasmas because they don't suffer from motion blur (or poor off-axis response, or poor contrast ratios)". With the intense marketing of 120Hz and 240Hz, many consumers assumed plasma was lagging behind, fitting into their erroneous preconceptions that plasma is somehow an "older" technology.
Instead, all three plasma manufactures (LG, Panasonic and Samsung), have adopted the "600Hz" claim. As Obi-Wan Kenobi said in Star Wars Episode VI: Return of the Jedi, "... what I told you was true, from a certain point of view".
From a certain point of view? As an engineer once eloquently explained to me, plasmas create light with time. Each pixel in a plasma has only two states: on or off. (In that way, they're a completely digital device, unlike LCD, which can still be analog, but that's fodder for an entirely different article.)
Because plasma pixels only have two states, they create different levels of brightness by flashing more or less often. This is where the 600Hz comes in. In the most basic explanation, plasmas break up each frame of video into 10 subfields (60Hz x 10 = 600). If the pixel is supposed to be bright white, it flashes once for each of those subfields. If it's supposed to be 50 per cent bright (50 IRE, or medium gray), it flashes for half of those 10 subfields. When it's supposed to be dark, it doesn't flash at all.
Digital Light Processing (DLP), which is used in projectors, works by a similar principle; each mirror is either on (facing the lens) or off (facing away).
In reality, it's a bit more complicated than this, but this is the general idea. There are other pros and cons to this method that are beyond the scope of this article, but if anyone really wants me to dive into it, let me know.
So "600Hz" is more or less a marketing thing, but it's not untrue. The fact is, plasmas don't suffer from motion blur like LCDs do, so they don't need higher refresh rates.
If you're looking for motion blur in your own TV, keep in mind that, in some cases, there's going to be blur in the source. This is most common with movies shot on film. Fast motion will blur on film because of its low frame rate.
Personally, I notice motion blur most in close-ups. When an actor's face fills the screen, for a moment, he or she will be stationary and you'll see every bit of facial detail. Then they'll move slightly, and the image will blur. I see this across all different types of source material.
Refresh rate is how often the TV shows a new image. All modern video is either 24 frames per second (movies and most TV shows), 60 fields per second (1080i video), or 60 frames per second (720p video). Higher refresh rates are used to increase the apparent motion resolution of LCDs. The 600Hz of plasmas is largely marketing, but is technically how they work.
If you're annoyed by motion blur, you'll be better off getting the highest-refresh-rate LCD that you can get, or sticking with plasma (or OLED).
However, not everyone notices or is bothered by motion blur. I do, and it's one of the main reasons I prefer plasma over LCD (the other being contrast ratio). As mentioned earlier, my colleage David doesn't notice or get bothered by it. We both have highly critical eyes when it comes to TVs, but because the perception of motion blur is so subjective, we're both right.
Do also note that, because your source is 24fps or 60fps, you do not need special HDMI cables for a 100 or 200Hz TV. If the salesperson tells you that, he or she is either clueless or lying. But for more on that, check out "Why all HDMI cables are the same".
Are you bothered by motion blur? Comment below; I'm curious.