Print 80 comment(s) - last by titanmiller.. on Oct 22 at 6:23 PM

4K resolution hardware will officially be called ultra HD

The Consumer Electronics Association (CEA) has announced the official name for the next generation 4K high-definition display technology: "Ultra High-Definition" or "Ultra HD." According to the CEA, the name is intended to infer the new format's superiority over conventional HDTV.

The CEA Board of Industry Leaders voted unanimously this week to recommend the previously mentioned names for the new next-generation HD resolution. Along with agreeing on a name, the CEA also outlined minimal performance characteristics to help consumers and retailers understand the benefit of the new technology set to begin rolling out this fall.

“Ultra HD is the next natural step forward in display technologies, offering consumers an incredibly immersive viewing experience with outstanding new levels of picture quality,” said Gary Shapiro, president and CEO, CEA. “This new terminology and the recommended attributes will help consumers navigate the marketplace to find the TV that best meets their needs.”

The core characteristics that the CEA agreed on include a minimum display resolution of at least 8 million active pixels with at least 3840 pixels horizontal and at least 2160 vertical pixels.

To meet the minimum needs the display will have to have an aspect ratio of at least 16 x 9. Devices meeting the specifications will also be required to have at least one digital input capable of carrying and presenting native 4K format video at 3840 x 2160 resolution without relying on upconverting.


Comments     Threshold

This article is over a month old, voting and posting comments is disabled

PC Monitors
By Lazarus52980 on 10/19/2012 9:26:12 AM , Rating: 2
I wonder what impact this will have on the PC monitor market? Now that I think about it... I wonder if this will impact the Xbox 720 and PS4. This could actually have a good trickle down effect on PC gaming as a whole.

RE: PC Monitors
By mcnabney on 10/19/12, Rating: -1
RE: PC Monitors
By Vinny141 on 10/19/2012 9:48:45 AM , Rating: 5
i don't think xbox "screwed up"

4K TVs wont be mainstream, i'm betting, for another 5-10 years.

xbox 360 did petty good without a blu-ray drive

RE: PC Monitors
By kattanna on 10/19/2012 11:08:15 AM , Rating: 5
4K TVs wont be mainstream, i'm betting, for another 5-10 years.

you still have a LOT of people who dont even have a 1080p TV right now..and they are ..relatively cheap. oh.. also remember that the first 1080p TV's came out way back in 2005

these new 4K TV's will be WOW expensive for years..and everything will have to be up converted from 1080 for a long time.

then.. think of current cable/sat TV services and there existing compression to feed you 1080i/p they will be very hard pressed to handle such feeds.

RE: PC Monitors
By inperfectdarkness on 10/20/2012 2:56:35 AM , Rating: 2
This. There will be so very little in media to compel consumers to upgrade to "ultra-hd", that it will take years before it becomes a mainstream standard. Heck, I don't think 1080p quite standard yet--since you can still find some TV's that don't support it.

Furthermore, games have stagnated at being programmed for 1080p (or lower, and being upscaled) for several years now. As of this post, I know of no game that even comes close to being programmed for 8Mp rendering. 2560x1600 has HALF the display resolution as this standard...and it's still twice the resolution of 1080p. Final Fantasy: The Spirits Within is probably the highest-definition source material already available for > HD viewing, and it supposedly only has twice the HD resolution as well--putting it on par with WQXGA as well.

I do have hopes that pushing the resolution envelope will eventually spread across the entire tech spectrum and improve the quality of media offered--but I also know that without some mandated push by the government, that it will trickle down very, very slowly. HD was somewhat easier to press because digital was necessitated by dictation. Pushing Ultra-HD isn't going to be nearly as easy--especially with virtually no media currently being produced with that magnitude of resolution.

RE: PC Monitors
By rdhood on 10/19/2012 11:13:15 AM , Rating: 1
4K TVs wont be mainstream, i'm betting, for another 5-10 years.

I think you are right, though your timeline is a little far out there.

If this unfolds a lot like 1080p & 3D, then 4K TVs will come first... THEN 4k content will begin to trickle in.... THEN 4k TVs will drop in price and start to become mainstream as people anticipate the arrival of much more 4k content, THEN the 4k content arrives in huge amounts, and somewhere along the line xboxXXX will support 4k.

1080p sets were introduced around 2005. It wasn't for another 3 years before enough 1080p content became available to finally push up sales of 1080p sets. The same type of thing is happening with 3D, though most people view 3D as a "want" rather than a "need".

5 years for 4k seems about right.

RE: PC Monitors
By PrinceGaz on 10/19/2012 12:47:46 PM , Rating: 1
The improvement from 480(NTSC) or 576(PAL) lines at 720 columns up to 1920x1080 full HD was a major improvement everyone could see, assuming decent eyesight.

I'm not convinced that 4K is really going to make any difference to most people's perception of the quality of what they are watching, and is likely to go the same way as 3D TVs-- something nice to have, but not worth paying extra for. As such, adoption of it is likely to be very slow, slower even than that for 3D TVs, because you will also need a whole new BD-player/cable/satellite box to provide the 4K video. I wouldn't be surprised if the 4K or UltraHD BD discs are also more expensive than normal BD discs, which are still more expensive than DVDs.

I'd love "retina"-resolution PC monitors to become readily available at affordable prices though.

RE: PC Monitors
By bsd228 on 10/19/2012 2:52:35 PM , Rating: 2
The just announced GoPro Hero3 will record 4k (albeit only at 15fps), and that's a $399 sku. The initial driver for 4k use will be SLR owners who have been having to display their 18,21,24,36mp pics at 2mp resolution - now they have ~8mp of screen.

No, the question becomes what will they name 8k screen in 15 years...super ultra hd? 8k may be the end game for technology as we know it now - they demo'd this at the Olympics covering the swimming events and people described it as damn near reality. As nice as current HD is with a good signal, you still won't call it looking at the real thing.

RE: PC Monitors
By titanmiller on 10/22/2012 6:18:30 PM , Rating: 2
I assume it will be UHD 4K and UHD 8K.

RE: PC Monitors
By Solandri on 10/19/12, Rating: 0
RE: PC Monitors
By Shadowself on 10/19/2012 4:18:14 PM , Rating: 5
You're clearly a person who knows nothing about resolution and what is perceived by a person (or even a non human digital sensor for that matter) in the real world. You've read somewhere the brain dead simple explanation that the *average* person's "pixel" resolution (based upon an individual rod or cone resolution) is approximately one arc minute. You've bought into the lame excuse for reality that anything at a higher resolution than that one arc minute is 100% useless.

In a single word: Bullshit.

There are many, many levels of perception that the average person "sees" that goes well beyond that of the basic one arc minute per pixel fallacy. Look up vernier resolution as just one one of *many* examples of higher perceived resolution. Some levels of perception by people are as much as 10 to 30 times better than that one arc minute per pixel.

Therefore the "you have to be X feet away to see Y resolution" is off by at least a factor of 10. Before you lose *any* perceived effect/benefit of a 4K screen that is 50" diagonal you have to be over 16 feet away.

The argument really comes down to, "Is the additional perceived resolution/clarity of the imagery worth it to ME?" That is a personal response based upon each individual's eyesight, values, and willingness to pay for it. But to say that you have to be two feet away from a 50" screen to see any benefit of 4K is absolutely not true.

And this is true on a non human/non personal level too. The classic example is the GOES satellites. Their sensors take a picture of the Earth at 1 km resolution at nadir (straight down). However, they routinely image the bridge going across Lake Pontchartrain. Clearly that bridge is no where near one kilometer wide (in reality it is under 0.1 km wide!) yet it shows up in GOES imagery! And the resolution of the GOES sensor is worse than 1 km per pixel at that latitude!

If the dumb GOES sensor can image a < 0.1 km object with > 1 km pixels then it must be clear (pun intended!) that the human mind with human vision at one arc minute per "pixel" in our vision sensors can see much, much better than that.

RE: PC Monitors
By Shadowself on 10/19/2012 6:21:11 PM , Rating: 2
I guess rating this way down is because reality is more difficult to understand than simple, wrong charts!

Too many people would rather believe the simple -- and wrong -- writings by pundits out there posting crap on the 'net and having lots and lots of people repeat it rather than pay attention what what's really there.

If enough people repeat it then it must be fact, right?

How many people posting and re-posting and re-re-posting the inaccurate information have actually designed real-time optical sensors for satellites? How many of these people have designed optical sensors and imagery interpretation systems for the military? How many of these people have have done research into optical imagery perception including such things as edge effects and super resolution effects of motion imagery (among others)?

People can continue to rate these posts down, but people need to know what reality is. It is not what is most often being posted and re-posted across the 'net.

RE: PC Monitors
By macca007 on 10/19/2012 8:05:13 PM , Rating: 3
Thank you,Spot on!
Same applies to this 30fps bullshit, I am getting sick of the argument that you don't need or can't tell the difference over 30FPS, BULLSHIT I can't.
If anyone is a gamer out there especially fps games I bet they will notice the difference playing on a 120hz monitor compared to a 60hz screen, Let alone 30fps. The more you play or use pc the more you start picking up slightest details be it resolution or speed.

RE: PC Monitors
By Solandri on 10/20/2012 3:26:53 AM , Rating: 2
Looks like the videophiles are out in force. I don't deny there will be a market for 4k TVs, just like there's a market for $5000 stereo components. I'm just saying for the typical TV buyer they'll be useless. Two of the major TV stations in LA broadcast in 1080i, the other two broadcast in 720p. I've challenged many people to tell me which two are 1080 and which are 720. Only one has been able to say correctly, and he admitted he was just guessing.

The 1 arc-minute of resolution is derived from the rayleigh limit. At a typical pupil diameter of 3 mm, the minimum separation before two objects blend into one is about 1/100 of a degree, or about 2/3rds an arc-minute. Even if you had a perfect retina, you can't distinguish apart objects smaller than this. It's the optical limit of your eye's lens.

Having binocular vision as well as being able to scan helps improve slightly over this. But we're talking about levels of detail where you'd have to strain to see differences. This is something that's relevant if you're examining scientific or crime scene photos. Not something you'll be doing while watching TV. Feel free to think you can see 10x higher resolution than this. Next time you're using a telescope try magnifying 10x higher than the theoretical max magnification and see if it really helps.

Their sensors take a picture of the Earth at 1 km resolution at nadir (straight down). However, they routinely image the bridge going across Lake Pontchartrain. Clearly that bridge is no where near one kilometer wide (in reality it is under 0.1 km wide!) yet it shows up in GOES imagery! And the resolution of the GOES sensor is worse than 1 km per pixel at that latitude!

... If that's the basis for your misguided beliefs, I don't even know what to say. Please read up on some sampling theory.

Having a resolving limit of 1 km doesn't mean you can't detect objects smaller than 1 km. It means you can't distinguish between objects smaller than 1 km. That is, if you imaged a 300m ship with the same reflectance characteristics as the bridge (same brightness and color), you wouldn't be able to distinguish it from the bridge. For the purposes of the GOES sensor, all objects 1km or smaller with the same reflectivity are the same.

In TV terms relevant to this discussion, at a 6 foot viewing distance, a pixel smudge on a 50" 1080p screen would be indistinguishable from 2x2 pixels on a 50" 4k screen. The additional detail is there on the 4k screen. It's just that like the GOES sensor, your eye cannot tell the 1080p smudge pixel apart from the sharper 2x2 pixels on the 4k screen.

RE: PC Monitors
By DarkUltra on 10/22/2012 4:25:45 PM , Rating: 2
It's not just about the number of pixels and cone cells, but about pixel blending and aliasing. What we need is a proper "blind" test where the viewers does not know what resolution they are looking at.

Like this one, but with pixels instead of framerate:

RE: PC Monitors
By chemist1 on 10/20/2012 9:24:57 PM , Rating: 2
I'd say you're pretty close to being right, for most TV viewers, when it comes purely to the effect of the pixel count (that 1080p provides sufficient pixel density). Though for those with a home theater, a little more resolution would help. I have a 98" diagonal screen; to achieve the THX-recommended HDTV viewing angle of 36 degrees (for an immersive experience), I sit 11' away. At that distance, obtaining a resolution of 1 arc minute would require a 13% higher pixel density (in each dimension) than what HDTV currently provides (plus slightly improved optics on my projector). And for those who want larger viewing angles, for an even more immersive movie experience, THX recommends 40 degrees as the ideal, which corresponds to human horizontal vision; however here resolution is more noticeably compromised with 1080p.

But the real problem is that even mere 1080p (never mind 4K) is not adequately provided by the current HDTV standard. First, to fit on Blu-Ray, there's compression of ~50:1 to 100:1 vs. a raw 1080p camera feed. This compression is done in a very sophisticated way, and much of the image is static from frame to frame and doesn't need to be reproduced each time, but this compression is nevertheless not visually lossless. Further, HDTV doesn't give us a full color space (relative to what the eye can distinguish), and provides only 8 bits of color depth, which is easily distinguishable from 10 bits. Additionally, because of chroma subsampling, we don't get full 1080p resolution of the individual colors -- on my projector, I can see the difference between a 1080p 4:4:4 image from a jpeg, vs a 1080p 4:2:0 image from a Blu-Ray. So the 1080p standard is in need of updating, irrespective of the issue of pixel density. For more information on this, you can read:

That's not to say they should have done things differently -- given the bandwidth limits of Blu-Ray, significant compromises had to be made. My point is rather that, looking forward to the "Ultra HD" standard, there's a lot more to it than just more pixels -- my understanding is that it has a full color space and a higher color bit depth, both of which are needed (not sure what they are doing about chroma subsampling). But to deliver this without suffering even more compression loss than what we have now with Blu-Ray, they are going to have to significantly increase the pipeline and further improve the compression schemes. If they don't, then Ultra HD will not live up to its potential --thus some feel we'd get better quality from Ultra HD if we devoted less bandwidth to increased pixel count vs. other areas. Myself, I'd like to see both increased pixels/bit depth/color space, and less compression loss, but it remains to be seen how well they can deliver that.

Finally, note that Blu-Ray is the best a consumer can get -- most 1080p sources have less bandwidth than Blu-Ray, and are thus even more compromised. So the other huge challenge with Ultra HD will be how to get the non-optimal sources (cable, streaming) up to snuff.

RE: PC Monitors
By Digimonkey on 10/19/2012 10:02:54 AM , Rating: 2
it looks like Microsoft has really screwed up on the 720. We already know the GPU in the 720 is equivalent to a $50 GPU you can buy NOW.

Your source for this information is where?

RE: PC Monitors
By Kefner on 10/19/2012 10:23:55 AM , Rating: 2
I could have sworn I heard the new Xbox was going to support 4K and multiple screens. For the life of me I don't remember where I read that, but will look for it when I get home this evening (at work now, don't have time to search), unless someone here has seen it and can link to it.

RE: PC Monitors
By Milliamp on 10/19/2012 12:10:27 PM , Rating: 2
I think multiple screens is a really good idea. They could make a cheaper version that supports 1 screen and split screen and a more expensive version that supports 2 player games on 2 full sized screens.

They can't boost the hardware specs in a refresh but they could probably add support for a 2nd display.

RE: PC Monitors
By Milliamp on 10/19/2012 12:36:43 PM , Rating: 2
Great, a new version of Blu Ray already. Assuming an average 1080p stream is 5 mb/s 4k is going to need 20mb/s of throughput (not bandwidth).

I guess if there is one positive side to 4k (UHD) it is that there really isn't much/any need for anything after that for home use.

RE: PC Monitors
By kattanna on 10/19/2012 12:48:53 PM , Rating: 2
I guess if there is one positive side to 4k (UHD) it is that there really isn't much/any need for anything after that for home use.

i heard that from people when 1080 first came out..


RE: PC Monitors
By geddarkstorm on 10/19/2012 1:01:09 PM , Rating: 3
Why do we buy the mountain? Because it's there.

RE: PC Monitors
By Milliamp on 10/19/2012 2:45:15 PM , Rating: 2
Yes but with 4k (UHD?) you have to be 5 feet away from an 80 inch screen, or 3 feet away from a 60 inch screen to notice based on this article:

With that said I find the calibration of that diagram pretty questionable. Based on that you have to be 15 feet on a 50" TV before the "full benefit" of 480p is visible. I think for many people these are living room distances and there is a huge difference between 480p and HD even if your eye sight is wonky.

I personally don't think I would notice much difference between 1080p an 4k on my 50" TV but if I replace it with an 80" TV I might or start using one as a PC monitor I might.

If TV manufacturers need something else to set a goal for instead of racing to the bottom in pricing I say go for it. Maybe by the time my kid is old enough to play games he will have a couple of 80" 4k TV's in his game room for his friends to play photo realistic games on. This doesn't serve much purpose for average use but it moves us a little closer to that.

RE: PC Monitors
By MrTeal on 10/19/2012 3:24:44 PM , Rating: 2
With that said I find the calibration of that diagram pretty questionable. Based on that you have to be 15 feet on a 50" TV before the "full benefit" of 480p is visible. I think for many people these are living room distances and there is a huge difference between 480p and HD even if your eye sight is wonky.

There's a big difference, but I think the benefit of going to a 1080p was exaggerated for most people by all the other changes that were made at the same time. You're changing a lot of variables when you're going from SDTV over analog component cables to a tube TV vs 1080p signal over digital cabling to a 1080p TV.

Try watching a 1080p Youtube video (particularly an animated one that was rendered at hi res like at full resolution, and then change your monitor's resolution to VGA (640x480). The difference you'll see just from the resolution increase isn't nearly as large as you see going from something like SDTV sports to the HD version.

4k is going to be pretty underwhelming, IMO. People will limited on the size of TVs they can use by their room, and the benefit from a 4k display on a 40-50" TV won't be a huge leap from 1080p without other features.

RE: PC Monitors
By Shadowself on 10/19/2012 4:35:04 PM , Rating: 2
The Blu-ray spec is that the video imagery is capped at 40 Mbps with audio capped at 8 Mbps.

The average over a movie with high action sequences (think of things like the recent Avengers movie) in 1080p is likely to be < 20 Mbps (honestly, I haven't checked). However, high action sequences with lots and lots happening in each frame will usually smack up against that 40 Mbps limit for several seconds. For these kinds of movies the post production guys generating the Blu-ray versions spend a lot of time to live within that cap and still have no apparent blocking of the imagery.

If they implement H.265 rather than H.264 and go to this "4K" Ultra HD standard (as different from the Digital Cinema 4K standard) then they'll still need double the bandwidth (H.265 proponents claim that H.265 requires half the bandwidth, on average, as H.264).

Therefore in order for the "next generation of Blu-ray" [or whatever they end up calling it] will need a video bandwidth cap of at least 80 Mbps with a total storage of 100 GB. If they don't go with H.265 but stay with H.264, you need to double that -- 160 Mbps for video and 200 GB for the disc size.

And just so you know, there have been 200 GB working units in labs for over four years and 1 TB versions working units in the labs for almost as long. They just haven't become commercialized because there really was no consumer need for them.

And to just stop any lingering arguments before they start... No, your satellite or cable provider does not deliver Blu-ray quality imagery. The imagery they provide is much more highly compressed. And yes, some people will claim they cannot tell the difference, but that is a personal level thing. It is not that there is no difference or that many people cannot tell the difference.

RE: PC Monitors
By Silver2k7 on 10/20/2012 2:59:49 AM , Rating: 2
There are blu-ray recordable discs with 100GB and 128GB avalible right now with 3 & 4 layers.

The question is will those be used as the new standard for the next generation video players for the home or will there be something else.

RE: PC Monitors
By Reflex on 10/19/2012 1:30:34 PM , Rating: 2
Yeah, ever since Microsoft announced the specs to the next Xbox I have felt much the a second, they haven't announced any new device or specs, have they? You are just talking out of your ass, aren't you?

Never mind then...

RE: PC Monitors
By aju on 10/19/2012 1:48:15 PM , Rating: 1
Most likely the PS4's support is similar to that found in current home theater receivers such as the Onkyo HT-RC460 which supports upscaling to 4K. Chances are that the game is rendered in 1080p by a similar $50 GPU and then upscaled to 4K. After all, it would be tough for a current high end video card from Nvidia or AMD to pull off decent frame rates powering over 8.2 million pixels and 4k has at a minimum 8,224,400 pixels. For example the EVGA Geforce GTX 680 Classified only can handle a minimum frame rate of 29.7 for Crisis Warhead on a resolution of 5760x1200 that is only 6.9 million pixels. Throw in another 1.2 million pixels in and you asking for very poor performance. That card currently costs $559.00 on Newegg. There is no way Sony would put in a $500.00 GPU in a PS4 and that is what it would take even a year from now to support true 4K rendering of games. Chances are it will just be a cheap upscaler chip. If that is the case why pay $24,999.99 for a Sony XBR-84X900 display if you are not really seeing a true 4K rendered game.

RE: PC Monitors
By titanmiller on 10/22/2012 6:23:23 PM , Rating: 2
I personally would rather Microsoft and Sony hold off 24 months so that they can properly implement 4K rendering.

RE: PC Monitors
By BZDTemp on 10/19/2012 4:36:22 PM , Rating: 2
There is a chance the next gen. consoles will offer 4K. Though the current ones may be able to display 1080p but most games run lower res and are then upscaled so even if 4K is supported it may be more in name than in actual real resolution.

Anyway I'm so hoping this will mean 4K monitors in 30" size or there about.

Considering I had a 21"(20" viewable) monitor giving me 2048x1536 over a decade ago and that 30" 2560x1600* has been the max for some years now it is high time we see an real improvement.

* Not counting the specialist monitors that cost $10,00 and more.

RE: PC Monitors
By BansheeX on 10/19/2012 4:58:15 PM , Rating: 2
What's ridiculous about your comment or the desire for 4k gaming in general, is that 1080p games don't look like 1080p recordings of reality. That tells you that resolution isn't the problem, it's polygons and physics. We're still a long way away from games looking real and driving the resolution higher is simply going to take horsepower away from where it should be going.

Oh and I'll take a 1080p OLED over a 4k LCD any day of the week no hesitation. This is just like the 3D novelty all over again. To hell with LCD, we need a real improvement in display quality.

RE: PC Monitors
By BZDTemp on 10/20/2012 1:51:06 PM , Rating: 2
I think you need to cool down a bit.

When asking for 4K in way I suggested that other improvements are also desirable. In fact I actually raised the subject that 4K support could be in name only with regards to the new consoles.

As for you 1080p OLED over a 4K LCD then all I can say you should wait till the choice is possible before making up your mind. Right now your statement is just meaningless name dropping.

"A lot of people pay zero for the cellphone ... That's what it's worth." -- Apple Chief Operating Officer Timothy Cook

Copyright 2016 DailyTech LLC. - RSS Feed | Advertise | About Us | Ethics | FAQ | Terms, Conditions & Privacy Information | Kristopher Kubicki