UHD 101: Demystifying 4K, UHD Blu Ray, wide color gamut, HDR, 4:4:4, 18Gbps and the rest!
UHD is the next generation of video formats, and has many potentially picture enhancing features over our current HD format. As we’ll explain, 4K resolution is only part of the UHD concept. There are other important components such as wider color gamuts (more colors), higher color bit depths (smoother color gradations) and brighter image highlights (HDR).
Here are the items covered in our introductory section. Read this if you want the high level details:
- What UHD Sources are there?
- What is the Ultra HD Blu Ray specification?
- What capabilities does my display need to have to enjoy UHD content?
- Does my AVR / Pre-Pro need to be HDMI2.0 / HDCP2.2?
Then we get more technical, and explain some of the aspects of UHD:
- 4K resolution
- Wide color gamut
- Color bit depth
- Color subsampling
- HDMI versions
- HDMI data rates
- HDMI cables
- Display certification standards
What UHD Sources are there?
At this point a few hardware options:
- UHD Blu Ray players, such as the new Samsung UBD-K8500. You can buy discs for around $30 from Amazon or Best Buy right now in the US!
- Kaleidescape Strato – these players can download UHD content from the Kaleidescape store. The movies are created directly from master files received from the content creators and are not simple copies of UHD Blu Ray discs…in fact there is more UHD content on the store than there is currently UHD Blu Rays! The Strato supports HDR10 but not Dolby Vision, as well as the lossless audio codecs and Dolby Atmos.
- Sony FMP-X10 player, which we believe is end-of-life and does not support HDR.
For a comprehensive list of sources and available content, see the “Master List of 4K“… thread over at AVSForum.com.
Note that most of the streaming services have compressed audio soundtracks, even the ones advertising Dolby Atmos like Vudu. Since high performance home theater is about video and audio, that only leaves two sources of note: UHD Blu Ray and Kaleidescape’s Store. For more read My Search for Higher-Quality Soundtracks in Streaming Movies.
What is the Ultra HD Blu Ray specification?
The Ultra HD Blu Ray spec is as follows (see the rest of this article for details on how to “decipher” these acronyms):
- Up to 4K resolution
- 4:2:0 color sub-sampling
- Up to 10 bit color
- Up to 60 frames per second
- Support for wide color gamuts (REC.2020)
- Support for HDR10 and Dolby Vision
- No 3D support
As many of these specifications are optional it seems that just because a disc is labeled Ultra HD Blu Ray does not mean it will have HDR, a wide color gamut or 10 bit color. Hopefully UHD Blu Ray players will have some kind of signal information menu to reveal what is actually on the disc.
Note also that many of the movies announced for release on UHD Blu Ray so far were actually not shot at 4K resolution! See this article for some insightful analysis.
What capabilities does my display need to enjoy UHD content?
- HDMI2.0 / HDCP2.2. First and foremost you need a display capable of receiving a HDMI2.0 / HDCP2.2 signal. Unfortunately for consumers some early TVs sold as 4K capable do not have the required HDMI2.0 / HDCP2.2 inputs!
- 4K resolution. UHD is 4x the resolution of HDTV, typically 3840×2160 (2160p), although the Sony PJs have slightly larger, since they use imaging chips derived from their cinema projectors.
Whilst not strictly necessary for UHD your display may also have:
- Wide color gamut (WCG). HDTV and Blu-Rays are mastered and distributed in the REC.709 color gamut. With UHD content can be distributed in either the REC.709 or REC.2020 color gamuts. If REC.2020 is used that does not necessarily mean the content has been created using such a wide color gamut. We expect most of the “wide color gamut” content for the next few years to actually have a color gamut similar to the commercial cinema P3 space. REC.2020 can, in this instance, be considered as the container, rather than necessarily the actual color gamut of the encoded content. With this being said, your display must accept a REC.2020 signal to display WCG content. Apparently some displays released to date such as the Sony VPL-VW1100ES do not accept REC.2020 even though they are marketed as being able to reproduce a color gamut approaching P3. Confusing, we know, but that is the state of things with respect to 4K / UHD in 2016!
- High dynamic range (HDR). This is the ability to display images with highlights that are much brighter than today. These highlights are used for things like reflections off cars. Content has to be specifically mastered for HDR, so not all UHD content will necessarily be HDR. There are also at least three competing standards for HDR, including HDR10 and Dolby Vision and no broadcast standard. There’s also a new gamma curve (more correctly an electro-optical transfer function or EOTF), SMPTE ST2084. Whilst the UHD Alliance has mandated brightness standards for TVs there are no such standards for projectors. In addition whilst there is a mastering standard for HDR (1000 nit for HDR10) there is no standard or accepted practice for how to scale to the output levels that your display is capable of. Suffice to say it is early days for HDR!
Things there are confusion over with respect to UHD displays:
- 18Gbps. This may be required if content makers start releasing movies with high frame rates. 4K @ 60 frames per second may require 18Gbps data rates if it is at bit depths of 10 or 12 bits. Otherwise 10.2Gbps is fine, even for 4K60 at 8 bit!
- 12 bit color. This refers to the color bit-depth. Old Blu-Ray is 8 bit, new UHD Blu Ray can support 10 bit. It’s not clear what bit depth is being used by the streaming services, but it’s probably 8 bit.
- 4:4:4. This refers to color sub-sampling. Whilst 4:4:4 is used in content mastering, UHD is distributed via 4:2:0.
- 3D. There is no UHD 3D at this time, at least via Blu Ray.
Does my AVR / Pre-Pro need to be HDMI2.0 / HDCP2.2?
If you want to use your AVR or Pre-Pro to switch UHD sources then it needs to have HDMI2.0 (2.0a for HDR)/ HDCP2.2.
If your AVR/Pre-Pro does not have this then the workaround is to run one HDMI cable from the source to the display and another from the source to the AVR. Both the Samsung UHD and Kaleidescape Strato players provide this functionality.
This section covers most of the background behind the UHD specifications. We’ve tried to make it comprehensive, but beware that this means quite a few technical details! Where relevant we’ve linked to places where you can do further reading.
If you find any technical inaccuracies or things that require further clarification please leave a comment. It’s likely there are a few!
Resolution refers to the number of pixels displayed. Our current HD standard is 1920 horizontal and 1080 vertical. The new UHD standard is four times this amount, or 3840 horizontal x 2160 vertical.
Confusingly commercial cinema 4K as specified by the Digital Cinema Initiative (DCI) and VESA (the standard used by still and video cameras) is 4096 x 2160, which is slightly wider than Quad HD / UHD at 3840 x 2160.
Wide Color Gamut (REC.709, P3, REC.2020)
Nearly all current consumer content, whether HDTV over cable or satellite, DVD or Blu Ray is created and distributed in the REC.709 color gamut.
The issue with this is that it only represents a small portion of the visible spectrum of colors. The standard used in commercial cinema has more colors and is called DCI P3. There is also an even wider color gamut called REC.2020 which is the “wide color gamut” standard used in UHD.
Note that just because content is labeled as UHD does not mean it has a wide color gamut. Much of the initial UHD content actually has the same REC.709 color gamut that we have today.
WCG content comes in a REC.2020 container. As we mentioned earlier not every display can accept REC.2020. If your display cannot accept REC.2020 then the source should “downconvert” to the REC.709 space. The new Samsung UHD Blu-Ray player does this.
There is some confusion over the P3 color space – whist the wide color gamut content actually encoded onto the disc may have a color gamut close or equivalent to the P3 space in UHD it is transmitted inside a REC.2020 container. There is no consumer P3 color space. The confusion arises because display manufacturers are now advertising % of P3, and this is also a specification that has been codified in the Ultra HD Premium TV certification. We guess the assumption being made is that despite the container being REC.2020 what we will see for the forseeable future is that the actual colors in the content are equivalent to the P3 space.
Color Bit Depth (8 bit, 10 bit, 12 bit)
Bit depth describes the number of potential values that the encoding of color information in a video signal can have.
Historically, Blu Ray has been 8 bit, which means 256 possible values for red, green and blue. UHD Blu Ray is 10 bit, giving 1024 values for RGB. 12 bit color provides 4096 values for RGB.
One important reason that we have moved to a 10 bit system for UHD Blu Ray is to reduce color banding. This is an image defect where bands of color are visible. It’s more important in the UHD world because of the expanded color space and hence the greater color variations.
HDMI2.0 supports 8, 10 and 12 bit color in various formats, as covered in the following sections.
Sometimes you will see references to 24, 30 and 36 bit color. These references relate to the total color bit depth for all RGB channels, where 24 bit = 8 bit red, 8 bit green, 8 bit blue and 36 bit = 12 bit red, 12 bit green, 12 bit blue.
Color Subsampling (4:2:0, 4:2:2, 4:4:4)
Consumer video is stored, transmitted, and processed in a color space called Y’CbCr. The three components stand for:
- Y’ = Luminance, or Luma, representing the brightness of the pixel
- Cb = Blue color difference
- Cr = Red color difference
This standard was defined back at the start of the color TV era as a way of including color information in the broadcast signal.
Encoding in the Y’CbCr color space allows the resolution of the color channels Cb and Cr to be reduced through color or chroma subsampling. This technique takes advantage of the fact that human vision is more sensitive to light differences than to color differences.
There are three main types of color subsampling used today. These are 4:4:4, 4:2:2 and 4:2:0.
If 4:4:4 is a full bandwidth signal, then 4:2:2 occupies 2/3rds the space and 4:2:0 occupies 1/2 the space. Blu-ray and UHD Blu-ray both store the video signal in the 4:2:0 format. This essentially means that each pixel has a Y’ signal, odd pixel lines have no Cb or Cr and then Cb and Cr are alternated on each even pixel line.
To get the video displayed on the TV or Projector it typically goes through the following conversions:
- 4:2:0 to 4:2:2
- 4:2:2 to 4:4:4
- Y’CbCr to RGB
Historically the source upsamples to 4:2:2, which is sent over HDMI, and then the display upsamples to 4:4:4 and converts to RGB. The reason for this sequence is that HDMI v1.4 and previous iterations did not support 4:2:0. HDMI2.0 does support 4:2:0 though only at 50/60 frames per second (FPS). At 24 FPS 10 bit only 4:4:4 and RGB are supported.
Note that there appears to be some confusion about exactly what is supported in the HDMI specification, even among manufacturers and industry participants. Some people we discussed this article said 4:2:0 was supported at 24/25/30, others said that as of 2.0a 4:2:2 was supported at 10 bit. In the absence of a clear industry wide understanding we will stick to the information published on the HDMI website.
There is no intrinsic benefit to the source upsampling to 4:4:4 or converting to RGB. With respect to UHD and HDMI it is actually beneficial if the conversion to 4:2:2, 4:4:4 and RGB is, as much as possible, left to the display, as this reduces the HDMI bandwidth requirements.
- Choosing a color space, by Spears & Munsil (note that some of the information for HDMI is out of date, as it was written for 1.4)
Placeholder for future section on HDMI handshaking between source and display and how this impacts data rates, color bit depth, color gamut and color space…
HDMI Versions (1.4, 2.0 and beyond)
There have been essentially three types of of “4K capable” HDMI chipsets on the market. These have been implemented into various TVs, projectors, processors, AVRs and switchers since the 4K came onto the market in 2013.
- HDMI 1.4 chipsets supported data rates up to 10.2Gbps. This means they could do 4K at up to 30 frames per second at the formats supported by the 1.4 standard (8 bit RGB or 4:4:4 and 12 bit 4:2:2). Whilst 1.4 is therefore theoretically 4K capable nearly all components that have it lack HDCP2.2, which is the copy protection scheme the industry has settled on for UHD.
- We then saw HDMI 2.0 chipsets that supported the new formats in the 2.0 standard but were still limited to data rates of 10.2Gbps. These new chipsets have HDCP2.2 so they can display some but potentially not all UHD content.
- Finally at the end of 2015 we started to see 18Gbps chipsets.
We’ve summarized this information together with the formats and bandwidth requirements in the table below:
This is important information, as we are seeing some sources such as the Samsung UHD Blu-Ray player that only support 4:4:4 or RGB at 10 bit, both of which require the whole video chain to be 18Gbps capable. They could easily have used 12 bit at 4:2:2:, which would have enabled compatibility with 10.2Gbps chipsets.
HDCP (1.4, 2.2)
HDCP stands for High-Bandwidth Digital Content Protection. The version used in UHD is 2.2.
HDMI data rates (10.2Gbps, 18 Gbps)
Part of the deal with UHD is the potential requirement of using HDMI chipsets and HDMI cables that support “18Gbps” data rates.
The table below summarizes the data rates for the different frame rates and formats that are part of the HDMI2.0 specification.
As you can see it is possible that a 10.2Gbps chipset and cable infrastructure can support UHD Blu Ray, assuming the transfer medium is 4:2:2 at 12 bit.
For short runs (say up to 6′ / 2m) most passive cables will be able to support the data rates required for UHD. Between 6′ and 15′ you’ll find some passive cables that can support UHD and others that can’t. Above 15′ you’ll very likely need an active cable.
There are a few independent companies that are providing HDMI cable certification for UHD data rates. These include:
- HDMI, “Premium Certified Cable”, at the 18Gbps rate
- UL Lab, “High Speed 4K Cable”, at the 10.2 and 18 Gbps rates
- DPL Labs, “DPL Seal of Approval”, at the 10.2 and 18 Gbps rates
HDBaseT (HDMI over category cable) is capable of data rates up to 10.2Gbps.
Display Certification Standards
Ultra HD Premium
The main display certification standard is one created by the UHD Alliance. This is an industry consortium of content creators (e.g. Hollywood Studios), distributors (e.g. Amazon, DirecTV) and hardware manufacturers (e.g. LG, TCL). They have created a set of specifications and a logo to help consumers.
It’s not clear at present if this “badge” will just get applied to consumer displays, or if it’s intended to be used across the content creation and distribution ecosystem as well.
To get a Ultra HD Premium “sticker” a display must meet a set of criteria, these are:
- Image Resolution: 3840×2160
- Color Bit Depth: 10-bit signal
- Color Palette (Wide Color Gamut)
- Signal Input: BT.2020 color representation
- Display Reproduction: More than 90% of P3 colors
- High Dynamic Range
- SMPTE ST2084 EOTF
- A combination of peak brightness and black level either:
- More than 1000 nits peak brightness and less than 0.05 nits black level
- More than 540 nits peak brightness and less than 0.0005 nits black level (note the interesting “fudge” here, clearly something included in the spec for low light output OLED TVs…)
- More than 1000 nits peak brightness and less than 0.05 nits black level
Interestingly, Sony, despite being a member of the UHD Premium alliance, has a different “sticker” that it is putting on it’s displays. We think they are doing this because they also have projectors, and to have a consistent “sticker” they need their own standard, since the projectors can’t hit the peak brightness / black level standards required for the UHD Premium “sticker”.
High Dynamic Range (HDR) Standards
There are multiple HDR standards at this point, and it is not clear which one will become dominant in the market. HDR10 and Dolby Vision appear to be the front runners, but there are others lurking in the wings such as Hybrid Log Gamma (HLG).
It’s quite the “evolving ecosystem” at this point. Even if you buy a display with both HDR10 and Dolby Vision (which limits you to flat panel TVs), the amount of light output the TV can put out will likely be on an upward trajectory for the next few years. The TVs you can buy now are limited to about 1,000 nits, but the Dolby Vision standard can see a future with 10,000 nit displays!
This section covers the two main standards at a high level. We’d also recommend reading the State of HDR article if you really want the technical details (warning, it’s even more technical than what we’ve written).
Be aware that there are a number of articles out there about HDR, many of which were written in 2015 before the slightly clearer picture that emerged late in 2015 after the publication of the SMPTE HDR Imaging Ecosystem report. As such you’ll likely find conflicting and incomplete information. It’s likely that some of the things we have written are incorrect or incomplete too, so please leave a comment if you find things we should update or clarify!
HDR10 is an open standard.
- Appears to be fewer standards around production and playback than Dolby Vision. For example there is no dynamic metadata like with the Dolby Vision standard that maps high brightness areas of the image to the display’s capabilities. It is left to each TV manufacturer to work out how to do that.
- Content typically mastered to 1,000 nits and 10 bit color.
- Requires HDMI2.0a.
- It is mandatory that UHD HDR Blu-Rays include HDR10 metadata, even if they also have Dolby Vision. This requirement does not extend to other sources of HDR content, such as VUDU, which we don’t believe has HDR10 metadata.
- In many 2016 TVs and some projectors support HDR10 – see this HDR-Capable Display list for a comprehensive and updated list.
- Specifies that the goal is for the cinematic master be done in 12 bit color and with 10,000 nit brightness. Note that there are no displays that can do 10k nits, so presently content is being mastered to 4k nits, primarily using the Dolby PRM-4220 monitor. It also seems like 12 bit is not a necessity for Dolby Vision, and 10 bit may be recommended for broadcast.
- The content creator can specify in metadata how to display the specular highlights. For example they can create a 4,000 nit master and then also create “variants” for displays that can only do 1,000 nits. Apparently this is an improvement over HDR10, which leaves it up to the TV to figure out how to map high brightness content to the display’s capabilities.
- Requires dedicated silicon, and therefore if a display does not have it on launch it is not going to get it via a firmware update.
- Does not require HDMI2.0a. The Vizio Reference TVs do not have 2.0a.
- The streaming service VUDU UHD is using Dolby Vision, as are Netflix for some of their original content shows (Marco Polo).
- It does not seem that any UHD Blu-Ray players have Dolby Vision at this time, and therefore there are also no Dolby Vision Blu-Ray discs…please correct us if we are wrong.
- Can be found in TVs from LG (ALL 2016 OLEDs – B, C, E and G Series), Vizio (Reference Series) and TCL
- No consumer projectors have Dolby Vision.