Advanced Video Coding


Advanced Video Coding, also referred to as H.264 or MPEG-4 Part 10, Advanced Video Coding, is a video compression standard based on block-oriented, motion-compensated integer-DCT coding. It is by far the most commonly used format for the recording, compression, and distribution of video content, used by 91% of video industry developers. It supports resolutions up to and including 8K UHD.
The intent of the H.264/AVC project was to create a standard capable of providing good video quality at substantially lower bit rates than previous standards, without increasing the complexity of design so much that it would be impractical or excessively expensive to implement. This was achieved with features such as a reduced-complexity integer discrete cosine transform, variable block-size segmentation, and multi-picture inter-picture prediction. An additional goal was to provide enough flexibility to allow the standard to be applied to a wide variety of applications on a wide variety of networks and systems, including low and high bit rates, low and high resolution video, broadcast, DVD storage, RTP/IP packet networks, and ITU-T multimedia telephony systems. The H.264 standard can be viewed as a "family of standards" composed of a number of different profiles, although its "High profile" is by far the mostly commonly used format. A specific decoder decodes at least one, but not necessarily all profiles. The standard describes the format of the encoded data and how the data is decoded, but it does not specify algorithms for encoding video that is left open as a matter for encoder designers to select for themselves, and a wide variety of encoding schemes has been developed. H.264 is typically used for lossy compression, although it is also possible to create truly lossless-coded regions within lossy-coded pictures or to support rare use cases for which the entire encoding is lossless.
H.264 was standardized by the ITU-T Video Coding Experts Group of Study Group 16 together with the ISO/IEC JTC1 Moving Picture Experts Group. The project partnership effort is known as the Joint Video Team. The ITU-T H.264 standard and the ISO/IEC MPEG-4 AVC standard are jointly maintained so that they have identical technical content. The final drafting work on the first version of the standard was completed in May 2003, and various extensions of its capabilities have been added in subsequent editions. High Efficiency Video Coding, a.k.a. H.265 and MPEG-H Part 2 is a successor to H.264/MPEG-4 AVC developed by the same organizations, while earlier standards are still in common use.
H.264 is perhaps best known as being the most commonly used video encoding format on Blu-ray Discs. It is also widely used by streaming Internet sources, such as videos from Netflix, Hulu, Prime Video, Vimeo, YouTube, and the iTunes Store, Web software such as the Adobe Flash Player and Microsoft Silverlight, and also various HDTV broadcasts over terrestrial, cable, and satellite systems.
H.264 is protected by patents owned by various parties. A license covering most patents essential to H.264 is administered by a patent pool administered by MPEG LA.
The commercial use of patented H.264 technologies requires the payment of royalties to MPEG LA and other patent owners. MPEG LA has allowed the free use of H.264 technologies for streaming Internet video that is free to end users, and Cisco Systems pays royalties to MPEG LA on behalf of the users of binaries for its open source H.264 encoder.

Naming

The H.264 name follows the ITU-T naming convention, where the standard is a member of the H.26x line of VCEG video coding standards; the MPEG-4 AVC name relates to the naming convention in ISO/IEC MPEG, where the standard is part 10 of ISO/IEC 14496, which is the suite of standards known as MPEG-4. The standard was developed jointly in a partnership of VCEG and MPEG, after earlier development work in the ITU-T as a VCEG project called H.26L. It is thus common to refer to the standard with names such as H.264/AVC, AVC/H.264, H.264/MPEG-4 AVC, or MPEG-4/H.264 AVC, to emphasize the common heritage. Occasionally, it is also referred to as "the JVT codec", in reference to the Joint Video Team organization that developed it. Some software programs internally identify this standard as AVC1.

History

Overall history

In early 1998, the Video Coding Experts Group issued a call for proposals on a project called H.26L, with the target to double the coding efficiency in comparison to any other existing video coding standards for a broad variety of applications. VCEG was chaired by Gary Sullivan. The first draft design for that new standard was adopted in August 1999. In 2000, Thomas Wiegand became VCEG co-chair.
In December 2001, VCEG and the Moving Picture Experts Group formed a Joint Video Team, with the charter to finalize the video coding standard. Formal approval of the specification came in March 2003. The JVT was chaired by Gary Sullivan, Thomas Wiegand, and Ajay Luthra. In July 2004, the Fidelity Range Extensions project was finalized. From January 2005 to November 2007, the JVT was working on an extension of H.264/AVC towards scalability by an Annex called Scalable Video Coding. The JVT management team was extended by Jens-Rainer Ohm. From July 2006 to November 2009, the JVT worked on Multiview Video Coding, an extension of H.264/AVC towards 3D television and limited-range free-viewpoint television. That work included the development of two new profiles of the standard: the Multiview High Profile and the Stereo High Profile.
Throughout the development of the standard, additional messages for containing supplemental enhancement information have been developed. SEI messages can contain various types of data that indicate the timing of the video pictures or describe various properties of the coded video or how it can be used or enhanced. SEI messages are also defined that can contain arbitrary user-defined data. SEI messages do not affect the core decoding process, but can indicate how the video is recommended to be post-processed or displayed. Some other high-level properties of the video content are conveyed in video usability information, such as the indication of the color space for interpretation of the video content. As new color spaces have been developed, such as for high dynamic range and wide color gamut video, additional VUI identifiers have been added to indicate them.

Fidelity range extensions and professional profiles

The standardization of the first version of H.264/AVC was completed in May 2003. In the first project to extend the original standard, the JVT then developed what was called the Fidelity Range Extensions. These extensions enabled higher quality video coding by supporting increased sample bit depth precision and higher-resolution color information, including the sampling structures known as Y′CBCR 4:2:2 and 4:4:4. Several other features were also included in the FRExt project, such as adding an 8×8 integer discrete cosine transform with adaptive switching between the 4×4 and 8×8 transforms, encoder-specified perceptual-based quantization weighting matrices, efficient inter-picture lossless coding, and support of additional color spaces. The design work on the FRExt project was completed in July 2004, and the drafting work on them was completed in September 2004.
Five other new profiles intended primarily for professional applications were then developed, adding extended-gamut color space support, defining additional aspect ratio indicators, defining two additional types of "supplemental enhancement information", and deprecating one of the prior FRExt profiles that industry feedback indicated should have been designed differently.

Scalable video coding

The next major feature added to the standard was Scalable Video Coding. Specified in Annex G of H.264/AVC, SVC allows the construction of bitstreams that contain layers of sub-bitstreams that also conform to the standard, including one such bitstream known as the "base layer" that can be decoded by a H.264/AVC codec that does not support SVC. For temporal bitstream scalability, complete access units are removed from the bitstream when deriving the sub-bitstream. In this case, high-level syntax and inter-prediction reference pictures in the bitstream are constructed accordingly. On the other hand, for spatial and quality bitstream scalability, the NAL is removed from the bitstream when deriving the sub-bitstream. In this case, inter-layer prediction is typically used for efficient coding. The Scalable Video Coding extensions were completed in November 2007.

Multiview video coding

The next major feature added to the standard was Multiview Video Coding. Specified in Annex H of H.264/AVC, MVC enables the construction of bitstreams that represent more than one view of a video scene. An important example of this functionality is stereoscopic 3D video coding. Two profiles were developed in the MVC work: Multiview High profile supports an arbitrary number of views, and Stereo High profile is designed specifically for two-view stereoscopic video. The Multiview Video Coding extensions were completed in November 2009.

3D-AVC and MFC stereoscopic coding

Additional extensions were later developed that included 3D video coding with joint coding of depth maps and texture, multi-resolution frame-compatible stereoscopic and 3D-MFC coding, various additional combinations of features, and higher frame sizes and frame rates.

Versions

Versions of the H.264/AVC standard include the following completed revisions, corrigenda, and amendments. Each version represents changes relative to the next lower version that is integrated into the text.

Applications

The H.264 video format has a very broad application range that covers all forms of digital compressed video from low bit-rate Internet streaming applications to HDTV broadcast and Digital Cinema applications with nearly lossless coding. With the use of H.264, bit rate savings of 50% or more compared to MPEG-2 Part 2 are reported. For example, H.264 has been reported to give the same Digital Satellite TV quality as current MPEG-2 implementations with less than half the bitrate, with current MPEG-2 implementations working at around 3.5 Mbit/s and H.264 at only 1.5 Mbit/s. Sony claims that 9 Mbit/s AVC recording mode is equivalent to the image quality of the HDV format, which uses approximately 18–25 Mbit/s.
To ensure compatibility and problem-free adoption of H.264/AVC, many standards bodies have amended or added to their video-related standards so that users of these standards can employ H.264/AVC. Both the Blu-ray Disc format and the now-discontinued HD DVD format include the H.264/AVC High Profile as one of three mandatory video compression formats. The Digital Video Broadcast project approved the use of H.264/AVC for broadcast television in late 2004.
The Advanced Television Systems Committee standards body in the United States approved the use of H.264/AVC for broadcast television in July 2008, although the standard is not yet used for fixed ATSC broadcasts within the United States. It has also been approved for use with the more recent ATSC-M/H standard, using the AVC and SVC portions of H.264.
The CCTV and Video Surveillance markets have included the technology in many products.
Many common DSLRs use H.264 video wrapped in QuickTime MOV containers as the native recording format.

Derived formats

is a high-definition recording format designed by Sony and Panasonic that uses H.264.
AVC-Intra is an intraframe-only compression format, developed by Panasonic.
XAVC is a recording format designed by Sony that uses level 5.2 of H.264/MPEG-4 AVC, which is the highest level supported by that video standard. XAVC can support 4K resolution at up to 60 frames per second. Sony has announced that cameras that support XAVC include two CineAlta cameras—the Sony PMW-F55 and Sony PMW-F5. The Sony PMW-F55 can record XAVC with 4K resolution at 30 fps at 300 Mbit/s and 2K resolution at 30 fps at 100 Mbit/s. XAVC can record 4K resolution at 60 fps with 4:2:2 chroma sampling at 600 Mbit/s.

Design

Features

H.264/AVC/MPEG-4 Part 10 contains a number of new features that allow it to compress video much more efficiently than older standards and to provide more flexibility for application to a wide variety of network environments. In particular, some such key features include:
These techniques, along with several others, help H.264 to perform significantly better than any prior standard under a wide variety of circumstances in a wide variety of application environments. H.264 can often perform radically better than MPEG-2 video—typically obtaining the same quality at half of the bit rate or less, especially on high bit rate and high resolution video content.
Like other ISO/IEC MPEG video standards, H.264/AVC has a reference software implementation that can be freely downloaded. Its main purpose is to give examples of H.264/AVC features, rather than being a useful application per se. Some reference hardware design work has also been conducted in the Moving Picture Experts Group.
The above-mentioned aspects include features in all profiles of H.264. A profile for a codec is a set of features of that codec identified to meet a certain set of specifications of intended applications. This means that many of the features listed are not supported in some profiles. Various profiles of H.264/AVC are discussed in next section.

Profiles

The standard defines several sets of capabilities, which are referred to as profiles, targeting specific classes of applications. These are declared using a profile code and sometimes a set of additional constraints applied in the encoder. The profile code and indicated constraints allow a decoder to recognize the requirements for decoding that specific bitstream. By far the most commonly used profile is the High Profile.
Profiles for non-scalable 2D video applications include the following:
;Constrained Baseline Profile : Primarily for low-cost applications, this profile is most typically used in videoconferencing and mobile applications. It corresponds to the subset of features that are in common between the Baseline, Main, and High Profiles.
;Baseline Profile : Primarily for low-cost applications that require additional data loss robustness, this profile is used in some videoconferencing and mobile applications. This profile includes all features that are supported in the Constrained Baseline Profile, plus three additional features that can be used for loss robustness. The importance of this profile has faded somewhat since the definition of the Constrained Baseline Profile in 2009. All Constrained Baseline Profile bitstreams are also considered to be Baseline Profile bitstreams, as these two profiles share the same profile identifier code value.
;Extended Profile : Intended as the streaming video profile, this profile has relatively high compression capability and some extra tricks for robustness to data losses and server stream switching.
;Main Profile : This profile is used for standard-definition digital TV broadcasts that use the MPEG-4 format as defined in the DVB standard. It is not, however, used for high-definition television broadcasts, as the importance of this profile faded when the High Profile was developed in 2004 for that application.
;High Profile : The primary profile for broadcast and disc storage applications, particularly for high-definition television applications.
;Progressive High Profile : Similar to the High profile, but without support of field coding features.
;Constrained High Profile : Similar to the Progressive High profile, but without support of B slices.
;High 10 Profile : Going beyond typical mainstream consumer product capabilities, this profile builds on top of the High Profile, adding support for up to 10 bits per sample of decoded picture precision.
;High 422 Profile : Primarily targeting professional applications that use interlaced video, this profile builds on top of the High 10 Profile, adding support for the 4:2:2 chroma sampling format while using up to 10 bits per sample of decoded picture precision.
;High 444 Predictive Profile : This profile builds on top of the High 4:2:2 Profile, supporting up to 4:4:4 chroma sampling, up to 14 bits per sample, and additionally supporting efficient lossless region coding and the coding of each picture as three separate color planes.
For camcorders, editing, and professional applications, the standard contains four additional Intra-frame-only profiles, which are defined as simple subsets of other corresponding profiles. These are mostly for professional applications:
;High 10 Intra Profile : The High 10 Profile constrained to all-Intra use.
;High 422 Intra Profile : The High 4:2:2 Profile constrained to all-Intra use.
;High 444 Intra Profile : The High 4:4:4 Profile constrained to all-Intra use.
;CAVLC 444 Intra Profile : The High 4:4:4 Profile constrained to all-Intra use and to CAVLC entropy coding.
As a result of the Scalable Video Coding extension, the standard contains five additional scalable profiles, which are defined as a combination of a H.264/AVC profile for the base layer and tools that achieve the scalable extension:
;Scalable Baseline Profile : Primarily targeting video conferencing, mobile, and surveillance applications, this profile builds on top of the Constrained Baseline profile to which the base layer must conform. For the scalability tools, a subset of the available tools is enabled.
;Scalable Constrained Baseline Profile : A subset of the Scalable Baseline Profile intended primarily for real-time communication applications.
;Scalable High Profile : Primarily targeting broadcast and streaming applications, this profile builds on top of the H.264/AVC High Profile to which the base layer must conform.
;Scalable Constrained High Profile : A subset of the Scalable High Profile intended primarily for real-time communication applications.
;Scalable High Intra Profile : Primarily targeting production applications, this profile is the Scalable High Profile constrained to all-Intra use.
As a result of the Multiview Video Coding extension, the standard contains two multiview profiles:
;Stereo High Profile : This profile targets two-view stereoscopic 3D video and combines the tools of the High profile with the inter-view prediction capabilities of the MVC extension.
;Multiview High Profile : This profile supports two or more views using both inter-picture and MVC inter-view prediction, but does not support field pictures and macroblock-adaptive frame-field coding.
The Multi-resolution Frame-Compatible extension added two more profiles:
;MFC High Profile : A profile for stereoscopic coding with two-layer resolution enhancement.
;MFC Depth High Profile :
The 3D-AVC extension added two more profiles:
;Multiview Depth High Profile : This profile supports joint coding of depth map and video texture information for improved compression of 3D video content.
;Enhanced Multiview Depth High Profile : An enhanced profile for combined multiview coding with depth information.

Feature support in particular profiles

Levels

As the term is used in the standard, a "level" is a specified set of constraints that indicate a degree of required decoder performance for a profile. For example, a level of support within a profile specifies the maximum picture resolution, frame rate, and bit rate that a decoder may use. A decoder that conforms to a given level must be able to decode all bitstreams encoded for that level and all lower levels.
Level
Maximum
decoding speed
Maximum
frame size
Maximum video
bit rate for video
coding layer

Examples for high resolution
@ highest frame rate
Toggle additional details

11,4859964
128×96@30.9
176×144@15.0
1b1,48599128
128×96@30.9
176×144@15.0
1.13,000396192
176×144@30.3
320×240@10.0
352×288@7.5
1.26,000396384
320×240@20.0
352×288@15.2
1.311,880396768
320×240@36.0
352×288@30.0
211,8803962,000
320×240@36.0
352×288@30.0
2.119,8007924,000
352×480@30.0
352×576@25.0
2.220,2501,6204,000
352×480@30.7
352×576@25.6
720×480@15.0
720×576@12.5
340,5001,62010,000
352×480@61.4
352×576@51.1
720×480@30.0
720×576@25.0
3.1108,0003,60014,000
720×480@80.0
720×576@66.7
1,280×720@30.0
3.2216,0005,12020,000
1,280×720@60.0
1,280×1,024@42.2
4245,7608,19220,000
1,280×720@68.3
1,920×1,080@30.1
2,048×1,024@30.0
4.1245,7608,19250,000
1,280×720@68.3
1,920×1,080@30.1
2,048×1,024@30.0
4.2522,2408,70450,000
1,280×720@145.1
1,920×1,080@64.0
2,048×1,080@60.0
5589,82422,080135,000
1,920×1,080@72.3
2,048×1,024@72.0
2,048×1,080@67.8
2,560×1,920@30.7
3,672×1,536@26.7
5.1983,04036,864240,000
1,920×1,080@120.5
2,560×1,920@51.2
3,840×2,160@31.7
4,096×2,048@30.0
4,096×2,160@28.5
4,096×2,304@26.7
5.22,073,60036,864240,000
1,920×1,080@172.0
2,560×1,920@108.0
3,840×2,160@66.8
4,096×2,048@63.3
4,096×2,160@60.0
4,096×2,304@56.3
64,177,920139,264240,000
3,840×2,160@128.9
7,680×4,320@32.2
8,192×4,320@30.2
6.18,355,840139,264480,000
3,840×2,160@257.9
7,680×4,320@64.5
8,192×4,320@60.4
6.216,711,680139,264800,000
3,840×2,160@300.0
7,680×4,320@128.9
8,192×4,320@120.9

The maximum bit rate for the High Profile is 1.25 times that of the Constrained Baseline, Baseline, Extended and Main Profiles; 3 times for Hi10P, and 4 times for Hi422P/Hi444PP.
The number of luma samples is 16×16=256 times the number of macroblocks.

Decoded picture buffering

Previously encoded pictures are used by H.264/AVC encoders to provide predictions of the values of samples in other pictures. This allows the encoder to make efficient decisions on the best way to encode a given picture. At the decoder, such pictures are stored in a virtual decoded picture buffer. The maximum capacity of the DPB, in units of frames, as shown in parentheses in the right column of the table above, can be computed as follows:
Where MaxDpbMbs is a constant value provided in the table below as a function of level number, and PicWidthInMbs and FrameHeightInMbs are the picture width and frame height for the coded video data, expressed in units of macroblocks. This formula is specified in sections A.3.1.h and A.3.2.f of the 2017 edition of the standard.

Level
1
1b
1.1
1.2
1.3
2
2.1
2.2
3
3.1
3.2
4
4.1
4.2
5
5.1
5.2
6
6.1
6.2
MaxDpbMbs
396
396
900
2,376
2,376
2,376
4,752
8,100
8,100
18,000
20,480
32,768
32,768
34,816
110,400
184,320
184,320
696,320
696,320
696,320

For example, for an HDTV picture that is 1,920 samples wide and 1,080 samples high, a Level 4 decoder has a maximum DPB storage capacity of floor = 4 frames. Thus, the value 4 is shown in parentheses in the table above in the right column of the row for Level 4 with the frame size 1920×1080.
It is important to note that the current picture being decoded is not included in the computation of DPB fullness. Thus, a decoder needs to actually have sufficient memory to handle one frame more than the maximum capacity of the DPB as calculated above.

Implementations

In 2009, the HTML5 working group was split between supporters of Ogg Theora, a free video format which is thought to be unencumbered by patents, and H.264, which contains patented technology. As late as July 2009, Google and Apple were said to support H.264, while Mozilla and Opera support Ogg Theora. Microsoft, with the release of Internet Explorer 9, has added support for HTML 5 video encoded using H.264. At the Gartner Symposium/ITXpo in November 2010, Microsoft CEO Steve Ballmer answered the question "HTML 5 or Silverlight?" by saying "If you want to do something that is universal, there is no question the world is going HTML5." In January 2011, Google announced that they were pulling support for H.264 from their Chrome browser and supporting both Theora and WebM/VP8 to use only open formats.
On March 18, 2012, Mozilla announced support for H.264 in Firefox on mobile devices, due to prevalence of H.264-encoded video and the increased power-efficiency of using dedicated H.264 decoder hardware common on such devices. On February 20, 2013, Mozilla implemented support in Firefox for decoding H.264 on Windows 7 and above. This feature relies on Windows' built in decoding libraries. Firefox 35.0, released on January 13, 2015 supports H.264 on OS X 10.6 and higher.
On October 30, 2013, Rowan Trollope from Cisco Systems announced that Cisco would release both binaries and source code of an H.264 video codec called OpenH264 under the Simplified BSD license, and pay all royalties for its use to MPEG LA for any software projects that use Cisco's precompiled binaries, thus making Cisco's OpenH264 binaries free to use. However, any software projects that use Cisco's source code instead of its binaries would be legally responsible for paying all royalties to MPEG LA. Current target CPU architectures are x86 and ARM, and current target operating systems are Linux, Windows XP and later, Mac OS X, and Android; iOS is notably absent from this list, because it doesn't allow applications to fetch and install binary modules from the Internet. Also on October 30, 2013, Brendan Eich from Mozilla wrote that it would use Cisco's binaries in future versions of Firefox to add support for H.264 to Firefox where platform codecs are not available.
Cisco published the source to OpenH264 on December 9, 2013.

Software encoders

Hardware

Because H.264 encoding and decoding requires significant computing power in specific types of arithmetic operations, software implementations that run on general-purpose CPUs are typically less power efficient. However, the latest quad-core general-purpose x86 CPUs have sufficient computation power to perform real-time SD and HD encoding. Compression efficiency depends on video algorithmic implementations, not on whether hardware or software implementation is used. Therefore, the difference between hardware and software based implementation is more on power-efficiency, flexibility and cost. To improve the power efficiency and reduce hardware form-factor, special-purpose hardware may be employed, either for the complete encoding or decoding process, or for acceleration assistance within a CPU-controlled environment.
CPU based solutions are known to be much more flexible, particularly when encoding must be done concurrently in multiple formats, multiple bit rates and resolutions, and possibly with additional features on container format support, advanced integrated advertising features, etc. CPU based software solution generally makes it much easier to load balance multiple concurrent encoding sessions within the same CPU.
The 2nd generation Intel "Sandy Bridge" Core i3/i5/i7 processors introduced at the January 2011 CES offer an on-chip hardware full HD H.264 encoder, known as Intel Quick Sync Video.
A hardware H.264 encoder can be an ASIC or an FPGA.
ASIC encoders with H.264 encoder functionality are available from many different semiconductor companies, but the core design used in the ASIC is typically licensed from one of a few companies such as Chips&Media, Allegro DVT, On2, Imagination Technologies, NGCodec. Some companies have both FPGA and ASIC product offerings.
Texas Instruments manufactures a line of ARM + DSP cores that perform DSP H.264 BP encoding 1080p at 30fps. This permits flexibility with respect to codecs while being more efficient than software on a generic CPU.

Licensing

In countries where patents on software algorithms are upheld, vendors and commercial users of products that use H.264/AVC are expected to pay patent licensing royalties for the patented technology that their products use. This applies to the Baseline Profile as well.
A private organization known as MPEG LA, which is not affiliated in any way with the MPEG standardization organization, administers the licenses for patents applying to this standard, as well as other patent pools, such as for MPEG-4 Part 2 Video, HEVC and MPEG-DASH. The patent holders include Fujitsu, Panasonic, Sony, Mitsubishi, Apple, Columbia University, KAIST, Dolby, Google, JVC Kenwood, LG Electronics, Microsoft, NTT Docomo, Philips, Samsung, Sharp, Toshiba and ZTE, although the majority of patents in the pool are held by Panasonic, Godo Kaisha IP Bridge and LG Electronics.
On August 26, 2010, MPEG LA announced that royalties won't be charged for H.264 encoded Internet video that is free to end users. All other royalties remain in place, such as royalties for products that decode and encode H.264 video as well as to operators of free television and subscription channels. The license terms are updated in 5-year blocks.
Since the first version of the standard was completed in May 2003 and the most commonly used profile was completed in June 2004, a substantial number of the patents that originally applied to the standard have been expiring, although one of the US patents in the MPEG LA H.264 pool lasts at least until 2027.
In 2005, Qualcomm sued Broadcom in US District Court, alleging that Broadcom infringed on two of its patents by making products that were compliant with the H.264 video compression standard. In 2007, the District Court found that the patents were unenforceable because Qualcomm had failed to disclose them to the JVT prior to the release of the H.264 standard in May 2003. In December 2008, the US Court of Appeals for the Federal Circuit affirmed the District Court's order that the patents be unenforceable but remanded to the District Court with instructions to limit the scope of unenforceability to H.264 compliant products.