Fact-checked by Grok 2 weeks ago

Video codec

A video codec is a technology that compresses and decompresses , enabling efficient and transmission by reducing size while maintaining acceptable . It functions as both an encoder, which reduces the bitrate of raw video by exploiting redundancies such as spatial and temporal correlations, and a , which reconstructs the video for playback. This process is essential for applications ranging from streaming services to , as requires immense — for instance, raw video at 60 typically demands 6–12 Gbps. The development of video codecs traces back to the 1980s, with the 's H.120 standard (revised in 1988) marking an early digital video coding effort that introduced basic motion-compensated inter-frame coding in its second version for videoconferencing. The breakthrough came in 1990 with , the first commercially successful standard, designed for audiovisual services at p × 64 kbit/s and establishing the hybrid block-based coding architecture still used today. Subsequent advancements included in 1996 for low-bitrate mobile and conferencing applications, and collaborative efforts between 's (VCEG) and ISO's (MPEG), such as in 1995, which enabled digital television and DVD formats. By the early 2000s, H.264/AVC (Advanced Video Coding), standardized in 2003 by the Joint Video Team (JVT), became the dominant codec, offering roughly double the compression efficiency of prior standards and powering over 80% of internet video traffic due to its balance of quality and computational demands. This was followed by H.265/HEVC (High Efficiency Video Coding) in 2013, which achieved about twice the compression of H.264 for ultra-high-definition content like 4K, supporting emerging formats such as HDR and wide color gamuts. More recently, the landscape has diversified with royalty-free alternatives: AV1, finalized in 2018 by the Alliance for Open Media, provides an 18-30% bitrate reduction over HEVC and is optimized for web streaming by platforms like YouTube and Netflix. By 2025, AV1 adoption has grown significantly in streaming, with ongoing research into AI-enhanced coding for further efficiency. Meanwhile, ITU-T's H.266/VVC (Versatile Video Coding), completed in 2020, delivers up to 50% better compression than HEVC, targeting immersive applications including 8K, VR, and 360-degree video. Video codecs continue to evolve to address surging demands, with video comprising about 82% of global internet traffic as of 2024, driven by mobile devices and high-resolution streaming. Ongoing work by JVET focuses on extensions for screen content and enhanced tools for low-latency encoding, while emerging standards like EVC (Essential Video Coding) offer baseline royalty-free profiles for broadcast and IP delivery. These advancements balance compression efficiency, licensing models, and hardware compatibility to support diverse ecosystems from consumer electronics to professional production.

Core Concepts

Definition and Purpose

A video codec is software or that implements algorithms for compressing and decompressing data, specifically targeting the moving picture component of content rather than audio signals handled by separate audio codecs. The primary purpose of a video codec is to reduce the massive volume of raw data—typically hundreds of megabits per second (e.g., 270 Mbps for )—into a compact suitable for efficient storage on , transmission over , and playback on devices, all while preserving acceptable visual to enable applications such as streaming, , and portable consumption. At its core, a video codec consists of an encoder, which transforms uncompressed raw video frames into a serialized by applying techniques like and , and a , which reverses this process to reconstruct approximate original frames from the bitstream for . Fundamental terminology includes , which compresses individual frames independently (analogous to still-image methods like ), and inter-frame coding, which exploits temporal redundancy by predicting changes between consecutive frames using to achieve higher efficiency. Video codecs predominantly employ , where quantization discards less perceptible data to shrink file sizes significantly, though lossless variants exist that retain all original information at the cost of lower compression ratios.

Compression Principles

Video compression relies on exploiting redundancies inherent in visual data to reduce bitrate while maintaining acceptable . These redundancies include spatial correlations within individual , temporal similarities across consecutive , and statistical patterns in pixel values that can be efficiently encoded. The process typically involves for spatial , predictive coding for temporal , quantization to control data loss, and to further compact the . These principles form the foundation of modern video codecs, enabling significant data reduction for storage and transmission. Spatial compression addresses intra-frame redundancy by transforming pixel data into a where energy is concentrated in fewer coefficients, allowing selective discarding of less perceptible high-frequency components. A key technique is the (DCT), applied to 8x8 blocks, which converts spatial information into coefficients representing average () and varying () frequencies. The 2D DCT for an 8x8 block is given by: F(u,v) = \frac{1}{4} C(u) C(v) \sum_{x=0}^{7} \sum_{y=0}^{7} f(x,y) \cos\left[\frac{(2x+1)u\pi}{16}\right] \cos\left[\frac{(2y+1)v\pi}{16}\right] where f(x,y) is the pixel value at position (x,y), u and v are frequency indices from 0 to 7, and C(k) = \frac{1}{\sqrt{2}} for k=0 and 1 otherwise. This transform concentrates most energy in low-frequency coefficients, which are then quantized to remove insignificant details, reducing data volume while introducing minimal visible distortion. Temporal compression exploits inter-frame redundancy by predicting current frames from previously encoded ones, primarily through . This involves dividing frames into blocks (typically 16x16 or smaller) and estimating motion vectors via block-matching algorithms, which search for the best-matching block in a reference frame to minimize prediction error. Common block-matching methods, such as full search or three-step search, compute the (SAD) or (MSE) between candidate blocks and the current block to find optimal displacement vectors. The residual error between the predicted and actual block is then encoded, significantly lowering the bitrate for sequences with smooth motion. Entropy coding further compresses the quantized transform coefficients and motion data by assigning shorter binary codes to more frequent based on their , approaching the theoretical limit. uses a tree constructed from symbol frequencies, where rarer symbols receive longer codes, while achieves higher efficiency by encoding entire sequences into a single fractional number within a [0,1) interval, dynamically updating probability intervals for each symbol. For instance, in a simple binary model, if a symbol has probability p, its code length approximates -\log_2 p bits, enabling lossless compaction of the residual data without additional distortion. Rate-distortion optimization guides the compression process by balancing the trade-off between bitrate (R) and reconstruction distortion (D), aiming to minimize D for a given R or vice versa. This is conceptualized through the rate-distortion curve, which plots achievable distortion levels against corresponding bitrates for a source, with the curve's shape determined by the source entropy and distortion measure (e.g., MSE). In video coding, decisions like quantization step size are selected to operate near the curve's convex hull, ensuring efficient resource use without exhaustive computation of the full curve. Psycho-visual models incorporate human visual system (HVS) characteristics to enhance compression efficiency by prioritizing perceptually important information. The HVS exhibits lower to high spatial frequencies, color differences (), and subtle changes in uniform areas, allowing codecs to allocate fewer bits to these elements—such as by a factor of 2 in format—while preserving details. This masking of imperceptible details reduces artifacts and improves subjective quality at low bitrates. Quantitative psycho-visual measures, derived from HVS models, further refine quantization by weighting errors based on and . In video compression, lossy methods dominate due to the high data volumes of uncompressed , introducing irreversible distortions through quantization to achieve practical bitrates (e.g., 0.35 bits per for HDTV). Common artifacts include blocking from coarse quantization of adjacent blocks and blurring from over-suppression of high frequencies, which become noticeable at low bitrates but can be mitigated via deblocking filters. , relying solely on without quantization, preserves all data but yields only modest ratios (typically 2:1 for video), insufficient for most applications, highlighting the necessary between and in lossy schemes.

Historical Development

Early Analog and Digital Pioneers

The origins of video compression trace back to the analog era of the 1950s and 1970s, when television broadcasting standards incorporated modulation techniques to transmit video signals efficiently over constrained channel bandwidths. The NTSC (National Television System Committee) standard, adopted in 1953 for color television in North America, encoded chrominance signals using quadrature amplitude modulation on a 3.58 MHz subcarrier, allowing color information to share the 6 MHz broadcast bandwidth with luminance without requiring additional spectrum. This approach effectively compressed color data by interleaving it with the monochrome signal, ensuring backward compatibility with existing black-and-white receivers while minimizing bandwidth expansion. Similarly, the PAL (Phase Alternating Line) standard, developed in the late 1950s and first implemented in 1967 across much of Europe and Asia, alternated the phase of the chrominance subcarrier per line to reduce color distortion, operating within a 7-8 MHz bandwidth for 625-line broadcasts and representing an evolution in analog signal efficiency for international TV transmission. These standards addressed early challenges in analog video by optimizing signal representation, though they relied on inherent modulation rather than explicit digital processing. Research during this period laid groundwork for more sophisticated analog compression methods. In 1952, engineers at developed Differential Pulse-Code Modulation (DPCM), an early predictive technique that estimated pixel values from prior samples to reduce redundancy in video signals, marking one of the first systematic approaches to reduction in analog-to-digital conversion experiments. By the 1960s, advanced practical video transmission with the PicturePhone, publicly demonstrated at the , which captured and sent black-and-white video at 250-line resolution and 30 frames per second over dedicated twisted-pair lines. However, the system's uncompressed analog video required about 1 MHz of —over 300 times that of voice —prompting rudimentary via scan conversion and signal filtering to partially mitigate limitations, though commercial deployment remained limited due to these constraints. In the early , Sony's format, introduced in 1982 as a half-inch professional videotape system, further exemplified analog-era efficiencies by separating (Y) and (C) into component signals, enabling higher sampling rates and reduced compared to composite formats like , thus achieving implicit through improved signal integrity and storage density on ferric-oxide tape. The shift toward digital compression gained momentum in the 1980s amid the rollout of (ISDN), which offered digital channels at multiples of 64 kbit/s but imposed strict bandwidth limits for video applications, typically 64-384 kbit/s for feasible transmission. The 's H.261 standard, initiated in 1984 and approved in 1990 following intensive 1988 development, became the inaugural codec, tailored for videophones and videoconferencing over ISDN lines at p × 64 kbit/s bitrates to overcome these limitations. It pioneered hybrid coding with (DCT) for spatial compression within frames and block-based for temporal prediction across frames, enabling acceptable quality at low rates like 128 kbit/s for quarter-CIF resolution. Key events included ' foundational videotelephony research from the , which informed H.261's focus on real-time, low-latency encoding. Paralleling this, the (MPEG) was formed in January 1988 under ISO by Leonardo Chiariglione and Hiroshi Yasuda, with initial objectives to standardize coded representations of moving pictures at around 1.5 Mbit/s for digital storage media, bridging telephony and needs.

Standardization and Digital Evolution

The standardization of video codecs has been driven by collaborative efforts among international bodies, primarily the and the Joint Technical Committee 1 (JTC1), with the under ISO/IEC playing a pivotal role in multimedia standards. These organizations have jointly developed many codecs to ensure across global applications, from storage media to broadcasting. Alliances like the have further specified codec usage in consumer formats, mandating standards such as H.264/AVC for high-definition optical discs to promote widespread adoption. The early marked the transition to standards, beginning with in 1992, formalized as ISO/IEC 11172, which targeted compression of VHS-quality video and audio for digital storage media at bitrates up to about 1.5 Mbit/s, enabling applications like Video CDs (VCDs) on CD-ROMs. This was followed by in 1994, defined in ISO/IEC 13818 and H.262, which extended support to standard-definition (SD) and high-definition (HD) broadcasting and , providing scalable profiles for professional and consumer use in cable, satellite, and terrestrial transmission. Building on these, the introduced in 1996 as an enhancement to the earlier for videoconferencing, optimizing low-bitrate communication (below 64 kbit/s) over PSTN and early connections through improved and optional negotiable modes. The 2000s saw a surge in efficiency and adoption, led by in 2003, jointly standardized as H.264 and ISO/IEC 14496-10 (MPEG-4 Part 10), which achieved roughly double the compression of at similar quality levels, dominating Blu-ray Disc playback and internet streaming with profiles like Main for broadcast and High for HD content. This paved the way for higher resolutions, with (HEVC/H.265) approved in 2013 by H.265 and ISO/IEC 23008-2 ( Part 2), offering up to 50% bitrate reduction over H.264 to support 4K ultra-high-definition (UHD) video in streaming and broadcasting. Further advancing this trajectory, Versatile Video Coding (VVC/H.266), finalized in 2020 as H.266 and ISO/IEC 23090-3 (MPEG-I Part 3), targets 8K and immersive formats with 30-50% efficiency gains over HEVC, accommodating higher frame rates and wider color gamuts. In parallel, open-source initiatives emerged to counter proprietary licensing, with releasing in 2010 as a royalty-free successor to earlier formats, integrated into the container for web video, followed by in 2013, which improved compression by 30-50% for HD streaming on platforms like . The (AOMedia) then unveiled in 2018, a royalty-free codec developed collaboratively by industry leaders including , , and , achieving comparable efficiency to HEVC without licensing fees to foster broader deployment. These efforts reflect a dual path of licensed, ITU/ISO-led standards for regulated industries and open alternatives for web-scale applications, continually evolving to meet demands for higher resolutions and constraints.

Technical Design

Encoding Process

The encoding process in video codecs transforms raw video data into a compressed by exploiting spatial and temporal redundancies through a series of sequential operations. This workflow typically begins with pre-processing the input , followed by prediction to generate residuals, transformation and quantization of those residuals, , and finally bitstream assembly, all modulated by rate control mechanisms to meet target bitrates and compatibility constraints. Intra-frame operations focus on spatial prediction within a single frame to reduce redundancy, while inter-frame operations use and compensation across frames for temporal efficiency, as detailed in standards like H.264/AVC. Pre-processing prepares the raw video for compression by converting the and applying . , often in RGB format, is converted to , where (Y) carries most detail and (Cb, Cr) components can be subsampled (e.g., ) to reduce data volume without significant perceptual loss, as specified in ITU-R BT.601. filters, such as temporal or spatial smoothing, are applied to mitigate artifacts like or sensor noise, enhancing compression efficiency by minimizing high-frequency components that consume bitrate without adding value. Prediction forms the core of intra- and inter-frame operations, generating a by subtracting a predicted from the original. For , spatial prediction uses neighboring pixels within the same to estimate values, employing directional modes (e.g., , vertical) to capture local correlations. Inter-frame coding, conversely, relies on to identify temporal similarities: the current is divided into s (typically 16x16 macroblocks or smaller partitions), and for each , a matching in a reference is searched within a defined . The full exhaustively evaluates all candidate positions in the search using a distortion metric like (SAD), yielding motion vectors that describe displacement for compensation. The residual from prediction undergoes transformation to concentrate energy into fewer coefficients, followed by quantization to discard less perceptible details. Block-based discrete cosine transform (DCT) is commonly applied, converting spatial residuals into frequency-domain coefficients; alternatively, wavelet transforms offer multi-resolution analysis in some codecs for better handling of varying content. Quantization then scales these coefficients using a step size (Q_step) determined by a quantization parameter (QP), with the formula: \text{Quantized coefficient} = \round\left( \frac{\text{DCT coefficient}}{Q_\text{step}} \right) This scalar process reduces precision, controlling bitrate at the cost of minor quality loss, where higher QP values yield coarser quantization. Entropy coding compresses the quantized coefficients and motion vectors into a compact representation using variable-length or arithmetic codes, such as context-adaptive binary arithmetic coding (CABAC) in advanced codecs. The bitstream is then formed by inserting headers—sequence parameter sets (SPS), picture parameter sets (PPS), and slice headers—that define frame structure, prediction modes, and metadata. Video is organized into pictures (I for intra-only, P for predicted, B for bi-directional), grouped into slices for error resilience and parallel processing. Rate control ensures the output adheres to constraints like limits, employing bitrate (CBR) for steady data flow in or (VBR) for quality optimization in storage, where complex scenes allocate more bits. Buffer management, via models like the video buffering verifier () in MPEG standards or hypothetical reference decoder (HRD) in H.264, prevents overflow/underflow by regulating quantization and skipping during encoding. Profiles and levels impose constraints on the encoding process to ensure across devices. Profiles define supported features (e.g., for low-complexity, High for advanced tools like 8x8 transforms), while levels cap parameters like , , and bitrate (e.g., Level 3.1 supports up to at 14 Mbps), tailoring the for specific applications without altering core steps.

Decoding Process

The decoding process in video codecs reverses the applied during encoding, transforming a compressed into a sequence of reconstructed video frames suitable for display or further processing. This involves several interdependent steps that ensure fidelity to the original video while managing computational efficiency and robustness to errors. Representative examples from standards like H.264/AVC illustrate these operations, where the operates on (NAL) units containing video coding layer (VCL) data and supplemental enhancement information. Bitstream parsing begins with entropy decoding to extract structural elements from the compressed data. Using methods such as context-adaptive variable-length coding (CAVLC) or (CABAC), the decoder interprets the to retrieve headers, motion vectors, and quantized transform coefficients. (SPS) and (PPS) provide global and frame-specific parameters, such as , level, and , while slice headers define boundaries for independent processing units. Motion vectors, encoded with variable precision (e.g., quarter-pixel in H.264), and quantized coefficients, scanned in zigzag order, are decoded to prepare for reconstruction. This parsing ensures the bitstream's syntax is correctly interpreted without loss of essential data. Inverse quantization and transformation reconstruct the residual signal from the parsed coefficients. Quantized coefficients, scaled during encoding to reduce bitrate, undergo inverse quantization by multiplying each coefficient by a quantization step size Q_{\text{step}}, which depends on the quantization parameter (QP). The dequantized coefficients are then transformed back to the spatial domain using an inverse (IDCT) or equivalent integer approximation. For a 4x4 block in H.264/AVC, this yields the block via: \text{Residual block} = \text{IDCT}(\text{Quantized coefficients} \times Q_{\text{step}}) This step approximates the original residual, with the integer transform matrix ensuring exact reversibility in the to avoid drift. generates the predicted portion of the frame by applying decoded motion vectors to reference frames stored in a decoded picture (DPB). For inter-predicted blocks, the shifts and pixels from previously decoded frames, supporting variable block sizes (e.g., 4x4 to 16x16 in H.264) and multiple reference frames for improved accuracy. Sub-pixel interpolation, often using a 6-tap , refines predictions at quarter-sample precision, such as b = (-E + 5F - 10G + 20H + 20I - 5J + K)/32 for half-sample positions. The reconstructed block is then formed by adding the motion-compensated prediction to the decoded residual. Post-processing enhances the reconstructed frames to mitigate compression artifacts. In-loop deblocking filters, applied adaptively across boundaries, reduce visible discontinuities by averaging pixels based on QP-dependent thresholds (e.g., boundary strength and clipping values \alpha and \beta). For instance, in H.264/AVC, the filter processes luma and chroma edges separately, improving visual quality by 5-10% in terms of (PSNR). Additional deringing techniques, such as smaller transform sizes (e.g., 4x4 instead of 8x8), suppress high-frequency oscillations around edges. These operations occur within the decoding loop to influence future predictions. Error resilience mechanisms handle bitstream corruptions, particularly in error-prone environments like streaming over networks. Techniques such as slice-level allow the to isolate and conceal errors within affected slices, replacing lost s with spatial or temporal interpolations from neighboring . Flexible ordering (FMO) and redundant slices provide alternative paths for recovery, while partitioning separates headers, motion, and texture for graceful degradation. These features ensure partial usability of the video even under 1-5% . Synchronization maintains temporal alignment during playback by processing timestamps embedded in the bitstream. The hypothetical reference decoder (HRD) model in standards like H.264 uses coded picture buffer (CPB) removal times and decoded picture buffer (DPB) management to regulate frame rates and buffer delays, preventing overflows or underflows. Access unit delimiters and picture order counts (POC) further ensure frames are output in the correct sequence, supporting variable frame rates up to 75 Hz.

Algorithms and Standards

Video codecs rely on sophisticated prediction algorithms to minimize redundancy in video data. Intra-prediction exploits spatial correlations within a single frame by estimating pixel values based on neighboring blocks, with H.264/AVC defining nine directional modes for 4x4 luma blocks, including vertical, horizontal, and diagonal predictions, plus a DC mode using the average of adjacent pixels. Inter-prediction, conversely, leverages temporal correlations across frames through ; H.264/AVC supports multiple reference frames, up to 16 in P- and B-slices, allowing selection of the most suitable prior frame for block matching to enhance prediction accuracy and compression efficiency. These mechanisms reduce the residual that requires further encoding, forming the core of video compression frameworks. Transform coding further compacts the prediction residuals by converting them into the . While earlier standards like H.264/AVC employ an integer approximation of the (DCT) for 4x4 and 8x8 blocks to approximate energy compaction, HEVC advances this with larger integer transforms up to 32x32, using separable core transforms based on DCT-like kernels that maintain invertibility without floating-point operations, thereby improving coding efficiency for high-resolution content. More recent standards like (H.266) extend this with transforms up to 64x64 and enhanced separability for better efficiency in 8K and immersive video. To mitigate artifacts from block-based processing, modern codecs incorporate in-loop filters applied post-reconstruction. The adaptive in H.264/AVC and HEVC analyzes boundaries between blocks to adjust pixel values based on quantization parameters and edge strength, reducing visible blocking without excessive blurring. HEVC extends this with sample adaptive offset (SAO), which applies either edge offset or band offset to residual samples, compensating for quantization distortions and yielding up to 5% bit-rate savings in subjective quality tests. Standardization ensures interoperability across devices and applications, with bodies like and ISO/IEC defining profiles and levels. Profiles, such as H.264/AVC's Baseline profile optimized for low-latency applications like video conferencing by omitting B-frames and CABAC , tailor features to use cases, while levels impose constraints on , , and —e.g., Level 4.1 caps at @30fps with 20 Mbps—to guarantee decoder capabilities. , specified in the standards, verifies implementation fidelity through test sequences and compliance. Codec performance is evaluated via complexity metrics and compression benchmarks. Encoding and decoding complexity increases with newer standards like HEVC compared to H.264/AVC, often necessitating for high-resolution formats like . Compression ratios highlight efficiency gains; HEVC achieves roughly 50% better bit-rate reduction than H.264/AVC at equivalent quality, as demonstrated in joint collaborative team tests where HEVC encoded UHD sequences at half the while preserving PSNR. Patent licensing models influence codec adoption. As of 2025, many essential patents for H.264/AVC have expired in major jurisdictions (e.g., in January 2025, in 2024), reducing royalty obligations, while remaining patents in some regions are managed by Via Licensing Alliance with structured fees and caps. HEVC continues to rely on patent pools like Via Licensing Alliance, aggregating essential patents and charging royalties to facilitate use. In contrast, from the is , with members committing to license patents on fair, reasonable, and non-discriminatory terms without monetary compensation, promoting open-source implementations and reducing barriers for web and streaming applications.

Applications and Use Cases

Media Production and Editing

In professional media production workflows, video codecs are integral from the capture stage, where onboard camera encoding prioritizes high-fidelity preservation for subsequent . Codecs such as and enable raw-like quality in and 8K captures by providing intra-frame with minimal , supporting playback and multistream directly from camera files. ProRes, for instance, is in cameras like the for log-encoded footage, maintaining 12-bit depth per channel to retain and color detail during initial recording. DNxHD similarly facilitates efficient onboard encoding in production cameras, with bitrates up to 440 Mbit/s in variants like DNxHD 444, ensuring compatibility with high-resolution sensors without introducing visible artifacts. Editing software integration relies on intermediate codecs—lightly compressed or visually lossless formats—to facilitate non-destructive manipulation during cuts, transitions, and effects application. These codecs, including ProRes 422 HQ and DNxHR, are transcoded from camera originals early in the pipeline to avoid from repeated encodes, as their frame-independent structure prevents error propagation across timelines. In applications like or , ProRes supports up to 33 simultaneous streams for real-time editing, while DNxHR handles 8K workflows with reduced decoding complexity, preserving spatial and temporal integrity for iterative revisions. This approach ensures that color corrections and VFX composites remain faithful to the source. Specific workflow steps often culminate in from uncompressed or intermediate edit masters to formats like H.264 for review proxies or interim sharing. Productions typically maintain masters in ProRes or DNxHD at high bitrates (e.g., 220-500 Mbit/s for /) before converting to H.264 at 10-20 Mbit/s for client , using integrated tools in software to automate the process without altering the primary assets. This preserves the master’s quality for final output while enabling efficient collaboration, as H.264’s long-GOP efficiency suits bandwidth-constrained reviews without compromising the production chain. Quality preservation hinges on high-bit-depth support in codecs, where 10-bit and 12-bit processing is standard to maintain gradient smoothness in color grading and HDR workflows. SMPTE recommendations specify at least 10-bit depth for wide color gamut (WCG) content in production paths, supporting 4:2:2 or 4:4:4 chroma subsampling to minimize banding in shadows and highlights during grading sessions. 12-bit variants, as in ProRes 4444 or DNxHR 444, offer further precision for noise-free CGI integration and animation, with 12-bit mastering reducing quantization errors in file formats like MXF. These depths ensure perceptual uniformity in tools like DaVinci Resolve, where lower bitrates could otherwise introduce visible artifacts in post. In and TV production, industry standards dictate codec applications for standardized interoperability. The Interoperable Master Format (IMF), per SMPTE ST 2067, employs for image essence in , supporting 8-12 bit depths and resolutions up to UHD with progressive or interlaced scanning for archival masters. This format ensures license-free, high-quality packaging for global distribution, with codestream constraints aligned to ISO/IEC 15444-1 for reversible or . For TV, EBU guidelines endorse intermediate codecs like DNxHD (120-185 Mbit/s, 10-bit) and AVC-I in HDTV workflows, achieving quasi-transparent quality across 4-5 generations in , as validated in multi-pass tests exceeding 100 Mbit/s thresholds.

Distribution and Streaming

Video codecs are integral to the distribution and streaming of video content, enabling efficient transmission over networks by compressing data to minimize usage while preserving quality. In , protocols such as (HLS) and () segment video into short clips encoded with codecs like H.264 (, AVC) and HEVC (, H.265), allowing clients to switch between quality levels in real-time based on available . HLS, developed by Apple, mandates encoding in H.264/AVC or HEVC/H.265 for compatibility across devices, supporting segmented transport streams or fragmented MP4 containers to facilitate seamless playback transitions. Similarly, , standardized by MPEG, is codec-agnostic but commonly employs H.264 and HEVC for its media presentation description files, which reference multiple bitrate variants to adapt to fluctuating network speeds, ensuring reduced rebuffering events in applications like online video platforms. As of 2025, AV1 adoption has expanded, with platforms like and using it for a significant portion of streams, achieving 20-30% bitrate savings over HEVC. Broadcast standards further highlight codec efficiency in fixed-bandwidth environments. The (NextGen TV) standard specifies HEVC as a primary video codec for ultra-high-definition (UHD) broadcasts and has approved (H.266) as an additional option as of July 2025, constraining profiles and levels to support at up to 120 frames per second while enabling higher compression ratios than predecessor H.264-based systems. This allows broadcasters to deliver immersive content over terrestrial signals with improved . In and , HEVC is widely adopted under (Digital Video Broadcasting) guidelines, compressing high-definition and UHD channels to fit within constrained capacities, thereby supporting more simultaneous streams without quality degradation. Content delivery networks (CDNs) and optimize codec selection for low-latency delivery, where processing video closer to users minimizes transport delays. (AOMedia Video 1), an open-source offering up to 30% better than HEVC, is increasingly used in such setups; , for example, rolled out support for in 2023, enabling delivery at lower bitrates to reduce buffering on variable connections via its global CDN infrastructure. This choice enhances edge caching efficiency, as smaller file sizes accelerate content propagation and playback initiation. Assessing streamed video quality during distribution relies on objective metrics like Peak Signal-to-Noise Ratio (PSNR), which measures pixel-level distortion in decibels, and Structural Similarity Index (SSIM), which evaluates perceived structural, luminance, and contrast fidelity on a scale from 0 to 1. These metrics guide codec tuning in pipelines, with PSNR above 30 dB and SSIM exceeding 0.9 typically indicating acceptable quality for streaming, helping providers benchmark compression against network-induced artifacts. In live scenarios, they inform real-time adjustments to maintain viewer satisfaction. A key challenge in video distribution is accommodating variable conditions, such as fluctuations or peak-hour , which can cause stalls or quality drops. Netflix addresses this through per-title encoding, analyzing each video's complexity to generate custom bitrate ladders—often using optimization on PSNR curves—resulting in up to 20% bitrate savings or equivalent quality at lower rates compared to uniform encoding. This approach ensures robust delivery across diverse conditions, from low-bandwidth environments to high-speed links, without over-provisioning resources. Emerging codecs like (H.266) are being tested for distribution, offering up to 50% better compression than HEVC for 8K and immersive streaming applications as of 2025.

Consumer Devices and Hardware

Video codecs play a crucial role in mobile devices, where on-device encoding and decoding must balance quality, speed, and power consumption. Smartphones commonly rely on H.264 (AVC) for video processing in applications like , as it benefits from widespread that minimizes battery drain during capture and playback. This codec's efficiency stems from optimized decoding pipelines in mobile SoCs, allowing apps to handle editing and sharing without excessive energy use, particularly on resource-constrained devices. For instance, devices often HEVC content to H.264 to avoid high computational costs, preserving battery life for extended sessions. In home entertainment systems, codecs enable high-resolution playback on dedicated hardware. Blu-ray players support HEVC (H.265) for Ultra HD discs, delivering content with enhanced compression that maintains visual fidelity while fitting within disc capacity limits. Streaming boxes like integrate HEVC for streaming, recommending it for UHD encodings up to level 5.1 and bitrates of 25 Mbps, ensuring smooth playback on compatible models without straining processing resources. These devices also handle H.264 for broader compatibility, allowing seamless integration with existing libraries of HD content. With (H.266) hardware integration emerging in 2025, future devices will support even higher efficiencies for 8K and content. Gaming consoles leverage specialized codecs for low-latency applications, particularly in scenarios. NVIDIA's NVENC hardware encoder, integrated into GPUs used in services like , supports H.264 and HEVC for real-time encoding, offloading the CPU to maintain high frame rates during streaming. This setup enables consoles to deliver immersive experiences over networks, with NVENC's dedicated cores ensuring minimal performance impact for interactive gameplay. Compatibility challenges arise with legacy devices, where newer codecs like HEVC or may not be supported, necessitating fallbacks to older standards such as . , a , serves as a reliable option for web-based video on outdated hardware, with broad browser support including and , though limits it to WebRTC contexts. Developers often provide multiple sources—such as / alongside MP4/H.264—to ensure playback on these systems without overhead. The efficiency of modern codecs directly influences by optimizing storage and download times on consumer devices. For 4K videos on smartphones, HEVC reduces file sizes by approximately 40-50% compared to H.264 at equivalent quality, allowing more content to fit on limited internal storage—such as a 1-minute clip dropping from 400 MB to under 200 MB. This also shortens download durations over mobile networks, cutting data usage and buffering waits, which is critical for users streaming or sharing high-resolution media.

Notable Codecs

Legacy and Widely Adopted Codecs

, standardized as ISO/IEC 13818-2 and H.262, became a cornerstone of in the 1990s, widely adopted for discs and terrestrial broadcast television. It supports resolutions up to high-definition () formats like , with maximum bitrates of up to 19.4 Mbps in the ATSC A/53 standard for U.S. over-the-air HD transmission. This codec's block-based and (DCT) techniques enabled efficient compression for standard-definition (SD) content, but it proved inefficient for HD due to higher required bitrates—often 15-19 Mbps for acceptable quality—compared to successors, leading to greater demands in broadcast and storage. Despite these limitations, MPEG-2's ubiquity in legacy infrastructure ensures its continued use in some and systems. H.264/AVC, defined in ITU-T H.264 and ISO/IEC 14496-10, emerged in 2003 as a major advancement, achieving widespread adoption with over 90% market share in online video by the mid-2010s due to its superior compression efficiency. It offers multiple profiles tailored to applications, including the High 4:2:2 Profile, which supports 10-bit per channel 4:2:2 for professional workflows like broadcast contribution and , enabling better color fidelity for editing without full 4:4:4 overhead. Licensing is managed through the Via Licensing Alliance (formerly MPEG LA), which administers a covering essential patents from multiple contributors, with royalties applied to encoders, decoders, and content distribution exceeding certain thresholds. H.264/AVC typically provides 50% bitrate savings over for equivalent subjective quality, as verified in NTIA subjective tests for HDTV, making it ideal for Blu-ray discs, streaming, and mobile video. VP8, originally developed by On2 Technologies as a in 2008, was acquired and open-sourced by in 2010 under a BSD-like license to promote web video. It is primarily used within the , which combines VP8 video with or audio, facilitating efficient multiplexing for online delivery. VP8 gained traction in video adoption, with native support in browsers like , , and by 2011, enabling to serve VP8-encoded content without plugins and supporting the alternative to H.264 in web standards. DivX and Xvid, both implementations of the MPEG-4 Part 2 Advanced Simple Profile (ASP), rose to prominence in the early for compressing full-length movies onto CDs or for early distribution. DivX, initially a hacked version of Microsoft's MPEG-4 codec released in , evolved into a commercial product by DivX, Inc., while Xvid emerged as its open-source reverse-engineered counterpart in 2001, offering near-identical performance with greater customization. These codecs became staples in file-sharing networks during the and early era, allowing users to encode and share high-quality video at bitrates around 700-1500 kbps for content, far more efficient than prior formats like , though limited by block artifacts at low bitrates compared to later standards.

Modern and Emerging Codecs

, developed by and released in 2013 as the successor to , is a video codec offering approximately 50% better compression efficiency than H.264/AVC for similar quality levels. It incorporates advanced features like larger block sizes, improved , and support for 12-bit and , making it suitable for and 8K resolutions. has been widely adopted for web streaming, particularly by , which uses it for the majority of its and content as of 2025, and is supported natively in major browsers and devices, contributing to the shift toward open-source codecs in online video delivery. High Efficiency Video Coding (HEVC), also known as H.265, represents a significant advancement in video , achieving approximately 50% bitrate reduction compared to its predecessor H.264/AVC while maintaining equivalent video . This efficiency stems from enhanced modes, larger tree units, and improved intra- and inter- techniques, enabling support for resolutions up to 8K. HEVC has seen widespread in 4K UHD Blu-ray discs, where it facilitates high- playback at average bitrates around 80 Mbit/s. However, its deployment has been hampered by complex patent licensing structures involving multiple pools and licensors, leading to fragmented royalty agreements and higher implementation costs. AOMedia Video 1 (), developed by the (AOMedia), emerged as a alternative to HEVC, offering around 30% better compression efficiency for the same quality level. Backed by major industry players including , , and , AV1 leverages advanced tools like extended partitioning and transform skips to optimize encoding for internet streaming. Netflix began rolling out AV1 encoding in the early 2020s, which has boosted viewing hours by 5% and reduced quality switches by 38%. Its open-source nature has accelerated hardware integration in devices like modern smartphones and smart TVs, positioning AV1 as a dominant choice for web-based video delivery. Versatile Video Coding (VVC), standardized as H.266 by the ITU in 2020, builds on HEVC to deliver about 50% bitrate savings at equivalent perceptual quality, particularly for high-resolution content. introduces flexible partitioning, affine motion models, and enhanced filtering to handle demanding applications like 8K video and 360-degree immersive formats, reducing bandwidth needs for ultra-high-definition streaming. Developed jointly by the Joint Video Exploration Team (JVET), it supports a broader range of bit depths and color formats, making it suitable for future broadcast and environments, though its higher computational complexity—up to 10 times that of HEVC—poses encoding challenges. Among emerging standards, (EVC), part of MPEG-5 and finalized in 2020 as ISO/IEC 23094-1, offers a baseline royalty-free profile alongside an enhanced profile with optional patented tools, achieving up to 30% bitrate reduction over H.264 in basic configurations. Supported by companies like , , and , EVC emphasizes straightforward licensing with a limited set of essential patents, facilitating easier adoption in resource-constrained devices without sacrificing core efficiency gains. Low Complexity Enhancement Video Coding (), standardized as MPEG-5 Part 2 in , functions as an enhancement layer atop existing codecs like H.264 or HEVC, improving compression by 20-50% through low-overhead upscaling and detail restoration without requiring full recoding of legacy streams. This approach allows incremental upgrades to older infrastructure, preserving compatibility while boosting quality for mobile and low-bandwidth scenarios. AI-based innovations are pushing codec boundaries, with prototypes like Netflix's neural network-driven —introduced in 2023—using VMAF-guided optimization to preserve perceptual during resolution reduction, achieving bandwidth savings comparable to traditional methods but with scene-adaptive precision. These end-to-end neural codecs employ for tasks such as residual prediction, outperforming conventional hybrids in subjective metrics. Looking ahead, integration, particularly neural motion estimation, promises further gains by replacing block-based searches with learned models, reducing artifacts in dynamic scenes. Sustainability trends emphasize power-efficient designs, with AI-assisted adaptive streaming frameworks targeting reduced energy consumption in encoding and transmission, aligning video tech with environmental goals.

Implementation Aspects

Software and Open-Source Tools

Software implementations of video codecs enable flexible encoding and decoding through libraries and tools that operate independently of specialized , facilitating into diverse applications and workflows. These open-source solutions emphasize , extensibility, and , allowing developers to customize builds for specific needs such as processing or high-quality archiving. FFmpeg stands as a cornerstone open-source framework, featuring a command-line tool that supports a wide array of codecs—over 100 in total—for tasks including media files and streaming content across networks. At its core lies libavcodec, a versatile library that provides a generic framework for encoding and decoding video, audio, and subtitle streams, with a supporting custom compilations to include only required components. This design enables efficient resource usage in embedded systems or large-scale servers, while maintaining compatibility with numerous formats. Prominent examples include and , open-source encoders developed by the project for the H.264/AVC and HEVC/H.265 standards, respectively. x264 delivers high-performance H.264 encoding, capable of processing multiple 1080p streams in on , through tunable presets that encoding speed against , such as "ultrafast" for processing or "veryslow" for optimal efficiency. Similarly, x265 extends these capabilities to HEVC, offering bitrate reductions of 25–50% over H.264 at equivalent via analogous preset options and advanced optimizations like parallel threading. Both encoders integrate seamlessly with FFmpeg, enhancing its utility for professional video workflows. For development, these tools provide robust APIs that support integration into applications like the , which leverages for decoding and playback across platforms including Windows, , and macOS. This cross-platform compatibility ensures consistent behavior in diverse environments, from desktop software to mobile apps. Open-source nature fosters ongoing enhancements through community contributions; for instance, the libaom library, released in 2018 as the reference encoder by the , has driven royalty-free advancements in next-generation compression, with iterative improvements in speed and efficiency via collaborative development.

Hardware Acceleration and Integration

Hardware acceleration for video codecs leverages dedicated to offload computationally intensive encoding and decoding tasks from general-purpose CPUs, enabling processing of high-resolution content such as and 8K video. This approach utilizes application-specific integrated circuits () and programmable GPUs to perform operations like , , and more efficiently than software implementations. By integrating these accelerators directly into processors or as co-processors, systems achieve lower and higher throughput, which is essential for applications demanding seamless playback and streaming. Dedicated , such as 's Quick Sync Video, provide hardware support for H.264/AVC and HEVC (H.265) encoding and decoding on integrated graphics in processors starting from the 2nd generation series. Quick Sync employs fixed-function pipelines optimized for these codecs, allowing for multiple simultaneous sessions without taxing the host CPU. Similarly, AMD's (VCN) architecture, found in GPUs and APUs, supports H.264/AVC and HEVC encode/decode through dedicated media engines, with VCN generations improving efficiency for up to 8K resolutions. These prioritize power-constrained environments like laptops and desktops by minimizing thermal output during prolonged encoding tasks. GPU-based acceleration extends these capabilities through units, exemplified by NVIDIA's NVENC encoder integrated with cores. NVENC, available on RTX GPUs from the Turing architecture onward, handles H.264, HEVC, and encoding, with the Ada Lovelace generation (RTX 40-series) delivering support at up to 8K60 with enhanced compression efficiency over software methods. The subsequent Blackwell architecture (RTX 50-series, released in 2025) introduces the 9th-generation NVENC with further enhancements, including up to 60% faster encoding speeds. This parallelism allows GPUs to process multiple frames or streams concurrently, making them suitable for professional workflows involving batch encoding. In system-on-chip (SoC) designs for mobile and embedded devices, hardware acceleration is tightly integrated for on-device processing. Qualcomm's Snapdragon platforms, such as the Snapdragon 8 Gen series, incorporate video processing units (VPUs) that support 8K HEVC decoding at 60 FPS, enabling efficient playback on smartphones without excessive battery drain. These SoCs combine ASICs for codec operations with AI-enhanced image signal processors to handle real-time video pipelines in power-sensitive scenarios. The primary benefits of include substantial reductions in CPU utilization—often offloading 90-100% of codec workloads—and improved power efficiency, with specialized VPUs achieving up to 3x better energy use compared to CPU-based encoding for streams. For instance, GPU-accelerated encoding on hardware can process video several times faster than equivalent software solutions while maintaining comparable quality, with speedups of 2-5x typical on high-end GPUs. These gains are particularly impactful in streaming and , where sustained high-bitrate processing is required. Despite these advantages, hardware acceleration faces challenges such as , where proprietary implementations like Quick Sync or NVENC limit interoperability across ecosystems, potentially requiring specific drivers or APIs. Support for emerging codecs like (VVC/H.266) remains limited as of 2025, with most hardware focused on and HEVC; widespread VVC adoption is hindered by the need for new silicon generations and inconsistent device compatibility.

Codec Packs and Container Formats

Codec packs are bundled collections of audio and video codecs, filters, and decoders designed to enhance multimedia playback compatibility on operating systems like Windows, particularly through frameworks such as . The , for instance, provides a modular set of components including LAV Filters and ffdshow, enabling users to play a wide range of formats that may not be natively supported by default media players. Similarly, the (CCCP) focuses on filters tailored for niche content like , incorporating tools such as Haali Media Splitter and VSFilter to handle rare or specialized video streams without requiring extensive configuration. These packs facilitate playback of uncommon formats by installing necessary decoders, but users must select configurations carefully to avoid conflicts with system codecs. Container formats serve as wrappers that encapsulate compressed video, audio, , and other data streams into a single file, allowing for organized storage and playback. The MP4 format, based on the (ISOBMFF) defined in MPEG-4 Part 12, commonly packages H.264 (AVC) or video codecs alongside audio, supporting efficient streaming and broad device compatibility. In contrast, the (MKV) container offers greater flexibility by accommodating multiple video, audio, and subtitle tracks within one file, making it ideal for complex media like multilingual releases or director's cuts. The container, developed by the WebM Project, pairs , , or video with or audio, prioritizing royalty-free web delivery and integration with video elements. These containers play a critical role in demultiplexing interleaved streams during playback, where a demuxer separates video, audio, and subtitle data for independent decoding by respective components. They also ensure synchronization by embedding timestamps that align audio and video presentation, preventing desync issues in content. Additionally, containers support embedding for details like markers, artwork, and encoding parameters, enhancing user and file management. Browser compatibility for modern containers has improved significantly, with providing native support for video in MP4 files as of version 70 and beyond, enabling efficient streaming without plugins by 2025. This adoption extends to other browsers like and , though legacy formats may still require fallbacks for older hardware. Distributing codec packs or containers with patented codecs, such as H.264 in MP4, raises legal challenges due to licensing requirements from pools like , which mandate royalties for encoders and certain distributions to avoid infringement. Open alternatives like the Ogg container, which pairs video with audio under a fully permissive license, address these issues by offering patent-free options for distributions and web embedding.

References

  1. [1]
    [PDF] THE VIDEO CODEC LANDSCAPE IN 2020 - ITU
    The video codec landscape in 2020 includes VVC, EVC, and AV1. VVC is ahead of its competitors, and the landscape is rapidly evolving.
  2. [2]
    video codec - an overview | ScienceDirect Topics
    A video codec is a coder/decoder system that compresses and decompresses video data to maximize data flow through networks or storage spaces.Missing: authoritative | Show results with:authoritative
  3. [3]
    Visual Coding & VCEG - ITU
    Jul 15, 2006 · ITU-T H.261 Video codec for audiovisual services at p x 64 kbit/s was the first commercially-successful digital video coding standard, and introduced the modern ...
  4. [4]
  5. [5]
    [PDF] Overview of International Video Coding Standards (preceding H.264 ...
    Jul 22, 2005 · Two organizations have dominated video compression standardization: • ITU-T Video Coding Experts Group (VCEG).
  6. [6]
  7. [7]
    High-performance compression of visual information-a tutorial ...
    Two different compression categories must be distinguished: lossless and lossy. Lossless compression is achieved if no distortion is introduced in the coded ...
  8. [8]
    Video Compression: Principles, Practice, and Standards
    – Video contains much spatial and temporal redundancy. • Spatial redundancy: Neighboring pixels are similar. • Temporal redundancy: Adjacent frames are similar.
  9. [9]
    [PDF] JPEG Image Compression using the Discrete Cosine Transform - arXiv
    Nov 1, 2019 · The. DCT equation is the summation of the input function and cosine functions over 8X8 block that is being compressed. DCT(u, v) = 1. 4. C(u)C ...
  10. [10]
  11. [11]
    [PDF] Evaluation of Huffman and Arithmetic Algorithms for Multimedia ...
    Huffman and arithmetic coding are the two most important entropy coding in image compression standards. In this paper, we are planning to answer the ...
  12. [12]
    [PDF] A Practical Approach for Rate-Distortion-Perception Analysis ... - arXiv
    Apr 30, 2021 · Compression algorithms are optimized and evaluated based on Shannon's rate-distortion framework, where the goal is to minimize the distortion, ...
  13. [13]
  14. [14]
    [PDF] Rethinking Lossy Compression: The Rate-Distortion-Perception ...
    Lossy compression techniques are ubiquitous in the modern- day digital world, and are regularly used for communicating and storing images, video and audio.
  15. [15]
  16. [16]
    The History of Video Compression Standards, From 1929 Until Now
    Jun 8, 2021 · Video compression may seem very modern, but it has a long history that goes all the way back to 1929, when it was first suggested for analog video!
  17. [17]
    Picturephone - Engineering and Technology History Wiki
    Apr 1, 2019 · Bell Labs resumed video telephone research in 1956. From this work came the April 1964 debut of the Picturephone as a major attraction at ...
  18. [18]
    Televisuality, Gender, and the Short Life of the Picturephone
    Sep 18, 2021 · THE TELEPHONE OF TOMORROW AND THE DESIRES OF YESTERDAY. For Bell Laboratories, the Picturephone ... The video signal was compressed so it ...
  19. [19]
    The History of Sony Betacam SP - Datarecovery.com
    Jun 1, 2022 · Launched in August 1982, the original Betacam was an analog component video format that utilized ferric-oxide tape. The cassettes were ...
  20. [20]
    The true history of MPEG's first steps - Leonardo's Blog
    Dec 28, 2019 · MPEG was established in January 1988, with its first project being video coding at 1.5 Mbit/s. The first meeting was in May 1988, and the first ...
  21. [21]
    MPEG-1 standard
    MPEG-1. ISO/IEC 11172. Coding of moving pictures and associated audio for digital storage media at up to about 1,5 Mbit/s. A suite of standards for audio-video ...Missing: 1992 | Show results with:1992
  22. [22]
    Celebrating 10 years of WebM and WebRTC
    May 27, 2020 · Starting from VP8 in 2010, the WebM Project has delivered up to 50% video bitrate savings with VP9 in 2013 and an additional 30% with AV1 in ...<|separator|>
  23. [23]
    AV1 Features - Alliance for Open Media
    AV1 Features. ROYALTY-FREEPermalink. Interoperable and open. UBIQUITOUSPermalink. Scales to any modern device at any bandwith. FLEXIBLEPermalink.Missing: 2018 | Show results with:2018
  24. [24]
    [PDF] Overview of the H.264/AVC video coding standard - Circuits and ...
    H.264/AVC is the newest international video coding standard, aiming for enhanced compression and network-friendly video for various applications.
  25. [25]
    [PDF] Pre-processing techniques to improve HEVC subjective quality
    Apr 9, 2017 · It is also well known that pre-processing algorithms can be used to improve the video compression efficiency by means of noise reduction and the ...<|control11|><|separator|>
  26. [26]
    Search Algorithms for Block-Matching in Motion Estimation
    These algorithms estimate the amount of motion on a block by block basis, ie for each block in the current frame, a block from the previous frame is found.
  27. [27]
    CBR vs. VBR: Differentiating Constant Bitrate from Variable Bitrate
    Jul 18, 2025 · A CBR file format maintains the same bitrate throughout the entire file, providing a consistent data stream, while VBR full form adjusts the ...
  28. [28]
    [PDF] Overview of the H.264 / AVC Video Coding Standard
    264/AVC design includes a new feature consisting of picture types that allow exact synchronization of the decoding process of some decoders with an ongoing ...
  29. [29]
    [PDF] Overview of the H.264/AVC Video Coding Standard
    • 9 modes. • Used in texture area. – Intra 16x16. • 4 modes. • Used in flat ... 4x4 Intra Prediction Mode. • Mode 5. • Mode 6. Q A B C D E F G H. I a b c d. J ...<|control11|><|separator|>
  30. [30]
    Multiple Reference Frame - an overview | ScienceDirect Topics
    In H.264, multiple frames (up to a maximum of 16) can be used for prediction. The frames selected are identified in a reference picture index. This ...
  31. [31]
    [PDF] Core Transform Design in the High Efficiency Video Coding (HEVC ...
    Similar to previous video coding standards, HEVC specifies two-dimensional transforms resembling the inverse discrete cosine transform (IDCT) for all transform ...
  32. [32]
    Understanding in-loop filtering in the HEVC video standard
    Jun 25, 2013 · As shown, it is a cascading of two stages - namely de-blocking filtering (DBLK) and Sample adaptive offset (SAO) filtering to remove blocking ...Missing: deblocking | Show results with:deblocking
  33. [33]
    (PDF) Complexity analysis of next-generation HEVC decoder
    This paper analyzes the complexity of the HEVC video decoder being developed by the JCT-VC community. The HEVC reference decoder HM 3.1 is profiled with ...
  34. [34]
    Compression performance of high efficiency video coding (HEVC ...
    This paper presents the results of compression comparison tests between the current state of the emerging High Efficiency Video Coding (HEVC) draft standard ...<|control11|><|separator|>
  35. [35]
    AVC/H.264 - ViaLa - Via Licensing Alliance
    Via LA's AVC/H.264 Patent Portfolio License provides access to essential patent rights for the AVC/H.264 (MPEG-4 Part 10) digital video coding standard.Licensors · Licensees · FAQ · Submit a Patent
  36. [36]
    [PDF] Apple ProRes
    Apple ProRes is one of the most popular codecs in professional post‑production. The ProRes family of video codecs has made it both.
  37. [37]
    [PDF] Avid High Resolution Workflows Guide
    Transcoding takes the original essence and converts it to either Avid DNxHR, DNxHD or DNxSD media in a different codec with a data rate of choice. If you ...
  38. [38]
    How to Choose the Right Codec for Every Project - Frame.io Insider
    Feb 15, 2017 · The first is the “proxy” workflow or “offline edit.” This means that you are transcoding your captured footage into an intermediate format, ...
  39. [39]
    [PDF] Study Group Report High-Dynamic-Range (HDR) Imaging Ecosystem
    It is expected that at least 10-bit representation will generally be required for support of HDR/WCG content in codecs, signal paths, file formats and in ...
  40. [40]
    [PDF] SMPTE TECHNICAL SPECIFICATION
    Implementations shall support the combinations of JPEG 2000 IMF profiles (as specified in ISO/IEC 15444-. 1:2016) and image frame dimensions listed in Table ...
  41. [41]
    [PDF] HDTV production codec tests - EBU tech
    Whilst HDTV equipment in the past has targeted tape-based solutions, the user requirements for modern HDTV production workflows are file-based, non-linear ...
  42. [42]
    A/341, “Video – HEVC”Approved: 17 July 2025 - ATSC.org
    This Standard describes the video coding constraints on ITU-T Rec. H.265 | International Standard ISO/IEC 23008-2 (“HEVC”) when it is used for video ...
  43. [43]
    [PDF] H.265-HEVC-Tutorial-2014-ISCAS.pdf
    Ultra-‐HD 4K broadcast expected for Japan in. 2014. London Olympics Opening and Closing. Ceremonies shot in Ultra-‐HD 8K. → Need higher throughput! • ...
  44. [44]
    - YouTube
    May 27, 2025 · We've officially launched AV1 for live streams on YouTube! AV1 ... Your codec will start with av01 if you're watching an AV1 live stream.
  45. [45]
    Quality metrics for live streaming video | AWS for M&E Blog
    Mar 6, 2024 · The quality metrics used are peak-signal-to-noise ratio (PSNR), structural similarity index (SSIM), and perceptual video quality score (VMAF).
  46. [46]
    Per-Title Encode Optimization - Netflix TechBlog
    Dec 14, 2015 · Per-title encoding allows us to deliver higher quality video two ways: Under low-bandwidth conditions, per-title encoding will often give you ...
  47. [47]
    TikTok Video Size Guide: Best Dimensions for 2025 - Riverside
    Dec 12, 2024 · Recommended Codec: TikTok recommends using the H. 264 codec for video and the AAC codec for audio. These two codecs are both optimized for high ...
  48. [48]
    Delivering Video Content for Safari | Apple Developer Documentation
    GIFs can be up to 12 times as expensive in bandwidth and twice as expensive in energy use when compared to a modern video codec. Instead, use H.264-encoded MP4 ...
  49. [49]
    Compatible media transcoding - Android Developers
    Compatible media transcoding on Android automatically converts videos like HEVC to AVC when needed, but is off by default and is computationally expensive.
  50. [50]
    Blu-ray Disc Association Completes Ultra HD Blu-ray Specification ...
    May 13, 2015 · The Blu-ray Disc Association (BDA) has announced the completion of the Ultra HD Blu-ray specification and released the new logo that will ...
  51. [51]
    Streaming specifications - Roku Developer
    HEVC 4K encodings. Roku recommends HEVC for UHD encodings. Roku 4K devices support HEVC up to level 5.1 and bitrates up to 25Mbps. UHD encodings should ...
  52. [52]
    NVIDIA Video Codec SDK
    NVIDIA GPUs contain an on-chip hardware-accelerated video encoder (NVENC), which provides video encoding for H.264, HEVC (H.265) and AV1 codecs. The software ...Missing: consoles | Show results with:consoles
  53. [53]
    NVIDIA NVENC Obs Guide | GeForce News
    Jan 30, 2025 · GeForce RTX GPUs have dedicated hardware encoders (NVENC), letting you capture and stream content without impacting GPU or CPU performance.Missing: consoles | Show results with:consoles
  54. [54]
    Web video codec guide - Media | MDN
    This guide introduces the video codecs you're most likely to encounter or consider using on the web, summaries of their capabilities and any compatibility ...
  55. [55]
    High-Efficiency Video Coding (HEVC) - Amazon Web Services
    The step up to H.265 (or HEVC) requires more computing power than H.264, but is considerably more efficient, and offers improved video quality at lower ...
  56. [56]
    Explained: What is High Efficiency Video Codec (HEVC) and why ...
    Jan 26, 2022 · ... 4K video file recorded in the HEVC format will consume relatively less storage space compared to the same video file recorded in the regular H.Missing: impact | Show results with:impact
  57. [57]
    Video Compression and Its Impact on Streaming Costs - CacheFly
    Jun 25, 2025 · Higher compression efficiency means less data to store and transmit, reducing CDN costs. For example, switching from H.264 to HEVC can ...Missing: benchmarks | Show results with:benchmarks
  58. [58]
    What Is HEVC (H.265)? - Streaming Media
    H.265/High Efficiency Video Coding (HEVC) is the successor codec to H.264, which, like H.264, is jointly developed by the ISO/IEC Moving Picture Experts Group.
  59. [59]
    [PDF] Report ITU-R BT.2386-5 (03/2024) - Digital terrestrial broadcasting
    • RAI 1 HD, RAI 2 HD, RAI 3 (HD and SD, see § 8.1.10.2), RAI News 24. • ≈ 20 Mbit/s. • Coverage > 98%. –. Mux A and B: national. • Thematic channels (RAI Sport ...
  60. [60]
    HEVC Licensing: Misunderstood, Maligned, and Surprisingly ...
    Apr 22, 2025 · Google trained Gemini on its AV1-friendly content, and remarkably, AV1 remains royalty-free as of 2025. That reflects the predominant sentiment ...
  61. [61]
    AV1 vs HEVC: Know Exactly Which Codec to Choose - VideoProc
    Aug 29, 2025 · While AV1 offers next-generation compression efficiency and royalty-free licensing, HEVC (H.265), developed by the MPEG group under the Joint ...
  62. [62]
    AV1 at Netflix: Redefining Video Encoding for a New Era of Streaming
    “AV1 also reduces the bandwidth needed for 4K streaming, leading to a 5% increase in 4K viewing hours, and minimizes noticeable quality down-switches by 38%, ...Av1 At Netflix: Redefining... · Advantages Of Av1 · Av1 For All Industry...
  63. [63]
    H.266 / VVC - Fraunhofer Heinrich-Hertz-Institut
    Compared to its predecessor H.265/HEVC, VVC achieves about 50% bit-rate reduction at the same subjective video quality for a wide range of video content and ...
  64. [64]
    The video codec landscape 2020 - ITU
    Jun 8, 2020 · This paper provides an analysis of the coding tools of VVC and EVC, stable since January 2020, and of AV1 stable since 2018. The quality and ...
  65. [65]
    Samsung Electronics, Huawei and Qualcomm Welcome the ...
    May 8, 2020 · The main goal of the EVC standard is to provide a significantly improved compression capability over existing video coding standards with timely ...
  66. [66]
    [PDF] low complexity enhancement video coding (lcevc) - ITU
    Rather than being a replacement for existing video coding schemes, LCEVC is designed to leverage existing (and future) codecs to enhance their performances ...
  67. [67]
    Low Complexity Enhancement Video Codec - LCEVC - A New ...
    LCEVC is a new video standard by MPEG. It specifies an enhancement layer which, when combined with a base video encoded with a separate codec, produces an ...How LCEVC Works · Contact · Integrations · Resources
  68. [68]
    Improving our Video Encodes for Legacy Devices - Netflix Research
    Nov 28, 2023 • Aditya Mavlankar, Software Engineer - Video & Image Encoding ... VMAF (Video Multi-Assessment Fusion) is a quality metric that combines ...Missing: AI | Show results with:AI
  69. [69]
    [PDF] Motion Information Propagation for Neural Video Compression
    In most existing neural video codecs, the information flow therein is uni-directional, where only motion coding provides motion vectors for frame coding.
  70. [70]
    Towards AI-Assisted Sustainable Adaptive Video Streaming Systems
    Jun 4, 2024 · This survey provides a comprehensive overview of the video streaming lifecycle, content delivery, energy and Video Quality Assessment (VQA) metrics and models.
  71. [71]
  72. [72]
    Libavcodec Documentation - FFmpeg
    The libavcodec library provides a generic encoding/decoding framework and contains multiple decoders and encoders for audio, video and subtitle streams, and ...
  73. [73]
    x264, the best H.264/AVC encoder - VideoLAN
    ### Summary of x264 Key Features
  74. [74]
    x265 Documentation - Read the Docs
    x265 Documentation¶. Introduction · About HEVC · About x265 · LEGAL NOTICES · Command Line Options · Executable Options · Logging/Statistic Options ...
  75. [75]
  76. [76]
    H.265/HEVC Hardware Encoding and Decoding Support - Intel
    Intel® Quick Sync Video uses the dedicated media processing capabilities of Intel® Graphics Technology to decode and encode fast. This enables the processor ...
  77. [77]
    [PDF] AMD CDNA™ 2 ARCHITECTURE
    The VCN blocks support H.264/AVC, HEVC, VP9 and JPEG for decode, as well as H.264/AVC and HEVC for encode3.
  78. [78]
    Improving Video Quality and Performance with AV1 and NVIDIA Ada ...
    Jan 18, 2023 · NVIDIA NVENC AV1 offers substantial compression efficiency with respect to H.264 and HEVC at better performance.Nvidia Nvenc Av1 Performance · Psnr Score · Split Encoding 8k60
  79. [79]
    Qualcomm Snapdragon 888 Plus 5G Mobile Platform
    8K encode and decode capabilities allow ultra-detailed video capture and viewing, seamlessly and without large data consumption. Plus, the Qualcomm Spectra ...
  80. [80]
    Hardware-Accelerated Video Streaming Efficiency & Savings
    Jul 21, 2025 · By replacing CPU-based encoders with hardware-accelerated VPUs, streaming platforms can cut transcoding costs by up to 90%.Missing: load | Show results with:load
  81. [81]
    Accelerated Video Encoding: G&L Systemhaus illustrates how ...
    Jul 9, 2025 · Specialised VPUs can boost energy efficiency by up to 3× over CPU encoding, keeping real-time streaming speeds and broadcast quality while ...
  82. [82]
    5 Benefits: VideoProc Converter AI's Full Hardware Acceleration
    Jan 27, 2024 · Learn how full GPU hardware acceleration in VideoProc Converter AI speeds up large (4K) video conversion, editing, and processing for faster ...
  83. [83]
    The State of the Video Codec Market 2025 - Streaming Media
    Mar 28, 2025 · I'm here to help you decide whether it's time to go all in on AV1, VVC, LCEVC, or EVC or whether it's better to stick with H.264, VP9, and HEVC.Missing: vendor lock-
  84. [84]
    Codec Guide: K-Lite Codec Pack - For Windows 11 / 10 / 8.1 / 7
    The K-Lite Codec Pack is a free software bundle for high quality playback of all your music and video files. It is easy to use, but also very flexible with ...
  85. [85]
    Download DirectShow FilterPack 5.1 - Free-Codecs.com
    DirectShow FilterPack is a lightweight collection of filters which allows you to play the most popular multimedia formats.
  86. [86]
    CCCP Codec Pack - Freewaregenius.com
    (A codec pack is a collection of Directshow filters that enable media players and media editing applications to decode/read/play and sometimes encode/write ...
  87. [87]
    Container File Formats: Definitive Guide (2023) - Bitmovin
    Jun 14, 2022 · ISO Base Media File Format (ISOBMFF, MPEG-4 Part 12) is the base of the MP4 container format. ISOBMFF is a standard that defines time-based ...
  88. [88]
    About MKV file format - MakeMKV
    MKV stands for "Matroska Video". Matroska is a container file format, capable of holding unlimited number of video, audio and subtitile tracks.
  89. [89]
    Frequently Asked Questions - The WebM Project
    WebM files consist of video streams compressed with the VP8 or VP9 video codec, audio streams compressed with the Vorbis or Opus audio codecs, and WebVTT text ...
  90. [90]
    Video Container File Formats: A Developer's Guide - FastPix
    Sep 19, 2025 · During playback, a demuxer (demultiplexer) separates the streams, allowing the media player to decode and render the video, audio, and subtitles ...Missing: syncing | Show results with:syncing
  91. [91]
    What Are Container File Formats (Media Containers)? - Cloudinary
    Container formats are file formats that encapsulate multimedia data such as audio and video streams, along with additional information such as metadata, ...Common Video Container... · Container Formats Concepts · Single Coding Formats Vs...
  92. [92]
    AV1 video format | Can I use... Support tables for HTML5, CSS3, etc
    AV1 video format ; Chrome. 4 - 66 : Not supported. 67 - 69 : Not supported. 70 - 141 : Supported. 142 : Supported ; Edge *. 12 - 17 : Not supported. 18 - 115 : ...
  93. [93]
    Know Your Rights: H.264, patent licensing, and you - Engadget
    May 4, 2010 · As an end user, you'll never have to think about your legal liability over H.264, because there's no need for you to be licensed unless you're distributing ...