Fact-checked by Grok 2 weeks ago

Video Coding Experts Group

The Video Coding Experts Group (VCEG) is an informal group of technical experts operating under the Telecommunication Standardization Sector () Study Group 21, dedicated to the development of international standards for the compression coding of and still images to enable efficient and in applications. VCEG's work traces its origins to the early 1980s, with the group's foundational efforts culminating in the first video coding standard, H.120, published in 1984 for low-bitrate videoconferencing over primary digital group , and revised in 1988. This marked the beginning of a series of H.26x standards, including (1988) for p×64 kbit/s videoconferencing and (1996) for low-bitrate video communication, which laid the groundwork for modern video technologies. In recognition of its pioneering contributions, VCEG's efforts were voted the most influential work of in 2006 and received the ITU 150th Anniversary Award in 2015. A hallmark of VCEG's approach has been its close collaboration with the /International Electrotechnical Commission (ISO/IEC) Joint Technical Committee 1 Subcommittee 29 Working Group 11 (MPEG), resulting in jointly developed standards that dominate global video deployment. Notable joint achievements include H.262 (, 1995) for broadcasting, H.264/ (AVC, 2003) which accounts for a significant portion of traffic, and H.265/ (HEVC, 2013) offering up to 50% bitrate reduction over AVC for high-definition content. VCEG also contributes to still image standards through partnerships with ISO/IEC JTC1/SC29/WG1, supporting families like and . In recent years, VCEG has focused on next-generation technologies through the Joint Video Experts Team (JVET), co-led with MPEG, which finalized H.266/ () in 2020 to support ultra-high-definition video, 360-degree immersive content, and screen content with improved compression efficiency. Ongoing activities include extensions to for enhanced scalability and explorations into neural network-based video coding, as highlighted in joint ITU-MPEG workshops on future video technologies incorporating advanced and . Following a January 2025 joint workshop, VCEG and MPEG are preparing a Call for Evidence and Proposals for a new generation of standards beyond .

Overview

Role and Mandate

The Video Coding Experts Group (VCEG) operates as Question 6/21 within the Telecommunication Standardization Sector (ITU-T) Study Group 21 (SG21), effective from the 2025-2028 study period, following the consolidation of the former Study Group 16 (SG16) on multimedia technologies with Study Group 9 (SG9) on broadband cable and television, as decided at the World Telecommunication Standardization Assembly (WTSA) in 2024. This positioning underscores VCEG's role in advancing standardized coding techniques for visual and related signals amid evolving ecosystems. VCEG's primary mandate is to develop and maintain ITU-T Recommendations on coding methods for visual, speech, audio, and other signal data, tailored for both conversational services—such as real-time videoconferencing and requiring low-delay —and non-conversational applications like streaming and broadcast television. This includes producing standards that support of video sequences, still images, graphics, stereoscopic and multi-view content, light fields, point clouds, volumetric imagery, computer-generated displays, , 360-degree video, and virtual/ (VR/AR) elements, with an emphasis on flexibility across transport mechanisms like the , networks, and ITU-T's H.222.0 multiplexing framework. Central to VCEG's efforts is the pursuit of high compression efficiency that preserves audiovisual quality, enabling efficient bandwidth utilization for diverse applications including , video streaming, and mobile communications, while balancing trade-offs in , quality, delay, and . These standards also incorporate features for error resilience and the integration of digital signing to ensure multimedia content authenticity and synchronization within coded streams. VCEG collaborates closely with ISO/IEC JTC 1/SC 29/WG 11 (MPEG) on joint initiatives, such as the Joint Video Experts Team (JVET), to create hybrid standards like (). Since its origins in 1984 with the development of the H.120 , VCEG has aimed to standardize methods for and image to minimize requirements in global telecommunication networks, fostering and efficiency in visual data transmission.

Organizational Structure

The Video Coding Experts Group (VCEG) operates as a rapporteur group under Question 6/21 of ITU-T Study Group 21 (SG21), focusing on visual, audio, and signal technologies. SG21 was established for the 2025-2028 study period to address , content delivery, and related applications. VCEG convenes through multiple meetings annually, typically including 3-4 plenary sessions aligned with SG21 schedules, supplemented by interim rapporteur group meetings, such as virtual sessions and in-person gatherings like the one held in , , in June–July 2025. Leadership of VCEG is provided by a and associate rapporteurs, elected by participants to guide technical coordination, manage contributions, and oversee the development of outputs. The current rapporteur is Gary Sullivan from Dolby Laboratories, , supported by associate rapporteurs who assist in specific areas of video coding standardization. These roles ensure efficient progression of work, including the integration of collaborative efforts with bodies like the Joint Video Experts Team (JVET). Participation in VCEG is open to ITU-T Member States, Sector Members (such as industry entities including , , , and Fraunhofer), Associates, and academic institutions, fostering contributions from diverse stakeholders. Contributions from these participants are rigorously reviewed through mechanisms like core experiments—structured tests to evaluate proposed technologies—and calls for evidence, which solicit demonstrations of compression performance beyond existing standards, such as in the development of H.266/ (). Decision-making in VCEG follows ITU-T's consensus-based approach, where agreements are reached through general discussion and resolution of objections among experts, avoiding formal except for approving final Recommendations by Member States. Documents are classified into contributions (formal inputs for discussion), temporary documents (informal working papers), and standards drafts, which evolve toward ITU-T Recommendations through iterative review. This structure evolved from VCEG's prior placement under Question 6/16 of 16 (SG16), following decisions at the World Telecommunication Standardization Assembly (WTSA-24) in , which merged SG16 and SG9 into SG21 to enhance focus on technologies.

History

Formation and Early Developments

The Video Coding Experts Group (VCEG) traces its origins to 1984, when it emerged as part of the (ITU, then known as the CCITT) efforts to standardize visual over emerging networks. This initiative responded to the transition from analog to telecommunications, particularly the rollout of (ISDN), which promised higher bandwidth for applications. Initial meetings under the CCITT's XV focused on developing codecs for videoconferencing, led by the Specialists Group on Coding for Visual , chaired by Sakae Okubo of NTT. The group's mandate emphasized low-bitrate video compression to enable real-time transmission within the constraints of early infrastructure, such as primary rate ISDN channels at rates up to 2.048 Mbit/s. The first outcome of this work was the H.120 standard, published by the CCITT in 1984 and revised in 1988, marking the inaugural international recommendation for coding. H.120 targeted ISDN-based videophones and videoconferencing at p×64 kbit/s rates, where p is an integer multiplier (typically up to 30 for full primary rate). It employed differential (DPCM) with conditional replenishment to compress video signals with limited motion, suitable for the era's low-bandwidth channels. This standard facilitated the transmission of monochrome video at resolutions like 256×240 for or 256×288 for PAL, but its performance was limited by the absence of advanced prediction techniques, resulting in modest compression efficiency. Building on H.120's foundations, the group advanced to in 1990, informally known as the "Px64" standard, which introduced a hybrid coding framework that became pivotal for subsequent video standards. utilized motion-compensated (DCT) coding, combining block-based motion prediction with transform-based residual compression to achieve better efficiency at bitrates from 64 kbit/s to 2 Mbit/s. It targeted resolutions of QCIF (176×144 pixels) and (352×288 pixels), enabling real-time decoding on 1980s hardware like early processors. Key challenges included optimizing for computational constraints—such as limiting processing to ensure decoding within 33 ms per frame—and handling transmission errors over ISDN without robust error correction, which often led to visible artifacts in low-bitrate scenarios. These early developments laid the groundwork for block-based hybrid coding paradigms still in use today.

Key Milestones and Evolution

The Video Coding Experts Group (VCEG) achieved a significant in 1994 with the joint development of H.262 alongside the (MPEG), resulting in the MPEG-2 Video standard, which became foundational for broadcasting and DVD storage by supporting and higher bit rates up to 20 Mbps. This collaboration marked VCEG's shift from standalone efforts toward integrated standardization with ISO/IEC, addressing the growing demand for multimedia applications beyond initial uses. In 1996, VCEG released , a pivotal advancement over that enhanced compression efficiency for low-bit-rate video transmission, making it suitable for emerging videoconferencing with features like negotiable options for adaptability. Over the subsequent decade, H.263 evolved through multiple extensions, culminating in Version 3 in 2005, which incorporated improvements in error resilience and scalability to meet the needs of and mobile networks. The 2000s saw VCEG deepen its partnership with MPEG through the formation of the Joint Video Team (JVT) in 2001, leading to the completion of in 2003, which offered roughly double the compression efficiency of prior standards and supported diverse profiles for broadcasting, streaming, and mobile devices. This era reflected VCEG's adaptation to the proliferation of , emphasizing robustness against transmission errors amid the rise of internet-based video. Entering the 2010s, VCEG collaborated via the Joint Collaborative Team on Video Coding (JCT-VC), established in 2010, to develop H.265/High Efficiency Video Coding (HEVC) finalized in 2013, achieving about 50% better compression than H.264/AVC to handle high-definition and ultra-high-definition content efficiently. In 2015, the Joint Video Exploration Team (JVET) was formed to explore technologies beyond HEVC, incorporating scalability and advanced error resilience for multimedia ecosystems driven by broadband expansion. Overall, VCEG's progression from telephony-centric standards like H.261 to versatile multimedia solutions mirrored the transition to widespread digital broadband, prioritizing interoperability and efficiency for global video deployment. In 2024, the World Telecommunication Standardization Assembly consolidated ITU-T Study Group 16 (encompassing VCEG) with Study Group 9 into the new Study Group 21 for the 2025-2028 period, continuing multimedia coding work under an expanded mandate.

Video Coding Standards

Early Video Standards

The Video Coding Experts Group (VCEG) established its foundational work with Recommendation ITU-T H.120, approved in 1984 and revised in 1988, marking the first international standard for digital video compression. The 1984 version (v1) utilized differential pulse code modulation (DPCM) and conditional replenishment for intra-frame coding of changed areas, processing video signals without inter-frame prediction or motion compensation to compress monochrome or color video for videoconferencing. The 1988 revision (v2) introduced motion compensation for changed pixels and background prediction. Operating at fixed bit rates of 1.544 Mbit/s for 525-line/60 fields per second systems (NTSC) and 2.048 Mbit/s for 625-line/50 fields per second systems (PAL), H.120 was designed for transmission over primary digital group channels, achieving limited compression suitable primarily for static or low-motion videophone scenarios. Advancing beyond intra-frame limitations, VCEG developed H.261 in 1990 (initial draft 1988), introducing the hybrid coding paradigm that combined block-based motion compensation with discrete cosine transform (DCT) for inter-frame prediction, enabling more efficient compression of temporal redundancies. Targeted at audiovisual services over integrated services digital network (ISDN) lines, H.261 supported flexible bit rates of p \times 64 kbit/s, where p = 1 to 30, commonly ranging from 384 to 1536 kbit/s for practical deployments. The algorithm divided frames into 16×16 luminance macroblocks (with 8×8 chrominance blocks in 4:2:0 sampling), applying motion compensation on 16×16 or 8×8 blocks followed by 8×8 DCT transformation and scalar quantization; a post-loop filter was incorporated to attenuate blocking artifacts arising from block boundaries in the reconstructed frames. Motion estimation employed full-search block matching, selecting the displacement vector \mathbf{v} = (v_x, v_y) that minimizes the sum of absolute differences (SAD) over a predefined search window: \text{SAD}(\mathbf{v}) = \sum_{m=0}^{N-1} \sum_{n=0}^{N-1} \left| f(t, x + m, y + n) - f(t-1, x + m + v_x, y + n + v_y) \right| where f(t, \cdot, \cdot) denotes the pixel intensity at time t, and N = 16 or 8 depending on the block size. This structure enabled compression ratios of 80:1 to 100:1 for quarter common intermediate format (QCIF) video, supporting real-time videotelephony with acceptable quality over ISDN channels. To address the needs of emerging low-bit-rate applications like streaming, VCEG finalized in 1996 as an evolution of , incorporating five baseline picture formats—sub-QCIF (128×96), QCIF (176×144), CIF (352×288), 4CIF (704×576), and 16CIF (1408×1152)—along with support for custom formats up to 31 distinct resolutions for flexibility in diverse systems. Key enhancements included optional negotiable modes such as unrestricted motion vector mode, which permitted vectors to extend beyond picture boundaries using or replication to improve prediction accuracy in edge regions, and advanced prediction mode, featuring overlapped block and up to four 8×8 motion vectors per for finer granularity in handling complex motion. These innovations, built on the hybrid framework with added deblocking filters and options, boosted compression efficiency, achieving ratios up to 100:1 for QCIF at bit rates as low as 20 kbit/s, making pivotal for early web video codecs like and enabling widespread adoption in ISDN-based telephony and dial-up streaming. These early VCEG standards provided the core architectural principles—intra/inter prediction, , and —that influenced subsequent developments, such as H.264.

Advanced Video Standards

The Video Coding Experts Group (VCEG), in collaboration with the (MPEG) through the Joint Video Team (JVT), developed H.264/ (AVC) in 2003 as a major advancement for compression. This standard introduced key innovations such as multiple reference frames for , allowing up to 16 reference pictures to improve prediction accuracy; Context-Adaptive Binary (CABAC) for encoding, which provides higher efficiency than previous methods; and spatial intra-prediction modes to reduce redundancy within frames. These features enabled up to 50% better compression efficiency compared to while maintaining similar video quality. H.264/AVC also defined profiles tailored to applications, including the Main profile for broadcast and streaming, and the High profile for high-definition content in mobile and professional environments. Building on H.264/AVC, VCEG and MPEG advanced to H.265/ (HEVC) in 2013 via the JVT, targeting even greater efficiency for ultra-high-definition video. HEVC employs larger Coding Tree Units (CTUs) up to 64×64 pixels, compared to the 16×16 macroblocks in H.264/AVC, which reduces overhead and supports higher resolutions. It features flexible partitioning through a structure for Coding Units (CUs) and Prediction Units (PUs), including asymmetric modes, alongside advanced motion vector prediction using spatial and temporal candidates in merge and AMVP modes. Rate-distortion optimization in HEVC encoders minimizes the cost function: J = D + \lambda R where D represents (e.g., sum of squared differences), R is the , and \lambda is the balancing the trade-off. H.264/AVC saw widespread adoption, serving as the primary for Blu-ray Disc and streaming, enabling efficient delivery of broadcast-quality content. HEVC, in turn, became the standard for Ultra HD Blu-ray, supporting higher bit depths and resolutions with reduced bandwidth needs. Extensions enhanced versatility: (MVC) for H.264/AVC enabled stereoscopic , while Scalable HEVC (SHVC) added layered scalability to HEVC for adaptive streaming. Technical advances in these standards included provisions for to address in hardware implementations. HEVC introduced tiles for independent rectangular regions and slices for segmented decoding, facilitating multi-core processing without interdependencies. These mechanisms reduced and enabled efficient encoding/decoding on consumer devices. Such innovations laid the groundwork for subsequent video coding efforts.

Modern and Emerging Video Standards

The Versatile Video Coding (VVC) standard, formally known as H.266 and developed through the Joint Video Exploration Team (JVET) comprising VCEG and MPEG, was finalized in 2020 to achieve substantial compression improvements for high-resolution content. VVC delivers 30-50% bitrate savings over its predecessor HEVC while maintaining equivalent video quality, as demonstrated in JVET's core experiments using BD-rate metrics. This efficiency stems from advanced tools such as affine motion compensation, which models complex non-translational motion using a two- or three-parameter affine model for sub-block prediction, and decoder-side motion vector derivation (DMVD), which refines motion vectors at the decoder without additional signaling overhead to enhance merge mode accuracy. Additionally, palette mode optimizes coding for screen-sharing and graphics content by representing blocks with a small set of representative colors, reducing residual data. VVC supports ultra-high definitions up to 8K resolution at 30 frames per second with bitrates under 100 Mbps, enabling practical transmission of immersive and broadcast-quality video. Extensions to VVC address emerging applications, including adaptations for through multi-model coding that projects spherical content onto multiple equirectangular frames for efficient compression. Integration with compression leverages VVC's core engine in video-based point cloud coding (V-PCC), where 3D geometry and attributes are projected into 2D video streams for encoding, achieving higher fidelity for volumetric media compared to prior HEVC-based approaches. Ongoing amendments focus on low-latency enhancements, such as gradual decoder refresh (GDR) without picture reordering to minimize in streaming, and refined parameter sets for ultra-low latency applications like vehicular networks and live events. These extensions maintain while expanding VVC's versatility for immersive and interactive formats. Since 2021, VCEG through JVET has explored next-generation codecs beyond , with the Enhanced Compression Model () serving as a for incremental improvements targeting 20-50% additional gains via AI-assisted tools like neural network-based prediction and in-loop filtering. Efforts emphasize support for 16K resolutions and machine-optimized coding for analytics-driven applications. As of 2025, video coding efforts have moved to 21. A Joint Call for Evidence was issued in July 2025, with submissions evaluated in October 2025. Depending on results, a Call for Proposals will follow, targeting a new standard (potentially H.267) operational around 2029-2030, with AI integration emphasized in collaborations like those between , , and Fraunhofer HHI for 6G-era applications. JVET tests of ECM versions confirm approximately 25-30% BD-rate savings over (and thus ~60-70% over HEVC) in configurations, underscoring the trajectory for post-VVC advancements.

Image Coding Standards

The Video Coding Experts Group (VCEG), in collaboration with ISO/IEC JTC 1/SC 29/WG 1 (the group), has contributed to several still standards, focusing on spatial redundancy reduction through techniques like block-based transforms and without temporal elements. These efforts emphasize efficient for professional photography, archiving, and high-dynamic-range () content, supporting both lossless and lossy modes to preserve quality while minimizing file sizes. A key contribution is ITU-T Recommendation H.273 (first published in 2014), which defines coding-independent code points (CICP) for identifying video signal types, including still images, and supports extensions through color space transformations such as from to as specified in BT.2100. This standard enables interoperability for still images by standardizing for color primaries, transfer characteristics, and matrix coefficients, facilitating with existing decoders. JPEG XT (ISO/IEC 18477 series, published starting 2015) is a backward-compatible extension of the baseline standard designed for and scalable still image compression, developed by the JPEG committee. JPEG XT employs a base layer using traditional JPEG discrete cosine transform (DCT) coding, augmented by residual coding for enhanced and , targeting continuous-tone photographic content with support for up to 16 bits per component. This approach achieves approximately 20-30% better compression efficiency than for HDR images at equivalent quality levels, particularly in professional archiving where lossless modes are critical. Additionally, VCEG co-developed the Main Still Picture in H.265/HEVC (ITU-T Recommendation H.265, 2013), which applies intra-only coding tools—such as angular intra modes and larger transforms (up to 64x64)—exclusively to single-frame images, excluding inter-frame motion . This leverages -based transforms and spatial to reduce redundancy within the image, offering lossy and lossless options suitable for high-resolution stills in applications like . Evaluations indicate HEVC intra coding provides 16-25% bitrate savings over for standard dynamic range images, with even greater gains for content when combined with H.273 color mappings. These intra tools draw directly from HEVC's video intra-frame mechanisms but operate independently for still image use.

Current Work and Collaborations

Ongoing Projects

As of 2025, the Video Coding Experts Group (VCEG), operating under Study Group 21 (SG21) Question 6/21, continues to advance amendments to (VVC, ITU-T H.266) through the Joint Video Experts Team (JVET). These amendments, initiated post-finalization in 2020, include enhancements to in-loop filtering technologies such as (ALF) and cross-component ALF (CC-ALF), as well as provisions for transform skip modes to improve coding efficiency for specific types like screen content. The work targets further with AI-driven pipelines, with ongoing refinements expected to culminate in additional recommendation updates by 2026, building on versions like H.266 (V3) released in 2023. VCEG's exploration of next-generation video coding has progressed since 2023 via a dedicated post-JVET exploration team, emphasizing neural network-based coding tools for superior compression and adaptability. This includes Exploration Experiments (EE) such as EE1 on neural network-based video coding, launched around 2021 and continuing into 2025, which evaluates models for prediction, transformation, and filtering to achieve gains beyond . Sustainability aspects are integrated through focuses on energy-efficient encoding algorithms, aiming to reduce computational overhead for deployment in resource-constrained environments while targeting at least 20-50% bitrate reduction over at equivalent quality. Under SG21's broader multimedia framework, VCEG projects integrate video coding with emerging applications, such as efficient representation for immersive environments including scenarios, where high-fidelity volumetric and data require optimized compression. This involves coordination with other SG21 questions on systems and services, ensuring seamless for IP-based and delivery networks. Specific outputs include advancements in coding for machines, as noted in statements, to support analytics in video streams. Recent VCEG activities in 2024-2025 have centered on JVET meetings and workshops advancing a call for evidence through the aforementioned exploration experiments, evaluating proposals for 50% or greater efficiency improvements over baselines. Key events include the JVET-AM meeting in in June 2025, reviewing and enhanced compression results, and the joint ITU-ISO/IEC workshop in from July 1-3, 2025, which discussed integration and standards beyond VVC. The inaugural SG21 meeting in (January 13-24, 2025) established priorities for visual coding under the new structure, with follow-up sessions in October 2025 further progressing these efforts. Ongoing challenges for VCEG include balancing patent licensing frameworks—where ITU-T standards like incorporate royalty-bearing essential patents—against demands for more accessible, royalty-free alternatives in open ecosystems, while prioritizing hardware-accelerated implementations for edge devices to enable low-latency, power-efficient processing in and mobile applications. These considerations are addressed through collaborative test conditions in the exploration experiments, ensuring practical deployability without excessive complexity.

Partnerships and Joint Efforts

The Video Coding Experts Group (VCEG) has maintained a long-term partnership with the (MPEG) since the 1990s, beginning with the collaborative development of the H.262 video coding standard, which was jointly published as ITU-T Recommendation H.262 and ISO/IEC 13818-2 ( Video). This cooperation continued with the formation of the Joint Video Team (JVT) in 2001, which developed the (AVC) standard, finalized in 2003 as ITU-T H.264 and ISO/IEC 14496-10 (MPEG-4 Part 10). Subsequent joint efforts included the Joint Collaborative Team on Video Coding (JCT-VC) from 2010 to 2015 for (HEVC, ITU-T H.265 | ISO/IEC 23008-2), and the Joint Video Experts Team (JVET) from 2015 to 2020 for (VVC, ITU-T H.266 | ISO/IEC 23090-3). These partnerships enable dual approval under and ISO/IEC, facilitating widespread adoption by combining ITU's telecommunication focus with ISO's broader industry branding. VCEG's contributions to , such as the H.264/MVC extension developed jointly with MPEG, have supported advancements in immersive technologies, including MPEG Immersive Video (MIV, ISO/IEC 23090-12) finalized in , which enables (6DoF) rendering for volumetric content captured by multiple views. Shared testing frameworks, like core experiments conducted within these joint teams, have streamlined development by evaluating proposals collaboratively, ensuring efficient for diverse applications. The ISO branding, as seen with AVC's designation as MPEG-4 Part 10, has broadened in and streaming services. VCEG maintains ties with the 3rd Generation Partnership Project () through the adoption of its standards in mobile video profiles, such as HEVC Main 10 Profile for UHD in 3GPP Release 14 and VVC profiles for enhanced efficiency in multimedia transmission. Similarly, the Digital Video Broadcasting () Project incorporates VCEG-developed codecs like HEVC for UHD broadcast in and added support in 2022 to enable 8K delivery within existing bandwidth constraints. These integrations promote , with VCEG providing the core coding tools tailored for mobile and broadcast environments via liaisons and profile specifications. In recent years, VCEG has pursued enhanced collaborations on AI integration in video coding, highlighted by the January 2025 joint SG21 and ISO/IEC JTC 1/SC 29 on future video incorporating advanced signal processing and techniques beyond . With the establishment of 21 (SG21) in 2025, which merged former SG16 (home to VCEG) and SG9 (focused on broadcast and cable systems), coordination on systems integration has intensified, supporting holistic standards that combine with delivery architectures.