Fact-checked by Grok 2 weeks ago

MPEG-1

MPEG-1 is a suite of international standards (ISO/IEC 11172) developed by the (MPEG) for the of moving pictures and associated audio, targeted at digital storage media with bit rates up to approximately 1.5 Mbit/s to enable VHS-quality video and CD-quality audio playback. The standard consists of five parts: Part 1 (Systems) defines the and of audio and video streams; Part 2 (Video) specifies a compressed representation of progressive video sequences, typically at resolutions like 352×240 pixels () with support for intraframe (I), predictive (P), and bidirectional (B) pictures using motion-compensated (DCT) coding; Part 3 (Audio) outlines three hierarchical layers (I, II, and III) for high-quality audio encoding at sampling rates of 32, 44.1, or 48 kHz and channels; Part 4 addresses compliance testing; and Part 5 provides reference software simulations. Established in 1988 under ISO/IEC JTC 1, the MPEG working group finalized MPEG-1 in 1993 following initial approvals in 1991, building on earlier video coding efforts like to address the need for efficient storage and retrieval of audiovisual content on emerging media such as CD-ROMs. Key features include hybrid coding techniques for video—combining temporal prediction with spatial frequency transformation—and perceptual coding for audio to minimize audible artifacts, enabling compression ratios of about 26:1 for video and 6:1 for audio while supporting features like , /rewind, and editing. Originally designed for applications like on personal computers, video on , and low-bitrate video transmission, MPEG-1 found widespread adoption in (VCD) format, which stores up to 74-80 minutes of standard-definition video on a single CD, as well as early streaming and due to its low requirements and broad with players. The audio component, particularly Layer III (commonly known as ), revolutionized digital music distribution by allowing high-fidelity sound in compact files, though the full standard's video capabilities laid foundational techniques for subsequent MPEG versions like MPEG-2. As an , MPEG-1 remains relevant for legacy media preservation and low-resource environments, with reference implementations available for decoding and encoding.

Introduction

Definition and Scope

MPEG-1, formally known as ISO/IEC 11172, is an developed by the (MPEG) for the of video and audio data. It enables the encoding of raw at Video Home System (VHS) quality and (CD) quality audio into a combined bitrate of approximately 1.5 Mbit/s, facilitating efficient storage and playback of content. This standard was conceived in the late to address the need for practical digital compression. The MPEG-1 specification is structured into five parts, each addressing a specific aspect of the and delivery process. Part 1 (Systems) defines the and of audio and video streams into a single . Part 2 (Video) specifies the algorithms for . Part 3 (Audio) outlines the coding methods for associated sound. Part 4 (Conformance) provides testing procedures to ensure compliance with the 's requirements. Part 5 (Reference Software) includes software implementations for encoding and decoding as a reference for verification. MPEG-1 was primarily targeted at applications involving digital storage on , where the effective bitrate aligns closely with single-speed data rates of around 1.4 Mbit/s. Additionally, its design supports transmission over digital channels with capacities such as 1.544 Mbit/s, corresponding to primary multiplex rates in regions like the and .

Design Objectives

The MPEG-1 standard was developed with the primary objective of enabling efficient compression of and audio for storage on media such as , targeting a total bitrate of up to 1.5 Mbit/s to fit within the constraints of early digital storage capacities. Specifically for video, the design aimed to compress raw digital footage at 30 frames per second and 352×240 pixel resolution—equivalent to VHS-quality Source Input Format ()—down to under 1.5 Mbit/s, achieving compression ratios around 26:1 while preserving acceptable visual quality suitable for interactive multimedia applications. For audio, the objectives focused on compressing stereo sound sampled at rates like kHz to bitrates between 128 and 384 kbit/s, providing near-transparent quality without perceptible loss for most listeners and enabling with the compressed video stream. This range supported CD-quality ratios of approximately 6:1, ensuring compatibility with the overall system bitrate limits. Key non-compression goals emphasized practical usability for storage and transmission, including support for to video segments within about 0.5 seconds, resilience to bit errors common in optical media like , and precise audio-video synchronization to maintain lip-sync and temporal alignment during playback. These features were integral to the five-part standard structure, which encompasses systems, video, audio, and to facilitate interoperable delivery.

Historical Development

Origins in Compression Research

The development of MPEG-1 originated from foundational research in the 1980s on video and techniques, particularly hybrid (DCT)-based coding for video and perceptual models for audio, aimed at enabling efficient storage and transmission of multimedia on emerging digital media like compact discs. Early video coding efforts built on intraframe DCT compression, which transforms spatial data into frequency coefficients to exploit redundancies, combined with differential (DPCM) for prediction, as explored in projects like the IVICO initiative starting in 1984, which integrated DCT with rudimentary to handle interframe dependencies. These hybrid approaches reduced bandwidth needs significantly; for instance, motion-compensated DCT schemes in late-1980s experiments achieved compression ratios that brought video bitrates down from tens of megabits per second to around 1-2 Mbit/s for acceptable quality, laying the groundwork for block-based processing in 16x16 macroblocks. On the audio side, perceptual coding models emerged from psychoacoustic research, focusing on human to discard inaudible signal components and achieve high-fidelity compression at lower bitrates. The MUSICAM (Masking-pattern adapted Universal Subband Integrated Coding And Multiplexing) project, funded under the Eureka EU147 initiative for (DAB) from 1987, developed subband filtering and bit allocation based on masking thresholds, enabling stereo audio compression to 192-384 kbit/s with near-transparent quality. Contributions from institutions like France's CCETT, Germany's IRT, and refined these models through subjective listening tests, emphasizing polyphase filter banks for efficient , which directly influenced the layered structure of later audio codecs. Early experiments in the 1980s further advanced video efficiency by estimating and subtracting interframe motion vectors, reducing temporal redundancy; prototypes from and European labs demonstrated that block-matching algorithms could predict pixel displacements, improving by up to 50% over static intraframe methods alone. These disparate efforts across video and audio, driven by needs for broadcast and storage applications, highlighted the necessity for unified standards. To consolidate this research, the (MPEG) was established in January 1988 under ISO/IEC JTC1/SC2 in , initiated by Leonardo Chiariglione and Hiroshi Yasuda to coordinate international efforts on integrated audiovisual coding. By May 1988, the group's first meeting in attracted 29 experts, transitioning oversight to the newly formed SC29 (Coding of Audio, Picture, Multimedia and Hypermedia Information) and WG11, focusing on synchronizing video and audio streams for applications like playback at 1.5 Mbit/s. This formation bridged ongoing European and Japanese projects, setting the stage for a cohesive standard without delving into formal ratification processes.

Standardization Process

The (MPEG), established under the (ISO) and (IEC) Joint Technical Committee 1/Subcommittee 2 in early 1988, held its inaugural meeting in May 1988 in , , with initial participation from 29 experts representing industry and academia. This gathering marked the start of collaborative efforts to develop a standard for compressed and audio suitable for storage media like . The group rapidly expanded, involving over 100 experts in subsequent meetings focused on practical implementation challenges. The standardization process unfolded through distinct phases, beginning with requirements definition in , which outlined target bit rates around 1.5 Mbit/s for VHS-quality video and audio while emphasizing complexity for playback. This led to a call for proposals, followed by intensive testing from 1990 to 1991, where submitted codecs underwent subjective evaluations by panels of viewers and listeners to assess perceptual quality. A draft proposal emerged in 1990, incorporating hybrid motion-compensated techniques selected from competitive submissions. Central to refinement were iterations of the verification model (initially termed simulation model), which integrated audio and video components through collaborative core experiments to optimize performance and . These models were validated via tests on real hardware decoders, confirming feasibility for consumer devices with limited processing power. The process culminated in the committee draft stage by late 1991 and final approval in 1992, with the complete MPEG-1 standard published as ISO/IEC 11172 in August 1993.

Patent and Licensing

Key Patents and Holders

The core patents underpinning MPEG-1 video compression, particularly those involving the (DCT) for spatial compression and for temporal prediction, were originally held by major electronics firms including Corporation, Koninklijke Electronics N.V., and Thomson Consumer Electronics. These patents formed the foundational for implementing the video encoding algorithms standardized in MPEG-1 Part 2, enabling efficient compression of streams at bitrates suitable for delivery. and , in particular, contributed key innovations in motion-compensated prediction and block-based DCT processing, while Thomson advanced related hardware implementations critical for consumer devices like Video CDs. For the audio components of MPEG-1, the patents for Layer II—derived from the MUSICAM (Masking-pattern adapted Universal Subband Integrated Coding and Multiplexing) algorithm—were held by and the research institute CCETT (now part of Labs), with contributions from the Institut für Rundfunktechnik (IRT). These entities licensed their through Sisvel Technology S.r.l., emphasizing techniques that achieved high-quality stereo audio at around 192 kbit/s. In contrast, the patents for Layer III, based on the Adaptive Spectral Perceptual Entropy Coding (ASPEC) scheme, were primarily owned by the , Bell Laboratories, and the (MIT), with additional input from Thomson-Brandt and . Fraunhofer's perceptual coding advancements, refined through collaborative EU147 project efforts, enabled superior compression efficiency for Layer III, supporting bitrates as low as 128 kbit/s while preserving audio fidelity. Essential patents for MPEG-1 audio Layers I-III were collectively licensed through Sisvel's MPEG Audio program, while video patents were typically licensed individually from the holders.

Expiration Status

All essential patents covering the MPEG-1 standard, including those for its video and audio components, expired by 2018, rendering the technology royalty-free for implementations worldwide. The final core patents for MPEG-1 Audio Layer III (), held by Fraunhofer IIS and , expired worldwide on December 30, 2017, while earlier video-related patents, such as US 4,472,747 listed in the ISO patent database, had expired in 2003. This shift to royalty-free status has significantly encouraged the adoption and distribution of MPEG-1 in software and open-source projects, removing previous legal barriers that restricted inclusion in distributions like . Open-source decoders, such as those in FFmpeg, can now be freely integrated without patent licensing concerns, fostering broader support for MPEG-1 playback in multimedia applications. While pure MPEG-1 implementations face no ongoing patent obligations, developers working on derivative technologies or systems combining MPEG-1 with later standards like should verify the status of those extensions, though patents have similarly expired globally by 2025. Original key holders, including Fraunhofer for audio and various contributors to the , no longer enforce royalties on the standard.

Systems Integration (Part 1)

Elementary Streams and Packets

In MPEG-1, elementary streams represent the fundamental output of individual encoders for a single type of media, such as video or audio, consisting of a continuous sequence of coded data units without additional or overhead. These streams are self-contained bitstreams that adhere to the defined in ISO/IEC 11172-2 for video or ISO/IEC 11172-3 for audio, ensuring compatibility with decoders while maintaining a near-real-time flow suitable for storage or transmission at up to about 1.5 Mbit/s. For instance, a video elementary stream comprises sequences of units like I-frames, P-frames, and B-frames, each representing complete images or predicted differences. To facilitate handling and synchronization in a system context, elementary streams are packetized into Packetized Elementary Streams (PES), where the continuous data is divided into discrete packets, each beginning with a header followed by contiguous bytes from the stream. PES packets have variable lengths, typically ranging from tens to hundreds of kilobytes depending on the application, allowing flexibility for error-free environments like digital storage media. The packet header starts with a 24-bit start code prefix (0x000001) to delineate packet boundaries, followed by a stream_id byte that uniquely identifies the media type—such as 0xE0 for video or 0xC0 for audio—to enable demultiplexing at the receiver. Optional fields in the header include packet length, markers for stuffing bytes, and crucially, timestamps for timing control. Timing and synchronization within PES packets rely on Presentation Time Stamps () and Decoding Time Stamps (DTS), which are 33-bit values encoded in the header to align media presentation and decoding with a clock. The PTS specifies the exact time a presentation unit (e.g., a video or audio ) should be displayed or played, while the DTS indicates when an access unit must be decoded, particularly important for non-display-order decoding in B-frames where decoding precedes presentation. Both timestamps operate at a precision of 90 kHz, derived from a 27 MHz clock divided down, providing sub-millisecond accuracy for lip-sync between audio and video; for example, PTS values increment in 90 kHz units to reflect the intended output timing in the target decoder model. If DTS is absent, it is inferred to equal PTS, simplifying processing for streams without reordering needs. System-level synchronization across packets is further ensured by the System Clock Reference (SCR), embedded periodically in the program stream headers, which samples the encoder's 27 MHz clock and conveys it as a 33-bit value at 90 kHz to initialize and correct the decoder's internal clock. This , along with extension fields for finer , allows decoders to lock onto the sender's timing within a tolerance of 30 , preventing drift over long playback durations. PES packets containing these elements form the building blocks that are subsequently multiplexed into higher-level program streams for combined audio-video delivery.

Program Streams and Multiplexing

In MPEG-1, program streams serve as the primary for combining one or more compressed elementary streams—such as video and audio—into a single optimized for reliable and retrieval on like CD-ROMs. Defined in the systems layer of ISO/IEC 11172-1, these streams employ variable-length packets for use in error-free environments. Unlike fixed-length formats, the variable packet sizes accommodate the irregular of compressed content, enabling efficient use of storage space while supporting playback rates up to approximately 1.5 Mbit/s. The process integrates multiple elementary s by packetizing them into Packetized Elementary Streams (PES) and interleaving these within larger pack structures. Each PES packet encapsulates data from a single elementary , prefixed by a header that includes a stream identifier and optional extension fields for timing information. Packs, in turn, wrap one or more PES packets along with system headers that specify overall parameters, such as sizes and initial values. This hierarchical interleaving ensures that data from different streams arrives at the in the correct order, with the process governed by the System Target (STD) model—a hypothetical reference that defines capacities to prevent overflow or underflow during demultiplexing. For the video component, the Video Buffering Verifier () model further constrains the multiplexed by simulating a input of specified size, verifying that the arrival rate avoids discontinuities in playback. Synchronization across streams relies on a of timestamps embedded in the pack and packet headers to align decoding and . The System Clock Reference (SCR) in each pack header provides a reference value sampled from the encoder's 27 MHz system clock (encoded as a 90 kHz with 300x precision), initializing and periodically correcting the decoder's System Time Clock (STC) to within 30 ppm tolerance. Time Stamps () in PES headers indicate the intended display time for a presentation unit relative to the STC, while Decoding Time Stamps (DTS) specify the decoding start time, particularly for video frames requiring reordering. These timestamps, also based on the 90 kHz clock, ensure audio-video lip-sync by enforcing simultaneous of units with matching values, with fields required at intervals not exceeding 700 ms to maintain continuity. The STD model incorporates these timestamps to regulate buffer occupancy, guaranteeing that decoding delays remain under 1 second for seamless playback.

Video Compression (Part 2)

Color Space and Resolution

MPEG-1 video encoding operates in the color space, where the Y component represents and the Cb and Cr components represent differences, facilitating efficient compression by prioritizing luminance detail. This standard employs , reducing the resolution of the Cb and Cr components to one-quarter of the Y component's resolution, which aligns with human and enables lower bitrates without significant perceptual loss. Each component is quantized to 8-bit precision, allowing for 256 levels per sample to balance quality and computational efficiency in the target storage media. The supported spatial resolutions are constrained to the Source Input Format (SIF), specifically 352 × 240 pixels for NTSC systems or 352 × 288 pixels for PAL systems, derived from the Common Intermediate Format (CIF) to suit CD-ROM storage capacities. Temporal resolution is limited to progressive frame rates of 29.97 frames per second for NTSC or 25 frames per second for PAL, ensuring compatibility with broadcast standards while avoiding the complexity of interlaced scanning. The maximum video bitrate is capped at 1.856 Mbit/s to fit within the overall 1.5 Mbit/s system constraint when combined with audio, promoting reliable playback on early digital media players. MPEG-1 mandates only, eschewing interlaced formats to simplify decoding and reduce artifacts in the primary application of video CDs. The default is 4:3 for broadcast compatibility, though signaling for 16:9 is supported to accommodate emerging display technologies without altering the core pixel grid. These specifications collectively ensure that MPEG-1 delivers VHS-equivalent quality at constrained bitrates, optimized for non-real-time storage and retrieval.

Frame Types and GOP Structure

In MPEG-1 video , as defined in ISO/IEC 11172-2, pictures are categorized into distinct types to balance spatial and temporal redundancy reduction while enabling efficient decoding and . The primary frame types are intra-coded (I-frames), predictive-coded (P-frames), and bidirectionally predictive-coded (B-frames), each employing where applicable to exploit inter-frame correlations. I-frames are self-contained pictures encoded solely through intra-frame techniques, requiring no to other frames for decoding; they serve as anchor points for subsequent predictions and facilitate error recovery or scene changes. P-frames are forward-predicted from a preceding I- or P-frame using motion vectors to estimate , transmitting only the differences after compensation, which typically halves the volume compared to I-frames. B-frames achieve the highest efficiency by interpolating from both preceding and succeeding I- or P-frames, using bidirectional motion vectors, but they are not used as references to avoid error propagation, resulting in about one-quarter the of I-frames. Additionally, D-frames provide a specialized intra-coded option limited to the DC coefficients of blocks, enabling low-detail, rapidly decodable pictures for applications like fast-forward playback in video . These frames are organized into Groups of Pictures (GOPs), which form the fundamental unit for access and decoding in MPEG-1 bitstreams, always beginning with an I-frame to support . A typical GOP pattern, such as I B B P B B P B B I, sequences frames to optimize compression—often spanning 9 to 15 pictures—while GOP length is influenced by target resolutions like or QCIF to maintain bitrates around 1.5 Mbit/s. GOPs may be flagged as closed, where all pictures decode independently without referencing the next GOP, ideal for or splicing, or open, permitting predictions across GOP boundaries for enhanced efficiency at the cost of greater dependency. The following table summarizes the key characteristics of MPEG-1 frame types for clarity:
Frame TypeCoding MethodReference DependencyCompression EfficiencyPrimary Use Case
I-frameIntra (spatial only)None (self-contained)Lowest, error recovery
P-framePredictive (forward motion)Past I- or P-frameMediumTemporal prediction
B-frameBidirectional predictivePast and future I- or P-framesHighestMaximum
D-frameIntra ( coefficients only)NoneVery high (low quality)Fast playback modes

Macroblocks and Motion Vectors

In MPEG-1 video coding, the image is partitioned into , which serve as the fundamental units for and compensation. Each consists of a 16×16 block of luma samples accompanied by two 8×8 blocks ( and ) in format, effectively dividing the chroma resolution to half that of luma in both horizontal and vertical directions. This structure aligns with the used in the standard, enabling efficient processing of video at resolutions like SIF (Source Input Format). Macroblocks support several coding modes to balance compression efficiency and quality, particularly in predicted () and bidirectionally predicted (B) frames within a (GOP). In intra (I) mode, a macroblock is encoded without reference to other frames, relying solely on spatial redundancy within the current picture. modes include forward prediction for P-frames using a single motion vector relative to the previous I- or P-frame, backward prediction for B-frames relative to the next frame, and interpolated prediction in B-frames that averages forward and backward compensated blocks. , applicable in P- and B-frames, inherit the motion vector from the previous without transmitting residual data, reducing bitrate for static regions. Motion vectors describe the displacement of a from its best-matching position in a reference frame, computed via block-matching algorithms during encoding. They achieve half-pel accuracy through of reference s, allowing finer motion representation than integer-pel shifts. The search for the optimal vector typically occurs within a ±15.5 range around the 's position in P- and B-frames, using full or logarithmic search methods as specified in the reference simulation model to minimize (MAD) or similar distortion metrics. To optimize bitrate, motion vectors are differentially encoded relative to the vector of the left or above neighboring , with the difference (motion vector difference, MVD) variable-length coded using Huffman tables tailored to predicted frame types. This exploits spatial correlations in motion, ensuring vectors for B-frames can reference either forward, backward, or both directions as per the macroblock mode.

DCT Transform and Quantization

In MPEG-1 video compression, the (DCT) is applied to 8×8 blocks of motion-compensated prediction residuals to decorrelate the data in the spatial domain and concentrate energy into lower-frequency coefficients. This transform operates on pixel differences ranging from -255 to 255, producing coefficients typically between -2048 and 2047, which facilitates efficient subsequent compression by exploiting spatial redundancy. The two-dimensional forward DCT for an 8×8 block f(x,y) is defined as: F(u,v) = \frac{1}{4} C(u) C(v) \sum_{x=0}^{7} \sum_{y=0}^{7} f(x,y) \cos\left[ \frac{(2x+1) u \pi}{16} \right] \cos\left[ \frac{(2y+1) v \pi}{16} \right] where C(u) = \frac{1}{\sqrt{2}} for u = 0 and 1 otherwise, and similarly for C(v). The inverse DCT (IDCT) follows a symmetric form to reconstruct the spatial block, enabling decoders to approximate the transform for reduced while adhering to precision requirements that limit reconstruction errors to at most 1 unit. Quantization follows the DCT to achieve lossy compression by scaling and rounding the coefficients, using a uniform scalar quantizer with 31 possible step sizes controlled by a quantization parameter Q ranging from 1 (finest) to 31 (coarsest). This process divides each coefficient F(u,v) by the product of Q and an element from a predefined weighting matrix W(u,v), which varies for luminance (luma) and chrominance (chroma) components to apply finer quantization to low frequencies and coarser to high frequencies, aligning with human visual sensitivity. Separate default matrices are specified for intra-coded and inter-coded blocks, with the luma matrix emphasizing preservation of detail in brighter areas and the chroma matrix allowing greater quantization due to reduced perceptual importance of color differences. After quantization, the resulting 8×8 matrix of coefficients is reordered using a zigzag scan pattern, which traverses from low to high frequencies in a diagonal manner to group nonzero values at the beginning for more efficient encoding of sparse data. Dequantization at the reverses this by multiplying the quantized values by Q \times W(u,v), followed by the IDCT, introducing controlled primarily in high-frequency components that are less visible to the eye.

Entropy Coding

In MPEG-1 video compression, entropy coding employs Huffman-based variable-length coding (VLC) to further reduce redundancy in the quantized data, assigning shorter codewords to more probable symbols such as small motion vector differences, common macroblock types, and frequent run-level pairs of DCT coefficients. This step follows the quantization of DCT-transformed blocks, where the resulting coefficients are scanned in a zigzag order to group low-frequency values together before encoding. The standard defines 11 distinct VLC tables to optimize coding efficiency for different elements of the bitstream. Motion vectors are differentially coded relative to predictors and then mapped to tables that favor small differences, typically ranging from -16 to +15.5 in half-pel precision, with longer codes for larger displacements. Macroblock types, indicating modes like intra-coded or predicted with , use separate tables tailored to intra (I), predicted (P), and bidirectional (B) frames, ensuring efficient representation of common configurations such as skipped or forward-predicted blocks. For DCT coefficients, run-level encoding represents sequences as pairs where "run" denotes the number of consecutive zero coefficients and "level" the amplitude of the next nonzero value; these pairs are then VLC-coded using tables that prioritize short runs and small levels. The end-of-block (EOB) , signaling no further nonzero coefficients in the block, is encoded with a compact 2-bit code (10 in binary). Rare events, such as long runs exceeding 40 zeros or levels beyond ±127, trigger an mechanism: a dedicated 14-bit code (0000 0 0000001 0) followed by explicit 6-bit run and 8-bit level fields (or 16 bits for linear level coding in some cases), allowing representation of any valid quantized value without table expansion. The resulting variable-length codes are packed into the starting from the most significant bit, with achieved by appending zero bits as needed to reach multiples of 8 bits; this ensures and prevents of start codes, which are forbidden 23-bit sequences beginning with 0x000001. bits are inserted only when necessary, typically after slice or picture headers, to maintain a constant through buffering.

Audio Compression (Part 3)

Layer I Encoding

Layer I is the simplest audio compression layer in the MPEG-1 standard (ISO/IEC 11172-3), optimized for real-time encoding and decoding in applications like and compression, where low and moderate data reduction are prioritized over high efficiency. It supports or mono audio at sampling frequencies of 32 kHz, 44.1 kHz, and 48 kHz, with bit rates ranging from 32 kbit/s to 448 kbit/s in increments that facilitate compatibility with various transmission channels. At a typical bitrate of 384 kbit/s, Layer I can compress CD-quality audio ( at 44.1 kHz) by a factor of about 4:1 while preserving perceptual transparency for most signals. The encoding process begins with a critically sampled 32-subband polyphase mirror filter (PQMF) bank, which decomposes the input PCM audio into 32 uniform subbands spanning 0 to half the sampling frequency. The polyphase implementation employs a low-pass filter with 512 taps to achieve adequate stop-band attenuation and minimize , processing 12 samples per subband for a total of 384 subband samples per frame. This design ensures perfect reconstruction in the absence of quantization, with a frame duration of 8 ms at 48 kHz sampling. The psychoacoustic Model 1 then analyzes the signal using a 512-point FFT to estimate tonal and noise-like components, computing simultaneous masking via the and critical bands, as well as temporal masking to detect transients and prevent pre-echo artifacts by shaping noise in the . Bit allocation is determined by comparing subband signal energy to the masking threshold, assigning 0 to 15 bits per subband to keep quantization noise inaudible. Quantization follows, applying 32 scalefactors to normalize subband sample amplitudes before non-uniform quantization, which uses a companding-like curve to emphasize low-level signals. The quantized samples and scalefactors are formatted into the bitstream without a bit reservoir, relying instead on fixed-rate padding for constant bitrate output and ensuring decoder simplicity. Side information includes a 4-bit allocation field per subband and 6-bit scalefactor values, with optional CRC for error detection; subband samples are encoded in fixed-length codes based on their allocated bits, packed sequentially by subband to facilitate parallel decoding. Layer I's straightforward approach laid the foundation for subsequent layers, which build upon it with enhanced perceptual modeling and entropy coding for better efficiency at lower bitrates.

Layer II Encoding

MPEG-1 Audio Layer II encoding enhances the framework introduced in Layer I by incorporating more sophisticated rate control and stereo processing, achieving better compression for mid-bitrate applications. It derives from the MUSICAM (MUSeum Integrated auDio Coding And Multiplexing) , jointly developed by , CCETT, and the Institute for Broadcasting Technology (IRT) under the European 147 project in the late 1980s. This foundation enabled Layer II to support bitrates from 128 to 384 kbit/s, targeting applications like and video soundtracks with improved perceptual quality over Layer I at equivalent rates. The core transforms the input audio into 32 equally spaced subbands using a critically sampled polyphase , which includes reduction through a 512-tap prototype to suppress imaging artifacts and ensure low across the 0-20 kHz audible range. Each processes 1152 time-domain samples (36 samples per subband), organized into three of 384 samples each, with each further divided into three scalefactor bands for finer in quantization and bit allocation. This structure allows dynamic scaling of subband signals based on perceptual importance, grouping 12 subband samples into one band for the first two and 6 for the third to optimize encoding overhead. A bit reservoir mechanism, with a capacity of up to 511 bits, facilitates flexible rate control by allowing unused bits from previous or subsequent to be stored and borrowed, accommodating transient signals that demand higher instantaneous bitrates without exceeding the target stream rate. For multichannel audio, Layer II employs joint stereo modes to exploit interchannel correlations, including intensity stereo—where high-frequency subbands (above 4-8 kHz) are coded with a single mono signal plus channel-specific intensity factors to preserve spatial image—and middle-side (MS) stereo, which transforms left and right channels into sum (M) and difference (S) signals for more efficient coding of correlated content. The psychoacoustic Model 2 analyzes the signal via a 1024-point FFT over 23 ms windows to derive masking thresholds, incorporating simultaneous masking (frequency-domain spread around tonal and noise-like components) and sequential masking (temporal pre- and post-masking effects up to 5 ms before/after a masker). These models guide bit allocation to noise-fill inaudible regions, ensuring quantization noise remains below the masking curve. At 192 kbit/s for 44.1 kHz stereo audio, Layer II delivers transparent quality comparable to uncompressed CD audio for most program material, as validated in subjective listening tests during standardization.

Layer III Encoding

Layer III, the most sophisticated audio compression layer in MPEG-1, originated from the Adaptive Spectral Perceptual Entropy Coding (ASPEC) algorithm developed by researchers at Fraunhofer Institute for Integrated Circuits (IIS), AT&T Bell Labs, Thomson-Brandt, and Centre National d'Etudes des Telecommunications (CNET) in the late and early . This approach achieved high by integrating perceptual modeling with advanced , enabling near-transparent audio at bit rates as low as 128 kbit/s for signals. The core of Layer III encoding is a hybrid filter bank that builds on the 32 equal-width subband polyphase filter bank from Layers I and II, followed by a Modified Discrete Cosine Transform (MDCT) applied to each subband signal. The MDCT operates with variable window sizes—36 samples for long blocks (yielding 18 spectral coefficients per subband) and 12 samples for short blocks (6 coefficients)—with special start and stop windows for transitions between block types and a mixed mode combining long blocks for low frequencies and three short blocks for higher frequencies to handle transients, allowing adaptive resolution for stationary versus non-stationary audio content with 50% overlap between blocks. This hybrid structure provides finer frequency resolution (up to 576 total spectral lines per granule) and better aliasing cancellation compared to prior layers, enhancing efficiency at low bit rates. Quantized spectral coefficients are grouped into 21 scalefactor bands approximating critical bands of human hearing, with scale factors adjusted iteratively to minimize perceptual distortion based on the psychoacoustic model. The psychoacoustic model Type I, used in Layer III, employs 1024-sample analysis windows to compute masking thresholds by identifying tonal and noise-like components, enabling advanced bit allocation that prioritizes audible spectral regions. Entropy coding applies Huffman variable-length codes to the spectral lines, organized into regions of big values, count1 (values -1, 0, 1), and zeros, optimizing the bitstream for perceptual irrelevance. A bit reservoir mechanism allows bit borrowing across granules (up to 511 bits), providing flexibility for variable-rate encoding within fixed frame constraints and supporting the high compression ratios of Layer III. For error detection, an optional 16-bit () is included in the header, verifying in .

Compliance and Tools (Parts 4 and 5)

Conformance Testing Procedures

MPEG-1 conformance testing procedures, defined in ISO/IEC 11172-4, establish methodologies to verify that bitstreams and decoders adhere to the syntactic, semantic, and performance requirements outlined in Parts 1 (Systems), 2 (Video), and 3 (Audio) of the standard. These procedures focus on ensuring interoperability by testing decoder capabilities, such as arithmetic precision and decoding accuracy, against predefined bitstream characteristics like picture size, frame rate, and bitrate. Central to these procedures are synthetic test bitstreams provided in Part 4, which include sequences for video, audio, and systems designed with precisely known outputs to enable objective evaluation of ._Compliance_Testing.zip) For video testing, procedures assess core decoding functions with pass/fail criteria requiring that the reconstruction error for each sample does not exceed 1 least significant bit (LSB) when compared to the reference output, ensuring negligible deviation in decoded images. Audio conformance testing uses similar synthetic bitstreams to assess layer-specific decoding (I, II, III), verifying output samples against known values with tight error tolerances defined in the standard. For systems-level testing, procedures emphasize management verification to confirm that the decoder's input and output buffers operate within specified sizes and overflow/underflow limits, alongside checks for accuracy to ensure precise audio-video via presentation and decoding timestamps. These tests collectively guarantee that compliant decoders handle multiplexed streams without timing discrepancies exceeding 90 kHz clock resolution. Reference software from MPEG-1 Part 5 may be employed to generate additional test cases or validate results during conformance assessment.

Reference Implementation Software

The software for MPEG-1 is specified in ISO/IEC 11172-5:1998, which provides a C-language simulation of encoders and decoders for the core components of the standard. This technical report serves as the official baseline implementation, enabling developers and researchers to understand and verify the algorithms defined in Parts 1 (Systems), 2 (Video), and 3 (Audio) of ISO/IEC 11172. The software is designed primarily to illustrate the coding processes rather than to achieve high performance or real-time operation, using to ensure portability and accuracy in across different platforms. Key components include a video encoder and decoder based on the MPEG-1 video test model, supporting intra-coded (I), predictive-coded (P), and bidirectionally predictive-coded (B) frames with (DCT) processing and . For audio, the implementation covers all three layers—Layer I, Layer II, and Layer III ()—with subband filtering, psychoacoustic modeling, and for bitstream generation and decoding. The systems component handles and demultiplexing of video, audio, and into Program Streams, including buffer management and timestamping to maintain synchronization. All modules are integrated to process complete MPEG-1 bitstreams, with the source code available for non-commercial use via public download from the ISO website. This reference software plays a critical role in validating conformance to the MPEG-1 standard, allowing implementations to be tested against the provided algorithms and generated bitstreams for bit-exact decoding. It also acts as a foundational reference for developing encoders and decoders, ensuring while permitting optimizations for speed or constraints. often involves comparing outputs from this software against official test bitstreams to confirm adherence to the specification.

Applications and Formats

Primary Use Cases

MPEG-1 video compression found its primary application in the (VCD) format, which enabled the distribution of on standard compact discs with a resolution of 352x240 pixels for systems. Introduced in , VCDs allowed consumers to play movies and other video content on CD players equipped with MPEG-1 decoders, achieving approximately 74 minutes of playback per disc at bit rates around 1.15 Mbit/s. This format gained significant popularity in and other regions during the as an affordable alternative to tapes, particularly for home entertainment and . Additionally, MPEG-1 was employed in early transmission over networks, such as ISDN lines, supporting low-bit-rate delivery of VHS-quality video for applications like video conferencing and initial trials. In audio applications, MPEG-1 Layer III, commonly known as , revolutionized portable music by enabling efficient of CD-quality audio to files as small as one-tenth the original size, facilitating storage and transfer on early digital devices. Standardized in 1992, supported bit rates from 32 to 320 kbit/s and became the format for digital music players in the late and early , powering devices like the and later the . Meanwhile, MPEG-1 Layer II served as the core for (DAB), a terrestrial radio deployed across and other regions starting in the mid-1990s, providing stereo audio at 192 kbit/s within multiplexed transmissions for ensemble services. Combined video and audio capabilities of MPEG-1 underpinned interactive media formats like (Compact Disc Interactive), for which MPEG-1 support was added in 1993 via a Digital Video Cartridge add-on to existing players launched by in 1991. This enabled , graphics, and sound for educational, gaming, and reference applications on dedicated players. CD-i discs using MPEG-1 supported resolutions up to 352x288, with up to 72 minutes of video per disc and user interactions in titles such as encyclopedias and simulations. In the 1990s, MPEG-1 also powered systems, particularly in , where VCD-based players displayed synchronized lyrics and video clips alongside or Layer II audio tracks, making home and commercial karaoke accessible via affordable media. These uses demonstrated MPEG-1's versatility in bridging storage constraints with multimedia delivery during the transition to digital consumer electronics.

File Extensions and Containers

MPEG-1 content is typically stored using specific file extensions that indicate the type of stream, whether multiplexed or elementary. For multiplexed files combining video and audio in a program stream, the extensions and .mpeg are commonly used. Audio-only files corresponding to the three layers of MPEG-1 Audio are identified by .mp1 for Layer I, for Layer II, and for Layer III. Elementary streams, which contain only video or audio without , employ .m1v for video and .m1a for audio. The primary container format for MPEG-1 is the MPEG Program Stream (MPEG-PS), defined in ISO/IEC 11172-1 for synchronized video, audio, and optionally other data streams suitable for storage media. For broader compatibility, MPEG-1 elementary or program streams can be encapsulated in alternative containers such as or file formats. These container choices facilitate integration with legacy playback systems and applications like (VCD), which relies on MPEG-PS files. MPEG-PS files are identified by specific header signatures, including the pack start code 0x000001BA, which marks the beginning of each pack header and synchronizes the stream. This 32-bit value, consisting of a 24-bit start code prefix (0x000001) followed by the pack identifier (0xBA), ensures proper parsing of the multiplexed .

Legacy Impact

Modern Relevance

Following the expiration of MPEG-1 patents (primarily by , with the last remaining until in some jurisdictions), the has assumed a prominent role in archival , where its open and status facilitates long-term storage without licensing constraints. As of 2025, MPEG-1 is entirely worldwide, supporting unrestricted use in open-source and archival applications. Institutions such as the and the UK National Archives document MPEG-1 as a stable format for of legacy content at VHS-equivalent quality, though higher standards like are preferred for long-term archiving. In embedded systems, particularly resource-constrained devices like legacy industrial equipment and sensors, MPEG-1 persists for its low computational demands and compatibility with minimal hardware. Open-source tools, including FFmpeg and , continue to support MPEG-1 encoding and decoding, enabling developers to handle legacy content in modern workflows without proprietary dependencies. Despite this enduring utility, MPEG-1's inherent limitations—such as a maximum resolution of 352x240 pixels and a bitrate capped at approximately 1.5 Mbit/s—render it unsuitable for contemporary high-definition (HD) or 4K applications, where higher fidelity and efficiency are essential. It has been largely superseded by advanced standards like H.264/AVC for video compression and AAC for audio, which offer superior quality at comparable or lower bitrates, driving the shift in mainstream media production and streaming. These shortcomings highlight MPEG-1's obsolescence in bandwidth-intensive environments, confining its practical deployment to scenarios where backward compatibility or simplicity outweighs performance needs. MPEG-1 maintains niche persistence through its audio layer, which remains integral to portable music players in 2025, with devices like the FiiO M21 and NW-A306 supporting playback for offline listening in environments restricting use, such as workplaces. (VCD) formats, leveraging MPEG-1 video, endure in developing regions for affordable media distribution due to low-cost production and playback hardware. This practical continuity underscores MPEG-1's foundational influence on subsequent MPEG standards, though its direct applications have diminished.

Influence on Successor Standards

MPEG-1's video compression techniques, particularly the use of (DCT) for and block-based for inter-frame , formed the foundational basis for subsequent standards including , MPEG-4, and the H.26x series. In , the 8x8 DCT block processing from MPEG-1 was retained and extended to support formats, enabling efficient spatial redundancy reduction at higher up to 20 Mbit/s for broadcast applications. in MPEG-1, which exploited temporal redundancy through forward in P-frames and bi-directional in B-frames, was enhanced in with dual prime prediction modes and field-based macroblocks, improving compression efficiency for . These elements also influenced the standard, where MPEG-1's concepts were adapted for low-bit-rate video over the H.26x framework. The (GOP) structure introduced in MPEG-1, which organizes frames into sequences starting with intra-coded I-frames followed by predicted P- and bi-directionally coded B-frames, directly carried over to and became integral to streaming protocols. In , the GOP was made more flexible with configurable parameters like N=9 and M=3 for typical structures, supporting closed and open GOPs to facilitate editing, , and error resilience in transport streams. This influenced later streaming standards by providing a for temporal and efficient decoding in bandwidth-constrained environments. For audio, the perceptual model in MPEG-1 Layer III (), which estimates masking thresholds based on psychoacoustic principles to shape quantization noise, significantly shaped the design of (AAC) in and MPEG-4. AAC refined MP3's perceptual calculation and simultaneous masking model, incorporating higher (up to 1024 lines versus MP3's 576) and temporal noise shaping (TNS) to better control pre-echo artifacts and achieve equivalent quality at approximately 70% of the , such as 96 kbit/s versus 128 kbit/s for stereo audio. This evolution allowed AAC to support multichannel audio and lower bit rates while maintaining perceptual transparency, building directly on MP3's filterbank-based architecture using (MDCT). MPEG-1's approach, employing a polyphase filterbank with 32 equally spaced bands for frequency-domain analysis, influenced the perceptual strategies in (AC-3). AC-3 adopted similar psychoacoustic modeling for deriving quantization masking thresholds across 64 non-uniform subbands, adapting MPEG-1's dynamic bit allocation to adjust resolution based on and available bits, which enhanced efficiency for multichannel cinema and broadcast audio. Additionally, concepts like intensity from MPEG-1 informed AC-3's high-frequency above 2 kHz, reducing data transmission while preserving spatial cues through techniques such as the Karhunen-Loève transform. In the systems domain, MPEG-1's timestamping mechanisms, including presentation time stamps () and decoding time stamps (DTS) referenced to a 90 kHz system clock, provided the groundwork for in MPEG-2 transport streams () and later adaptive streaming protocols like . MPEG-2 extended this with program clock references () at 27 MHz for higher precision, multiple elementary streams into fixed-size packets with sequence numbers for error detection and seamless playback across error-prone channels. The packet-based from MPEG-1, which combined audio, video, and clock information into a single stream, evolved in MPEG-2 TS to support and , enabling broadcast standards like and ATSC. These principles influenced by incorporating timeline-based in media presentation descriptions (), where segments use inherited timestamping for adaptive bitrate switching over HTTP.

References

  1. [1]
    MPEG-1 standard
    A suite of standards for audio-video and systems particularly designed for digital storage media.
  2. [2]
    ISO/IEC 11172-1:1993 - Information technology — Coding of moving ...
    CHF 199.00 In stock 2–5 day deliverySpecifies the system layer of the coding. Was developed principally to support the combination of the video and audio coding methods defined in ISO/IEC 11172-2 ...
  3. [3]
    ISO/IEC 11172-2:1993 - Information technology — Coding of moving ...
    CHF 221.00 In stock 2–5 day deliverySpecifies the coded representation of video for digital storage media and the decoding process. Is primarily applicable to digital storage media supporting ...
  4. [4]
    ISO/IEC 11172-3:1993 - Information technology — Coding of moving ...
    2–5 day deliveryISO/IEC 11172-3:1993 specifies the coded representation of high quality audio for storage media at about 1.5 Mbit/s, for media like CD, DAT and magnetic hard ...
  5. [5]
    MPEG-1 Video Coding (H.261) - Library of Congress
    Dec 27, 2022 · ISO/IEC 11172; first approvals in 1991. Five parts have been published; parts 1, 2, and 3 are central. See list of ISO documents in Format ...<|separator|>
  6. [6]
    Video | MPEG
    MPEG-1 Video 1 IntroductionISO/IEC 11172-2 specifies a video codec which was originally designed for the application domain of video for CD storage.
  7. [7]
    Moving Picture Experts Group (MPEG)-1 - VA.gov
    Moving Picture Experts Group (MPEG)-1 defines a coding standard of moving pictures and associated audio for digital storage media at up to about 1.5 Mbit/s.
  8. [8]
    The MPEG Metamorphoses - Leonardo's Blog
    Feb 23, 2020 · The net bitstream from CD – 1.4 Mbit/s – is close to the 1.544 Mbit/s of the primary digital multiplex in USA and Japan. Therefore it was ...
  9. [9]
    [PDF] MPEG: A Video Compression Standard for Multimedia Applications
    1980's technology made possible full-motion video over networks. – Television and Computer Video seen moving closer.
  10. [10]
    [PDF] A tutorial on MPEG/audio compression - IEEE Multimedia
    The MPEG standard addresses the compression of synchronized video and audio at a total bit rate of about 1.5 megabits per second (Mbps).
  11. [11]
    [PDF] MPEG-1 and MPEG-2 Digital Video Coding Standards
    For accessing video from storage media the MPEG-1 video compression algorithm was designed to support important functionalities such as random access and fast ...Missing: design | Show results with:design
  12. [12]
    Inside MPEG-1 - Riding the Media Bits
    Aug 20, 2015 · The title of MPEG-1 is rather convoluted: “Coding of moving pictures and associated audio for digital storage media at up to about 1,5 Mbit/s”.
  13. [13]
    The true history of MPEG's first steps - Leonardo's Blog
    Dec 28, 2019 · The first MPEG meeting took place in Ottawa in May 1988 with 29 people in attendance. At the 4th MPEG meeting in December 1988 in Hannover, the ...
  14. [14]
  15. [15]
    Moving Picture Experts Group (MPEG) - Scholarpedia
    Jan 22, 2009 · In the 1980s many were working on video and audio coding. Nippon Hoso Kyokai (NHK) developed and deployed an innovative hybrid (analogue/digital) ...Missing: origins | Show results with:origins
  16. [16]
    Digital Audio Broadcasting - DAB | SpringerLink
    Mar 15, 2010 · The MUSICAM method, used up to the present for audio compression in MPEG-1 and MPEG-2 layer II and is also used in DAB or, to put it more ...
  17. [17]
    Video Coding History — Vcodex BV
    The DCT was to become the transform of choice for video codecs from the 1980s onwards.Missing: origins | Show results with:origins
  18. [18]
    MPEG Basics
    Starting from its first meeting in May 1988 when 29 experts participated, MPEG has grown to an unusually large ISO Working Group. Usually some 350 experts from ...
  19. [19]
    History of MPEG - Courses
    MPEG - 1. The MPEG-1 standard, established in 1992, is designed to produce reasonable quality images and sound at low bit rates. MPEG-1 consists of 4 parts:.
  20. [20]
  21. [21]
    MPEG licensing basics - EE Times
    Mar 18, 2005 · Some 24 different companies own the 650 worldwide essential patents for the MPEG-2 video standard for compressing digital information.
  22. [22]
    Response To Koninklijke Philips Electronics, N.V.'s, Sony ...
    Jun 25, 2015 · Sony and Pioneer have granted Philips nonexclusive, sublicensable licenses on their "essential" patents to enable Philips to grant licenses "to ...
  23. [23]
    [PDF] Managing Intellectual Property Using Patent Pools - Rudi Bekkers
    Table 1 shows an overview of three optical disc generations, their patent pools, and the main characteristics of these pools. The Compact Disc. In 1979 Philips ...
  24. [24]
  25. [25]
    Sisvel's Impact on MPEG Audio Standard
    The MPEG Audio patents owned by Sisvel formed the backbone of the MPEG Audio standard, which is essential to the existence of MP3 players.Missing: Layer CCETT
  26. [26]
    ASPEC. Adaptive Spectral Perceptual Entropy Coding of high ...
    ASPEC is the joint proposal of AT&T Bell Laboratories, Thomson Consumer Electronics, Fraunhofer Society and CNET for the audio coding standard which is ...Missing: developers MIT
  27. [27]
    Patent Fights Are a Legacy of MP3's Tangled Origins
    Mar 5, 2007 · Until now, the most prominent holder of MP3 patents has been the Fraunhofer ... Aspec, which went on to become the basis for MPEG-1 Audio ...
  28. [28]
    Patents and MP3 - MP3-Tech.org
    All developers and publishers of MPEG-audio layer 3 (MP3) encoders based on ISO-source must pay a license fee to Fraunhofer.Missing: 1 ASPEC Bell Labs MIT
  29. [29]
    Via LA Licensing
    Transparent and Collaborative Global Patent Licensing. The industry's largest patent pool administrator with dozens of IP licensing programs.MPEG-2 · VC-1 · MPEG-4 Visual · MPEG-2 Systems
  30. [30]
    [PDF] Patent pools in high-tech industries - IAM Media
    The. MPEG LA® Licensing Model offers fair, reasonable, non-discriminatory access to essential intellectual property from multiple patent owners under a single ...<|separator|>
  31. [31]
    Alive and Kicking – mp3 Software, Patents and Licenses
    May 18, 2017 · The mp3 licensing program ended, but existing licensees can still use the software. New users can contact Fraunhofer directly, but some patents ...Missing: ASPEC | Show results with:ASPEC
  32. [32]
    MPEG-1 - Just Solve the File Format Problem
    Dec 28, 2023 · This standard is old enough that any patents that may have encumbered it in the past are now expired. US patents 4,472,747 and 5,214,678 are ...<|control11|><|separator|>
  33. [33]
    MPEG-1 and MPEG-2 are now completely free/open for Linux ...
    Apr 12, 2019 · Yeah, that's how it is for patents. It's copyrights that have 70-100 year terms (70 years + the life of the author, to be specific; depends on ...Which year should we treat H.264 patents as expiring? - RedditPatents notwithstanding, what features from the H264/265 ... - RedditMore results from www.reddit.com
  34. [34]
    Patent List - MPEG-2 Systems - Via LA Licensing
    Please note that the last US patent expired February 13, 2018, and patents are presently active in Malaysia. Download.
  35. [35]
    [PDF] A Guide to MPEG Fundamentals and Protocol Analysis - Tektronix
    MPEG-2 is used for video, MPEG-1, layer 2 audio is still in use as the ... The Picture Quality Analyzer performs direct comparisons of video frames.
  36. [36]
    [PDF] ARIB STD-B24 Version 5.1-E1
    ISO/IEC 11172-2 shall be used for MPEG-1 Video coding with constraints specified in Table 4-1. ... YCBCR 4:2:0. Input pixel depth. 8 bit. Scanning method.
  37. [37]
    [PDF] MPEG-1 Video
    • ISO/IEC 11172-2: Video. • ISO/IEC 11172-3: Audio. • ISO/IEC 11172-4 ... Coding for Intra DC. • VLC and VLI for Intra DC differential values. Page 14. 14.
  38. [38]
    [PDF] Spring 2008 Basic Video Compression Techniques H.261, MPEG-1 ...
    – The current image frame is referred to as Target Frame. – A match is sought between the macroblock in the Target Frame and the most similar macroblock in ...<|control11|><|separator|>
  39. [39]
    [PDF] MPEG: a video compression standard for multimedia applications
    The DCT basis function (or subband decomposition) is suffi- ciently well-behaved to allow ef- fective use of psychovisual crite- ria. (This is not the case with.
  40. [40]
    [PDF] E cient Algorithms for MPEG Video Compression
    Dec 11, 2001 · Like H.261, the MPEG standards employ a hybrid video coding scheme that combines BMMC with 2D-DCT coding. Unlike H.261, the MPEG stan- dards ...
  41. [41]
    [PDF] MPEG video coding A simple introduction - EBU tech
    In MPEG coding, spatial redundancy is removed by processing the digitized signals in 2-D blocks of. 8 pixels by 8 lines (taken from either one field or two, ...
  42. [42]
    [PDF] 10. MPEG-1
    Motion vector range. < −64 to +63.5 pels (using half-pel vectors), etc. Input buffer size (in VBV mode). ≤ to 327 680 bits. Bitrate. ≤ 1 856 000 bits/second ...
  43. [43]
    [PDF] MPEG-1
    • Separate MB-type VLC tables for I, P, and B pictures. • Quantization tables. • VLC supports large range of DCT coefficients. 18-899/Spring 1998/Chen.
  44. [44]
    Audio | MPEG
    MPEG-1 Layer I or II Audio is a generic subband coder operating at bit rates in the range of 32 to 448 kb/s and supporting sampling frequencies of 32, 44.1 and ...Missing: tap | Show results with:tap
  45. [45]
    ISO-MPEG-1 Audio: A generic standard for coding of high-quality ...
    ISO-MPEG-1 Audio: A generic standard for coding of high-quality digital audio · audio coding · concealment · data compression · data reduction · Datenkompression.
  46. [46]
    [PDF] - Low delay audio coding for broadcasting applications - ITU
    ISO/IEC 11172-3 (MPEG-1 Audio) Layer II and MPEG-1 Layer III ... MPEG-1 Layer II uses a 32-channel polyphase filter bank (PQMF) with a length of 512 taps to map.
  47. [47]
    [PDF] Psychoacoustics
    All three layers are based on a subband coder using 32 equally wide frequency bands. Page 28. 28/42. MPEG-1, layer I ... Static Huffman coding (fixed codewords) ...
  48. [48]
    The MPEG Representation of Digital Media - SpringerLink
    MPEG-1 Layer I is the simplest layer and it typically operates at ... the analysis filterbank results in 32 subband samples in each of 64 equal-width sub-.
  49. [49]
    MPEG-1 and MPEG-2 Layer II Audio Encoding - Library of Congress
    Mar 26, 2024 · MPEG-1 Audio Layer II is defined in ISO/IEC 11172-3. An extension of MPEG-1 Layer II is defined in ISO/IEC 13818-3, which adds more sampling ...
  50. [50]
    [PDF] MPEG Audio
    Maximal downsampling. • Q should be based on signal-to-masking ratio (SMR). • Ear's critical bands are not uniform, but logarithmic.
  51. [51]
    MPEG: Motion Picture Experts Group - Audio
    MPEG layer 1 samples the PCM data in groups (frames) of 384 to form the spectral powers, whereas layer 2 uses frames of 1152 samples (ie. one frame every 24 ms) ...
  52. [52]
    [PDF] MPEG Digital Audio Coding Standards
    The main motivations for low bit rate coding are the need to minimize transmission costs or to provide cost-efficient storage, the demand to transmit over.
  53. [53]
    [PDF] A Tutorial on MPEG/Audio Compression
    The tests showed that even with a 6-to-1 compression ratio (stereo, 16 bits/sample, audio sampled at 48 kHz compressed to 256 kbits/sec) and under optimal ...
  54. [54]
    Conformance testing | MPEG
    MPEG conformance testing verifies if bitstreams and decoders meet requirements of ISO/IEC 11172, using procedures to test compliance.
  55. [55]
    ISO/IEC TR 11172-5:1998 - Information technology
    Coding of moving pictures and associated audio for digital storage media at up to about 1,5 Mbit/s — Part 5: Software simulation.Missing: MPEG- | Show results with:MPEG-
  56. [56]
    MPEG-1 Reference Software
    This International Standard provides a C language software simulation of an encoder and decoder for Part 1 (Systems), Part 2 (Video), and Part 3 (Audio) of ISO ...
  57. [57]
  58. [58]
    [PDF] Digital Video Systems Video coding standards: 3/2/17
    H.261 (1990): Video conferencing. MPEG-1 (1992): Non-interactive applications, VCD. MPEG-2 and H.262 (1996): TV broadcast, DVD. H.263 (Nov. 1995; Sept. 1997 ...<|separator|>
  59. [59]
    The impact of MPEG standards - Leonardo's Blog
    May 14, 2019 · MP3 (MPEG-1 Audio Layer III) brought a revolution in the music world ... Repeating the “MP3 use case for video” was the ambition of many.Missing: portable | Show results with:portable
  60. [60]
  61. [61]
    [PDF] TS 103 466 - V1.2.1 - Digital Audio Broadcasting (DAB) - ETSI
    From 384 consecutive input audio samples, 12 consecutive sub-band samples are generated for each of the 32 sub-bands. syncword: 12-bit code embedded in the MPEG ...
  62. [62]
  63. [63]
    [PDF] Legal and patent issues
    Dec 2, 2009 · MPEG-1 Layer 3 audio has at least three separate companies that claim to have patents, Alcatel-Lucent, Thompson and AudioMPEG. All their claimed ...Missing: ASPEC Bell
  64. [64]
    Details: File format summary - The National Archives
    Nov 14, 2023 · MPEG-1 is a lossy compression standard for video and audio, intended ... Digital Preservation Department / The National Archives. Source ...
  65. [65]
    [PDF] MPEG Digital Video Coding Standards
    Identical to the MPEG-1 standard, the MPEG-2 coding algorithm is based on the general Hybrid DCT/DPCM coding scheme as outlined in. Figure 5, incorporating a ...
  66. [66]
    FFmpeg
    Their support will help sustain the maintainance of the FFmpeg project, a critical open-source software multimedia component essential to bringing audio and ...Download FFmpeg · Documentation · About · Contact Us
  67. [67]
    HandBrake: Open Source Video Transcoder
    HandBrake is a open-source tool, built by volunteers, for converting video from nearly any format to a selection of modern, widely supported codecs.Downloads · Features · Community · NewsMissing: 2025 | Show results with:2025
  68. [68]
    Best MP3 player 2025: top portable hi-res music ... - TechRadar
    Sep 13, 2025 · 1. Activo P1. The best hi-res audio player for most people · 64GB ; 2. FiiO JM21. The best entry-level model ; 3. Astell & Kern A&norma SR35. The ...
  69. [69]
    [PDF] MP3 and AAC Explained
    stereo coding. AES 17th International Conference on High Quality Audio Coding. 10. Page 11. Karlheinz Brandenburg. MP3 and AAC explained. • Copy protection.Missing: ASPEC | Show results with:ASPEC
  70. [70]
    None
    ### Summary of MP3 Perceptual Coding Influence on AAC
  71. [71]
    [PDF] Design of the Audio Coding Standards for MPEG and AC-3
    This dissertation studies the design of audio standards: MPEG-1/2 and AC-3. The perceptual audio coder like MPEG-1/2 and AC-3 can be analyzed through filterbank ...Missing: influence | Show results with:influence
  72. [72]
    [PDF] Understanding Timelines within MPEG Standards
    DTS and PTS are equally present in the MP2P and. MP2T and have the same meaning as in MPEG-1. As in MPEG-1, in MPEG-2 the equations to obtain the. PTS and DTS ...
  73. [73]
    What would MPEG be without Systems? - Leonardo's Blog
    Mar 17, 2019 · Within the Service Channel bits 1-8 are used by the Frame Alignment Signal (FAS) and bits 9-16 are used by the Bit Alignment Signal (BAS).Missing: stuffing | Show results with:stuffing