Fact-checked by Grok 2 weeks ago

Sound quality

Sound quality is the perceptual evaluation of a sound's characteristics in relation to listener expectations and the intended context of use, arising from the interaction between the physical properties of the sound signal and human auditory perception. It is not an intrinsic property of the sound itself but emerges subjectively, often described as the adequacy of the sound for its purpose, such as in audio reproduction systems where deviations from expected fidelity can degrade the experience. Key aspects of sound quality include both objective technical metrics and subjective perceptual dimensions. Objective measures quantify signal integrity through parameters like total harmonic distortion (THD), which assesses nonlinear distortions introducing unwanted harmonics; signal-to-noise ratio (SNR), indicating the level of background noise relative to the desired signal; and frequency response, evaluating how evenly an audio system reproduces sounds across the audible spectrum (typically 20 Hz to 20 kHz). These metrics provide a foundation for engineering assessments but do not fully capture human judgment. Subjectively, sound quality is multidimensional, encompassing attributes such as clarity (distinctness of elements), spaciousness (sense of width and depth), distortion (perceived impurities), harshness (unpleasant sharpness), and balance in , , and strengths, as identified in perceptual studies of music reproduction. In practice, sound quality assessment combines these approaches via standardized methods, including psychoacoustic models like , , and , which predict subjective responses to complex sounds, and advanced tools such as for objective-perceptual correlation in audio systems. These evaluations are essential across applications, from consumer audio devices and automotive to live event acoustics, where optimizing quality enhances user satisfaction and contextual appropriateness.

Fundamentals

Definition and Scope

Sound quality is the subjective perceptual evaluation of a sound's characteristics in relation to listener expectations and the intended context of use, including aspects such as clarity, spatial imaging, and the perceived presence of distortions or . This concept encompasses both measures, which quantify physical properties of the signal through standardized metrics, and subjective perceptions, which depend on human auditory responses and contextual factors. The distinction between these approaches is foundational, as evaluations provide repeatable data while subjective ones capture listener preferences and experiences. The historical roots of sound quality trace back to 19th-century acoustics, where Hermann von Helmholtz's seminal work on timbre in On the Sensations of Tone (1863) linked perceptual qualities of sound to their harmonic structure, laying groundwork for understanding audio fidelity beyond mere pitch and loudness. In the 20th century, concepts evolved with advancements in recording and broadcasting; electrical recording in the 1920s improved fidelity over acoustic methods, while standardization efforts by organizations like the Audio Engineering Society in the mid-century established benchmarks for broadcast audio, such as those for AM and FM transmission, to ensure consistent reproduction across systems. These developments shifted focus from basic sound capture to optimized perceptual reproduction in mass media. Sound quality intersects multiple disciplines, including audio engineering for , psychoacoustics for modeling human hearing, human-computer interaction for interface design in , and environmental acoustics for assessing ambient soundscapes. For instance, in , it emphasizes speech intelligibility to ensure clear communication amid , prioritizing bandwidth efficiency for voice transmission. In contrast, hi-fi systems in consumer audio prioritize , aiming to preserve and spatial imaging for immersive listening experiences. This broad scope highlights sound quality's role in applications ranging from technical transmission to perceptual enhancement.

Perceptual Psychology

The perceptual psychology of sound quality examines how the human auditory system processes acoustic signals to form judgments of auditory pleasantness, clarity, and realism, rooted in psychoacoustic principles that bridge physical sound properties and subjective experience. Human hearing typically spans frequencies from 20 Hz to 20 kHz, though sensitivity varies with age and intensity, influencing overall sound quality perception by determining the audible spectrum for musical harmonics, speech formants, and environmental cues. Just noticeable differences (JNDs) represent the smallest detectable changes in sound attributes, serving as thresholds for perceived alterations in quality; for pitch, the JND is approximately 0.3% of the base frequency (e.g., about 1 Hz at 500 Hz), for loudness around 0.5–1 dB, and for timbre roughly 5–10% variation in spectral envelope, highlighting how subtle deviations can degrade perceived fidelity. Masking effects play a central role in psychoacoustic foundations, where one sound obscures another's perception, affecting quality by altering effective signal detectability. Simultaneous masking occurs when a louder raises the detection of a nearby frequency occurring at the same time, often within the same , reducing perceived detail in complex audio like . Temporal masking, conversely, involves a preceding or following sound influencing sensitivity, such as post-masking where a brief loud burst elevates thresholds for subsequent faint signals up to 200 ms later, which can mask transients critical to rhythmic clarity. These effects, quantified through psychophysical experiments, explain why noise floors or compression artifacts diminish quality in perceptual terms. Auditory models formalize these processes, with the Fletcher-Munson curves—now refined as ISO equal-loudness contours—illustrating how perceived loudness varies nonlinearly across frequencies at different levels; for instance, low frequencies require higher intensities (up to 20–30 more) to match the loudness of midrange tones at 40 , guiding balanced in audio reproduction to maintain quality. Critical bands, proposed by Zwicker, divide the audible spectrum into 24 frequency regions (each about 100–300 Hz wide, increasing with ) where energy integration occurs, akin to auditory banks that process information for and masking analysis, enabling models to predict perceived or roughness in sounds. These models underpin quality assessment by simulating cochlear mechanics and . Cognitive biases further shape sound quality judgments beyond raw sensory input, introducing subjective overlays that influence preferences. Listeners often favor the "warmth" associated with analog sound—perceived as fuller due to subtle even-order distortions—over digital's "sterile" precision, a rooted in familiarity and expectation rather than measurable superiority, as tests reveal minimal inherent differences when artifacts are controlled. effects exacerbate this, where an initial positive impression (e.g., from visual equipment or ) positively skews ratings of unrelated attributes like spatial or tonal , leading to overestimation of overall quality in non- evaluations. Such biases underscore the interplay between and higher-order in audio appraisal. Binaural hearing enhances spatial quality perception by exploiting interaural time differences (ITDs up to 700 μs) and level differences (ILDs up to 20 ), allowing localization accuracy within 1–2° azimuthally, which contributes to immersive and source separation in stereophonic reproduction. This dichotic processing not only aids in rejection but elevates perceived , as disruptions in cues (e.g., from playback) reduce spatial coherence and overall quality ratings.

Technical Factors

Signal Fidelity Metrics

Signal fidelity metrics quantify the accuracy with which an audio system reproduces the original signal's characteristics, focusing on parameters that ensure faithful representation without alteration beyond inherent system limits. These metrics provide a theoretical foundation for assessing how well a system maintains the signal's temporal, spectral, and amplitude properties, essential for high-quality audio reproduction. Key among them are , , and , alongside foundational standards like the Nyquist theorem for sampling. Frequency response describes how evenly an audio system reproduces signals across the audible spectrum, typically defined by its and characteristics. For full-range audio, the bandwidth requirement aligns with the human of approximately 20 Hz to 20 kHz, ensuring all perceptible frequencies are captured without significant . refers to the gradual decrease in gain at the band's edges, often specified at the -3 dB points, which can impact tonal balance by emphasizing or diminishing certain frequency components—such as excessive low-frequency leading to a thinner sound or high-frequency resulting in muffled highs. Dynamic range measures a system's capacity to handle the full span of amplitude variations in the signal, from the quietest discernible levels to the loudest peaks without compression or clipping. It is calculated as the ratio of the maximum signal amplitude to the minimum detectable signal, expressed in decibels using the formula: \text{Dynamic range (dB)} = 20 \log_{10} \left( \frac{\max \text{ signal}}{\min \text{ signal}} \right) This metric establishes the peak-to-peak amplitude handling capability, where higher values indicate greater fidelity in preserving subtle details alongside intense transients. Linearity assesses the system's consistent response to input variations, encompassing both and domains. linearity ensures that output scales proportionally with input across the frequency range, maintaining uniform without deviation that could alter perceived volume or . linearity requires a constant group delay, where all frequencies experience equal time shift, preserving shape and transient accuracy critical for natural reproduction. A key metric for evaluating linearity is distortion (IMD), which quantifies nonlinear interactions between multiple frequencies, producing spurious sum and difference tones that degrade signal purity; low IMD levels confirm robust linearity. The Nyquist theorem provides a fundamental standard for sampling audio signals, stating that to accurately reconstruct a continuous signal, the sampling rate must be at least twice the highest frequency component in the . For audio limited to 20 kHz, this implies a minimum sampling rate of 40 kHz, preventing and ensuring theoretical in digital representation.

Noise and

Noise in audio systems refers to any unwanted random electrical or acoustic signals that degrade the clarity of the intended sound. Common types include , which has equal across all frequencies, resulting in a flat spectrum that sounds like persistent hiss; , characterized by equal power per and a -3 dB per octave roll-off, often used in testing due to its similarity to ; and thermal noise, also known as Johnson-Nyquist noise, arising from the random thermal agitation of charge carriers in conductors and resistors, which is inherently white and unavoidable in electronic components. The impact of noise is quantified by the signal-to-noise ratio (SNR), which measures the ratio of the desired signal power to the background noise power, typically expressed in decibels for audio applications. The formula for SNR in voltage terms, common in audio engineering, is: \text{SNR} = 20 \log_{10} \left( \frac{\text{RMS value of signal}}{\text{RMS value of noise}} \right) Higher SNR values indicate better quality, with professional audio systems often targeting above 90 dB to ensure imperceptible noise. Noise sources can be environmental, such as 50/60 Hz hum induced by electromagnetic interference from power lines, or electronic, like crosstalk in amplifiers where signals from one channel leak into another, causing channel separation degradation and perceived smearing of stereo imaging. Distortion, unlike , involves predictable alterations to the signal , often nonlinear, that introduce unwanted components and reduce . Harmonic distortion occurs when the output contains integer multiples (s) of the input , quantified by (THD), calculated as the ratio of the root-sum-square of harmonic powers to the fundamental power. The is: \text{THD} = \frac{ \sqrt{ \sum_{h=2}^{N} P_h } }{ P_1 } \times 100\% where P_h is the power of the h-th and P_1 is the fundamental power. Nonlinear effects like clipping arise when the signal exceeds the system's , producing abrupt truncation and high odd-order harmonics that sound harsh. For high-fidelity audio, acceptable THD levels are typically below 0.1% across the audible band to avoid audible coloration. To mitigate quantization noise in digital-to-analog conversion, dithering introduces low-level noise to linearize the process, while noise shaping redistributes this noise to higher frequencies outside human hearing sensitivity, effectively improving perceived dynamic range without altering the core signal.

Analog and Digital Aspects

Analog Sound Characteristics

Analog sound systems represent audio signals as continuous waveforms, faithfully capturing the natural variations in amplitude and frequency of the original sound without discretization. This continuous representation occurs through physical media such as vinyl records, where a stylus traces helical grooves modulated by the audio signal, and magnetic tape, where varying magnetic fields on an oxide-coated substrate encode the waveform. In vinyl playback, repeated stylus contact causes groove wear through friction, progressively degrading the signal by flattening the groove walls and introducing distortion, particularly in high-frequency content. Similarly, magnetic tape experiences gradual signal loss over time due to oxide particle shedding and magnetization decay, leading to reduced fidelity with each playback or storage period. A distinctive trait of analog systems is the perceived "warmth" often attributed to even-order harmonic distortion generated by amplifiers, which produce musically sympathetic that enhance tonal richness without harshness. These even-order harmonics, primarily the second and fourth, arise from the nonlinear response of under signal load, adding subtle and depth that many listeners associate with organic sound character. In contrast to systems' precise but sometimes sterile , this analog warmth contributes to a more immersive listening experience in applications like playback and recording. Analog systems are prone to several inherent limitations that affect sound quality. Wow and refer to low-frequency () and high-frequency () speed variations in playback mechanisms, such as turntable platters or tape transport reels, causing audible pitch instability—wow manifests as slow wobbling, while adds a tremolo-like . Hiss in arises from in the particles and signal residue, becoming more prominent during quiet passages or when tape saturation compresses dynamic peaks, limiting the . To mitigate groove overload and in , the curve is applied during recording, attenuating low frequencies by up to 20 dB and boosting highs, with inverse compensation during playback to restore flat response. Historically, the introduction of the in 1948 by revolutionized analog audio, enabling up to 30 minutes of playback per side at 33⅓ RPM on 12-inch microgroove , compared to the prior 78 RPM discs' 3-5 minutes. This format expanded to approximately 70 , allowing greater musical expression through reduced surface noise and finer grooves, though still constrained by analog media's physical limits like wear and speed inconsistencies.

Digital Audio Representation

Digital audio representation involves converting continuous analog sound waves into discrete numerical values through two primary processes: sampling and quantization. This discretization enables the storage, transmission, and manipulation of audio in digital systems, preserving the essential characteristics of the original signal within the limits of the chosen parameters. The sampling process captures the of the at regular intervals, determined by the sampling rate. According to the Nyquist-Shannon sampling theorem, to accurately reconstruct the original signal without loss of information, the sampling rate must be at least twice the highest frequency component in the signal; for human hearing, which extends up to approximately 20 kHz, a minimum rate of 40 kHz is required. Failure to adhere to this can result in , where higher frequencies masquerade as lower ones, distorting the audio; this is mitigated by applying an —a —prior to sampling to attenuate frequencies above half the sampling rate. Following sampling, quantization assigns each sample amplitude to the nearest level from a , defined by the . A of n bits provides 2n possible levels, with the spacing between levels introducing quantization noise, which limits the signal's . The theoretical (SNR) for uniform quantization is given by: \text{SNR} = 6.02n + 1.76 \, \text{dB} For example, a 16-bit depth yields approximately 96 of , sufficient for most consumer applications as it exceeds the human ear's to variations. The most common format for representing these quantized samples is (PCM), an uncompressed method that stores each sample as a value, typically in linear fashion for straightforward processing. To optimize storage and bandwidth, audio files often employ compression: lossless formats like reduce file size by up to 50-70% through and entropy encoding without discarding any data, ensuring bit-perfect reconstruction. In contrast, lossy formats such as achieve higher compression ratios (often 10:1 or more) by leveraging perceptual coding, which analyzes the psychoacoustic model of human hearing to remove or quantize less audible components, such as those masked by louder sounds. A foundational standard for digital audio is the Compact Disc (CD) format, established through collaboration between and , which specifies stereo PCM at a 44.1 kHz sampling rate and 16-bit depth, allowing about 74-80 minutes of playback on a 120 mm disc while capturing frequencies up to 20 kHz with a 96 dB . Subsequent advancements have led to , described by the as providing extended resolution in bandwidth, , time, and spatial acuity beyond CD specifications. These commonly include 24-bit depth for over 144 dB and sampling rates of 96 kHz or higher, enabling greater in professional recording and playback systems. However, the audible benefits of over CD quality remain a subject of debate, with some studies indicating subtle perceptual differences under controlled conditions while others find them indistinguishable for most listeners.

Evaluation Methods

Objective Measurements

Objective measurements in sound quality involve the use of precise and standardized procedures to quantify audio performance without relying on human , enabling repeatable and comparable results across systems. These methods assess parameters such as , distortion levels, and noise floors by analyzing electrical or acoustic signals in controlled environments. Key tools include spectrum analyzers, which decompose signals into frequency components to evaluate harmonic distortion and noise spectra; oscilloscopes, which visualize time-domain waveforms to detect clipping or transient anomalies; and specialized audio precision analyzers, such as those from Audio Precision, that simultaneously measure (SNR) and (THD) with high accuracy, often achieving resolutions below -120 dB for professional applications. Test signals are to these evaluations, providing controlled inputs to isolate specific audio characteristics. Pure s at various frequencies are commonly used to measure , as they reveal and products when analyzed; for instance, a 1 kHz can quantify THD by comparing output harmonics to the . Swept s, which vary frequency over time, assess and reveal resonances or roll-offs in systems like loudspeakers. For loudness normalization, the BS.1770 standard employs integrated metering with test signals that simulate program material, calculating perceived in loudness units relative to () to ensure consistent playback across broadcasts. Standardized procedures enhance the reliability of these measurements, with organizations like the (AES) providing guidelines for test setups, including input levels, bandwidth limits, and environmental controls to minimize variables. , a frequency-dependent standardized in IEC 61672, is applied to noise measurements to approximate human hearing sensitivity, weighting mid-frequencies more heavily and thus providing a perceived noise level in dBA that correlates with objective hiss or hum in audio equipment. The development of noise reduction in the late 1960s by introduced pre-emphasis and techniques that boosted by up to 20 dB, fundamentally impacting measurement practices by necessitating specialized decoders and analyzers to verify expansion accuracy and residual noise, as outlined in early Dolby technical manuals.

Subjective Assessments

Subjective assessments of sound quality rely on human listeners to evaluate perceived audio characteristics, incorporating elements of such as masking thresholds that influence detectability of s. Common test types include the double-blind triple-stimulus method, akin to ABX testing, where listeners compare a reference signal (A), a test signal (B), and an unknown (X) to detect small differences, using a five-grade from imperceptible (5.0) to very annoying (1.0). For intermediate quality evaluations, the (Multiple Stimuli with Hidden Reference and Anchor) method presents several stimuli simultaneously, including a hidden reference and low-quality anchors (e.g., low-pass filtered signals), with listeners rating each on a 0-100 continuous quality divided into categories like excellent to bad. Preference scaling, often using 1-5 ratings, allows direct comparison of audio variants to gauge overall appeal. Protocols for these tests emphasize standardized conditions to ensure reliability, as outlined in ITU-R Recommendation BS.1116, which specifies listening environments, equipment, and procedures for assessing small impairments. Trained listeners, typically experts with prior experience in analytic listening, are preferred for their consistency and ability to detect subtle artifacts, requiring at least 10 participants; non-expert or naive listeners (minimum 20) may suffice for broader population representation but often need training to align with expert judgments. Training sessions, lasting up to three hours, familiarize participants with test signals, grading scales, and equipment to minimize bias and enhance repeatability. Influencing factors include room acoustics, which must meet strict criteria such as a time of approximately 0.25 seconds (adjusted for room volume) between 200 Hz and 4 kHz, early reflection attenuation of at least 10 within 15 , and no higher than NR-10 to prevent perceptions. , which can degrade judgment accuracy, is mitigated by limiting sessions to 20-30 minutes with rest periods equal to or longer than the session duration, and capping trials at 10-15 per sitting. Perceptual models like , defined in BS.1387, simulate subjective tests by computationally modeling human auditory perception to predict quality degradation, outputting an Objective Difference Grade (ODG) that correlates with subjective ratings without requiring live listeners. This approach uses psychoacoustic principles, such as excitation patterns and models, to evaluate codecs and distortions, validated against databases of human assessments for applications in development and monitoring.

Applications and Enhancements

In Recording and Playback Systems

In professional recording chains, the serves as the initial capture device, where its is critical for accurate sound reproduction. condenser microphones aim for a flat across the audible spectrum (typically 20 Hz to 20 kHz) with minimal coloration, measured according to standards like IEC 60268-4 for equipment. Following the , the amplifies the signal while introducing minimal ; high-quality preamps achieve an equivalent input (EIN) of -128 to -130 A-weighted (150 Ω source). During multitrack mixing, maintaining headroom—typically 6 to 12 dB below 0 on the master bus—prevents inter-sample clipping and preserves , allowing subsequent processing without introducing that could degrade overall fidelity. In playback systems, speaker drivers are optimized through crossover networks to ensure a flat overall , directing low frequencies to woofers, mids to drivers, and highs to tweeters within their efficient bandwidths. These networks, often employing Linkwitz-Riley filters of second or fourth order, maintain phase coherence and magnitude flatness within ±1 across the crossover region to avoid peaks or dips that alter tonal balance. For , impedance matching between the amplifier's output (ideally <1/8 of the ) and the driver load is essential to preserve the intended ; mismatches above this ratio can cause roll-off or emphasis in low-impedance designs (e.g., 32 Ω), reducing accuracy in and . System integration from recording to playback involves managing end-to-end fidelity losses, where cumulative noise and distortion from multiple stages—such as analog-to-digital conversion, transmission, and output—can reduce signal-to-noise ratio by 10-20 dB if not controlled, though digital chains remain lossless without re-quantization. For instance, vinyl mastering limits dynamic range to 55-70 dB due to groove constraints and surface noise, often requiring more conservative compression compared to digital streaming formats that support 90-96 dB, resulting in greater perceived punch in streamed audio but potential warmth from vinyl's analog imperfections. Studio monitors are standardized to a reference level of 85 dB SPL at 1 meter using pink noise at -20 dBFS, providing a consistent calibration point for mix translation across environments as per AES practices. In wireless playback, Bluetooth codecs like aptX (up to 352 kbps, 16-bit/48 kHz) deliver perceptually superior quality to the mandatory SBC (max 328 kbps, 16-bit/44.1 kHz) by reducing compression artifacts and latency to under 40 ms, enabling clearer highs and tighter bass in mobile systems; more recent developments include LE Audio with the LC3 codec, offering improved efficiency and quality in low-latency scenarios as of 2025.

Quality Improvement Techniques

Noise reduction techniques have been pivotal in enhancing sound quality by mitigating unwanted background interference, particularly in analog and systems. The Dolby A system, introduced in 1966 by Laboratories, employs a four-band compressor-expander architecture that boosts low-level signals during recording and restores them during playback, reducing tape hiss in professional recording environments. B, developed for consumer applications in 1968, uses a single high-frequency band to suppress tape hiss, pre-emphasizing soft high frequencies to overpower noise without significantly altering louder signals, thus improving clarity in compact cassette decks. Building on this, C, launched in the early 1980s, applies multi-band processing with spectral skewing and double expansion for enhanced and greater in consumer tape systems while reducing sensitivity to playback mismatches. In digital domains, spectral subtraction serves as a foundational noise reduction method, estimating the noise spectrum from speech pauses and subtracting it from the noisy signal's magnitude spectrum to recover cleaner audio. This technique, pioneered in a 1979 IEEE paper, effectively suppresses broadband acoustic noise in speech signals, enhancing intelligibility by targeting the power spectral density without altering the phase, though it may introduce minor musical noise artifacts in low-signal-to-noise scenarios. Equalization techniques further refine sound quality by compensating for acoustic anomalies. Parametric equalization (PEQ) for room correction involves designing filters that adjust the in-room of loudspeakers to achieve perceptual flatness, tuning parameters such as , peak gain, and () through optimization algorithms like least-squares nonlinear fitting. For instance, a 12-band PEQ can smooth peaks and dips measured via responses, reducing room-induced distortions and improving . Dynamic range compression enhances perceived and consistency, with multiband variants dividing the audio spectrum into 3-5 frequency bands via crossovers and applying independent compression to each, preventing while preserving overall . In mixing and mastering, multiband compressors target issues like sibilance in the 5-8 kHz range or low-frequency pumping in below 120 Hz, allowing precise that boosts average levels by 3-6 without clipping. Advanced digital processing includes in digital-to-analog converters (DACs), which interpolates lower-rate signals (e.g., 44.1 kHz) to higher rates using slow filters, generating ultrasonic images that act as to average out DAC non-linearities and reduce . This approach improves , yielding clearer highs and reduced time-domain smearing compared to sharp filters. Post-2020 developments in AI-based leverage neural networks, such as models, to reverse degradations like or compression artifacts; for example, conditional frameworks like CDiffuSE (2021) and (2023) train deep networks on denoising tasks to iteratively restore speech and music by modeling probabilistic clean signal generation. High-definition formats have also driven quality improvements. (SACD), introduced by in 1999, employs (DSD) encoding at 2.8224 MHz with 1-bit , enabling exceeding 120 dB and up to 100 kHz for more analog-like fidelity in playback systems. Similarly, , launched in 2012, introduces object-based spatial audio with height channels and dynamic rendering, supporting up to 128 audio objects to create immersive soundscapes that enhance directional accuracy and envelopment in and home environments. These techniques collectively address sources like tape hiss or room modes, elevating overall perceptual fidelity.

References

  1. [1]
    Objective perceptual audio quality measurement methods | NHK STRL
    As an objective quality measure for audio signals, metrics such as the SN (signal to noise) ratio and total harmonic distortion have been used traditionally.
  2. [2]
  3. [3]
    [PDF] General methods for the subjective assessment of sound quality - ITU
    It includes, but is not restricted to, such things as timbre, transparency, stereophonic imaging, spatial presentation, reverberance, echoes, harmonic ...
  4. [4]
    AES Journal Forum » Correlation Between Subjective and Objective ...
    Aug 1, 1974 · The aim of the research reported here was to disclose, if possible, connections between a subjective evaluation of quality loudspeakers and ...
  5. [5]
    [PDF] Method for objective measurements of perceived audio quality - ITU
    The objective measurement method described in this Recommendation measures audio quality and outputs a value intended to correspond to perceived audio quality.
  6. [6]
    [PDF] An Exploration of Musical Timbre - Stanford CCRMA
    After demonstrating that the complex periodic vibrations of musical and vocal sounds consist of sets of harmonics, Helmholtz showed that the ear can distinguish ...
  7. [7]
    An Audio Timeline - Audio Engineering Society
    Dedicated to the preservation of over a century of audio history, the Committee is developing a broad-based history of audio engineering and the audio industry.Missing: quality | Show results with:quality
  8. [8]
    [PDF] FROM PSYCHOACOUSTICS TO SOUND QUALITY ENGINEERING
    In this paper psychophysical methods useful for both psychoacoustics and sound quality engineering will be discussed. Models of basic psychoacoustic magnitudes ...
  9. [9]
    Effects of Sound Quality on the Accuracy of Telephone Captions ...
    Dec 14, 2022 · However, little is known about the effects of degraded telephone audio on the intelligibility of ASR captioning. This research note investigates ...
  10. [10]
    Guide to HiFi: What is High Fidelity Audio? - AVIXA
    Aug 8, 2025 · HiFi preserves the original recording's integrity, minimizing distortion, maintaining an even frequency response, and achieving a dynamic range ...Missing: musical | Show results with:musical
  11. [11]
    Extended High Frequency Thresholds in College Students - NIH
    Human hearing is sensitive to sounds from as low as 20 Hz to as high as 20,000 Hz in normal ears. However, clinical tests of human hearing rarely include ...Missing: authoritative | Show results with:authoritative
  12. [12]
    Loudness, Its Definition, Measurement and Calculation
    Author & Article Information. Harvey Fletcher , W. A. Munson. Bell Telephone Laboratories. J. Acoust. Soc. Am. 5, 82–108 (1933). https://doi.org/10.1121 ...Missing: URL | Show results with:URL
  13. [13]
    Subdivision of the Audible Frequency Range into Critical Bands ...
    E. Zwicker; Subdivision of the Audible Frequency Range into Critical Bands (Frequenzgruppen), The Journal of the Acoustical Society of America, Volume 33, ...
  14. [14]
    [PDF] Mitigating the Halo Effect: Managing the Wow Factor in Music ...
    This article focuses on addressing one distinct inter-‐rater effect: the halo effect. The halo effect occurs when impressions of the quality of a performance ...<|separator|>
  15. [15]
    Binaural Signal Processing in Hearing Aids - PMC - PubMed Central
    Sep 24, 2021 · The brain integrates information received from each ear and then translates the differences into a unified perception of a single sound arriving ...
  16. [16]
    [PDF] ISO 16976-7:2023 - iTeh Standards
    4 Range of hearing and speech. Humans with normal hearing can usually hear sound pressure waves in a frequency range of about. 20 Hz to 20 000 Hz, but the ear ...
  17. [17]
    How Speaker Frequency Response Impacts Sound Quality
    Feb 26, 2025 · A loudspeaker's frequency response is a key factor in determining its sound quality and accuracy. This response measures how well a loudspeaker reproduces ...
  18. [18]
    Dynamic Range - an overview | ScienceDirect Topics
    whereas the system dynamic range is. (8.13) System dynamic range = 20 log 10 Largest permissible signal voltage Smallest detected signal voltage. In system ...
  19. [19]
  20. [20]
    FAQ | What is linear phase and why should anyone care about it?
    Jun 4, 2024 · Linear phase means all frequencies are delayed by the same amount, preserving the time relationship between fundamental and harmonics.
  21. [21]
    What Is Intermodulation Distortion - An Engineers Guide - Keysight
    Even small nonlinear effects can produce distortion that grows with signal power and directly impacts system linearity. It is useful to distinguish between ...
  22. [22]
    [PDF] Communication In The Presence Of Noise - Proceedings of the IEEE
    Using this representation, a number of results in communication theory are deduced concern- ing expansion and compression of bandwidth and the threshold effect.
  23. [23]
    Managing Noise in the Signal Chain, Part 1 - Analog Devices
    Aug 7, 2014 · The three sources of white noise in semiconductor devices are thermal, shot, and avalanche noise. Thermal Noise. Thermal noise, also called ...
  24. [24]
    Noise in Audio Amplifiers - Elliott Sound Products
    It is also known as Johnson noise, named after the man who discovered the phenomenon in 1928. ... Thermal noise is random, so two 1V noise voltages sum to 1.414V ...
  25. [25]
    Signal to Noise Ratio Calculator - Apex Waves
    SNR is calculated using the formula: SNR (dB) = 10 * log10(P_signal / P_noise) ... In audio engineering, for example, a higher SNR indicates clearer sound.
  26. [26]
  27. [27]
    [PDF] Reducing Crosstalk in Directpath Headphone Amplifiers
    Large variations in sensing this potential cause reduced performance of the amplifier, specifically noise, distortion and crosstalk. Some amplifiers are ...
  28. [28]
    Understanding, Calculating, and Measuring Total Harmonic ...
    Feb 20, 2017 · Total harmonic distortion (THD) is a measurement that tells you how much of the distortion of a voltage or current is due to harmonics in the signal.
  29. [29]
    What is total harmonic distortion (THD)? - Audiophile ON
    Aug 16, 2025 · THD is calculated by comparing the sum of the powers of all harmonic frequencies above the fundamental frequency to the power of the fundamental ...
  30. [30]
    Does Dithering Matter In Audio Production?
    Sep 27, 2024 · Noise shaping enhances the dithering process by focusing the noise in frequency bands where it is less likely to be perceived. Applications ...
  31. [31]
    Preservation Self-Assessment Program (PSAP) | Phonograph Record
    Deterioration: All grooved disc media is susceptible to warpage, breakage, groove wear, and surface contamination. Types of surface contamination include ...
  32. [32]
    Analogue Warmth - Sound On Sound
    Even‑order harmonic distortion tends to sound musically sympathetic, smooth, and bright in a constructive way. Many simple valve‑based circuits (including most ...
  33. [33]
    Wow and Flutter | AVAA - AV Artifact Atlas
    Wow manifests itself by slow pitch variation resulting from small speed variations; flutter refers to rapid speed variations.
  34. [34]
    Noise Reduction and Tape Hiss - HyperPhysics Concepts
    A persistent random noise signal from the residual magnetization of the oxide granules limits the fidelity of magnetic tape recording. Because of the small size ...
  35. [35]
  36. [36]
    Inside the Archival Box: The First Long-Playing Disc | Now See Hear!
    Apr 13, 2019 · Columbia Records released the first long-playing microgroove record, spinning at 33 1/3 revolutions per minute and holding about 23 minutes each ...
  37. [37]
    The history of the LP - Hi-Fi+
    Dec 20, 2023 · The LP has a dynamic range of 65dB (or possibly 70dB, depending on which source you trust) and bandwidth that's said to cover 7Hz to 50kHz.
  38. [38]
    PCM, Pulse Code Modulated Audio - The Library of Congress
    Apr 26, 2024 · Pulse code modulation was originally developed in 1939 as a method for transmitting digital signals over analog communications channels. The ...
  39. [39]
  40. [40]
    FLAC - What is FLAC? - Xiph.org
    FLAC stands for Free Lossless Audio Codec, an audio format similar to MP3, but lossless, meaning that audio is compressed in FLAC without any loss in quality.Features · Downloads · Changelog · Using FLACMissing: specification | Show results with:specification
  41. [41]
    [PDF] MP3 and AAC Explained
    Perceptual encod- ing is a lossy compression technique, i.e. the decoded file is not a bit-exact replica of the original digital audio data. Perceptual coders ...
  42. [42]
    [PDF] Communications - Philips
    Mar 6, 2009 · Furthermore, the following impor- tant parameters were fixed during the. Philips-Sony meetings: a 44.1 kHz sam- pling frequency, a 16-bit ...
  43. [43]
    AES Technical Committee: High resolution audio
    High resolution audio refers to extended resolution in bandwidth, dynamic range, time, and spatial acuity. The committee focuses on technology, signal ...
  44. [44]
  45. [45]
  46. [46]
  47. [47]
    Frequency Range | SCHOEPS Microphones
    The standard IEC 60268-4 ("Sound system equipment - Part 4 ... microphone is the frequency response, from which ultimately the frequency range is derived.
  48. [48]
    Understanding Microphone Preamplifier Noise - Sound Devices
    Jun 27, 2024 · This number is properly measured using 150 ohms as an input terminator. The very best EIN that can be achieved is -133 dBV, since this is noise ...Missing: standard | Show results with:standard
  49. [49]
  50. [50]
    Does vinyl sound better than streaming? - SoundGuys
    Nov 4, 2024 · Dynamic range. Digital files allow over 90dB of difference between the loudest and softest sounds, compared to vinyl's 70dB dynamic range.
  51. [51]
    [PDF] An Integrated Approach to Metering, Monitoring and Levelling
    The top 50% of the physical scale is devoted to only the top 6 dB of dynamic range, and the meter's useable dynamic range is only about 13 dB. Not realizing ...
  52. [52]
    SBC vs aptX: Which Bluetooth Codec Is The Best? - RTINGS.com
    Aug 15, 2017 · SBC is default, aptX is optional with better encoding. SBC has higher latency, aptX-LL reduces it. SBC audio quality is average, aptX is ...
  53. [53]
    [PDF] A CENTURY OF INNOVATION AN ABRIDGED TIMELINE OF THE ...
    Dolby Laboratories. (Dolby) is founded in. London by Ray Dolby. (above). One year later, the company introduces. A-type noise reduction for music recording.Missing: function | Show results with:function
  54. [54]
    Q. What is different about the varieties of Dolby noise reduction?
    Dolby A uses four bands, B is single-band, C is multi-band, SR is sophisticated multi-band, S is a halfway house, and HX is not noise reduction.
  55. [55]
    Noise Reduction in Tape Recording - HyperPhysics
    The Dolby systems for noise reduction employ circuitry which pre-emphasizes high frequencies before they are recorded onto tape in order to make them larger.
  56. [56]
    Suppression of acoustic noise in speech using spectral subtraction
    Abstract: A stand-alone noise suppression algorithm is presented for reducing the spectral effects of acoustically added noise in speech.Missing: seminal | Show results with:seminal
  57. [57]
    Automated Design of Audio Filters for Room Equalization - MathWorks
    The variables being tuned by the optimization algorithm are typical audio parametric EQ parameters: Center Frequency, Filter Bandwidth, and Peak Gain. Use ...
  58. [58]
    Recording Magazine Resources: Multiband Dynamics
    Explore multiband compressors. Today, multiband dynamics processing is vitally important in nearly all audio fields, including speakerphone design, ...Eq Problems · Making It Loud · The Loudness Button
  59. [59]
    [PDF] Theory of Upsampled Digital Audio - MLSSA
    Upsampling DACs use slow roll-off filters, which reduce time smearing, leading to improved sound quality compared to sharp filters.
  60. [60]
    Diffusion Models for Audio Restoration Invited paper for the SPM ...
    To address this problem, audio restoration methods aim to recover clean sound signals from the corrupted input data. We present here audio restoration ...
  61. [61]
    Sony Launches the First Super Audio CD Player
    Apr 6, 1999 · Sony Marketing (Japan) Inc. today announced plans to launch the world's first Super Audio CD player [SCD-1] in the Japanese market in May 1999.
  62. [62]
    [PDF] Dolby Atmos Specifications
    Since launching in 2012, Dolby Atmos has revolutionized how moviegoers experience entertainment and was honored with a. Scientific and Engineering Award from ...