Fact-checked by Grok 2 weeks ago

Nyquist–Shannon sampling theorem

The Nyquist–Shannon sampling theorem, also known as the cardinal theorem of sampling or the sampling theorem, is a foundational result in that establishes the conditions under which a continuous-time bandlimited signal can be perfectly reconstructed from a sequence of its samples. Specifically, it states that if a signal contains no frequencies higher than a maximum value W cycles per second, it is completely determined by its values at points spaced $1/(2W) seconds apart, provided the sampling rate exceeds twice the to prevent . The theorem originated with Harry Nyquist's 1928 analysis of telegraph transmission, where he determined that the necessary frequency range for distortionless signaling does not exceed the signaling speed, implying that a equal to half the sampling rate suffices for complete signal representation through independent sinusoidal components. Nyquist's work, published as "Certain Topics in Telegraph Transmission Theory" in the Transactions of the , laid the groundwork by linking the number of signal elements to the required frequency components, approximately half the signaling rate. This insight was later formalized and proven rigorously by in his 1949 paper "Communication in the Presence of Noise," published in the Proceedings of the IRE, which integrated it into the broader framework of and demonstrated exact via sinc for bandlimited functions. Central to the theorem is the concept of the , defined as twice the highest frequency in the signal, which serves as the critical threshold for faithful . Sampling below this rate introduces , where higher frequencies masquerade as lower ones, distorting the signal irreversibly without prior . Reconstruction from samples is achieved through the , which sums scaled sinc functions centered at each sample point, ensuring no information loss for signals strictly bandlimited to below half the sampling frequency. The theorem's implications extend across , enabling efficient analog-to-digital and digital-to-analog conversions in applications such as audio recording (e.g., compact discs sampling at 44.1 kHz to capture up to 22.05 kHz), , medical imaging, and systems. It underpins modern technologies by guaranteeing that discrete representations preserve continuous signal fidelity, provided filters limit the input appropriately. Extensions of the theorem address non-uniform sampling, multidimensional signals, and noisy environments, but the core principle remains essential for avoiding information loss in sampled systems.

Fundamentals

Statement of the Theorem

The Nyquist–Shannon sampling theorem, also known as the cardinal theorem of sampling or Whittaker–Shannon–Kotelnikov–Nyquist theorem, addresses the conditions under which a continuous-time signal can be perfectly reconstructed from its discrete samples. A bandlimited signal is defined as one whose Fourier transform X(\omega) is zero for all frequencies |\omega| > 2\pi B, where B is the bandwidth in hertz, meaning the signal contains no energy outside the frequency range [-B, B]. The theorem states that if a continuous-time signal x(t) is bandlimited to frequency B, then it can be completely reconstructed from its uniformly spaced samples x(nT), taken at a sampling rate f_s = 1/T \geq 2B samples per second, without loss of information. The minimum such sampling rate of 2B samples per second is known as the , while the is B Hz (half the sampling rate). The theorem was first implied in the context of telegraph transmission by in 1928, who showed that a of bandwidth W can transmit up to $2W independent pulses per second without , and rigorously proved by in 1949 for arbitrary bandlimited functions. Perfect reconstruction under the theorem requires several conditions: uniform sampling at intervals T \leq 1/(2B), an infinite number of samples extending over all time, absence of or other distortions in the sampling , and, in some formulations, a strict f_s > 2B to ensure no overlap at the boundary frequencies. Violation of the sampling rate condition leads to , where higher frequencies masquerade as lower ones. The reconstruction is achieved via ideal sinc interpolation, expressed by the formula: x(t) = \sum_{n=-\infty}^{\infty} x\left(\frac{n}{f_s}\right) \operatorname{sinc}\left(f_s \left(t - \frac{n}{f_s}\right)\right), where \operatorname{sinc}(u) = \sin(\pi u)/(\pi u) is the normalized . This formula weights each sample by a shifted , whose frequency response acts as an ideal to recover the original bandlimited signal. The factor of two in the arises because sampling at f_s produces periodic replicas of the signal's spaced by f_s; a rate of at least $2B ensures these replicas do not overlap, allowing the to be isolated via .

Aliasing Phenomenon

Aliasing refers to the distortion in a sampled signal where higher-frequency components are incorrectly represented as lower frequencies, resulting from the periodic replication of the signal's during the sampling process. This phenomenon arises because sampling a continuous-time signal at a rate f_s creates multiple copies of the original centered at multiples of f_s, and if f_s is less than twice the signal's B, these replicas overlap, causing frequencies above the to "fold" into the . Such overlap prevents the unique identification and perfect reconstruction of the original signal, directly contravening the Nyquist–Shannon sampling theorem's requirement for distortion-free recovery. In the frequency domain, the spectrum of the sampled signal X_s(f) is obtained by multiplying the original spectrum X(f) by a Dirac comb, which generates replicas at intervals of the sampling frequency f_s = 1/T, where T is the sampling period. This results in the key relation: X_s(f) = \frac{1}{T} \sum_{k=-\infty}^{\infty} X\left(f - \frac{k}{T}\right) If the replicas do not overlap—ensured when f_s > 2B—the baseband spectrum remains undistorted and can be isolated for reconstruction; otherwise, the superposition corrupts the low-frequency content with contributions from higher frequencies. A classic example of aliasing is the observed in motion pictures or video, where the spokes of a rotating appear to rotate backwards or stand still due to the sampling being insufficient relative to the 's , causing the perceived motion to alias into the opposite direction. In audio processing, aliasing can manifest as unintended low-frequency tones when high-frequency sounds are sampled inadequately, similar to moiré patterns in where fine details create false fringes. To detect aliasing, one can analyze the for unexpected low-frequency artifacts or use diagnostic signals exceeding half the sampling rate; prevention typically involves applying low-pass filters before sampling to attenuate frequencies above f_s/2, ensuring no overlap in the replicated spectra.

Nyquist Frequency

The , denoted as f_n, is defined as half the sampling frequency f_s, such that f_n = \frac{f_s}{2}. This represents the highest frequency component in a continuous-time signal that can be accurately captured and reconstructed from its discrete samples without distortion, assuming the signal is bandlimited. In the context of the Nyquist–Shannon sampling theorem, the sampling frequency must satisfy f_s > 2B, where B is the of the signal (its maximum content), implying f_n \geq B. Theoretical reconstruction is possible when equality holds (f_n = B) for strictly bandlimited signals, but this requires an ideal , which is unattainable in practice and leads to instability due to filter imperfections and . Consequently, real-world systems typically employ f_s slightly greater than $2B to provide a for practical filter . In digital signal processing systems, the Nyquist frequency establishes the upper limit of the for the sampled signal; any frequency components exceeding f_n will fold back into the (0 to f_n) as , distorting the representation. filters are thus essential prior to sampling to attenuate frequencies above f_n, ensuring the input signal adheres to the bandlimit. A prominent example is the (CD) audio format, which uses a sampling rate of 44.1 kHz, yielding a Nyquist frequency of 22.05 kHz. This exceeds the typical human up to 20 kHz, providing a margin to accommodate filter transition bands while preventing of ultrasonic components. The Nyquist frequency differs from the signal's bandwidth B, which characterizes the inherent frequency extent of the original continuous signal, whereas f_n is a property of the sampling process that must be at least as large as B to enable faithful digitization.

Mathematical Derivations

Proof via Poisson Summation

The Poisson summation formula serves as a powerful tool for deriving the Nyquist–Shannon sampling theorem by linking the discrete samples of a signal in the time domain to periodic replicas of its Fourier transform in the frequency domain. For a square-integrable function x(t) with Fourier transform X(f) = \int_{-\infty}^{\infty} x(t) e^{-i 2\pi f t} \, dt, the Poisson summation formula states that \sum_{n=-\infty}^{\infty} x(nT) = \frac{1}{T} \sum_{k=-\infty}^{\infty} X\left(\frac{k}{T}\right) e^{i 2\pi k t / T}, evaluated at t = 0 for the sampling context, yielding \sum_{n=-\infty}^{\infty} x(nT) = \frac{1}{T} \sum_{k=-\infty}^{\infty} X(k/T). This relation, originally due to Poisson and extended in Fourier analysis, reveals the inherent periodicity introduced by sampling at interval T. To derive the sampling theorem, assume x(t) is bandlimited such that X(f) = 0 for |f| > B. The impulse-sampled signal is then x_s(t) = \sum_{n=-\infty}^{\infty} x(nT) \delta(t - nT), where \delta is the . The Fourier transform of x_s(t) is X_s(f) = \sum_{n=-\infty}^{\infty} x(nT) e^{-i 2\pi f n T}. Applying the in the distributional sense, X_s(f) simplifies to a periodic repetition of the original spectrum: X_s(f) = \frac{1}{T} \sum_{k=-\infty}^{\infty} X\left(f - \frac{k}{T}\right). This expression shows that sampling replicates X(f) at intervals of $1/T in the , scaled by $1/T. If the sampling period satisfies T \leq 1/(2B), or equivalently, the sampling frequency f_s = 1/T \geq 2B, the spectral replicas do not overlap within the |f| < B. Under this condition, the original X(f) can be recovered perfectly from X_s(f) by applying an ideal low-pass filter with cutoff B and gain T, isolating the central replica without distortion. The reconstructed signal is then obtained via the inverse Fourier transform: x(t) = T \int_{-B}^{B} X_s(f) \, e^{i 2\pi f t} \, df, which, given the non-overlapping replicas, equals \int_{-B}^{B} X(f) \, e^{i 2\pi f t} \, df, yielding x(t). This step confirms that the continuous-time signal is uniquely recoverable from its samples. This Fourier-analytic approach highlights the theorem's connection to broader principles in harmonic analysis, demonstrating why uniform sampling suffices for bandlimited signals by exploiting the duality between summation in time and periodicity in frequency. It underscores the role of the sampling rate in preventing spectral folding, though it assumes ideal bandlimiting and infinite-duration signals. In practice, this relates to the discrete-time Fourier transform, where finite samples approximate the infinite sum, but deviations from ideality introduce errors.

Shannon's Information Theory Approach

Claude Shannon provided a foundational proof of the sampling theorem in his 1949 paper "Communication in the Presence of Noise," where he integrated it into the broader framework of to analyze communication systems affected by noise. In this work, Shannon linked the sampling process to the concept of , demonstrating how bandlimited signals can be represented efficiently for transmission over noisy channels. By viewing signals geometrically in a multidimensional space, he showed that sampling at the ensures lossless encoding of the signal's information content without excess redundancy. The core argument posits that a bandlimited signal with bandwidth W (in cycles per second) observed over a time interval T possesses exactly $2WT degrees of freedom, meaning it requires no more than $2W independent samples per second to fully specify its information. This finite dimensionality arises because the signal's frequency content is confined, limiting the number of independent parameters needed for its description. Sampling at a rate f_s = 2W captures these degrees of freedom precisely, allowing exact reconstruction in the absence of noise, while lower rates lead to information loss and higher rates introduce unnecessary samples. Shannon's derivation begins by representing the bandlimited signal in a space of dimension $2WT, where the signal can be expanded using an orthogonal basis of approximately $2WT functions. He proposed sinc functions, specifically \frac{\sin 2\pi W (t - n/(2W))}{\pi (t - n/(2W))} for n = 0, 1, \dots, 2WT - 1, which form an orthogonal set for large T and span the space of bandlimited signals. The coordinates of the signal in this basis correspond to the values at the sampling points spaced $1/(2W) apart, proving that these $2WT samples uniquely determine the signal, as any bandlimited function is a linear combination of this basis with coefficients given by the samples. In the context of noisy channels, Shannon derived the maximum information rate as C \leq W \log_2 (1 + S/N) bits per second, where S is signal power and N is noise power in W. For the noiseless case (N = 0), this capacity becomes infinite, but the sampling theorem establishes that exact recovery is possible at f_s = 2W, as the samples fully encode the signal's infinite-precision values without distortion. Unlike frequency-domain proofs relying on spectral replication and aliasing avoidance, Shannon's approach emphasizes the intrinsic information content and finite dimensionality of bandlimited signals, rigorously justifying the Nyquist rate as the minimal sufficient rate for complete representation. This perspective underscores that the theorem is not merely about avoiding overlap in the frequency domain but about efficiently capturing the signal's degrees of freedom with no more or less than necessary.

Applications and Variations

Multidimensional Signals

The Nyquist–Shannon sampling theorem generalizes seamlessly to multidimensional signals, where the principle remains that a bandlimited function in N-dimensional space can be perfectly reconstructed from its samples on a lattice if the sampling density is sufficiently high. This extension, formalized for arbitrary dimensions, applies to signals such as two-dimensional images or three-dimensional volumetric data, whose Fourier transforms are confined to a bounded region in the multidimensional frequency domain. The critical sampling density is determined by the Lebesgue measure (volume) of this spectral support: for a support of measure V in N-dimensional frequency space, the lattice must have a density of at least V points per unit volume in the spatial domain to ensure no information loss. In two dimensions, relevant for imaging applications, the theorem specifies that a signal bandlimited to a region of area A in the (f_x, f_y) frequency plane requires a sampling rate of at least A samples per unit area. This holds regardless of the support's shape, though the optimal lattice geometry varies; for irregular or non-rectangular supports, the minimal density may approach this bound more efficiently than uniform rectangular grids. The adaptation preserves the core idea from one dimension—avoiding through adequate sampling—but the effective Nyquist rate now depends on the geometry of the spectral support, such as its area for isotropic cases or adjusted for anisotropic extensions. For instance, if the bandwidth is anisotropic, with different extents in x and y directions, the sampling lattice must account for the elongated support to prevent distortion. For uniform rectangular grid sampling of two-dimensional images, the condition translates to specific intervals Δx and Δy. Assuming the signal is bandlimited to frequencies |f_x| ≤ B_x and |f_y| ≤ B_y, the sampling must satisfy 1/(Δx Δy) ≥ 4 B_x B_y to capture the full bandwidth without aliasing. Reconstruction is then achieved via a two-dimensional : f(x, y) = \sum_{m=-\infty}^{\infty} \sum_{n=-\infty}^{\infty} f(m \Delta x, n \Delta y) \cdot \operatorname{sinc}\left( \frac{\pi (x - m \Delta x)}{\Delta x} \right) \cdot \operatorname{sinc}\left( \frac{\pi (y - n \Delta y)}{\Delta y} \right), where the sinc function is defined as \operatorname{sinc}(t) = \sin(t)/t. This separable form leverages the rectangular lattice's alignment with Cartesian coordinates, enabling efficient computation in digital imaging systems. Alternative lattices, such as hexagonal grids, offer improved efficiency for certain bandwidth shapes, particularly circular or isotropic supports common in natural images. A hexagonal arrangement achieves the Nyquist density with approximately 13% fewer samples than a square grid for the same aliasing-free reconstruction, due to its denser packing in the frequency domain that better tiles the spectral replicas. This reduction stems from the hexagonal lattice's fundamental domain covering 86.6% of the area required by square sampling for equivalent bandwidth containment. In practice, hexagonal sampling enhances resolution in applications like remote sensing and computer vision, though it complicates processing compared to rectangular grids. Key applications of multidimensional sampling include digital imaging, where cameras sample scenes on rectangular or hexagonal arrays to meet Nyquist criteria for high-fidelity capture, and magnetic resonance imaging (MRI), where k-space is sampled in two or three dimensions to reconstruct anatomical volumes without aliasing artifacts. In MRI, adherence to the theorem ensures that the field-of-view matches the sampling grid, preventing wrap-around effects in the reconstructed images. Challenges arise with anisotropic bandwidths, where the spectral support is not symmetric, requiring adaptive lattices to maintain the minimal density; suboptimal choices can lead to inefficient oversampling or incomplete coverage. Additionally, aliasing in two dimensions manifests as moiré patterns—interference fringes from overlapping spectral replicas—particularly evident when sampling fine periodic structures like fabrics or grids in photography. These patterns underscore the need for pre-sampling anti-aliasing filters to enforce bandlimiting before discretization.

Non-Baseband Sampling

Non-baseband signals, commonly referred to as bandpass signals, have their frequency content restricted to an interval [f_1, f_2] where f_1 > 0, and the bandwidth is defined as W = f_2 - f_1. These signals differ from signals, which occupy frequencies from 0 to some maximum f_{\max}, by having energy concentrated away from . The Nyquist–Shannon sampling theorem adapts to bandpass signals by specifying that the minimum sampling rate f_s must be at least twice the , f_s \geq 2W, rather than twice the highest frequency $2f_2. This adaptation enables , where f_s < 2f_2, provided the sampling avoids aliasing overlaps between the original spectrum and its replicas shifted by multiples of f_s. The condition ensures that the signal can be reconstructed perfectly from samples if the replicas are positioned without intrusion into the signal band. The precise allowable sampling rates are determined by the band position to prevent overlap. For an integer k \geq 1, the sampling frequency must satisfy \frac{2 f_2}{k+1} \leq f_s \leq \frac{2 f_1}{k}, which creates guard bands around the signal spectrum to isolate it from aliased components. For the lowest-order case (k=1), this yields f_s \geq 2W directly, but higher k allows even lower rates under stricter positioning. These constraints arise from the periodic replication of the spectrum in the frequency domain upon sampling. In practice, bandpass sampling is applied to radio frequency (RF) signals in communication systems, where high carrier frequencies with narrow modulation bandwidths permit efficient digitization at rates far below the carrier frequency, reducing hardware complexity. Another prominent example is quadrature sampling for in-phase (I) and quadrature (Q) demodulation, where the bandpass signal is mixed with cosine and sine carriers to produce lowpass equivalents, each sampled at f_s = 2W; this effectively captures the complex envelope without full-rate sampling of the carrier. However, successful reconstruction demands accurate prior knowledge of the band's location [f_1, f_2] to select an appropriate f_s; misalignment can cause replicas to overlap, leading to irreversible aliasing distortion. Sensitivity to band position and noise amplification in reconstruction filters further limits applicability in imprecise environments.

Nonuniform Sampling Methods

Nonuniform sampling refers to the process of acquiring signal values at irregular time instants t_n, where the intervals between consecutive samples vary, in contrast to the equally spaced intervals of uniform sampling. For a bandlimited signal with bandwidth B, perfect reconstruction remains possible provided the average sampling rate satisfies \lim_{T \to \infty} \frac{N(T)}{2T} \geq 2B, with N(T) denoting the number of samples in the interval [-T, T]. This average rate condition ensures that the samples capture sufficient information density, though the irregularity introduces additional considerations for reconstruction stability. Extensions of the Nyquist–Shannon theorem to nonuniform sampling, notably the Beutler-Yao theorem, establish conditions for error-free and stable recovery of bandlimited signals from irregularly spaced samples. Beutler's work demonstrates that reconstruction is feasible if the sampling set forms a "total" set with density exceeding the Nyquist rate, avoiding large gaps that could lead to information loss. Yao and Thomas further analyzed stability, showing that nonuniform expansions converge uniformly under bounded perturbations in sampling locations, provided the average density meets or exceeds $2B and the maximum gap between samples remains finite. These results generalize the uniform case by emphasizing density over regularity, enabling reconstruction via series expansions analogous to the cardinal series but adapted for irregular grids. Central to nonuniform sampling is the Landau rate, which defines the minimal average sampling density of $2B necessary for stable reconstruction of signals bandlimited to a set of total measure B. This lower bound, derived from density theorems for entire functions, implies that while nonuniformity allows flexibility, the overall sample count over long intervals must not fall below this threshold to prevent instability. Reconstruction typically employs iterative methods grounded in frame theory, where samples generate a frame for the signal space, allowing stable recovery through least-squares projections or oblique dual frames. Frame-based approaches, applicable in shift-invariant spaces containing bandlimited functions, ensure geometric convergence and robustness to noise when the frame bound exceeds unity. Alternatively, Prony's method can iteratively fit exponential sums to approximate bandlimited signals from nonuniform points, particularly effective for finite-duration or periodic cases. Practical examples of nonuniform sampling arise in analog-to-digital converters (ADCs), where clock jitter induces irregular timing, yet reconstruction is viable if the average rate adheres to the Nyquist criterion. This jitter, often modeled as Gaussian perturbations, serves as a precursor to compressive sensing by enabling sub-Nyquist exploration in sparse scenarios, though here it maintains the full rate on average. Hardware advantages include deliberate nonuniformity in time-interleaved ADCs, which reduces aperture jitter sensitivity and lowers overall system costs by allowing slower individual channels to achieve equivalent high rates through interleaving. Reconstruction from nonuniform samples presents challenges, primarily increased computational complexity due to the need for solving ill-conditioned linear systems or iterative optimizations, as opposed to the closed-form sinc interpolation in uniform cases. Error bounds are tightly linked to measures of irregularity, such as the discrepancy between the empirical sampling distribution and uniform density, or the maximal gap size; larger discrepancies amplify reconstruction errors, with bounds scaling proportionally to the perturbation norm. These factors necessitate careful design to balance density sufficiency with gap control for practical stability.

Undersampling with Constraints

The Nyquist–Shannon sampling theorem traditionally mandates a sampling rate f_s \geq 2B for perfect reconstruction of bandlimited signals with bandwidth B, but undersampling below this rate becomes feasible when additional constraints on the signal structure are imposed, such as sparsity in a known basis or other parametric forms. These priors enable sub-Nyquist sampling by exploiting the signal's low intrinsic dimensionality, allowing recovery through optimization techniques rather than direct interpolation. A prominent approach is compressed sensing, pioneered by Donoho and independently by Candès, Romberg, and Tao, which targets signals that are k-sparse in an orthonormal basis, meaning they have at most k nonzero coefficients in a representation of length N. In this framework, a signal can be recovered from m linear measurements where m \geq c k \log(N/k) for some constant c > 0, far fewer than the N samples required without sparsity. is achieved via , such as \ell_1-norm minimization: \hat{x} = \arg\min \| x \|_1 \quad \text{subject to} \quad y = \Phi x, where y are the measurements, \Phi is the measurement matrix, and the restricted isometry property ensures stable reconstruction provided the sparsity basis and measurement basis are incoherent. Other methods include sampling of signals with finite rate of innovation (FRI), introduced by Vetterli et al., which applies to parametric signals like sums of Diracs or exponentials with a finite number of degrees of freedom per unit time, say \rho, allowing sampling at rates slightly above $2\rho using annihilating filters or pronys methods for reconstruction. Multi-coset sampling, developed by Mishali and Eldar, employs multiple parallel samplers with periodic undersampling patterns to capture spectrally sparse multiband signals, enabling sub-Nyquist rates when the signal occupies a small portion of a wide spectrum, with recovery via basis pursuit or greedy algorithms assuming unknown band locations. These techniques find applications in radar systems for sparse target detection and for spectrum sensing of sparse wideband signals, where the assumption of incoherence between the sparsity and measurement domains is crucial for reliable recovery. However, they require precise prior knowledge of the signal model, and reconstruction is generally but not perfectly invertible in the presence of , degrading performance as noise levels increase.

Historical Development

Early Contributions

The foundations of the Nyquist–Shannon sampling theorem emerged in the through mathematical explorations of from points. In 1841, developed a trigonometric interpolation formula for periodic functions from equally spaced samples, serving as an early precursor to signal reconstruction techniques, though not framed in terms of frequency limitations. Similarly, Joseph-Louis Lagrange's method, introduced in the , provided a systematic way to approximate continuous functions using a of observations, influencing subsequent approaches to in analysis. By the early , the rise of and electrical communications spurred interest in representing signals efficiently, leading to more targeted results on bandlimited functions. In 1915, formulated the cardinal series expansion, demonstrating that certain bandlimited analytic functions could be uniquely interpolated from their values at uniform sampling points spaced at intervals of 1/(2B), where B is the , using shifted sinc functions. This work, published in the Proceedings of the Royal Society of , established a mathematical basis for sampling without loss of information for functions of exponential type. In 1933, Vladimir A. Kotelnikov presented the paper "On the of 'Ether' and Wire in Electrical Communications" at the First All-Union Conference on Questions of Communications in Leningrad, proving that a continuous-time signal bandlimited to W can be perfectly reconstructed from uniform samples taken at rate 2W, motivated by limits in transmission systems; however, the Russian-language publication was overlooked in until the . These developments occurred amid growing needs in wired and wireless telegraphy, where discrete pulse representations were essential for efficient signaling, yet lacked a cohesive theorem uniting them. Parallel concepts appeared in optics, as Ernst Abbe's 1873 diffraction limit described the minimal resolvable distance in microscopy as λ/(2NA)—where λ is wavelength and NA is numerical aperture—imposing a spatial frequency constraint akin to the temporal sampling requirement for avoiding aliasing. Nyquist's 1928 analysis of pulse transmission rates in telegraph channels later bridged these mathematical precursors to practical engineering applications.

Attribution and Recognition

The Nyquist–Shannon sampling theorem derives its name from the foundational contributions of and , whose works established the core principles of sampling bandlimited signals. In , Nyquist published "Certain Topics in Telegraph Transmission Theory," where he analyzed the transmission of pulses over telegraph s and derived that, to avoid , the pulse rate must be at least twice the B of the channel, expressed as f_s \geq 2B. This result, though applied to discrete pulse signaling rather than continuous waveforms, laid the groundwork for the sampling in communication systems. Claude Shannon extended and formalized Nyquist's insight in his 1949 paper "Communication in the Presence of Noise," proving that any bandlimited signal with bandwidth W can be completely reconstructed from samples taken at a rate of at least 2W per second, using the cardinal series (sinc interpolation) for perfect recovery. Shannon's treatment generalized the theorem to arbitrary continuous-time signals, integrating it into information theory and emphasizing its implications for noise-free transmission. Independent discoveries preceded and paralleled these efforts, notably by Vladimir A. Kotelnikov, who in 1933 presented "On the Carrying Capacity of 'Ether' and Wire in Electrical Communications" at a Leningrad conference, rigorously stating the sampling theorem for both lowpass and bandpass signals at twice the highest frequency. Kotelnikov's work, however, remained largely overlooked in the West due to its publication in Russian with limited circulation among a student audience and the geopolitical isolation of Soviet research at the time. Parallel independent discoveries occurred, including by Fumio Someya in Japan around 1940, though similarly overlooked until later due to wartime conditions. These contributions were constrained by wartime secrecy and specialized venues, contributing to their delayed recognition. The naming convention evolved from these origins: "Nyquist rate" emerged in 1930s engineering literature to denote the minimum signaling rate of 2B, popularized in telephony and control systems contexts. Shannon's formulation inspired the term "Shannon sampling theorem" in mathematical and circles by the late 1940s. The hyphenated "Nyquist–Shannon sampling theorem" gained prominence in 1950s textbooks and reviews, reflecting the complementary and theoretical perspectives, such as in S. Goldman's 1953 "Information Theory." Debates over priority have persisted, with claims highlighting Kotelnikov's earlier comprehensive proof and the overlooked contributions from other regions. Kotelnikov's work gained formal recognition in the West starting in the through translations and citations in international literature. Contemporary scholarship emphasizes inclusivity by increasingly citing these diverse origins to foster a more complete historical narrative.

References

  1. [1]
    What Is the Nyquist Theorem - MATLAB & Simulink - MathWorks
    The Nyquist theorem holds that a continuous-time signal can be perfectly reconstructed from its samples if it is sampled at a rate greater than twice its ...Missing: fundamentals primary sources
  2. [2]
    [PDF] Communication In The Presence Of Noise - Proceedings of the IEEE
    THE SAMPLING THEOREM. Let us suppose that the channel has a certain bandwidth in cps starting at zero frequency, and that we are allowed to use this channel ...
  3. [3]
    [PDF] Certain topics in telegraph transmission theory
    The paper discusses frequency bands, minimum band width, distortionless transmission, steady-state characteristics, and the relationship between frequency band ...
  4. [4]
    Certain Topics in Telegraph Transmission Theory - IEEE Xplore
    This paper discusses calculating telegraph signal distortion using steady-state characteristics, criteria for distortionless transmission, and minimum ...
  5. [5]
    Nyquist and Shannon's Sampling Theorems - NI
    ### Summary of Nyquist and Shannon Sampling Theorems
  6. [6]
    [PDF] AN-236 An Introduction to the Sampling Theorem - Texas Instruments
    In the sample data system band-limiting is achieved by attenuating those frequencies greater than the Nyquist frequency to a level undetectable or invisible to ...
  7. [7]
  8. [8]
    The Nyquist-Shannon Sampling Theorem
    A precise statement of the Nyquist-Shannon sampling theorem is now possible. Given a continuous-time signal x with Fourier transform X where X(ω ) is zero ...
  9. [9]
    [PDF] Chapter 5: Sampling and Quantization
    This phenomenon is called aliasing, since one frequency appears to be a different frequency if the sampling rate is too low.
  10. [10]
    [PDF] Communication in the Presence of Noise* - MIT
    A precise meaning will be given later to the requirement of reliable resolution of the M signals. II. THE SAMPLING THEOREM. Let us suppose that the channel ...<|control11|><|separator|>
  11. [11]
    [PDF] Sampling and Reconstruction
    This phenomenon is called aliasing, presumably because it implies that any discrete-time sinusoidal signal has many continuous-time identities (its “identity” ...
  12. [12]
    The Wagon Wheel Effect - Computer Science | UC Davis Engineering
    Dec 30, 2024 · The wagon-wheel effect (also called the stroboscopic effect) is an optical illusion in which a wheel appears to rotate differently from its true rotation.
  13. [13]
    [PDF] Sampling and Aliasing
    When we sample at a rate which is less than the Nyquist rate, we say we are undersampling and aliasing will yield misleading results. • If we are sampling a 100 ...
  14. [14]
    [PDF] MT-002: What the Nyquist Criterion Means to Your ... - Analog Devices
    The mathematical basis of sampling was set forth by Harry Nyquist of Bell. Telephone Laboratories in two classic papers published in 1924 and 1928, respectively ...
  15. [15]
    Explanation of 44.1 kHz CD sampling rate - CS@Columbia
    Jan 9, 2008 · The CD sampling rate has to be larger than about 40 kHz to fulfill the Nyquist criterion that requires sampling at twice the maximum analog ...Missing: source | Show results with:source
  16. [16]
    [PDF] The Shannon Sampling Theorem and Its Implications
    It is common to formulate Poisson's summation formula with only B = 1/2. f ( x + n 2B ) e−2πimx·2B dx. f(y)e−2πimy·2B = 2B ˆ f(2Bm).
  17. [17]
    Communication in the Presence of Noise | IEEE Journals & Magazine
    ... Proceedings of the IRE >Volume: 37 Issue: 1. Communication in the Presence of Noise. Publisher: IEEE. Cite This. PDF. C.E. Shannon. All Authors. Sign In or ...
  18. [18]
    Sampling and reconstruction of wave-number-limited functions in N ...
    This paper includes some preliminary results of research conducted by Mr. Petersen under the guidance of Dr. Middleton, in partial fulfillment of the ...
  19. [19]
    The Sampling Theorem in Higher Dimensions - SpringerLink
    Multidimensional signals include black and white images, which can be depicted as two dimensional functions with zero corresponding to black and one as white.
  20. [20]
    Compressed sensing MRI: a review from signal processing ...
    Mar 29, 2019 · Hence, in order to obtain an image without aliasing artifacts, k-space samples need to satisfy the Nyquist sampling criterion. Despite this ...
  21. [21]
    [PDF] Moire Patterns And Two-Dimensional Aliasing In Line Scanner Data ...
    If the low pass filtering is inadequate, then the resulting Moire' patterns arise from two-dimensional aliasing. This type of distortion is particularly ...Missing: 2D | Show results with:2D
  22. [22]
    The theory of bandpass sampling | IEEE Journals & Magazine
    Abstract: The sampling of bandpass signals is discussed with respect to band position, noise considerations, and parameter sensitivity.Missing: paper | Show results with:paper
  23. [23]
    [PDF] The Sampling Theorem and the Bandpass Theorem
    It has been revealed that the highest detectable frequency in the sampled data is the Nyquist frequency of π radians per interval. An alternative statement is ...
  24. [24]
    [PDF] Non-Uniform Sampling in Statistical Signal Processing - DiVA portal
    Non-uniform sampling comes natural in many applications, due to for example imperfect sensors, mismatched clocks or event-triggered phenomena. Examples.
  25. [25]
    Error-Free Recovery of Signals from Irregularly Spaced Samples
    IEEE Transactions on Signal Processing, Vol. 48, No. 2 | 1 Jan 2000. Filterbank reconstruction of bandlimited signals from nonuniform and generalized samples.
  26. [26]
  27. [27]
    [PDF] Minimum rate sampling and reconstruction of signals with ... - Microsoft
    We give a characterization of the signals that can be reconstructed at exactly the minimum rate once a nonuniform sampling pattern has been fixed.
  28. [28]
    Nonuniform Sampling and Reconstruction in Shift-Invariant Spaces
    This article discusses modern techniques for nonuniform sampling and reconstruction of functions in shift-invariant spaces. It is a survey as well as a ...Missing: challenges | Show results with:challenges
  29. [29]
    A Prony-like method for non-uniform sampling - ScienceDirect
    Prony's approach uses two separate least-squares solutions, each of order n p , with the first least-squares solution resulting in the frequencies of the signal ...
  30. [30]
    [PDF] NONUNIFORM SAMPLING DRIVER DESIGN FOR OPTIMAL ADC ...
    Deliberate nonuniform sampling promises increased equivalent sampling rates with reduced overall hardware costs of the DSP system. The equivalent sampling rate ...Missing: advantages | Show results with:advantages
  31. [31]
    Numerical analysis of the non-uniform sampling problem
    We give an overview of recent developments in the problem of reconstructing a band-limited signal from nonuniform sampling from a numerical analysis view ...
  32. [32]
    Sparsity and incoherence in compressive sampling - IOPscience
    [4] Candès E J, Romberg J and Tao T 2006 Robust uncertainty principles: exact signal reconstruction from highly incomplete frequency ...Missing: paper | Show results with:paper
  33. [33]
    [PDF] Sampling signals with finite rate of innovation - CUHK EE
    Abstract—Consider classes of signals that have a finite number of degrees of freedom per unit of time and call this number the rate of innovation.
  34. [34]
    [PDF] Blind Multiband Signal Reconstruction: Compressed Sensing for ...
    We begin by quoting Landau's theorem for the minimal sampling rate of an arbitrary sampling method that allows known-spectrum perfect reconstruction. We then ...
  35. [35]
  36. [36]
    [0709.1563] Blind Multi-Band Signal Reconstruction - arXiv
    Sep 11, 2007 · Our approach assumes an existing blind multi-coset sampling method. The sparse structure of multi-band signals in the continuous frequency ...
  37. [37]
    Lagrange Interpolation and Sampling Theorems - SpringerLink
    The aim of this chapter is to discuss the relationship between Lagrange interpolation and sampling theorems.
  38. [38]
  39. [39]
    Sampling Theory - Cambridge University Press & Assessment
    Covering the fundamental mathematical underpinnings together with key principles and applications, this book provides a comprehensive guide to the theory ...
  40. [40]
    The Diffraction Barrier in Optical Microscopy | Nikon's MicroscopyU
    The diffraction-limited resolution theory was advanced by German physicist Ernst Abbe in 1873 (see Equation (1)) and later refined by Lord Rayleigh in 1896 ...
  41. [41]
    [PDF] Certain Topics in Telegraph Transmission Theor-y - Monoskop
    It differs from one signal in other words, the total number of components is element to another but is the same in all parts of the usually infinite. Page 4 ...
  42. [42]
    Vladimir A. Kotelnikov - Engineering and Technology History Wiki
    Mar 1, 2016 · Professor Kotelnikov led the formulation and proof of the sampling theorem ... There is little doubt that Kotelnikov's 1933 paper was the first to ...
  43. [43]
    [PDF] The Establishment of Sampling as a Scientific Principle—A Striking ...
    During the period 1928–1949, several engineers contributed to the establishment of a sampling principle. They did this virtually independently.Missing: 1841 precursor
  44. [44]
    [PDF] A Chronology of Interpolation: From Ancient Astronomy to Modern ...
    Abstract—This paper presents a chronological overview of the developments in interpolation theory, from the earliest times to the present date.
  45. [45]
    The Nyquist-Shannon Sampling Theorem, Published Nine Years ...
    The Nyquist-Shannon sampling theorem, implied by Nyquist's work, was proved by Claude Shannon in 1949, though the manuscript was received in 1940.
  46. [46]
    [PDF] Vladimir Aleksandrovich Kotelnikov: Pioneer of the sampling ...
    VLADIMIR ALEKSANDROVICH KOTELNIKOV: PIONEER OF THE SAMPLING THEOREM, ... There is little doubt that Kotelnikov's 1933 paper was the first to address ...
  47. [47]