The Whittle likelihood is a frequency-domain approximation to the exact Gaussian likelihood function for second-order stationarytime series, enabling efficient parameter estimation by leveraging the periodogram and a parametric model for the spectral density.[1] Introduced by mathematician Peter Whittle in his 1951 PhD thesis, with further development in his 1953 paper on the analysis of multiple stationarytime series,[2][3] it simplifies the computation of the negative log-likelihood from O(n³) operations—required for exact evaluation via the Cholesky decomposition of the covariance matrix—to O(n log n) using the fast Fourier transform.[4]The approximation is given by the formula-2 \log L(\theta) \approx 2n \log(2\pi) + \sum_{j=1}^{m} \left[ \log f_\theta(\omega_j) + \frac{I(\omega_j)}{f_\theta(\omega_j)} \right],where n is the sample size, m \approx n/2 is the number of Fourier frequencies, f_\theta(\omega_j) is the spectral density at frequency \omega_j under parameter \theta, and I(\omega_j) is the periodogram ordinate.[4] This form treats the periodogram values as approximately independent exponential random variables with mean f_\theta(\omega_j), which holds asymptotically for large n.[1] Maximizing the Whittle likelihood yields estimators that are consistent and asymptotically normal, asymptotically equivalent to exact maximum likelihood estimators for Gaussian ARMA and fractional ARIMA processes.[4]Originally developed for univariate time series, the Whittle likelihood has been extended to multivariate, spatial, and nonstationary settings, including debiased versions to correct finite-sample bias in parameter estimates.[5] It is widely applied in spectral analysis, econometrics, and signal processing—such as estimating long-memory parameters like the Hurst exponent in fractal time series or covariance structures in physiological data—due to its balance of statistical efficiency and low computational cost.[6]
Introduction
Overview and Definition
The Whittle likelihood serves as a pseudolikelihood approximation to the exact Gaussian likelihood function for a stationary Gaussian time series, formulated in the frequency domain by leveraging the periodogram as a nonparametric estimate of the power spectrum and the parametric spectral density of the process. This approach treats the discrete Fourier transforms at distinct frequencies as approximately independent complex Gaussian random variables, enabling a simplification of the intractable multivariate Gaussian density in the time domain.[7]The Whittle log-likelihood is given byL_W(\theta) = - \sum_{k=1}^m \left[ \log f_\theta(\omega_k) + \frac{I(\omega_k)}{f_\theta(\omega_k)} \right],up to additive and multiplicative constants independent of \theta, where f_\theta(\omega_k) denotes the spectral density of the process parameterized by \theta, evaluated at the Fourier frequencies \omega_k = 2\pi k / n for k = 1, \dots, m with m \approx n/2 and sample size n; I(\omega_k) is the periodogram ordinate at \omega_k, computed as I(\omega_k) = \frac{1}{n} \left| \sum_{t=1}^n X_t e^{-i \omega_k t} \right|^2.[8] This expression arises from approximating the quadratic form and determinant in the exact Gaussian likelihood via their spectral representations, up to constants independent of \theta.[7]A primary advantage of the Whittle likelihood lies in its computational efficiency, requiring O(n \log n) operations through the fast Fourier transform (FFT) to compute the periodogram, in contrast to the O(n^3) cost of exact likelihood evaluation via direct matrix inversion for general stationary covariances.[9] It proves particularly useful in scenarios where the dependence structure renders the exact likelihood intractable, such as in high-dimensional or complex parametric models for time series data.[8] The method was originally developed by Peter Whittle in the early 1950s for spectral estimation in stationary processes.
Historical Development
The Whittle likelihood emerged in the post-World War II period, a time of rapid advancement in statistical methods for analyzing stationary time series, spurred by applications in signal processing and economics. This era saw increased focus on spectral methods, influenced by Andrey Kolmogorov's 1941 work on prediction theory for stationary processes and Norbert Wiener's 1949 treatise on extrapolation and smoothing of stationary time series, which formalized the spectral representation of such processes. Peter Whittle, a New Zealand-born mathematician pursuing his PhD at Uppsala University, introduced the core idea in his 1951 thesis, "Hypothesis Testing in Time Series Analysis," where he proposed approximating the likelihood for Gaussian stationary processes in the frequency domain to facilitate parameter estimation.[10]Whittle's foundational contributions were detailed in his 1953 publications, including "Estimation and Information in Stationary Time Series," which derived an approximation to the Gaussian log-likelihood by representing the quadratic form in the time domain via Parseval's theorem in the frequency domain, enabling computationally efficient inference for univariate processes. That same year, in "The Analysis of Multiple Stationary Time Series," he extended the approach to multivariate cases, laying the groundwork for joint spectralestimation across multiple series.[11] These works positioned the Whittle approximation as a practical tool for spectrum-based estimation, contrasting with the computationally intensive exact likelihood methods prevalent at the time.By the early 1960s, the Whittle likelihood gained traction for estimating parameters in autoregressive moving average (ARMA) models, as evidenced by James Durbin's 1960 review, which highlighted its utility in fitting time series models through frequency-domain approximations and discussed its alignment with classical least-squares principles. Durbin's subsequent analyses further explored the asymptotic properties of such estimators, contributing to the method's theoretical robustness. Although initially underappreciated, Whittle's innovation influenced later extensions, including spatial variants for geostatistical data.
Mathematical Foundations
Gaussian Likelihood in Time Series
A stationary Gaussian time series \{X_t, t \in \mathbb{Z}\} is a stochastic process where each X_t follows a Gaussian distribution, the mean \mathbb{E}[X_t] = 0 is constant (without loss of generality by centering), and the covariance function \Gamma(h) = \mathbb{E}[X_t X_{t+h}] depends solely on the time lag h. This weak stationarity ensures that the joint distribution of any finite collection of observations is multivariate Gaussian with a covariance matrix that exhibits a Toeplitz structure, reflecting the time-invariant dependence pattern.For a finite sample X = (X_1, \dots, X_n)^T \sim \mathcal{N}(0, \Gamma_\theta), where \Gamma_\theta is the n \times n Toeplitz covariance matrix with entries (\Gamma_\theta)_{i,j} = \Gamma_\theta(|i-j|) parameterized by \theta, the exact log-likelihood function is given byL(\theta) = -\frac{n}{2} \log(2\pi) - \frac{1}{2} \log \det \Gamma_\theta - \frac{1}{2} X^T \Gamma_\theta^{-1} X.This expression arises directly from the density of the multivariate Gaussian distribution and serves as the basis for maximum likelihood estimation of the parameters \theta.Evaluating this log-likelihood poses significant computational challenges for large n, as computing the determinant and inverse of \Gamma_\theta requires O(n^3) operations using standard methods, rendering it intractable for high-dimensional data. While specialized algorithms, such as the Levinson-Durbin recursion, can reduce the cost to O(n^2) for certain structured covariances like those in autoregressive moving average (ARMA) models, the complexity escalates for processes with long-memory features, where covariances decay slowly (e.g., \Gamma(h) \sim h^{2d-1} for fractional differencing parameter d \in (0, 0.5)), leading to ill-conditioned matrices and numerical instability.The spectral density function provides a frequency-domain representation of the covariance structure, defined asf_\theta(\omega) = \frac{1}{2\pi} \sum_{h=-\infty}^{\infty} \Gamma(h) e^{-i h \omega}, \quad \omega \in [-\pi, \pi],which is nonnegative, integrable, and integrates to the variance \int_{-\pi}^{\pi} f_\theta(\omega) \, d\omega = \Gamma(0). This formulation bridges the time-domain covariances to the frequency domain, facilitating approximations that exploit asymptotic independence of periodogram ordinates at distinct frequencies.
Derivation of the Whittle Approximation
The derivation of the Whittle approximation begins with the exact Gaussian log-likelihood for a zero-mean stationary time series X = (X_1, \dots, X_n)^T observed under a parametric spectral density f_\theta(\omega), where the covariance matrix \Gamma = \Gamma(\theta) is Toeplitz and positive definite.The log-likelihood is given by L(\theta) = -\frac{n}{2} \log(2\pi) - \frac{1}{2} \log \det \Gamma - \frac{1}{2} X^T \Gamma^{-1} X. To approximate this in the frequency domain, the first step applies Parseval's theorem, which equates the energy in the time domain to that in the frequency domain for Fourier transforms. Specifically, the quadratic form X^T \Gamma^{-1} X is rewritten using the discrete Fourier transform of X, leveraging the fact that for large n, the eigenvectors of the Toeplitz matrix \Gamma are approximately the Fourier basis vectors \exp(i t \omega_k) for frequencies \omega_k = 2\pi k / n, k = 0, \dots, n-1. This transforms the quadratic term into a sum approximation \sum_{k=0}^{n-1} \frac{I(\omega_k)}{f_\theta(\omega_k)}, where I(\omega_k) is the periodogram ordinate, under the assumption of circular stationarity or high-sample-size conditions that justify the Fourier diagonalization.[4]The second step approximates the determinant \log \det \Gamma. For Toeplitz matrices generated by the spectral density f_\theta(\omega), Szegő's theorem provides the asymptotic relation \frac{1}{n} \log \det \Gamma \approx \frac{1}{2\pi} \int_{-\pi}^{\pi} \log(2\pi f_\theta(\omega)) d\omega as n \to \infty, assuming f_\theta(\omega) is positive and continuous. This yields \log \det \Gamma \approx n \log(2\pi) + n \frac{1}{2\pi} \int_{-\pi}^{\pi} \log f_\theta(\omega) d\omega + o(n), enabling a frequency-domain expression for the normalizing constant in the likelihood.The third step discretizes the frequency integrals using a Riemann sum over the Fourier frequencies \omega_k. The continuous integral for the determinantterm becomes \frac{1}{2\pi} \int_{-\pi}^{\pi} \log f_\theta(\omega) d\omega \approx \frac{1}{n} \sum_{k=0}^{n-1} \log f_\theta(\omega_k), and similarly for the quadratic term, \sum_{k=0}^{n-1} \frac{I(\omega_k)}{f_\theta(\omega_k)}, where the periodogram is I(\omega_k) = \frac{1}{n} \left| \sum_{t=1}^n X_t e^{-i t \omega_k} \right|^2. This discretization ignores endpoint corrections at \omega = 0 and \omega = \pm \pi for large n, relying on the high-frequency approximation where aliasing effects are negligible for processes with smooth spectra.[4]Combining these steps, the approximate Whittle log-likelihood isL_W(\theta) \approx -n \log(2\pi) - \frac{1}{2} \sum_{k=0}^{n-1} \log f_\theta(\omega_k) - \frac{1}{2} \sum_{k=0}^{n-1} \frac{I(\omega_k)}{f_\theta(\omega_k)},which serves as a computationally efficient surrogate for maximum likelihood estimation under the assumptions of Gaussianity, stationarity, and large sample size. This form was originally proposed by Whittle in his foundational work on time series hypothesis testing.[4]
Properties and Accuracy
Approximation Error Bounds
The Whittle likelihood approximation introduces several sources of error in finite samples, primarily stemming from the discretization of the continuous integral representation of the Gaussian log-likelihood via a Riemann sum over Fourier frequencies, aliasing effects at low frequencies due to the periodic extension of the finite observed series, and boundary effects in the periodogram estimation arising from the truncation of the infinite process to a finite sample.[7][8]A foundational quantitative bound on the approximation error was established by Durbin, who showed that the difference between the exact Gaussian log-likelihood L(\theta) and the Whittle log-likelihood L_W(\theta) satisfies |L(\theta) - L_W(\theta)| = O(\log n / n) for sample size n, assuming sufficient smoothness of the parametric spectral density f_\theta.[12] This rate reflects the dominant contribution from the logarithmic term in the periodogram variance at low frequencies, which diminishes slowly relative to higher-order terms.Hannan provided refinements to this analysis, developing higher-order asymptotic expansions for autoregressive processes that achieve an error of O_p(1/n) under conditions ensuring consistent estimation of the spectral density, such as finite fourth moments and spectral continuity. These expansions highlight improved accuracy for short-memory AR models by accounting for additional covariance structure beyond the zeroth-order approximation.Worst-case errors in the Whittle approximation are particularly pronounced for long-memory processes, such as fractional ARIMA models, where low-frequency bias dominates due to the slow decay of the spectral density near zero, leading to amplified discrepancies between the periodogram and f_\theta at those frequencies.
Asymptotic Behavior and Consistency
The Whittle maximum likelihood estimator (MLE), defined as \hat{\theta}_W = \arg\max_\theta L_W(\theta), where L_W(\theta) denotes the Whittle log-likelihood, exhibits strong consistency as the sample size n \to \infty for stationarytime series under standard assumptions of ergodicity, stationarity, and parameter identifiability.[13] This result holds because the Whittle approximation converges in probability to the true Gaussian log-likelihood, ensuring that the maximizer \hat{\theta}_W approaches the true parameter \theta_0 almost surely.[13]Under mild regularity conditions, such as the existence of moments and spectral density smoothness, the Whittle MLE is asymptotically normal: \sqrt{n} (\hat{\theta}_W - \theta_0) \xrightarrow{d} N(0, I(\theta_0)^{-1}), where the asymptotic covariance matrix is the inverse of the Fisher information matrix I(\theta_0) = \int_{-\pi}^{\pi} \left( \frac{\partial \log f(\omega; \theta_0)}{\partial \theta} \right) \left( \frac{\partial \log f(\omega; \theta_0)}{\partial \theta^T} \right) \frac{d\omega}{2\pi f(\omega; \theta_0)} and f(\omega; \theta) is the spectral density parameterized by \theta.[14] This distribution arises from the central limit theorem applied to the score function in the frequency domain, with the information matrix capturing the expected curvature of the Whittle log-likelihood.[14]For autoregressive moving average (ARMA) models under Gaussianity, the Whittle MLE achieves asymptotic efficiency equivalent to the exact Gaussian MLE, attaining the Cramér-Rao lower bound with variance I(\theta_0)^{-1}. However, in non-Gaussian or spatial settings, the efficiency may not hold fully, as the Whittle approximation assumes Gaussianity and univariate stationarity, leading to potential discrepancies in the asymptotic variance.Recent theoretical advances have reconciled the Whittle and exact Gaussian likelihoods more precisely, showing that their difference converges at the rate O(1/n) under \beta-mixing conditions for dependent data, thereby preserving the asymptotic normality and efficiency of \hat{\theta}_W even in misspecified models.[8] This reconciliation, established for AR(1) processes and extendable to broader classes, underscores the robustness of Whittle-based inference in large samples.[8]
Extensions
Spatial Whittle Likelihood
The spatial Whittle likelihood provides an approximation to the Gaussian likelihood for stationary random fields observed on a d-dimensional lattice, extending the univariate time series formulation to multivariate spatial processes. For a Gaussian random field with parameterized spectral density matrix f_\theta(\omega), the method leverages the discrete Fourier transform to compute a periodogram I(\omega_k) at Fourier frequencies \omega_k, yielding a computationally efficient pseudo-likelihood that avoids direct evaluation of the full covariance matrix.[15]The approximate log-likelihood is expressed asL_W(\theta) = -\frac{N}{2} \log(2\pi) - \frac{1}{2} \sum_k \log \det f_\theta(\omega_k) - \frac{1}{2} \sum_k \trace\left( I(\omega_k) f_\theta(\omega_k)^{-1} \right),where N denotes the number of observation sites and the sum runs over the distinct Fourier frequencies. For univariate scalar fields common in spatial statistics, f_\theta(\omega) and I(\omega_k) are scalars, simplifying the determinant and trace terms to f_\theta(\omega_k) and I(\omega_k)/f_\theta(\omega_k), respectively. This formulation, originally adapted for spatial lattices with missing values, requires O(N \log^2 N) operations via fast Fourier transforms, facilitating inference in high-dimensional settings.[16]A key development for practical implementation with Matérn covariograms occurred in the work of Guinness and Fuentes, who introduced circulant embedding of approximate covariances to compute the aliased spectral densities efficiently using FFTs, even on large lattices where exact Matérn evaluations are prohibitive due to aliasing. Their quasi-Matérn approximation ensures positive definiteness and rapid evaluation without iterative aliasing sums, supporting parameter estimation for covariance functions like K_\theta(h) = \sigma^2 \frac{2^{1-\nu}}{\Gamma(\nu)} (\|h\|/\lambda)^\nu K_\nu(\|h\|/\lambda). When d=1, the spatial Whittle likelihood recovers the classical time series approximation.[17]This likelihood handles irregular boundaries and incomplete data through tapering functions, such as cosine tapers, which weight observations to mitigate spectral leakage and boundary effects while preserving asymptotic properties. In geostatistics, it enables scalable analysis of environmental datasets, including climate applications like interpolating photosynthetically active radiation from satellite imagery with cloud-induced missing values.[16][18]
Debiased and Penalized Variants
The debiased Whittle likelihood addresses the boundary bias inherent in the standard Whittle approximation, which arises from spectral leakage and aliasing effects in finite samples. This bias is corrected by replacing the spectral density f_\theta(\omega) in the likelihood with its expected value under the model, \bar{f}_n(\omega; \theta), computed via convolution with the Fejér kernel to account for blurring. Techniques such as data tapering with discrete prolate spheroidal sequences or differencing the process further mitigate broadband bias, significantly reducing bias compared to the unmodified Whittle estimator.The debiased likelihood is given by\ell_D(\theta) = -\sum_{\omega \in \Omega} \left\{ \log \bar{f}_n(\omega; \theta) + \frac{I(\omega)}{\bar{f}_n(\omega; \theta)} \right\},where I(\omega) is the periodogram and \Omega is the set of Fourier frequencies, often adjusted by excluding low or high frequencies or applying weights to focus on the reliable spectral range. This formulation achieves O_p(1/n) bias for a range of stationary processes, including non-Gaussian cases, while maintaining the n^{-1/2} convergence rate of maximum likelihood estimators.Penalized variants extend the Whittle likelihood by incorporating a roughness penalty on the log-spectral density to enforce smoothness, particularly beneficial in high-dimensional spatial settings where nonparametric estimation is prone to overfitting. The estimator minimizes a criterion combining the Whittle negative log-likelihood with an L2-style penalty term, such as \lambda \int |\frac{d^2}{d\omega^2} \log f(\omega)|^2 d\omega, where \lambda is tuned via cross-validation. This approach unifies kernel and spline-based methods and yields consistent estimates under non-Gaussian assumptions for regularly spaced spatial data.[19]A spatial extension of the debiased Whittle likelihood targets covarianceparameterestimation in large regular grids, correcting for boundary effects and aliasing in multidimensional settings. Simulations on Matérn covariance models demonstrate improved estimation variance relative to standard Whittle methods, especially for range parameters on grids with significant missing data. These refinements enhance accuracy in non-stationary data analysis, such as estimating fractal exponents in long-memory processes.[20]Recent developments as of 2025 include extensions to gravitational wave background analysis and modeling on linear networks using Whittle-Matérn fields.[21][22]
Parameter estimation using the Whittle likelihood involves maximizing the approximate log-likelihood function L_W(\theta) = -\frac{1}{2\pi} \int_{-\pi}^{\pi} \left[ \log f_\theta(\omega) + \frac{I(\omega)}{f_\theta(\omega)} \right] d\omega with respect to the model parameters \theta, where f_\theta(\omega) is the parametric spectral density and I(\omega) is the periodogram of the observed time series. This maximization is typically performed numerically due to the lack of closed-form solutions, employing iterative optimization techniques such as the Newton-Raphson method, which updates parameter estimates via the gradient and Hessian of L_W(\theta), or the expectation-conditional maximization (ECM) algorithm for handling complex dependencies in the likelihood.[23] Initial parameter values are often obtained from simpler approximations like the Yule-Walker equations, which provide moment-based estimates for autoregressive components to ensure convergence.A representative example is the estimation of parameters in an autoregressive AR(p) model, where the spectral density takes the formf_\theta(\omega) = \frac{\sigma^2}{|1 - \sum_{j=1}^p \phi_j e^{-i j \omega}|^2},with \theta = (\phi_1, \dots, \phi_p, \sigma^2). The Whittle likelihood is then maximized to yield estimates \hat{\phi} and \hat{\sigma}^2, which capture the model's autoregressive structure efficiently in the frequency domain.[24] These estimates leverage the periodogram I(\omega), computed via the fast Fourier transform (FFT) for O(n log n) time complexity, providing an efficient frequency-domain alternative to time-domain methods like Kalman filter-based exact maximum likelihood, which scale as O(n) but require model-specific state-space implementations. For models with nuisance parameters, such as higher-order terms in ARMA processes, profile likelihood techniques concentrate the objective function by optimizing over the nuisances first, reducing the dimensionality of the main optimization.[24]Whittle estimators are asymptotically efficient, achieving the Cramér-Rao lower bound under Gaussian assumptions.
Spectrum and Fractal Analysis
The Whittle likelihood functions as a contrast for nonparametric estimation of the spectral density f(\omega) of stationary Gaussian processes, extending Peter Whittle's foundational work on fitting spectra through approximate likelihood maximization in the frequency domain. This approach treats the periodogram ordinates as asymptotically independent exponential random variables with means given by the spectral density, enabling efficient estimation without assuming a parametric form for f(\omega). For power-law spectra common in long-memory processes, estimation proceeds by minimizing the Whittle contrastfunction\sum_k \left[ \log f(\omega_k) + \frac{I(\omega_k)}{f(\omega_k)} \right],where I(\omega_k) denotes the periodogram at Fourier frequencies \omega_k, and the sum is over a suitable bandwidth of frequencies.[25] This formulation yields consistent nonparametric estimators, particularly when combined with penalization to control smoothness and avoid overfitting.[26]In fractal time series analysis, the Whittle maximum likelihood estimator (MLE) provides a robust method for estimating the Hurst exponent H (0 < H < 1) of fractional Brownian motion (fBm) or its increment process, fractional Gaussian noise (fGn), by leveraging the power-law decay of the spectral density f(\omega) \propto 1/|\omega|^{2H-1} at low frequencies.[27] The estimation equates to a log-periodogram regression, where the negative log-likelihood is minimized over H using the observed periodogram values against the theoretical spectrum, often implemented via numerical optimization in tools like MATLAB.[27] This semiparametric approach excels in capturing monofractal scaling behavior and has been detailed in recent tutorials for practical application.The Whittle MLE proves particularly effective for scaling analysis in physiological signals, such as heart rate variability (HRV), where it quantifies long-range correlations indicative of autonomic nervous system dynamics or disease states using datasets like those from PhysioNet.[27][28] For short time series prone to bias from finite-sample effects or non-stationarities, debiasing techniques—such as differencing to remove trends and adjusting H estimates—enhance accuracy without sacrificing the method's asymptotic efficiency.[27] Compared to wavelet-based estimators for the Hurst exponent, the Whittle approach offers superior computational speed for large sample sizes (e.g., n > 2^{15}), as its frequency-domain evaluation scales linearly with the number of periodogram ordinates, avoiding the multiresolution decompositions required in wavelet methods.[29]
Signal Detection
In signal detection tasks, the Whittle likelihood facilitates hypothesis testing through the likelihood ratio statistic \Lambda = 2 [ L_W(\theta_1) - L_W(\theta_0) ], where L_W(\theta) denotes the Whittle log-likelihood under parameter vector \theta. Under the null hypothesis H_0: \theta = \theta_0, this statistic asymptotically follows a chi-squared distribution with degrees of freedom equal to the difference in dimensionality between \theta_1 and \theta_0, enabling threshold-based decisions for detecting deviations from a baseline model.[30]A prominent application arises in detecting a known deterministic signal embedded in stationary Gaussian noise, where the alternative hypothesis posits a spectral density f(\omega) = |S(\omega)|^2 + \sigma^2, with S(\omega) as the Fourier transform of the signal and \sigma^2 the noise variance. The null hypothesis corresponds to pure noise (|S(\omega)|^2 = 0), and the Whittle likelihood ratio test assesses signal presence by maximizing the likelihood under each hypothesis, often yielding the matched filter as the optimal detector when the noise spectrum is known. This framework underpins robust detection in domains like gravitational wave analysis, where it maximizes signal-to-noise ratios for transient events.[31][32]The Whittle likelihood has been employed in electroencephalography (EEG) for detecting oscillatory components indicative of neural activity, such as alpha or theta rhythms amid background noise. By fitting local spectral models via penalized Whittle approximations, it identifies significant periodicities associated with cognitive states or pathologies like seizures, offering nonparametric flexibility for nonstationary brain signals.[33]For change-point detection in piecewise stationary series, adaptive variants segment the time series into locally stationary blocks, approximating the overall likelihood as a product of segment-specific Whittle likelihoods to identify abrupt spectral shifts. Recent extensions in network physiology apply this to multivariate physiological signals, such as cardiorespiratory interactions, enabling detection of regime changes during health transitions without strong parametric assumptions.[34][6]Monte Carlo simulations demonstrate that Whittle-based likelihood ratio tests achieve superior type I error control compared to time-domain generalized likelihood ratio tests, particularly for high-frequency signals where frequency-domain approximations capture rapid oscillations more effectively. These studies highlight improved power and finite-sample robustness in nonstationary settings.[30]