Unscented transform
The unscented transform (UT) is a deterministic sampling method for approximating the effect of a nonlinear transformation on a random variable characterized by its mean and covariance, enabling the propagation of probability distributions through nonlinear functions without explicit linearization. Developed as a core component of the unscented Kalman filter (UKF), it selects a minimal set of sigma points—typically 2n + 1 points for an n-dimensional state—that fully capture the first two moments (mean and covariance) of the distribution, applies the nonlinear function to these points, and reconstructs the transformed mean and covariance via weighted averages.[1] Introduced by Simon J. Julier and Jeffrey K. Uhlmann in the late 1990s, the UT emerged as a solution to the limitations of the extended Kalman filter (EKF), which relies on first-order Taylor series approximations that often lead to inaccuracies, biases, and numerical instabilities in highly nonlinear systems. The foundational work appeared in their 1997 paper presenting a novel extension of the Kalman filter, where the UT serves as the mechanism for sigma-point propagation to achieve equivalent performance to the linear Kalman filter while generalizing seamlessly to nonlinear dynamics.[1] Subsequent refinements, including scaling parameters for arbitrary dimensions and higher-order moment capture, were detailed in their 2002 and 2004 publications, solidifying the UT's role in modern nonlinear estimation theory.[2][3] Mathematically, for a random variable x with mean μ_x and covariance P_xx, the UT generates sigma points χ_i using χ_0 = μ_x and χ_i = μ_x + (√((n + κ)P_xx))_i (and symmetrically for negative roots), where κ is a scaling parameter; these points are then transformed via y = f(x) to yield Y_i = f(χ_i), with the output mean μ_y = ∑ W_i^m Y_i and covariance P_yy = ∑ W_i^c (Y_i - μ_y)(Y_i - μ_y)^T, using weights W_i tuned for second-order accuracy. This approach achieves at least third-order accuracy for Gaussian inputs (compared to the EKF's first-order) and scales cubically with state dimension, making it computationally efficient for moderate n.[1][3] The UT's advantages include its Jacobian-free implementation, which simplifies coding and avoids derivative-related errors, as well as robustness to discontinuities and non-Gaussian noise, outperforming the EKF in scenarios with strong nonlinearities. Extensions like the scaled UT address high-dimensional challenges, while variants incorporate square-root decompositions for numerical stability in ensemble filters.[2][3] In applications, the UT underpins the UKF for real-time state estimation in fields such as autonomous vehicle navigation, target tracking, spacecraft reentry prediction, and sensor fusion, where it has demonstrated superior performance in propagating uncertainties through complex dynamics like polar-to-Cartesian coordinate transformations. Its influence extends to particle filters, multilevel estimation, and hybrid methods, establishing it as a cornerstone of probabilistic robotics and control engineering.[3]Introduction
Definition and Overview
The unscented transform (UT) is a deterministic sampling method for approximating how a nonlinear function affects a random variable characterized by its mean and covariance. It operates by selecting a small, carefully chosen set of sigma points that fully capture the first two moments (mean and covariance) of the input distribution, typically modeled as Gaussian. These sigma points are propagated through the nonlinear function, and the transformed points are then used to compute weighted estimates of the output mean and covariance, yielding a Gaussian approximation of the propagated distribution. This approach provides a direct way to handle nonlinear transformations without relying on series expansions or derivatives.[4] A core distinction of the UT lies in its use of deterministic rather than stochastic sampling; unlike Monte Carlo techniques, which draw numerous random samples and suffer from high variance unless oversampled, the UT selects points systematically to achieve accurate moment matching with minimal computational cost—specifically, 2n + 1 points for an n-dimensional state. The method assumes a Gaussian input but excels in preserving the mean and covariance up to second-order terms, offering higher accuracy for mildly nonlinear systems compared to first-order linearization methods. Importantly, the UT imposes no differentiability requirement on the nonlinear function, making it robust for applications where analytical derivatives are unavailable or unreliable.[4] In essence, the basic workflow involves generating sigma points around the input mean scaled by the covariance, applying the nonlinear transformation to each point, and recovering the output moments via predefined weights that ensure unbiased estimates. Originating in the late 1990s, the UT was developed primarily to address challenges in nonlinear state estimation, providing a simpler and more reliable alternative to traditional approximation techniques in filtering and prediction tasks.[4]Historical Development
The unscented transform was first proposed in the mid-1990s by Simon J. Julier and Jeffrey K. Uhlmann as a deterministic sampling method to improve upon the linearization approximations used in the extended Kalman filter (EKF) for nonlinear state estimation, particularly motivated by observations of EKF's suboptimal performance in nonlinear state estimation tasks such as those in robotics and navigation.[5] Their initial work, presented at the 1995 American Control Conference, introduced a set of carefully chosen sigma points to propagate mean and covariance through nonlinear functions, offering higher-order accuracy without requiring analytical Jacobians.[5] This approach stemmed from Uhlmann's PhD research at the University of Oxford, where he sought alternatives to EKF's sensitivity to strong nonlinearities.[5] A pivotal milestone came in 1997 with the publication of their seminal paper, which formalized the unscented transform and integrated it into the unscented Kalman filter (UKF) framework, demonstrating superior performance over EKF in benchmark nonlinear problems.[6] Julier played a key role in developing the sigma point scaling mechanisms, while Uhlmann emphasized the practical motivations from real-world filtering challenges.[5] The name "unscented" was chosen arbitrarily by Uhlmann, inspired by a deodorant label he saw while working late one evening, to avoid naming the method after himself, such as the "Uhlmann filter."[5] Subsequent refinements in the early 2000s introduced tunable parameters, such as alpha for spread control and kappa for higher-moment matching, enhancing flexibility for diverse applications.[2] Early adoption focused on navigation and control systems, but post-2000 extensions broadened its scope, including adaptations for non-Gaussian distributions via the unscented particle filter and methods to handle high-dimensional states through scaled sigma points.[7][2] By the 2010s, the transform saw refinements for robustness in discontinuous systems and convergence guarantees.[8] Post-2020 developments have increasingly incorporated the unscented transform into machine learning for uncertainty propagation in neural networks and policy optimization, such as in unscented autoencoders for variational inference and expansion-compression variants for reinforcement learning.[9][10]Theoretical Foundations
Probability Distributions and Moments
The multivariate Gaussian distribution, also referred to as the multivariate normal distribution, is a fundamental probability distribution in multiple dimensions, fully parameterized by its mean vector \mu \in \mathbb{R}^n and positive semi-definite covariance matrix P \in \mathbb{R}^{n \times n}.[11] The mean vector \mu represents the central tendency or expected value of the random vector, while the covariance matrix P quantifies the variance along each dimension and the covariances between dimensions, capturing linear dependencies among the variables.[12] This parameterization assumes Gaussianity, a common approximation in estimation problems like tracking and sensor fusion, where real-world uncertainties are modeled as Gaussian to enable tractable computations despite potential deviations from true normality.[13] Moments provide a systematic way to characterize the shape and properties of probability distributions beyond the mean and variance. The first moment corresponds to the mean vector, indicating the location of the distribution, while the second central moment yields the covariance matrix, describing the dispersion and correlations.[11] Higher-order moments, such as the third (linked to skewness, measuring asymmetry) and fourth (linked to kurtosis, measuring tail heaviness and peakedness), offer insights into deviations from symmetry and normality that influence the distribution's behavior.[12] These higher moments are particularly relevant for nonlinear effects, as transformations can induce non-zero skewness or excess kurtosis, altering the distribution in ways not captured by mean and covariance alone.[14] For multivariate Gaussians, odd-order central moments above the first are zero, resulting in zero skewness, and the kurtosis is invariant, which simplifies analysis under Gaussian assumptions but underscores limitations when higher moments evolve under nonlinearity.[15] For linear transformations of random variables, the moments of a multivariate Gaussian propagate exactly, preserving the Gaussian form. Consider a random vector x \sim \mathcal{N}(\mu_x, P_x) subjected to a linear transformation y = H x, where H is an m \times n matrix. The resulting distribution is y \sim \mathcal{N}(\mu_y, P_y), with \begin{align*} \mu_y &= H \mu_x, \\ P_y &= H P_x H^T. \end{align*} [16] This exact propagation forms the basis for uncertainty handling in linear systems. In filtering applications, such as the Kalman filter for dynamic state estimation, the mean-covariance pair encapsulates the predicted state and its uncertainty, enabling recursive fusion of predictions and measurements to refine estimates of system behavior under Gaussian noise assumptions.[13]Challenges of Nonlinear Transformations
Nonlinear transformations distort the mean and covariance of probability distributions in ways that linear approximations cannot capture accurately. When a random variable with a given mean and covariance passes through a nonlinear function, the resulting distribution often exhibits biases due to the introduction of higher-order terms, such as skewness and kurtosis, which alter the propagated moments beyond simple affine mappings.[17] This distortion arises because nonlinearities amplify asymmetries in the distribution, leading to systematic errors in state estimation tasks like tracking or sensor fusion.[18] The extended Kalman filter (EKF) addresses nonlinearity through local linearization via Taylor series expansion, but this approach is limited by first-order truncation errors that neglect higher-order derivatives. These omissions cause the approximated mean and covariance to deviate from their true values, particularly in systems with strong nonlinearities, where the bias can accumulate over iterations.[17] Moreover, computing the required Jacobians—first-order partial derivatives of the nonlinear functions—poses significant challenges, as they are analytically intensive or infeasible for high-dimensional or complex models, increasing implementation complexity and potential for errors.[17] In practice, these limitations manifest as filter divergence, where the estimated covariance becomes underestimated, failing to reflect the true uncertainty and leading to overconfident predictions that drift from reality.[17] The EKF is also highly sensitive to initial conditions; poor starting estimates can exacerbate linearization errors, causing rapid degradation in performance, as seen in examples where small offsets in state variables result in large estimation discrepancies.[17] Additionally, the reliance on differentiability excludes applications involving non-differentiable functions, such as those with discontinuities or piecewise definitions common in real-world models like obstacle avoidance or switching dynamics.[19] From a statistical viewpoint, propagating distributions through nonlinearities is inherently underdetermined, as infinitely many probability distributions can share the same mean and covariance, yet higher moments are crucial for accurate transformation. This ambiguity necessitates moment-matching approximations, but traditional methods like the EKF often fail to preserve even the first two moments reliably under nonlinearity, motivating advanced techniques for better fidelity.[17]Core Mechanism
Sigma Point Generation
The unscented transform approximates the effect of a nonlinear function on a random variable characterized by its mean \mu and covariance matrix P by propagating a carefully selected set of deterministic sample points, known as sigma points, through the function. These sigma points are generated to exactly capture the mean and covariance of the input distribution, enabling higher-order accuracy without requiring derivatives. In the standard formulation for an n-dimensional state, the symmetric sigma point set consists of $2n + 1 points.[1][3] The central sigma point is placed at the mean: \chi_0 = \mu The remaining $2n points form symmetric pairs around the mean, offset along the principal axes scaled by the square root of the covariance: \chi_i = \mu + \left( \sqrt{(n + \lambda) P} \right)_i, \quad i = 1, \dots, n \chi_{i+n} = \mu - \left( \sqrt{(n + \lambda) P} \right)_i, \quad i = 1, \dots, n Here, \left( \sqrt{(n + \lambda) P} \right)_i denotes the i-th column (or row) of the matrix square root. This construction ensures the sample mean and covariance of the sigma points match \mu and P exactly.[1][3] The scaling parameter \lambda controls the distribution of the sigma points and influences the approximation of higher-order moments: \lambda = \alpha^2 (n + \kappa) - n where \alpha \in (0, 1] determines the spread of the point set relative to the true distribution (smaller \alpha concentrates points closer to the mean), and \kappa is a secondary parameter typically set to $0 or $3 - n to incorporate information about higher moments. For Gaussian inputs, setting n + \lambda = 3 (e.g., \alpha = 1, \kappa = 3 - n) ensures the transform is accurate up to the third order in the Taylor series expansion of the mean and second order for the covariance. Values of \lambda \geq 0 are preferred to maintain numerical stability, as negative \lambda can lead to points that are too close together.[2][3] The matrix square root \sqrt{(n + \lambda) P} is computed using the Cholesky decomposition, which factors P = LL^T where L is lower triangular, yielding \sqrt{P} = L for positive semi-definite P. This method is numerically efficient and stable, avoiding issues with symmetric but indefinite factorizations, though care must be taken when \lambda < 0 by adjusting the scaling to preserve positive semi-definiteness.[3][20] While the symmetric $2n + 1 set is the most widely used due to its balance of accuracy and simplicity, variants exist to reduce computational cost or handle specific distributions. Minimal sigma point sets employ only n + 1 points, such as spherical simplex configurations, which place points on a hypersphere to match the mean and covariance while minimizing the number of samples. These are selected when computational resources are limited and the input is Gaussian, as they capture the first two moments exactly under appropriate scaling, providing at least second-order accuracy for Gaussian inputs, though they may not achieve the third-order accuracy of the symmetric set. The choice between minimal and symmetric sets depends on the need to approximate higher moments: the symmetric set better preserves odd moments like skew for Gaussians, ensuring third-order accuracy, whereas minimal sets prioritize efficiency for second-order fidelity.[3][21]Propagation and Moment Recovery
The propagation step in the unscented transform involves applying the nonlinear function f to each of the $2n + 1 sigma points \chi_i, where n is the dimension of the input random variable, yielding the transformed points y_i = f(\chi_i) for i = 0, \dots, 2n.[3] This deterministic mapping propagates the sigma points through the nonlinearity, capturing the effects of the transformation on the distribution without requiring analytical derivatives or linearization.[3] To recover the statistical moments of the output distribution, a set of weights is applied to the transformed points. The mean weights are defined as W_m^{(0)} = \lambda / (n + \lambda) for the central point and W_m^{(i)} = 1 / [2(n + \lambda)] for i = 1, \dots, 2n, where \lambda = \alpha^2 (n + [\kappa](/page/Kappa)) - n incorporates scaling parameters \alpha and \kappa.[3] The output mean is then computed as the weighted average: \mu_y = \sum_{i=0}^{2n} W_m^{(i)} y_i. The covariance weights differ only for the central point: W_c^{(0)} = W_m^{(0)} + (1 - \alpha^2 + \beta), with W_c^{(i)} = W_m^{(i)} otherwise, where \beta accounts for prior knowledge of the distribution's higher-order moments (e.g., \beta = 2 for Gaussian inputs).[3] The output covariance is recovered via: P_y = \sum_{i=0}^{2n} W_c^{(i)} (y_i - \mu_y)(y_i - \mu_y)^T. For joint transformations involving input-output cross-terms, the cross-covariance is similarly obtained as P_{xy} = \sum_{i=0}^{2n} W_c^{(i)} (\chi_i - \mu_x)(y_i - \mu_y)^T.[3] In applications such as filtering, where process and measurement noises are present, the unscented transform handles additivity by treating noises separately to maintain computational efficiency and accuracy. Additive noise assumptions allow the noise covariances to be directly added to the propagated state covariance without inclusion in the sigma points, avoiding the need for full state augmentation in every step.[3] This approach is particularly useful when noises are uncorrelated with the state, ensuring the second-order accuracy of the moment estimates.[3]Implementation Details
Parameter Tuning
The unscented transform employs three key tunable parameters—\alpha, \kappa, and \beta—to balance accuracy in approximating moments of transformed distributions, particularly for nonlinear functions. The parameter \alpha (where $0 < \alpha \leq 1) controls the spread of sigma points around the mean, with smaller values concentrating points closer to the mean to better capture highly nonlinear transformations, while larger values increase spread for more linear cases.[22] The secondary scaling parameter \kappa influences the overall distribution of points and is often set to 0 for simplicity or to $3 - n (where n is the state dimension) to achieve third-order accuracy for Gaussian inputs in low dimensions.[22] Meanwhile, \beta incorporates prior knowledge about the distribution's higher moments, with a value of 0 assuming no prior information and 2 optimizing for Gaussian distributions by better matching kurtosis.[22] Tuning these parameters involves guidelines that emphasize their impact on bias-variance trade-offs and higher-moment fidelity. Typically, \alpha ranges from $10^{-3} to 1, where low values (e.g., near $10^{-3}) reduce bias in mean and covariance estimates for strong nonlinearities but may increase variance by under-sampling the distribution's tails, while higher \alpha enhances robustness to multimodality at the cost of potential overestimation in linear regimes.[22] Adjustments to \kappa and \beta fine-tune this: positive \kappa expands the point cloud to improve higher-order moment recovery, but excessive values can amplify numerical errors, and \beta = 2 prioritizes fourth-order accuracy for Gaussian assumptions, though deviations may be needed for heavy-tailed distributions to minimize kurtosis mismatch.[22] Overall, tuning seeks to minimize approximation error in propagated moments, with sensitivity analyses revealing that \alpha most strongly affects mean bias in nonlinear settings, while \kappa and \beta dominate covariance and higher-moment stability.[23] Empirical rules provide practical starting points, such as the widely adopted defaults \alpha = 0.001, \kappa = 0, and \beta = 2, which offer a conservative spread suitable for many state estimation tasks while ensuring positive weights for the central sigma point.[24] These defaults stem from simulations showing robust performance across moderate nonlinearities, though application-specific sensitivity analysis—e.g., via Monte Carlo evaluations of mean squared error—is recommended to adjust for dimensionality or nonlinearity degree, often yielding \alpha reductions for high-dimensional problems.[23] Advanced considerations include avoiding negative weights, which can arise if \lambda = \alpha^2 (n + \kappa) - n < 0 leads to non-positive central weight W_0^{(m)} = \lambda / (n + \lambda), potentially causing indefinite covariance matrices; this is mitigated by enforcing \kappa \geq 0 or scaling adjustments to ensure all weights remain non-negative.[25] Refinements in the scaled unscented transform introduce dimension-dependent choices, such as scaling \alpha inversely with n in high-dimensional systems to counteract the effects of high dimensionality on sigma-point spread, preserving accuracy without negative weights through modified weight normalization.[2]Numerical Stability and Computation
The unscented transform incurs a computational complexity of O(n^3) per propagation, dominated by the Cholesky decomposition of the n \times n covariance matrix and the evaluation of weighted outer products across $2n+1 sigma points to recover the transformed covariance.[3] This scaling aligns with that of the extended Kalman filter, rendering the method efficient for low-dimensional states where n < 20, but less viable for high-dimensional problems without optimizations.[3] Key stability concerns arise from ill-conditioned covariance matrices, which can distort sigma point generation and lead to inaccurate moment matching, especially in high dimensions.[25] Square root computations, typically via Cholesky factorization, are prone to numerical errors under finite-precision arithmetic, amplifying inaccuracies in the propagation step.[3] Additionally, non-positive definite covariances P pose issues, as the transform presupposes positive definiteness; suboptimal tuning parameters (e.g., \beta < 0) may yield negative weights, producing unphysical covariance estimates.[25] Mitigation strategies include the scaled unscented transform (SUT), which incorporates auxiliary scaling to enhance matrix conditioning and ensure positive semidefiniteness, often at no extra computational cost.[2] For robust factorization, alternatives to direct Cholesky—such as QR decomposition or singular value decomposition—can prevent breakdown in near-singular cases, though they may increase overhead. Floating-point precision can be bolstered by employing double-precision arithmetic, regularization of small eigenvalues (e.g., adding \epsilon I with \epsilon \approx 10^{-12}), or square-root filter formulations that avoid explicit covariance inversion.[25] Implementations are supported in established libraries, such as Python's FilterPy, which provides a dedicatedunscented_transform function for sigma point propagation and moment recovery.[26] MATLAB's Control System Toolbox offers the unscentedKalmanFilter object, facilitating integration with nonlinear estimation workflows.[27] Furthermore, the independence of sigma point evaluations enables parallelization, which can substantially reduce propagation time on multi-core systems for nonlinear function calls.[28]
Theoretical Properties
Optimality Analysis
The unscented transform (UT) exhibits third-order optimality for Gaussian input distributions transformed through nonlinear functions, where it exactly matches the mean and covariance of the output distribution up to the third-order terms in the Taylor series expansion. This level of accuracy arises because the symmetrically chosen sigma points capture not only the first two moments but also the third-order central moments of the input Gaussian, enabling precise propagation without linearization errors inherent in first-order approximations.[6][29] In general, given only the mean and covariance of the input distribution, the UT represents one of infinitely many deterministic sampling schemes that exactly reproduce these moments using a minimal set of 2n + 1 sigma points, with the symmetric configuration minimizing higher-order approximation errors for symmetric distributions like Gaussians by preserving odd central moments up to the third order. Taylor series analysis of the propagation step demonstrates that the resulting errors are confined to fourth-order and higher terms. However, in high dimensions, the fixed number of sigma points may fail to adequately represent the input distribution, leading to increased approximation errors.[30][6] Despite these properties, the UT's optimality is limited for vector-valued outputs, where fourth-order errors emerge in the covariance due to uncaptured cross-derivative terms in the multivariate Taylor expansion, and the analysis assumes the nonlinear function is analytic to justify the series validity.[6]Comparison to Alternative Methods
The unscented transform (UT) offers distinct advantages over the extended Kalman filter (EKF) by avoiding the explicit computation of Jacobians, thereby mitigating linearization errors that can lead to biased mean and covariance estimates in highly nonlinear systems. In contrast to the EKF's first-order Taylor approximation, the UT achieves at least third-order accuracy for means and second-order for covariances through deterministic sigma-point sampling, resulting in empirically lower mean squared error (MSE) in simulations of nonlinear dynamics, such as radar tracking or localization tasks.[31][32] These gains stem from the UT's ability to capture higher-order effects without derivative calculations, making it more robust in scenarios prone to EKF divergence.[33] Compared to Monte Carlo methods and particle filters, the UT provides a deterministic approximation using a minimal set of sigma points (typically 2n+1 for n-dimensional states), enabling faster computation without the randomness or resampling overhead of stochastic sampling. While particle filters excel in handling non-Gaussian distributions through adaptive particle numbers, the UT's fixed-sample approach yields higher accuracy for low-dimensional Gaussian assumptions at a fraction of the cost—often comparable MSE to Monte Carlo simulations but with orders-of-magnitude lower runtime in uncertainty propagation tasks like state estimation.[34] However, the UT is less flexible for strongly multimodal posteriors, where particle methods can better represent diverse hypotheses.[35] Relative to Taylor series-based methods, such as higher-order expansions or statistical linear regression, the UT serves as an efficient sample-based alternative that implicitly incorporates equivalent higher-order terms without requiring explicit Hessian or derivative computations, which escalate rapidly in dimensionality.[36] Statistical linear regression, a first- or second-order Taylor variant, demands analytical linearization and can suffer from increased computational burden for higher moments, whereas the UT maintains O(n^3) complexity while matching or exceeding accuracy in moment propagation for smooth nonlinearities.[37] This makes the UT preferable in real-time applications where derivative-free operation is essential, though Taylor methods may offer tighter bounds in analytically tractable cases.[38] In recent developments post-2020, the UT contrasts with deep learning surrogates like neural networks for nonlinear propagation, where data-driven models can approximate complex transformations but require extensive training data and lack interpretability. The UT's model-based, parameter-light design ensures transparency and no overfitting risks, outperforming black-box neural propagators in low-data regimes or when physical models are available, as seen in hybrid filters combining UT with neural enhancements for tasks like motion prediction.[39] However, neural surrogates may surpass the UT in capturing highly irregular nonlinearities from large datasets, though at the expense of generalization outside trained conditions.[40]Practical Examples
Two-Dimensional Transformation
To illustrate the unscented transform in a low-dimensional setting, consider a two-dimensional Gaussian random variable representing Cartesian coordinates (x, y) with mean \bar{\mathbf{x}} = \begin{bmatrix} 12.3 \\ 7.6 \end{bmatrix} and covariance matrix P_{\mathbf{x}\mathbf{x}} = \begin{bmatrix} 1.44 & 0 \\ 0 & 2.89 \end{bmatrix}. This distribution is propagated through a nonlinear transformation to polar coordinates, defined by the functions r = \sqrt{x^2 + y^2} (radial distance) and \theta = \tan^{-1}(y/x) (angle in radians, assuming the first quadrant for simplicity). This example demonstrates how the unscented transform approximates the mean and covariance of the output distribution \mathbf{y} = [r, \theta]^\top. The unscented transform generates a set of $2n + 1 = 5 sigma points for n=2, using the scaling parameter \lambda = [1](/page/1) (derived from defaults such as \alpha = 1, \kappa = [0](/page/0), yielding \lambda = \alpha^2(n + \kappa) - n = [1](/page/1)). The sigma points \mathcal{X}^{(i)} are computed as \mathcal{X}^{(0)} = \bar{\mathbf{x}} and \mathcal{X}^{(i)} = \bar{\mathbf{x}} + \sqrt{n + \lambda} \cdot \mathbf{P}_{\mathbf{x}\mathbf{x}}^{1/2}_{j} (and the negative for i = n+[1](/page/1) to $2n), where \mathbf{P}_{\mathbf{x}\mathbf{x}}^{1/2} is the Cholesky decomposition (diagonal here: \sqrt{1.44} = [1.2](/page/1.2), \sqrt{2.89} = [1.7](/page/1.7)) and \sqrt{n + \lambda} = \sqrt{3} \approx [1.732](/page/1.732). The resulting sigma points (rounded to two decimals) are:| Index i | \mathcal{X}^{(i)}_x | \mathcal{X}^{(i)}_y |
|---|---|---|
| 0 | 12.30 | 7.60 |
| 1 | 14.38 | 7.60 |
| 2 | 12.30 | 10.54 |
| 3 | 10.22 | 7.60 |
| 4 | 12.30 | 4.66 |
| Index i | \mathcal{Y}^{(i)}_r | \mathcal{Y}^{(i)}_\theta (rad) |
|---|---|---|
| 0 | 14.459 | 0.554 |
| 1 | 16.262 | 0.486 |
| 2 | 16.202 | 0.708 |
| 3 | 12.737 | 0.640 |
| 4 | 13.153 | 0.364 |