Fact-checked by Grok 2 weeks ago

Additive model

An additive model is a statistical model in which the of the response variable Y given the predictors \mathbf{X} = (X_1, \dots, X_p) is expressed as E[Y \mid \mathbf{X} = \mathbf{x}] = \alpha + \sum_{j=1}^p f_j(x_j), where \alpha is an intercept and each f_j is a , univariate of the j-th predictor. This generalizes the classical model by replacing linear terms \beta_j x_j with flexible nonlinear functions f_j(x_j), while maintaining the additive structure that assumes each predictor contributes independently to the response without interactions. Additive models are semiparametric, combining elements (like the additive form) with nonparametric for the functions f_j, which are typically estimated using methods such as splines, kernels, or . The functions are often centered such that E[f_j(X_j)] = 0 for , and estimation proceeds iteratively via the backfitting , which applies univariate smoothers to partial residuals until . This approach provides interpretability, as the effect of each predictor can be examined separately through the estimated f_j, unlike fully nonparametric models that may suffer from the curse of dimensionality. The concept of additive models has roots in earlier work on separable functions in dating back to the mid-20th century, but gained prominence in modern statistics through developments in the for handling nonlinear data while preserving model simplicity. They form the foundation for generalized additive models (GAMs), which extend the framework to response distributions beyond the normal, such as or , via a link function in likelihood-based settings. Additive models are widely applied in fields like environmental modeling, , and bioinformatics for their balance of flexibility and computational efficiency, with software implementations available in languages like through packages such as mgcv.

Overview

Definition and Core Principles

An additive model in statistics is a approach that expresses the of the response as a of smooth s, each depending on a single predictor , plus a . Formally, for a response Y and predictors X_1, \dots, X_p, the model is given by E(Y \mid X_1, \dots, X_p) = \beta_0 + f_1(X_1) + \dots + f_p(X_p), where each f_j(\cdot) is an unspecified smooth capturing the marginal of X_j, and the functions are typically centered such that E[f_j(X_j)] = 0. This formulation approximates the true by its closest additive structure in the least-squares sense, minimizing the expected squared . The core principle of additivity posits that the effects of the predictors on the response are independent, with no interactions between them; thus, the total effect is the simple sum of individual contributions. Each smooth function f_j is estimated from the data using flexible nonparametric techniques, such as splines or smoothers, allowing the model to adapt to nonlinear relationships without assuming a specific parametric form. This semiparametric nature enables the model to reveal complex marginal effects that might be obscured in more rigid frameworks. In contrast to linear regression models, where the effects are strictly linear as E(Y \mid X) = \beta_0 + \beta_1 X_1 + \dots + \beta_p X_p, additive models replace the linear terms \beta_j X_j with arbitrary smooth functions f_j(X_j), providing greater flexibility for capturing nonlinearity while maintaining interpretability through additivity. Basic assumptions include the smoothness of the underlying functions f_j, independence of the additive effects across predictors, and an additive error structure with finite variance, often assuming predictors are bounded (e.g., in [0,1]) and observations are independent. These models can be extended to generalized additive models for non-normal response distributions via a link function.

Historical Development

The concept of additive models has earlier roots in econometrics, dating back to analyses of separable functions by Leontief (1947). The modern statistical formulation traces back to the emergence of nonparametric regression in the 1970s, which sought flexible alternatives to rigid parametric forms for modeling relationships in data. This development drew heavily on foundational kernel smoothing techniques, such as the Nadaraya-Watson estimator proposed independently in 1964, enabling local weighted averaging to approximate unknown regression functions. Spline methods, popularized in the same era through works on smoothing splines, further supported additive structures by allowing piecewise polynomial approximations that maintained computational feasibility in multiple dimensions. A pivotal advancement occurred in 1981, when Jerome H. Friedman and Werner Stuetzle introduced additive models within the framework of projection pursuit regression, specifically as a means to mitigate the curse of dimensionality in high-dimensional datasets by decomposing the regression surface into sums of univariate functions. This approach built directly on nonparametric foundations, offering a balance between flexibility and interpretability without the full complexity of fully nonparametric multivariate smoothing. The concept evolved further with the introduction of generalized additive models (GAMs) by and in 1986, extending the additive structure to accommodate generalized linear models and handle diverse response distributions like or outcomes through a link function applied to the additive predictor. In the 1990s, these models gained traction in contexts, with practical implementations emerging in software such as , which facilitated their application in and predictive modeling. Modern implementations advanced in the through the mgcv, developed by Simon N. Wood, which introduced efficient methods for automatic smoothness selection and mixed-effects extensions of GAMs, making the models more accessible for statistical computing. Key contributions from , Hastie, and Tibshirani are synthesized in their influential 2001 text, The Elements of Statistical Learning, which contextualized additive models within broader and inference paradigms, significantly boosting their adoption across statistics and .

Statistical Foundations

Mathematical Formulation

The additive model provides a flexible framework for by expressing the response variable as a sum of functions of individual predictors, without interactions between them. For a consisting of n observations (Y_i, X_{i1}, \dots, X_{ip}) for i = 1, \dots, n, the model is formulated as Y_i = \alpha + \sum_{j=1}^p f_j(X_{ij}) + \varepsilon_i, where \alpha is term, each f_j is an unknown capturing the effect of predictor X_j, and the errors \varepsilon_i are and identically distributed as \varepsilon_i \sim N(0, \sigma^2). This setup assumes additivity, meaning the expected response E(Y \mid X_1, \dots, X_p) = \alpha + \sum_{j=1}^p f_j(X_j) depends only on univariate contributions from each predictor, eschewing multiplicative or higher-order terms. To represent the smooth functions f_j, basis expansions are commonly employed, such as splines, where f_j(x) = \sum_{m=1}^{M_j} \beta_{jm} b_{jm}(x). Here, b_{jm}(x) denotes the m-th for the j-th predictor (e.g., or truncated power basis), \beta_{jm} are coefficients to be estimated, and M_j is the number of basis terms, chosen to provide sufficient flexibility while controlling . Smoothness of f_j is enforced through a roughness penalty, typically the integrated squared J(f_j) = \lambda_j \int [f_j''(x)]^2 \, dx, where \lambda_j \geq 0 is a smoothing parameter that balances fit and complexity; larger \lambda_j yields smoother functions. For , since the model is invariant to constant shifts across the f_j (as they can absorb into \alpha), a is imposed: \int f_j(x) \, dF_j(x) = 0 or equivalently E[f_j(X_j)] = 0, where F_j is the of X_j. This ensures unique solutions up to the null space of the penalty. Under the Gaussian error assumption, estimation proceeds by minimizing the penalized objective \sum_{i=1}^n \left( Y_i - \alpha - \sum_{j=1}^p f_j(X_{ij}) \right)^2 + \sum_{j=1}^p \lambda_j \int [f_j''(x)]^2 \, dx. This criterion directly corresponds to the negative log-likelihood (up to constants) for the normal , incorporating both data fidelity and penalties. Key assumptions underpinning the model include the additivity of predictor effects, the twice-differentiable of each f_j, and the and homoscedasticity of the errors \varepsilon_i with constant variance \sigma^2. These ensure the model's interpretability and the consistency of estimates under appropriate conditions. For non-Gaussian responses, the formulation extends via generalized additive models with link functions, but the core additive structure remains.

Estimation Methods

The backfitting algorithm is a key iterative procedure for estimating the component functions f_j in an additive model, where each f_j is fitted while holding the others fixed. The process begins with an initial estimate, such as the constant term \alpha = \mathbb{E}(Y), and proceeds by cyclically updating each f_j(X_j) through smoothing the residuals Y - \alpha - \sum_{k \neq j} f_k(X_k) against X_j, using methods like or splines; this alternation continues until convergence. Originally proposed for additive models by and Stuetzle in 1981 and adapted for generalized additive models (GAMs) by Hastie and Tibshirani, the algorithm operates as a contraction mapping in , ensuring convergence to a unique solution under mild conditions on the smoothers. For generalized additive models, the local scoring algorithm extends backfitting to handle non-Gaussian responses via . In each iteration, a working response Z = \eta + (Y - \mu) / (\mu' \cdot g'(\mu)) is constructed, where \eta = \sum f_j(X_j) is the current additive predictor, \mu = g^{-1}(\eta), and g is the link function; this is then smoothed component-wise with weights to update the f_j. Developed by Hastie and Tibshirani, this method generalizes the backfitting approach for distributions, maintaining additivity while approximating the maximum likelihood estimates. Penalized regression approaches estimate the f_j by minimizing a penalized objective \sum (Y_i - \alpha - \sum f_j(X_{ij}))^2 + \sum \lambda_j \int (f_j''(t))^2 dt, often using P-splines or thin-plate splines for the components, with smoothing parameters \lambda_j selected via generalized cross-validation (GCV) or (AIC) to balance fit and smoothness. These can be framed as mixed effects models, solvable via , providing efficient computation for high-dimensional settings. Wood's implementation in the R package mgcv employs this framework for GAMs, integrating automatic smoothness selection. Computationally, these methods scale well to models with many predictors p by avoiding the curse of dimensionality inherent in full nonparametric fits, as each component is estimated marginally; for instance, backfitting requires O(np) operations per cycle for n observations, far preferable to grid-based alternatives. Implementations like the gam package in facilitate practical use, supporting various smoothers and response distributions. Convergence of iterative algorithms like backfitting and local scoring is typically assessed by monitoring the change in component functions, stopping when \|f_j^{\text{new}} - f_j^{\text{old}}\| < \delta for a small tolerance \delta, such as $10^{-3}, or when the relative change in deviance falls below a threshold; theoretical guarantees ensure monotonic decrease in the objective under convex smoothers.

Variants and Extensions

Generalized Additive Models

Generalized additive models (GAMs) extend the additive model framework by integrating the structure of generalized linear models (GLMs) to accommodate response variables with non-normal distributions, such as or data. In this setup, the linear predictor is formulated as \eta = g(\mu) = \alpha + \sum_{j=1}^p f_j(X_j), where \mu = E(Y) represents the of the response Y, g is the monotonic link function (for example, the logit function g(\mu) = \log(\mu / (1 - \mu)) for responses), \alpha is an intercept term, and each f_j is an unspecified smooth univariate function of the covariate X_j. This formulation preserves the additive property while allowing flexible, nonparametric modeling of individual predictor effects on the scale of the linear predictor. Central to GAMs are the univariate smooth functions f_j, which operate on the linear predictor scale to maintain compatibility with GLM assumptions, and estimation via , which relies on maximizing an expected log rather than a full likelihood, enabling robust fitting for distributions where variance depends on the . These smooths are typically estimated using nonparametric techniques like scatterplot smoothers or splines, capturing nonlinear relationships without assuming a specific form. Building on basic additive models, GAMs introduce the link function to transform the response appropriately for non-Gaussian cases. Fitting GAMs involves iterative procedures such as local scoring, which generalizes the (IRLS) algorithm from GLMs by applying weighted smoothing to an adjusted response variable at each step, often combined with backfitting to cyclically update the smooth functions until convergence. The effective for each f_j is quantified as the of the associated smoother S_j, offering a data-driven measure of and model complexity analogous to parametric in GLMs. Augmented backfitting variants enhance this process by incorporating penalties for during iterations. A practical example is the logistic GAM for outcomes, where the probability is modeled as P(Y=1 \mid X) = \frac{1}{1 + \exp(-[f_1(\text{age}) + f_2(\text{nodes})])}, with smooth functions f_1 and f_2 flexibly describing how age and number of positive axillary nodes nonlinearly influence the log-odds of 5-year survival; this approach has been applied to datasets like Haberman's survival data, revealing curved effects not captured by linear terms. Model diagnostics in GAMs adapt GLM techniques for non-normal responses, including plots of Pearson residuals \left( (Y - \mu)/\sqrt{V(\mu)} \right) against fitted values or linear predictors to detect outliers and heteroscedasticity, as well as partial residual plots that add the estimated f_j to residuals excluding that term, aiding assessment of functional form and additivity assumptions.

Functional Additive Models

Functional additive models (FAMs) represent an extension of additive models to functional data analysis, where predictors are infinite-dimensional objects such as curves, spectra, or trajectories observed over a continuum, rather than discrete scalars. In the scalar-on-function setting, the model is typically formulated as Y_i = \alpha + \sum_{j=1}^p \int f_j(X_{ij}(t)) \, dt + \epsilon_i, where Y_i is a scalar response, X_{ij}(t) denotes the j-th functional predictor for the i-th observation, and f_j are smooth unknown functions defined over the functional domain. A more general form for functional responses allows for varying coefficient structures, expressed as Y_i(t) = \alpha(t) + \sum_{j=1}^p \int \beta_j(t, X_{ij}(s)) \, ds + \epsilon_i(t), accommodating interactions between the argument t and the predictor values. These formulations leverage tools from functional data analysis (FDA) to handle the inherent smoothness and dependence in functional predictors. Estimation in FAMs proceeds by first representing the functional predictors X_j(t) via dimension reduction techniques, such as functional principal component analysis (FPCA) or basis expansions (e.g., B-splines or Fourier bases), to project them onto a finite set of scores \xi_{jk}. An additive model is then fitted to these scores using nonparametric smoothers, like local polynomials or splines, yielding E(Y | X) = \alpha + \sum_k f_k(\xi_k), where the f_k capture nonlinear effects. For generalized responses, penalized iteratively reweighted least squares (P-IRLS) with functional penalties, such as tensor-product splines, enforces smoothness and identifiability, with smoothing parameters selected via criteria like generalized cross-validation (GCV). This approach integrates seamlessly with existing GAM software frameworks adapted for functional inputs. FAMs find application in analyzing high-frequency data, such as growth curves to predict health outcomes, spectral data in , or trajectories in (e.g., diffusion tensor imaging for cognitive scores). For instance, they model associations between time-varying risk factors like blood pressure trajectories and scalar disease status. Compared to scalar generalized additive models, FAMs better accommodate the serial dependence within each functional predictor, exploiting FDA techniques like FPCA to mitigate the curse of dimensionality while preserving interpretability through additive components. Despite these strengths, FAMs face challenges from computational demands, as functional integrations and high-dimensional smoothing require substantial resources, particularly for large datasets or complex interactions. Additional hurdles include ensuring through regularization to avoid and handling issues in bivariate forms, often resolved via centering constraints on basis coefficients.

Applications

In Regression Analysis

Additive models provide a flexible for by allowing each predictor to enter the model through a , nonlinear function, enabling the capture of complex relationships without assuming a specific form. This approach is particularly useful for predictive tasks where linear assumptions fail, such as in , where the of wages can be modeled as the of functions of years of experience and level, along with linear terms for other factors like . By decomposing the response into additive components, these models improve predictive accuracy while maintaining interpretability compared to fully nonparametric alternatives. A prominent case study involves environmental modeling of concentrations, where daily maximum levels in from the are regressed on meteorological variables using a (GAM) of the form = f_1() + f_2() + f_3() + ε. The smooth functions f_1, f_2, and f_3 reveal nonlinear patterns, such as increasing nonlinearly with up to a and decreasing with higher speeds due to effects; this model was fitted via backfitting and demonstrated superior fit to linear alternatives. Model selection in additive regression involves testing the significance of each smooth component f_j to determine variable inclusion, commonly via an approximate that compares the fit with and without the term to assess if it deviates significantly from zero. The smoothing parameters, which govern the wiggliness of each f_j, are chosen to optimize a criterion like generalized cross-validation (GCV), balancing goodness-of-fit against and directly influencing the model's predictive performance and smoothness of estimated effects. For interpretation, the marginal effects of predictors are visualized through plots of the estimated smooth functions f_j, illustrating how the response changes with each while averaging over the others; these plots, often called partial effect plots, facilitate understanding of nonlinearities, such as accelerating then plateauing returns to in wage models. An empirical illustration appears in the 1970s Boston housing dataset, where an additive model for median house values uncovers nonlinear effects in variables such as the average number of rooms and the proportion of units occupied by lower status residents, along with negative linear effects of crime rates and pupil-teacher ratios on prices, highlighting spatial economic patterns.

In Time Series Decomposition

In time series analysis, the decomposes a observed series Y_t into three main components: Y_t = T_t + S_t + I_t, where T_t represents the trend component capturing the smooth, long-term progression or decline in the data, S_t denotes the seasonal component reflecting periodic patterns that repeat over known intervals, and I_t is the irregular or component encompassing random, non-systematic fluctuations. This assumes that the seasonal variations maintain a constant regardless of the trend level, making it suitable for series where the magnitude of does not scale with the overall level of the data. Estimation in classical additive decomposition typically begins with smoothing the series to isolate the trend using a centered of order $2m+1, where m is half the seasonal period, to eliminate short-term fluctuations while preserving the long-term direction. The detrended series, obtained by subtracting the estimated trend from Y_t, is then averaged over each seasonal cycle to extract the seasonal component, with the remaining residuals forming the irregular component after further adjustment to ensure the seasonal estimates sum to zero over each period. A more flexible extension is the STL (Seasonal-Trend decomposition procedure based on Loess) algorithm, which employs locally weighted (LOESS) for robust, iterative fitting of the trend and seasonal components, allowing for non-constant seasonal periods and handling outliers through repeated cycles of smoothing and deseasonalizing. Developed by , , McRae, and Terpenning, STL iteratively refines the by first applying LOESS to estimate the trend, then extracting from the residuals, and finally computing the , with parameters tunable for and robustness. An illustrative example is the monthly international airline passenger dataset from 1949 to 1960, analyzed by Box and Jenkins, where an additive decomposition reveals a clear upward trend and annual seasonality, but the fit is suboptimal compared to a multiplicative model because the seasonal amplitude increases proportionally with the rising trend in passenger volumes. In contexts where seasonality scales with the trend—such as in growing economic series—the additive model may fail to capture the proportional variations, necessitating a switch to multiplicative decomposition Y_t = T_t \times S_t \times I_t for better representation.

Advantages and Limitations

Key Advantages

Additive models provide enhanced interpretability compared to fully nonlinear methods like neural networks, as each component function f_j(x_j) isolates the marginal effect of an predictor, facilitating clear and understanding of how variables influence the response without entangled interactions. This additive structure preserves the transparency of linear models while allowing for more nuanced insights into predictor behaviors. A key strength lies in their flexibility to model nonlinear relationships through nonparametric functions, avoiding the rigid assumptions of linear models and thereby reducing in scenarios where linearity does not hold. Unlike black-box alternatives, this approach balances expressiveness with simplicity, enabling effective capture of complex patterns such as monotonic increases or thresholds in predictor effects. In terms of dimensionality, additive models handle moderate numbers of predictors (e.g., 10-20 variables) more effectively than fully nonparametric techniques by enforcing additivity, which sidesteps the curse of dimensionality and maintains stable estimates even as the predictor space grows. This property makes them suitable for real-world datasets where interactions are sparse or unnecessary. Computationally, the backfitting procedure for estimating additive models is efficient, achieving linear O(n) in sample size n, which outperforms the higher-order costs of full multivariate smoothing methods. Empirical evidence supports these benefits, with studies showing that generalized additive models (GAMs) yield lower mean squared error (MSE) than generalized linear models (GLMs) in simulations involving misspecified nonlinear relationships. For example, in Hastie and Tibshirani's analysis of prognosis data, adding smooth terms to the model reduced deviance by 5.6 units relative to linear specifications, uncovering important nonlinear effects in age and tumor size.

Common Limitations

One primary limitation of additive models, including generalized additive models (GAMs), is the strict of additivity, which posits that the effects of predictors on the response variable are independent and can be expressed as a sum of individual smooth functions without interactions. This can lead to biased estimates if the true underlying involves non-additive effects or interactions between predictors, as the model fails to capture synergies or antagonisms among variables. For instance, in scenarios where predictor interactions are biologically or economically meaningful, such as in ecological modeling, the additivity constraint may result in oversimplified representations that underestimate complexity. Although the original backfitting could be computationally intensive for large datasets and many predictors, often demanding substantial and time—for example, fitting full GAMs to datasets exceeding observations was historically infeasible on standard hardware due to numerical complexities in non-parametric —modern software implementations, such as the bam() in R's mgcv package, have significantly improved , allowing efficient fitting to very large datasets (e.g., millions of observations) on standard hardware. Recent developments, including doubly algorithms (e.g., DSGAM, 2023) and variational approximations, further enhance for applications. Additionally, in high-dimensional settings where p is large relative to n, additive models exhibit degraded statistical performance and increased variance, limiting their applicability without sparsity-inducing extensions. Additive models are also prone to , especially when smoothing parameters are selected via methods like generalized cross-validation without additional penalties, leading to overly flexible fits that capture noise rather than signal. This risk is heightened in low-sample-size regimes, where insufficient may prevent reliable of functions, or when extrapolating beyond the range, as the non-parametric components lack inherent mechanisms for handling unseen values. Proper regularization, such as penalizing effective , is essential to mitigate these issues but requires careful tuning.

References

  1. [1]
    [PDF] Additive Models - Statistics & Data Science
    the generalization is to use the additive model for things like the ...
  2. [2]
    [PDF] Generalized Additive Models
    Abstract. Likelihood-based regression models such as the normal linear regression model and the linear logistic model, assume a linear (or some.
  3. [3]
    Additive Model - an overview | ScienceDirect Topics
    Additive models have a long history in statistics and econometrics. They were discussed already by Leontief (1947), who analyzed so-called separable functions. ...
  4. [4]
    Additive Regression and Other Nonparametric Models - Project Euclid
    ... 1985 Additive Regression and Other Nonparametric Models. Charles J. Stone · DOWNLOAD PDF + SAVE TO MY LIBRARY. Ann. Statist. 13(2): 689-705 (June, 1985). DOI ...
  5. [5]
    Projection Pursuit Regression - jstor
    Friedman and Stuetzle: Projection Pursuit Regression 821 at the San Jose measuring station. Three projections were accepted. Figures 2a through 2c show the ...
  6. [6]
    Generalized Additive Models - Project Euclid
    We introduce the class of generalized additive models which replaces the linear form ∑βjXj ∑ β j X j by a sum of smooth functions ∑sj(Xj) ∑ s j ( X j ) . The sj ...
  7. [7]
    Generalized Additive Models - 1st Edition - D.R. Cox - T.J. Hastie - R
    In stock Free deliveryThis book describes an array of power tools for data analysis that are based on nonparametric regression and smoothing techniques.
  8. [8]
    The Elements of Statistical Learning - SpringerLink
    They are prominent researchers in this area: Hastie and Tibshirani developed generalized additive models and wrote a popular book of that title.
  9. [9]
    [PDF] Linear Smoothers and Additive Models Andreas Buja; Trevor Hastie
    Apr 1, 2008 · The additive model and its normal equations. The additive model (15) ... penalized least-squares criterion (20) Q(f) = 11 y - f +112 + Cf ...
  10. [10]
    [PDF] 1986, Vol. 1, No. 3, 297–318 - Generalized Additive Models
    This paper is nontechnical for the most part, with an emphasis on the techniques and their illustration through examples. In Section 2, we review the linear.
  11. [11]
    GAMs with integrated model selection using penalized regression ...
    This paper presents the basic mathematical and numerical approach to GAMs implemented in the R package mgcv, and includes two environmental examples.
  12. [12]
    Convergence of the backfitting algorithm for additive models
    The backfitting algorithm is an iterative procedure for fitting additive models in which, at each step, one component is estimated keeping the other ...
  13. [13]
    Functional Generalized Additive Models - PMC - PubMed Central
    We introduce the functional generalized additive model (FGAM), a novel regression model for association studies between a scalar response and a functional ...
  14. [14]
    [PDF] Functional Additive Models - UC Davis Statistics
    The combination of functional principal component score predictors and additive models therefore emerges as a particularly natural and flexible data-adaptive.
  15. [15]
    [PDF] Generalized Additive Model Selection - Trevor Hastie
    ... additive model setting. There have ... p j=1, setting wi = 0.25. Inner loop: Run the coordinate descent procedure of §4.1 on the penalized least squares.
  16. [16]
    3.4 Classical decomposition | Forecasting: Principles and ... - OTexts
    There are two forms of classical decomposition: an additive decomposition and a multiplicative decomposition. These are described below for a time series with ...
  17. [17]
    [PDF] STAT481/581: Introduction to Time Series Analysis - UNM Math
    There are two forms of classical decomposition: an additive decomposition and a multiplicative decomposition. Let m be the seasonal period. 25. Page 28 ...
  18. [18]
    [PDF] STL: A Seasonal-Trend Decomposition Procedure Based on Loess
    Abstract: STL is a filtering procedure for decomposing a time series into trend, time series into trend, seasonal, and remainder components. STL has a simple ...
  19. [19]
    Evaluation of GLM and GAM for estimating population indices from ...
    GAMs have been shown to fit inherently nonlinear relationships with lower prediction error than GLMs (Austin et al., 2006; Li et al., 2011; Moisen and Frescino, ...
  20. [20]
    [PDF] Comparative performance of generalized additive models and ...
    have been unable to fit GAM models to full datasets because of memory limitations imposed by the numerical complexities of this technique. This prompted our ...
  21. [21]
    General additive models address statistical issues in diffusion MRI
    A second limitation is the propensity of a GAM to overfit the data, particularly when using a generalized cross-validation method. Fortunately, we can place a ...