Fact-checked by Grok 2 weeks ago

Confirmatory factor analysis

Confirmatory factor analysis (CFA) is a multivariate statistical used within to test whether the covariances or correlations observed among a set of measured variables align with a predefined theoretical model of underlying latent factors. Unlike , which identifies potential factor structures from data without prior hypotheses, CFA evaluates the adequacy of a researcher-specified model, including the number of factors, their interrelationships, and the loadings of observed variables onto those factors. This method accounts for measurement error and provides goodness-of-fit indices, such as the comparative fit index (CFI) and error of approximation (RMSEA), to assess how well the model reproduces the sample data. CFA was pioneered by Karl G. Jöreskog in the 1960s as an extension of traditional , introducing a general maximum likelihood approach to confirm hypothesized structures in psychometric and research. Building on early work in dating back to Karl Pearson's contributions in 1901, Jöreskog's framework formalized the testing of a priori models, enabling rigorous validation of theoretical constructs like or attitudes. By the 1970s and 1980s, CFA became integral to software such as LISREL and EQS, facilitating its widespread adoption in fields including , , and health sciences. In practice, CFA involves specifying a model, estimating parameters via methods like maximum likelihood under assumptions of multivariate , and interpreting fit to confirm or refine the . It is commonly applied to validate instruments, such as surveys assessing primary healthcare , where it confirms whether items load onto intended dimensions like timeliness or . Key advantages include its ability to handle correlated factors and test invariance across groups, though challenges like model misspecification or small sample sizes can affect reliability, necessitating careful theoretical grounding and sample sizes typically exceeding 200.

Overview

Definition and Purpose

Confirmatory factor analysis (CFA) is a multivariate statistical method used to test whether a set of observed variables reflects a hypothesized structure of underlying latent factors, as specified by a researcher to . Developed as a hypothesis-testing approach, CFA evaluates the extent to which the support preconceived theoretical models, allowing researchers to confirm or reject assumptions about factor structures, loadings, and inter-factor relationships. This technique is widely applied in to validate instruments, in sciences to assess theoretical constructs such as attitudes or behaviors, and in to refine survey scales for consumer insights. The core components of a CFA model include latent variables, also known as , which represent unobservable constructs inferred from the data; observed indicators, which are the measured variables that serve as proxies for these ; factor loadings, interpreted as coefficients indicating the strength of the relationship between a factor and its indicators; unique terms capturing measurement unreliability or specificity; and factor covariances, which describe associations between multiple latent . In contrast to exploratory methods, CFA imposes a priori constraints on these elements to test specific theoretical predictions. CFA models are often visualized using path diagrams, where latent factors are depicted as circles or ovals, observed indicators as rectangles or squares, single-headed arrows from factors to indicators representing factor loadings, single-headed arrows from indicators to error terms for unique variances, and double-headed arrows between factors for covariances. For instance, in a simple one-factor model assessing , multiple observed test items—such as scores, tasks, and arithmetic problems—would load onto a single latent factor representing general cognitive ability, with the diagram illustrating how these indicators converge to measure the underlying while accounting for errors.

Historical Development

The roots of confirmatory factor analysis (CFA) lie in the broader development of , which began with Charles Spearman's introduction of the concept in 1904 to explain correlations among cognitive abilities through a single general factor, known as the g factor. In the 1930s, Louis L. Thurstone extended this framework by proposing multiple orthogonal factors to account for diverse mental abilities, shifting away from Spearman's unitary model and laying groundwork for more complex latent variable structures. CFA emerged as a distinct hypothesis-testing approach in the late 1960s, primarily through the work of Karl G. Jöreskog at the . Jöreskog's 1967 contributions introduced maximum likelihood methods for with specified constraints, enabling the testing of predefined factor structures rather than exploratory derivation. This was formalized in his 1969 paper, which presented a general confirmatory maximum likelihood procedure for factor models, marking the first explicit use of the term "confirmatory factor analysis." In the 1970s, Jöreskog integrated CFA into the (SEM) framework, developing the LISREL software with Dag Sörbom to estimate latent variable models via covariance analysis. This advancement built on by allowing researchers to specify and test relationships between observed indicators and latent factors within larger causal models. The widespread adoption of CFA accelerated in the 1980s, driven by computational improvements in personal computers and accessible software like LISREL, which facilitated its use in for validating instruments. By the 1990s, CFA gained prominence in through user-friendly tools such as , developed by James L. Arbuckle and integrated into , enabling graphical model specification and broader application in scale development and construct validation. Post-2000, advancements in Bayesian CFA addressed limitations of traditional , such as handling non-normality and small samples, with methods incorporating prior distributions for robust inference in complex datasets. In the 2020s, CFA has expanded to contexts, such as integrating with word embeddings for measurement models in textual data, leveraging computational power for high-dimensional analyses in fields like and social sciences.

Exploratory Factor Analysis

Exploratory factor analysis (EFA) is a data-driven statistical method used to uncover the underlying factor structure of a set of observed variables without preconceived hypotheses about the number or nature of the factors. It aims to identify latent constructs by examining patterns of correlations among variables, typically through techniques such as principal components analysis, which accounts for total variance, or principal axis factoring, which focuses on shared (common) variance among variables. These methods, originating from early psychometric work, decompose the correlation matrix to reveal potential dimensions, with factor extraction often guided by criteria like eigenvalues greater than 1 for retaining factors. In contrast to confirmatory factor analysis (CFA), which is theory-driven and tests specific, predefined models—such as a fixed number of factors and constrained factor loadings—EFA adopts an inductive, exploratory approach where all variables are permitted to load on all factors, and the structure emerges from the data. While CFA imposes a hypothesized simple structure to validate theoretical constructs, EFA allows unrestricted loadings and relies on rotation methods (e.g., varimax) to achieve interpretability, making it suitable for initial scale development or when little is known about the data's dimensionality. This fundamental difference underscores EFA's role in generation versus CFA's emphasis on testing. A common workflow involves using EFA to identify potential factor structures in an initial dataset, followed by CFA to confirm those structures in a separate validation sample, ensuring generalizability and reducing overfitting risks. EFA's advantages include its flexibility for discovering unanticipated patterns and lower demands on sample size (typically 5-10 cases per variable), but it suffers from subjectivity in decisions like factor retention and potential for unstable solutions without cross-validation. In comparison, CFA provides rigorous validation but requires stronger theoretical justification, larger samples (often 200+), and careful model specification to avoid identification issues. For instance, applying EFA to a battery of personality questionnaire items might reveal the five broad factors of the model (e.g., , ) through emergent loading patterns, whereas CFA would subsequently test whether those items conform to a prespecified structure with cross-loadings constrained to zero. This sequential approach leverages EFA's exploratory power to inform CFA's confirmatory rigor, enhancing the reliability of psychometric instruments.

Structural Equation Modeling

Structural equation modeling (SEM) represents a comprehensive multivariate statistical framework that combines confirmatory factor analysis (CFA) as the measurement model with path analysis as the structural model to test hypothesized relationships among unobserved latent variables and their observed indicators. This integration allows researchers to account for measurement error in observed variables while examining complex causal structures among latent constructs, such as those in psychological, sociological, or economic theories. Within SEM, CFA serves as the foundational measurement submodel, specifying how observed variables load onto underlying latent factors and assessing the validity of this factor structure. SEM extends beyond CFA by incorporating a structural component that hypothesizes directed paths or regressions between latent factors, enabling tests of theoretical relationships like mediation or moderation among constructs. In contrast, standalone CFA evaluates only the adequacy of the measurement model without specifying inter-factor relations, whereas SEM fully articulates both measurement and structural components to validate comprehensive theoretical models. A practical illustration of this distinction appears in organizational research, where CFA might first confirm latent factors of (measured by items on pay, supervision, and ) and turnover intention (measured by items on intent to quit and job search behavior), after which the structural model tests whether job satisfaction negatively predicts turnover intention. This approach, which unified CFA's factor analytic tradition with path analysis for , was pioneered by Karl Jöreskog through the development of the LISREL framework in the 1970s, providing the first general method for estimating covariance structures in linear models with latent variables. Like CFA, SEM relies on several key assumptions, including linearity in the relationships between latent and observed variables, as well as the absence of omitted variables that could bias parameter estimates or lead to specification errors. Violations of these assumptions, such as nonlinear effects or unmodeled confounders, can compromise the validity of inferences drawn from the model.

Model Specification

Statistical Model

The confirmatory factor analysis (CFA) model posits a linear relationship between a set of observed variables and underlying latent factors. The general structural equation is expressed as \mathbf{Y} = \boldsymbol{\Lambda} \boldsymbol{\xi} + \boldsymbol{\varepsilon}, where \mathbf{Y} is a p \times 1 vector of observed variables, \boldsymbol{\Lambda} is a p \times k factor loading matrix linking the observed variables to the latent factors, \boldsymbol{\xi} is a k \times 1 vector of latent factors, and \boldsymbol{\varepsilon} is a p \times 1 vector of unique error terms. Under this formulation, the error terms are assumed to have zero , E(\boldsymbol{\varepsilon}) = \mathbf{0}, and a diagonal , \text{Cov}(\boldsymbol{\varepsilon}) = \boldsymbol{\Theta}, reflecting uncorrelated unique variances across indicators. Similarly, the latent factors are assumed to have zero , E(\boldsymbol{\xi}) = \mathbf{0}, and covariance matrix \text{Cov}(\boldsymbol{\xi}) = \boldsymbol{\Phi}, which captures correlations among factors if they are not orthogonal. The elements of the factor loading matrix \boldsymbol{\Lambda} quantify the direct effects or, in standardized form, correlations between each latent factor and its associated observed indicators; these are theoretically specified, with certain entries fixed at zero to enforce a measurement where indicators load only on designated factors. From these specifications, the model implies a covariance for the observed variables given by \boldsymbol{\Sigma} = \boldsymbol{\Lambda} \boldsymbol{\Phi} \boldsymbol{\Lambda}' + \boldsymbol{\Theta}, where \boldsymbol{\Sigma} is the p \times p population reproduced by the model parameters. The associated with the model are determined by the number of unique elements in the observed minus the number of free parameters to be estimated, yielding df = \frac{p(p+1)}{2} - q, where q includes the nonzero elements of \boldsymbol{\Lambda}, \boldsymbol{\Phi}, and \boldsymbol{\Theta}. For illustration, consider a simple two- CFA model with four observed variables Y_1, Y_2, Y_3, Y_4, where Y_1 and Y_2 load exclusively on the first \xi_1, and Y_3 and Y_4 load exclusively on the second \xi_2, with all cross-loadings constrained to zero. The loading takes the form \boldsymbol{\Lambda} = \begin{pmatrix} \lambda_{11} & 0 \\ \lambda_{21} & 0 \\ 0 & \lambda_{32} \\ 0 & \lambda_{42} \end{pmatrix}, leading to the system of equations Y_1 = \lambda_{11} \xi_1 + \varepsilon_1, Y_2 = \lambda_{21} \xi_1 + \varepsilon_2, Y_3 = \lambda_{32} \xi_2 + \varepsilon_3, Y_4 = \lambda_{42} \xi_2 + \varepsilon_4, and allowing for a nonzero covariance \phi_{12} between \xi_1 and \xi_2.

Identification Conditions

In confirmatory factor analysis (CFA), a model is identified if its parameters—typically the factor loading matrix \Lambda, the factor covariance matrix \Phi, and the error covariance matrix \Theta—can be uniquely estimated from the sample covariances of the observed variables. This ensures that the mapping from data to parameters is one-to-one, allowing for consistent and finite estimates under standard estimation methods like maximum likelihood. Necessary conditions for identification include the t-rule, which requires that the number of unique elements in the sample (i.e., p(p+1)/2, where p is the number of observed variables) is at least as large as the number of free parameters in the model. Additionally, scale identification must be imposed for each latent , as factor models are inherently scale-invariant; this is typically achieved by fixing one loading per factor to 1 or setting the factor variance to 1. These conditions are essential but not sufficient, as they only guarantee that the model is not underidentified in a basic counting sense. Sufficient conditions provide stronger assurances of . For standard CFA models, the three-indicator rule states that a is identified if it is measured by at least three indicators, each loading exclusively on that , with no correlated errors among indicators and unconstrained covariances. For simpler cases, such as models with two indicators per , identification can be achieved if the error covariances are diagonal and there is at least one nonzero between factors. Identification can be local or global. Local identification ensures a unique solution in the neighborhood of the true parameter values, often verified by checking the of the information matrix or using empirical tests like multiple starting values in . Global identification, which implies local, requires that the solution is unique across the entire parameter space and may involve exhaustive checks, such as testing vanishing tetrad differences (products of pairs of covariances expected to be zero under the model). Distinguishing between them is crucial, as local identification may hold while global does not, leading to potential multiple distant solutions. Underidentification occurs when these conditions fail, resulting in non-unique solutions where parameters cannot be distinguished from the , often yielding estimates or improper values like negative error variances (known as Heywood cases). For instance, a with fewer than two indicators typically leads to underidentification, as the loading and error variance become empirically indistinguishable. To address underidentification, analysts impose additional constraints, such as equality restrictions on loadings across groups or fixing cross-loadings to zero, while ensuring they are theoretically justified to avoid model misspecification. These strategies, when applied judiciously, can transform an underidentified model into an one without altering its substantive meaning.

Assumptions and Estimation

Key Assumptions

Confirmatory factor analysis (CFA) relies on several key statistical and theoretical assumptions to ensure valid inference and model interpretation. One fundamental assumption is multivariate normality of the observed variables, which posits that the joint of the indicators follows a . This assumption is particularly critical when using maximum likelihood (ML) estimation, as violations can lead to biased parameter estimates and underestimated standard errors, potentially inflating Type I error rates in tests. For instance, data from Likert-scale items often exhibit non-normality due to their ordinal nature and bounded response options, which can distort factor loadings and covariances if treated as continuous. Another core assumption is the of unique errors, meaning that the error terms (often represented by the diagonal elements of the error Θ) are uncorrelated with each other and with the latent factors. This local independence implies that any observed correlations among indicators are fully explained by the common factors, and violations—such as correlated residuals—may indicate omitted variables, misspecified factors, or unmodeled method effects that require model revision. CFA also assumes a simple factor structure where indicators load primarily on one theorized , with no cross-loadings unless explicitly specified based on prior . This restriction enforces the hypothesized model by constraining non-salient loadings to zero, promoting and interpretability; allowing untheorized cross-loadings can obscure the distinctiveness of factors and complicate model . Adequate sample size is essential for reliable and standard errors in CFA, with guidelines recommending a minimum of 5 to 10 cases per estimated and an overall sample of at least 200 for basic models to achieve and power. Smaller samples risk unstable solutions, especially in complex models with many parameters, while larger ones (e.g., over 400) enhance precision but may not be necessary for well-specified models. Theoretically, CFA requires a priori model specification driven by substantive theory rather than exploratory , ensuring that the , loadings, and covariances are hypothesized in advance to test specific hypotheses. This contrasts with exploratory approaches and ties directly into estimation procedures, where violations of these assumptions may necessitate robust alternatives like to adjust standard errors and fit statistics under non-normality.

Estimation Procedures

Confirmatory factor analysis (CFA) parameters, including factor loadings, unique variances, and factor covariances, are estimated by minimizing a fit function that measures the discrepancy between the observed sample S and the model-implied \Sigma(\theta), where \theta denotes the parameter vector. The most widely used estimation method is maximum likelihood (ML), which assumes multivariate of the observed variables to ensure the consistency and asymptotic efficiency of the estimates. Under this assumption, ML minimizes the fit function F_{\text{ML}}(\theta) = \log|\Sigma(\theta)| + \text{tr}\left[S \Sigma(\theta)^{-1}\right] - \log|S| - p, where p is the number of observed variables, yielding unbiased estimates in large samples. For non-normal or categorical data, where the assumption fails, (WLS) provides an asymptotically distribution-free alternative by weighting the discrepancies according to the inverse of the asymptotic of the sample statistics. The WLS fit function is F_{\text{WLS}}(\theta) = (s - \text{vech}(\Sigma(\theta)))' W^{-1} (s - \text{vech}(\Sigma(\theta))), with s = \text{vech}(S) the vectorized form of the sample covariance matrix and W the weight matrix, often estimated from polychoric correlations for ordinal indicators. Diagonally weighted least squares (DWLS) approximates WLS for computationally intensive large models by using only the diagonal elements of W, reducing estimation time while maintaining reasonable performance for factor loadings and correlations in ordinal data analyses. To address non-normality in ML estimation, robust corrections such as the Satorra-Bentler scaled chi-square adjust the test statistic and standard errors by incorporating a scaling factor derived from the kurtosis of the data, improving the validity of inference in small or non-normal samples. Additionally, bootstrapping generates empirical standard errors and confidence intervals by resampling the data with replacement and re-estimating the model multiple times, offering a non-parametric robust alternative without relying on normality. Estimation typically proceeds iteratively using algorithms like Newton-Raphson or quasi-Newton methods, starting from initial values (e.g., from ) and continuing until , defined by criteria such as a parameter change less than 0.001 or a below a small . For example, in a two- CFA model with three indicators per assessing verbal and spatial abilities, estimation might yield loadings of 0.85, 0.78, and 0.82 for verbal and 0.90, 0.88, and 0.75 for spatial, along with a of 0.45, assuming a sample size of 500 and multivariate normality.

Model Fit Assessment

Absolute Fit Indices

Absolute fit indices assess the degree of correspondence between the observed sample covariance matrix S and the model-implied covariance matrix \Sigma implied by the confirmatory factor analysis model, without comparing to a null or baseline model. These indices provide a direct measure of absolute model fit, focusing on overall discrepancy rather than relative improvement. The chi-squared test statistic (\chi^2) serves as the foundational absolute fit measure in confirmatory factor analysis, testing the null hypothesis that the model-implied covariance matrix \Sigma reproduces the sample covariance matrix S exactly. It is computed as \chi^2 = (N-1) F_{ML}, where N is the sample size and F_{ML} is the minimum value of the maximum likelihood discrepancy function F_{ML} = \log|\Sigma| - \log|S| + \trace(\Sigma^{-1} S) - p, with p denoting the number of observed variables. The degrees of freedom (df) are given by df = p(p+1)/2 - t, where t is the number of freely estimated parameters. A non-significant \chi^2 p-value (typically p > 0.05) suggests that the model fits the data adequately, indicating no significant discrepancy between S and \Sigma. However, the \chi^2 test has notable limitations: it is highly sensitive to sample size, often rejecting well-fitting models in large samples (N > 400) due to its asymptotic nature, and assumes multivariate normality of the data, which can inflate the statistic under violations. To address its power against misspecification, researchers conduct power analyses for the \chi^2 test, estimating the probability of detecting population discrepancies of varying magnitudes given the df and N. The error of approximation (RMSEA) quantifies the discrepancy per degree of freedom, adjusting for model and providing an estimate of fit in the . It is calculated as \text{RMSEA} = \sqrt{ \frac{\chi^2 / df - 1}{N - 1} }, with a 90% often reported to assess precision; values close to zero indicate better approximation of the covariance structure. RMSEA values below 0.06 suggest good fit, while those below 0.08 indicate acceptable fit, particularly when combined with other indices. This index is less sensitive to sample size than \chi^2 and does not assume exact fit, making it suitable for evaluating approximate model adequacy. The standardized (SRMR) measures the average size of standardized residuals between the observed and model-implied covariances, offering a straightforward fit . It is defined as \text{SRMR} = \sqrt{ \frac{1}{t} \sum_{i \leq j} \epsilon_{ij}^2 }, where t = p(p+1)/2 is the number of unique elements in the , and \epsilon_{ij} = (s_{ij} - \sigma_{ij}) / \sqrt{\sigma_{ii} \sigma_{jj}} are the standardized residuals. SRMR values below 0.08 indicate good model fit, as they reflect minimal average discrepancy after standardization for scale. Unlike unstandardized versions, SRMR facilitates comparability across models with differing variable scales. The goodness-of-fit index (GFI) and adjusted goodness-of-fit index (AGFI) evaluate the relative amount of observed s reproduced by the model. GFI is computed as \text{GFI} = 1 - \frac{\sum (s_{ij} - \sigma_{ij})^2}{\sum s_{ij}^2}, representing the proportion of variance and covariance in S accounted for by \Sigma; values above 0.90 suggest good fit. AGFI adjusts GFI for model parsimony using \text{AGFI} = 1 - \frac{ \frac{p(p+1)}{2} (1 - \text{GFI}) }{df}, where p is the number of observed variables, penalizing less models; thresholds mirror GFI at >0.90 for acceptable fit. However, both indices are sensitive to non- in the data and do not perform well under violations of multivariate normality, leading to recommendations against their primary use in favor of more robust alternatives. Guidelines for interpreting absolute fit indices emphasize combining multiple measures for robust evaluation, as no single index is definitive. and Bentler recommend RMSEA < 0.06 and SRMR < 0.08 alongside \chi^2 assessments, while cautioning against reliance on GFI/AGFI to their limitations. These cutoffs, derived from simulations under varied conditions, balance Type I and Type II error rates in model rejection decisions.

Relative Fit Indices

Relative fit indices, also known as incremental or comparative fit indices, evaluate the improvement in model fit relative to a baseline model, typically the null model assuming complete among observed variables with no correlations or factor structure. This comparison quantifies the proportional reduction in lack of fit achieved by the target confirmatory factor analysis model over the baseline, providing a measure of relative adequacy rather than absolute fit. The Normed Fit Index (NFI), introduced by Bentler and Bonett, is calculated as: \text{NFI} = \frac{\chi^2_{\text{null}} - \chi^2_{\text{model}}}{\chi^2_{\text{null}}} where \chi^2_{\text{null}} is the chi-square statistic for the null model and \chi^2_{\text{model}} is that for the target model. Values range from 0 to 1, with higher values indicating better relative fit; Bentler and Bonett suggested that NFI values above 0.90 indicate good fit, while values around 0.80 may suggest acceptable fit in some contexts, though lower values like 0.75 signal poor relative improvement. However, NFI tends to underestimate fit in small samples (e.g., N < 200) due to its sensitivity to sample size and lack of parsimony adjustment, making it less reliable without corroboration from other indices. The Non-Normed Fit Index (NNFI), also known as the Tucker-Lewis Index (TLI), addresses some limitations of the NFI by incorporating a penalty for model complexity through . It is computed as: \text{NNFI} = \frac{ (\chi^2_{\text{null}} / \text{df}_{\text{null}}) - (\chi^2_{\text{model}} / \text{df}_{\text{model}}) }{ 1 - \chi^2_{\text{null}} / \text{df}_{\text{null}} } where df denotes . Unlike the NFI, NNFI values can exceed 1 but generally fall between 0 and 1 for well-fitting models; thresholds above 0.95 are considered excellent, 0.90 to 0.95 acceptable, and below 0.90 indicative of poor relative fit. This index favors parsimonious models by rewarding improvements in the per degree of freedom ratio, making it particularly useful for comparing nested models in confirmatory factor analysis. The Comparative Fit Index (CFI), developed by , builds on the NNFI to provide a more robust measure less affected by sample size variations. Its formula is: \text{CFI} = 1 - \frac{ \max(\chi^2_{\text{model}} - \text{df}_{\text{model}}, 0) }{ \max(\chi^2_{\text{null}} - \text{df}_{\text{null}}, 1) } which normalizes the relative reduction in noncentrality while bounding values between 0 and 1. CFI values greater than 0.95 indicate excellent relative fit, 0.90 to 0.95 suggest good fit, and below 0.90 point to inadequate improvement over the baseline; for example, a CFI of 0.97 reflects strong relative performance, whereas 0.85 indicates substantial room for model refinement. Its resilience to sample size makes it one of the most widely recommended relative indices in applications. To account for model parsimony, adjustments such as the modify the NFI by multiplying it by the parsimony ratio, defined as df_model / df_null. Thus, \text{PNFI} = \text{NFI} \times \frac{\text{df}_{\text{model}}}{\text{df}_{\text{null}}} This penalizes overly complex models that achieve fit through excessive parameters, with values above 0.50 often considered acceptable for parsimonious relative fit, though it should be interpreted alongside unadjusted indices. Similar parsimony-corrected versions exist for other relative indices, emphasizing the trade-off between fit and simplicity in model evaluation. In practice, relative fit indices should be used in conjunction with absolute fit measures to provide a comprehensive , as they highlight incremental gains but do not evaluate standalone discrepancy. For instance, a model with NFI = 0.92, NNFI = 0.94, and CFI = 0.96 alongside good absolute fit supports retention, while values like NFI = 0.78, NNFI = 0.82, and CFI = 0.85 may warrant respecification even if absolute indices are marginal.

Applications and Extensions

Practical Applications

Confirmatory factor analysis (CFA) is widely applied in psychology to validate personality scales, such as the Big Five inventory, where it confirms the underlying factor structure of traits like extraversion and neuroticism from self-report items. For instance, CFA has been used to test the unidimensionality of the revised NEO Personality Inventory, supporting its five-factor model in normative samples despite some cross-loading issues. Similarly, in assessing clinical tools like the Minnesota Multiphasic Personality Inventory (MMPI), CFA validates the restructured clinical scales by examining how items load onto latent dimensions of psychopathology, such as emotional dysfunction, in clinical populations. In depression research, CFA evaluates the unidimensionality of measures like the Patient Health Questionnaire-9 (PHQ-9), often revealing deviations from a single-factor model across time points or populations, which informs scale refinement for accurate symptom assessment. In the social sciences, CFA tests latent factors underlying attitudes toward policies, such as political trust in institutions, by confirming measurement equivalence across diverse regimes and enabling comparisons of economic beliefs or dissatisfaction dimensions from survey data. For example, multi-group CFA has identified distinct factors for trust in central versus local governance, supporting policy analysis on public perceptions of economic reforms. In marketing, CFA analyzes survey data to validate constructs like brand loyalty, where items measuring repurchase intention and preference load onto latent factors influenced by brand awareness and image, as demonstrated in studies of consumer behavior in competitive markets like smartphones. In health research, CFA confirms quality-of-life factors in patient-reported outcomes, such as the PROMIS framework, by verifying how physical, mental, and social health domains interrelate through item responses, with models showing improved fit when allowing cross-loadings for symptoms like fatigue. A practical case study in education involves applying CFA to test scores from reading achievement assessments, confirming whether items load onto a single "ability" factor to evaluate construct validity and inform curriculum policy decisions. For instance, in analyzing reading achievement tests for third-year university English majors, CFA supported a one-factor model, validating the scale's use for assessing reading ability and informing instructional strategies. However, applications often face challenges like cross-cultural invariance testing, where CFA assesses configural invariance (same factor structure across groups) and metric invariance (equivalent item loadings), ensuring measures like personality scales are comparable internationally without bias. In the 2020s, recent trends integrate CFA with techniques for handling high-dimensional data, such as textual , where word embeddings serve as indicators in confirmatory models to validate latent constructs in large-scale surveys or analyses. These hybrids enhance scalability for complex datasets, as seen in semi-confirmatory approaches that automate selection in sparse, high-dimensional settings. CFA findings can also extend briefly to for inferring causal relationships among latent variables in applied contexts.

Software and Implementation

Several software packages facilitate the implementation of confirmatory factor analysis (CFA), catering to varying levels of user expertise and analytical needs. Mplus stands out for its flexibility in specifying advanced models, such as multilevel and latent growth analyses, and includes Bayesian estimation options for handling complex data structures. The lavaan package in serves as a robust, free, and open-source tool for CFA and broader , enabling users to fit models with customizable estimation methods like maximum likelihood. , an extension of , provides a user-friendly that allows model specification via path diagrams, ideal for researchers preferring visual over syntactic approaches. LISREL, one of the earliest dedicated programs for , remains a historical standard for covariance-based CFA, though it requires more manual input compared to modern alternatives. Implementing CFA in these software packages follows a structured to ensure reliable results. First, specify the model by defining latent factors and their relationships to observed indicators, either through syntax in programs like lavaan or Mplus, or via diagrams in . Second, input the data as raw observations or precomputed covariance/saturation matrices, depending on the software's capabilities and data characteristics. Third, estimate the model parameters using methods such as maximum likelihood, monitoring for . Fourth, verify model identification by ensuring sufficient constraints (e.g., fixing one loading per factor to 1) and check for estimation to avoid Heywood cases or negative variances. Fifth, assess overall model fit using indices like , RMSEA, and CFI. Sixth, if respecification is needed, add theoretically justified paths or covariances based on modification indices, rather than purely empirical adjustments. Interpreting CFA output focuses on key parameter estimates and diagnostics to validate the measurement model. Standardized factor loadings exceeding 0.40 typically indicate statistically significant and substantively meaningful associations between indicators and latent factors, with values closer to 0.70 or higher reflecting strong . Modification indices, which quantify the expected decrease in if a parameter is freed, can highlight potential improvements, such as residual covariances among errors; however, they must be interpreted cautiously to prevent capitalizing on chance and undermining model generalizability. Best practices in CFA implementation emphasize and robustness to enhance and validity. Always report a comprehensive set of fit indices (e.g., CFI, RMSEA, SRMR) alongside estimates, and justify sample size adequacy, as CFA generally requires at least 200–300 cases for stable , with larger samples preferred for complex models. For , employ full maximum likelihood (FIML) estimation, available in most packages, to utilize all available without listwise deletion . To improve accessibility, free graphical interfaces like and JASP integrate CFA modules, often powered by lavaan, allowing point-and-click model building without coding. Recent software updates continue to expand capabilities; for instance, Mplus Version 9, released in October 2025, introduced enhancements for multilevel and three-level , including applications to CFA. As an example, the following lavaan syntax specifies and fits a simple one-factor CFA model where a latent factor loads on three observed items:
r
model <- 'factor1 =~ item1 + item2 + item3'
fit <- cfa(model, data = your_dataset)
summary(fit, standardized = TRUE, fit.measures = TRUE)
This code fixes the first loading to 1 for , estimates the remaining parameters, and outputs standardized loadings, variances, and fit measures for .

References

  1. [1]
    [PDF] 200-31: Exploratory or Confirmatory Factor Analysis? - SAS Support
    Confirmatory factor analysis (CFA) is a statistical technique used to verify the factor structure of a set of observed variables. CFA allows the researcher ...
  2. [2]
    An Overview of Confirmatory Factor Analysis and Item Response ...
    Confirmatory factor analysis attempts to account for the covariation among items (ignoring error variance), whereas common factor analysis accounts for the “ ...
  3. [3]
    [PDF] Confirmatory Factor Analysis - arXiv
    In confirmatory factor analysis, the researcher first develops a hypothesis about what factors they believe are underlying the used measures and may impose ...<|control11|><|separator|>
  4. [4]
  5. [5]
    [PDF] confirmatory factor analysis and structural equation modeling
    Confirmatory factor analysis (CFA) is used to study the relationships between a set of observed variables and a set of continuous latent variables.
  6. [6]
    A GENERAL APPROACH TO CONFIRMATORY MAXIMUM ...
    A GENERAL APPROACH TO CONFIRMATORY M A v i U T I . . LIKELIHOOD FACTOR ANALYSIS*. J M U M. K. G. JOKESKOG. EDUCATIONAL TESTING SERVICE. We describe a general ...
  7. [7]
    Confirmatory Factor Analysis (CFA) in R with lavaan - OARC Stats
    Purpose. This seminar will show you how to perform a confirmatory factor analysis using lavaan in the R statistical programming language.
  8. [8]
    [PDF] A Primer on Structural Equation Models: Part 1. Confirmatory Factor ...
    As with our path diagram for a single factor model, we have circles for latent variables (i.e., factors and errors), squares for measured variables, single ...
  9. [9]
    Confirmatory Factor Analysis (CFA): A Detailed Overview
    Confirmatory Factor Analysis (CFA) is a sophisticated statistical technique used to verify the factor structure of a set of observed variables.
  10. [10]
    LISREL: A General Computer Program for Estimating a Linear ... - ETS
    Author(s):: Joreskog, Karl G.; Van Thillo, Marielle ; Publication Year: 1972 ; Report Number: RB-72-56 ; Source: ETS Research Bulletin ; Document Type: Report ...Missing: history | Show results with:history
  11. [11]
    Factor analysis and its extensions. - APA PsycNet
    This chapter is an attempt to give an account of the developments of factor analysis as a statistical method in the last 50 years and its extensions from ...
  12. [12]
    Sage Research Methods - Confirmatory Factor Analysis
    Confirmatory factor analysis is a statistical procedure for testing HYPOTHESES about the commonality among VARIABLES.
  13. [13]
    [PDF] IBM® SPSS® Amos™ 26 User's Guide
    IBM SPSS Amos (Analysis of Moment Structures) is an easy-to-use program for visual SEM. With Amos, you can quickly specify, view, and modify your model.
  14. [14]
    A tutorial on the Bayesian approach for analyzing structural equation ...
    This article introduces a Bayesian approach in analyzing structure equation models. This approach can utilize prior information for achieving better ...<|separator|>
  15. [15]
    Evaluating Model Fit in Bayesian Confirmatory Factor Analysis ... - NIH
    The BRMSEA is well suited to evaluate model fit in large sample Bayesian CFA models by taking sample size and model complexity into account.
  16. [16]
    A Practical Introduction to Factor Analysis - OARC Stats - UCLA
    Exploratory factor analysis (EFA) is method to explore the underlying structure of a set of observed variables, and is a crucial step in the scale development ...
  17. [17]
    None
    ### Summary of https://web.pdx.edu/~newsomj/pmclass/EFA%20and%20CFA.pdf
  18. [18]
    [PDF] MULTIPLE FACTOR ANALYSIS - Statistics
    We have described a method of multiple factor analysis. 28. Page 21. 426. L. L. THURSTONE by which it is possible to ascertain how many general, inde- pendent ...
  19. [19]
    [PDF] Comparisons of Exploratory and Confirmatory Factor Analysis - ERIC
    Comparisons of Exploratory and Confirmatory Factor Analysis ... serve as a generalized measure of the reliability of the model (Joreskog & Sorbom, 1986,.
  20. [20]
    An Introduction to Structural Equation Modeling - SpringerLink
    Nov 4, 2021 · Structural equation modeling is a multivariate data analysis method for analyzing complex relationships among constructs and indicators.
  21. [21]
    A Tutorial on Structural Equation Modeling for Analysis of ...
    Structural Equation Modeling (SEM) is a very general approach to analyzing data in the presence of measurement error and complex causal relationships.
  22. [22]
    [PDF] Confirmatory Factor Analysis & Structural Equation Models
    How the latent ...Missing: definition | Show results with:definition
  23. [23]
    Structural Equation Modeling - Statistics Solutions
    Learn how Structural Equation Modeling (SEM) integrates factor analysis and regression to analyze complex relationships between variables.
  24. [24]
    (PDF) Structural equation modelling (SEM): Confirmatory factor ...
    Jun 2, 2017 · PDF | This presentation is intended for the Structural Equation Modeling (SEM) workshop series that I have been conducting since May 2017.
  25. [25]
    Job satisfaction, burnout, and turnover intention among primary care ...
    Jan 15, 2020 · This study aims to examine the relationships between job satisfaction, burnout and turnover intention, and explore the predictors of turnover intention.
  26. [26]
    Structural Equation Modeling: What It Is and When to Use It
    Oct 2, 2024 · Linearity: SEM assumes linear relationships between the latent constructs and the observed variables. It is not suitable for non-linear datasets ...
  27. [27]
    [PDF] Introduction to Structural Equation Modeling Using Stata
    – SEM assumes that no relevant variables are omitted from any equation in the model. – Omitted variable bias can arise in linear regression if ... model.Missing: linearity | Show results with:linearity
  28. [28]
    Statistical Assumptions Underlying Structural Equation Modeling
    This chapter considers the major assumptions associated with structural equation modeling. These include, multivariate normality, completely random missing ...Statistical Assumptions... · Nonnormality · Missing Data · Specification ErrorMissing: linearity | Show results with:linearity
  29. [29]
    Confirmatory Factor Analysis - IDENTIFICATION AND ESTIMATION
    In the context of CFA, identification refers to whether or not it is possible to find unique estimates for all the free and constrained parameters (i.e., ...
  30. [30]
    [PDF] Confirmatory Factor Analysis: Identification and estimation
    Recommendation for identification. 13. • Rule out quickly any model that doesn't pass t-rule. • See if any of the sufficient conditions applies. • Whenever ...
  31. [31]
    A general approach to confirmatory maximum likelihood factor ...
    We describe a general procedure by which any number of parameters of the factor analytic model can be held fixed at any values and the remaining free param.
  32. [32]
    A General Approach to Confirmatory Maximum Likelihood Factor ...
    We describe a general procedure by which any number of parameters of the factor analytic model can be held fixed at any values and the remaining free ...
  33. [33]
    Asymptotically distribution-free methods for the analysis of ... - PubMed
    Asymptotically distribution-free methods for the analysis of covariance structures. Br J Math Stat Psychol. 1984 May:37 ( Pt 1):62-83. doi: ...Missing: CFA | Show results with:CFA
  34. [34]
    Fitting Large Factor Analysis Models With Ordinal Data - PMC
    The term diagonal weighted least squares (DWLS) has been used to broadly define a collection of estimation techniques for ordinal data that use the diagonal of ...
  35. [35]
    [PDF] Scaling Corrections for Statistics in Covariance Structure Analysis
    May 27, 1988 · A family of scaling corrections aimed to improve the chi-square approxi- mation of goodness-of- t test statistics in small samples, ...
  36. [36]
    An Empirical Evaluation of Alternative Methods of Estimation ... - NIH
    Abstract. Confirmatory factor analysis (CFA) is widely used for examining hypothesized relations among ordinal variables (e.g., Likert-type items).Missing: paper | Show results with:paper
  37. [37]
    Chi‐square for model fit in confirmatory factor analysis - Alavi - 2020
    Apr 22, 2020 · There are limitations with using the chi-square statistic as a model fit index. First, it is sensitive to sample size with larger sample sizes ...THE CHI-SQUARE FIT INDEX · LIMITATIONS FOR THE CHI... · SUMMARY AND...
  38. [38]
    [PDF] Cutoff criteria for fit indexes in covariance structure analysis
    Nov 3, 2009 · This article examines the adequacy of the "rules of thumb" conventional cutoff crite- ria and several new alternatives for various fit ...
  39. [39]
    SEM: Measuring Model Fit (David A. Kenny)
    Oct 6, 2024 · Root Mean Square Error of Approximation (RMSEA) This absolute measure of fit is based on the non-centrality parameter. Its computational ...
  40. [40]
    Using the Standardized Root Mean Squared Residual (SRMR) to ...
    SRMR is the square root of the average of squared standardized residual variances and covariances, used to assess how well a model reproduces observed ...
  41. [41]
    [PDF] Some Clarification and Recommendations on Fit Indices
    The adjustments are to penalize models that are less parsimonious, so that simpler theoretical processes are favored over more complex ones. The more complex ...
  42. [42]
    The Big Five Inventory-2: Confirmatory Factor Analysis ... - Redalyc
    In this study, we aimed to address this issue by providing evidence of validity of a recent scale based on this model of personality: the Big Five ...
  43. [43]
    Confirmatory factor analysis of the revised NEO personality inventory
    Results from the CFA with a normative sample of 1000 adults indicated poor fit between the obtained factor structure and the hypothesized dimensions ...
  44. [44]
    The MMPI-2 Restructured Clinical Scales in the Assessment of ... - NIH
    This study examined the psychometric properties of the Minnesota Multiphasic Personality Inventory-2 (MMPI-2) Restructured Clinical Scales (RCSs)
  45. [45]
    Can We Trust Measures of Political Trust? Assessing Measurement ...
    Jul 4, 2016 · A number of studies implement a multiple group confirmatory factor analysis (MGCFA) of political trust models in the European Social Survey (ESS) ...
  46. [46]
    A two-phase confirmatory factor analysis and structural equation ...
    Next, the survey data have been collected and analysed using Confirmatory Factor Analysis (CFA), and the relationship between the constructs has been measured ...
  47. [47]
    Confirmatory Factor Analysis of the Patient-Reported Outcomes ...
    The initial PROMIS domain framework specified that physical function and symptoms such as pain and fatigue indicate Physical Health (PH); depression, anxiety, ...
  48. [48]
    [PDF] A Case Study of a Reading Achievement Test for Third - ERIC
    Results from examining the competing CFA models revealed that the one-factor model was suitable to explain the underlying structure of the reading ...
  49. [49]
    Measurement Invariance Conventions and Reporting: The State of ...
    If configural invariance is supported, the next step is to test for metric invariance, or equivalence of the item loadings on the factors. Metric invariance ...
  50. [50]
    Confirmatory Factor Analysis with Word Embeddings - ResearchGate
    Sep 6, 2025 · To this end, this work introduces a statistical approach that provides tools to implement an informed measurement process on unstructured ...Missing: hybrid | Show results with:hybrid
  51. [51]
    Muthén & Muthén, Mplus Home Page
    Mplus Version 9 includes corrections to minor problems that have been found since the release of Version 8.11 in May, 2024 and the following new features.Version History · Mplus User's Guide Excerpts · Pricing · Mplus Demo Version
  52. [52]
  53. [53]
    Mplus Version History
    Mplus has had seven major version updates since 1998, with the current version being 8. The latest version is 8.11 released May 3, 2024.
  54. [54]
    A CFA example – lavaan.org
    We start with a simple example of confirmatory factor analysis, using the cfa() function, which is a user-friendly function for fitting CFA models.<|control11|><|separator|>