Fact-checked by Grok 2 weeks ago

Scree plot

A scree plot is a graphical tool in , consisting of a line plot that displays the eigenvalues of principal components or factors in decreasing order against their corresponding component numbers, aiding in the determination of the optimal dimensionality of data by identifying how many components capture significant variance. Introduced by psychologist Raymond B. Cattell in 1966 as the "scree test" for , it visualizes the point at which additional components contribute diminishing returns in explained variance, often resembling a of loose rocks—hence the name derived from geological . In (), the plot orders eigenvalues from largest to smallest on the y-axis, with the x-axis representing component indices, allowing analysts to retain components that account for the bulk of the data's variability while discarding noise. The primary interpretation method is the "elbow rule," where one identifies a sharp bend or in the beyond which eigenvalues level off, suggesting minimal additional from further components; for instance, if the plot flattens after the third component, retaining the first three may explain over 80% of variance. This subjective yet widely adopted criterion helps balance model parsimony and explanatory power, though alternatives like parallel analysis or the Kaiser criterion (retaining eigenvalues ≥1) are sometimes used alongside it to reduce ambiguity. plots are implemented in statistical software such as and , and their use extends beyond to , emphasizing their role in for high-dimensional datasets in fields like , , and .

Definition and Purpose

Definition

A scree plot is a graphical representation in , typically depicted as a line plot with the component numbers on the x-axis and the eigenvalues (or singular values) derived from the or of a on the y-axis, arranged in descending . The term "" originates from , where it refers to loose rock debris or that accumulates at the of a steep or cliff, serving here as a for the diminishing eigenvalues that represent variance after the of major principal components. In , the scree plot visualizes the proportion of total variance explained by each successive component, aiding in the assessment of without implying specific interpretive rules. This tool plays a fundamental role in () for by highlighting the relative importance of components based on their associated eigenvalues.

Purpose in Dimensionality Reduction

The scree plot serves as a visual diagnostic tool in techniques, such as () and (), to identify the optimal number of principal components or factors by plotting the eigenvalues in descending order and observing the point where the explained variance begins to level off sharply. This approach enables analysts to retain only those components that capture the majority of the data's variability, thereby simplifying complex datasets while minimizing information loss. By facilitating the selection of a of components that account for substantial variance, the scree plot helps prevent in subsequent modeling tasks, as retaining excessive minor components—often representing random rather than meaningful structure—can lead to models that perform poorly on unseen data. via scree-guided component selection thus promotes more robust generalizations, particularly in high-dimensional settings where dominates lower-variance directions. A key objective of the scree plot is to strike a balance between preserving a high proportion of total variance, typically aiming for 80-90% cumulative explained variance, and maintaining model to enhance interpretability and computational efficiency. This trade-off ensures that the reduced representation captures essential patterns without unnecessary complexity, supporting reliable inference in downstream analyses. In practice, the scree plot is widely applied in for refining latent variable models from with numerous items, and in bioinformatics for distilling profiles involving thousands of variables into manageable dimensions that highlight biological signals.

Construction

Steps to Create a Scree Plot

To create a scree plot from a multivariate , begin by preparing the and following a structured computational process rooted in .
  1. Compute the covariance or correlation matrix of the dataset. Start with a data matrix of p variables and n observations, centering the variables by subtracting their means to place the data cloud at the origin. If the variables have differing scales or units, standardize them (divide by standard deviations) and compute the p \times p matrix; otherwise, use the to capture raw variances. This matrix summarizes the pairwise relationships among variables and serves as the input for .
  2. Perform eigenvalue decomposition on the matrix. Apply to the or to extract the eigenvalues \lambda_1 \geq \lambda_2 \geq \dots \geq \lambda_p and corresponding eigenvectors. These eigenvalues quantify the variance captured by each successive principal component, with the sum of all eigenvalues equaling the total variance in the data (trace of the ).
  3. Plot the eigenvalues against component indices. Arrange the eigenvalues in descending order and create a with the component numbers (1 to p) on the x-axis and the eigenvalue magnitudes on the y-axis. Use a line plot connecting the points or a bar for , which reveals the distribution of variance across components.
  4. Optionally, incorporate cumulative variance or scree ratios. To enhance interpretability, overlay a line showing the cumulative proportion of total variance explained by the first k components (computed as \sum_{i=1}^k \lambda_i / \sum_{i=1}^p \lambda_i) or plot ratios of consecutive eigenvalues (\lambda_k / \lambda_{k+1}) to highlight drops in importance. These additions are not part of the core but aid in assessing dimensionality.
For instance, in a with 10 variables, the scree plot might display eigenvalues such as 5.2, 3.1, 1.4, 0.7, 0.5, and smaller values thereafter, exhibiting a sharp initial drop that tapers off, typical of where the first few components dominate variance explanation.

Mathematical Basis

The mathematical foundation of the scree plot lies in the of the derived from the data, which underpins () and . Consider a with n observations and p variables, where the \Sigma is a p \times p symmetric positive semi-definite . The eigenvalue solves the \Sigma \mathbf{v} = \lambda \mathbf{v}, where \lambda_i (for i = 1, \dots, p) are the eigenvalues and \mathbf{v}_i are the corresponding eigenvectors, forming an that diagonalizes \Sigma as \Sigma = V \Lambda V^T, with V the of eigenvectors and \Lambda the of eigenvalues ordered such that \lambda_1 \geq \lambda_2 \geq \dots \geq \lambda_p \geq 0. These eigenvalues represent the variance explained by each principal component, as the i-th principal component is the of the onto the eigenvector \mathbf{v}_i, capturing \lambda_i units of variance along that direction. The total variance in the dataset equals the of \Sigma, which is the sum of all eigenvalues: \operatorname{trace}(\Sigma) = \sum_{i=1}^p \lambda_i. In the context of a scree plot, these eigenvalues are plotted in decreasing order against their indices to visualize the of variance across components. The proportion of total variance explained by the i-th component is given by \frac{\lambda_i}{\sum_{j=1}^p \lambda_j}, allowing assessment of each component's relative contribution; cumulatively, the first k components explain \frac{\sum_{i=1}^k \lambda_i}{\sum_{j=1}^p \lambda_j}. When using a R instead of the (common for variables with differing scales), the eigenvalues are scaled by the process, where R = D^{-1/2} \Sigma D^{-1/2} with D diagonal containing variances; this adjusts the eigenvalue magnitudes but preserves their ordering and proportional interpretations, though the plot's vertical scale reflects correlations rather than raw variances. This framework assumes linear relationships in the data, as the decomposition operates in the original feature space; for nonlinear structures, kernel variants map data to a higher-dimensional space where linear applies, effectively performing nonlinear decomposition via functions.

Interpretation

Identifying the Elbow

The elbow in a scree plot refers to the point along the eigenvalue curve where the decline bends sharply, shifting from a steep drop-off that captures substantial signal variance to a more gradual, asymptotic flattening indicative of or variation. This visual feature arises because eigenvalues are ordered in decreasing magnitude, with the initial ones representing the principal components that explain the majority of the data's variability. Visually, the scree plot exhibits a series of large eigenvalues at the outset, followed by a rapid decrease that levels off after the first few components, forming the characteristic shape. In many practical applications, this transition often appears after the initial 2 to 4 components, though the exact location depends on the dataset's underlying structure. Analysts inspect the plot for a clear discontinuity in the , where the line segments change from steep to nearly , signaling the retention point for meaningful components. The identification process relies on subjective visual assessment: after constructing the plot with component numbers on the x-axis and eigenvalues on the y-axis, the analyst examines the curve for the most pronounced bend. For instance, consider a hypothetical scree plot with eigenvalues of 10, 8, 3, 1, and 0.5 for the first five components; the sharp drop after the third eigenvalue (from 3 to 1) would indicate the at component 3, suggesting retention of the first three. This method emphasizes qualitative judgment over rigid thresholds, allowing flexibility across diverse datasets. Due to its reliance on human interpretation, the elbow method introduces inter-analyst variability, as different researchers may perceive the bend at slightly varying points on the same . This subjectivity underscores the need for supplementary validation in component retention decisions.

Rules for Component Retention

Several established rules provide objective guidelines for retaining principal components using scree plots, focusing on eigenvalue thresholds and variance contributions to complement visual assessment. The Kaiser criterion, proposed by Henry F. Kaiser, recommends retaining components with eigenvalues greater than 1 when analyzing matrices, as such components account for more variance than an individual variable. This rule assumes that eigenvalues below 1 represent trivial contributions relative to the average variable variance of 1 in a matrix. Another widely adopted approach is the cumulative variance rule, which retains components until they collectively explain a substantial portion of the total variance, typically 70% to 90%, depending on the application's tolerance for information loss. This threshold ensures that the retained components capture the dominant structure in the data while discarding noise, with higher percentages like 90% favored in fields requiring high fidelity, such as . Parallel analysis offers a more robust, simulation-based method by generating eigenvalues from multiple random datasets with the same dimensions as the original data and retaining only those components whose eigenvalues exceed the 95th of the random eigenvalues at each position. Developed by John L. Horn, this technique accounts for sampling variability and performs well in empirical studies across various factor structures. Ian T. Jolliffe proposed a modified eigenvalue , suggesting retention of components with eigenvalues greater than 0.7 times the average eigenvalue (which equals 0.7 for correlation matrices), providing a less stringent alternative to the rule that still filters out minor components. This criterion, introduced in Jolliffe's 1972 work, balances inclusivity and parsimony by retaining components that contribute meaningfully beyond random variation. These rules integrate with the scree plot by allowing researchers to verify cutoffs against the plotted eigenvalues; for instance, the point where the threshold of 1 intersects the plot can align with other criteria to confirm retention decisions, with the elbow serving as a brief complementary visual check.

History and Etymology

Origin of the Term

The term "scree" originates from , where it describes a sloping mass of loose, broken rock fragments, often referred to as talus, that accumulates at the base of a steep cliff or mountain slope due to and . This natural formation represents debris that has slid down from higher elevations, forming a relatively stable but insignificant pile compared to the main geological structure above. In statistics, the term "scree plot" was coined by psychologist Raymond B. Cattell in 1966, drawing directly from this geological metaphor to visualize the eigenvalues in factor analysis. Cattell, renowned for his extensive use of factor analysis in personality psychology to identify underlying traits such as the 16 personality factors, introduced the concept to distinguish significant factors from trivial ones. He analogized the plot's steep initial decline—representing major eigenvalues or substantive factors—to a mountain's face, with the subsequent flat "scree" line depicting minor eigenvalues as rubble or debris at the base, which should be discarded as noise or insignificant influences. In his seminal paper, Cattell explicitly described the scree as "a straight line of rubble and boulders which forms at the pitch of sliding stability at the foot of a mountain," emphasizing its role in guiding the retention of only meaningful components. The term first appeared in Cattell's article "The Scree Test for the Number of ," published in the journal Multivariate Behavioral Research, marking its debut in statistical literature as a practical tool for in psychological . This innovation stemmed from Cattell's ongoing efforts to refine factor extraction methods amid the complexities of measurement, where distinguishing signal from error was paramount.

Development in Statistics

Following its introduction by Raymond B. Cattell in 1966, the scree plot experienced notable adoption in through Ian T. Jolliffe's 1986 book Principal Component Analysis, where it was highlighted as an essential graphical method for assessing the number of meaningful components by plotting eigenvalues in descending order. The marked a pivotal period for the scree plot's standardization, as it was integrated into major statistical software packages such as and the newly developed environment (first released in 1995), which facilitated its routine use in and data analysis workflows. The technique evolved in parallel with methodologies throughout the late , and by the , it had become a standard diagnostic in contexts for guiding and decisions.

Applications

In Principal Component Analysis

In (PCA), the scree plot serves a critical role in the post-decomposition phase by visualizing the eigenvalues associated with each principal component, enabling researchers to identify and retain those components that explain the maximal variance in the . This graphical plots the eigenvalues in descending against the component number, highlighting the relative of each component and aiding in without losing essential information. The typical integrating a scree plot into involves first performing the decomposition on standardized data to obtain the , followed by generating the scree plot to inspect the "" point where the explained variance begins to level off. Researchers then retain the number of components up to this elbow, such as the first three in a where subsequent eigenvalues form a nearly straight line, ensuring the selected components capture a substantial portion of the total variance, often over 80%. This process streamlines analysis by focusing computational and interpretive efforts on the most informative dimensions. A prominent application appears in , where scree plots facilitate the reduction of high-dimensional data—often comprising thousands of —down to dozens of principal components that retain the majority of biological variation. For instance, in analyzing , a scree plot might reveal that the first two components explain over 50% of the variance, allowing researchers to summarize complex expression profiles into manageable scores for downstream tasks like signature validation. In , scree plots are used in to analyze multivariate datasets, such as variables in river systems. For example, in studying the River environmental data, a scree plot can help determine the number of principal components that capture key variations in factors like , nutrients, and pollutants, often retaining the first few components that explain the majority of ecological variance. The scree plot complements biplots in by emphasizing variance distribution across components, whereas biplots overlay variable loadings and observation scores to illustrate relationships between variables and samples. This distinction ensures the scree plot's focus remains on eigenvalue-based selection rather than detailed variable contributions. Furthermore, scree plots aid in addressing in by revealing redundant dimensions through the rapid decline in eigenvalues after initial components, indicating that correlated variables are consolidated into fewer uncorrelated principal components. This helps mitigate issues like inflated variance estimates in models derived from the original multicollinear data, promoting more stable and interpretable results.

In Factor Analysis and Extensions

In factor analysis (FA), the scree plot serves as a visual diagnostic tool to determine the number of common factors by plotting the eigenvalues derived from the correlation matrix, where these eigenvalues reflect the amount of shared variance among observed variables explained by each factor. Unlike principal component analysis, which captures total variance, FA eigenvalues in methods like principal axis factoring incorporate estimated communalities on the diagonal of the reduced correlation matrix, emphasizing latent structures of common variance rather than unique or error components. originally developed the scree test specifically for FA applications, aiming to identify the "elbow" point beyond which additional factors contribute negligibly to explaining intercorrelations among variables. A prominent application of scree plots in FA occurs in psychological research, particularly when analyzing questionnaire data to uncover personality traits. For instance, in exploratory FA of the NEO Personality Inventory, a scree plot reveals an elbow supporting five factors corresponding to the traits (neuroticism, extraversion, , agreeableness, and ), guiding researchers to retain factors that account for substantial shared variance in self-reported behaviors. This approach helps validate latent constructs by ensuring the retained factors represent meaningful psychological dimensions rather than noise. The scree plot extends to advanced FA techniques, such as exploratory (ESEM), where it informs the initial extraction in a framework that allows cross-loadings while integrating structural relations. In confirmatory , the plot provides a preliminary validation of the hypothesized structure by comparing observed eigenvalues against expected patterns from theory-driven models. Additionally, the scree plot guides pre-rotation decisions in , such as determining the number of before applying orthogonal rotations like varimax, which maximizes the variance of squared loadings to enhance interpretability without altering the total shared variance captured.

Limitations and Alternatives

Key Criticisms

One major criticism of the scree plot is its inherent subjectivity in identifying the "," which leads to variability in interpretations across analysts and poor of results. Even when using the average judgments of trained raters, the method exhibits only moderate reliability, as different observers may select different points as the elbow based on . The scree plot is also insensitive to certain data characteristics, such as high levels of or scenarios where eigenvalues are approximately equal, resulting in the absence of a clear elbow and misleading conclusions about dimensionality. In noisy datasets, the plot often fails to distinguish signal from random variation, leading to inconsistent estimates of the number of meaningful components. Simulation studies have demonstrated that the scree plot underperforms relative to more objective methods like parallel analysis, particularly in recovering the true number of components across diverse data structures. For instance, Peres-Neto et al. (2005) evaluated 20 stopping rules in extensive simulations and found the scree test to have lower accuracy, especially when compared to parallel analysis, which consistently outperformed it in terms of recovery rates. Additionally, the scree plot oversimplifies the decision process by ignoring key factors like sample size, performing poorly when n < 100 due to increased sampling variability that obscures eigenvalue patterns. This limitation exacerbates errors in small-sample contexts, where the plot's visual cues become unreliable without adjustments for finite population effects. Cattell's scree has faced dated since the 1970s and 1980s for tending to inflate the number of retained factors, often overestimating dimensionality by one or more components in empirical and simulated . Zwick and Velicer (1986) highlighted this through comparative simulations, showing the scree test's propensity to retain extraneous dimensions compared to statistical alternatives.

Alternative Techniques

Parallel analysis, proposed by , serves as a statistical complement to the scree plot by generating eigenvalues from multiple random matrices of the same dimensions as the observed , typically under the of no structure, and retaining only those components whose observed eigenvalues exceed the 95th of the simulated distribution. This resampling approach provides an objective threshold for dimensionality, outperforming subjective in simulations across various data structures. The Kaiser-Guttman rule offers a simpler , recommending retention of components with eigenvalues greater than 1, on the grounds that such values indicate variance explanation surpassing that of a single variable in the original . Often used alongside plots, this rule assumes an identity under randomness, where eigenvalues equal 1, thus flagging deviations as meaningful structure. Velicer's minimum average partial (MAP) test provides a unique non-graphical alternative by iteratively extracting components and computing the average squared partial among , selecting the number of at which this average reaches its minimum, thereby isolating common variance from unique and error components. This method emphasizes the off-diagonal elements of the residual after partialling out successive , offering robustness in factor analytic contexts without relying on eigenvalue thresholds. To enhance the of traditional plots, bootstrapped variants resample the with to generate intervals around eigenvalues, allowing identification of reliable components through eigenvector stability or eigenvalue variability, as demonstrated in ecological datasets where it refined retention decisions. Jackson's approach, for instance, correlates original eigenvectors with those from bootstrap replicates to assess axis robustness, addressing sampling variability inherent in visual methods. These techniques collectively promote greater objectivity over scree-based decisions; for example, the nFactors integrates analysis, , and other heuristics to yield consensus recommendations on component retention, facilitating automated and replicable analyses in exploratory models.

References

  1. [1]
    Scree Plot - an overview | ScienceDirect Topics
    A scree plot is defined as a graphical representation that displays the residual variance as a function of the number of extracted eigenvectors, ...
  2. [2]
    The Scree Test For The Number Of Factors - Taylor & Francis Online
    (1966). The Scree Test For The Number Of Factors. Multivariate Behavioral Research: Vol. 1, No. 2, pp. 245-276.
  3. [3]
    Interpret all statistics and graphs for Principal Components Analysis
    To visually compare the size of the eigenvalues, use the scree plot. The scree plot can help you determine the number of components based on the size of the ...
  4. [4]
  5. [5]
    Principal Component Analysis: A Method for Determining the ... - NIH
    Hence the name scree (geological debris at the bottom of a cliff) plot has been tied to PCA. Other criteria are commonly used for the choice of essential ...
  6. [6]
    Scree Plot: Determining the Number of Components in Data Analysis
    The term scree plot draws its name from the geological concept of scree, which describes loose rock materials building up against mountains.
  7. [7]
    Scree plot - Principal component analysis (PCA) - Analyse-it
    A scree plot visualizes the dimensionality of the data. The scree plot shows the cumulative variance explained by each principal component.Missing: authoritative sources
  8. [8]
    The Scree Test For The Number Of Factors - PubMed
    The Scree Test For The Number Of Factors. Multivariate Behav Res. 1966 Apr 1;1(2):245-76. doi: 10.1207/s15327906mbr0102_10. Author. R B Cattell. PMID: 26828106 ...Missing: Raymond original paper
  9. [9]
    Does PCA reduce overfitting? - Quora
    Oct 3, 2019 · Though that, PCA is aimed to reduce dimensionality, which leads to a smaller model and possibly reduce the chance of overfitting. So, in case ...
  10. [10]
    Principal Component Analysis (PCA): Explained Step-by-Step | Built In
    Principal component analysis (PCA) is a dimensionality reduction and machine learning method used to simplify a large data set into a smaller set.
  11. [11]
    16.3 - Choosing the number of eigencomponents | STAT 555
    The usual rule of thumb is to select enough PCs to explain over 90% of the variance, or until the scree plot flattens out.<|separator|>
  12. [12]
    How Dimension Reduction works—ArcGIS Pro | Documentation
    The output eigenvalues table comes with a customized line chart, called a scree plot, to show the percent of variance maintained by each component. In the scree ...Missing: bioinformatics | Show results with:bioinformatics
  13. [13]
    Chapter 9 Dimensionality reduction | Orchestrating Single-Cell ...
    Dimensionality reduction aims to reduce the number of separate dimensions in the data. This is possible because different genes are correlated.
  14. [14]
    [PDF] PRINCIPAL COMPONENTS ANALYSIS (PCA)
    Dec 5, 2019 · 3) A scree plot is useful for understanding how variance is distributed among the principal components, and it should be the first step in ...
  15. [15]
    None
    Summary of each segment:
  16. [16]
    [PDF] A QUICK PRIMER ON EXPLORATORY FACTOR ANALYSIS
    The scree plot is a two-dimensional graph with factors on the x-axis and eigenvalues on the y-axis. Eigenvalues are produced by the analysis and represent the ...
  17. [17]
    Nonlinear Component Analysis as a Kernel Eigenvalue Problem
    A new method for performing a nonlinear form of principal component analysis is proposed. By the use of integral operator kernel functions, one can efficiently ...
  18. [18]
    [PDF] A Monte Carlo Study Of The Accuracy Of The Scree Plot
    May 1, 2005 · Hutchinson and Bandalos, (1997) pointed that Monte Carlo studies are commonly used to study the behavior of statistical tests and psychometric ...
  19. [19]
    Principal component analysis: a review and recent developments
    Apr 13, 2016 · Principal component analysis (PCA) is a technique for reducing the dimensionality of such datasets, increasing interpretability but at the same time minimizing ...
  20. [20]
    Parallel Analysis: a method for determining significant principal ...
    PCA eigenvalues from the data greater than PA eigenvalues from the corresponding random data can be retained. All components with eigenvalues below this ...
  21. [21]
    [PDF] Exploratory Factor Analysis=1See last slide for copyright information.
    Scree plots. Testing. 32 / 38. Page 33. Scree Plots. In geology, “scree” is the pile of rock and debris often found at the foot of a mountain cliff or volcano.
  22. [22]
    Mass wasting
    These loose angular rock fragments that accumulate at the base of a slope are referred to as talus or scree. Talus is notoriously unstable to climb down or up, ...Missing: definition | Show results with:definition
  23. [23]
  24. [24]
    [PDF] The Scree Test For The Number Of Factors - Sci-Hub
    Jun 10, 2010 · CONCEPTION OF THE PROBLEM. A brief, easily applicable test for determining the number of factors to extract in factor analytic experiments ...
  25. [25]
    Principal Component Analysis Guide & Example - Statistics By Jim
    Dimensionality reduction​​ In this manner, you avoid overfit models. And, because PCA produces uncorrelated components, it also addresses multicollinearity. ...Principal Components... · Factor Analysis Vs Pca · Reasons To Use Pca<|control11|><|separator|>
  26. [26]
    Characteristics and Validation Techniques for PCA-Based Gene ...
    Feb 6, 2017 · Principal Component Analysis (PCA) can be used to summarize a gene signature into a single score. Our hypothesis is that gene signatures can be ...
  27. [27]
    Principal components analysis in clinical studies - PMC - NIH
    In multivariate analysis, independent variables are usually correlated to each other which can introduce multicollinearity in the regression models.Principal Components... · Figure 2 · Advanced Visualization Tools
  28. [28]
    Factor Analysis: A Short Introduction, Part 1
    The eigenvalue is a measure of how much of the common variance of the observed variables a factor explains. Any factor with an eigenvalue ≥1 explains more ...
  29. [29]
    Evaluating the Evidence for the General Factor of Personality across ...
    Once again, an inspection of the scree plot and the interpretability of several solutions suggested a Big Five structure. As before, the inter-factor ...
  30. [30]
    EXPLORATORY STRUCTURAL EQUATION MODELING IN ...
    Jan 31, 2022 · ESEM should be, then, used to confirm the factor structure that EFA suggested. In cases in which the factor structure is already defined and ...
  31. [31]
    12.6 - Final Notes about the Principal Component Method | STAT 505
    Unlike the competing methods, the estimated factor loadings under the principal component method do not change as the number of factors is increased.<|control11|><|separator|>
  32. [32]
    [PDF] Using Horn's Parallel Analysis Method in Exploratory Factor ... - ERIC
    Parallel analysis is based on random data generation, which is parallel to the actual data set, using the Monte Carlo. Simulation Technique to determine the ...
  33. [33]
    Stopping Rules in Principal Components Analysis: A Comparison of ...
    Dec 1, 1993 · Heuristic procedures included: retaining components with eigenvalues (λ) > 1 (i.e., Kaiser—Guttman criterion); components with bootstrapped λ > ...Missing: original | Show results with:original
  34. [34]
    [PDF] STOPPING RULES IN PRINCIPAL COMPONENTS ANALYSIS
    The scree plot consistently estimated one dimension more than the number of simulated dimensions. Bartlett's test of sphericity showed inconsistent results.