Fact-checked by Grok 2 weeks ago

Studentized range distribution

The Studentized range distribution is the continuous of the studentized range Q, defined as the (maximum minus minimum) of k sample means from independent populations with equal variances and equal sample sizes, divided by the estimated of a single mean, where the for the error estimate is \nu. This arises in the context of comparing multiple means following an analysis of variance (ANOVA), providing critical values for determining significant differences while controlling the . The concept of the studentized range was first introduced by D. Newman in 1939, who derived its distribution for samples from a population and expressed it in terms of an independent estimate of the standard deviation, building on earlier suggestions by "" (William Sealy ). In 1952, M. Keuls extended its application to multiple comparisons in ANOVA, proposing a step-down procedure that uses quantiles from this distribution to test ordered means. John W. Tukey further developed the framework in 1949 for comparing individual means post-ANOVA, and comprehensive tables of the distribution's percentage points were published by H. Leon Harter in 1960, facilitating practical computations. The distribution is central to several multiple comparison procedures, including Tukey's Honestly Significant Difference (HSD) test, which employs a single-step approach using the studentized range to compare all pairwise means while maintaining the overall Type I error rate at a specified level \alpha. It is also integral to the Newman-Keuls test, a more powerful but less conservative step-down method for ordered comparisons. Quantiles from the studentized range distribution depend on three parameters: the number of groups k, the sample size per group n (assuming balanced design), the degrees of freedom \nu, and \alpha, and are tabulated or computed numerically for use in these tests under the assumption of normality and homogeneity of variances. For unbalanced designs, modified procedures such as the Tukey-Kramer method adjust the critical values based on this distribution. Modern statistical software implements these calculations, enabling robust post-hoc analyses in experimental designs.

Definition

Probability density function

The studentized range Q is defined as the range of k independent standard normal random variables divided by an independent estimate of the standard deviation, specifically Q = \frac{R}{S}, where R is the among the k normals and S = \sqrt{\chi^2_\nu / \nu} with \chi^2_\nu following a with \nu , independent of the normals. This arises in the context of k sample means from normal populations with equal sample sizes n, where the range of the means is studentized by the pooled s / \sqrt{n}, with s^2 based on \nu = N - k and N = kn total observations; the distribution of Q is free of n and \sigma^2. The probability density function of Q for q > 0, with parameters k > 1 (number of groups) and \nu > 0 (degrees of freedom), is given by f(q \mid k, \nu) = \frac{k(k-1) \nu^{\nu/2}}{2^{\nu/2 - 1} \Gamma(\nu/2)} \int_0^\infty s^\nu \exp\left( -\frac{\nu s^2}{2} \right) \left[ \int_{-\infty}^\infty \phi(z) \phi(q s + z) \left[ \Phi(q s + z) - \Phi(z) \right]^{k-2} \, dz \right] ds, where \phi and \Phi denote the standard normal and , respectively. Here, k represents the number of independent samples or groups compared, while \nu reflects the associated with the estimate of the variance. This integral form for the PDF is derived from the joint distribution of the order statistics of the k standard normal variables, which yields the marginal density of the range R as f_R(r) = k(k-1) \int_{-\infty}^\infty \phi(z) \phi(r + z) [\Phi(r + z) - \Phi(z)]^{k-2} \, dz for r > 0, reflecting the probability that one observation is at the lower extreme, another at the upper, and the remaining k-2 lie between them. Transforming to the studentized form Q = R / S then involves integrating the product of the range density, the density of S, and the Jacobian factor s, leading to the double integral representation above after substitution.

Cumulative distribution function

The cumulative distribution function (CDF) of the studentized range Q, denoted F(q \mid k, \nu), where k is the number of populations or samples and \nu is the , gives the probability that the studentized range does not exceed the value q. This function is expressed through an representation involving the standard normal \Phi: F(q \mid k, \nu) = \int_0^\infty \left[ k \int_{-\infty}^\infty \phi(u) \left[ \Phi(u + q \sqrt{t}) - \Phi(u) \right]^{k-1} \, du \right] \frac{ (\nu/2)^{\nu/2} t^{\nu/2 - 1} \exp(-\nu t / 2) }{ \Gamma(\nu/2) } \, dt, where \phi is the standard normal probability density function. This form arises from conditioning on the scaled variance estimate t \sim \chi^2(\nu)/\nu and integrating the conditional distribution of the range of k standard normal variables. The CDF is strictly monotonically increasing in q for q > 0, starting from the boundary condition F(0 \mid k, \nu) = 0 since the range is nonnegative, and approaching 1 as q \to \infty. This monotonicity ensures that probabilities accumulate smoothly over the support [0, \infty). In practice, due to the complexity of the double integral, the CDF has historically been evaluated using tabulated values rather than direct computation. The , $1 - F(q \mid k, \nu), represents the upper-tail probability and is particularly relevant in hypothesis testing for multiple comparisons, where it determines the likelihood of observing a exceeding a critical threshold under the . Early evaluations of the CDF faced significant computational challenges, requiring laborious methods on early computers, which prompted the creation of comprehensive tables for various k and \nu values before the advent of efficient numerical algorithms and software.

Quantiles and critical values

The quantile function of the studentized range distribution gives the value q_{\alpha}(k, \nu) such that the probability P(Q > q_{\alpha}(k, \nu)) = \alpha, where Q is the depending on the number of samples k and \nu. This inverse of the is essential for determining thresholds in . Critical values q_{\alpha}(k, \nu) are commonly tabulated for significance levels such as \alpha = 0.05 and \alpha = 0.01, with parameters k ranging from 2 to 10 and \nu from small values like 5 up to . These tables facilitate applications in multiple procedures. For illustration, the following shows selected critical values for \alpha = 0.05:
\nuk=2k=3k=4k=5
53.645.706.987.80
\infty2.773.644.124.40
Similar tables exist for \alpha = 0.01, where values are larger; for example, q_{0.01}(2, 5) = 5.24 and q_{0.01}(2, \infty) = 3.70. As \nu \to \infty, the quantiles q_{\alpha}(k, \nu) converge to those of the range of k independent standard normal random variables, reflecting the stabilization of the denominator in the studentized range. For non-tabulated combinations of k, \nu, and \alpha, linear interpolation between table entries—often with respect to $1/\nu—provides a practical approximation, though more advanced methods like Aitken interpolation may improve accuracy for precise computations.

Properties

Moments and characteristic function

The moments of the studentized range distribution Q_{k,\nu}, where k is the number of populations and \nu is the , are derived by integration with respect to its f(q; k, \nu). Since the density is expressed as a , the moments lack closed-form expressions and are typically computed numerically or approximated asymptotically. The first moment, or , is given by E[Q] = \int_0^\infty q f(q; k, \nu) \, dq, or equivalently for this non-negative , E[Q] = \int_0^\infty [1 - F(q; k, \nu)] \, dq, where F is the . The studentized range can be decomposed as Q = R / V, where R is the range of k i.i.d. standard normal random variables (independent of V), and V = \sqrt{\chi^2_\nu / \nu} follows the distribution of the square root of a scaled chi-squared random variable. This independence allows $1 - F(q; k, \nu) = E[1 - F_R(q V)], where F_R is the cdf of R. The cdf of R has the exact integral form F_R(r) = k \int_{-\infty}^\infty [\Phi(x + r/2) - \Phi(x - r/2)]^{k-1} \phi(x) \, dx, with \Phi and \phi denoting the cdf and pdf of the normal distribution, respectively. Thus, E[Q] admits an exact integral expression involving the cdf, amenable to numerical evaluation. For large \nu, V \to 1 in probability, so the moments of Q approach those of R. In this regime, E[Q] \approx E[R] \approx 2 \sqrt{2 \log k} for large k, derived from applied to the order statistics of the normals. The variance follows from the second moment, \operatorname{Var}(Q) = E[Q^2] - [E[Q]]^2, with E[Q^2] = 2 \int_0^\infty q [1 - F(q; k, \nu)] \, dq. Asymptotically for large \nu and large k, \operatorname{Var}(Q) \approx \operatorname{Var}(R) \approx \pi^2 / (6 \log k), reflecting the Gumbel limiting distribution of the normalized extremes (with variance \pi^2 / 6 for each endpoint, leading to total asymptotic variance \pi^2 / (6 \log k) after adjustment by the scaling factor $1 / \sqrt{2 \log k} for the two endpoints). Higher moments E[Q^r] for r \geq 3 are obtained analogously via E[Q^r] = r \int_0^\infty q^{r-1} [1 - F(q; k, \nu)] \, dq or by of the density. These can also be derived from the \phi(t) = E[e^{itQ}] = \int_0^\infty e^{itq} f(q; k, \nu) \, dq, with the r-th given by i^{-r} \frac{d^r}{dt^r} \phi(t) \big|_{t=0}. Using the , \phi(t) = E_V [\phi_R(t / V)], where \phi_R is the characteristic function of R, which itself integrates over the joint density of the order statistics. This form facilitates analysis in limit theorems, such as convergence to extreme value distributions for varying k or \nu. Asymptotic expansions for moments as k \to \infty or \nu \to \infty build on refinements of the Gumbel approximation, including subleading terms like -\frac{\log \log k + \log (4\pi)}{2 \sqrt{2 \log k}} for the expected maximum.

Special cases and limiting behavior

When k = 2, the studentized range Q reduces to a scaled version of the of a Student's t-distributed with \nu , specifically Q = \sqrt{2} |T| where T \sim t_{\nu}. This equivalence arises because the range between two group means, when standardized by the of a single mean, corresponds to the two-sample t-statistic scaled by \sqrt{2} to account for the of the difference between means. As \nu \to \infty, the studentized range distribution converges to the distribution of the range of k independent standard normal random variables, since the sample standard deviation s converges in probability to the true standard deviation \sigma = 1 (after ). This limiting distribution, often denoted as the range distribution for normal samples, has been tabulated and studied for its use in multiple comparisons under known variance. For large k, implies that Q / \sqrt{2 \log k} \to 2 in probability, reflecting the asymptotic behavior of the range of k standard normals where the maximum deviates approximately \sqrt{2 \log k} from the mean and the minimum by a symmetric amount. This normalization highlights the growth rate of the studentized range as the number of comparisons increases, connecting it to Gumbel-type extreme value distributions after further centering and scaling. At small \nu, such as \nu = 1, the distribution exhibits increased variability with heavy tails resembling those of a scaled Cauchy distribution (the t_1 case for k=2), leading to wider critical values and less precise inferences in multiple comparisons due to the high uncertainty in estimating the error variance. As \nu \to 0, the distribution degenerates to a point mass at zero, as the error degrees of freedom approach insufficiency for variance estimation, rendering the studentization undefined and the range effectively uninformative.

Applications

Multiple comparison tests

The studentized range distribution is fundamental to post-hoc multiple comparison tests in (ANOVA), enabling controlled assessments of pairwise differences among k group means after a significant overall . These procedures leverage the distribution to establish critical differences that balance power and error control in comparing means from normal populations. Tukey's Honestly Significant Difference (HSD) test utilizes the studentized range to evaluate whether observed differences between sample means are statistically significant. The test's rejection region for the of equal population means between groups i and j is defined as |\bar{y}_i - \bar{y}_j| > q_{\alpha}(k, \nu) \cdot \frac{s}{\sqrt{n}}, where q_{\alpha}(k, \nu) denotes the upper \alpha- of the studentized range distribution with k means and \nu for the error mean square, s is the pooled from the ANOVA, and n is the common sample size per group. This formulation scales the range of ordered means by the estimated , providing a conservative for . The HSD test controls the (FWER) at the nominal \alpha level across all \binom{k}{2} pairwise comparisons, meaning the probability of declaring at least one false significant difference when all population means are equal does not exceed \alpha. This strong control is achieved through the joint distribution properties of the studentized range, making the procedure suitable for exploratory analyses where multiple inferences are drawn simultaneously. Duncan's multiple range test extends this framework by using adjusted critical values derived from the studentized range distribution, tailored to the number of means in each subset comparison rather than the full set of k groups. These adjustments yield protection levels that increase with subset size, resulting in a less stringent test than HSD while still basing decisions on the same underlying distribution. Both tests require equal sample sizes n across groups, normality in the underlying populations, and homogeneity of variances to ensure the validity of the pooled error estimate and the distribution's applicability. As an illustrative application, consider a one-way ANOVA involving k=4 treatment groups, each with n=10 observations, where the error \nu = 36 and the pooled standard deviation s=2.5. Following a significant ANOVA result, the HSD critical difference is q_{0.05}(4, 36) \cdot \frac{2.5}{\sqrt{10}} \approx 3.81 \cdot 0.79 \approx 3.01; thus, any pairwise difference exceeding 3.01 units would be deemed significant at \alpha=0.05, allowing researchers to identify specific differing groups while maintaining FWER control.

Analysis of variance extensions

The studentized range distribution extends beyond foundational pairwise multiple comparisons in analysis of variance (ANOVA) to more general scenarios involving linear contrasts and complex experimental designs. In these extensions, the distribution provides s for constructing simultaneous inference procedures that control the while accommodating broader sets of hypotheses. Simultaneous intervals for all linear contrasts can be formed using the studentized range statistic applied to normalized combinations of treatment means, specifically for contrasts of the form \sum c_i \bar{y}_i where \sum c_i^2 = 1. The half-width of such intervals is given by q_{k, \nu, \alpha} \sqrt{\text{MSE}/n} / \sqrt{2}, where q_{k, \nu, \alpha} is the from the studentized range distribution with [k](/page/K) groups and \nu error , ensuring joint coverage of 100(1-\alpha)% for the set of all such normalized contrasts. This approach leverages the fact that the maximum absolute value of these studentized normalized contrasts follows the studentized range distribution under the of equal means. When sample sizes are unequal, the standard studentized range procedures require adjustment to maintain validity. The Spjøtvoll-Stoline method extends Tukey's T-method by incorporating an effective number of comparisons k' (typically between 2 and k), which approximates the studentized range critical value as q_{k', \nu, \alpha} scaled by the harmonic mean of sample sizes or similar adjustments to the standard error. This yields conservative simultaneous confidence intervals for pairwise differences, with performance superior to Scheffé's method when imbalance is moderate and focus is on selected contrasts; simulations show it controls the error rate closely while offering higher power than unadjusted alternatives. Nonparametric analogs of the studentized range arise in rank-based multiple comparison tests, particularly for ordered alternatives, where the test statistic is the range of average ranks studentized by an estimate of scale. These methods, rooted in Hayter's one-sided studentized range test, replace sample means with rank sums and use asymptotic approximations or tables for critical values, providing robust without assumptions; for example, in the , the maximum range divided by a pooled rank variance follows a analogous to the studentized range under the . Such procedures maintain type I error control and detect shifts in location with efficiency comparable to versions for non-normal data. In balanced incomplete block designs (BIBD), the studentized range is applied to compare treatment effects after adjusting for block variability, treating the design as a fixed-effects ANOVA with the range of estimated treatment means \hat{\tau}_i. Critical values from the studentized range distribution with effective degrees of freedom \nu = bk - b - v + 1 (where b is the number of blocks, k the block size, and v the number of treatments) are used for simultaneous intervals on \hat{\tau}_{\max} - \hat{\tau}_{\min}, scaled by \sqrt{\text{MSE} \cdot (1/k + 1/\lambda)} where \lambda is the pairwise concurrence parameter; this ensures balanced power for all pairwise treatment contrasts while accounting for incomplete blocking. Power analysis for tests based on the studentized range distribution reveals that increases with sample size n and error variance estimate , but decreases with the number of groups k, as larger k inflates the q_{k, \nu, \alpha}. The exact , derived for the one-way ANOVA model, depends on the least favorable of means (e.g., equal spacing within the ), allowing sample size computation to achieve desired \beta under restricted alternatives; for instance, approaches 1 as the noncentrality parameter ( of means over deviation) grows, but minimal guaranteed is lowest when means are at extremes.

Connection to Student's t-distribution

The studentized range distribution exhibits a direct mathematical connection to Student's t-distribution, particularly evident when the number of samples k = 2. In this case, the studentized range Q is exactly equivalent to \sqrt{2} |T_{\nu}|, where T_{\nu} follows Student's t-distribution with \nu degrees of freedom, typically \nu = N - k in the context of analysis of variance with equal sample sizes. This equivalence arises because, for two samples, the range reduces to the absolute difference between the two sample means, and the studentization factor aligns precisely with the denominator of the two-sample t-statistic, scaled by \sqrt{2}. Consequently, the quantiles of the studentized range for k=2 are identical to those of the scaled t-distribution, facilitating straightforward computation using standard t-tables adjusted by the scaling factor. For general k > 2, the studentized range can be expressed as the range of k independent standard variables divided by an independent factor \sqrt{\chi^2_{\nu}/\nu}, where \chi^2_{\nu} is a -squared random with \nu ; this denominator is the same scaling that defines the t-distribution from a single variate. Equivalently, Q represents the range among k correlated t-distributed statistics, as each studentized sample mean follows a t-distribution with the shared denominator introducing dependence across the statistics. This structure underscores the probabilistic link to the t-distribution while accounting for multiplicity through the extremal range. Early computational tables for the studentized range leveraged this connection by repurposing t-distribution tables for the k=2 case, simply applying the \sqrt{2} scaling to obtain critical values without needing separate computations. For k > 2, however, the distribution deviates, exhibiting thicker tails than the single t-distribution due to the increased variability from considering the extremal range across multiple samples, which amplifies the probability of large deviations relative to a pairwise t-test. This tail behavior enhances the conservatism of multiple comparison procedures like Tukey's HSD, ensuring control of family-wise error rates.

Relation to uniform range distribution

The unstudentized range W is defined as the difference between the of k and identically distributed (i.i.d.) random variables from a specified distribution, such as the on [0,1] or the with known standard deviation \sigma = 1. For a general continuous distribution with (CDF) F and (PDF) f, the PDF of W is given by g(w) = k(k-1) \int_{-\infty}^{\infty} [F(u + w) - F(u)]^{k-2} f(u) f(u + w) \, du, where the integral is over the support where F(u + w) > F(u). For the specific case of k i.i.d. uniform[0,1] random variables, this simplifies to g(w) = k(k-1) w^{k-2} (1 - w) for $0 < w < 1. In contrast, for standard normals with known \sigma = 1, the PDF involves a more intricate with the normal density and CDF, reflecting the unbounded support and leading to heavier tails in the distribution compared to the uniform case. The studentized range Q = W / s, where s is an independent estimate of the standard deviation based on \nu degrees of freedom (typically from a pooled sample variance following a \chi^2_\nu / \nu distribution), introduces additional variability due to the randomness in s. This studentization effect creates dependence between the numerator and the scaling factor, as s can vary, resulting in heavier tails for the studentized range distribution relative to the unstudentized version; small values of s inflate Q, increasing the probability of extreme values. The joint distribution arises from the product measure of the unstudentized range and the scaled chi distribution for $1/s, leading to a more dispersed overall form that is crucial for robust inference under unknown variance. As the degrees of freedom \nu become large, the sample standard deviation s converges in probability to the true \sigma = 1, causing the studentized range distribution to asymptotically approach that of the unstudentized range from standard normals. For large k, both distributions exhibit similar tail behavior, with quantiles approximating \sqrt{2 \log k}, providing a scaling factor that normalizes the range to a limiting extreme value distribution; thus, the studentized range Q behaves approximately like the unstudentized range scaled by this factor in the tails. This convergence underscores their shared extremal properties for normal populations under stable variance estimation. In applications, the unstudentized range finds use in scenarios with known variance, such as statistical quality control charts where process \sigma is predetermined, allowing direct monitoring of variability through range limits without estimation error. Conversely, the studentized range is essential for statistical inference in multiple comparison procedures, like Tukey's honest significant difference test, where unknown variance requires scaling by s to maintain valid control of error rates across comparisons. Moment comparisons highlight key differences: for the unstudentized range from k i.i.d. uniform[0,1] variables, the expected value is simply E[W] = (k-1)/(k+1), reflecting the bounded support and linear simplicity of the . For the studentized range under normal assumptions, moments are more intricate, involving infinite series or integrals over the joint density of the range and chi-squared components, with E[Q] increasing with k and decreasing with \nu, often tabulated rather than closed-form due to the added variability from studentization.

Derivation

General form from order statistics

The studentized range distribution arises in the context of order statistics from the group sample means across multiple groups. Consider k independent groups, each with n observations denoted X_{ij} for i=1 to k and j=1 to n, where each X_{ij} \sim N(\mu_i, \sigma^2) independently. Under the null hypothesis that all group means are equal (\mu_i = \mu for all i), the group sample means \bar{X}_i are i.i.d. N(\mu, \sigma^2 / n). Standardizing by the known \sigma yields \sqrt{n} (\bar{X}_i - \mu)/\sigma \sim N(0,1), so the range of the standardized means R' = \max_i Z_i - \min_i Z_i, where Z_i \sim N(0,1) i.i.d. The pooled standard deviation s is derived from the mean square error (MSE) in the one-way ANOVA model, given by s^2 = \frac{1}{\nu} \sum_{i=1}^k \sum_{j=1}^n (X_{ij} - \bar{X}_i)^2, where \nu = k(n-1) is the error degrees of freedom. The studentized range statistic is then Q = \frac{\max_i \bar{X}_i - \min_i \bar{X}_i}{s / \sqrt{n}} = \frac{R' \cdot \sigma / \sqrt{n}}{s / \sqrt{n}} = R' / (s / \sigma), or equivalently Q = R' / \sqrt{\chi^2_\nu / \nu} since s / \sigma \sim \sqrt{\chi^2_\nu / \nu}. Under normality, the vector of group means and the pooled s are independent. The probability structure of Q is built upon the distribution of the range R' of k i.i.d. standard normals. For i.i.d. observations from a continuous distribution with pdf \phi and cdf \Phi (standard normal), the cdf of R' is P(R' \leq r) = k \int_{-\infty}^{\infty} \left[ \Phi(u + r) - \Phi(u) \right]^{k-1} \phi(u) \, du. The cdf of Q is then P(Q \leq q) = E\left[ P\left( R' \leq q \cdot U \,\middle|\, U \right) \right] = \int_0^{\infty} P\left( R' \leq q u \right) f_U(u) \, du, where U = \sqrt{\chi^2_\nu / \nu} and f_U is its pdf (a scaled chi distribution). Substituting the cdf of R' yields the double integral representation P(Q \leq q) = \int_0^{\infty} k \left[ \int_{-\infty}^{\infty} \left[ \Phi(u + q u) - \Phi(u) \right]^{k-1} \phi(u) \, du \right] f_U(u) \, du. This structure emphasizes the role of order statistics of the k standardized means in establishing the foundational probability integral for the studentized range under the normal model. While the exact form relies on the normality assumption for independence and the joint density, generalizations to non-normal distributions can be approached via resampling methods such as bootstrapping, which approximates the sampling distribution of Q by generating empirical replicates from the observed data. However, the focus here remains on the precise integral-based derivation from order statistics of the group means under the normal model.

Specific form for normal populations

When the underlying population follows a normal distribution, the studentized range distribution takes a specific form derived from the order statistics of k standard normal random variables Z_1, \dots, Z_k. The range is defined as R = \max_i Z_i - \min_i Z_i, and an independent random variable with \nu degrees of freedom yields the studentized range Q = R / \sqrt{\chi_\nu^2 / \nu}. This construction assumes the samples are drawn from N(\mu, \sigma^2), with the Z_i obtained after standardization by the known or estimated \sigma. The resulting distribution of Q is pivotal for normal populations and independent of \mu and \sigma. The cumulative distribution function (CDF) of Q is given by the double P(Q \leq q) = \int_0^\infty k \left[ \int_{-\infty}^\infty \left[ \Phi(u + q u) - \Phi(u) \right]^{k-1} \phi(u) \, du \right] f_U(u) \, du, for q > 0, where f_U(u) is the pdf of U = \sqrt{\chi^2_\nu / \nu}, \Phi denotes the CDF of the standard , and \phi its pdf. The pdf f(q) can be obtained by differentiating the CDF with respect to q. This expression arises from the joint distribution of the order statistics under , with the inner capturing the conditional cdf of the scaled range. The inner integral employs the standard normal CDF \Phi, which can be expressed in terms of the error function \erf(z) = \frac{2}{\sqrt{\pi}} \int_0^z e^{-w^2} dw as \Phi(z) = \frac{1}{2} \left[1 + \erf\left(z / \sqrt{2}\right)\right], facilitating numerical evaluation while preserving the closed-form integral structure. No simple closed-form pdf exists, and the distribution is typically evaluated numerically or via tables. Validation of the expression occurs in special cases; for instance, when k=2, the distribution simplifies to a scaled Student's t-distribution with \nu degrees of freedom, specifically Q / \sqrt{2} \sim |t_\nu|, aligning with the known two-sample comparison under normality.

History

Origins in multiple comparisons

The studentized range distribution originated in the context of early 20th-century agricultural experiments, where statisticians sought reliable methods to compare multiple treatment means under limited sample sizes. Ronald A. Fisher developed analysis of variance (ANOVA) in the 1920s at the Rothamsted Experimental Station in , motivated by the need to analyze field trials involving numerous crop varieties, fertilizers, and soil conditions. While ANOVA provided an for overall group differences, follow-up pairwise comparisons posed challenges, as repeated significance tests inflated error rates, prompting informal approaches to control Type I errors across multiple inferences. Prior to formalization, and his collaborator explored range-based ideas in for bioassays and agricultural designs, using the difference between the largest and smallest sample means as a simple metric for detecting effects in experiments like yield comparisons or potency assays. These methods were ad hoc, often approximating via t-tests or unadjusted ranges, without an exact , resulting in conservative procedures to mitigate false positives in multi- settings. Their tables for biological and agricultural research supported such analyses but highlighted the limitations of lacking precise critical values for ranges in normal populations. The studentized range statistic, defined as the range divided by an independent estimate of the standard deviation, was suggested by W. S. Gosset () in his 1927 paper "Errors of Routine Analysis," where it appeared in discussions of small-sample variability from normal distributions in industrial . The distribution was first derived by D. Newman in 1939, who calculated percentage points for the statistic Q. A pivotal advancement for its application occurred in 1949 when John W. Tukey proposed using the studentized range as a core tool for multiple comparisons in his paper "Comparing Individual Means in the Analysis of Variance," published in Biometrics. Tukey applied the statistic to perform simultaneous pairwise tests following a significant ANOVA, explicitly designed to control the family-wise error rate at a specified level, such as 5%, across all comparisons. This built directly on Newman's work and earlier ideas by Gosset. Early adoption of Tukey's method encountered significant hurdles due to the absence of exact distribution tables for the studentized range under varying and group sizes, forcing practitioners to rely on conservative approximations like bounding the statistic with t-distribution quantiles or unstudentized values. By the 1950s, these limitations were addressed through the development of exact forms and computational tables, transitioning the approach from informal heuristics to a rigorous, implementable integrated into standard statistical software and texts for ANOVA extensions.

Key developments and contributors

In 1952, M. Keuls extended the application of the studentized range to multiple comparisons in ANOVA, proposing a step-down procedure that uses quantiles from the distribution to test ordered means, forming the basis of the Newman-Keuls test. In 1955, David B. Duncan developed the multiple range test, extending the application of the studentized range to perform ordered multiple comparisons with adjusted critical values based on protection levels derived from . During the 1960s, significant advancements in computational accessibility came from the compilation of extensive tables for the quantiles of the studentized range distribution. H. Leon Harter published comprehensive tables covering up to 24 groups (k=24) and degrees of freedom (ν) up to 120, facilitating practical use in analysis of variance procedures. Earlier foundational tables had been provided by E. S. Pearson and H. O. Hartley in 1942, tabulating the probability integral for smaller sample sizes up to n=20, which laid the groundwork for these expansions. By the 1980s, the studentized range was integrated into statistical software packages such as and precursors to , allowing automated calculation of critical values and p-values. Contributions from P. H. Ramsey and colleagues addressed extensions for unequal sample sizes (n), modifying the studentized range statistic to maintain control over family-wise error rates in unbalanced designs. Key figures in these developments include D. Newman for the initial derivation, for application to multiple comparisons, M. Keuls and David Duncan for practical testing extensions, and table compilers like E. S. Pearson and H. O. Hartley, whose work ensured the distribution's widespread adoption in statistical practice. Post-2000, classical developments have been complemented by Bayesian extensions incorporating prior distributions on range parameters for robust multiple comparisons, and robust versions adapting the studentized range to non-normal data via trimmed means or bootstrapping.

Numerical methods

Computation of distribution functions

The probability density function (PDF) and cumulative distribution function (CDF) of the studentized range distribution are expressed through double integrals that incorporate the standard normal density and the chi-squared distribution. Numerical evaluation of these functions relies on quadrature methods, such as Gauss-Legendre quadrature, to approximate the integrals with high precision. For instance, a 16-point Legendre formula is commonly applied to the inner integral over the normal density, while adaptive quadrature handles the outer integral to adjust for varying parameter values and ensure accuracy. These methods are essential for computing the PDF and CDF across a range of the number of groups k, degrees of freedom \nu, and the studentized range value q. A foundational approach for exact computation is provided by Algorithm AS 190, which evaluates the CDF via numerical of the double integral representation: F(q; k, \nu) = k \int_0^\infty \int_{-\infty}^\infty \left[ \Phi(y + q \sqrt{u}) - \Phi(y) \right]^{k-1} \phi(y) f(u; \nu) \, dy \, du, where \Phi and \phi are the standard normal CDF and PDF, respectively, and f(u; \nu) is the PDF of \chi^2_\nu / \nu. The algorithm uses adaptive limits and error controls to achieve relative accuracy of about $10^{-5}, and quantiles are obtained by inverting the CDF using a . Recurrence relations derived from differential equations satisfied by the CDF enable efficient updates for sequential values of q or k, reducing computational overhead in multiple evaluations. Monte Carlo simulation offers a flexible, though approximate, for estimating the distribution functions, particularly useful for validation or non-standard parameters. The procedure generates large numbers of independent samples from the (size k), computes the sample range R, estimates the standard deviation s from pooled samples with \nu , and forms the studentized range Q = R / s. The empirical CDF is then the proportion of simulated Q values less than or equal to a given q. With $10^6 or more replications, this yields estimates accurate to within a few percent in the central region, though variance is higher in the tails. For large \nu > 100,000 or large k, asymptotic expansions facilitate rapid approximations. When \nu is large, the distribution converges to that of the range of k standard normals, computed via a single over , avoiding the chi-squared component. Higher-order approximations, such as those using Edgeworth series expansions around the normal limit, incorporate and terms to improve tail accuracy for finite but large parameters. These are particularly effective for quick estimates in high-dimensional settings. Numerical methods for the studentized range can face convergence challenges, especially for small \nu < 5 or large k > 20, where the integrands exhibit sharp peaks or slow , leading to potential underestimation of tail probabilities. In such cases, increasing the number of points, employing adaptive step-sizing, or switching to asymptotic forms mitigates errors, with relative accuracies typically maintained above $10^{-4} through careful implementation.

Approximations and software tools

Several approximations have been developed to compute quantiles and distribution functions of the studentized range distribution efficiently, avoiding the computational intensity of exact . One early approach involves approximations for upper quantiles, as outlined in Peiser (1943), which utilizes inverse functions to estimate critical values. More comprehensive tables and approximation methods for probabilities and quantiles are provided in (1981), including extensions for multiple comparisons that rely on and t-distribution approximations for practical tabularization. A notable non-iterative approximation for upper quantiles was proposed by Hayter (1999), which derives studentized range quantiles directly from accurate t-distribution quantiles through simple arithmetic operations, achieving high precision across a wide range of parameters. For large degrees of freedom (ν > 100,000), asymptotic approximations based on infinite degrees of freedom are employed, simplifying the cumulative distribution function to a form amenable to standard normal computations. Software implementations facilitate practical computation of the studentized range distribution. In , the base stats package provides functions such as ptukey() for the , qtukey() for quantiles, and ptukey() for tail probabilities, supporting Tukey's honest significant difference tests. The multcomp package extends this with glht() for general linear hypotheses involving studentized range-based multiple comparisons. In , scipy.stats.studentized_range offers probability density, , and quantile functions, while statsmodels.stats.multicomp.tukeyhsd() implements Tukey's test using these distributions. PROC GLM computes studentized range statistics via the LSMEANS statement with multiple comparison options, and the PROBMC procedure handles partitioned and full range probabilities. Updated tables for critical values extend classical works like those of Harter (1969) and Pearson and Hartley (1972) to higher parameters and are available through resources such as the Real Statistics ResourcePack for Excel on CRAN mirrors and online statistical appendices. These tables cover α levels from 0.001 to 0.10, with k up to 100 and df up to , surpassing earlier limitations like those in Dunn and Clark (1971). Validation studies confirm the accuracy of these approximations relative to exact integrals. Hayter's (1999) method yields relative errors below 0.075% across 4,738 test cases (r > 2), with over 98% of errors under 0.02%, and maximum absolute errors limited to three units in the fourth decimal place for tabulated values. Similarly, implementations in and base functions report errors under 1% for quantile estimates when compared to high-precision numerical integrations, particularly for moderate to large ν.

References

  1. [1]
    THE DISTRIBUTION OF RANGE IN SAMPLES FROM A NORMAL ...
    THE DISTRIBUTION OF RANGE IN SAMPLES FROM A NORMAL POPULATION, EXPRESSED IN TERMS OF AN INDEPENDENT ESTIMATE OF STANDARD DEVIATION. D. NEWMAN.
  2. [2]
    scipy.stats.studentized_range — SciPy v1.16.2 Manual
    A studentized range continuous random variable. As an instance of the rv_continuous class, studentized_range object inherits from it a collection of generic ...
  3. [3]
    Univariate Distributions - JuliaStats
    ... Γ(ν/2)1​21−ν/2xν− ... The studentized range distribution has probability density function: f ( q ...
  4. [4]
    Tables of Range and Studentized Range
    ### Summary of Content on Studentized Range Distribution
  5. [5]
    TABLES OF RANGE AND STUDENTIZED RANGE1 - Project Euclid
    The idea of using the ratio of the range to an independent estimate s of the population standard deviation (studentized range) was first proposed by “Stu-.Missing: original | Show results with:original
  6. [6]
    Q Table for Tukey's HSD - Stat@Duke
    TABLE B.5-. The studentized range statistic (q)*. *The critical values for q corresponding to alpha = .05 (top) and alpha = .01 (bottom). df for Error Term ...
  7. [7]
    None
    ### Summary of Special Cases of the Studentized Range (k=2 or Limits) from Tukey (1949)
  8. [8]
    Comparing Individual Means in the Analysis of Variance - jstor
    COMPARING INDIVIDUAL MEANS IN THE. ANALYSIS OF VARIANCE*. JOHN W. TUKEY. Princeton University. The practitioner of the analysis of variance often wants to draw ...
  9. [9]
    Statistical notes for clinical researchers: post-hoc multiple comparisons
    Tukey's HSD procedure provides the simplest way to control αFW and is considered as the most preferable method when all pairwise comparisons are performed. The ...
  10. [10]
    Tukey's Honestly Significant Difference (HSD) - Sage Knowledge
    The studentized range distribution is similar to the Student's i ... Expected Value · Fixed-Effects Models · Inclusion Criteria · Influence ...
  11. [11]
    Inferences for Contrasts and Treatment Means
    The estimates of μ2−μ1,μ2−μ3,μ2−μ4 are ˉy2.−ˉy1.=289.75,ˉy2.−ˉy3.=427.50,ˉy2.−ˉy4.=364.25. Hence the simultaneous 95% confidence intervals are 289.75±95.63, ...
  12. [12]
    An Extension of the T-Method of Multiple Comparison to Include the ...
    Apr 5, 2012 · Theory and Method. An Extension of the T-Method of Multiple Comparison to Include the Cases with Unequal Sample Sizes. Emil Spjøtvoll ...
  13. [13]
    Nonparametric multiple-comparison methods for simply ordered ...
    Jun 15, 2007 · The first is a rank-based method, which traces its roots to the one-sided Studentized-range test by Hayter, and the other is a two-stage method ...
  14. [14]
    [PDF] Chapter 16 - ANOVA
    In a balanced incomplete block design, all the pairwise differences are identifiable and have the same standard error. Pairwise differences are more likely.
  15. [15]
    The Power Function of the Studentised Range Test - Project Euclid
    In this paper we investigate the power function of the Studentised range test for comparing the means of normal populations in the one-way fixed effects ...Missing: studentized | Show results with:studentized
  16. [16]
  17. [17]
    [PDF] Applied linear statistical models - Statistics - University of Florida
    ... Studentized Range Distribution 746. Simultaneous Estimation 747. Simultaneous ... joint density function of YI and Y2, and h (Y2) is the marginal density.
  18. [18]
    [PDF] ORDER STATISTICS IN STATISTICAL INFERENCE - CORE
    Apr 2, 2018 · For a comprehensive study of the order statistics, their distributions and moments etc. one may refer to David (1981; Chapter 2,3). Page 24. 14.
  19. [19]
    Ronald Fisher, a Bad Cup of Tea, and the Birth of Modern Statistics
    Aug 6, 2019 · At the time, the early 1920s, Fisher worked at an agricultural research station north of London. A short, slight mathematician with rounded ...
  20. [20]
    Introduction to Fisher (1926) The Arrangement of Field Experiments
    Fisher introduced the subdivision of sums of squares now known as an analysis of variance (anova) table (1923), derived the exact distribution of the (log of ...
  21. [21]
    [PDF] Statistical tables for biological, agricultural and medical research
    AND. FRANK YATES, C.B.E., Sc.D., F.R.S.. HEAD OF STATISTICAL DEPARTMENT ... SIR RONALD FISHER'S sudden death in July 1962 delayed the com- pletion of ...Missing: bioassays 1930s
  22. [22]
    The Principles of Biological Assay - jstor
    Fisher and Yates (1947, Table VIII 2) have given a table to assist rapid esti- mation by this method. So high an efficiency is satisfactory for most ...
  23. [23]
    Comparing individual means in the analysis of variance.
    Comparing individual means in the analysis of variance. · J. Tukey · Published in Biometrics 1 June 1949 · Mathematics.
  24. [24]
    Probabilities and Upper Quantiles for the Studentized Range - jstor
    Dunlap et al. (1977) published an algorithm for probabilities for the studentized range. It requires an evaluation of the integrand at 1032 nodes to produce ...
  25. [25]
    Externally studentized normal midrange distribution - SciELO
    This work aimed to derive analytic equations to distribution of the externally studentized midrange, obtaining the cumulative distribution, probability density ...
  26. [26]
    [PDF] nag_prob_studentized_range (g01emc) - Numerical Algorithms Group
    The Studentized range statistic can be used as part of a multiple comparisons procedure such as the. Newman–Keuls procedure or Duncan's multiple range test (see ...
  27. [27]
    Simultaneous Statistical Inference - Rupert G. Jr. Miller - Google Books
    A new table of critical points for the studentized maximum modulus is included in this second edition as an addendum. The original edition included the table by ...
  28. [28]
    An accurate, non-iterative approximation for studentized range ...
    This article proposes a non-iterative method of approximating upper quantiles of the studentized range, suitable for this context.Missing: original | Show results with:original
  29. [29]
    The Studentized Range Distribution - R
    Description. Distribution function of the studentized range, R / s R/s R/s , where R is the range of a standard normal sample and d f × s 2 df \times s^2 df×s2 ...Missing: density | Show results with:density
  30. [30]
    PROBMC Function - SAS Help Center
    Aug 20, 2025 · Computing the Partitioned Range. RANGE applies to the distribution of the studentized range for n group means. PARTRANGE applies to the ...
  31. [31]
    Studentized Range q Table | Real Statistics Using Excel
    The following tables provide the critical values for q(k, df, α) when α = .10, .05, .025, 01, .005 and .= 001.Missing: derivation | Show results with:derivation
  32. [32]
    An accurate, non-iterative approximation for studentized range ...
    Sep 30, 2025 · Given an accurate quantile from the student t distribution, only a few arithmetic operations yield a studentized range quantile with accuracy ...Missing: origin | Show results with:origin