Fact-checked by Grok 2 weeks ago

Half-normal distribution

The half-normal distribution is a continuous defined on the non-negative real numbers, representing the distribution of the of a following a with zero and \sigma > 0. It serves as a special case of both the (with \mu = 0) and the , and is equivalent to the chi distribution with one degree of freedom when \sigma = 1. The (PDF) of the half-normal distribution is given by
f(x \mid \sigma) = \frac{\sqrt{2}}{\sigma \sqrt{\pi}} \exp\left( -\frac{x^2}{2\sigma^2} \right)
for x \geq 0, and zero otherwise, where the is at x = 0 and the density is strictly decreasing. The (CDF) is
F(x \mid \sigma) = \erf\left( \frac{x}{\sigma \sqrt{2}} \right) = 2\Phi\left( \frac{x}{\sigma} \right) - 1,
for x \geq 0, where \erf is the and \Phi is the standard CDF. Key moments include the \mu = \sigma \sqrt{2/\pi} \approx 0.7979\sigma, the variance \sigma^2 (1 - 2/\pi) \approx 0.3634\sigma^2, \sqrt{2}(4 - \pi)/(\pi - 2)^{3/2} \approx 0.9953, and excess $8(\pi - 3)/(\pi - 2)^2 \approx 0.8784. Higher-order moments are \mathbb{E}[X^n] = 2^{n/2} \sigma^n \Gamma((n+1)/2) / \sqrt{\pi} for n \geq 1.
This distribution finds applications in statistical modeling of non-negative phenomena, such as measurement errors, under in , and process deviations in charts. It is also used as a prior for variance parameters in and in for estimating failure rates. Parameter estimation typically employs maximum likelihood, with the \hat{\sigma} satisfying \hat{\sigma} = \sqrt{2/\pi} \cdot \mathbb{E}[X] for the sample mean in large samples.

Definition

Probability density function

The half-normal distribution is defined as the distribution of the of a following a with mean zero and variance \sigma^2. If X \sim \mathcal{N}(0, \sigma^2), then Y = |X| follows a half-normal distribution with \sigma > 0. The probability density function of Y is f(y; \sigma) = \frac{\sqrt{2 / \pi}}{\sigma} \exp\left( -\frac{y^2}{2 \sigma^2} \right), \quad y \geq 0, and f(y; \sigma) = 0 for y < 0. The support of the distribution is the non-negative real line [0, \infty). This PDF arises by folding the symmetric normal density over the positive axis, which doubles the probability mass on y \geq 0 due to the reflection of the negative tail, and the normalizing constant \sqrt{2 / \pi} ensures the total probability integrates to 1 over the half-line, as derived from the integral of the standard normal density from 0 to \infty, which equals $1/2, combined with the folding factor of 2. Qualitatively, the PDF exhibits a bell-like shape truncated at zero, with its mode at y = 0 where the density achieves its maximum value of \sqrt{2 / \pi} / \sigma, and it decreases monotonically toward zero as y increases to infinity.

Cumulative distribution function

The cumulative distribution function of the half-normal distribution with scale parameter \sigma > 0 is F(y; \sigma) = \begin{cases} 0 & \text{if } y < 0, \\ \operatorname{erf}\left( \frac{y}{\sigma \sqrt{2}} \right) & \text{if } y \geq 0, \end{cases} where \operatorname{erf}(z) = \frac{2}{\sqrt{\pi}} \int_0^z e^{-t^2} \, dt denotes the error function. This expression arises from integrating the probability density function over [0, y], which corresponds to the probability that the absolute value of a zero-mean normal random variable with standard deviation \sigma falls between 0 and y. Equivalently, it equals the normal cumulative distribution function evaluated from -y to y, scaled by the folding at zero: F(y; \sigma) = 2\Phi(y / \sigma) - 1 for y \geq 0, where \Phi is the cumulative distribution function of the standard normal distribution. The quantile function (inverse cumulative distribution function) is y_p = \sigma \sqrt{2} \, \operatorname{erf}^{-1}(p) for $0 < p < 1, which facilitates the generation of random variates. At the boundaries, F(0; \sigma) = 0 and \lim_{y \to \infty} F(y; \sigma) = 1.

Properties

Moments

The moments of the half-normal distribution Y, defined as the distribution of |X| where X \sim \mathcal{N}(0, \sigma^2), can be derived by integrating powers of Y against its probability density function or by leveraging the even moments of the underlying normal distribution. The mean is given by
\mathbb{E}[Y] = \sigma \sqrt{\frac{2}{\pi}}.
This follows from the integral \mathbb{E}[Y] = \int_0^\infty y \cdot \frac{\sqrt{2}}{\sigma \sqrt{\pi}} \exp\left( -\frac{y^2}{2\sigma^2} \right) \, dy, which, via the substitution u = y^2 / (2\sigma^2), simplifies to \sigma \sqrt{2/\pi}.
The variance is
\mathrm{Var}(Y) = \sigma^2 \left( 1 - \frac{2}{\pi} \right).
It is computed as \mathbb{E}[Y^2] - (\mathbb{E}[Y])^2, where \mathbb{E}[Y^2] = \sigma^2 since Y^2 = X^2 follows a chi-squared distribution with one degree of freedom.
Higher-order moments exhibit positive skewness and moderate excess kurtosis, reflecting the distribution's asymmetry and heavier right tail compared to the normal. The skewness coefficient is
\gamma_1 = \sqrt{2} \frac{4 - \pi}{(\pi - 2)^{3/2}} \approx 0.995,
a parameter-free measure derived from the third central moment normalized by the cube of the standard deviation.
The excess kurtosis is
\gamma_2 = \frac{8(\pi - 3)}{(\pi - 2)^2} \approx 0.870,
indicating leptokurtic tails relative to the normal distribution; the raw kurtosis is thus approximately 3.870. This is obtained from the fourth central moment divided by the fourth power of the standard deviation.
The general raw moment for k > -1 is
\mathbb{E}[Y^k] = \sigma^k \cdot 2^{k/2} \cdot \frac{\Gamma\left( \frac{k+1}{2} \right)}{\sqrt{\pi}}.
This closed-form expression arises from evaluating the integral \int_0^\infty y^k f_Y(y) \, dy using the substitution t = y^2 / (2\sigma^2), which transforms it into a Gamma function integral: \mathbb{E}[Y^k] = 2^{k/2} \sigma^k \int_0^\infty t^{(k-1)/2} e^{-t} \, dt / \sqrt{\pi} = \sigma^k 2^{k/2} \Gamma((k+1)/2) / \sqrt{\pi}. For positive integer k, it equivalently expresses in terms of double factorials.

Characteristic function and generating functions

The characteristic function of a half-normal random variable Y \sim \mathrm{HN}(0, \sigma^2), defined as \phi_Y(t) = \mathbb{E}[e^{itY}], is expressed as \phi_Y(t) = e^{-\sigma^2 t^2 / 2} \left[1 + \erf\left(i \frac{\sigma t}{\sqrt{2}}\right)\right], where \erf denotes the extended to complex arguments via . This form arises from the integral representation \phi_Y(t) = \sqrt{\frac{2}{\pi}} \int_0^\infty e^{itx} e^{-x^2/(2\sigma^2)} \, dx / \sigma, leveraging the symmetry of the underlying and properties of the . An equivalent expression uses the \Phi of the standard normal: \phi_Y(t) = 2 e^{-\sigma^2 t^2 / 2} \Phi(i \sigma t). The complex ensures the function is well-defined, though numerical evaluation requires care due to the imaginary argument. The M_Y(t) = \mathbb{E}[e^{tY}] follows from of the , yielding M_Y(t) = e^{\sigma^2 t^2 / 2} \left[1 + \erf\left(\frac{\sigma t}{\sqrt{2}}\right)\right] for all real t, reflecting the positive support of the distribution. This expression is obtained by substituting t \to -it in \phi_Y(t), exploiting the entire nature of the . The domain of convergence is the entire real line, unlike distributions with unbounded negative support, but the half-line restriction introduces challenges in asymptotic expansions for large |t|, where the approaches its asymptotic limits slowly. The cumulant-generating function is K_Y(t) = \log M_Y(t), from which cumulants are derived as successive derivatives at t=0; the first cumulant equals the mean \sigma \sqrt{2/\pi}, and the second equals the variance \sigma^2 (1 - 2/\pi), consistent with the moments of the distribution. These generating functions facilitate theoretical analyses, such as convolutions of half-normal variables, which appear in limit theorems for non-negative processes like absolute deviations or folded errors in statistical models. For instance, the supports singularity analysis in bivariate generating functions, yielding half-normal limits for parameters in combinatorial structures under zero drift. The restricted support complicates direct application of some inversion techniques compared to full-line distributions, necessitating for inversion formulas.

Parameter estimation

Method of moments

The method of moments for estimating the σ of the half-normal distribution involves computing the sample moments and setting them equal to the corresponding moments. The mean is E[Y] = σ √(2/π) and the second raw moment is E[Y²] = σ². To implement this, calculate the sample mean m₁ = (1/n) ∑{i=1}^n y_i and the sample second raw moment m₂ = (1/n) ∑{i=1}^n y_i², where n is the sample size and y_i are the observations. Using the first moment matching, the estimator is obtained by solving m₁ = \hat{σ} √(2/π), yielding \hat{σ} = m₁ √(π/2). This estimator is unbiased, since the sample mean is unbiased for E[Y] and the transformation is linear, and it is consistent as n → ∞ by the . An alternative estimator uses the second raw moment matching, solving m₂ = \hat{σ}², giving \hat{σ} = √m₂. This estimator is the same as the maximum likelihood estimator. It is unbiased for σ² but biased downward for σ due to the concavity of the function. It is consistent as n → ∞. The first moment estimator is simple to compute but less efficient than the maximum likelihood estimator for small samples, as it does not utilize the full shape information of the distribution.

Maximum likelihood estimation

The maximum likelihood estimator (MLE) for the scale parameter \sigma of the half-normal distribution is derived from the log-likelihood function for an independent and identically distributed sample y_1, \dots, y_n > 0. The probability density function of the half-normal distribution is f(y \mid \sigma) = \sqrt{\frac{2}{\pi}} \frac{1}{\sigma} \exp\left(-\frac{y^2}{2\sigma^2}\right) for y > 0, leading to the log-likelihood \ell(\sigma) = n \log\left(\sqrt{\frac{2}{\pi}} \frac{1}{\sigma}\right) - \frac{1}{2\sigma^2} \sum_{i=1}^n y_i^2 = \frac{n}{2} \log\left(\frac{2}{\pi}\right) - n \log \sigma - \frac{1}{2\sigma^2} \sum_{i=1}^n y_i^2. Maximizing \ell(\sigma) with respect to \sigma > 0 yields the MLE \hat{\sigma}_{\text{MLE}} = \sqrt{\frac{1}{n} \sum_{i=1}^n y_i^2}, which corresponds to the root mean square of the sample. This estimator arises directly from the invariance property of applied to the with known mean zero, where the half-normal is a special case; the squared observations y_i^2 match those of the underlying , preserving the MLE for the scale. The MLE \hat{\sigma}_{\text{MLE}} is consistent, with n \hat{\sigma}_{\text{MLE}}^2 / \sigma^2 following a \chi^2_n distribution scaled appropriately, converging to the true \sigma as n \to \infty. Under standard regularity conditions, it is asymptotically efficient and normally distributed: \sqrt{n} (\hat{\sigma}_{\text{MLE}} - \sigma) \xrightarrow{d} \mathcal{N}\left(0, \frac{\sigma^2}{2}\right), where the asymptotic variance \sigma^2 / 2 derives from the I(\sigma) = 2n / \sigma^2. For datasets potentially including zeros or subject to censoring (e.g., in residuals truncated at zero), modifications to the likelihood are required to account for the or censoring mechanism, though explicit forms depend on the specific setup.

Applications

In statistical diagnostics

The half-normal plot serves as a key diagnostic tool in for assessing the of residuals in models and identifying significant effects in experimental designs. It involves plotting the ordered absolute values of residuals against the theoretical quantiles of the half-normal distribution, where a straight-line pattern indicates that the residuals follow a half-normal distribution, supporting the assumption of underlying in the errors. In regression diagnostics, the half-normal plot is particularly useful for detecting deviations from normality and pinpointing outliers or influential observations. For instance, points that deviate substantially from the straight line suggest non-normal behavior or potential anomalies in the data. This approach extends to generalized linear models, where simulated envelopes around the plot provide a reference band derived from multiple simulations under the assumed model, aiding in the visual identification of systematic departures. A prominent application is Daniel's plot, employed in the analysis of factorial experiments within (DOE) to evaluate factor effects. Here, the absolute values of estimated effects are ranked and plotted against half-normal quantiles; effects aligning with the line are deemed negligible due to , while those straying indicate active factors. In unreplicated two-level designs, this method allows practitioners to distinguish real effects from random variation without formal tests. Simulation-based testing enhances the half-normal plot by generating half-normal variables from the fitted model to construct confidence s, enabling comparison of observed residuals or effects against expected distributions for detection. If observed points fall outside the envelope, it signals potential outliers or model misspecification, providing a robust, non-parametric way to validate assumptions. This technique is especially valuable in ANOVA contexts, where effects from designs are ranked and plotted against half-normal quantiles to isolate significant factors amid experimental noise. The half-normal plot was introduced by Cuthbert Daniel in 1959 specifically for interpreting two-level experiments, revolutionizing the visual screening of effects in industrial and scientific studies.

In physical modeling

The half-normal distribution is employed in physical modeling to represent the magnitudes of symmetric errors in measurements, where the underlying errors follow a but only non-negative deviations are physically meaningful, such as in for constrained positive quantities. For instance, when evaluating uncertainties in experimental data, the of normally distributed errors yields a half-normal distribution, allowing for the quantification of deviation scales without negative values that lack physical interpretation. This approach is particularly useful in Bayesian Type A evaluations under the Guide to the Expression of Uncertainty in Measurement (), where the half-normal arises from truncating a at zero to reflect physical boundaries. In , the half-normal distribution models the distribution of failure times or material strengths when the underlying is governed by a , but only positive excursions are relevant, such as in scenarios involving or tensile strengths. The σ in this context interprets the variability of the underlying normal , representing the standard deviation of the symmetric deviations before folding to the positive domain. In , the half-normal distribution describes the amplitude of , particularly after detection or operations, such as in the of white where the magnitude follows a half-normal form. This is evident in ultrasonic guided wave detection for weak Doppler signals, where the noise amplitude's half-normal behavior aids in assessments and detection thresholds. Similarly, in acoustics, intensity fluctuations due to spatial variations in room measurements, like times, can be modeled with half-normal distributions to account for uncertainties from normal-distributed phase or amplitude perturbations. In , such as interferometric , half-normal models arise from the absolute values of normal phase errors, influencing stability analyses of wavefront reconstructions.

Relation to the normal distribution

The half-normal distribution is defined as the distribution of the of a following a with mean zero. Specifically, if X \sim \mathcal{N}(0, \sigma^2), then Y = |X| has a half-normal distribution with \sigma > 0. This construction preserves the scale \sigma of the underlying while restricting the support of Y to the non-negative real line [0, \infty)./05:_Special_Distributions/5.13:_The_Folded_Normal_Distribution) The cumulative distribution function of Y is derived from the symmetry of the normal distribution via the probability integral transform: F_Y(y) = P(Y \leq y) = 2 \Phi\left( \frac{y}{\sigma} \right) - 1, \quad y \geq 0, where \Phi denotes the cumulative distribution function of the standard normal distribution \mathcal{N}(0,1). For y < 0, F_Y(y) = 0. This relation highlights the half-normal as a folded version of the centered normal, capturing only the positive deviations. Samples from the half-normal distribution can be straightforwardly generated by drawing from \mathcal{N}(0, \sigma^2) and applying the absolute value operation. In contrast to the symmetric centered at zero, the half-normal is asymmetric, with its achieving its maximum () at zero and monotonically decreasing thereafter. This peaked behavior at the reflects the folding , which concentrates mass near zero compared to the bell-shaped normal. If the underlying normal has a non-zero \mu \neq 0, the yields the more general ; the half-normal is the specific case where \mu = 0./05:_Special_Distributions/5.13:_The_Folded_Normal_Distribution)

Connections to chi and Rayleigh distributions

The half-normal distribution is equivalent to the distribution with one degree of freedom. Specifically, if Z \sim \mathcal{N}(0, 1), then |Z| follows the standard distribution with k = 1, which is identical to the standard half-normal distribution with \sigma = 1. For the general case where the underlying normal has variance \sigma^2, the half-normal random variable is \sigma times a standard chi(1) variate. This equivalence extends to moments: the mean of the half-normal distribution with parameter \sigma is \sigma \sqrt{2/\pi}, matching the mean of the chi(1) distribution scaled by \sigma, while the variance is \sigma^2 (1 - 2/\pi). The half-normal thus serves as the k=1 special case within the broader chi distribution family, which generalizes the norm of a k-dimensional multivariate . The Rayleigh distribution connects to the half-normal through this shared framework, as the Rayleigh with parameter \sigma is the chi distribution with k=2, representing the norm of a two-dimensional standard multivariate normal scaled by \sigma. In contrast to the half-normal's mean of \sigma \sqrt{2/\pi}, the Rayleigh has mean \sigma \sqrt{\pi/2} and variance \sigma^2 (4 - \pi)/2, highlighting their distinct roles despite common origins. In applications to multivariate normal distributions with zero mean, the absolute value of any single marginal component follows a half-normal distribution, whereas the full vector norm follows a chi distribution with degrees of freedom equal to the dimension; for dimension 2, this yields the Rayleigh.