Fact-checked by Grok 2 weeks ago

Logistic distribution

The logistic distribution is a continuous in and , characterized by a \mu \in \mathbb{R} and a s > 0, with f(x; \mu, s) = \frac{1}{s} \cdot \frac{e^{-(x - \mu)/s}}{(1 + e^{-(x - \mu)/s})^2} and F(x; \mu, s) = \frac{1}{1 + e^{-(x - \mu)/s}}. This S-shaped CDF, known as the or , arises naturally in models of bounded growth, and the distribution is symmetric about \mu, unimodal, and bell-shaped like distribution but with heavier tails that allow for more extreme values. For the standard case (\mu = 0, s = 1), the , , and coincide at 0, with variance \pi^2 / 3 \approx 3.2899. The logistic distribution traces its origins to the mid-19th century work of Belgian mathematician Pierre François Verhulst, who introduced the logistic growth model in 1838 to describe with limited resources, where the CDF form emerged from the solution to the \frac{dP}{dt} = rP(1 - P/K). This model was rediscovered independently in the early 20th century by Raymond Pearl and Lowell Reed in 1920 for U.S. population forecasting, leading to broader statistical adoption. By the 1950s, David Cox formalized its use in , a for binary outcomes where the log-odds follow a linear predictor, revolutionizing applications in , , and . Beyond regression, the distribution models phenomena with sigmoid growth patterns, such as in for population limits and for failure times (e.g., mechanical component lifetimes). Its mathematical tractability—closed-form CDF and moments—facilitates quantile estimation and , making it a practical to the normal distribution in scenarios requiring heavier tails without excessive complexity. Variants like the generalized logistic extend it for or in advanced modeling.

Definition

Probability density function

The probability density function of the logistic distribution is f(x \mid \mu, s) = \frac{\exp\left( \frac{x - \mu}{s} \right)}{s \left( 1 + \exp\left( \frac{x - \mu}{s} \right) \right)^2}, where \mu \in \mathbb{R} is the location parameter and s > 0 is the scale parameter. This density is symmetric about \mu, producing a bell-shaped curve that peaks at x = \mu. The logistic distribution exhibits heavier tails than the normal distribution with the same variance. The variance is \frac{\pi^2 s^2}{3}. In the standard parameterization (\mu = 0, s = 1), the random variable follows the distribution of the log-odds \log\left( \frac{p}{1-p} \right) for p drawn from a on [0, 1]. The PDF is derived by differentiating the of the logistic distribution.

Cumulative distribution function

The (CDF) of the logistic distribution with \mu and s > 0 is given by F(x; \mu, s) = \frac{1}{1 + e^{-(x - \mu)/s}}, \quad x \in \mathbb{R}. This CDF represents the standard , which produces an S-shaped curve that models cumulative probabilities for bounded outcomes ranging from 0 to , such as growth processes approaching saturation. The function is strictly increasing over line, with F(\mu) = 1/2, and exhibits asymptotic where F(x) \to 0 as x \to -\infty and F(x) \to [1](/page/1) as x \to \infty. The logistic distribution derives its name from the logistic function's early role in 19th-century models of , first introduced by Pierre François Verhulst in 1838 to describe patterns in biological systems.

Properties

Moments and cumulants

The logistic distribution with location parameter \mu and scale parameter s > 0 has mean \mathbb{E}[X] = \mu and variance \mathrm{Var}(X) = s^2 \pi^2 / 3. Due to the of the distribution about \mu, all odd central moments are zero, and thus the is 0. The is $21/5 = 4.2, yielding an excess kurtosis of $6/5 = 1.2, which indicates that the logistic distribution is leptokurtic relative to the normal distribution (excess kurtosis 0). The central moments m_n = \mathbb{E}[(X - \mu)^n] can be derived by direct integration of x^n against the or via the M(t) = \mathbb{E}[e^{tX}], which for the standard case (\mu = 0, s = 1) is M(t) = \Gamma(1 - it) \Gamma(1 + it), using the . For the standard logistic Z, the odd moments are m_n = 0 for odd n. For the general case, these scale as m_n = s^n m_n^{(Z)}, where m_n^{(Z)} denotes the corresponding moment of the standard . For example, the second is m_2 = \pi^2 / 3 and the fourth is m_4 = 7 \pi^4 / 15 for the standard case. The cumulants \kappa_n are obtained as the coefficients in the Taylor expansion of the cumulant-generating function K(t) = \log M(t). The first cumulant is \kappa_1 = \mu, the second is \kappa_2 = s^2 \pi^2 / 3, the third is \kappa_3 = 0 (consistent with zero skewness), and higher cumulants follow from derivatives of K(t), which involve polygamma functions due to the gamma function representation of the moment-generating function. For instance, the fourth cumulant is \kappa_4 = 2 s^4 \pi^4 / 15. To highlight differences in shape, the first four central moments of the standard logistic distribution (mean 0, variance \pi^2 / 3 \approx 3.29) can be compared to those of the standard (mean 0, variance 1). For fair scale comparison, consider the scaled logistic with variance 1 (achieved by dividing by \sqrt{\pi^2 / 3}):
MomentStandard NormalScaled Logistic (Var=1)
Mean (\mu_1)00
Variance (\mu_2)11
(\mu_3 / \mu_2^{3/2})00
Excess (\mu_4 / \mu_2^2 - 3)01.2
This table illustrates the shared symmetry but heavier tails of the logistic distribution.

Quantile function

The , also known as the inverse , of the logistic distribution with \mu and s > 0 is given by Q(p; \mu, s) = \mu + s \ln\left(\frac{p}{1-p}\right), \quad 0 < p < 1. This expression provides the value x such that the cumulative probability up to x equals p. The formula arises from the algebraic inversion of the cumulative distribution function. Setting F(x; \mu, s) = p and solving for x yields the logit transformation scaled and shifted by the parameters, as the logistic CDF is the inverse of this operation. Key quantiles include the median Q(0.5; \mu, s) = \mu, reflecting the distribution's symmetry around the location parameter. The interquartile range, Q(0.75; \mu, s) - Q(0.25; \mu, s) = 2s \ln 3 \approx 2.197s, measures the spread of the central 50% of the distribution and scales linearly with s. In simulation, the quantile function enables efficient generation of logistic random variables through inverse transform sampling: if U is a standard uniform random variable on (0,1), then X = \mu + s \ln\left(\frac{U}{1-U}\right) follows the logistic distribution with parameters \mu and s. This method is particularly useful for and bootstrapping in statistical modeling. The closed-form expression ensures numerical stability across p \in (0,1), with the logit-based computation avoiding overflow or cancellation issues common in other distributions, even for extreme quantiles near 0 or 1. This property underpins its role in quantile regression, where it relates to the logit link function for modeling probabilities.

Characteristic function

The characteristic function of a logistic random variable X with location parameter \mu and scale parameter s > 0 is \phi_X(t) = e^{i \mu t} \frac{\pi s t}{\sinh(\pi s t)}, where \sinh denotes the hyperbolic sine function. An equivalent representation employs the gamma function: \phi_X(t) = e^{i \mu t} \Gamma(1 - i s t) \Gamma(1 + i s t). This equivalence follows from the identity |\Gamma(1 + i y)|^2 = \pi y / \sinh(\pi y) for real y, which is a consequence of the reflection formula for the gamma function. To derive the characteristic function, compute the Fourier transform of the probability density function, \phi_X(t) = \mathbb{E}[e^{i t X}] = \int_{-\infty}^{\infty} e^{i t x} f_X(x) \, dx, where f_X(x) = \frac{e^{-(x-\mu)/s}}{s [1 + e^{-(x-\mu)/s}]^2}. One rigorous method avoids direct contour integration by expanding the standard logistic density (s=1, \mu=0) as an alternating infinite series of Laplace densities: f(x) = \sum_{k=1}^{\infty} (-1)^{k-1} \frac{k}{2} e^{-k |x|}. The characteristic function then emerges as the corresponding series summation, leveraging the known characteristic function of the Laplace distribution and identities from integral tables. For the general case, apply the location-scale transformation to adjust for \mu and s. The hyperbolic form highlights key analytic : \phi_X(t) is defined and continuous for all real t, but its analytic to the features simple poles along the imaginary at t = i k / s for nonzero integers k, arising from the zeros of \sinh(\pi s t). These poles, absent in the entire (pole-free) Gaussian characteristic function of the , underscore the logistic's heavier tails and departure from Gaussianity. Derivatives yield the cumulants via \kappa_n = \frac{1}{i^n} \frac{d^n}{dt^n} \log \phi_X(t) \big|_{t=0}. Using the gamma representation, \log \phi_X(t) = i \mu t + \log \Gamma(1 - i s t) + \log \Gamma(1 + i s t), the derivatives involve polygamma functions, as the digamma function \psi(z) = \frac{d}{dz} \log \Gamma(z) and higher-order polygamma functions \psi^{(n-1)}(z) = \frac{d^n}{dz^n} \log \Gamma(z) appear in the expansion.

Parameterizations and estimation

Location-scale parameterization

The logistic distribution is commonly expressed in location-scale form using a μ ∈ ℝ, which serves as the mean and , and a positive s > 0, which controls the dispersion of the distribution. The standard deviation of this parameterization is given by s π / √3, reflecting the fixed variance of π²/3 for the standard logistic distribution (with μ = 0 and s = 1) scaled by s². This form facilitates modeling symmetric, bell-shaped data with heavier tails than the normal distribution, maintaining the characteristic S-shaped under shifts and stretches. An alternative rate parameterization replaces the scale s with a rate parameter β = 1/s > 0, which emphasizes the steepness of the distribution's rise and is particularly useful in contexts involving growth rates or odds ratios. In this variant, the adjusts to f(x) = β e^{-β (x - μ)} / [1 + e^{-β (x - μ)}]^2, preserving the location μ while inverting the for interpretive convenience in rate-based models. The logistic distribution belongs to the location-scale family, exhibiting shape invariance under affine transformations: if X follows a logistic distribution with parameters μ and s, then for any a ≠ 0 and b ∈ ℝ, the transformed variable Y = aX + b follows a logistic distribution with parameters aμ + b and |a|s. This property ensures that the distribution's qualitative features, such as and , remain unchanged after linear rescaling and shifting. Historically, the logistic form traces back to Pierre-François Verhulst's work on , where he introduced a growth model incorporating a rate parameter to describe self-limiting toward a , laying the foundation for the distribution's parameterization in ecological and demographic applications. In Verhulst's original formulation, the growth rate r corresponds conceptually to the inverse scale (akin to β), highlighting the distribution's roots in bounded growth processes. Conversions between the scale and rate parameterizations are straightforward, as summarized below:
Scale Parameter (s)Rate Parameter (β)
s > 0β = 1/s
Standard (s = 1)β = 1
Variance = s² π² / 3Variance = π² / (3 β²)
In , the rate parameter β relates to the change in log-odds per unit predictor, with odds ratios given by e^β.

Maximum likelihood estimation

The maximum likelihood estimates of the \mu and s > 0 for the logistic distribution are obtained by maximizing the L(\mu, s \mid x_1, \dots, x_n) = \prod_{i=1}^n \frac{1}{s} \cdot \frac{\exp\left( -\frac{x_i - \mu}{s} \right)}{\left(1 + \exp\left( -\frac{x_i - \mu}{s} \right)\right)^2}, or equivalently, the log-likelihood \ell(\mu, s \mid x_1, \dots, x_n) = -n \log s - \frac{1}{s} \sum_{i=1}^n (x_i - \mu) - 2 \sum_{i=1}^n \log \left(1 + \exp\left( -\frac{x_i - \mu}{s} \right)\right). Setting the partial derivatives (score equations) to zero yields a of two coupled s with no closed-form : \sum_{i=1}^n \frac{1}{1 + \exp\left( \frac{x_i - \hat{\mu}}{ \hat{s} } \right)} = \frac{n}{2}, \sum_{i=1}^n z_i \frac{\exp(z_i) - 1}{1 + \exp(z_i)} = n, where z_i = (x_i - \hat{\mu}) / \hat{s} for i = 1, \dots, n, and \hat{\mu}, \hat{s} denote the maximum likelihood estimates. The first equation implies that \hat{\mu} is approximately the sample for large n, while \hat{s} requires solving the second , which can involve functions in certain derivations or approximations. Since the equations lack a closed form, numerical methods such as Newton-Raphson are typically employed to solve the system jointly, starting from initial values like the sample mean and standard deviation (adjusted for the logistic variance \pi^2/[3](/page/3)). Under standard regularity conditions (including finite and differentiability of the ), the maximum likelihood (\hat{\mu}, \hat{s}) is consistent and asymptotically : \sqrt{n} ((\hat{\mu}, \hat{s}) - (\mu, s)) \xrightarrow{d} \mathcal{N}(0, I(\mu, s)^{-1}), where I(\mu, s) is the Fisher information matrix. Relative to the method of moments, which equates sample moments to theoretical ones (\hat{\mu}_\text{MOM} = \bar{x}, \hat{s}_\text{MOM} = \frac{\sqrt{3} \hat{\sigma}}{\pi}, where \hat{\sigma} is the sample standard deviation), the maximum likelihood approach yields asymptotically more efficient estimators.

Applications

Logistic regression

Logistic regression is a used for problems, where the probability of the outcome being in one category is modeled using the (CDF) of the logistic distribution. Specifically, for a response Y \in \{0, 1\} and predictors X, the model is specified as P(Y=1 \mid X) = F(X \beta), where F is the logistic CDF and \beta is the vector of coefficients. This formulation arises because the logistic CDF, F(z) = \frac{1}{1 + e^{-z}}, produces a sigmoid-shaped probability curve that is bounded between 0 and 1. Taking the inverse (log-odds) transform yields the link function: \log\left(\frac{p}{1-p}\right) = X \beta, where p = P(Y=1 \mid X), allowing the log-odds to be expressed as a of the predictors. This model can be interpreted through a latent variable framework, where an unobserved continuous variable Y^* = X \beta + \epsilon determines the observed binary outcome via Y = 1 if Y^* > 0 and Y = 0 otherwise, with \epsilon following a standard logistic distribution (mean 0, variance \pi^2 / 3). The logistic error assumption provides symmetric tails and ensures the difference in probabilities between logit and probit models is typically small, except in the extremes. Historically, logistic regression was introduced by David Cox in 1958 as a method for analyzing binary sequences in dose-response studies, building on earlier uses of the logistic function in bioassay and population growth models. Cox proposed the logit link to handle non-linear relationships in binary data, enabling maximum likelihood estimation and facilitating comparisons across groups. Parameter estimation for \beta is typically performed via maximum likelihood, as no closed-form solution exists. A common algorithm is (IRLS), which approximates the likelihood through weighted linear regressions, updating weights based on current probability estimates until convergence. Alternatively, gradient-based methods like can be used for large datasets, minimizing the negative log-likelihood. Compared to the , which uses the normal CDF, offers advantages in interpretability and computational simplicity. The coefficients directly correspond to changes in log-, exponentiating to odds ratios (e^{\beta_j}) that quantify multiplicative effects on the odds of the outcome for a unit change in predictor j, holding others constant. The logistic CDF also has a closed algebraic form, easing numerical evaluations over the probit's integral-based .

Natural and social sciences

In physics, the logistic distribution models processes exhibiting bounded patterns, such as particle spread in constrained environments where the follows a logistic . For instance, processes derived from birth-death mechanisms have been used to describe logistic in physical systems, capturing effects in propagation dynamics. Additionally, the Fermi-Dirac distribution in quantum statistics for fermions adopts the form of the logistic , representing the average occupation number of energy states as f(E) = \frac{1}{1 + e^{(E - \mu)/kT}}, which approximates the of the logistic distribution scaled appropriately. In , the , particularly through its logarithmic transformation as the , fits frequency and rainfall data due to its heavier tails compared to the normal distribution, enabling better of extreme events. This approach was evaluated for frequency analysis in the late , building on earlier U.S. Army Corps of Engineers studies from the 1970s that explored various heavy-tailed distributions for hydrologic extremes in river basin planning. In , the logistic distribution underpins extensions of the Verhulst equation for , where the cumulative distribution function's shape models bounded growth toward in finite environments. Originally proposed by Pierre-François Verhulst in 1838 to describe human population increase, the deterministic logistic model \frac{dN}{dt} = rN \left(1 - \frac{N}{K}\right) has been generalized to diffusion processes incorporating noise, allowing simulation of fluctuations around the logistic mean in species populations. These variants reveal persistence and thresholds influenced by environmental variability. In the social sciences, the logistic distribution informs the for competitive outcomes, such as chess, where win probabilities are modeled using the to estimate relative player strengths based on rating differences. Developed by and formalized in his 1978 analysis, the expected score for player A against B is E_A = \frac{1}{1 + 10^{(R_B - R_A)/400}}, leveraging the logistic curve for its mathematical tractability over the normal distribution in pairwise comparisons. For income distribution modeling, variants like the log-logistic have been applied to capture the skewed, heavy-tailed nature of wage data across socioeconomic systems, providing insights into patterns in empirical datasets. A notable case study involves epidemic modeling, where the logistic distribution approximates the trajectory of infectious cases in (Susceptible-Infectious-Recovered) frameworks, particularly when incidence follows a logistic form to account for saturation in contacts. This connection allows the cumulative infected proportion to follow a generalized , facilitating parameter estimation from outbreak data as seen in analyses of dynamics. The briefly references the bounded growth from initial spread to thresholds in such models.

Extreme value connections

The logistic distribution maintains a significant connection to extreme value theory through its representation as a location-scale transformation of the difference between two independent Gumbel-distributed random variables. Specifically, if G_1 and G_2 are independent standard Gumbel random variables with cumulative distribution function (CDF) F_G(g) = \exp(-\exp(-g)), then the random variable L = G_1 - G_2 follows a standard logistic distribution with CDF F_L(l) = \frac{1}{1 + \exp(-l)}. This relationship underscores the logistic distribution's utility in modeling differences between extreme values, such as in paired comparisons or choice models where extremes represent maximum utilities. The logistic distribution exhibits a min-stable property within , particularly when considering stability under minima with respect to certain random sample sizes. It is the unique symmetric distribution that is both maximum-stable and minimum-stable with respect to sequences of positive integer-valued random variables, such as geometric distributions. For and identically distributed (i.i.d.) logistic random variables X_1, \dots, X_n, the minimum M_n = \min(X_1, \dots, X_n) has CDF F_{M_n}(x) = 1 - [1 - F(x)]^n, where F is the logistic CDF; while this does not exactly replicate the logistic form for fixed n, the property holds precisely for random n from appropriate distributions, preserving the logistic shape up to location-scale adjustments. This stability characterizes the logistic among symmetric distributions and facilitates modeling of order statistics in extreme scenarios. In , the logistic distribution serves as a parent distribution in the domain of attraction of the for both maxima and minima, due to its exponential tail behavior. The normalized minima of i.i.d. logistic variables, specifically M_n^* = a_n (M_n - b_n) with normalizing constants a_n = 1/s and b_n = \mu - s \log n (for \mu and scale s), converge in distribution to a standard for minima with CDF \exp(-\exp(x)). This limiting form enables the logistic to model the tails of minima in large samples, providing a bridge to broader extreme value applications. These extreme value connections find practical use in , where the logistic distribution models failure times of components with monotonically increasing hazard rates that approach a constant in the upper tail. For instance, it has been applied to lifetime data of mechanical valves.

Normal and other approximations

The logistic distribution with \mu and s > 0 is frequently approximated by a through moment matching, yielding a with \mu and variance \sigma^2 = \pi^2 s^2 / 3, which exactly equals the variance of the logistic distribution. This approximation leverages the similar bell-shaped forms of their probability density functions, though the logistic has heavier tails ( of 4.2 compared to the normal's 3). The standard logistic density, f(x) = e^{-x} / (1 + e^{-x})^2, can also be expressed as f(x) = \frac{1}{4} \sech^2(x/2), relating it directly to the hyperbolic squared function and facilitating certain analytical or numerical approximations involving hyperbolic identities. For sums of independent and identically distributed logistic random variables, the central limit theorem implies convergence in distribution to a normal after standardization, as the logistic has finite variance. The Berry–Esseen theorem quantifies the rate of this convergence, bounding the supremum distance between the cumulative distribution functions by C \rho / (\sigma^3 \sqrt{n}), where \rho = \mathbb{E}[|X - \mu|^3] is the third absolute central moment (approximately 9.3 for the standard logistic), \sigma is the standard deviation (\pi / \sqrt{3}), n is the number of terms, and C \approx 0.4748 is a universal constant. This bound ensures the approximation error decreases as O(1/\sqrt{n}), making it reliable for moderate to large n. In finite-sample settings, the logistic distribution can be approximated by a to better account for its heavier tails relative to the normal. Pingel (2014) proposes such an approximation for a standard logistic random variable (mean 0, variance \pi^2 / 3) using a t-distribution with scaled , demonstrating improved fit in applications like the of parameters. Edgeworth expansions enhance the normal approximation by including higher cumulants, such as the logistic's zero and excess of 1.2, yielding series corrections that particularly refine tail probabilities; for instance, the first-order expansion adjusts the density via weighted by cumulants divided by factorials of powers of n. These approximations are selected based on context: the logistic form is favored for its closed-form cumulative distribution function and interpretability in models like logistic regression, whereas the normal offers greater tractability for large-sample inference and computational efficiency.

References

  1. [1]
    [PDF] Logistic Distribution - Paul Johnson Homepage
    Jun 10, 2013 · The Logistic distribution is a continuous probability density function that is symmetric and uni-modal. It is similar in appearance to the ...
  2. [2]
    [PDF] arXiv:2108.07036v1 [math.ST] 16 Aug 2021
    Aug 16, 2021 · The logistic distribution apparently found its origin in the mid-nineteenth century in the writings of Verhulst (1838,. 1845). Since then it ...
  3. [3]
    [PDF] The Origins of Logistic Regression - Tinbergen Institute
    Analyses linking binary discrete responses to several covariates became known as logistic regression. This wider acceptance was greatly helped by the advent of ...
  4. [4]
    The Logistic Distribution
    The logistic distribution has been used for growth models, and is used in a certain type of regression known as the logistic regression.Logistic Probability Density... · Confidence Bounds · Bounds On The Parameters
  5. [5]
    [PDF] LOGPDF
    Mar 20, 1997 · The standard form of the logistic probability density function is: ... LOGCDF. = Compute the logistic cumulative distribution function.
  6. [6]
    Probability Playground: The Logistic Distribution
    The cdf of the logistic distribution is very similar to the that of the normal distribution, with the same property that F(μ - x) = 1 - F(μ + x).
  7. [7]
    Probability Playground: The Standard Logistic Distribution
    The standard logistic distribution is a logistic distribution with location parameter 0 and scale parameter 1. It represents the log odds of a probability p ...Missing: definition | Show results with:definition
  8. [8]
    [PDF] LOGCDF
    Mar 20, 1997 · Compute the standard logistic (i.e, mean=0, sd=π/sqrt(3)) cumulative distribution function. DESCRIPTION. The standard form of the logistic ...
  9. [9]
  10. [10]
    [PDF] The Origins of Logistic Regression
    The logistic function was discovered anew in 1920 by Pearl and Reed in a study of the population growth of the United States. They were unaware of Verhulst's ...
  11. [11]
    Logistic Distribution -- from Wolfram MathWorld
    The logistic distribution is implemented in the Wolfram Language as LogisticDistribution[mu, beta]. The mean, variance, skewness, and kurtosis excess are. mu, = ...
  12. [12]
    The Standard Logistic Distribution - Random Services
    It follows that the excess kurtosis of Z is kurt ( Z ) − 3 = 6 5 . Related Distributions. The standard logistic distribution has the usual connections with the ...
  13. [13]
    Quantile-Parameterized Distributions for Expert Knowledge Elicitation
    Mar 31, 2025 · For example, distribution based on logit may exhibit greater numerical stability because of its simple closed-form quantile function, which does ...
  14. [14]
    (PDF) A Commentary on the Logistic Distribution - ResearchGate
    Sep 27, 2021 · The paper provides a series representation of the logistic probability density function in terms of differently scaled double exponential ...
  15. [15]
    The early origins of the logit model - ScienceDirect.com
    This paper describes the origins of the logistic function and its history ... In logistic regression this is generalized to several determinants of P; Z=x ...
  16. [16]
    Fitting Logistic Parameters via MLE - Real Statistics Using Excel
    Describes how to find logistic distribution parameters that best fit a data set using maximum likelihood estimation (MLE) in Excel.Basic Concepts · Example · Newton's Method
  17. [17]
    The Regression Analysis of Binary Sequences - jstor
    Cox's paper seems likely to result in a much wider acceptance of the logistic function as a regression model. I have never been a partisan in the probit v ...
  18. [18]
    Log Odds and the Interpretation of Logit Models - PMC - NIH
    Many researchers prefer to estimate logit rather than probit models because of the odds ratio interpretation of the logit coefficients. The odds for individual ...
  19. [19]
    Logistic Growth Described by Birth-Death and Diffusion Processes
    We consider the logistic growth model and analyze its relevant properties, such as the limits, the monotony, the concavity, the inflection point, ...
  20. [20]
    The Fermi-Dirac Distribution - HyperPhysics
    The Fermi-Dirac distribution applies to fermions, particles with half-integer spin which must obey the Pauli exclusion principle.Missing: logistic | Show results with:logistic
  21. [21]
    Log-logistic flood frequency analysis - ScienceDirect.com
    The log-logistic (LLG) distribution is evaluated for flood frequency analysis. Some of its properties and methods of parameter estimation are given.
  22. [22]
    Verhulst and the logistic equation (1838) - SpringerLink
    In 1838 the Belgian mathematician Verhulst introduced the logistic equation, which is a kind of generalization of the equation for exponential growth.
  23. [23]
    [1506.01137] Stochastic dynamics and logistic population growth
    Jun 3, 2015 · It depends on two parameters, the intrinsic growth rate and the carrying capacity. These parameters can be estimated for different populations ...Missing: parameterization | Show results with:parameterization
  24. [24]
    [PDF] Stochastic Models of Wage Distributions: Empirical Comparison
    Log-logistic distribution is the probability distribution of a random variable whose logarithm has a logistic distribution. The formula (3) represents the ...
  25. [25]
    [PDF] 3 Logit
    The difference between two extreme value variables is distributed logistic. That is, ifεnj andεni are iid extreme value, thenε∗ nji = εnj − εni follows ...
  26. [26]
    Characterization of the logistic and loglogistic distributions by ...
    Jul 14, 2016 · The logistic distribution is proved to be the only symmetric distribution which is maximum stable with respect to each member of a sequence of ...
  27. [27]
    New Logistic Family of Distributions: Applications to Reliability ...
    Aug 18, 2025 · In reliability engineering, effectively modeling lifetime data is crucial for understanding failure mechanisms and predicting product ...
  28. [28]
    Handbook of the Logistic Distribution - 1st Edition - N. Balakrishnan
    Free deliveryDec 3, 2019 · This book highlights various theoretical developments on logistic distribution, illustrates the practical utility of these results, ...Missing: pdf | Show results with:pdf
  29. [29]
    [PDF] A Compendium of Common Probability Distributions - Rice Statistics
    The logistic distribution is sometimes called the Sech-squared distribution. Characterizations. 1. The logistic law of growth is described by the following ...
  30. [30]
    [PDF] Berry–Esseen Bounds for Independent Random Variables
    The upper bound on the smallest possible value of C has decreased from Esseen's original estimate of 7.59 to its current value of. 0.4785 by Tyurin (2010).Missing: logistic | Show results with:logistic<|control11|><|separator|>
  31. [31]
    Some approximations of the logistic distribution with application to ...
    Some approximations of the logistic distribution with application to the covariance matrix of logistic regression. Author links open overlay panel. Ronnie ...