Fact-checked by Grok 2 weeks ago

Poisson distribution

The Poisson distribution is a discrete probability distribution that models the number of events occurring within a fixed interval of time or space, where these events happen independently and at a constant average rate, making it ideal for describing rare or random occurrences such as arrivals or defects. It is characterized by a single parameter, λ (lambda), which represents the expected number of events in the interval, serving as both the mean and variance of the distribution. The probability mass function for the Poisson distribution is given by
P(X = k) = \frac{e^{-\lambda} \lambda^k}{k!}
where k is a non-negative integer (0, 1, 2, ...), e is the base of the natural logarithm (approximately 2.71828), and \lambda > 0.
Named after the , the distribution was introduced in his 1837 work Recherches sur la probabilité des jugements en matière criminelle et en matière civile, where it arose as an approximation to the for large numbers of trials with small success probabilities./05:_Discrete_Probability_Distributions/5.06:_Poisson_Distribution) An early empirical demonstration came from Ladislaus Bortkiewicz in 1898, who applied it to analyze the frequency of horse-kick deaths among officers, showing a close fit to the observed data. The distribution's is the largest less than or equal to λ (or both λ and λ-1 if λ is an ), and its lacks a simple closed form, often requiring numerical computation. Key assumptions for the Poisson distribution include that are , the λ remains constant over the interval, and the probability of more than one in a very small subinterval is negligible. It approximates the when the number of trials n is large and the success probability p is small such that np = λ is fixed, providing a simpler model for count data. In practice, the maximum likelihood estimator for λ is the sample mean of the observed counts. The Poisson distribution finds wide applications in fields like for modeling customer arrivals, reliability engineering for defect counts, and for disease incidence rates, as well as in for predicting rare events like stock trades. It extends to generalized forms, such as the compound Poisson distribution for summed random variables, and serves as the basis for in analyzing count-based outcomes in generalized linear models.

History and Motivation

Historical Development

The Poisson distribution has roots in early attempts to approximate binomial probabilities under conditions of rare events. In 1711, provided an approximation for the when the probability of success p is small and the number of trials n is large, yielding P(K = k) \approx e^{-\mu} \frac{\mu^k}{k!} where \mu = np. This formula, derived in his work De Mensura Sortis seu; de Probabilitate Eventuum in Ludis a Casibus Dependentium, effectively described the distribution of rare occurrences, though de Moivre did not explicitly recognize it as a distinct limiting form. The distribution was formally introduced by in 1837 as part of his treatise Recherches sur la probabilité des jugements en matière criminelle et en matière civile, where he derived it as a limit of the and termed it the "law of small numbers" to model infrequent events in legal and social contexts. Poisson's derivation emphasized its application to probabilistic judgments in jury decisions and civil matters, establishing it as a tool for analyzing sparse data. Its adoption as a practical statistical tool gained prominence in the late through Ladislaus Bortkiewicz's 1898 monograph Das Gesetz der kleinen Zahlen, which applied Poisson's law to empirical data, including the famous analysis of deaths from horse kicks in the (1875–1894), demonstrating close fits between observed frequencies and theoretical predictions. This work popularized the distribution among statisticians for rare event modeling. The term "Poisson distribution" emerged in the early , honoring despite de Moivre's precursor contributions. By the mid-20th century, the distribution evolved into a foundational element of stochastic processes, with the Poisson process formalized in the 1940s by figures like , enabling its use in modeling continuous-time random events such as arrivals in and point processes.

Derivation from Limit

The describes the number of successes in a fixed number of independent trials, each with success probability p. The (PMF) for a random variable X \sim \text{[Binomial](/page/Binomial)}(n, p) is given by P(X = k) = \binom{n}{k} p^k (1-p)^{n-k}, \quad k = 0, 1, \dots, n. To derive the Poisson distribution, consider the limit as the number of trials n \to \infty and the success probability p \to 0, while keeping the expected number of successes \lambda = np fixed and finite. This scenario models , where successes are infrequent but the total rate \lambda remains constant. Substitute p = \lambda / n into the binomial PMF: P(X = k) = \binom{n}{k} \left(\frac{\lambda}{n}\right)^k \left(1 - \frac{\lambda}{n}\right)^{n-k} = \frac{n!}{k!(n-k)!} \cdot \frac{\lambda^k}{n^k} \cdot \left(1 - \frac{\lambda}{n}\right)^{n-k}. Rewrite the binomial coefficient as \binom{n}{k} = \frac{n(n-1)\cdots(n-k+1)}{k! \, n^k}. As n \to \infty, the term \frac{n(n-1)\cdots(n-k+1)}{n^k} \to 1 for fixed k, since each factor (n-j)/n \to 1 for j = 0, 1, \dots, k-1. Additionally, \left(1 - \frac{\lambda}{n}\right)^n \to e^{-\lambda} by the standard definition of the exponential function, and \left(1 - \frac{\lambda}{n}\right)^{-k} \to 1. Thus, \lim_{n \to \infty} P(X = k) = \frac{\lambda^k}{k!} \cdot e^{-\lambda}, which is the PMF of the Poisson distribution with parameter \lambda: P(Y = k) = e^{-\lambda} \frac{\lambda^k}{k!}, \quad k = 0, 1, 2, \dots. A more rigorous expansion for the binomial coefficient can use Stirling's approximation n! \approx \sqrt{2\pi n} (n/e)^n, but the direct suffices for fixed k. This derivation interprets the Poisson distribution as arising from many trials with infinitesimally small success probabilities, such as modeling events in small time intervals or spatial units where occurrences are rare but aggregate to a fixed rate \lambda. For example, the number of successes in n trials each with success probability \lambda/n approximates a \text{Poisson}(\lambda) random variable when n is large.

Definition

Probability Mass Function

The Poisson distribution is a discrete probability distribution that expresses the probability of a given number of events occurring in a fixed interval of time or space if these events occur with a known constant mean rate and independently of the time since the last event. The (PMF) for a Poisson random variable X with parameter \lambda > 0 is given by P(X = k \mid \lambda) = \frac{e^{-\lambda} \lambda^k}{k!} for k = 0, 1, 2, \dots, and P(X = k) = 0 for k < 0. The support of the distribution is the set of non-negative integers, reflecting its application to counting discrete events. The term e^{-\lambda} serves as the normalizing constant, ensuring that the infinite sum of probabilities equals 1, since \sum_{k=0}^{\infty} \frac{\lambda^k}{k!} = e^{\lambda}, so e^{-\lambda} e^{\lambda} = 1. The probabilities satisfy a recursive relation: P(X = k+1) = P(X = k) \cdot \frac{\lambda}{k+1} for k = 0, 1, 2, \dots. The shape of the PMF is right-skewed for small values of \lambda (e.g., \lambda \leq 5), with the mode at 0 or nearby integers, and becomes approximately symmetric as \lambda increases (e.g., \lambda \geq 25), though it remains slightly skewed.

Parameter and Support

The Poisson distribution is parameterized by a single positive real number λ, which represents the average rate of occurrence of events within a fixed unit interval, such as time or space, and serves as both the mean and variance of the distribution. This λ quantifies the expected number of events in that interval, making it essential for modeling rare or random events like arrivals or defects. For the distribution to be well-defined, λ must be strictly greater than zero; as λ approaches 0, the probability mass concentrates entirely at zero, reflecting scenarios with negligible event likelihood. Conversely, as λ tends to infinity, the distribution approximates a normal distribution via the central limit theorem, useful for large-scale event modeling. The support of the Poisson random variable X is the set of non-negative integers, denoted as X \in \{0, 1, 2, \dots\}, which aligns with its application to count data where negative or fractional values are impossible. This discrete domain ensures the distribution captures whole-number occurrences, such as the number of emails received in an hour. Regarding units, λ is dimensionless when modeling pure counts but can be scaled to represent rates, for instance, events per hour or per square kilometer, depending on the context of the unit interval. A notable special case occurs when λ = 1, where the probability of zero events is e^{-1} \approx 0.3679, illustrating the likelihood of no occurrences in an interval expected to have exactly one event on average.

Properties

Moments and Central Tendency

The Poisson distribution, denoted as X \sim \text{Poisson}(\lambda) where \lambda > 0 is the rate parameter, exhibits several key measures of and dispersion that characterize its shape and behavior. The , or , of a Poisson is exactly equal to its parameter \lambda, reflecting the average number of events in the fixed . This property arises directly from the and underscores the distribution's role in modeling count data with a stable long-run average. The variance of X is also \lambda, a hallmark feature known as equidispersion, where the spread of the distribution matches its . This equality implies that the standard deviation is \sqrt{\lambda}, and it distinguishes the Poisson from overdispersed alternatives like the . In practical applications, such as or rare event modeling, this variance-mean equality facilitates straightforward without additional parameters. For measures of beyond the , the of the Poisson distribution—the value with the highest probability—depends on whether \lambda is an . If \lambda is not an , the is uniquely \lfloor \lambda \rfloor; if \lambda is an , both \lambda and \lambda - 1 are modes, each with equal probability. This unimodal structure, peaking near \lambda, aligns with the distribution's asymmetry for small \lambda and near-symmetry for large \lambda. The , which divides the probability mass into two equal parts, is typically \lfloor \lambda \rfloor for most values, though a refined for large \lambda is \lfloor \lambda + 1/3 - 1/(50\lambda) \rfloor, providing a more precise central location in asymptotic analyses. These location parameters highlight the Poisson's utility in discrete settings where exact counts matter, such as in ecology for species abundance. Regarding higher-order moments, the skewness \gamma_1 of the Poisson distribution is \lambda^{-1/2}, indicating right-skewness that diminishes as \lambda increases, approaching zero for large \lambda and resembling a normal distribution. This decreasing skewness explains the distribution's transition from a highly asymmetric form (e.g., for \lambda = 1, skewness ≈ 1) to near-symmetry (e.g., for \lambda = 100, skewness ≈ 0.1), a property exploited in central limit theorem applications for Poisson approximations. Complementing this, the coefficient of variation—defined as the standard deviation divided by the mean—is $1/\sqrt{\lambda}, which similarly decreases with \lambda, quantifying the relative variability and aiding comparisons across different rate scales in fields like reliability engineering. These moments can be derived using the probability generating function, though their direct computation from the mass function confirms their parameter dependence.

Generating Functions

The probability generating function (PGF) of a Poisson random variable X with parameter \lambda > 0 is defined as G_X(s) = \mathbb{E}[s^X] for |s| \leq 1, and it equals e^{\lambda(s-1)}. This form arises from substituting the probability mass function into the expectation: G_X(s) = \sum_{k=0}^{\infty} s^k \cdot e^{-\lambda} \frac{\lambda^k}{k!} = e^{-\lambda} \sum_{k=0}^{\infty} \frac{(\lambda s)^k}{k!} = e^{-\lambda} e^{\lambda s} = e^{\lambda(s-1)}. The PGF encapsulates the distribution's probabilities and facilitates analysis of sums of independent Poissons. The moment generating function (MGF) of X is M_X(t) = \mathbb{E}[e^{tX}] for t \in \mathbb{R}, given by e^{\lambda(e^t - 1)}. It is derived similarly: M_X(t) = \sum_{k=0}^{\infty} e^{tk} \cdot e^{-\lambda} \frac{\lambda^k}{k!} = e^{-\lambda} \sum_{k=0}^{\infty} \frac{(\lambda e^t)^k}{k!} = e^{-\lambda} e^{\lambda e^t} = e^{\lambda(e^t - 1)}. Moments of X are obtained by evaluating derivatives of the MGF at t = 0; for instance, the first derivative M_X'(t) = \lambda e^t \cdot e^{\lambda(e^t - 1)} yields \mathbb{E}[X] = M_X'(0) = \lambda. Higher-order derivatives provide further moments, confirming properties like the raw moments through successive differentiation. The cumulant generating function is the natural logarithm of the MGF, K_X(t) = \log M_X(t) = \lambda(e^t - 1). Its Taylor series expansion around t = 0 has coefficients that are the s of X, and for the Poisson, all cumulants equal \lambda. In particular, the first is the \lambda, and the second is the variance \lambda, establishing the equidispersion property where variance equals the .

Sums of Independent Variables

A fundamental property of the Poisson distribution is its closure under addition for independent random variables. Specifically, if X_1, X_2, \dots, X_n are independent random variables where X_i \sim \text{Poisson}(\lambda_i) for i = 1, \dots, n, then their sum S = \sum_{i=1}^n X_i follows a Poisson distribution with parameter \sum_{i=1}^n \lambda_i. This result can be proved using probability generating functions (PGFs). The PGF of a \text{Poisson}(\lambda) random variable is G(s) = e^{\lambda (s-1)}. Since the X_i are independent, the PGF of S is the product of the individual PGFs: G_S(s) = \prod_{i=1}^n G_{X_i}(s) = \prod_{i=1}^n e^{\lambda_i (s-1)} = e^{\left( \sum_{i=1}^n \lambda_i \right) (s-1)}, which is the PGF of a \text{Poisson}\left( \sum_{i=1}^n \lambda_i \right) random variable. This property extends naturally to Poisson processes. The superposition of independent Poisson processes with rates \lambda_1, \dots, \lambda_n—that is, the combined process counting events from all sources—is itself a Poisson process with rate \sum_{i=1}^n \lambda_i. For example, consider defects in manufactured items arising from multiple independent sources, such as different production stages, each modeled as a Poisson process with rate \lambda_i. The total number of defects observed follows a Poisson distribution with equal to the sum of the individual rates, facilitating aggregated analysis.

Inequality and Tail Bounds

The Poisson distribution exhibits strong concentration properties around its mean \lambda, making it amenable to tail bounds that quantify the probability of deviations. Concentration inequalities such as Chernoff bounds provide exponential upper bounds on these tail probabilities, leveraging the moment generating function E[e^{tX}] = e^{\lambda(e^t - 1)} for X \sim \text{Poisson}(\lambda). For the upper tail, a standard Chernoff bound states that for \delta > 0, P(X \geq (1 + \delta)\lambda) \leq \exp\left( -\frac{\lambda \delta^2}{2 + \delta} \right). This bound is derived by optimizing the Chernoff parameter t = \ln(1 + \delta) and is widely used due to its simplicity and tightness for moderate \delta. A more precise form, without simplification, is P(X \geq a) \leq \exp(a - \lambda (\lambda / a)^a) for a > \lambda, obtained by minimizing over t > 0. For the lower tail, the Chernoff bound yields P(X \leq (1 - \delta)\lambda) \leq \exp(-\lambda \delta^2 / 2) for $0 < \delta < 1. This follows from applying Markov's inequality to e^{-tX} with t = \ln(1 - \delta), providing an exponential decay that mirrors the upper tail but with a slightly looser constant. These bounds highlight the sub-Gaussian-like behavior of the Poisson despite its unbounded support. Hoeffding's inequality, originally for bounded independent random variables, adapts to the Poisson distribution via its representation as a limit of binomials or through direct moment bounds. An improved version tightens the by a factor of at least two, yielding exponential upper tail estimates of the form P(X \geq k) \leq \exp\left( - \frac{[\Phi^{-1}(1 - R)]^2}{2(m/k + k/m - 2)} \right) for X \sim \text{Poisson}(m) and k \geq m, where \Phi is the standard normal CDF and R relates to the tail probability. This adaptation enhances applicability in scenarios requiring uniform bounds across parameters. For the upper tail, Stirling's approximation k! \approx \sqrt{2\pi k} (k/e)^k facilitates computable bounds, particularly for large k > \lambda. A practical upper bound is P(X \geq k) \leq \frac{\lambda}{k+1} \cdot e^{-\lambda} \frac{\lambda^k}{k!}, which is numerically stable and avoids overflow for large \lambda. More refined Stirling-based estimates incorporate normal approximations for near-median tails, such as P(X \geq \lfloor \lambda \rfloor + n) \leq \exp(1/8\lambda) \cdot (1 + 1/\lambda) \cdot \sqrt{\lambda} \cdot \Phi\left( (n - 3/2)/\sqrt{2\lambda} \right) for n \geq 1 and \lambda \geq 2, bridging local and global tail behavior. These inequalities underpin for the Poisson distribution, where the rate function I(x) = x \ln(x/\lambda) - x + \lambda governs the P(X \geq x) \sim \exp(-\lambda I(x/\lambda)) for x \gg \lambda. In rare event modeling, such as queueing systems or , these bounds enable precise estimation of overflow probabilities, with applications to Poisson approximations in compound processes ensuring asymptotic equivalence under factorial cumulant constraints. For instance, in estimating large deviations for sums approximating Poissons, the relative error is controlled by terms like (x - \lambda)^{3/2}/\Delta, where \Delta > 0 bounds higher moments.

Connection to Binomial Distribution

The Poisson distribution with parameter \lambda arises as a limiting case of the binomial distribution \operatorname{Bin}(n, p) when the number of trials n is large, the success probability p is small, and the product \lambda = np is held fixed. In this regime, the probability mass function of the binomial converges pointwise to that of the Poisson, providing an exact limit in the sense of distribution. Conversely, the binomial can be viewed as a finite-n approximation to the Poisson when modeling scenarios with many independent rare events. The quality of this approximation is quantified by the distance d_{\mathrm{TV}}(\operatorname{Bin}(n,p), \operatorname{Pois}(\lambda)). A simple bound is d_{\mathrm{TV}} \leq p, which scales as \lambda / n. A practical for employing the Poisson approximation is to use it when np < 5 and p < 0.1, ensuring the conditions of large n and small p are met for reasonable accuracy. For instance, in manufacturing quality control, the number of defective items in a large batch of n products, each with a small defect probability p, follows approximately a \operatorname{Pois}(np) distribution; an example is inspecting 1000 light bulbs with a 0.1% defect rate, where \lambda = 1 and the Poisson simplifies computations for rare defects.

Role in Poisson Point Processes

The Poisson point process provides a fundamental framework in stochastic geometry where the Poisson distribution arises as the marginal law governing point counts. In a homogeneous Poisson point process defined on a space such as \mathbb{R}^d with constant intensity \lambda > 0, the number of points falling within any bounded region B of finite measure |B| (such as area in two dimensions or volume in three) follows a Poisson distribution with parameter \lambda |B|. This means the probability of observing exactly n points in B is given by P(N(B) = n) = \frac{(\lambda |B|)^n e^{-\lambda |B|}}{n!}, \quad n = 0, 1, 2, \dots A key property is the independence of counts across disjoint regions: if B_1, B_2, \dots, B_k are mutually disjoint bounded sets, then the random variables N(B_1), \dots, N(B_k) are independent, each Poisson-distributed with parameters \lambda |B_i|. This construction generalizes to the inhomogeneous , where the \lambda(\mathbf{x}) varies with location \mathbf{x}. Here, the number of points in a bounded region B is Poisson-distributed with mean equal to the integrated \int_B \lambda(\mathbf{x}) \, d\mathbf{x}, reflecting the expected total "rate" over the region. Counts in disjoint regions remain independent under this setup. In the temporal setting, where the process models events over time, the homogeneous case reduces to the classic , a process with interarrival times; the number of arrivals in a fixed interval of length t is then Poisson(\lambda t). The connection extends to spatial statistics, where Poisson point processes serve as null models for analyzing point patterns in fields like and , assuming complete spatial . Additionally, these processes underpin shot noise models, originally developed to describe fluctuations in electrical currents due to discrete emissions, as introduced by Campbell in his studies of thermionic . Superposition of independent Poisson point processes yields another Poisson point process with intensity equal to the sum of the individual .

Generalizations and Variants

The bivariate Poisson distribution provides a natural extension of the univariate Poisson distribution to model pairs of correlated count variables, such as the number of events occurring in two related processes. It is constructed by letting X = X_1 + Z and Y = X_2 + Z, where X_1, X_2, and Z are independent Poisson random variables with parameters \lambda_1, \lambda_2, and \lambda_0 \geq 0, respectively; the shared component Z induces positive correlation between X and Y. The marginal distributions of X and Y are Poisson with means \lambda_1 + \lambda_0 and \lambda_2 + \lambda_0, while the correlation coefficient is \rho = \lambda_0 / \sqrt{(\lambda_1 + \lambda_0)(\lambda_2 + \lambda_0)}. This model is particularly useful for bivariate count data exhibiting dependence, such as insurance claims in multiple categories or sports match outcomes. The construction was first detailed by Holgate in 1964, with further developments in estimation and applications appearing in subsequent works. The arises as a key overdispersed generalization of the Poisson distribution, accommodating scenarios where the variance exceeds the mean, which is common in real-world count data due to unobserved heterogeneity or clustering. Parameterized by mean \lambda > 0 and dispersion r > 0, it has P(K = k) = \binom{k + r - 1}{k} \left( \frac{\lambda}{\lambda + r} \right)^k \left( \frac{r}{\lambda + r} \right)^r for k = 0, 1, 2, \dots, yielding mean \lambda and variance \lambda + \lambda^2 / r. As r \to \infty, the distribution converges to the Poisson with \lambda, while smaller r increases overdispersion. This makes it a flexible alternative for modeling counts like incidences or traffic accidents, where Poisson assumptions fail. The negative binomial's role as an overdispersed Poisson variant is extensively analyzed in contexts by Cameron and (1998). The compound Poisson distribution generalizes the Poisson by considering the total as a random sum S = \sum_{i=1}^N Y_i, where N \sim \mathrm{[Poisson](/page/Poisson)}(\lambda) represents the number of clusters or events, and the Y_i are independent and identically distributed positive random variables (independent of N) with G_Y(s) = \mathbb{E}[s^Y]. The of S is then G_S(s) = \exp\left( \lambda (G_Y(s) - 1) \right), which highlights its infinitely divisible nature and utility in aggregating risks. If the Y_i are themselves Poisson, S follows a Neyman Type A distribution; more generally, it models compound events like total claim amounts in , where N counts claims and Y_i their sizes. This framework underpins much of and stochastic modeling. In non-commutative probability theory, the free Poisson distribution emerges as the free independence analog of the classical Poisson distribution, developed by Voiculescu in the mid-1980s to study operator algebras and random matrices. Unlike the classical case, free probability replaces commuting variables with freely independent non-commuting ones, leading to the free Poisson law—also known as the Marchenko-Pastur distribution—with R-transform \psi(z) = \lambda / (1 - z) for parameter \lambda > 0. It arises as the limiting spectral distribution of Wishart matrices and captures "free compounding" effects, with mean \lambda and variance \lambda(1 + \lambda). This variant has impacted random matrix theory and free stochastic processes, providing tools for non-commutative limits absent in classical probability. Voiculescu's foundational work (1985) established free probability, with the free Poisson detailed in subsequent developments. The Weibull-Poisson distribution extends the Poisson framework by compounding it with a Weibull mixing , yielding a flexible model for counts or lifetime data where events follow a non-homogeneous pattern. Defined via the or as a Poisson process with Weibull-distributed intensities, it accommodates monotone increasing or decreasing failure rates, making it suitable for analyzing censored count data in or , such as recurring failures over time. The resulting has support on non-negative integers and can exhibit under- or depending on Weibull shape parameters. This generalization was introduced by Cancho et al. (2011) for competing risks and long-term modeling.

Parameter Estimation

Maximum Likelihood Estimation

The maximum likelihood estimator (MLE) for the parameter \lambda of a Poisson distribution is derived from a sample of n independent and identically distributed observations x_1, x_2, \dots, x_n, each following \text{Poisson}(\lambda). The likelihood function is given by L(\lambda \mid x_1, \dots, x_n) = \prod_{i=1}^n \frac{e^{-\lambda} \lambda^{x_i}}{x_i!} = \frac{e^{-n\lambda} \lambda^{\sum x_i}}{\prod x_i!}. To maximize this, consider the log-likelihood \ell(\lambda) = -n\lambda + \left( \sum x_i \right) \log \lambda - \sum \log(x_i!), which is differentiated with respect to \lambda and set to zero: \frac{d\ell(\lambda)}{d\lambda} = -n + \frac{\sum x_i}{\lambda} = 0, yielding the MLE \hat{\lambda} = \bar{x} = \frac{\sum x_i}{n}, the sample . This estimator is unbiased, satisfying E[\hat{\lambda}] = \lambda, and achieves the Cramér-Rao lower bound, making it the (MVUE). Under standard regularity conditions, the MLE is asymptotically : as n \to \infty, \sqrt{n} (\hat{\lambda} - \lambda) \xrightarrow{d} \mathcal{N}(0, \lambda). For example, if counts of —such as arrivals at a service point over n fixed time intervals—are observed as x_1, \dots, x_n, then \hat{\lambda} provides the maximum likelihood estimate of the average arrival rate per interval.

Confidence Intervals for λ

intervals for the parameter λ of a Poisson distribution quantify the uncertainty in estimating the rate based on observed counts. For a random sample of n independent observations X_1, \dots, X_n from a Poisson(λ) distribution, the sample \bar{X} = \frac{1}{n} \sum_{i=1}^n X_i is the maximum likelihood estimator for λ, and intervals are typically constructed around an observed value \bar{x}. For large n or large λ, where the central limit theorem applies, a normal approximation provides a simple interval: \bar{x} \pm z_{\alpha/2} \sqrt{\bar{x}/n}, with z_{\alpha/2} the (1 - \alpha/2)-quantile of the standard normal distribution. This Wald interval relies on the asymptotic normality of \sqrt{n} (\bar{X} - \lambda) / \sqrt{\lambda} \approx \mathcal{N}(0, 1). However, it performs poorly for small n or small λ due to skewness and variance instability. Exact methods avoid approximations by inverting equal-tailed tests using the cumulative distribution function of the Poisson, often leveraging its relationship to the chi-squared distribution. For a single observation k \sim \mathrm{Poisson}(\lambda), the cumulative probability satisfies \sum_{j=0}^{k-1} e^{-\lambda} \lambda^j / j! = P(\chi^2_{2k} \geq 2\lambda), and similarly for the upper tail. The Garwood interval, a central exact method, uses these relations to yield: \begin{aligned} \text{Lower limit} &= \frac{1}{2} \chi^2_{\alpha/2, \, 2k}, \\ \text{Upper limit} &= \frac{1}{2} \chi^2_{1 - \alpha/2, \, 2(k+1)}, \end{aligned} where \chi^2_{p, \, \nu} is the p-quantile of the chi-squared distribution with \nu degrees of freedom. This is conservative, guaranteeing coverage at least $1 - \alpha. For n observations with total count S = \sum X_i \sim \mathrm{Poisson}(n\lambda), apply the formula with k replaced by S to obtain bounds for n\lambda, then divide by n for λ. This approach is the Poisson analog of the Clopper-Pearson interval for binomial proportions. The discrete nature of the Poisson leads to conservativeness in exact intervals, prompting adjustments like the mid-p method. This subtracts half the probability mass at the observed count from the tail probabilities before inverting the test, producing shorter intervals with expected coverage closer to the nominal $1 - \alpha without falling below it in most cases. For the Garwood setup, the mid-p lower limit solves \sum_{j=0}^{k-1} P(j; \lambda_L) + \frac{1}{2} P(k; \lambda_L) = \alpha/2, and similarly for the upper. It reduces overcoverage while maintaining good properties. Exact and mid-p intervals tend to be conservative for small λ (e.g., λ < 2), with actual coverage exceeding $1 - \alpha due to the step function of the discrete cumulative distribution, leading to wider intervals than necessary. Simulations confirm this overcoverage, particularly for low counts, where the Garwood interval's coverage can reach 0.96 or higher for nominal 95%. To address this, simulation-based alternatives, such as parametric bootstrapping or profile likelihood methods, generate empirical intervals tailored to achieve near-nominal coverage and shorter lengths, especially in small-sample settings.

Bayesian Inference

Prior and Posterior Distributions

In Bayesian inference for the Poisson distribution, the parameter λ represents the of occurrence, and the likelihood function arises from independent observations X_1, \dots, X_n \sim \text{Poisson}(\lambda), yielding a product of Poisson probabilities. The conjugate prior distribution for λ is the gamma distribution, parameterized as \text{Gamma}(\alpha, \beta), with probability density function proportional to \lambda^{\alpha-1} e^{-\beta \lambda} for \lambda > 0, where \alpha > 0 is the and \beta > 0 is the . This conjugacy ensures that the posterior distribution remains in the gamma family, facilitating analytical updates. Given the prior \lambda \sim \text{Gamma}(\alpha, \beta) and data x_1, \dots, x_n, the posterior distribution is \text{Gamma}\left(\alpha + \sum_{i=1}^n x_i, \beta + n\right), where the shape parameter updates by adding the total count of events \sum x_i and the rate parameter updates by adding the number of observations n. The posterior mean, which serves as the Bayes estimator under squared error loss, is then \frac{\alpha + \sum_{i=1}^n x_i}{\beta + n}, providing a weighted average that shrinks the sample mean toward the prior mean \alpha / \beta. For non-informative priors, the Jeffreys prior for the Poisson parameter λ is derived from the Fisher information and takes the form proportional to \lambda^{-1/2}, equivalent to an improper \text{Gamma}(1/2, 0) distribution. This prior is improper because its integral over (0, \infty) diverges, but it leads to proper posteriors for n \geq 1 and is invariant under reparameterization, making it suitable for objective Bayesian analysis. In hierarchical models involving multiple rates, such as when observations arise from distinct Poisson processes with parameters \lambda_1, \dots, \lambda_m, a common approach assumes each \lambda_i \sim \text{Gamma}(\alpha, \beta) independently, forming a Poisson-gamma hierarchical structure. This setup allows borrowing strength across groups by pooling information, with the marginal distribution for each count following a negative binomial, and enables full Bayesian inference on the shared hyperparameters \alpha and \beta to account for heterogeneity in rates.

Credible Intervals

In for the Poisson distribution, credible intervals for the rate parameter λ are derived from the , which is typically Gamma(α', β') under a conjugate Gamma , providing a probability statement about the location of λ given the data. The equal-tailed captures 100(1-α)% of the by taking the α/2 and 1-α/2 quantiles of the Gamma(α', β') distribution, often computed using the inverse . This interval is symmetric in probability tails and straightforward to calculate numerically. The highest posterior density (HPD) interval, in contrast, is the shortest interval that contains 100(1-α)% of the posterior mass, defined as the set {θ : π(θ | x) ≥ k} where k is chosen to achieve the desired coverage under a unimodal posterior. For the Gamma posterior, this can be found numerically by solving for equal density points at the boundaries. For large α', the Gamma(α', β') posterior can be approximated by a with mean α'/β' and variance α'/β'^2, yielding an approximate of mean ± z_{α/2} √(variance), where z_{α/2} is the standard . This asymptotic approach simplifies computation when the posterior is concentrated. As an example, consider a uniform improper , equivalent to , 0), which yields a posterior Gamma(1 + ∑x_i, n) for n independent Poisson observations with total events ∑x_i; a 95% equal-tailed is then obtained via the 0.025 and 0.975 quantiles of this posterior using the inverse CDF. Compared to frequentist intervals, Bayesian s, particularly those using informative priors, often exhibit superior coverage probabilities in small samples, approaching or exceeding nominal levels (e.g., 95%) where classical methods like approximations may undercover for low λ values.

Applications

Modeling Rare Events

The Poisson distribution serves as the law of , approximating the probability of observing k occurrences in a fixed when events are and each has a small probability, under the condition that the expected number of events \lambda remains finite as the number of trials increases.https://planetmath.org/lawofrareevents Specifically, for a process with n trials and success probability p = \lambda / n, as n \to \infty, the converges to P(K = k) \approx e^{-\lambda} \frac{\lambda^k}{k!}, where K counts the events.https://math.iisc.ac.in/~gadgil/MA261/notes/chapter-26.html This limit captures scenarios where events are infrequent yet their aggregate expectation is constant, such as defects in manufacturing or arrivals in queueing systems. A classic illustration involves meteor strikes on Earth, where the number of impacts exceeding a certain size in a given period follows a Poisson distribution due to their sporadic and independent nature, with \lambda estimated from historical crater records.https://www.scribbr.com/statistics/poisson-distribution/ Similarly, in genetics, the number of mutations in a human genome per generation adheres to this model; the per-base-pair germline mutation rate is approximately $1.2 \times 10^{-8}, making mutations rare at each site across the roughly 3 billion base pairs, yielding an expected total of about 70–80 de novo single-nucleotide variants as of 2025.https://www.nature.com/articles/s41586-025-08922-2 The model's validity hinges on the independence of events and the absence of clustering; it breaks down if occurrences are contagious or exhibit dependence, such as in spreads where one event triggers others.https://planetmath.org/lawofrareevents Historically, Ladislaus Bortkiewicz applied this in 1898 to analyze deaths from kicks in the across 200 corps-years, finding the counts followed a Poisson distribution with \lambda \approx 0.61, demonstrating the for infrequent military accidents. In risk analysis, the Poisson distribution models insurance claims from catastrophes like hurricanes or earthquakes, where claims arise from rare, high-severity events assumed to occur independently.https://www.casact.org/sites/default/files/database/proceed_proceed73_73146.pdf For instance, with an expected claim frequency of 0.1 per year, premiums are calculated using Poisson probabilities to ensure against zero or multiple losses in a policy period.https://www.casact.org/sites/default/files/database/proceed_proceed73_73146.pdf

Counting in Biology and Physics

In , the Poisson distribution models the random occurrence of mutations in bacterial populations, as demonstrated by the Luria-Delbrück experiment, which showed that mutations arise spontaneously at a constant rate per rather than in response to selection, following a Poisson process that leads to high variance in mutant counts across parallel cultures due to early "jackpot" events. This validates the assumption of independent, rare events in , where the number of mutations in a lineage approximates a Poisson distribution with parameter proportional to the number of divisions. Similarly, firing rates are often characterized as inhomogeneous Poisson processes, where the inter-spike intervals are exponentially distributed, and the count of spikes in a fixed time window follows a Poisson distribution with mean equal to the of the rate function, capturing the irregular yet rate-dependent spiking observed in cortical neurons. In physics, the Poisson distribution describes the statistics of counts, as first quantified by Rutherford and Geiger in their 1910 experiments on emissions from , where the number of scintillations in successive time intervals of 7.5 seconds closely matched the Poisson , confirming the randomness and independence of decay events. This model extends to arrivals in for coherent light sources, such as lasers, where the number in a over a time follows a Poisson distribution with equal to the average intensity, reflecting the classical wave-like nature of the field and equality of and variance in the count statistics. Astronomy employs the spatial to model galaxy counts in large survey volumes under the assumption of homogeneity, where the expected number of galaxies in a cubic region is proportional to its volume, and fluctuations around this mean follow a distribution, serving as a for testing clustering in cosmic . However, real galaxy distributions often exhibit relative to Poisson, indicating positive correlations. In , Poisson assumptions for species counts are frequently violated by due to clumped distributions, where individuals aggregate in patches influenced by heterogeneity or , leading to variance exceeding the mean and necessitating alternatives like the to account for the aggregation parameter. Such deviations highlight limitations of the model in patchy environments, as seen in or population surveys. Modern applications include single-cell RNA sequencing, where read counts per gene per cell are modeled as Poisson-distributed under the assumption of independent mRNA capture events, with the rate parameter reflecting true expression levels scaled by sequencing depth, though zero-inflation and overdispersion often require extensions for accurate inference. This framework enables quantification of transcriptional variability across heterogeneous cell populations.

Regression and Overdispersion Models

Poisson regression extends the Poisson distribution to model count data as a function of predictor variables. In this framework, each observation y_i is assumed to follow a Poisson distribution with mean \lambda_i, where the logarithm of the mean is modeled as a linear predictor: \log(\lambda_i) = \mathbf{x}_i^T \boldsymbol{\beta}. This formulation positions as a (GLM) with the canonical log link function, ensuring that predicted means are positive. The model parameters \boldsymbol{\beta} are estimated by maximizing the log-likelihood function, typically via iteratively reweighted least squares (IRLS) or Newton-Raphson algorithms, which iteratively update estimates until convergence. These methods leverage the exponential family structure of the Poisson distribution for efficient computation. A key assumption of the Poisson model is equidispersion, where the conditional variance equals the mean (\mathrm{Var}(y_i \mid \mathbf{x}_i) = \lambda_i); however, many real-world datasets exhibit overdispersion, with variance exceeding the mean. Overdispersion often arises from unobserved heterogeneity or clustering effects and can lead to underestimated standard errors if unaddressed. It can be detected through deviance-based goodness-of-fit tests, which compare observed and expected deviance under the null of no overdispersion, or via score tests that assess auxiliary regressions for evidence of extra variation. To handle , negative binomial regression serves as a robust alternative, parameterizing the Poisson mean \lambda_i similarly while introducing a dispersion parameter r > 0 to allow greater flexibility in the variance: \mathrm{Var}(y_i \mid \mathbf{x}_i) = \lambda_i + \frac{\lambda_i^2}{r}. As r \to \infty, the model reduces to the standard Poisson, making it a natural extension for overdispersed counts. The parameters are estimated using similar maximum likelihood techniques adapted for the negative binomial likelihood. When datasets feature an excess of zeros beyond what a standard Poisson predicts—often due to structural factors like non-response or absence of risk—zero-inflated Poisson (ZIP) models address this by mixing a degenerate distribution at zero with a Poisson component. Specifically, the probability of a zero is \pi_i + (1 - \pi_i) e^{-\lambda_i}, where \pi_i captures the inflation probability, typically modeled via a link on covariates, and the non-zero part follows a truncated Poisson. Parameter estimation proceeds via expectation-maximization or direct maximization of the mixture likelihood. Poisson regression and its extensions find prominent applications in modeling disease incidence, where counts of cases in regions or time periods are regressed on demographic or environmental covariates to estimate relative risks. For instance, it has been used to analyze rates in relation to exposure levels. In web traffic modeling, these approaches predict page views or user sessions as functions of marketing variables or search trends, accommodating the count nature of digital interactions.

Computational Methods

Numerical Evaluation

The (PMF) of the Poisson distribution can be computed directly using the formula P(X = k; \lambda) = e^{-\lambda} \frac{\lambda^k}{k!} for small values of k and \lambda, where direct evaluation avoids overflow issues. For larger k, is achieved by computing the log-PMF as \log P(X = k; \lambda) = -\lambda + k \log \lambda - \log \Gamma(k+1), where \log \Gamma(k+1) approximates \log(k!) using the log-gamma function or Stirling's series to handle large factorials without overflow. An efficient recursive algorithm starts with P(X = 0; \lambda) = e^{-\lambda} and iterates forward using the relation P(X = k+1; \lambda) = P(X = k; \lambda) \cdot \frac{\lambda}{k+1} for k \geq 0, which avoids repeated computations and is suitable for evaluating a sequence of probabilities around the . The (CDF) is given by F(k; \lambda) = P(X \leq k) = \sum_{i=0}^{k} e^{-\lambda} \frac{\lambda^i}{i!}, which can be expressed in terms of the upper as F(k; \lambda) = 1 - \frac{\Gamma(k+1, \lambda)}{k!}, where \Gamma(s, x) = \int_x^\infty t^{s-1} e^{-t} \, dt; this relation leverages numerical libraries for incomplete gamma evaluation to compute tail probabilities accurately. In software implementations, the SciPy library provides the scipy.stats.poisson class, which includes methods like pmf for the PMF and cdf for the CDF, utilizing optimized recursive and gamma-based computations internally. Similarly, R's base statistics package offers dpois for the PMF and ppois for the CDF of the Poisson distribution, employing recursive algorithms for efficiency in vectorized operations. For large \lambda (typically \lambda > 10), the Poisson distribution is approximated by a with \lambda and variance \lambda, so P(a \leq X \leq b) \approx \Phi\left( \frac{b + 0.5 - \lambda}{\sqrt{\lambda}} \right) - \Phi\left( \frac{a - 0.5 - \lambda}{\sqrt{\lambda}} \right), where \Phi is the standard normal CDF and the of \pm 0.5 improves accuracy for discrete-to-continuous approximation.

Random Variate Generation

Generating pseudo-random variates from the Poisson distribution with parameter \lambda > 0 is essential for simulations, statistical modeling, and . Several algorithms have been developed to produce these integer-valued samples efficiently, balancing computational speed and accuracy across different ranges of \lambda. These methods typically rely on uniform random number generators and exploit properties of the Poisson distribution, such as its relation to the or approximations to continuous distributions for large \lambda. The method is a foundational technique applicable to the Poisson distribution. It involves generating a uniform random variate U \sim \text{Uniform}(0,1) and finding the smallest non-negative k such that the (CDF) satisfies F(k; \lambda) \geq U, where F(k; \lambda) = \sum_{i=0}^{k} e^{-\lambda} \lambda^i / i!. This requires iterative computation of the partial sums of the until the threshold is met, making it straightforward to implement but potentially inefficient for large \lambda due to the need for many terms in the sum. For small to moderate \lambda (typically \lambda < 10), Knuth's algorithm provides an efficient alternative based on sequential products of uniform variates. The procedure initializes k = 0 and S = 1, then repeatedly generates U \sim \text{Uniform}(0,1) and updates S \leftarrow S \cdot U and k \leftarrow k + 1 until S < e^{-\lambda}, at which point it returns k - 1. This method leverages the equivalence between the Poisson process and exponential interarrival times, ensuring exact sampling with an expected number of uniform draws proportional to \lambda + 1. It was introduced in the context of seminumerical algorithms and remains a benchmark for low-\lambda generation due to its simplicity and bounded runtime. When \lambda is large (say, \lambda > 100), direct methods become impractical, and approximation-based techniques are preferred. One widely used approach combines the X \approx \mathcal{N}(\lambda, \lambda) with to ensure exact Poisson variates. A candidate Y \sim \mathcal{N}(\lambda, \lambda) is generated, and it is accepted as the output if it falls within the (non-negative integers) and a U \sim \text{Uniform}(0,1) satisfies U \leq p(Y; \lambda) / g(Y), where p(\cdot; \lambda) is the Poisson probability mass at the rounded Y and g(\cdot) is a bounding function (often derived from the normal density). If rejected, the process repeats. This method achieves high rates for large \lambda because the Poisson distribution converges to the normal, with the rejection step correcting for discreteness and tail discrepancies; its improves as \lambda grows, with expected acceptances near 1. Another exact method employs acceptance-rejection with a proposal from a translated exponential distribution, particularly effective for moderate \lambda. The proposal distribution is a shifted exponential, such as q(x) = \lambda e^{-\lambda (x - \lfloor \lambda \rfloor)} for x \geq \lfloor \lambda \rfloor, truncated and adjusted to majorize the Poisson tail, with acceptance probability proportional to the ratio of the Poisson mass to the proposal density at the candidate integer. This exploits the Poisson's relation to the gamma distribution (as the waiting time for the \lambda + 1-th event in a Poisson process follows an exponential), allowing efficient sampling in the modal region and beyond. Seminal implementations, such as the PTPE algorithm, use piecewise exponential envelopes for even better performance across a range of \lambda. In modern computational libraries, such as in , the numpy.random.poisson function implements a approach for vectorized generation of multiple variates. For small \lambda, it employs variants of Knuth's method; for larger values, it switches to with , ensuring for high-dimensional simulations while maintaining exactness. This implementation is optimized for speed, supporting array outputs and handling edge cases like \lambda = 0 (always returning 0).

References

  1. [1]
    1.3.6.6.19. Poisson Distribution - Information Technology Laboratory
    Probability Mass Function, The Poisson distribution is used to model the number of events occurring within a given time interval. The formula for the ...
  2. [2]
    Poisson Distribution - an overview | ScienceDirect Topics
    Poisson distribution is defined as a discrete distribution that measures the probability of a given number of events occurring in a specified time frame.<|control11|><|separator|>
  3. [3]
    Poisson Distribution: Definition & Uses - Statistics By Jim
    The Poisson distribution is a discrete probability distribution that describes probabilities for counts of events that occur in a specified observation space.Missing: mathematics | Show results with:mathematics
  4. [4]
    De Moivre's Poisson Approximation to the Binomial - Bellhouse - 2009
    Oct 29, 2009 · In his first work on probability, written in 1711, Abraham De Moivre looked at the problem of finding the number of trials required in a ...
  5. [5]
    Siméon-Denis Poisson (1781 - 1840) - Biography - MacTutor
    , an important work on probability published in 1837, the Poisson distribution first appears. The Poisson distribution describes the probability that a ...
  6. [6]
    Bortkewitsch's Horse-Kicks and the Generalised Linear Model - jstor
    The 1898 monograph gives the famous horse-kick data (Bortkewitsch, 1898, pp. 23-25) whose analysis has featured in so many text-book accounts of the Poisson.
  7. [7]
    [PDF] A note on the history of the Poisson process. 1 Introduction
    Sep 23, 2019 · This note reviews how the contemporary concept of the Poisson process. (sometimes called the Poisson random measure) evolved through out the.
  8. [8]
    [PDF] Poisson distribution
    The Poisson distribution is a limiting case of the binomial distribution which arises when the number of trials n increases indefinitely whilst the product μ = ...
  9. [9]
    [PDF] Theorem The Poisson(µ) distribution is the limit of the binomial(n, p ...
    Theorem The Poisson(µ) distribution is the limit of the binomial(n, p) distribution with. µ = np as n → с. Proof Let the random variable X have the ...
  10. [10]
    Lesson 12: The Poisson Distribution - STAT ONLINE
    It is important to keep in mind that the Poisson approximation to the binomial distribution works well only when n is large and p is small. In general, the ...
  11. [11]
    [PDF] Poisson Distribution, Continuous Random Variables - Washington
    Poisson: X ∼ Poi(λ) iff X has the following probability mass function: pX (k) = e−λ λk k! , k = 0, 1,... E [X] = λ and Var (X) = λ. An example of a ...<|control11|><|separator|>
  12. [12]
    [PDF] Milestones of Probability Theory 1 Introduction 2 The Poisson ...
    2.2 Properties of the Poisson Distribution. As a sanity check on our distribution, the probability values (2) had better sum to 1. Using the. Taylor expansion ...
  13. [13]
    [PDF] Poisson Variables
    Next, a recursion relation is derived for the probability, denoted P(n + 1;t), for there to be n + 1 events in a time t, which will be based on the probability ...
  14. [14]
    [PDF] Chapter 4 The Poisson Distribution
    The Poisson distribution is named after Simeon-Denis Poisson (1781–1840). In addition, poisson is French for fish. In this chapter we will study a family of ...
  15. [15]
    Poisson distribution | Properties, proofs, exercises - StatLect
    X has a Poisson distribution with parameter $lambda $ if its probability mass function is [eq3] where $x!$ is the factorial of x . Relation to the ...
  16. [16]
    [PDF] Poisson Distribution Probability generating functions
    Poisson Distribution. Definition 1. A random variable X has Poisson distribution with mean λ if. P(X = k) = λke−λ k! , for k = 0,1,2,... The next exercise ...
  17. [17]
    [PDF] Moment generating functions - UConn Undergraduate Probability OER
    mX+Y (t) = mX(t)mY (t) = ea(et−1)eb(et−1) = e(a+b)(et−1), which is the moment generating function of a Poisson with parameter a+b, therefore X +Y is a Poisson ...<|separator|>
  18. [18]
    [PDF] 1 Cumulants
    The Poisson distribution with mean µ has moment generating function exp(µ(eξ − 1)) and cumulant generating function µ(eξ − 1). Con- sequently all the cumulants ...
  19. [19]
    5. Thinning and Superposition - Random Services
    Complementary to splitting or thinning a Poisson process is superposition : if we combine the random points in time from independent Poisson processes, then we ...
  20. [20]
    [PDF] Introduction to Probability Models: Eighth Edition by Sheldon M. Ross.
    Oct 26, 2008 · Introduction to Probability Models: Eighth Edition by Sheldon M ... We will first prove that the sum of two Poisson random variables is a Poisson ...
  21. [21]
    [PDF] Tail Bounds and Applications - CMU School of Computer Science
    We start by illustrating how the Chernoff bound can be used to bound the tail of. 𝑋 where 𝑋 ∼ Poisson(𝜆), as in (18.2). Let 𝑋 ∼ Poisson(𝜆). For 𝑡 > 0,. E 𝑒𝑡𝑋 =.
  22. [22]
    Improved Inequalities for the Poisson and Binomial Distribution and ...
    Dec 23, 2013 · Exponential bounds—which improve upon the Chernoff-Hoeffding inequalities by a factor of at least two—are also obtained for both distributions.
  23. [23]
    [PDF] '<-I)=I o - Stanford University
    Upper bounds on the left and right tails of the Poisson distribution are given. These bounds can be easily computed in a numerically stable way, even when ...
  24. [24]
    On Large Deviations in the Poisson Approximation - SIAM.org
    This paper proves a general lemma comparing the behavior of probabilities of large deviations P ( X ≧ x ) of a random variable X against the Poisson ...
  25. [25]
    [PDF] Total variation distance between measures
    Feb 15, 2005 · I adapted the results on the total variation and relative entropy distances between Binomial and Poisson distributions from Reiss (1993, p 25).<|control11|><|separator|>
  26. [26]
    [PDF] STAT 511 - Lecture 6: The Binomial, Hypergeometric, Negative ...
    Feb 5, 2019 · ▷ An acceptable rule of thumb is to use this approximation when n ≥ 50 and np < 5. Levine. STAT 511. Page 26. Example. ▷ In a certain ...
  27. [27]
    12.4 - Approximating the Binomial Distribution | STAT 414
    Five percent (5%) of Christmas tree light bulbs manufactured by a company are defective. ... That is, if there is a 5% defective rate, then there is a 26.5% ...
  28. [28]
    An Introduction to the Theory of Point Processes - SpringerLink
    Point processes and random measures find wide applicability in telecommunications, earthquakes, image analysis, spatial point patterns, and stereology.
  29. [29]
    Poisson Processes - J. F. C. Kingman - Oxford University Press
    This book attempts to redress the balance. It records the author's fascination with the beauty and wide applicability of Poisson processes in one or more ...
  30. [30]
    [PDF] Chapter 2 - POISSON PROCESSES - MIT OpenCourseWare
    A Poisson process is a simple and widely used stochastic process for modeling the times at which arrivals enter a system.
  31. [31]
    None
    ### Key Properties of the Homogeneous Poisson Point Process
  32. [32]
    Estimation for the bivariate Poisson distribution - Oxford Academic
    This paper is concerned with the estimation of the covariance parameter of the bivariate Poisson distribution. It is shown that the method of moments has low ...
  33. [33]
    On bivariate Poisson regression models - ScienceDirect.com
    The bivariate Poisson is the most widely used model for bivariate counts. It was proposed by Holgate (1964) and presented by Johnson and Kotz (1969). The ...<|control11|><|separator|>
  34. [34]
    [math/0501274] Free extreme values - arXiv
    Jan 18, 2005 · Abstract: Free probability analogs of the basics of extreme-value theory are obtained, based on Ando's spectral order.
  35. [35]
    [PDF] Free Probability Theory
    Free probability theory was created by Dan Voiculescu around 1985, motivated by his efforts to understand special classes of von Neumann algebras. His dis-.
  36. [36]
    A new compounding life distribution: the Weibull–Poisson distribution
    In this paper, a new compounding distribution, named the Weibull–Poisson distribution is introduced. The shape of failure rate function of the new compounding ...
  37. [37]
    1.5 - Maximum Likelihood Estimation | STAT 504
    Setting the first derivative equal to 0 gives the solution: λ ^ = ∑ i = 1 n x i n . Thus, for a Poisson sample, the MLE for is just the sample mean.
  38. [38]
    Statistical Inference for the Poisson Distribution
    Sep 6, 2025 · 5 Maximum Likelihood Estimator. The usual estimator of the parameter μ is μ ^ = x . This turns out to also be the maximum likelihood ...
  39. [39]
    [PDF] COMPARISON OF CONFIDENCE INTERVALS FOR THE POISSON ...
    We attempt to perform an exhaustive review of the existing methods for obtaining confidence intervals for the Poisson parameter and present an extensive.
  40. [40]
    [PDF] Lecture 21 — Prior distributions 21.1 Conjugate priors and improper ...
    ∼ Poisson(λ), then a conjugate prior for λ is Gamma(α, β), and the corresponding posterior given X1 = x1,...,Xn = xn is Gamma(s + α, n + β) where s = x1 + ...
  41. [41]
    [PDF] Chapter 9 The exponential family: Conjugate priors - People @EECS
    9.0.3 Poisson distribution and gamma priors. As another example of a discrete exponential family distribution let us consider the Poisson distribution: p(x|θ) ...
  42. [42]
    [PDF] The Gamma/Poisson Bayesian Model
    The Gamma/Poisson Bayesian Model. ▻ If our data X1,...,Xn are iid Poisson(λ), then a gamma(α, β) prior on λ is a conjugate prior. Likelihood: L(λ|x) = n. Y i ...
  43. [43]
    [PDF] One-parameter models - MyWeb
    Note that the exponential distribution also has the gamma as a conjugate prior, leading to the posterior distribution θ|y ∼ Gamma α + n, β +. X i yi.
  44. [44]
    [PDF] Jeffrey's prior derivations
    The Jeffreys' prior is thus π(θ) ∝pI(θ) ∝ θ-1/2. (13). This is an improper prior. It can be seen as the limiting distribution of the prior θ ∼ Gamma(1/2,b).
  45. [45]
    [PDF] Bayesian Data Analysis Third edition (with errors fixed as of 20 ...
    ... Bayesian Data Analysis. Third edition. (with errors fixed as of 20 February 2025). Andrew Gelman. Columbia University. John B. Carlin. University of Melbourne.
  46. [46]
    [PDF] BAYESIAN AND NON-BAYESIAN INTERVAL ESTIMATORS FOR ...
    We consider Bayesian credible intervals for the Poisson distribution under different priors. 3.1. Posterior distributions under different priors. The efficiency ...
  47. [47]
    [PDF] Lecture 20 — Bayesian analysis 20.1 Prior and posterior distributions
    The interval from the 0.05 to the 0.95 quantile of the Gamma(s + α, n + β) distribution forms a 90% Bayesian credible interval for λ.
  48. [48]
    [PDF] MAS3301 Bayesian Statistics Problems 5 and Solutions - Ncl
    (c) Using a normal approximation to the posterior distribution, based on the posterior mean and variance, find a 95% symmetric posterior credible interval for λ ...
  49. [49]
    [PDF] Coverage Probabilities of Poisson Interval Estimates
    In Figure 1, the proportion of simulated 95% confidence intervals that contain true value of λ happens to be exactly 95/100. Figure 1. One hundred 95% ...
  50. [50]
    [PDF] Statistics for the Luria-Delbrück distribution - arXiv
    Since the number of divisions is large and the probability of mutation is small, the total number of mutations approximately follows the Poisson distribution.<|separator|>
  51. [51]
    Differences in Spike Train Statistics between Parietal Cortical Areas
    May 14, 2009 · The variability of neuronal responses is proportional to the mean in many brain areas, which suggests that neural responses might follow a Poisson distribution.
  52. [52]
    [PDF] Poissonian Distributions in Physics: Counting Electrons and Photons
    Jan 30, 2021 · It is a discrete distribution that gives the probability of a number of events. Poisson distribution is used as a model to have the number of ...
  53. [53]
    The Distribution of the Galaxies - W.C. Saslaw
    Thus by the late 1950s there was a consensus that galaxies did not have a Poisson distribution, even though the description of their clustering was unclear.
  54. [54]
    An Overview of Modern Applications of Negative Binomial Modelling ...
    Many approaches have been developed for analysing overdispersed counts, including quasi-Poisson and Poisson log-normal models [13], generalised Poisson models ...
  55. [55]
    The Poisson distribution model fits UMI-based single-cell RNA ...
    Jun 17, 2023 · A major challenge is that scRNA-seq data typically contain a large number of zero counts for gene/cell combinations (often exceeding 90%) [12].
  56. [56]
    Regression-based tests for overdispersion in the Poisson model
    Regression-based tests for overdispersion in the Poisson model☆. Author ... Cameron and Trivedi, 1985. A.C. Cameron, P.K. Trivedi. Regression based tests ...
  57. [57]
    USE OF POISSON REGRESSION MODELS IN ESTIMATING ...
    A more general application of Poisson regression is its utility in modeling disease rates according to postulated etlologic mechanisms of exposures or according ...
  58. [58]
    A Poisson Regression Examination of the Relationship between ...
    Aug 6, 2025 · In this paper, Poisson regressions are used to explore the relationship between the online traffic to a specific website and the search volumes ...
  59. [59]
    [PDF] On the Computation of Poisson Probabilities - UPCommons
    In (Kemp and Kemp, 1991, sect. 3), the starting probability is Pbλ+0.5c(λ), which is approximated using formulas developed in (Kemp, 1988b).
  60. [60]
    [PDF] stirling's formula - keith conrad
    Stirling's formula first arose from correspondence between Stirling and DeMoivre in the. 1720s about DeMoivre's work on approximating a binomial distribution by ...
  61. [61]
    [PDF] approximation of the inverse Poisson cumulative distribution function
    In this paper we have derived a number of approximations of the inverse incomplete gamma function which can be used to implement the inverse Poisson cumulative ...Missing: relation | Show results with:relation
  62. [62]
    scipy.stats.poisson — SciPy v1.16.2 Manual
    scipy.stats.poisson# ; logcdf(k, mu, loc=0). Log of the cumulative distribution function. ; sf(k, mu, loc=0). Survival function (also defined as 1 - cdf , but sf ...1.7.01.13.1
  63. [63]
    The Poisson Distribution - R
    dpois gives the density, ppois is the cumulative distribution function, and qpois is the quantile function of the Poisson distribution. rpois generates random ...
  64. [64]
    28.2 - Normal Approximation to Poisson | STAT 414
    Now, let's use the normal approximation to the Poisson to calculate an approximate probability. First, we have to make a continuity correction. Doing so, we ...Missing: lambda | Show results with:lambda
  65. [65]
    [PDF] Poisson-Minibatching for Gibbs Sampling with Convergence Rate ...
    Nov 21, 2019 · Inverse transform sampling is a clas- sical method that generates samples from a uniform distribution and then transforms them by the inverse of ...
  66. [66]
    The art of computer programming, volume 2 (3rd ed.)
    The art of computer programming, volume 2 (3rd ed.): seminumerical algorithmsNovember 1997. Author: Author Picture Donald E. Knuth. Stanford Univ., Stanford ...
  67. [67]
    Computer Generation of Poisson Deviates from Modified Normal ...
    AHRENS, J.H., AND DIETER, U. Computer methods for samphng from gamma, beta, Poisson and binomial distributions Computing 12 (1974), 223-246. Google Scholar.
  68. [68]
    Poisson Random Variate Generation - jstor
    Knuth, D. E. (1969) The Art of Computer Programming, vol. 2, Seminumerical Algorithms, p. 121. Reading: Addison-Wesley. Kronmal, R. A. and Peterson, A. V. (1979) ...