Monotone likelihood ratio
In statistics, the monotone likelihood ratio (MLR) is a property exhibited by certain parametric families of probability distributions, characterized by the likelihood ratio f_{\theta_2}(x) / f_{\theta_1}(x) being non-decreasing in a real-valued statistic Y(x) whenever \theta_1 < \theta_2 and the densities are positive.[1] This condition ensures that the family admits uniformly most powerful (UMP) tests for one-sided composite hypotheses, such as H_0: \theta \leq \theta_0 versus H_1: \theta > \theta_0, where the optimal test rejects the null based on exceeding a threshold of the statistic Y(x).[2] The MLR property facilitates powerful inference by aligning the ordering of observations with the parameter space, minimizing type I and type II errors in a controlled manner.[1] Prominent examples of distributions possessing the MLR include the exponential family (such as the exponential and gamma distributions), the uniform distribution on (0, \theta) with respect to the maximum order statistic, the Poisson distribution with respect to the sum of observations, and the normal distribution for testing the mean with known variance.[1][3][2] Additionally, noncentral versions of the t, chi-squared, and F distributions exhibit MLR in their noncentrality parameters, extending the property's utility in more complex testing scenarios.[4] The concept, formalized in foundational works on hypothesis testing such as Karlin and Rubin (1950), underpins much of modern statistical decision theory by providing a criterion for the existence of optimal tests without requiring full specification of alternative parameters.[1][5]Definition and Intuition
Formal Definition
A parametric family of probability distributions \{P_\theta \mid \theta \in \Theta\} is said to have the monotone likelihood ratio (MLR) property in a statistic T if, for all \theta_1 < \theta_2 in \Theta, the likelihood ratio \frac{L(\theta_2; x)}{L(\theta_1; x)} is a non-decreasing function of T(x) for every x in the support of the distribution. Here, L(\theta; x) denotes the likelihood function, which corresponds to the density f(x \mid \theta) with respect to a dominating measure \mu (such as Lebesgue measure for continuous distributions or counting measure for discrete distributions) under P_\theta. This formulation assumes that \Theta is an open interval of the real line, T is a sufficient statistic for \theta, and the family \{P_\theta\} is dominated by \mu, ensuring the existence of densities f(x \mid \theta). For the discrete case, where the distributions have probability mass functions p(x \mid \theta) with respect to the counting measure, the ratio \frac{p(x \mid \theta_2)}{p(x \mid \theta_1)} is non-decreasing in T(x), allowing for plateaus where the ratio remains constant (i.e., equality holds over intervals of T(x)). Similarly, in the continuous case with densities f(x \mid \theta) with respect to Lebesgue measure, \frac{f(x \mid \theta_2)}{f(x \mid \theta_1)} is non-decreasing in T(x), again permitting equality in subregions. These cases unify under the dominated family framework, where the indicator for equality ensures the property holds weakly (non-strictly) to accommodate non-strict monotonicity. Equivalently, the MLR property can be expressed in terms of the conditional expectation of the likelihood ratio given T = t: \Lambda(\theta_1, \theta_2; t) = \mathbb{E}\left[ \frac{L(\theta_2; X)}{L(\theta_1; X)} \,\Big|\, T = t \right], which is non-decreasing in t for \theta_1 < \theta_2. By the sufficiency of T, this conditional expectation equals the ratio of the induced densities of T under \theta_2 and \theta_1, providing a precise condition on the marginal distribution of the sufficient statistic.Intuitive Explanation
The monotone likelihood ratio (MLR) property captures the idea that, within a parameterized family of probability distributions, the evidence supporting a larger parameter value θ₂ over a smaller one θ₁ grows steadily stronger as a relevant summary statistic T (derived from the data) increases. Intuitively, this monotonicity reflects an ordered structure in the data: higher observed values of T tilt the balance more decisively toward θ₂, making the comparison between parameter values predictable and consistent across the range of possible observations. This property simplifies inference by ensuring that the likelihood ratio Λ(θ₁, θ₂; x) = f(x | θ₂) / f(x | θ₁)—where f denotes the density or mass function—behaves in a non-decreasing manner with respect to T(x), as formalized in the preceding definition.[2] This ordering implies that the family of distributions is well-behaved for one-sided comparisons, where larger data realizations align intuitively with higher parameter values, enhancing the reliability of decisions like rejecting a null hypothesis in favor of an alternative. For instance, in settings where θ might represent a mean or rate parameter, the MLR ensures that accumulating evidence from the data reinforces the case for larger θ without reversals or ambiguities as T grows. Such intuitive alignment is crucial for practical statistical procedures, as it allows tests to leverage this monotonicity for maximal efficiency.[2] The MLR property builds on the Neyman-Pearson lemma from the 1930s and was formalized by Samuel Karlin and Herman Rubin in 1956 to handle optimal tests for composite hypotheses.[6] In families lacking MLR, however, the likelihood ratio may fluctuate non-monotonically with T, resulting in tests whose rejection regions cannot be simply threshold-based and often lack uniform optimality, thereby complicating inference and reducing power in one-sided scenarios.[2]Basic Example
A simple and illustrative example of the monotone likelihood ratio (MLR) property arises in the context of independent Bernoulli trials, each with success probability \theta, where the sufficient statistic T is the total number of successes k in n trials.[7] The likelihood function for observing k successes is L(\theta; k) = \binom{n}{k} \theta^k (1 - \theta)^{n - k}. For two values \theta_1 < \theta_2, the likelihood ratio is given by \frac{L(\theta_2; k)}{L(\theta_1; k)} = \left( \frac{\theta_2}{\theta_1} \right)^k \left( \frac{1 - \theta_2}{1 - \theta_1} \right)^{n - k}. To verify monotonicity, consider the ratio of consecutive values: \frac{\frac{L(\theta_2; k+1)}{L(\theta_1; k+1)}}{\frac{L(\theta_2; k)}{L(\theta_1; k)}} = \frac{\theta_2 (1 - \theta_1)}{\theta_1 (1 - \theta_2)} > 1, since \theta_2 > \theta_1 implies \frac{\theta_2}{\theta_1} > 1 and \frac{1 - \theta_1}{1 - \theta_2} > 1. Thus, the likelihood ratio is strictly increasing in k.[7] This demonstrates the MLR property because a larger value of k (more successes) provides progressively stronger evidence in favor of the higher \theta_2 over \theta_1, as the ratio grows with k.[7] The Bernoulli example is particularly useful for highlighting MLR in binary outcome settings, which are foundational in introductory statistical inference.[7]Distributions Exhibiting MLR
Common Parametric Families
Several well-known parametric families of distributions possess the monotone likelihood ratio (MLR) property, which facilitates the construction of uniformly most powerful tests for one-sided hypotheses. These families are often one-parameter exponential families or closely related, where the likelihood ratio is monotone in a sufficient statistic T. Examples include the binomial, Poisson, normal (with known variance), exponential, gamma (with fixed shape), Weibull (with fixed shape parameter), and uniform distributions on [0, \theta].[8] For the binomial distribution \text{Bin}(n, p) with fixed n and parameter p, the likelihood ratio for p_2 > p_1 is increasing in the number of successes k, as it takes the form \left(\frac{p_2}{p_1}\right)^k \left(\frac{1-p_1}{1-p_2}\right)^{n-k}, where the first factor increases with k while the second decreases but at a slower rate.[8] For the Poisson distribution \text{Po}(\lambda), the ratio for \lambda_2 > \lambda_1 is \left(\frac{\lambda_2}{\lambda_1}\right)^k e^{(\lambda_1 - \lambda_2)}, which increases in the observation k since the exponential term is constant and less than 1, but outweighed by the increasing power term.[8] In the normal distribution \mathcal{N}(\mu, \sigma^2) with known \sigma^2 and parameter \mu, the ratio for \mu_2 > \mu_1 is monotone increasing in the sample mean \bar{x}, reflecting the location family's shift invariance.[8] The exponential distribution with rate \lambda (or scale $1/\lambda) has an MLR in the sum of observations \sum x_i, where for \lambda_2 > \lambda_1, the ratio \left(\frac{\lambda_2}{\lambda_1}\right)^n e^{(\lambda_1 - \lambda_2) \sum x_i} decreases in the sum, but equivalently increases when reparameterized by scale.[8] For the gamma distribution \text{Gamma}(\alpha, \beta) with fixed shape \alpha and scale parameter \beta, the likelihood ratio for \beta_2 > \beta_1 is increasing in \sum x_i, as the form involves \left(\frac{\beta_1}{\beta_2}\right)^{n\alpha} e^{(\beta_2 - \beta_1) \sum x_i / \beta_1 \beta_2} adjusted for the scale, but monotone due to the exponential term dominating.[8] Similarly, the Weibull distribution with fixed shape k and scale parameter \lambda exhibits MLR in \sum \log x_i, where the ratio for \lambda_2 > \lambda_1 increases in this statistic, akin to the exponential case since Weibull is a transformation of exponential.[8] Finally, the uniform distribution on [0, \theta] has MLR in the maximum order statistic X_{(n)}, with the ratio for \theta_2 > \theta_1 being (\theta_1 / \theta_2)^n if X_{(n)} \leq \theta_1 and infinite otherwise, which is nondecreasing in X_{(n)}.[8] A key reason many of these families satisfy MLR is that they belong to the class of one-parameter exponential families where the natural parameter is monotone and the log-partition function is convex, ensuring the likelihood ratio is increasing in the sufficient statistic (further details in the section on exponential families).[8]| Family | Parameter(s) | Sufficient Statistic T |
|---|---|---|
| Binomial(n, p) | p (n fixed) | Number of successes \sum k_i |
| Poisson(λ) | λ | Sum \sum x_i |
| Normal(μ, σ²) | μ (σ² known) | Sample mean \bar{x} |
| Exponential(λ) | λ (rate) | Sum \sum x_i |
| Gamma(α, β) | β (α fixed) | Sum \sum x_i |
| Weibull(k, λ) | λ (k fixed) | Sum of logs \sum \log x_i |
| Uniform[0, θ] | θ | Maximum X_{(n)} |