Fact-checked by Grok 2 weeks ago

False discovery rate

The false discovery rate (FDR) is a statistical framework in multiple testing that controls the expected proportion of incorrectly rejected hypotheses among all rejected hypotheses, offering a balance between discovering true effects and limiting false positives. Introduced by Yoav Benjamini and Yosef Hochberg in 1995 as an alternative to controlling the (FWER)—which bounds the probability of any false rejection—FDR is less stringent and thus more powerful when many hypotheses are false, making it suitable for large-scale analyses. The seminal Benjamini-Hochberg (BH) procedure implements FDR control through a step-up : p-values from individual tests are sorted in ascending order, and the largest index k such that the k-th is at most (k/m)q (where m is the number of tests and q is the FDR level) determines the rejection , rejecting all hypotheses up to that point. This approach guarantees FDR control at level q under the assumption of independence, later extended to certain forms of positive dependence, with simulations demonstrating substantial power gains over FWER methods like the . FDR has become a in high-dimensional , particularly in and , where thousands to millions of tests are common, such as in via or genome-wide association studies (GWAS) to identify disease-related variants. Extensions like Storey's q-value method estimate the proportion of true nulls to adaptively adjust for , enhancing power, while modern adaptive procedures incorporate covariates (e.g., effect sizes or genomic features) to further refine control in complex scenarios. Challenges remain in dependent tests and sparse signals, but FDR's flexibility has driven its adoption across fields including , astronomy, and sciences, with methods continuing to evolve as of 2025.

Background and Definitions

Multiple Hypothesis Testing

Multiple hypothesis testing arises when researchers perform several statistical tests simultaneously, often denoted as m tests, each evaluating a H_{0i} for i = 1, \dots, m, with corresponding observed p-values p_i measuring the evidence against each null. This scenario is common in fields requiring high-throughput analysis, where the independence or dependence among tests can complicate interpretation. In multiple testing, error rates extend beyond single-hypothesis scenarios. Type I errors occur as false positives, quantified by V, the number of true null hypotheses incorrectly rejected; the proportion of such errors among the true nulls is V / m_0, where m_0 is the number of true nulls. Type II errors manifest as false negatives, representing missed detections of true alternatives, though these are less directly controlled in traditional frameworks. The cumulative impact of these errors escalates with m, as the probability of at least one Type I error increases substantially without adjustment. Traditional approaches prioritize controlling the (FWER), defined as the probability of making one or more false positives across the family of tests, \Pr(V \geq 1). A conservative method for strong FWER control is the , which adjusts the significance level to \alpha / m per test, ensuring the overall FWER does not exceed \alpha. However, this adjustment drastically reduces statistical power, particularly for large m, as it demands extremely low p-values for rejection and overlooks potential dependencies among tests. Consider an illustration from : suppose m = 1000 genes are tested for expression between two conditions, with each test at \alpha = 0.05. If all hypotheses are true, an uncorrected approach expects about 50 false positives purely by chance, severely inflating the error rate and complicating downstream biological validation. Such scenarios underscore the limitations of conservative controls like Bonferroni in high-dimensional settings. The need for less stringent error control becomes evident in large-scale applications like , where m can reach tens of thousands; here, methods like the false discovery rate offer a more powerful alternative to FWER while addressing the multiplicity issue.

False Discovery Rate

In the context of multiple hypothesis testing, the false discovery rate (FDR) provides a measure of error control that addresses the proportion of incorrect rejections among all rejections, rather than focusing solely on the occurrence or average number of errors. Formally, the FDR is defined as \text{FDR} = \mathbb{E}[Q], where Q = V/R if R > 0 and Q = 0 if R = 0. Here, V denotes the number of false positives (true null hypotheses incorrectly rejected), R = V + S is the total number of rejections, and S is the number of true positives (false null hypotheses correctly rejected). This definition is equivalent to \text{FDR} = \mathbb{E}[V/R \mid R > 0] \cdot P(R > 0), emphasizing the expected proportion of false discoveries conditional on at least one rejection, weighted by the probability of discoveries. A related variant is the positive false discovery rate (pFDR), defined as \text{pFDR} = \mathbb{E}[V/R \mid R > 0], which conditions explicitly on the event of at least one rejection and is particularly useful in scenarios where discoveries are anticipated. The FDR controls the expected proportion of false rejections among all rejections at a specified level q, meaning that procedures achieving FDR control ensure \text{FDR} \leq q. This interpretation allows researchers to bound the average fraction of erroneous discoveries while permitting some false positives to increase the detection of true effects. The FDR occupies an intermediate position between the per-comparison error rate (PCER), defined as \text{PCER} = \mathbb{E}[V]/m where m is the total number of tests, and the , which controls the probability of at least one false rejection via \text{FWER} = P(V \geq 1). Unlike the conservative FWER, which strictly limits any false positives at the expense of power, or the lenient PCER, which only averages errors across tests, the FDR strikes a balance by allowing a controlled proportion of false positives to enhance overall discovery power in large-scale testing. FDR control typically assumes independence among test statistics or positive regression dependence to ensure the bound holds; under independence, a conservative estimate satisfies \text{FDR} \leq q. These assumptions facilitate practical implementation while maintaining theoretical guarantees on error rates.

Historical Development

Technological Motivations

The advent of high-throughput technologies in the late and , particularly DNA microarrays in , generated vast datasets requiring the simultaneous testing of thousands of , far exceeding the scale of traditional statistical analyses where m >> 100. These platforms enabled researchers to measure levels across entire genomes in a single experiment, such as probing 10,000 or more genes for differential expression between conditions, but this scale amplified the challenges of multiple hypothesis testing. Similar demands arose in fields like , where functional MRI scans tested activation across numerous brain voxels, necessitating robust controls for multiplicity to avoid spurious findings. Traditional methods for controlling the (FWER), such as the , proved overly conservative in these high-dimensional settings, dividing the significance level α by the number of tests (e.g., α/m for m=), which often resulted in near-zero discoveries even when a small proportion of true effects existed. This stringency stifled biological insights, as studies typically expected only a small fraction of s—around 1-5%—to be differentially expressed, yet FWER adjustments failed to detect many genuine signals due to reduced statistical , particularly with limited sample sizes common in early experiments (e.g., n=5-10 per group). Conversely, applying no correction to p-values led to thousands of false positives; for instance, at α=0.05 across tests, hundreds of spurious gene hits could emerge by alone, overwhelming downstream validation and interpretation in 1990s analyses. The core motivation was thus to achieve a practical balance: maximizing power to uncover true signals in sparse high-dimensional data while controlling false positives at a tolerable proportional rate, rather than an absolute one, to support discovery-driven sciences like . This need spurred early explorations of adaptive approaches, including to estimate the proportion of true nulls, as uncorrected or overly strict procedures hindered progress in identifying biologically relevant patterns. These pressures culminated around , coinciding with the formal introduction of FDR control as a response to the burgeoning microarray era.

Key Contributions in Literature

Early precursors to the formalization of the false discovery rate (FDR) appeared in the , addressing the challenges of analyzing large sets of p-values in multiple testing. Schweder and Spjøtvoll (1982) proposed plotting ranked p-values against their expected under the to visually assess deviations, offering an informal graphical method to estimate the proportion of true hypotheses (π₀) by fitting a line to the upper tail of the plot. This approach highlighted the potential to distinguish true signals from noise without strict control of error rates. Building on this, Soric (1989) introduced the concept of the proportion of false discoveries among significant results, framing it as an E(V/R) where V is the number of false positives and R the total rejections, and suggested a method to bound this proportion for effect-size estimation in multiple tests. The seminal contribution came with Benjamini and Hochberg (1995), who formally defined the FDR as the expected proportion of false discoveries among all discoveries, E(V/R), and proposed a simple step-up procedure to control it at a specified level q under the assumption of independent test statistics. This work proved that their procedure controls the FDR while offering greater power than traditional methods, making it practical for high-dimensional data. As of 2025, the paper had garnered over 120,000 citations, establishing it as one of the most influential in statistics. Subsequent expansions addressed limitations in the original framework, particularly regarding test dependencies and estimation of true nulls. Benjamini and Yekutieli (2001) extended FDR control to general dependence structures by deriving a conservative adjustment based on the of inverse ranks, ensuring the procedure remains valid without assumptions, though at a potential cost to power. Storey (2002) introduced the positive FDR (pFDR) as a Bayesian-motivated variant and developed q-values as posterior probabilities of a being false, incorporating direct estimation of π₀ via a tuning parameter on the histogram to boost power in sparse signal settings. Later reviews synthesized these developments and underscored their evolution. Benjamini (2010) reflected on the FDR's origins, tracing its roots to graphical ideas like those of Schweder and Spjøtvoll, and highlighted how the Benjamini-Hochberg procedure's simplicity facilitated its rapid adoption, ranking it among the top-cited statistical papers due to its balance of theory and applicability. The FDR framework saw widespread adoption in by the early 2000s, where it became essential for analyzing data with thousands of tests, enabling reliable discovery while controlling false positives. Its influence extended to for imaging studies involving multiple brain regions and to for high-dimensional regression analyses, promoting rigorous error control across empirical sciences.

FDR Controlling Procedures

Benjamini–Hochberg Procedure

The Benjamini–Hochberg (BH) procedure is a seminal step-up method for controlling the false discovery rate (FDR) at a specified level q in multiple testing settings. Introduced as a practical to (FWER) controls like the , it rejects a controlled proportion of null hypotheses while maintaining higher power, particularly when many true s are present. The algorithm operates on m p-values p_1, \dots, p_m from m hypothesis tests. First, sort the p-values in non-decreasing order to obtain p_{(1)} \leq p_{(2)} \leq \dots \leq p_{(m)}. Then, identify the largest index k such that p_{(k)} \leq \frac{k}{m} q; if no such k exists, reject no hypotheses. Otherwise, reject the first k null hypotheses corresponding to p_{(1)}, \dots, p_{(k)}. This step-up approach starts from the smallest p-value and progressively increases the critical threshold, allowing more rejections as k grows. The procedure controls the FDR under the assumption that the p-values are or positively dependent on a of the alternatives (PRDS). was the original condition, ensuring FDR \leq q, while PRDS extends validity to certain dependent cases without adjusting the thresholds. It is conservative when few or no true alternatives exist, as the expected proportion of false discoveries V/R (where V is false rejections and R is total rejections) approaches zero. Under independence, the standard proof shows that conditional on rejecting k hypotheses, the expected number of false discoveries E[V \mid R = k] \leq (m_0 / m) q k, where m_0 is the number of true nulls. Thus, E[V / R \mid R > 0] \leq (m_0 / m) q \leq q, and \text{FDR} = E[V / R \mid R > 0] P(R > 0) \leq q. The step-up design enhances over step-down methods by rejecting more hypotheses when small p-values , though it remains less aggressive than adaptive procedures. For illustration, consider m = 5 tests with q = 0.05 and sorted p-values [0.005, 0.015, 0.025, 0.06, 0.1]. The thresholds are (1/5) \cdot 0.05 = 0.01, (2/5) \cdot 0.05 = 0.02, (3/5) \cdot 0.05 = 0.03, (4/5) \cdot 0.05 = 0.04, and (5/5) \cdot 0.05 = 0.05. Here, p_{(1)} = 0.005 \leq 0.01, p_{(2)} = 0.015 \leq 0.02, p_{(3)} = 0.025 \leq 0.03, but p_{(4)} = 0.06 > 0.04, so the largest k = 3. This rejects the hypotheses with p-values 0.005, 0.015, and 0.025, controlling at 5%.

Benjamini–Yekutieli Procedure

The extends the to control the under arbitrary dependence structures among test statistics, providing conservative FDR control where the may fail under negative or unknown dependencies. This adjustment ensures validity across all dependency scenarios, including those with negative associations that can inflate the beyond the nominal level in the original procedure. The algorithm mirrors the stepwise approach of the Benjamini–Hochberg procedure: sort the m p-values in ascending order as p_{(1)} \leq p_{(2)} \leq \cdots \leq p_{(m)}, then find the largest k such that p_{(k)} \leq \frac{k}{m} \cdot \frac{q}{c(m)}, where q is the target FDR level, and reject the k null hypotheses corresponding to the smallest p-values. Here, the adjustment factor is c(m) = \sum_{i=1}^m \frac{1}{i}, the m-th , which approximates \log m + \gamma with \gamma \approx 0.57721 the Euler-Mascheroni constant. The proof establishes FDR control at q by deriving a series of inequalities that bound the expected value of the false discovery proportion, incorporating c(m) to account for the worst-case dependence and ensuring the bound holds universally. While this guarantees robustness, the procedure is less powerful than the Benjamini–Hochberg method by the factor c(m) ≈ log m, as the stricter thresholds reduce the number of rejections, particularly in large-scale testing. For instance, with m = hypotheses, c(m) ≈ 7.48, which effectively scales down the target q by this amount and results in fewer discoveries compared to the Benjamini–Hochberg procedure under dependent conditions.

Storey Procedure

The Storey procedure introduces an adaptive approach to controlling the (FDR) by estimating the proportion of true hypotheses, denoted \pi_0, from the observed p-values, thereby improving compared to non-adaptive methods like the Benjamini–Hochberg . This estimation leverages the fact that under the , p-values are uniformly distributed on [0,1], while alternative p-values tend to be smaller; thus, p-values exceeding a tuning parameter \lambda (commonly \lambda = [0](/page/0).5) are assumed to primarily arise from true . The estimator is given by \hat{\pi}_0(\lambda) = \frac{\#\{p_i > \lambda\}}{m(1 - \lambda)}, where m is the total number of hypotheses and \#\{p_i > \lambda\} counts the p-values above \lambda. To account for variability, a bootstrap method can be employed to compute intervals for \hat{\pi}_0. The procedure controls the positive FDR (pFDR), defined as E[V/R \mid R > 0] where V is the number of false positives and R the number of rejections, providing a conservative bound such that the estimated pFDR is at least the true pFDR. Central to this is the q-value, an adjusted for each that represents the minimum pFDR at which it would be rejected. For ordered p-values p_{(1)} \leq \cdots \leq p_{(m)}, the q-value for the i-th smallest is q_{(i)} = \min_{j \geq i} \frac{\hat{\pi}_0 m p_{(j)}}{j}, and hypotheses are rejected if q_{(i)} \leq q for a target FDR level q. This formulation ensures monotonicity and interpretability, with the q-value indicating the expected proportion of false discoveries among rejections including that hypothesis. In practice, the algorithm proceeds as follows: first, compute the p-values and estimate \hat{\pi}_0 using the chosen \lambda; second, adjust the p-values to p'_i = \hat{\pi}_0 p_i; third, apply the Benjamini–Hochberg procedure to these adjusted p-values at level q, rejecting hypotheses where the adjusted ordered p-value p'_{(k)} \leq (k/m) q. This adjustment effectively scales the number of hypotheses to the estimated true nulls \hat{\pi}_0 m, leading to a looser threshold and more rejections when many nulls are present. The Benjamini–Hochberg procedure emerges as a special case when \hat{\pi}_0 = 1. The key advantages of the Storey procedure lie in its data-driven adaptivity, which boosts statistical particularly when \pi_0 is small (i.e., many true alternatives), without assuming or specific dependence structures beyond weak conditions. It provides a conservative guarantee that the pFDR does not exceed the target [q](/page/Q), and simulations demonstrate substantial power gains; for instance, with m = 1000, \pi_0 = 0.9, and target FDR [q](/page/Q) = 0.05, the effective number of hypotheses reduces to $0.9m, allowing approximately 10% more rejections than non-adaptive methods while maintaining control. This makes it especially suitable for high-dimensional settings like , where the proportion of true nulls is often high.

Recent and Advanced Procedures

Recent advancements in false discovery rate (FDR) control have extended classical procedures like the Benjamini-Hochberg method to address challenges in high-dimensional data, structured hypotheses, and domain-specific applications. These innovations prioritize exact control, adaptability to dependencies, and with modern computational frameworks, often building on foundational multiple testing techniques to enhance precision in fields such as and . One prominent development is the knockoff filter, introduced in 2015, which provides exact FDR control for variable selection in high-dimensional settings without assuming knowledge of the noise level or signal strength. The method constructs "knockoff" copies of the original features that mimic their joint distribution, enabling a swap-based statistic to distinguish true signals from noise while controlling the FDR at a user-specified level, such as 0.05, even under arbitrary correlations. This approach has been particularly influential in and , where it outperforms traditional thresholding by reducing false positives in sparse signal environments. Subsequent extensions, from 2015 onward, have integrated knockoff filters with pipelines for reproducible ; for instance, DeepPINK combines knockoffs with deep neural networks to identify predictive variables while maintaining FDR guarantees, demonstrating improved stability over lasso-based methods in simulated high-dimensional tasks. Similarly, DeepLINK applies knockoffs within inference for , achieving FDR control below the target (e.g., 0.1) across diverse datasets like eQTL mapping. In 2018, the structured FDR (sFDR) procedure was proposed to provide smoother error control by replacing the linear denominator R in the traditional FDR definition \mathbb{E}[V/R] with a non-decreasing s(R), such as s(R) = R^\gamma where $0 < \gamma < 1. This modification balances the conservatism of family-wise error rate (FWER) control with the power of standard FDR, yielding a more flexible criterion that adapts to the number of discoveries; for example, when \gamma = 1, it recovers the original FDR, while smaller \gamma values penalize large R less severely, improving performance in neuroimaging applications with spatially structured tests. Simulations in the original work show sFDR maintains control at the nominal level (e.g., 5%) across varying signal strengths, outperforming Benjamini-Hochberg in power for moderate discovery sets. A 2021 analysis highlighted issues with volcano plots, a common visualization tool in genomics that thresholds on both effect size (e.g., |t|-statistics) and FDR-adjusted p-values, leading to inflated FDR due to unaccounted selection bias. In RNA-seq simulations, this double-thresholding procedure resulted in actual FDR exceeding the target by up to 2-3 times (e.g., observed 0.15 vs. nominal 0.05), as the symmetric |t|-filtering ignores directional hypotheses and induces dependence among selected tests. The study recommends using one-sided p-values or effect-size-adjusted thresholds to mitigate this, with empirical evidence from GTEx data showing reduced false positives when applying one-sided tests prior to FDR correction. Extensions to group-structured hypotheses emerged in 2024, updating for pre-existing groups such as gene pathways in genomics, via hierarchical FDR control that applies separate thresholds within groups and across the hierarchy. This structured approach, detailed in a framework for multivariate testing, ensures FDR control at the group level (e.g., <0.05) while leveraging inter-group dependencies, as validated in simulations with correlated gene modules where it recovered 20-30% more true pathways than flat procedures. In drug response studies, hierarchical FDR properly bounds false discoveries across gene-environment interactions, maintaining validity in high-dimensional settings like . In record linkage tasks, a 2025 method introduced novel false discovery proportion (FDP) estimation for overlapping datasets, using empirical Bayes modeling to synthesize auxiliary records from marginal distributions and bound false matches in privacy-preserving scenarios. By estimating the proportion of true nulls via synthetic linkages, this approach controls FDP below 0.1 in simulations of administrative data merges, addressing challenges like partial overlaps where traditional FDR fails due to unobservable truths; it outperforms heuristic matching by reducing false links by 15-25% in real-world census applications.

Theoretical Properties

Adaptivity and Scalability

The false discovery rate (FDR) controlling procedures exhibit notable adaptivity by adjusting to the underlying data structure, particularly through estimation of the proportion of true null hypotheses, denoted as \pi_0. The and procedures are non-adaptive, implicitly assuming \pi_0 = 1, which can lead to conservative control when few true nulls are present (i.e., a low \pi_0). In contrast, adaptive procedures, such as the , estimate \pi_0 < 1 using the distribution of p-values, effectively increasing the sample size for power calculations and enhancing detection of true signals without violating FDR control. This data-driven adjustment improves power in scenarios where signals are sparse (high \pi_0), by estimating \pi_0 and adjusting the threshold to be less conservative when signals are present. FDR procedures also demonstrate strong scalability to large numbers of tests m, a critical feature for high-throughput applications like genomics. The BH procedure requires sorting p-values, achieving linearithmic time complexity O(m \log m), which enables efficient handling of m > 10^6 in genome-wide association studies and sequencing analyses. Efficient implementations in statistical software further support this, with benchmarks confirming run times under seconds for m = 10^5 across various FDR methods. Empirical simulations under validate the asymptotic behavior of these procedures, showing that FDR holds as m \to \infty, with the actual FDR converging to the nominal level. Adaptive methods outperform non-adaptive ones in detecting more true positives under sparse alternatives, where only a small fraction of hypotheses are false, as demonstrated in studies. A limitation arises in the BY procedure for ultra-large m, where the dependency adjustment factor c(m) = \sum_{i=1}^m 1/i (the m-th ) requires O(m) , potentially intensive for m \gg 10^6. However, approximations such as c(m) \approx \ln m + \gamma, with \gamma \approx 0.57721 the Euler-Mascheroni constant, mitigate this by reducing to constant-time evaluation while preserving FDR bounds.

Handling Dependencies

Dependencies among test statistics can significantly impact the control of the false discovery rate (FDR) in multiple testing procedures. Positive dependencies, such as positive dependence on a (PRDS) of true hypotheses, are common in applications like where test statistics from correlated gene expressions exhibit this structure; under PRDS, the Benjamini–Hochberg (BH) procedure controls the FDR at the nominal level q. In contrast, negative dependencies weaken the FDR control guarantees of the BH procedure, potentially leading to exceedance of the target rate, necessitating more conservative adjustments like the Benjamini–Yekutieli (BY) procedure, which incorporates a factor c(m) = \sum_{i=1}^m \frac{1}{i} to ensure control under arbitrary dependence structures. Theoretically, under arbitrary dependence, the BH procedure satisfies \mathrm{FDR}(\mathrm{BH}) \leq q \cdot c(m), providing a bound that accounts for the worst-case scenario, including negative associations. Simulations demonstrate that the BH procedure remains conservative under weak positive dependencies, meaning the actual FDR often falls below the nominal level, which can reduce power but ensures reliability in moderately correlated settings. To address unknown dependence empirically, permutation tests can estimate the dependence structure by resampling the data while preserving correlations, allowing for adjusted p-values that control FDR under general dependence. Similarly, the knockoff framework generates exchangeable "knockoff" statistics that mimic the original covariates' dependence, enabling exact FDR control at level q without assuming or specific dependence types. In neuroimaging, spatial correlations among voxel-based test statistics inflate the number of false positives V, as adjacent brain regions often share signal dependencies; here, the BY procedure or structured methods like topological FDR adjustments are recommended to mitigate these effects and maintain valid control.

Proportion of True Nulls

In false discovery rate (FDR) theory, the proportion of true null hypotheses, denoted \pi_0, represents the fraction of null hypotheses that are truly null among the total m hypotheses tested, formally defined as \pi_0 = m_0 / m, where m_0 is the number of true nulls. This parameter fundamentally influences FDR control by quantifying the sparsity of signals: a high \pi_0 (close to 1) indicates few true alternatives, making the signal sparse, whereas a low \pi_0 suggests denser effects. Adaptive FDR methods gain substantial power in high-\pi_0 scenarios by estimating \pi_0 < 1, avoiding the conservatism of procedures that implicitly assume \pi_0 = 1, such as the Benjamini-Hochberg method. Estimation of \pi_0 typically leverages the uniform distribution of p-values under true nulls. Storey's seminal \lambda-based estimator computes \hat{\pi}_0(\lambda) = \#\{p_i > \lambda\} / [m(1 - \lambda)], with \lambda = 0.5 as a common choice to balance bias and variance by excluding small p-values likely from alternatives. For greater robustness, especially against non-uniform patterns, spline methods fit a natural cubic spline to \hat{\pi}_0(\lambda) across a fine grid of \lambda values (e.g., from 0 to 0.95) and extrapolate to \lambda = 1, thereby smoothing fluctuations and reducing error. Histogram-based approaches further improve reliability by modeling the empirical density of ordered p-values in bins, fitting linear or retroflective patterns to derive conservative estimates that perform well under weak dependence. Biases in \pi_0 estimation arise primarily from the choice of \lambda: a low \lambda tends to overestimate \pi_0 due to inclusion of p-values from alternatives, leading to conservative rejections, while a high \lambda reduces this bias but increases variance, potentially allowing more liberal rejections. The conservative oracle \hat{\pi}_0 = 1 equates to the Benjamini-Hochberg procedure and sacrifices power. Under independence, these estimators exhibit asymptotic consistency, satisfying \mathbb{E}[\hat{\pi}_0] \to \pi_0 as m \to \infty, ensuring reliable FDR control in large-scale settings. The impact of estimation errors highlights the sensitivity of FDR outcomes to \pi_0. Simulations reveal that a 10% relative in \hat{\pi}_0 can change the number of rejections by 20-50%, with overestimation reducing discoveries (e.g., from 787 to 584 rejections in a with m=1000 and true \pi_0=0.5) and underestimation risking FDR .

Power and Performance

The average of false discovery rate (FDR) controlling procedures is defined as the expected proportion of true hypotheses that are correctly , denoted as \mathbb{E}[S / m_1], where S is the number of true discoveries and m_1 is the number of false hypotheses. This quantifies the of FDR methods to detect genuine signals while controlling the expected proportion of false positives among rejections. In contrast to (FWER) procedures, which prioritize avoiding any false positives at the cost of , FDR methods offer higher average , particularly when the number of tests m is large, as they allow a controlled fraction of errors. Comparisons between the Benjamini–Hochberg (BH) procedure and conservative alternatives like the Bonferroni method highlight substantial gains for FDR. Under independence, the BH procedure's power can be approximated asymptotically as p^* \frac{1 - (1 - \pi_1 \alpha)}{\pi_1}, where p^* is the proportion of rejected hypotheses, \pi_1 is the proportion of false nulls, and \alpha is the target FDR level; this reflects higher detection rates relative to FWER bounds. Simulations demonstrate that BH often yields 2–10 times more discoveries than Bonferroni in sparse scenarios with few signals (\pi_1 \approx 0.05–0.1), as the latter's threshold \alpha/m severely limits rejections. FDR procedures exhibit high in sparse signal settings when using adaptive variants, such as the Storey procedure, which estimates the proportion of true nulls \pi_0 to relax thresholds and boost detections without inflating errors. However, power declines under strong positive dependence among test statistics unless adjustments like the Benjamini–Yekutieli procedure are applied, as correlations inflate the effective number of independent tests and reduce the rejection threshold. Marginal power per test, assessed via the Type II error rate, further underscores these dynamics, with adaptive FDR maintaining near-optimal rejection probabilities for weak signals. Empirical studies in , such as differential expression analyses in data, illustrate these performance characteristics, where FDR methods detect up to 80% of true effects compared to approximately 20% for FWER controls like Bonferroni, especially in high-dimensional settings with thousands of genes. These gains stem from FDR's tolerance for some false positives, enabling broader exploration of sparse genomic signals while preserving interpretability.

False Coverage Rate

The false coverage rate (FCR) is a measure used in selective inference to control the expected proportion of false coverage statements among selected intervals, analogous to the false discovery rate (FDR) in hypothesis testing. Formally, it is defined as \text{FCR} = E\left[ \frac{V}{R} \mid R > 0 \right] P(R > 0), where R is the number of selected intervals, V is the number of those intervals that fail to cover their true parameters \theta, and the expectation is taken over the joint distribution of the data. This quantity relates to the FDR by extending the control of false positives in testing to errors in estimation after selection; under point hypotheses, the FCR marginalizes to the FDR when considering coverage failures as discoveries of non- effects. It addresses coverage errors in selective , where intervals are constructed only for parameters deemed significant, ensuring that the average among selected intervals remains valid despite conditioning on selection. Procedures for controlling the FCR adapt multiple testing methods to interval construction. A Benjamini–Hochberg (BH)-style step-up procedure selects parameters based on sorted p-values and then builds simultaneous confidence intervals at level $1 - (R \cdot q / m), where R is the number selected, q is the target FCR, and m is the total number of parameters; this controls the FCR at most q under assumptions. For settings with , the Benjamini–Yekutieli procedure adjusts the level by a factor accounting for positive dependency or general dependence, bounding the FCR by q or q \cdot \sum_{j=1}^m 1/j. In applications, FCR control is particularly useful post-FDR selection, such as constructing bands for sizes of significant genes in genomic studies; for instance, in a analysis, FCR-adjusted intervals for selected odds ratios maintain a 95% coverage while excluding null values for the selected set. A key property is that controlling the FCR at level [q](/page/Q) guarantees valid selective coverage, meaning the expected proportion of non-coverages among selected intervals is at most [q](/page/Q), providing on-average protection without requiring simultaneous coverage over all possible selections.

Family-Wise Error Rate Comparisons

The (FWER) is defined as the probability of committing at least one Type I error across a family of m simultaneous tests, formally P(V \geq 1) \leq \alpha, where V is the number of false positives and \alpha is the desired error rate. This contrasts with the false discovery rate (FDR), which instead controls the expected proportion of false positives among all rejected hypotheses, allowing a controlled number of errors proportional to discoveries rather than strictly bounding the probability of any error. Common methods for controlling the FWER include the , which divides the significance level by the number of tests (\alpha / m), providing a simple but conservative adjustment valid under arbitrary dependence structures. The Holm-Bonferroni step-down procedure improves upon this by sequentially adjusting p-values in a less conservative manner, rejecting hypotheses one at a time until the adjusted threshold is violated, while still strongly controlling the FWER. For independent tests, the offers a slightly more powerful alternative to Bonferroni, using $1 - (1 - \alpha)^{1/m} as the adjusted threshold. FWER control is stricter than FDR, ensuring no false positives occur with probability at least $1 - \alpha, but this conservativeness incurs a power loss scaling approximately as O(\alpha / m) for large m, making it suitable for small families of tests or scenarios where any false positive is intolerable, such as confirmatory clinical trials. In contrast, FDR procedures sacrifice this absolute guarantee for greater power in exploratory analyses involving large-scale testing, like genomics or -omics studies, where discovering true signals amid many tests is prioritized over zero error risk. Hybrid approaches, such as the Šidák-Holm step-down procedure, combine stepwise rejection with independence assumptions to achieve FWER with improved power over single-step methods like Bonferroni. Notably, when all hypotheses are true (\pi_0 = 1), FDR coincides with FWER , as the proportion of false discoveries equals the probability of any discovery. Empirical simulations demonstrate these trade-offs: for m = 1000 tests with moderate effect sizes, FWER methods like Bonferroni or Holm typically detect far fewer true positives—often by a factor of several times less—compared to FDR procedures at equivalent nominal levels (e.g., α = q = 0.05), while FWER maintains higher specificity by virtually eliminating false positives.

Bayesian Approaches

Bayesian approaches to false discovery rate (FDR) control frame the multiple testing problem within a probabilistic model that incorporates distributions on hypotheses, enabling the computation of posterior probabilities for decision-making. The core framework models the of p-values or test statistics as a mixture: f(p) = \pi_0 f_0(p) + \pi_1 f_1(p), where \pi_0 represents the proportion of true null hypotheses, f_0(p) is the density under the null (typically on [0,1] for p-values), and f_1(p) captures the alternative density. This two-groups model allows for on whether each hypothesis is null or alternative, providing a natural way to estimate error rates beyond frequentist thresholds. A key quantity in this framework is the local false discovery rate (lfdr), defined as the that the H_0 holds given the observed p_i: \text{lfdr}(p_i) = P(H_0 \mid p_i) = \frac{\pi_0 f_0(p_i)}{f(p_i)}. Hypotheses are rejected when \text{lfdr}(p_i) \leq 1 - q, where q is the target FDR level; the average lfdr over rejected hypotheses then approximates the realized FDR. This local perspective offers finer-grained control compared to global procedures, as it conditions on each individually. In practice, the mixture parameters \pi_0, f_0, and f_1 are unknown and estimated via , which borrow strength across tests to infer the prior from the data itself. For instance, Efron (2008) outlines nonparametric techniques, such as fitting the null component from the central region of the observed and using for the alternative, ensuring the procedure controls the marginal FDR under mild assumptions. These methods are particularly effective in high-dimensional settings like , where direct Bayesian computation is infeasible. Bayesian FDR methods excel at handling complex dependencies among tests by leveraging the full posterior , rather than assuming , which enhances power in correlated data scenarios. The q-value, while originally a frequentist construct, finds a Bayesian analog in the lfdr, serving as a posterior-expected FDR for each discovery. In applications such as , hierarchical priors on effect sizes further refine the model; for example, these priors account for varying protein abundances by shrinking estimates toward a global mean, improving FDR control in identification. The limma software implements empirical Bayes moderation of variances within this paradigm, stabilizing inferences for differential expression analysis.

Implementations and Applications

Software Packages

Several software packages facilitate the implementation of false discovery rate (FDR) procedures in statistical analysis, supporting methods like for multiple testing across diverse applications. These tools are available in popular programming languages and often integrate with domain-specific workflows, such as . In , the multtest package, part of , offers resampling-based multiple testing procedures, including , Benjamini-Yekutieli (BY), and methods for FDR control. The qvalue package estimates the proportion of true null (π₀) and computes q-values, featuring visualization tools like plots of q-values against p-values to assess FDR results. Additionally, fdrtool provides estimation of local false discovery rates (lfdr) from various test statistics, such as z-scores or p-values. Python libraries also support FDR corrections effectively. The statsmodels package includes the fdrcorrection function for BH and BY adjustments on p-values, suitable for independent or dependent tests. Pingouin offers FDR options in its multicomp function, implementing BH and BY corrections for post-hoc analyses. The scikit-posthocs package provides pairwise comparisons with Holm and BH (fdr_bh) corrections integrated into nonparametric tests. Beyond and Python, MATLAB's mafdr function estimates positive FDR for multiple hypothesis testing, particularly tailored for data using Storey's adaptive procedure. In applications, Bioconductor's limma package applies empirical Bayes moderation via eBayes and subsequent FDR adjustment in topTable for expression analysis. Key features across these packages include handling dependencies, such as the BY method in multtest for general or negatively dependent hypotheses. Many tools support visualization of q-values and seamless integration with high-throughput pipelines, especially Bioconductor packages for large-scale genomic data processing. Most implementations default to the BH procedure; for adaptive FDR, users should confirm π₀ estimation in packages like qvalue.

Practical Examples

In genomics, a common application of the false discovery rate (FDR) involves analyzing microarray data to identify differentially expressed genes. Consider a simulated experiment with m=20,000 genes across 12 samples, where 10% are truly differentially expressed; applying the Benjamini-Hochberg (BH) procedure at q=0.05 rejects approximately 500 genes while estimating 25 false positives, achieving an FDR of 5%. In contrast, the Bonferroni correction, which controls the family-wise error rate, rejects 0 genes in this scenario due to its stringent adjustment, highlighting FDR's greater power for high-dimensional data. In , FDR is frequently used for multiple testing across voxels in (fMRI) data to detect activated regions while accounting for spatial dependencies. The Benjamini-Yekutieli (BY) adjusts for positive dependence among tests and has been applied in fMRI analyses involving tens of thousands of voxels to control the FDR, for example at q=0.05, reducing false positives in correlated signals. Additionally, transformation-invariant FDR methods enable consistent identification of activated regions, such as the in . A typical for applying FDR in such analyses begins with loading raw data, such as counts or intensities, into statistical software. P-values are then computed for each using appropriate tests, like t-tests for differential expression or group comparisons. The procedure is applied to these p-values to generate q-values, which represent FDR-adjusted significance levels. Finally, q-value plots are interpreted to visualize the distribution of discoveries and confirm FDR control below the target threshold, often revealing a sharp drop-off in q-values for truly significant features. One key pitfall in FDR applications arises in volcano plots, which combine thresholds with filters and can inflate the FDR by up to 77% in small samples due to unadjusted subset selection. To mitigate this, adjusted thresholds via closed testing or focused procedures should be used to ensure FDR control over the filtered set.