Fact-checked by Grok 2 weeks ago

Active learning (machine learning)

Active learning is a subfield of in which a learning interactively selects the most informative unlabeled points from which it can obtain labels from an , typically a annotator, to improve model performance while minimizing the cost of labeling. This approach leverages the abundance of unlabeled and the expense of obtaining labels, enabling algorithms to achieve higher accuracy with fewer training examples compared to passive supervised learning. The core hypothesis is that by querying strategically chosen instances, the learner can focus on that maximally reduces uncertainty or error in the model. Active learning operates in various scenarios, including pool-based sampling, where queries are selected from a fixed set of unlabeled examples; stream-based selective sampling, which processes data sequentially and decides whether to query each instance; and membership query synthesis, where the algorithm generates new data points to label. Common querying strategies include uncertainty sampling, which prioritizes instances where the model's prediction is least confident (e.g., when the is closest to 0.5 for ); query-by-committee, which uses a group of models to identify examples on which they disagree; and expected error reduction, which selects queries anticipated to minimize overall future prediction error. These methods have been foundational since the early , with influential surveys highlighting their efficiency in reducing labeling efforts in controlled settings. In recent developments, has integrated with architectures, such as convolutional neural networks for tasks like and semantic segmentation, where techniques like entropy-based sampling outperform traditional methods. Applications span (e.g., text classification and ), materials science (e.g., regression tasks in AutoML), and scenarios, often incorporating human-inspired interactive querying to boost performance by 20-25%. Challenges persist, including handling class imbalance, ensuring fairness, and scaling to large datasets, with ongoing research emphasizing standardized benchmarks and robust uncertainty estimation via Bayesian approaches. Overall, remains a vital technique for resource-efficient in data-scarce labeling environments.

Fundamentals

Definition and Core Principles

Active learning is a subfield of in which a learning algorithm interactively queries an —typically a annotator—to obtain labels for carefully selected unlabeled points, thereby optimizing the learning process with reduced annotation effort. This semi-supervised paradigm enables the model to focus on the most informative examples, contrasting with traditional approaches that rely on fixed or randomly sampled datasets. The core idea is that strategic data selection can yield superior performance compared to labeling all available data indiscriminately. In passive supervised learning, the training data is predetermined or sampled randomly, often leading to inefficient use of labeling resources, especially when unlabeled data is abundant but expensive to annotate. Active learning addresses this by empowering the learner to choose data points that maximize informational gain, such as those near decision boundaries or in underrepresented regions of the input space. It further distinguishes itself from semi-supervised learning, which infers structure from unlabeled data without oracle queries, by emphasizing explicit label acquisition to resolve uncertainties. The framework comprises several essential components: the learner model, which updates based on acquired labels; an unlabeled data pool serving as the source for potential queries; the , responsible for providing accurate labels; and a that determines which instances to select for labeling. These elements operate iteratively, allowing the to adaptively refine the model. Formally, given an unlabeled D_u and a hypothesis space \mathcal{H}, the process begins with a small initial labeled set L. At each t, the algorithm selects an instance x_t \in D_u using the , obtains the label y_t from the , and incorporates the pair (x_t, y_t) into L to retrain the model. The overarching goal is to attain high predictive accuracy with substantially fewer labeled examples than required in settings.

Motivation and Historical Context

Active learning emerges as a critical paradigm in due to the substantial costs associated with labeling data, particularly in domains like and , where expert annotation is labor-intensive and prone to variability. These costs can limit the scale of training datasets, hindering model development in resource-constrained settings. By selectively querying the most informative instances for labeling, active learning mitigates these challenges, enabling efficient model training with fewer annotations. In systematic reviews, empirical evaluations show that strategies can reduce the screening workload by 64-92% while finding 95% of relevant records, achieving comparable performance to fully supervised baselines. This label efficiency translates to significant cost savings and faster iteration cycles. Beyond cost reduction, facilitates quicker convergence to high-accuracy models, enhances adaptability to imbalanced datasets or by prioritizing underrepresented samples, and promotes better in environments with scarce . These benefits are especially pronounced in high-stakes applications where obtaining exhaustive labels is impractical. The conceptual foundations of originated in the 1980s within , with Angluin's 1987 model of exact learning via membership and queries providing a formal framework for learner-oracle interactions to identify target concepts efficiently. This query-based approach laid the groundwork for subsequent developments in during the 1990s, when pool-based gained traction for practical applications like text classification. Pioneering work by Lewis and in 1994 introduced uncertainty sampling, demonstrating up to a 500-fold reduction in manual labeling for text classifiers by targeting instances with high predictive ambiguity. Similarly, McCallum and Nigam's 1998 integration of expectation-maximization with pool-based querying improved semi-supervised text categorization by leveraging unlabeled data pools. The early 2000s saw further advancements tailored to specific models, such as Tong and Koller's 2001 active learning algorithm for support vector machines, which version-space sampling to minimize labeling needs in text tasks. Interest surged in the 2010s amid the deep learning boom, as active learning addressed the data-hungry nature of neural networks; Burr Settles' 2010 survey synthesized these trends, emphasizing strategies for real-world deployment and label minimization. By the 2020s, active learning evolved through deeper integration with automated machine learning (AutoML) systems, automating query selection and model optimization for broader accessibility. A 2025 benchmark in Scientific Reports exemplifies this progress, evaluating active learning within AutoML for small-sample regression in materials science, where it accelerated property prediction with limited annotations.

Learning Scenarios

Pool-Based Active Learning

Pool-based active learning operates in a scenario where a large, fixed of unlabeled U is available upfront, alongside a small initial set of L. The learner iteratively selects a of k instances from U to query for labels from an , removes the selected instances from U, incorporates the newly into L, and retrains the model on the updated L. This setup assumes a closed-world where the U is representative but not necessarily exhaustive of the underlying data distribution, and the provides accurate, deterministic labels without noise or cost variability. The process typically begins with a minimal labeled set L (often bootstrapped from random sampling or heuristics) and proceeds in rounds until a labeling is exhausted or a performance threshold is reached. Key advantages include its alignment with real-world applications where unlabeled data is readily accessible in bulk, such as or text , enabling efficient pre-computation and storage for repeated model . It also facilitates lookahead strategies, where the entire can be evaluated to prioritize queries, potentially leading to more informative selections compared to scenarios without full access. Empirically, pool-based active learning is evaluated through learning curves plotting model performance (e.g., accuracy) against cumulative labeling budget, often demonstrating substantial reductions in required labels; for instance, in text classification tasks, it achieves 81% accuracy with 30 queries 73% for random sampling. Unlike stream-based active learning, which processes unlabeled data sequentially as it arrives without full access to the pool, pool-based methods rank and select from the static U, allowing batch queries and comprehensive informativeness assessments. In contrast to membership query synthesis, pool-based active learning restricts queries to existing instances in U, avoiding the generation of potentially unnatural points.

Stream-Based Active Learning

In stream-based active learning, unlabeled data instances arrive sequentially from a continuous , and the learner must make an immediate, irrevocable decision for each instance: query its at a or discard it without labeling. This scenario operates under tight constraints on memory and labeling budget, preventing the storage of past instances for later review. The approach assumes that unlabeled data is abundant and inexpensive to observe, but labels require or . Early foundational work on applying statistical models to this sequential selection process was introduced by Cohn, Ghahramani, and in 1996. The primary challenges in this setting stem from the lack of a fixed pool of candidates for batch selection, necessitating informativeness evaluation without hindsight. Decisions are final, increasing the risk of missing valuable instances, and the stream may exhibit non-stationarity, such as concept drift where the data distribution evolves over time. These factors demand algorithms that balance exploration of uncertain regions with of current while maintaining computational for high-velocity . The workflow typically proceeds as follows: upon arrival of an instance x, the current model computes an informativeness score, often using uncertainty estimation or expected improvement metrics; if the score surpasses a dynamic , the is queried and incorporated to the model incrementally; otherwise, the instance is discarded to conserve resources. Common acquisition strategies, such as uncertainty sampling, are adapted here by evaluating single points rather than sets. Variants include fixed-budget protocols, where querying halts once a predefined labeling is reached, prioritizing the most informative instances early in the stream, and reservoir-based methods that maintain a small, randomly sampled of recent unlabeled to enable limited lookahead or batch decisions without violating strict streaming constraints. These adaptations enhance robustness in resource-limited environments. This offers advantages in online applications, such as sensor networks for or web log analysis for , where flows continuously and real-time model adaptation is critical for performance.

Membership Query Synthesis

In membership query synthesis, the active learner generates arbitrary synthetic instances from the input space and queries an for their corresponding , unbound by any existing pool of unlabeled data. This scenario enables targeted exploration of the input space, such as probing regions near decision boundaries to refine the model's understanding of the target function. Unlike pool-based or stream-based settings, it allows the learner to propose any point x in the instance space and receive the y, facilitating precise boundary delineation in scenarios where natural data may be sparse or unrepresentative. The theoretical foundations trace back to Dana Angluin's model of exact learning from queries, introduced in 1987, where membership queries permit the learner to determine if a given input belongs to the unknown target concept, often paired with equivalence queries for verification. In this framework, applied to learning regular languages via finite automata, the approach guarantees exact identification of the target in polynomial time for certain concept classes. This exact-learning paradigm was later adapted to statistical by Cohn et al. in 1996, shifting the emphasis from perfect concept recovery to probabilistic models that minimize expected future error through query selection, integrating Bayesian principles to guide synthesis in neural networks and other statistical estimators. The typical workflow begins with the current model generating candidate instances, for example, by perturbing points near predicted decision boundaries or sampling from a learned over the input . These synthetic queries are then submitted to the —such as a , automated simulator, or robotic system—for labeling. The received labels are incorporated into the training set, prompting model retraining to update version or posterior distributions, iteratively narrowing . Key advantages include the ability to directly target potentially informative regions, bypassing biases in natural data distributions, which proves effective for exact identification within finite spaces. In automated domains, it can yield substantial efficiency gains; for instance, the robot scientist "Adam" used membership query synthesis to autonomously design and execute yeast functional genomics experiments, reducing costs by a factor of up to 100 compared to random selection and by a factor of three compared to naive methods. However, limitations arise from oracle constraints: human annotators often reject or inaccurately label implausible synthetics, like garbled text or unrealistic images, rendering the approach impractical for subjective or perceptual tasks. While less prevalent in traditional empirical settings after the early , membership query synthesis has seen renewed interest since the , particularly with generative models for tasks like semi-supervised sentence classification, and persists in hybrid symbolic systems for automated testing and scientific . Recent advancements (as of 2023) have integrated membership query synthesis with deep generative models, such as variational autoencoders, to produce more natural synthetic instances for labeling in semi-supervised settings, mitigating oracle rejection concerns.

Acquisition Strategies

Uncertainty Sampling

Uncertainty sampling is a fundamental acquisition strategy in active learning that selects unlabeled data points for annotation based on the current model's lack of confidence in its predictions for those points. Introduced in the context of text classification, this approach queries the instance x from an unlabeled pool U that maximizes the model's , often measured through posterior probabilities or , to efficiently refine the with minimal labeling effort. The strategy assumes access to a probabilistic classifier capable of outputting confidence scores, making it particularly suitable for scenarios where labeling costs are high and the learner aims to prioritize informative examples. Common uncertainty measures include least confidence, which selects points where the highest predicted probability is minimized, formally defined as $1 - \max_y p(y \mid x; \theta), where \theta denotes the model's parameters. Margin sampling extends this by focusing on the difference between the top two predicted probabilities, \max_y p(y \mid x; \theta) - \max_{y' \neq y} p(y' \mid x; \theta), capturing cases where the model is torn between closely competing classes. Entropy-based sampling, meanwhile, quantifies overall predictive via the entropy H(y \mid x; \theta) = -\sum_y p(y \mid x; \theta) \log p(y \mid x; \theta), providing a more comprehensive distribution-aware metric. The formal selection criterion unifies these as \arg\max_{x \in U} \text{[uncertainty](/page/Uncertainty)}(x; \theta), where the specific uncertainty function is chosen based on the task. This strategy applies to probabilistic classifiers such as and neural networks, which naturally produce calibrated probability outputs, and it remains computationally inexpensive since it requires only forward passes through the model without additional simulations. In practice, it excels on balanced datasets by rapidly improving accuracy; for instance, early experiments on text categorization tasks demonstrated up to a 500-fold reduction in required compared to random sampling. However, it can underperform on imbalanced datasets, as high-uncertainty points often cluster around majority classes, potentially overlooking rare instances and exacerbating class imbalance in the selected samples. Variants of uncertainty sampling incorporate methods to estimate more robustly, such as version space density, which leverages the density of hypotheses consistent with to weight individual model predictions and better handle distributional shifts.

Query by Committee

Query by committee (QBC) is an active learning acquisition strategy that trains a committee of multiple models, denoted as |C|, on the current and selects unlabeled instances from the pool U that elicit the maximum disagreement among committee predictions. The committee members can be generated through posterior sampling in Bayesian models or techniques such as bagging, ensuring diversity in the hypotheses considered. This method prioritizes exploration by targeting regions of the input space where the models diverge, thereby refining the learner's understanding efficiently. The approach was first formalized by Seung et al. in 1992 as a means to shrink the version space in . Disagreement is quantified using measures like vote entropy or across the committee's output distributions. A standard formal measure of disagreement D(x) for an instance x is D(x) = H\left( \frac{1}{|C|} \sum_{c \in C} p_c(y \mid x) \right) - \frac{1}{|C|} \sum_{c \in C} H\left( p_c(y \mid x) \right), where p_c(y \mid x) is the predictive over labels y from member c, and H(\cdot) is the Shannon function. A simpler variant, vote entropy, computes the directly on the averaged vote proportions when using hard classifications. The query is then selected as x^* = \arg\max_{x \in U} D(x), focusing on instances that maximize this . QBC provides robustness against biases inherent in single-model predictions by relying on collective disagreement, which promotes broader exploration of the hypothesis space. In Bayesian settings, it particularly captures epistemic uncertainty, reflecting the learner's ignorance about the true model rather than inherent data noise. Despite these strengths, the strategy's computational cost is substantial, as it requires repeated model trainings and predictions for each committee member during query selection. To address this, approximations leverage scalable ensembles: random forests form implicit committees via diverse decision trees, while dropout ensembles approximate posterior samples through stochastic forward passes on a single , reducing the need for independent trainings.

Expected Error Reduction

Expected error reduction is an active learning strategy that selects unlabeled instances by estimating the anticipated decrease in the learner's future upon labeling and incorporating that instance into the training set. This approach directly targets the optimization of model by simulating the impact of potential labels on error metrics, such as the over a held-out validation set or the Bayes . For each candidate instance x from the unlabeled , the method computes a weighted average over possible labels y, drawn from the current model's predictive distribution p(y|x), evaluating the error reduction after hypothetically retraining the model with each (x, y) pair. The formal objective is to query the instance that maximizes the expected reduction in error: \arg\max_{x} \mathbb{E}_{y \sim p(y|x)} \left[ \operatorname{Err}(\theta; L \cup \{(x,y)\}) - \operatorname{Err}(\theta; L) \right], where L is the current labeled training set, \theta are the model parameters, and \operatorname{Err}(\theta; D) denotes the error (e.g., 0-1 loss or log loss) on a validation set D. For 0-1 loss, this is approximated as $1 - \max_y p(y|x') averaged over validation points x', while log loss uses the -\sum_y p(y|x') \log p(y|x'). Computation typically involves sampling: for a subset of validation examples and possible labels, the model is retrained incrementally, and the expected loss is averaged. This strategy was introduced by Roy and McCallum, who demonstrated its application to naive Bayes classifiers and its superiority over uncertainty sampling and query-by-committee in text classification tasks, achieving approximately 77% accuracy on tasks using subsets of the 20 Newsgroups with only 16 additional queries (beyond initial labels) compared to 68 for query-by-committee. Variants of expected error reduction approximate the computationally expensive retraining step to make the method feasible for larger models. One such variant is expected model change (), which measures the anticipated update to the model parameters via the expected length of the gradient \|\nabla_\theta \log p(y|x)\|, weighted over possible labels; this avoids full retraining by leveraging insights and has been applied to tasks, where it selects instances that induce the largest parameter shifts. Another variant uses the expected between the current posterior and the updated posterior after labeling, providing a Bayesian information-theoretic for error reduction without explicit error computation; this approach, explored in early work on probabilistic models, aligns with minimizing in parameter estimates. Theoretically, expected error reduction is grounded in , as it greedily minimizes the anticipated , making it optimal in expectation for single-step error minimization under the assumption that the next query is the final one. In practice, it has shown robust performance gains, such as reducing the number of required by up to 75% in domains while maintaining high accuracy. However, its primary drawback is extreme computational cost, scaling with the pool size |U|, label space |Y|, and retraining overhead, often requiring O(|U| \cdot |Y|) model updates per query; approximations like validation sets, models, or incremental learners mitigate this but introduce .

Hybrid and Batch Strategies

Hybrid strategies in active learning combine multiple acquisition criteria to balance informativeness and representativeness, often integrating measures with or assessments to select samples that are both informative for the model and reflective of the data distribution. For instance, density-weighted sampling adjusts scores by the local of unlabeled points, prioritizing queries in densely populated regions to avoid over-sampling outliers while leveraging predictive . Core-set selection further exemplifies this approach by identifying compact subsets of unlabeled data that geometrically approximate the entire pool, ensuring selected points capture the manifold structure for efficient labeling. Batch-mode active learning extends these hybrids by querying multiple points (k > 1) per iteration, amortizing the fixed costs of oracle interactions such as human labeling rounds and enabling scalability for large datasets. Diversity metrics are crucial in batch selection to prevent redundancy; determinantal point processes (DPPs) model point repulsion via a kernel matrix, probabilistically sampling diverse subsets that maximize both quality (e.g., uncertainty) and coverage. Similarly, the k-center greedy algorithm iteratively selects the point farthest from the current batch in feature space, approximating facility location to ensure uniform representation across the data. Formal batch selection often optimizes a combined objective, such as maximizing the sum of individual uncertainties plus a diversity term weighted by λ, formulated as \arg\max_{B \subset U, |B|=k} \sum_{x \in B} u(x) + \lambda \cdot d(B), where u(x) denotes , d(B) measures batch , and U is the unlabeled pool; this hybrid score promotes both model improvement and reduced correlation among queried points. A seminal example is the minimum marginal hyperplane strategy for support vector machines (SVMs), which selects points minimizing the distance to the current , \min \frac{|w \cdot x + b|}{\|w\|}, thereby efficiently shrinking the version space; this can be batched by greedily adding the k closest margin points to accelerate convergence in text classification tasks. These strategies offer key advantages, including mitigation of labeling bottlenecks through queries and superior on correlated compared to sequential single-point selection, as batched helps explore broader regions of the input space. In recent deep active learning applications as of 2025, integrations like the Batch Active learning by Diverse, Gradient Embeddings () algorithm embed unlabeled points in space to jointly capture (via ) and (via k-DPP sampling), achieving robust gains in image and text domains with neural networks. More recent advances, particularly in (LLM) contexts, extend hybrid strategies to generative acquisition, where LLMs synthesize new examples and labels via prompting and to augment datasets efficiently, or use explanation-generation for rationale-guided diverse selection to improve label quality.

Theoretical Foundations

Generalization Bounds

In the realizable case, where there exists a hypothesis in the class consistent with all labels, active learning algorithms can achieve a query complexity of \tilde{O}(d \log(1/\epsilon)) labels to obtain an \epsilon-error classifier with high probability, where d is the VC dimension of the hypothesis class. This contrasts with passive learning, which requires \Theta(d / \epsilon) labels in the worst case to reach the same error level, as active methods selectively query points that refine the hypothesis space more efficiently. Such bounds highlight how active learning exploits the learner's ability to choose queries, reducing label complexity from linear in $1/\epsilon to polylogarithmic under margin or separability assumptions. In the agnostic setting, where no perfect hypothesis exists and the goal is to minimize excess risk relative to the best-in-class error, probably approximately correct (PAC)-style generalization bounds for active learning show that label complexity can match passive rates of O(d / \epsilon^2) but with fewer labels under low-noise conditions, such as Tsybakov margin assumptions. For instance, the confidence-weighted active learning (CAL) algorithm achieves \tilde{O}(d \theta^2 / \epsilon^2) label requests, where \theta is the disagreement coefficient capturing the geometry of the hypothesis class and data distribution. This coefficient, defined as \theta(\epsilon) = \sup \{ P(\mu(A)) / \epsilon : P(\mathrm{DISC}(A)) \leq \epsilon \} with \mathrm{DISC}(A) the region where a version space committee disagrees and \mu(A) its measure under the marginal distribution, bounds future generalization error by quantifying how disagreement regions shrink, thereby linking query selection to calibration of uncertainty estimates. Despite these guarantees, worst-case generalization bounds in active learning are often pessimistic, as they assume adversarial distributions and do not fully capture empirical label savings, which frequently surpass theoretical predictions by orders of magnitude in practice. Recent advances in the have extended these bounds to non-parametric models, such as kernel methods, achieving adaptive rates dependent on intrinsic dimensionality rather than dimension, and to deep networks, where is upper-bounded by combining query informativeness with training dynamics to ensure faster convergence than passive baselines. For deep nets, these results incorporate surrogate losses and alignment metrics to derive non-vacuous bounds, demonstrating label efficiency gains in high-dimensional settings. As of 2025, theoretical extensions to large language models incorporate disagreement coefficients adapted for sequential data, achieving improved bounds in non-i.i.d. settings.

Query Efficiency and Convergence

Active learning algorithms demonstrate faster convergence to low error rates compared to methods, particularly in settings with margin-based assumptions or low noise. Under Tsybakov's noise conditions with parameter κ > 1, achieves excess error rates of O(((log n)/n)^{κ/(2κ-2)}), which can exhibit exponential improvements in label usage relative to passive baselines, which achieve excess error rates of O(n^{-κ/(2κ)}) under similar conditions, potentially requiring exponentially more labels in low-noise regimes. For in generalized linear models, a two-stage active approach converges at rates matching passive learning's lower bounds but with significantly fewer labels, often requiring only a single interactive round to approach optimality. Query efficiency in active learning is commonly measured by the label efficiency ratio, defined as the number of labels queried by the active method divided by the number needed in to reach equivalent accuracy. This ratio highlights the reduction in annotation costs; for instance, uncertainty sampling variants achieve O(1/n) convergence in excess risk for , enabling models to attain target accuracies with 20-50% fewer labels than random sampling in low-data regimes. In online stream-based scenarios, efficiency is further quantified via , representing the cumulative suboptimality of decisions over time steps T, where active strategies minimize long-term performance gaps. Optimal stopping in active learning relies on criteria that track diminishing returns, such as monitoring marginal gains in validation accuracy or reductions in model after successive queries. When the incremental in accuracy falls below a predefined , querying halts to avoid unnecessary labeling; this approach ensures efficient without over-querying. Bayesian stopping rules, such as those based on expected , evaluate the anticipated reduction in posterior or gap before and after a potential query, providing a probabilistic basis for termination in optimization-aligned active learning. Theoretical analyses establish sublinear bounds of O(√T) for bandit-like active learning formulations, where the learner balances and in sequential , ensuring cumulative error grows slower than linear in the horizon T. Under ideal conditions with access to infinite queries, active methods converge to the Bayes optimal error rate, as the adaptive selection refines the hypothesis class toward the true posterior distribution. In pool-based settings, query strategies accelerate convergence by revisiting the entire unlabeled set, achieving faster error decay than stream-based methods, which suffer slower rates due to the one-pass, no-revisit constraint that limits access to prior instances. Recent benchmarks in (AutoML) for tasks underscore these gains, with query-by-committee strategies reaching performance parity to full-dataset baselines using only 30% of the labels, yielding 70-95% reductions in labeling effort across datasets like band gap predictions. Such results confirm active learning's practical efficiency in data-scarce , where convergence to target occurs 2-3 times faster than passive sampling.

Applications and Challenges

Real-World Applications

Active learning has been effectively applied in classification tasks, particularly in text categorization, where it significantly reduces the need for labeled data. In early work on and , pool-based active learning combined with expectation-maximization allowed models to achieve comparable accuracy using substantially fewer labeled examples, often requiring only a fraction of the annotations compared to random sampling. Similarly, in , active learning facilitates the selective annotation of imaging scans, such as radiology images, by prioritizing uncertain cases, which has been shown to improve labeling efficiency by up to 90% over random selection while maintaining high diagnostic accuracy. In , active learning enhances by leveraging uncertainty estimates to select informative images for labeling. For instance, uncertainty-aware strategies integrated with single-stage detectors like have been used to query frames with high prediction ambiguity, reducing annotation efforts in tasks such as UAV-based detection. This approach is particularly valuable in autonomous driving, where active learning optimizes data labeling for vast sensor datasets by focusing on edge cases like rare traffic scenarios, thereby accelerating model training for perception systems. Beyond these areas, active learning extends to other domains including (), , and (). In , core-set selection methods adapted for transformer models efficiently curate diverse subsets of text data, minimizing labeling while preserving performance in tasks like . For property prediction, recent benchmarks demonstrate that active learning strategies within (AutoML) pipelines can achieve models with small labeled datasets, outperforming in discovering novel compounds. In , active querying of state-action pairs enables efficient exploration, as seen in episodic demonstration frameworks that select trajectories to resolve model disagreements, improving policy learning with fewer human interventions. Integrations of active learning with broader ML ecosystems further amplify its utility. For example, Bayesian Active Learning by Disagreement (BALD) has been incorporated into deep neural networks for scalable , enhancing label efficiency in high-dimensional settings. Tools like Vizier support active learning in AutoML by optimizing data selection alongside hyperparameter tuning, streamlining end-to-end workflows for resource-constrained environments. Case studies illustrate these benefits: Encord's 2023 guide highlights active learning loops for video annotation in projects, enabling iterative model refinement with targeted human input to cut down on manual effort. Likewise, Lightly.ai's data flywheels combine active selection with curation to build efficient training cycles, as demonstrated in production pipelines for image classification. Real-world deployments often yield substantial savings in annotation time and cost, especially on crowdsourcing platforms like (MTurk). Studies integrating active learning with MTurk workflows report reductions of 30-80% in labeling requirements, as the approach prioritizes high-value samples and automates low-confidence predictions, making it cost-effective for large-scale tasks.

Practical Limitations and Extensions

One significant practical limitation of active learning arises from oracle noise and errors, where human annotators exhibit inconsistency due to fatigue, expertise gaps, or ambiguous examples, leading to mislabeled that degrades model performance. exacerbates this issue, as multiple labelers introduce varying levels of , necessitating robust aggregation techniques. Additionally, adversarial poisoning poses a , as malicious actors can exploit query strategies to inject harmful samples, compromising the of selected and causing model . High initial computational costs further hinder deployment, as strategies like uncertainty sampling require repeated model training and scoring on large pools, often demanding resources infeasible for applications. Scalability challenges are prominent in large datasets, where single-instance querying becomes inefficient, prompting the reliance on batch modes to parallelize selections and reduce overhead, though this introduces issues in query selection. The cold-start problem compounds this, particularly when starting with an empty or minimal labeled set, as initial models lack sufficient to make informative queries, resulting in suboptimal early . Violations of core assumptions, such as and identical distribution (i.i.d.), occur in non-i.i.d. streams where query strategies inadvertently bias sampling toward certain regions, skewing the effective training distribution away from the true data manifold. Moreover, the assumption of an exhaustive, fixed unlabeled pool is unrealistic in dynamic environments like , limiting applicability to static scenarios. Extensions to active learning address these limitations through adaptations for modern paradigms. In deep active learning, pseudo-labeling hybrids combine confident model predictions on unlabeled data with selective human queries, enhancing efficiency in high-dimensional spaces like image classification while mitigating cold-start via self-supervised pretraining. Federated active learning emerges as a privacy-preserving extension, enabling distributed querying across devices without centralizing data, which counters non-i.i.d. challenges in heterogeneous environments. Integration with further reduces oracle dependency by leveraging noisy heuristics or programmatic rules alongside active queries, as seen in frameworks that iteratively refine weak labels through targeted human feedback. Mitigations include robust strategies such as outlier detection to filter anomalous queries before labeling, preventing resource waste on irrelevant samples and bolstering resilience to or . Hybrid passive-active pipelines offer another approach, starting with random sampling to warm-start the model before transitioning to active selection, which alleviates cold-start and issues in large-scale deployments. Recent advancements as of 2025 include the use of large language models (LLMs) as AI-assisted oracles in , providing preliminary labels to unlabeled data to reduce human inconsistency and labeling costs, particularly in tasks.

References

  1. [1]
    [PDF] Active Learning Literature Survey - Burr Settles
    Jan 26, 2010 · This report provides a general introduction to active learning and a survey of the literature. This includes a discussion of the scenarios in ...
  2. [2]
  3. [3]
    [PDF] A comprehensive survey on deep active learning in medical image ...
    Active learning selects informative samples for annotation to reduce costs in medical image analysis, where high annotation costs limit large datasets.Missing: motivation | Show results with:motivation
  4. [4]
    [cmp-lg/9407020] A Sequential Algorithm for Training Text Classifiers
    Jul 24, 1994 · This method, which we call uncertainty sampling, reduced by as much as 500-fold the amount of training data that would have to be manually ...
  5. [5]
    [PDF] Support Vector Machine Active Learning with Applications to Text ...
    c 2001 Simon Tong and Daphne Koller. Page 2. Tong and Koller classifier that will eventually be used to classify the rest of the web. Since human expertise ...
  6. [6]
    Performance of active learning models for screening prioritization in ...
    Jun 20, 2023 · Active learning models reduce screening by 63.9-91.7%, find 95% relevant records, and the Naive Bayes + TF-IDF model performed best. Average ...
  7. [7]
    Active Learning Performance in Labeling Radiology Images Is 90 ...
    Nov 29, 2021 · The AI methodology of active learning (AL) can assist human labelers by continuously sorting the unlabeled images in order of information gain.
  8. [8]
    A comprehensive benchmark of active learning strategies ... - Nature
    Oct 23, 2025 · This benchmark study aims to evaluate various active learning (AL) strategies within AutoML in materials science regression tasks. The ...
  9. [9]
    [PDF] Advancements in Active Learning: Strategies for Imbalanced Class ...
    This approach can be particularly beneficial in scenarios where unlabeled data are abundant, but labels are scarce or expensive to obtain [19]. By intelligently ...
  10. [10]
    A study of deep active learning methods to reduce labelling efforts in ...
    Dec 15, 2023 · We show through the experiments the importance of using AL methods to reduce the amount of labelling needed to construct high-quality data sets ...
  11. [11]
    [PDF] Learning Regular Sets from Queries and Counterexamples*
    75, 87-106 (1987). Learning Regular Sets from Queries and Counterexamples*. DANA ANGLUIN. Department of Computer Science, Yale University,. P.O. Box 2158, Yale ...Missing: exact | Show results with:exact
  12. [12]
    Employing EM and Pool-Based Active Learning for Text Classification
    ICML '98: Proceedings of the Fifteenth International Conference on Machine Learning. Employing EM and Pool-Based Active Learning for Text Classification.
  13. [13]
    Based Active Learning - an overview | ScienceDirect Topics
    There are two types of active learning: stream-based and pool-based. In pool-based active learning, the best query is selected from the entire unlabeled set.
  14. [14]
  15. [15]
    Enhanced uncertainty sampling with category information for ...
    Traditional uncertainty sampling methods in active learning often neglect category information, leading to imbalanced sample selection in multi-class ...
  16. [16]
    [PDF] Toward Optimal Active Learning through Monte Carlo Estimation of ...
    This paper presents an active learning method that di- rectly optimizes expected future error. This is in con- trast to many other popular techniques that ...
  17. [17]
    Maximizing Expected Model Change for Active Learning in ...
    In this paper, we propose a new active learning framework for regression called Expected Model Change Maximization (EMCM), which aims to choose the examples ...
  18. [18]
    [PDF] Active learning Thomas P. Minka Abstract
    ... expected KL-divergence: ─ = /╟p(ЭЬD)╚(p(xЬЭ) ЬЬ q(x)). (1). = /╟ p(x, ЭЬ D ) logp(xЬЭ) q(x). (2). The first question is: what estimate q(x) should we choose ...
  19. [19]
    [PDF] Dual Strategy Active Learning - CMU School of Computer Science
    Combining uncertainty with the density of the underlying data is a good strategy to reduce the error quickly. However, after rapid initial gains, DWUS exhibits ...
  20. [20]
    Active Learning for Convolutional Neural Networks: A Core-Set ...
    Aug 1, 2017 · We define the problem of active learning as core-set selection, ie. choosing set of points such that a model learned over the selected subset is competitive ...
  21. [21]
    Batch Active Learning Using Determinantal Point Processes - arXiv
    Jun 19, 2019 · In this paper, we present a new principled batch active learning method using Determinantal Point Processes, a repulsive point process that enables generating ...
  22. [22]
    Deep Batch Active Learning by Diverse, Uncertain Gradient Lower ...
    Jun 9, 2019 · Our algorithm, Batch Active learning by Diverse Gradient Embeddings (BADGE), samples groups of points that are disparate and high-magnitude when represented in ...
  23. [23]
    [PDF] Coarse sample complexity bounds for active learning - UCSD CSE
    [5] S. Dasgupta, A. Kalai, and C. Monteleoni. Analysis of perceptron-based active learning. Eighteenth Annual. Conference on Learning Theory, 2005.
  24. [24]
    [PDF] The True Sample Complexity of Active Learning*
    C has VC dimension. 1 since knowing the identity of the node labeled 1 on level i is enough to determine the labels of all nodes on levels 0,...,i perfectly.
  25. [25]
    Coarse sample complexity bounds for active learning - NIPS papers
    We characterize the sample complexity of active learning problems in terms of a parameter which takes into account the distribution over the input space, the ...Missing: generalization | Show results with:generalization
  26. [26]
    [PDF] A Bound on the Label Complexity of Agnostic Active Learning
    We study the label complexity of pool-based active learning in the agnostic PAC model. Specifically, we derive general bounds on the number of label ...
  27. [27]
    [PDF] Theory of Active Learning
    Sep 22, 2014 · The objective in active learning is to produce a highly-accurate classifier, ideally using fewer labels than the number of random labeled data ...
  28. [28]
    [PDF] A tutorial on active learning
    The label complexity of CAL (mellow, separable) active learning can be captured by the the VC dimension d of the hypothesis and by a parameter θ called the ...
  29. [29]
    [2409.09078] Bounds on the Generalization Error in Active Learning
    Sep 10, 2024 · This paper establishes upper bounds on generalization error in active learning, suggesting combining informativeness and representativeness ...Missing: deep | Show results with:deep
  30. [30]
    [PDF] Deep Active Learning by Leveraging Training Dynamics
    Then, we demonstrate that higher alignment usually comes with a faster convergence speed and a lower generalization bound. Furthermore, with the help of the ...
  31. [31]
    [PDF] rates of convergence in active learning by steve hanneke
    We study the rates of convergence in generalization error achievable by active learning under various types of label noise. Additionally, we study the.
  32. [32]
    Convergence Rates of Active Learning for Maximum Likelihood ...
    Jun 8, 2015 · An active learner is given a class of models, a large set of unlabeled examples, and the ability to interactively query labels of a subset of ...
  33. [33]
    [PDF] Convergence of Uncertainty Sampling for Active Learning
    They were first proposed by Lewis &. Gale (1994) who experimentally show that a probabilistic model with uncertainty sampling can improve the perfor- mance ...<|separator|>
  34. [34]
    [PDF] Interactive Algorithms: Pool, Stream and Precognitive Stream
    In this work, our goal is to study the relationship between these two important settings: the pool-based setting and the stream-based setting.
  35. [35]
    AI-Assisted Data Labeling Using Active Learning Loops
    May 2, 2025 · You can loop forever, but most teams use one of three stop rules: (a) model meets production KPI, (b) marginal accuracy gain per loop falls ...
  36. [36]
    [PDF] A stopping criterion for Bayesian optimization by the gap of expected ...
    The proposed stopping criterion for Bayesian optimization is based on the difference between the expected minimum of simple regrets before and after evaluating ...
  37. [37]
  38. [38]
    Stream-based active learning with linear models - ScienceDirect.com
    Oct 27, 2022 · In this work, we propose a new strategy for the stream-based scenario, where instances are sequentially offered to the learner.Missing: machine seminal
  39. [39]
    [PDF] Pool-Based Active Learning for Text Classification - Kamal Nigam
    McCallum, S. Thrun, and. T. Mitchell. Learning to classify text from labeled and unlabeled documents. In AAAI-98, 1998.
  40. [40]
    Active Learning Performance in Labeling Radiology Images Is 90 ...
    Nov 30, 2021 · The AI methodology of active learning (AL) can assist human labelers by continuously sorting the unlabeled images in order of information gain.Missing: motivation | Show results with:motivation
  41. [41]
    [2503.11062] Active Learning from Scene Embeddings for End-to ...
    Mar 14, 2025 · We propose an active learning framework that relies on these vectorized scene-level features, called SEAD.
  42. [42]
    Episodic Active Reinforcement Learning from Demonstrations - arXiv
    Jun 5, 2024 · By querying episodic demonstrations as opposed to isolated state-action pairs, EARLY improves the human teaching experience and achieves better ...
  43. [43]
    A Practical Guide to Active Learning for Computer Vision | Encord
    Feb 1, 2023 · Discover Encord's in-depth guide to active learning in computer vision, optimizing your ML model's performance with strategic annotation ...
  44. [44]
    The Practitioner Guide to Active Learning in Machine Learning
    Active learning with Lightly (Building a data flywheel). Active learning works best when paired with smart data curation and strong model foundations. At ...How Does Active Learning... · Active Learning in Practice...
  45. [45]
    [PDF] Active Learning with Amazon Mechanical Turk - ACL Anthology
    Jul 27, 2011 · Two approaches that reduce the cost of annotation are active learning and crowd- sourcing. However, these two approaches have not been combined ...
  46. [46]
    Amazon SageMaker Ground Truth – Build Highly Accurate Datasets ...
    Nov 28, 2018 · Automated data labeling incurs Amazon SageMaker training and inference costs, but it can help to reduce the cost (up to 70%) and time that it ...
  47. [47]
    A Survey on Active Learning: State-of-the-Art, Practical Challenges ...
    In this paper, we provide a survey of recent studies on active learning in the context of classification.Missing: seminal | Show results with:seminal<|control11|><|separator|>
  48. [48]
    [PDF] Practical Obstacles to Deploying Active Learning - ACL Anthology
    1 Intuitively, by selecting training data cleverly, an active learner might achieve greater predictive performance than it would by choosing examples at random.
  49. [49]
    Active Learning Under Malicious Mislabeling and Poisoning Attacks
    Jan 1, 2021 · Our experimental results demonstrate that the proposed active learning method is efficient for defending against malicious mislabeling and data poisoning ...
  50. [50]
    To Address Cold Start Problem in Vision Active Learning - arXiv
    Oct 5, 2022 · This paper seeks to address the cold start problem by exploiting the three advantages of contrastive learning: (1) no annotation is required; (2) label ...
  51. [51]
    The non i.i.d. problem resulted from active learning query strategy
    Nov 12, 2020 · In such a case, you shouldn't expect your ML-model to perform well, especially since you are measuring performance on a test dataset that has ...Statistical learning when observations are not iid - Cross ValidatedOn the importance of the i.i.d. assumption in statistical learningMore results from stats.stackexchange.comMissing: violations | Show results with:violations
  52. [52]
    Active learning for data streams: a survey | Machine Learning
    Nov 20, 2023 · Online active learning is a paradigm in machine learning that aims to select the most informative data points to label from a data stream.
  53. [53]
  54. [54]
    Active learning and weak supervision - Snorkel AI
    Active learning and weak supervision. In Snorkel Flow, programmatic labeling combines active learning and weak supervision.
  55. [55]
    Outlier detection by active learning - ACM Digital Library
    This paper presents a novel outlier detection approach using classification and active learning, addressing issues with density estimation methods.
  56. [56]
    8 AI and machine learning trends to watch in 2025 | TechTarget
    Jan 3, 2025 · Explore the key trends shaping AI in 2025, from multimodal models and AI agents to security challenges and evolving regulatory landscapes.Missing: AutoML | Show results with:AutoML