Fact-checked by Grok 2 weeks ago

Blob detection

Blob detection is a fundamental technique in for identifying connected regions, or "blobs," in digital images that differ markedly from their surrounding areas in properties such as , color, or . These regions represent , localized features that are to certain image transformations, making them essential for tasks requiring robust feature extraction. In practice, blob detection often operates within a scale-space framework, where images are analyzed at multiple scales to detect blobs of varying sizes without prior knowledge of their dimensions. A key method is the Laplacian of Gaussian (LoG), which applies a scale-normalized Laplacian to convolve the image, identifying blobs as local maxima or minima in the response, with the scale corresponding to the blob's characteristic size (typically σ ≈ r/√2, where r is the radius). For computational efficiency, the Difference of Gaussians (DoG) approximates LoG by subtracting Gaussian-blurred versions of the image at adjacent scales, enabling detection of scale-invariant keypoints. Prominent algorithms incorporating blob detection include the , which uses DoG extrema to locate blobs and generates descriptors for matching, achieving high repeatability across viewpoint changes, illumination variations, and affine transformations. Blob detection finds widespread applications in image matching for stitching and , object recognition in cluttered scenes, motion tracking, robot navigation, and for anomaly detection. Early developments trace back to scale-space theory in the 1980s, with significant advancements in automatic scale selection formalized in the late .

Fundamentals

Definition and Principles

Blob detection is a fundamental technique in used to identify regions in digital images, known as blobs, which are locally connected areas exhibiting similar properties such as intensity, brightness, or color that distinguish them from the surrounding background. These blobs represent coherent structures, often corresponding to objects or features of interest, and are typically defined as regions where values are approximately constant or vary within a narrow range, enabling the isolation of salient image elements from noise or irrelevant details. The core principles of blob detection revolve around multi-scale analysis, which addresses the challenge of detecting blobs of varying sizes by examining the image at multiple resolutions, ensuring robustness to scale differences inherent in real-world scenes. This approach incorporates invariances to transformations such as , rotation, and affine changes, allowing detected blobs to remain consistent under geometric distortions common in . Prerequisites for effective blob detection include Gaussian filtering to smooth the image and suppress fine-scale noise while preserving blob structures, and scale-space theory, which provides a mathematical for representing images across a of scales through with Gaussian kernels of increasing widths. Historically, the foundations of blob detection trace back to early research in the 1980s, particularly David Marr's seminal work on the primal sketch, which introduced blobs as basic "place tokens" for capturing low-level image features like intensity changes and geometric structures at different scales. Marr's framework emphasized representing visible surface organizations through these primitives, laying the groundwork for subsequent multi-scale methods. The basic mathematical foundation treats blobs as local maxima or minima in representations, where significant structures persist across scales, enabling their detection without prior knowledge of size or shape.

Applications in Computer Vision

Blob detection plays a pivotal role in and tracking within systems, where it identifies and follows moving regions of interest in video streams to monitor activities such as movement or . In , it facilitates the localization of anomalies like tumors or lesions in modalities such as and MRI, enabling automated by isolating bright or dark regions indicative of pathological structures. For astronomy, blob detection aids in identifying celestial bodies, including and galaxies, by detecting localized intensity peaks in images, which supports cataloging and population studies of stellar distributions. In industrial inspection, it detects defects such as surface irregularities or contaminants on manufactured products, allowing for in real-time assembly lines through shape and size analysis of anomalous blobs. Beyond direct detection, blob detection serves as a foundational step in feature extraction for advanced tasks, including image matching, where it provides scale-invariant keypoints for aligning images from different viewpoints, and segmentation, by delineating object boundaries from background clutter. It also enhances applications by anchoring virtual overlays to stable blob features in dynamic environments, ensuring robust tracking amid motion or lighting changes. Notable examples include its integration in the (SIFT), which employs difference-of-Gaussians for blob-like keypoint detection to enable reliable across scales, and the (SURF) descriptor, which uses Hessian-based blob responses for faster feature description in resource-constrained settings. Furthermore, libraries like implement efficient blob detectors, such as SimpleBlobDetector, supporting real-time processing in video feeds for applications ranging from to interactive systems. Despite its utility, blob detection in practical applications faces challenges related to noise robustness, where environmental can produce false positives, necessitating preprocessing filters to maintain detection accuracy in low-contrast scenes. Computational efficiency is another concern for large-scale images, as multi-scale searches increase processing time, though approximations like integral images in mitigate this for real-time deployment. In salient tasks, blob-based methods have demonstrated improved performance on benchmark datasets like MSRA10K by emphasizing contrast-driven blob grouping for foreground isolation. Hessian-based approaches, in particular, offer precise localization for small blobs in , enhancing tumor boundary delineation in noisy volumes.

Scale-Space Methods

Laplacian of Gaussian

The Laplacian of Gaussian (LoG) is a multi-scale operator used in blob detection that combines Gaussian smoothing with the Laplacian to identify blob-like structures across different sizes in an image. The algorithm begins by convolving the input image I with a Gaussian kernel G_\sigma at various scales \sigma, which reduces noise while preserving scale-specific features. The Laplacian is then applied to this smoothed image, producing the LoG response that highlights regions of rapid intensity change, such as the boundaries of blobs, through zero-crossings. The key equation for the operator is given by \nabla^2 (G_\sigma * I), where \nabla^2 denotes the Laplacian, G_\sigma(x, y) = \frac{1}{2\pi\sigma^2} \exp\left(-\frac{x^2 + y^2}{2\sigma^2}\right) is the 2D parameterized by the \sigma, and * represents . This second-order operator yields positive values inside dark blobs on bright backgrounds and negative values inside bright blobs on dark backgrounds, enabling detection of both types. Blobs are identified as local maxima (or minima) of the absolute LoG response within a pyramid constructed by varying \sigma, where the scale at the extremum corresponds to the blob's characteristic size. A primary advantage of the method is its , as the multi-scale analysis automatically selects the optimal \sigma for each blob, making it robust to variations in object size without prior knowledge. Additionally, the Gaussian pre-smoothing mitigates sensitivity inherent to second-order derivatives, while the operator's isotropic nature detects circular or nearly circular blobs effectively. However, the approach has notable limitations, including high computational cost from performing convolutions at multiple scales, which scales poorly with and the number of scales. It also remains somewhat sensitive to in low-contrast regions, potentially leading to false positives if smoothing is insufficient. In practice, candidate blobs are selected by applying a to the response to weak extrema, followed by non-maxima suppression across both spatial locations and scales to retain only the most salient detections. This process ensures precise localization and scale estimation, though approximations like the are often employed to reduce computational demands without explicit Laplacian computation.

Difference of Gaussians

The (DoG) provides an efficient approximation to the Laplacian of Gaussian for multi-scale blob detection by emphasizing regions with strong local contrast across different image scales. This technique processes the input image through Gaussian blurring at two closely related scales and subtracts the results to produce a band-pass filtered response that highlights blob structures. The core algorithm convolves the input image I with Gaussian kernels at scales \sigma and k\sigma, where k is a multiplicative factor (e.g., k = 2^{1/s} with s = 3 intervals per , yielding k \approx 1.26), followed by subtraction to yield the response. This operation approximates the scale-normalized Laplacian while avoiding explicit second-order derivatives. The key for the at position (x, y) and scale \sigma is: D(x, y, \sigma) = \left( G(x, y, k\sigma) - G(x, y, \sigma) \right) * I(x, y) where G(x, y, \sigma) denotes the two-dimensional Gaussian function with standard deviation \sigma, and * represents convolution. Scale selection in DoG employs an octave pyramid structure with discrete sampling to cover a wide range of scales efficiently. Each octave spans a doubling of the scale (factor of 2), divided into s intervals (commonly s = 3), achieved by setting k = 2^{1/s}; after every octave, the image is resampled by downsampling to half size to maintain computational efficiency. Blob detection occurs by searching for local extrema in the , which serve as candidate blob centers. For each sampled point, the DoG value is compared against its 26 immediate neighbors in a $3 \times 3 spatial region across the current and two adjacent scales; points that are maxima or minima in this neighborhood are selected as scale-invariant keypoints corresponding to blobs. DoG's primary advantages include significantly faster computation than the exact Laplacian of Gaussian, as it relies on straightforward subtractions of precomputed Gaussian-blurred images, enabling its use in applications like image feature matching. However, as an approximation, it can suffer from errors at fine scales and reduced precision in detecting small blobs, exacerbated by high sensitivity to that may produce false positives. The DoG method was pioneered in the (SIFT) algorithm by David Lowe in 1999, where it underpins robust keypoint detection for tasks requiring . This approach contributes to blob detection by facilitating the identification of stable features across varying image resolutions.

Scale-Space Blobs and Grey-Level Blobs

In theory, blobs are represented as connected components within the scale-space volume, where an image is progressively blurred by convolving it with Gaussian kernels of increasing standard deviation, causing fine-scale structures to merge into larger ones as scale grows. These blobs evolve continuously across scales, with their boundaries defined by level-set contours that simplify hierarchically, enabling the capture of multi-scale image structures without predefined pyramid levels. Grey-level blobs are defined as regions in the exhibiting similar values, forming the units in this representation; as increases, these blobs merge or split based on topological changes in the grey-level , which can be modeled as a hierarchical known as the grey-level blob tree. In the blob tree, leaf nodes correspond to the smallest-scale blobs at fine resolutions, while internal nodes represent merged structures at coarser scales, with branches indicating the or of blobs during the scale evolution process. This tree explicitly encodes the nested relationships between blobs, allowing for the analysis of how smaller regions are subsumed into larger ones. Blob detection in this framework identifies centers at local maxima within the three-dimensional volume (two spatial dimensions plus scale), where the position and scale of a maximum indicate the blob's location and size. Stability is assessed by the depth of a blob in the , which measures the scale range over which it remains a distinct component before merging, providing a measure of salience against or irrelevant details. Key concepts include blob depth, defined as the maximum scale difference between a blob's birth and death in the tree, and its lifetime, which quantifies across scales and aids in selecting robust features. Nested blobs are handled naturally through the hierarchical , where child blobs represent substructures within parent blobs, facilitating the detection of multi-resolution patterns like concentric intensity variations. This approach offers advantages in handling multi-scale analysis implicitly through continuous blurring rather than discrete pyramids, and it provides robustness to by prioritizing deep, stable blobs that survive larger scales. However, constructing the full is computationally intensive, requiring extensive processing of the volume to track all topological events. For segmentation tasks, grey-level blob structures can be briefly integrated with algorithms to delineate boundaries based on tree-derived regions.

Hessian-Based Methods

Determinant of the Hessian

The Determinant of the Hessian () is a second-order approach for detecting isotropic blob-like structures in images by analyzing the local curvature through the of second-order derivatives. This method identifies regions where the image intensity exhibits significant second-order variation characteristic of compact, symmetric blobs, such as circular or nearly circular features. The H at a point in a Gaussian-smoothed L is defined as H = \begin{pmatrix} L_{xx} & L_{xy} \\ L_{xy} & L_{yy} \end{pmatrix}, where L_{xx}, L_{xy}, and L_{yy} represent the second-order partial derivatives with respect to the spatial coordinates x and y. candidates are then selected as local maxima of the \det(H) = L_{xx} L_{yy} - L_{xy}^2, computed across multiple scales to capture of varying sizes. Positive determinants indicate regions of suitable for bright or dark , depending on the sign of the . In scale-space implementation, the derivatives are approximated using Gaussian kernels at different standard deviations \sigma, enabling multi-scale detection where the is evaluated at progressively coarser resolutions. This ensures for isotropic structures by linking blob responses across scales via the representation. DoH offers rotational invariance for circular blobs, as the remains unchanged under orthogonal transformations, and provides precise sub-pixel localization through at detected maxima. However, it is sensitive to anisotropic shapes, where elongated structures produce weaker responses due to differing principal curvatures, often necessitating eigenvalue analysis of H to assess ellipticity by comparing the magnitudes of the eigenvalues \lambda_1 and \lambda_2 (e.g., requiring |\lambda_1| \approx |\lambda_2| for true blobs). For efficient computation, the FAST-Hessian detector approximates the multi-scale using integral images and box filters to speed up derivative calculations, reducing complexity while maintaining detection accuracy for applications. It can also be hybridized with the Laplacian for refined selection in certain implementations.

Hessian-Laplace Detector

The Hessian-Laplace detector integrates the of the , \det(H), to quantify blob strength with the Laplacian, defined as the of the , \trace(H), for precise selection. This approach builds on the strengths of both operators: \det(H) identifies regions of changes indicative of blob boundaries, while \trace(H) captures the overall variation to pinpoint the characteristic where the blob is most prominent. By decoupling blob strength from scale estimation, the enhances detection reliability in varying conditions. In the detection process, multi-scale representations of the image are constructed by convolving with Gaussian derivatives at different scales \sigma. The Hessian and Laplacian responses are computed and normalized to account for scale dependencies, typically as \sigma^4 \det(H) for the Hessian measure and \sigma^2 |\trace(H)| for the Laplacian. Interest points are selected as local maxima in both the spatial domain (for \sigma^4 \det(H)) and in (for \sigma^2 |\trace(H)|), ensuring stable blob centers. Introduced by Mikolajczyk and Schmid, the detector refines initial candidates through sub-pixel for accuracy. Compared to the pure of (DoH) approach, which relies solely on \det(H) for both strength and scale, the Hessian-Laplace method offers superior scale estimation by leveraging the Laplacian's sensitivity to blob interiors, thereby reducing false positives from edge-like structures or noise. Evaluations on the Affine dataset demonstrate its effectiveness, achieving up to 68% under scale factors of 1.4 and moderate viewpoint changes, outperforming DoG and LoG in matching accuracy across 160 real images with affine deformations. However, the detector's computational demands are notable, as it requires evaluating second-order across multiple scales and spatial locations, often necessitating efficient approximations for applications. Additionally, it performs best on near-circular blobs, with degrading for highly elliptical shapes due to its isotropic assumptions. These limitations highlight its suitability for controlled scenarios like texture analysis rather than arbitrary deformations.

Affine-Adapted Blob Detectors

Affine-adapted blob detectors build upon isotropic methods like the Hessian-Laplace detector by incorporating an iterative shape process to achieve invariance to affine transformations, which can deform blobs into ellipses due to viewpoint changes or perspective distortions. This adaptation estimates local affine parameters to warp anisotropic regions into circular shapes, enhancing repeatability in matching tasks across images subjected to linear distortions. The core process begins with detecting initial candidate points using an isotropic blob detector, such as the Hessian-Laplace, which identifies scale-invariant extrema based on the determinant of the . Affine parameters are then estimated iteratively from the of the , which capture the principal curvatures of the local intensity surface; these inform the second moment matrix \mu of the neighborhood, yielding the affine shape matrix A that parameterizes the linear transformation. The region is warped using A to normalize it to , with independent adjustment of and scales along the principal axes; this is repeated until the eigenvalues of the normalized are nearly equal ( measure Q > 0.95), typically converging in 4–10 iterations. A key advantage of these detectors is their robustness to affine changes, such as viewpoint rotations up to 70° or factors of 1.4–4.5, enabling reliable matching in wide-baseline stereo or scenarios. For example, the Hessian-Affine detector achieves rates of up to 68% under combined and affine distortions, outperforming non-adapted methods on sequences. Similarly, Affine-SIFT extends the SIFT descriptor by simulating affine warps during keypoint detection, further improving invariance for blob-like features in texture recognition. Evaluations on affine-covariant datasets, like the Oxford Affine Regions dataset, demonstrate their superior performance in metrics under real-world transformations. Despite these benefits, affine-adapted detectors face limitations, including convergence failures in low-contrast regions where initial eigenvalues differ greatly, discarding up to 40% of candidates. The iterative warping also increases computational demands, making them slower than isotropic counterparts for large-scale applications.

Region-Based Methods

Maximally Stable Extremal Regions (MSER) is a blob detection that identifies stable connected components in an image by analyzing thresholds, serving as a region-growing method particularly effective for detecting blobs with uniform properties. Introduced by Matas et al. in 2002, the algorithm processes the image by sorting pixels according to their values and incrementally adding them in increasing or decreasing order to form connected components using a union-find for efficiency. This thresholding approach generates extremal regions, which are contiguous sets of pixels where all interior pixels have intensities either strictly greater (for bright blobs) or less (for dark blobs) than the boundary pixels. The core of MSER lies in selecting those extremal regions that exhibit maximal across a range of thresholds, ensuring robustness to variations in intensity. is measured by the relative change in the 's area over a local range of intensity values \Delta, defined as the criterion q(\tau) = \frac{|R(\tau + \Delta) \setminus R(\tau - \Delta)|}{|R(\tau)|}, where R(\tau) denotes the at threshold \tau, and a is selected if q(\tau) reaches a local minimum below a user-defined \delta. This process, with near-linear O(n \log \log n) for an with n pixels, avoids explicit multi-scale while producing a hierarchical structure of nested regions akin to grey-level blobs. MSER offers key advantages, including invariance to affine transformations of intensities (such as monotonic changes in or ) and to affine geometric transformations, making it suitable for wide-baseline matching without requiring representations. It has been widely applied in text detection, where stable character regions are extracted robustly, and in tasks, achieving high repeatability in matching with average epipolar errors below 0.09 pixels. However, MSER is sensitive to , where even small perturbations can destabilize region boundaries and lead to erroneous detections. Additionally, in textured areas, it tends to over-detect numerous small extremal regions from background patterns, increasing false positives. To enhance utility as blob descriptors, post-processing often involves fitting ellipses to the detected regions using second-order central moments, providing compact affine-invariant representations for further matching or description.

Watershed-Based Algorithms

Watershed-based algorithms for blob detection interpret the image as a topographic surface where pixel intensities represent heights, simulating the flooding process from local minima (or maxima for inverted images) to delineate catchment basins that correspond to blobs. This principle segments the image into regions separated by ridges, analogous to watersheds in , allowing the identification of connected components of similar as potential blobs. In the context of blob detection, the algorithm floods the surface starting from intensity minima, with water levels rising until basins merge at saddle points, thereby outlining blob boundaries based on flows. Lindeberg's variant extends this to a multi-scale framework within representation, where the image is progressively smoothed using Gaussian kernels to analyze structures at varying resolutions. The approach detects grey-level by extracting local extrema across scales and linking them into hierarchical structures via extremal paths, capturing events such as blob creation, merging, or annihilation at bifurcations. Shallow basins are iteratively merged into deeper ones to form blob trees, which represent nested hierarchies and enable the detection of stable, significant by resolving scale-dependent mergers. This multi-scale operates on the gradient magnitude to define edges and employs a t for hierarchical segmentation, ensuring are localized at their most salient scales. The key steps involve computing the gradient magnitude to identify ridge lines as barriers, performing initial segmentation at fine scales to isolate small basins, and then progressively increasing the to merge adjacent basins based on their depth and . Bifurcation events are registered to track how blobs evolve, with merging criteria prioritizing deeper structures to avoid fragmentation. This process integrates with theory to estimate blob depth, using metrics like the effective scale r(t) = \log(p(t)/p_0) and normalized blob volumes, where p(t) denotes the probability density at scale t, providing a measure of a blob's persistence and significance across scales. These grey-level blob trees relate briefly to the broader blobs by organizing them hierarchically based on merging events. Advantages of watershed-based methods include their ability to handle nested blob structures inherent in natural images and robustness to grey-level variations through scale-space smoothing, which suppresses noise while preserving significant features. By focusing on regional descriptors derived from basin properties, the algorithms provide stable cues for blob localization without relying on local differential invariants. However, limitations arise from potential over-segmentation at fine scales due to noise-induced spurious minima, necessitating marker-based refinements or clipping levels (e.g., around 35% of intensity range) to constrain flooding. Additionally, parameter tuning for the scale sampling and merging threshold is required to balance detail retention and computational efficiency, as excessive smoothing can obscure shallow blobs.

Extensions and Modern Variants

Spatio-Temporal Blob Detectors

Spatio-temporal blob detectors extend traditional blob detection techniques to video sequences by treating the data as a space-time volume with coordinates (x, y, t), where t represents time. This approach applies representations or Hessian-based methods in three dimensions to identify dynamic blobs corresponding to moving objects or events, such as trajectories forming tube-like structures in the volume. By detecting local extrema in this spatio-temporal domain, these methods capture both spatial extent and temporal evolution, enabling the localization of interest points that are stable across scales. Key methods include the use of Laplacian or operators convolved with Gaussian kernels. In the framework, the video is represented as L(\mathbf{x}; s, \tau) = g(\mathbf{x}; s, \tau) * f(\mathbf{x}), where g is an anisotropic Gaussian with s and temporal \tau, and interest points are found as maxima of a spatio-temporal Harris response H = \det(\mu) - k \trace^3(\mu), with \mu being the second-moment of Gaussian . approaches, such as the dense scale-invariant detector, compute the determinant of the \det(H) = L_{xx} L_{yy} L_{tt} - L_{xt}^2 L_{yy} - L_{yt}^2 L_{xx} - L_{xy}^2 L_{tt} + 2 L_{xy} L_{xt} L_{yt} at normalized scales to identify centers, allowing efficient detection of tube-like structures via non-iterative search in 5D (x, y, t, s, \tau) . These detectors locate spatio-temporal extrema by seeking local maxima in the saliency measure, often followed by selection based on the scale-normalized Laplacian \nabla^2_{\text{norm}} L. estimation arises from the of detected along the time dimension, where the ratio of temporal to spatial scales approximates motion speed under assumptions of constant . These methods offer advantages in handling motion blur and non-rigid deformations inherent in video, providing robust features for tasks like action recognition and object tracking by capturing events such as walking or collisions as coherent spatio-temporal structures. For instance, Laptev's spatio-temporal interest points () demonstrate high repeatability in dynamic scenes, enabling classification of human actions on benchmark datasets. Applications extend to video surveillance, where detected blobs facilitate real-time event detection and identification in crowded environments. However, limitations include significantly increased computational demands due to processing, often scaling with video length and resolution, and reliance on assumptions of locally constant velocity, which can fail under acceleration or complex motions like camera shake.

Deep Learning-Based Approaches

Deep learning-based approaches to blob detection represent a paradigm shift from traditional handcrafted filters, such as or Hessian-based operators, toward end-to-end models that learn hierarchical features directly from data. These methods, primarily leveraging convolutional neural networks (CNNs) and variants like , enable pixel-wise prediction of blob regions, addressing limitations in noisy or complex scenes where classical techniques falter due to sensitivity to parameter tuning. By training on annotated datasets, these models achieve superior , particularly in biomedical imaging where blobs correspond to lesions or cells. A prominent example is the architecture, adapted for small blob detection through pixel-wise segmentation, often combined with traditional priors for enhanced performance. In one hybrid approach, generates probability maps of potential blobs, which are jointly constrained with Hessian-based convexity analysis to refine detections without post-processing, reducing over-detection and under-segmentation. Trained supervisedly on datasets like optical microscopy images with thousands of ground-truth annotations (augmented for noise and transformations), this method outperforms classical detectors like Laplacian of Gaussian in F-scores on fluorescent and MRI data, while being 35% faster. For generative alternatives, BlobDetGAN employs a CycleGAN framework for unpaired image-to-image translation, first denoising noisy inputs while preserving blob geometry, then segmenting blobs in two stages without labels. This 2022 method demonstrates effectiveness on synthetic and medical images, though it requires longer training times compared to newer contrastive models. Training typically relies on supervised learning with annotated datasets, such as those from medical lesion segmentation (e.g., breast histology images expanded via augmentation to over 27,000 samples), achieving high F1-scores like 98.82% for cancerous blob detection using recurrent neural networks integrated with morphological operations. Self-supervised variants, like NU-Net, address data scarcity by training on unlabeled bioimage collections (e.g., 12,000+ nuclear and cellular images across modalities) using perceptual losses for blob enhancement, improving downstream detection F1-scores from 0.54 to 0.72 without paired data. These approaches offer advantages in handling noisy, complex environments through transfer learning and adaptability, with recent integrations allowing CNN models to run via OpenCV's DNN module for efficient inference in real-time applications. Despite these gains, methods suffer from high data requirements for supervision and the "black-box" nature, lacking the interpretability of analytical classics, which can hinder trust in critical domains like . Recent advances from 2020-2025, including contrastive learning in BlobCUT (2023) for faster small-blob segmentation in 3D medical volumes and U-Net-based detectors in 2025 studies, underscore their growing impact in cancer detection and , with ongoing efforts toward and paradigms to mitigate limitations.

References

  1. [1]
    [PDF] Machine learning for blob detection in high-resolution ... - DiVA portal
    The aim of blob detection is to find regions in a digital image that dif- fer from their surroundings with respect to properties like intensity or shape.
  2. [2]
    [PDF] Lecture 10 Detectors and descriptors
    Feb 16, 2015 · Corner/blob detectors. Edges are useful as local features, but corners and small areas (blobs) are generally more helpful in computer vision ...
  3. [3]
    Feature Detection with Automatic Scale Selection
    Manufactured in The Netherlands. Feature Detection with Automatic Scale Selection. TONY LINDEBERG. Computational Vision and Active Perception Laboratory (CVAP), ...
  4. [4]
    [PDF] Distinctive Image Features from Scale-Invariant Keypoints
    Jan 5, 2004 · This paper presents a method for extracting distinctive invariant features from images that can be used to perform reliable matching between ...
  5. [5]
    [PDF] Lecture 4 Feature Detectors and Descriptors: Corners, Blobs and SIFT
    • We define the characteristic scale of a blob as the scale that produces ... International Journal of Computer Vision 30 (2): pp 77--116. Slide: S ...
  6. [6]
    (PDF) Scale-Space Theory in Computer Vision - ResearchGate
    Algorithms are needed for smoothing grey-level images, detecting grey-level blobs in smoothed grey-level images, registering bifurcations, linking grey-level ...
  7. [7]
    Automated blob detection using iterative Laplacian of Gaussian ...
    In image processing, blobs (a.k.a. particles [1] or dots [2]) can be defined as small structures whose visual properties, e.g. brightness or color, are ...
  8. [8]
    Scale-Space Theory in Computer Vision | SpringerLink
    Part of the The Springer International Series in Engineering and Computer Science book series (SECS, volume 256). Download book PDF.
  9. [9]
    Marr's Theory: From primal sketch to 3-D models
    Marr proposed to capture these organizations by using a set of ``place tokens'', or low level features, which correspond to oriented edges, bars, ends and blobs ...
  10. [10]
    Vision - MIT Press
    An influential book that offers a framework for understanding visual perception and considers fundamental questions about the brain and its functions.Missing: blobs | Show results with:blobs
  11. [11]
    People tracking in surveillance applications - ScienceDirect.com
    The analysis of a single blob position or trajectory can determine whether the person is standing in a forbidden area, running, jumping or hiding. Combining ...People Tracking In... · Introduction · Event Detection
  12. [12]
    Computer-Aided Tumor Detection Based on Multi-Scale Blob ...
    Dec 10, 2012 · Our results suggest that CADe systems based on multi-scale blob detection can be used to detect breast tumors in ABUS images. Published in: IEEE ...
  13. [13]
    Uncertainty-aware blob detection with an application to integrated ...
    Blob detection is a common problem in astronomy. One example is in stellar population modelling, where the distribution of stellar ages and metallicities in ...
  14. [14]
    A Generalized Laplacian of Gaussian Filter for Blob Detection and ...
    Jan 9, 2013 · The gLoG filter can not only accurately locate the blob centers but also estimate the scales, shapes, and orientations of the detected blobs.
  15. [15]
    [PDF] SURF: Speeded Up Robust Features
    SURF is a scale- and rotation-invariant interest point detector and descriptor, computed faster than previous methods, using integral images.
  16. [16]
    Blob Detection Using OpenCV ( Python, C++ ) | - LearnOpenCV
    Feb 17, 2015 · This beginner tutorial explains simple blob detection using OpenCV. C++ and Python code is available for study and practice.
  17. [17]
    Salient object detection via reciprocal function filter - IET Journals
    Jul 16, 2019 · MSRA10K, AUC, 0.9477, 0.8108, 0.8624, 0.6004 ... et al: ' Rich feature hierarchies for accurate object detection and semantic segmentation'.
  18. [18]
    Improved small blob detection in 3D images using jointly ... - Nature
    Jan 15, 2020 · In this research, we propose a joint constraint blob detector from U-Net, a deep learning model, and Hessian analysis, to overcome these problems and identify ...
  19. [19]
    Theory of edge detection | Proceedings of the Royal Society of ...
    A theory of edge detection is presented. The analysis proceeds in two parts. (1) Intensity changes, which occur in a natural image over a wide range of scales, ...
  20. [20]
    [PDF] Feature Detection with Automatic Scale Selection - DiVA portal
    See (Lindeberg 1996a) concerning scale selection mechanisms for detecting edges and ridges. Page 43. Feature detection with automatic scale selection. 41.
  21. [21]
    A Survey of Blob Detection Methods: Techniques, Evaluation, and ...
    Jun 24, 2024 · This survey provides a comprehensive overview of the most prominent blob detection methods, categorizing them into differential, region-based, ...
  22. [22]
    [PDF] Object Recognition from Local Scale-Invariant Features 1. Introduction
    An object recognition system has been developed that uses a new class of local image features. The features are invariant to image scaling, translation, ...Missing: 2004 | Show results with:2004
  23. [23]
    [PDF] Discrete Scale-Space Theory and the Scale-Space Primal Sketch
    We pro pose that the canonical way to construct a scale-space for discrete signals is by convolution with a kernel called the discrete analogue of the Gaussian.
  24. [24]
    [PDF] Detecting Salient Blob-Like Image Structures and Their Scales with ...
    The basic tools for the analysis will be scale- space theory, and a heuristic principle stating that blob-like structures which are stable in scale-space are ...
  25. [25]
    The Hessian Blob Algorithm: Precise Particle Detection in ... - Nature
    Jan 17, 2018 · The Hessian blob algorithm presented here uses two of the simplest blob detectors, the normalized Laplacian of Gaussian and determinant of the ...Missing: seminal | Show results with:seminal
  26. [26]
    [PDF] Scale & Affine Invariant Interest Point Detectors
    In this paper we give a detailed description of a scale and an affine invariant interest point detector introduced in Mikolajczyk and Schmid (2001, 2002). Our ...
  27. [27]
    Distinctive Image Features from Scale-Invariant Keypoints
    Download PDF · International Journal of ... About this article. Cite this article. Lowe, D.G. Distinctive Image Features from Scale-Invariant Keypoints.
  28. [28]
    A Comparison of Affine Region Detectors
    Oct 14, 2005 · The paper gives a snapshot of the state of the art in affine covariant region detectors, and compares their performance on a set of test images under varying ...
  29. [29]
    ASIFT: A New Framework for Fully Affine Invariant Image Comparison
    The method proposed in this paper, affine-SIFT (ASIFT), simulates all image views obtainable by varying the two camera axis orientation parameters.
  30. [30]
    [PDF] Robust Wide Baseline Stereo from Maximally Stable Extremal Regions
    The first contribution of the paper is the introduction of a new set of distinguished regions, the so called extremal regions. Extremal regions have two ...
  31. [31]
    [PDF] Gaussian Affine Feature Detector - arXiv
    Abstract—A new method is proposed to get image features' geometric information. Using Gaussian as an input signal, a theoretical optimal solution to ...Missing: seminal | Show results with:seminal<|control11|><|separator|>
  32. [32]
    [PDF] Extremal Region Selection for MSER Detection in Food Recognition
    contain too many ERs from the background. Figure 8 shows three examples from ... Based on the graph presented in Figure 10, the classification rate pattern over ...
  33. [33]
    [PDF] Space-time interest points - Computer Vision, 2003 ... - IRISA
    In this paper, we propose to extend the notion of spatial interest points into the spatio-temporal domain and show how the resulting features often reflect ...
  34. [34]
    An Efficient Dense and Scale-Invariant Spatio-Temporal Interest ...
    Aug 7, 2025 · Over the years, several spatio-temporal interest point detectors have been proposed. While some detectors can only extract a sparse set of ...
  35. [35]
  36. [36]
    Automatic Blob Detection Method for Cancerous Lesions in ... - NIH
    Mar 31, 2025 · We present a deep learning-based technique for breast cancer lesion detection, namely blob detection, which automatically detects hidden and inaccessible ...
  37. [37]
    [PDF] A Self-Supervised Smart Filter for Enhancing Blobs in Bioimages
    While supervised deep neural networks have become the dominant method for image analysis tasks in bioimages, truly versatile methods are not available yet ...
  38. [38]
    Deep learning: How OpenCV's blobFromImage works
    Nov 6, 2017 · In today's blog post we are going to take apart OpenCV's cv2.dnn.blobFromImage and cv2.dnn.blobFromImages preprocessing functions and understand how they work.Missing: integration | Show results with:integration
  39. [39]
    BlobCUT: A Contrastive Learning Method to Support Small Blob ...
    Nov 29, 2023 · Xu et al. [32] proposed BlobDetGAN, a CycleGAN-based model to solve blob detection in two steps: image denoising and object segmentation.