Fact-checked by Grok 2 weeks ago

U-Net

U-Net is a fully architecture designed for fast and precise , featuring a symmetric U-shaped structure with a contracting path that captures rich contextual information through successive downsampling and feature extraction, an expanding path that performs for localization, and skip connections that concatenate high-resolution features from the contracting path to the expanding path to preserve spatial details. Introduced in 2015 by Olaf Ronneberger, Philipp Fischer, and Thomas Brox from the , U-Net was developed to address the challenge of training effective segmentation models with limited annotated data, employing a training strategy that relies on extensive techniques such as elastic deformations, rotations, and scaling to generate diverse training samples from few images. The architecture uses repeated 3x3 convolutions with ReLU activations, max pooling for downsampling in the contracting path (doubling feature channels from 64 to 1024), transposed convolutions for in the expanding path (halving channels), and a final 1x1 convolution for pixel-wise classification, enabling end-to-end training with and weighted loss to handle class imbalance. U-Net demonstrated superior performance on the International Symposium on Biomedical Imaging (ISBI) challenges, achieving a warping error of 0.0003529 and Rand error of 0.0382 on the 2012 electron segmentation task, outperforming prior methods, and intersection over union () scores of 0.9203 for PhC-U373 and 0.7756 for DIC-HeLa in the 2015 cell tracking challenge, with segmentation times under 1 second for 512x512 images on a GPU. Its modular design, efficiency in low-data regimes, and ability to produce high-resolution outputs have made it a of medical image analysis, widely adopted for segmenting structures in modalities like CT scans, MRI, X-rays, and , and extended to non-biomedical applications such as analysis and general tasks.

Introduction

Overview

U-Net is a () architecture designed as an encoder-decoder model with skip connections, specifically tailored for semantic segmentation tasks. It enables pixel-wise classification of images, assigning a class label to each pixel to delineate objects with high precision, which is particularly valuable in applications requiring accurate boundary detection, such as biomedical imaging. The architecture features a characteristic U-shaped structure formed by a symmetric contracting path (encoder) and expanding path (). The contracting path captures rich contextual information through successive downsampling, while the expanding path performs to recover and enable precise localization of features. Skip connections between corresponding levels of these paths concatenate high-resolution features from the encoder to the , preserving fine details that might otherwise be lost in the encoding process. A key strength of U-Net lies in its effectiveness with limited training data, achieved through extensive techniques, including elastic deformations that simulate plausible variations in biological structures. In practice, it processes an input image—such as a scan—and produces a segmented output , where each is classified to highlight specific regions like cells or neuronal structures. This design has made U-Net a foundational model for , influencing numerous extensions in and .

Motivation and Design Principles

Traditional convolutional neural networks (CNNs) for faced significant challenges, particularly in preserving spatial information and handling limited training data. In fully convolutional networks, repeated downsampling through pooling layers leads to a loss of fine-grained spatial details, making precise pixel-wise localization difficult, while fully connected layers in classification-oriented CNNs exacerbate on small datasets typical in biomedical imaging. These issues were especially pronounced in biomedical applications, where segmenting intricate structures like cells in images requires high accuracy but annotated data is scarce—often limited to just tens of images due to the labor-intensive nature of labeling. U-Net was specifically developed to address these limitations in the context of the Symposium on Biomedical Imaging (ISBI) challenges, originating from the 2012 electron microscopy () segmentation competition and culminating in its application to the 2015 tracking . The architecture was motivated by the need to outperform prior sliding-window approaches, which processed images patch-by-patch and suffered from computational inefficiency and redundant computations due to overlaps. By enabling end-to-end training on very few images—such as 30 for stacks or 35 for of human cells—U-Net demonstrated superior performance, winning the ISBI 2015 through effective techniques like elastic deformations. The design principles of U-Net emphasize a balance between capturing rich contextual information and enabling precise localization, achieved through a U-shaped structure that combines a contracting path for downsampling to encode context and an expansive path for to recover . Skip connections from the contracting to the expansive path propagate high-resolution features, mitigating the loss of spatial information from pooling. As a fully convolutional network without dense layers, U-Net supports variable input sizes and efficient processing of arbitrarily large images via an overlapping-tile strategy, where context is extrapolated for boundary tiles using mirroring to ensure seamless segmentation. This approach allows the network to generalize well despite data scarcity, prioritizing both speed and accuracy in biomedical tasks.

Architecture

Contracting Path

The contracting path, also known as the encoder, forms the initial phase of the U-Net architecture, designed to capture multi-scale contextual by progressively reducing spatial dimensions while expanding the depth of representations. It comprises a series of repeating blocks, each consisting of two unpadded 3×3 convolutions followed by rectified linear unit (ReLU) activations, and concludes with a 2×2 max pooling operation with stride 2 for downsampling. This structure enables the network to extract hierarchical , starting from low-level edges and textures to higher-level semantic patterns. In the original U-Net, the contracting path spans four levels, with the number of feature channels doubling at each downsampling step to increase representational capacity: beginning with 64 channels after the first block, then 128, 256, and 512 channels at the subsequent levels, culminating in a layer at 1024 channels before transitioning to the expansive path. Each , being unpadded, slightly reduces the spatial dimensions within a block (e.g., from 572×572 to 568×568 at the input level for a typical 572×572 ), further emphasizing the focus on contextual aggregation over precise localization. The primary role of the contracting path is to provide high-level semantic essential for segmentation tasks, particularly in biomedical imaging where global scene understanding aids in delineating structures. By halving the at each level via max pooling, the path reduces computational load while the grows exponentially through the stacking of convolutions and the effects of pooling strides, allowing deeper layers to perceive the entire input image—critical for capturing the context of small objects within large fields of view. Mathematically, if the input spatial dimensions are H \times W, the resolution after l levels is H_l = H / 2^l and W_l = W / 2^l.

Expansive Path

The expansive path in the U-Net architecture serves as the , progressively restoring the of maps while refining the segmentation output. It begins with an operation on the feature maps from the bottleneck layer, followed by a 2×2 transposed (also known as up-convolution) that halves the number of feature channels at each level to mirror the contracting path's . This is then succeeded by two convolutions, each activated by a ReLU nonlinearity, which process the upsampled features to capture contextual details at increasing resolutions. At the final stage of the expansive path, a 1×1 maps the resulting 64-channel feature vectors to the desired number of output classes, producing the segmentation map; for segmentation tasks, this output is typically passed through a activation to yield probability maps between 0 and 1. The path's design facilitates precise localization by integrating low-resolution, semantically rich features from deeper layers with higher-resolution details, enabling the network to generate detailed boundary predictions essential for tasks like biomedical . To accommodate large input images that exceed memory constraints or the network's fixed , U-Net employs an overlapping-tile strategy in the expansive path. The image is divided into overlapping tiles of sufficient size to ensure even dimensions compatible with max-pooling operations in the contracting path; each tile is processed independently through the full network, and the resulting predictions are averaged across overlap regions. Boundary artifacts are mitigated by mirroring the tile content to extrapolate missing context, yielding a seamless, full-resolution segmentation map.

Skip Connections and Feature Fusion

Skip connections in U-Net refer to direct links that connect the feature maps from corresponding layers in the contracting path to the expansive path, bypassing the at the network's base. These connections enable the transfer of high-resolution features from earlier layers, which would otherwise be lost due to repeated downsampling operations. The fusion of features via skip connections occurs through of the upsampled feature maps from the expansive path with the corresponding skip connections from the contracting path, effectively doubling the number of channels. This concatenated output is then processed by convolutional layers to reduce the channel dimensionality and refine the features. In the original U-Net design, the skip connections are cropped to align with the upsampled maps, accounting for any border artifacts introduced during convolutions in the contracting path. A key benefit of these skip connections is their ability to counteract the information loss from downsampling by preserving and reintegrating fine-grained, spatially precise details into the coarser, context-rich representations of the expansive path. This mechanism enhances the network's localization accuracy, allowing precise boundary delineation in segmentation tasks. Mathematically, the feature map at decoder layer i can be represented as: \text{decoder feature}_i = \text{Conv}\left( \text{Concat}\left( \text{upsampled}_{i+1}, \text{skip}_i \right) \right) where \text{Concat} denotes channel-wise concatenation, \text{upsampled}_{i+1} is the upsampled feature from the previous decoder layer, \text{skip}_i is the feature map from the corresponding encoder layer, and \text{Conv} applies the convolutional processing. In the original implementation, skip connections are established at each resolution level between the contracting and expansive paths, facilitating end-to-end training of the fully convolutional network even with limited annotated data.

Training and Implementation

Data Preparation and Augmentation

In U-Net training, particularly for biomedical image segmentation, input data preparation begins with or multi-channel images, which are normalized to the range [0,1] to ensure stable gradient flow and consistent across diverse imaging modalities. Given the high resolution of typical biomedical scans and hardware memory limitations, images are processed using a tile-based strategy, dividing them into smaller, manageable patches—often selected to have even dimensions compatible with max-pooling operations—while maximizing GPU utilization by training on single-image batches rather than larger ensembles. To mitigate the scarcity of annotated biomedical data, plays a pivotal role in enhancing model generalization. Rigid transformations, including rotations, scaling, and translations, are applied to simulate positional variations in setups. deformations further introduce realistic distortions by generating random displacement vectors on a coarse grid, sampled from a Gaussian distribution with a standard deviation of 10 pixels, and smoothly interpolating these to per-pixel displacements using . Intensity-based augmentations address inconsistencies in image acquisition, such as differing staining protocols or scanner artifacts; these include gamma correction to adjust brightness non-linearly and Gaussian blurring to mimic resolution variations. Corresponding ground-truth labels consist of binary or multi-class segmentation masks, aligned pixel-wise with the input tiles to facilitate precise supervised learning of object boundaries. For inference on large images, an overlapping-tile strategy reconstructs seamless segmentations by processing tiles with substantial overlap—commonly 50% between adjacent patches—to provide full contextual information at boundaries, with missing regions extrapolated via mirroring and predictions blended across overlaps to eliminate artifacts.

Loss Functions and Optimization

In the original U-Net formulation, the loss function is defined as a pixel-wise softmax over the final feature map of the expansive path, combined with a loss to measure segmentation accuracy. This approach treats the segmentation task as a multi-class problem at each , where the softmax produces probability distributions over classes, and penalizes deviations from the labels. To address class imbalance common in biomedical imaging—such as sparse foreground objects against dominant backgrounds—the loss is weighted, assigning higher penalties to misclassified pixels near boundaries between touching instances; specifically, a weight map w(x) is applied as w(x) = w_c(x) + w_0 \cdot \exp\left( -\frac{(d_1(x) + d_2(x))^2}{2\sigma^2} \right), with w_0 = 10 and \sigma \approx 5 pixels, where d_1(x) and d_2(x) are distances to nearest boundaries. The coefficient loss has become a widely adopted alternative or complement to in U-Net training, particularly for its robustness to class imbalance in segmentation tasks. Defined as the negative of the Dice similarity coefficient, it is given by \mathcal{L}_{\text{Dice}} = 1 - \frac{2 |X \cap Y|}{|X| + |Y|}, where X and Y represent the predicted and segmentation masks, respectively, and | \cdot | denotes the (or sum of pixel values for soft predictions). This formulation directly optimizes spatial overlap, making it effective for imbalanced datasets where foreground regions are small. To leverage the strengths of both, a combined is frequently employed, typically as a weighted \mathcal{L} = \alpha \mathcal{L}_{\text{[CE](/page/CE)}} + (1 - \alpha) \mathcal{L}_{\text{[Dice](/page/D.I.C.E.)}}, with \alpha tuned (often around 0.5) to balance accuracy and overlap optimization while mitigating imbalance effects like foreground-background disparity. Such losses improve and generalization in U-Net models for segmentation. For optimization, U-Net is trained using stochastic gradient descent (SGD) with a high momentum of 0.99, processing one image per batch to accommodate limited annotated data. Learning rate scheduling, such as exponential decay or step-wise reduction, is commonly integrated to stabilize training and prevent overshooting minima, often starting from an initial rate of $10^{-4} to $10^{-2} and adjusting based on validation performance. Segmentation performance is evaluated using metrics tailored to overlap and per-class accuracy, including the over (), defined as \frac{|X \cap Y|}{|X \cup Y|}, which quantifies regional agreement, as well as (true positives over predicted positives) and (true positives over actual positives) to assess boundary delineation in imbalanced scenarios. These metrics highlight U-Net's efficacy, with reported values exceeding 90% on cell tracking benchmarks in the original implementation.

Variants and Extensions

Dimensional Extensions

The 3D U-Net extends the original 2D architecture to handle volumetric data by replacing all two-dimensional convolutional operations with their three-dimensional counterparts, utilizing 3×3×3 kernels for convolutions, 2×2×2 max pooling for downsampling, and 3D transposed convolutions for in the expansive path. This adaptation enables dense segmentation of sparse 3D annotations, making it suitable for processing volumes such as CT and MRI scans. A primary challenge in implementing U-Net is the substantial increase in and computational requirements compared to the version, as the volumetric processing approximately triples the number of parameters per convolutional layer and significantly increases the computational operations due to the additional spatial and larger feature maps. To address constraints, particularly for high-resolution volumes that exceed GPU limits, patch-based training is employed, where overlapping sub-volumes are extracted for training and inference via sliding windows to reconstruct the full output. Additionally, the architecture accommodates anisotropic voxels—common in modalities like where slice thickness differs from in-plane resolution—through input resampling or adaptive kernel strides to preserve spatial coherence across dimensions. Key differences from the U-Net include not only the dimensional shift in operations but also the need for data augmentation techniques, such as elastic deformations applied volumetrically, to generate diverse training samples from limited annotations. The contracting path mirrors the design with successive convolutions and pooling to capture hierarchical features, while skip connections fuse multi-scale feature maps to recover fine-grained details in the expansive path. Özgün Çiçek et al. introduced this extension in 2016 specifically for volumetric segmentation tasks, demonstrating its efficacy on electron microscopy data for segmenting kidney glomeruli in embryos, where it achieved precise boundary delineation from sparse labels. In performance evaluations, U-Net has shown superiority over slice-by-slice approaches in tasks requiring volumetric context, such as neuron reconstruction from light-sheet , by maintaining connectivity and improving overlap metrics like Dice coefficient. However, these gains come at the cost of higher GPU demands, often requiring multi-GPU setups or efficient strategies for large volumes, with times extending several times longer than equivalents on similar .

Architectural Modifications

One prominent architectural modification to the original U-Net is U-Net++, which introduces nested dense skip pathways to facilitate multi-scale feature aggregation. By densely connecting nodes across multiple levels, this variant reduces the semantic gap between low-level and high-level features, enabling more precise delineation in segmentation tasks. Additionally, U-Net++ incorporates deep supervision at various scales, allowing intermediate outputs to contribute to the final prediction and improving training stability. The architecture also supports of redundant connections post-training to enhance computational efficiency without significant performance loss. Another key variant is Attention U-Net, which augments the skip connections with gated blocks to selectively focus on relevant features while suppressing irrelevant background regions. These attention gates use additive mechanisms to weigh feature maps from the encoder before concatenation in the , thereby enhancing the model's ability to capture target structures of varying sizes and shapes, particularly in sparse or noisy medical images. This modification proves especially beneficial for organ segmentation, where irrelevant anatomical structures can otherwise dilute focus. Further modifications include the integration of residual blocks into the U-Net framework, as seen in Residual U-Net variants. These incorporate ResNet-style residual units within the encoder and decoder paths to mitigate vanishing gradients and enable the of deeper networks, leading to richer feature representations. For instance, the Recurrent Residual U-Net combines recurrent convolutions with residual connections to better handle sequential dependencies in data, improving overall segmentation accuracy. Similarly, V-Net employs residual units in its convolutional blocks to stabilize in volumetric contexts, allowing for effective of gradients through the network. These architectural changes collectively address limitations in the original U-Net, such as semantic inconsistencies in skip connections and challenges in gradient flow during deep training. By reducing these gaps, they enhance feature fusion and model generalization. For example, U-Net++ and U-Net have demonstrated improved scores on benchmarks like cell nucleus segmentation, with gains of up to 3-5% over the baseline U-Net, highlighting their impact on precise boundary detection in biomedical applications. More recent extensions, as of 2025, incorporate transformer mechanisms and state-space models for enhanced efficiency and accuracy, such as TransUNet for hybrid CNN-transformer segmentation and PUNet integrating Mamba for lightweight processing.

Applications

Biomedical Imaging

U-Net was originally developed for biomedical image segmentation, with its inaugural application in segmenting neuronal structures from electron microscopy (EM) images as part of the 2012 International Symposium on Biomedical Imaging (ISBI) challenge. In this context, the model achieved first place by outperforming previous methods, attaining a warping error of 0.0003529 on the provided test data, demonstrating its efficacy in precise boundary delineation for densely packed cells. This success highlighted U-Net's ability to handle the limited training data typical in biomedical datasets through extensive data augmentation techniques. It also excelled in the 2015 ISBI cell tracking challenge using phase contrast and DIC microscopy. In organ segmentation tasks, U-Net has been widely adopted for delineating structures like the and in and scans. For liver segmentation, variants of U-Net have reported similarity coefficients () exceeding 0.96 on datasets such as the Combined Healthy Abdominal Organ Segmentation () challenge, enabling accurate volumetric analysis for surgical planning. In liver tumor segmentation on the benchmark, U-Net-based models achieve mean scores around 0.70 for tumors while maintaining high performance on the surrounding liver tissue ( >0.95), facilitating early detection and treatment monitoring. Similarly, for kidney segmentation in images, multi-scale supervised U-Net architectures yield values of approximately 0.97, supporting applications in renal disease assessment and tumor localization. Nuclei detection in slides represents another key application, where U-Net excels at segmenting overlapping instances in hematoxylin and (H&E)-stained images. Dense U-Net variants address clustering and touching boundaries, achieving instance segmentation precision over 0.80 on datasets like the Multi-Organ Nuclei Segmentation challenge, which aids in cancer grading and assessment. By leveraging skip connections to preserve fine-grained details, these models mitigate challenges posed by variable and cellular density. U-Net's advantages in biomedical imaging stem from its robustness to noise, artifacts, and inter-patient variability inherent in medical scans, attributes enhanced by elastic deformations and intensity shifts during training. This resilience has contributed to the integration of U-Net-inspired architectures in FDA-cleared AI tools for diagnostic imaging, such as automated segmentation modules in commercial radiology software that support clinical workflows in hospitals. Notable case studies include retinal vessel segmentation in fundus images, where improved U-Net models attain accuracy rates above 0.95 and scores of 0.82, improving screening. In delineation using the Brain Tumor Segmentation (BraTS) dataset, U-Net variants segment enhancing tumor cores with values around 0.85 on multimodal MRI, enabling precise radiotherapy planning and survival prediction. As of 2025, U-Net variants continue to evolve, integrating into foundation models for multimodal biomedical segmentation tasks.

General Computer Vision

U-Net has found extensive application in semantic segmentation tasks within general computer vision, particularly for urban scene understanding on the Cityscapes dataset, where it enables precise labeling of elements such as roads, pedestrians, and vehicles critical for autonomous systems. Modified variants of U-Net, incorporating encoders like VGG16, have demonstrated superior performance on this dataset compared to ResNet50-based alternatives, achieving mean intersection over union (mIoU) scores that highlight its effectiveness in handling complex urban environments with diverse object scales and occlusions. These adaptations leverage U-Net's encoder-decoder structure to capture multi-scale features, ensuring robust pixel-wise classification in real-world driving scenarios. In instance segmentation, U-Net is often combined with detection frameworks like Mask R-CNN to refine object boundaries and achieve detailed per-instance masks, as evaluated on datasets such as COCO that provide rich annotations for and segmentation. Experimental analyses show that while U-Net excels in semantic segmentation, integrating it with Mask R-CNN—pre-trained on COCO—enhances overall performance in distinguishing individual instances, particularly in dynamic scenes with overlapping objects, yielding comparable F1 scores across both approaches. This combination exploits U-Net's precise boundary delineation alongside Mask R-CNN's instance detection capabilities, making it suitable for tasks requiring both class-agnostic and instance-specific outputs. Beyond urban and object-centric tasks, U-Net extends to analysis for , where it segments diverse terrain types like forests, areas, and bodies from overhead images. A modified U-Net model applied to data has improved accuracy in and by incorporating residual connections, outperforming baseline convolutional networks in pixel-level mapping. Similarly, temporal extensions of U-Net, such as 3D U-Net architectures, facilitate video segmentation by processing spatio-temporal volumes of , enabling motion component separation in general video sequences for applications like object tracking. Adaptations of U-Net for processing have been developed for and autonomous driving, including lightweight variants optimized for lane detection to support on-board inference with minimal latency. For instance, attention-augmented U-Net models achieve high accuracy in delineating lane markings under varying lighting and weather conditions, enabling reliable path planning in self-driving vehicles. These implementations often reduce model complexity while preserving U-Net's skip connections for sharp boundary preservation. Overall, U-Net variants have attained competitive state-of-the-art results on benchmarks like ADE20K and have been incorporated alongside frameworks like Detectron2 in segmentation pipelines for scalable deployment.

History and Impact

Original Development

U-Net was developed by Olaf Ronneberger, Philipp Fischer, and Thomas Brox at the Department and BIOSS Centre for Biological Signalling Studies, , . The architecture was introduced in their 2015 paper titled "U-Net: Convolutional Networks for Biomedical Image Segmentation," initially submitted to and later published in the proceedings of the Medical Image Computing and Computer-Assisted Intervention (MICCAI) conference. The primary motivation for U-Net arose in response to the segmentation challenges posed by the International Symposium on Biomedical Imaging (ISBI) 2015 Cell Tracking Challenge, which focused on tracking cells in microscopy images with limited annotated training data. The authors designed U-Net to enable precise biomedical image segmentation using fully convolutional networks, emphasizing efficiency for small datasets typical in microscopy applications. In initial experiments, the network was trained on 35 partially annotated transmitted light microscopy images for phase contrast (PhC-U373) and 20 for differential interference contrast (DIC-HeLa), leveraging extensive data augmentation such as elastic deformations, rotations, and scaling to generate diverse training samples despite the scarcity of annotations. These experiments demonstrated U-Net's effectiveness, achieving an intersection over union () score of 92% on the phase contrast cell segmentation task from the ISBI challenge, significantly outperforming prior methods. The model also succeeded on the ISBI 2012 electron microscopy segmentation task, outperforming prior methods with a warping error of 0.0003529 mm, and for the ISBI 2015 challenge, it won the cell tracking categories for phase contrast and DIC images, as well as the dental X-ray image segmentation challenge. To facilitate reproducibility and broader use, the authors released a full Caffe-based implementation along with pre-trained networks shortly after publication, hosted on the University of Freiburg's website, which spurred rapid adoption in the research community through subsequent open-source ports and frameworks.

Adoption and Further Developments

Since its introduction in 2015, the U-Net architecture has experienced exponential citation growth, surpassing 118,000 citations by 2025 and solidifying its status as a cornerstone in the biomedical literature. This rapid accrual reflects its broad applicability and influence, with the original paper serving as a frequent reference point for subsequent innovations in convolutional neural networks for segmentation tasks. U-Net's adoption has been widespread in major deep learning frameworks, including official integrations in via the MONAI library for and numerous implementations in / ecosystems. This accessibility has spurred community-driven developments, with over 100 papers annually proposing variants or extensions since the mid-2010s, as evidenced by publication trends in segmentation-focused conferences and journals. Key milestones include the U-Net extension in 2016, which adapted the for volumetric data while leveraging sparse annotations, and U-Net++ in 2018, introducing nested skip pathways for improved multi-scale feature aggregation. More recently, U-Net's influence extended to -based models like UNETR in 2021, which replaces the encoder with a transformer to capture global context in segmentation. Efforts to address U-Net's limitations have focused on mitigating through techniques like enhanced and dropout integration, particularly in data-scarce scenarios. For deployment efficiency on edge devices, lightweight variants such as EdgeMedNet have been developed, reducing parameters while preserving segmentation accuracy for resource-constrained environments like mobile neural compute sticks. Looking ahead, future developments emphasize hybrids integrating U-Net with diffusion models, such as Diffusion-CSPAM-U-Net, to enhance generative segmentation capabilities and handle noisy or incomplete data. Additionally, ethical considerations in AI-driven segmentation highlight the need to combat biases in medical datasets, including demographic disparities that can skew U-Net performance across patient groups, prompting calls for fairness-aware training protocols.

References

  1. [1]
    U-Net: Convolutional Networks for Biomedical Image Segmentation
    May 18, 2015 · In this paper, we present a network and training strategy that relies on the strong use of data augmentation to use the available annotated samples more ...
  2. [2]
    U-Net and its variants for medical image segmentation - arXiv
    Nov 2, 2020 · The success of U-net is evident in its widespread use in all major image modalities from CT scans and MRI to X-rays and microscopy. Furthermore, ...Missing: impact | Show results with:impact
  3. [3]
    Medical Image Segmentation Review: The success of U-Net - arXiv
    Nov 27, 2022 · U-Net is the most widespread image segmentation architecture due to its flexibility, optimized modular design, and success in all medical image modalities.Missing: impact | Show results with:impact
  4. [4]
    [PDF] U-Net: Convolutional Networks for Biomedical Image Segmentation
    In this paper, we build upon a more elegant architecture, the so-called “fully convolutional network” [9]. We modify and extend this architecture such that ...
  5. [5]
    A novel unified Inception-U-Net hybrid gravitational optimization ...
    Aug 14, 2025 · Pre-processing involves image resizing with cropping, followed by extensive data augmentation techniques including contrast adjustment, image ...
  6. [6]
    Systematic Evaluation of Image Tiling Adverse Effects on Deep ... - NIH
    Feb 7, 2020 · Inference was performed using tiles of the same size that was used when training the model, with a 50% overlap between tiles in both the ...Missing: strategy | Show results with:strategy
  7. [7]
    [PDF] A survey of loss functions for semantic segmentation - arXiv
    Sep 3, 2020 · In this paper, we have summarized 14 well-known loss functions for semantic segmentation and proposed a tractable variant of dice loss function ...
  8. [8]
    An effective expansion of dice loss for medical image segmentation
    Dice loss closes all positive instances predicted by a model to the ground truth, and is a powerful method for achieving a semantic segmentation because it can ...
  9. [9]
    U-Net Architecture for Prostate Segmentation: The Impact of Loss ...
    Compound loss: Compound loss functions are a combination of different types of loss functions, mostly cross-entropy and Dice similarity coefficient. This loss ...
  10. [10]
    3D U-Net: Learning Dense Volumetric Segmentation from Sparse ...
    Jun 21, 2016 · Abstract:This paper introduces a network for volumetric segmentation that learns from sparsely annotated volumetric images.
  11. [11]
    3D U-Net: Learning Dense Volumetric Segmentation from Sparse ...
    Oct 2, 2016 · This paper introduces a network for volumetric segmentation that learns from sparsely annotated volumetric images.
  12. [12]
    3D U-Net Improves Automatic Brain Extraction for Isotropic Rat ... - NIH
    Second, patch-based training could lose information/segmentation consistency or overfit the data if the patch size and number of training samples are imbalanced ...Missing: details memory
  13. [13]
    and Three-Dimensional-Based U-Net Architectures for Brain Tissue ...
    Jan 10, 2022 · We aimed to compare the performance of 2D- and 3D-based segmentation networks to perform brain tissue classification in anisotropic CT scans.Model Development And... · Image Preprocessing · 3d U-Net Structure And...
  14. [14]
    Ultrafast 3D segmentation of brain-wide optical neuronal volume
    High-resolution segmentation of 3D optical neuron image is crucial for individual neuron reconstruction and neural circuit deciphering.
  15. [15]
    A GPU-based computational framework that bridges neuron ... - Nature
    Sep 18, 2023 · We theoretically prove that the DHS implementation is computationally optimal and accurate. This GPU-based method performs with 2-3 orders of ...
  16. [16]
    UNet++: A Nested U-Net Architecture for Medical Image Segmentation
    Jul 18, 2018 · In this paper, we present UNet++, a new, more powerful architecture for medical image segmentation. Our architecture is essentially a deeply-supervised encoder ...
  17. [17]
    Attention U-Net: Learning Where to Look for the Pancreas - arXiv
    Apr 11, 2018 · We propose a novel attention gate (AG) model for medical imaging that automatically learns to focus on target structures of varying shapes and sizes.
  18. [18]
    Recurrent Residual Convolutional Neural Network based on U-Net ...
    Feb 20, 2018 · In this paper, we propose a Recurrent Convolutional Neural Network (RCNN) based on U-Net as well as a Recurrent Residual Convolutional Neural Network (RRCNN) ...
  19. [19]
    V-Net: Fully Convolutional Neural Networks for Volumetric Medical ...
    Jun 15, 2016 · In this work we propose an approach to 3D image segmentation based on a volumetric, fully convolutional, neural network.
  20. [20]
    Fully Automatic Liver and Tumor Segmentation from CT Image ... - NIH
    In this study, we obtained the best DSC, JSC, and ACC liver segmentation performance metrics on the CHAOS dataset as 97.86%, 96.10%, and 99.75%, respectively, ...
  21. [21]
    The Liver Tumor Segmentation Benchmark (LiTS) - ScienceDirect.com
    The best liver segmentation algorithm achieved a Dice score of 0.963, whereas, for tumor segmentation, the best algorithms achieved Dices scores of 0.674 (ISBI ...
  22. [22]
    MSS U-Net: 3D segmentation of kidneys and tumors from CT images ...
    We present a multi-scale supervised 3D U-Net, MSS U-Net to segment kidneys and kidney tumors from CT images.
  23. [23]
    DenseRes-Unet: Segmentation of overlapped/clustered nuclei from ...
    We proposed a model to segment overlapped nuclei from H&E stained images. U-Net model achieved state-of-the-art performance in many medical image segmentation ...
  24. [24]
    A dual decoder U-Net-based model for nuclei instance ... - Frontiers
    In this paper, we propose a novel architecture, consisting of one encoder and two decoders, to perform nuclei instance segmentation in H&E-stained histological ...
  25. [25]
    Artificial Intelligence-Enabled Medical Devices - FDA
    Jul 10, 2025 · The AI-Enabled Medical Device List is a resource intended to identify AI-enabled medical devices that are authorized for marketing in the ...Artificial Intelligence in... · 510(k) Premarket Notification · SoftwareMissing: Net biomedical
  26. [26]
    An improved U-net based retinal vessel image segmentation method
    Fundus images have disadvantages such as uneven brightness, poor contrast, and strong noise, requiring per-processing before input the network for training.
  27. [27]
    Brain Tumor Segmentation using U-Net - Kaggle
    The Brain Tumor Segmentation (BraTS) 2020 dataset is a collection of multimodal Magnetic Resonance Imaging (MRI) scans used for the segmentation of brain tumors ...
  28. [28]
  29. [29]
    Segmentation of Satellite Imagery using U-Net Models for Land ...
    Mar 5, 2020 · This paper uses a modified U-Net model for land cover classification from satellite imagery, aiming to increase accuracy and change detection. ...
  30. [30]
    AID-U-Net: An Innovative Deep Convolutional Architecture for ... - NIH
    Nov 25, 2022 · Achieving mIoU of 53.13% in PASCAL and 55.84 in ADE20K datasets. Running 3 times faster than FCN. Unbalance flexibility between contracting ...
  31. [31]
    Liver margin segmentation in abdominal CT images using U-Net ...
    Mar 13, 2025 · The core of our study involves the implementation of two advanced deep learning models, U-Net and Detectron2, which are applied to the prepared ...
  32. [32]
    Our U-net wins two Challenges at ISBI 2015
    Apr 16, 2015 · The Cell Tracking Challenge compares the performance of segmentation and tracking algorithms on a set of 13 very different microscopic time ...
  33. [33]
  34. [34]
    UNet and its Family: UNet++, Residual UNet, and Attention UNet
    Aug 21, 2025 · Paper: U-Net: Convolutional Networks for Biomedical Image Segmentation by Olaf Ronneberger et al. (MICCAI 2015) Citations: 118,000+ (as of 2025).
  35. [35]
    (PDF) U-Net and its variants for medical image segmentation: theory ...
    PDF | U-net is an image segmentation technique developed primarily for medical image analysis that can precisely segment images using a scarce amount of.
  36. [36]
    UNETR: Transformers for 3D Medical Image Segmentation - arXiv
    Mar 18, 2021 · UNETR uses a transformer encoder to learn sequence representations for 3D medical image segmentation, capturing global multi-scale information.
  37. [37]
    A deeper and more compact split-attention U-Net for medical image ...
    In this paper, we propose a deeper and more compact split-attention u-shape network, which efficiently utilises low-level and high-level semantic information.<|control11|><|separator|>
  38. [38]
    EdgeMedNet: Lightweight and Accurate U-Net for Implementing ...
    Jun 30, 2023 · We propose EdgeMedNet, which is one lightweight and accurate U-Net model to enable the efficient medical image segmentation on Intel/Movidius Neural Compute ...
  39. [39]
    Diffusion-CSPAM U-Net: A U-Net model integrated hybrid attention ...
    Apr 5, 2025 · This study aimed to develop and evaluate a Diffusion-CSPAM-U-Net model for the segmentation of brain metastases on CT images and thereby provide a robust tool ...
  40. [40]
    Bias in artificial intelligence for medical imaging - PubMed Central
    AI in medical imaging is at risk of being compromised by several types of biases, which could adversely affect patient outcomes. • Understanding that medical ...