Next Article in Journal
Challenges and Solution Directions for the Integration of Smart Information Systems in the Agri-Food Sector
Previous Article in Journal
A Novel Key Distribution for Mobile Patient Authentication Inspired by the Federated Learning Concept and Based on the Diffie–Hellman Elliptic Curve
Previous Article in Special Issue
Enhanced U-Net for Infant Brain MRI Segmentation: A (2+1)D Convolutional Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning-Based Medical Ultrasound Image and Video Segmentation Methods: Overview, Frontiers, and Challenges

1
College of Mathematical Medicine, Zhejiang Normal University, Jinhua 321004, China
2
School of Computer Science and Technology (School of Artificial Intelligence), Zhejiang Normal University, Jinhua 321004, China
3
Puyang Institute of Big Data and Artificial Intelligence, Puyang 457006, China
4
School of Mathematical Sciences, Zhejiang University, Hangzhou 310027, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2025, 25(8), 2361; https://doi.org/10.3390/s25082361
Submission received: 13 March 2025 / Revised: 3 April 2025 / Accepted: 5 April 2025 / Published: 8 April 2025

Abstract

:
The intricate imaging structures, artifacts, and noise present in ultrasound images and videos pose significant challenges for accurate segmentation. Deep learning has recently emerged as a prominent field, playing a crucial role in medical image processing. This paper reviews ultrasound image and video segmentation methods based on deep learning techniques, summarizing the latest developments in this field, such as diffusion and segment anything models as well as classical methods. These methods are classified into four main categories based on the characteristics of the segmentation methods. Each category is outlined and evaluated in the corresponding section. We provide a comprehensive overview of deep learning-based ultrasound image segmentation methods, evaluation metrics, and common ultrasound datasets, hoping to explain the advantages and disadvantages of each method, summarize its achievements, and discuss challenges and future trends.

1. Introduction

Continuous improvements in artificial intelligence, especially in deep learning techniques, are helping to identify, classify, and quantify patterns in clinical images [1]. Applying deep learning to lesion detection, classification, and other related tasks, deep learning demonstrates superior performance compared to traditional techniques and even surpasses some highly skilled medical professionals in certain tasks [2]. Deep learning is a type of machine learning that enables computers to learn from experience and comprehend the environment through a hierarchical conceptual framework [3]. Models of deep learning have the capability to learn and extract intricate patterns and features from vast datasets, making them exceptionally proficient in areas such as image recognition, natural language processing, medical diagnosis, and autonomous vehicles. Applying some deep learning models to medical images can provide very critical diagnostic evidence for clinical diagnosis. Deep learning has revolutionized medical image analysis through its applications in core computer vision tasks, including image classification, lesion detection [4], semantic segmentation, multimodal registration, reconstruction from sparse data, and cross-modality synthesis. Among these, the foundational triad—classification, detection, and segmentation—continues to dominate clinical implementations [5], while emerging tasks like deformable registration and pathology quantification are increasingly addressing complex diagnostic challenges. Medical ultrasound imaging is a safe, painless, and non-invasive diagnostic tool that has become well-established in clinical practice. This is primarily due to its advantages of low cost, portability, and widespread availability compared to other medical imaging modalities [6]. The segmentation of medical images plays a crucial role in clinical diagnosis, and accurate segmentation can facilitate more precise quantitative and qualitative analysis. It plays a crucial role in the diagnosis of various medical conditions, including internal organ tumors, and has also become a standard method for prenatal diagnosis [7]. Image segmentation involves the extraction of regions of interest from the background in an ultrasound image, which significantly influences the quality of the final analysis [8]. However, due to the complex imaging structure of ultrasound images, a variety of artifacts and noises, including high scattering noise due to the coherence of ultrasound imaging [9,10], low signal-to-noise ratios and intensity inhomogeneities, and the need for skilled operators, pose additional challenges to image segmentation. Figure 1 presents a series of representative ultrasound images obtained from a typical breast ultrasound examination.
A variety of methodologies have been proposed to address the aforementioned challenges. Some of the commonly used methods include the application of active contour models [11,12] for image and video segmentation based on levels set by spectral clustering and normalization criterion (NCut) [13], segmentation methods based on texture and shape prior [14], statistical shape models using Gabor filters [15], using a radial basis function neural network for classification to differentiate potential segmented regions [16], model initialization and efficient DDC (distance dynamic contour) contour refinement [9], the design of partial differential equation-based streams for maximum likelihood segmentation in scene targets [17], and a priori property profiles for fully automatic segmentation [18]. Despite extensive research into ultrasound image and video segmentation, it continues to be a prominent area of focus in current studies. In recent years, various deep learning-based ultrasound image and video segmentation methods have been proposed, aiming to leverage the promising potential of artificial intelligence to assist doctors in addressing the challenges posed by complex imaging structures, various artifacts, and noise in medical ultrasound image segmentation, thereby improving segmentation accuracy [19]. Therefore, this paper classifies deep learning approaches in ultrasound image segmentation based on the neural network structure as well as in combination with other recently emerged methods, including: U-Net neural networks and its variants, fully convolutional neural networks (FCNs), recurrent neural networks (RNNs), generative adversarial networks (GANs), deep reinforcement learning (DRL) methods, diffusion models, weakly supervised learning (WSL), and segment anything models (SAMs). Furthermore, this paper elaborates on several evaluation methods for image segmentation and introduces a number of high-quality, publicly available datasets. Figure 2 presents statistical data regarding the classification of the previously mentioned deep learning models.
Figure 1. Ultrasound images of breast. The ultrasound images of the left and right breasts, denoted as (a) and (b), respectively, show relatively low contrast. Images (c,d) have annotations on the areas of the breast nodules. Ultrasound image showing a lesion marked with measurement calipers. The crosses indicate the boundaries used to calculate the transverse (1) and longitudinal (2) diameters. The rectangular region highlights the area of interest for further analysis. https://datasetninja.com/breast-ultasound-images (accessed on 3 January 2025).
Figure 1. Ultrasound images of breast. The ultrasound images of the left and right breasts, denoted as (a) and (b), respectively, show relatively low contrast. Images (c,d) have annotations on the areas of the breast nodules. Ultrasound image showing a lesion marked with measurement calipers. The crosses indicate the boundaries used to calculate the transverse (1) and longitudinal (2) diameters. The rectangular region highlights the area of interest for further analysis. https://datasetninja.com/breast-ultasound-images (accessed on 3 January 2025).
Sensors 25 02361 g001
The main contributions of this paper are listed as follows:
  • To our knowledge, this is the first systematic review of deep learning-based medical ultrasound image and video segmentation methods, including common evaluation methods and datasets.
  • With a survey of more than 80 papers, we provide comprehensive introductions, analyses, and detailed statistics on recent and classical publications from different perspectives (such as methods and evaluation criteria) in the related sections, tables, and figures.
  • According to our survey and statistics, we reasonably put forward challenges and future trends in the Discussion and Conclusion section.
Figure 2. Deep learning is a branch of artificial intelligence and machine learning. According to the classification of the learning paradigm, it is classed as more detailed, secondary classification learning.
Figure 2. Deep learning is a branch of artificial intelligence and machine learning. According to the classification of the learning paradigm, it is classed as more detailed, secondary classification learning.
Sensors 25 02361 g002
The remainder of our paper is organized as follows: Section 2 introduces common architectures in deep learning and some recently emerging artificial intelligence models combined with deep learning. In Section 3, we discuss related work on deep learning-based medical ultrasound image and video segmentation methods and review publicly available medical ultrasound image and video datasets. In addition, for some deep learning methods with more data, the literature is summarized in a table, including the ROI, dataset, corresponding model, and so on. Section 4 presents common evaluation methods. In the Section 5 we draw conclusions by discussing issues related to deep learning methods for medical ultrasound image and video segmentation and speculate on future technical developments and trends in this research field.

2. Description of Deep Learning Methods Based on Learning Paradigm Classification

In recent years, deep learning has become the dominant paradigm in medical ultrasound image and video segmentation. To better capture the diversity of approaches and address concerns about classification clarity, we categorize existing methods based on learning paradigms, namely, supervised learning, semi-supervised learning, weakly supervised learning, unsupervised learning, deep reinforcement learning, diffusion models, and the segment anything model (SAM), which are advanced techniques applied in medical ultrasound image segmentation. Although SAM is pre-trained in a fully supervised manner on a large-scale dataset, it performs segmentation during inference using only a small amount of weak supervision (prompts). Therefore, it is categorized under the weakly supervised learning paradigm. Each paradigm reflects a different degree of reliance on annotated data and offers unique advantages and challenges.

2.1. Supervised Learning

Supervised learning methods require fully annotated data (i.e., pixel-level labels for each image) to train segmentation models. These methods typically offer the highest accuracy due to the richness of supervision signals. Prominent approaches are described below.

2.1.1. Fully Convolutional Neural Networks

FCNs are a type of deep learning model designed for semantic image segmentation tasks. A key feature of FCNs compared to traditional CNNs is their ability to accept input images of arbitrary size and output dense predictive maps of the corresponding size, which makes FCNs particularly suitable for image segmentation tasks. Long et al. proposed fully convolutional neural networks in 2015 [20], which opened a new chapter in the use of deep learning for image segmentation. FCNs adapt AlexNet, VGG networks, and GoogLeNet to full convolutional networks based on the end-to-end, pixel-to-pixel training of convolutional networks, starting from supervised pretraining without the need for complex processing in state-of-the-art semantic segmentation [21]. Global convolutional networks transfer their learned representations to the segmentation task by fine-tuning, defining a skip architecture used for more accurate segmentation. Since learning and inference are performed through intensive feed-forward computation and backpropagation, the in-network upsampling layer facilitates pixel-level prediction and learning in networks that incorporate subsampling pools. Figure 3 shows a typical FCN network architecture.

2.1.2. U-Net Convolutional Neural Networks

U-Net is a variant convolutional network model based on the traditional CNN convolutional network model, which is still one of the most commonly used network architectures due to its excellent performance since it was proposed by Ronneberger et al. in 2015 [22]. An early popular approach to deep learning for semantic segmentation was patch classification, which has limitations due to its requirement for fixed-size input images and relatively slow processing speed; at the same time, it generated large amounts of redundant data [23]. U-Net modifies and extends the traditional convolutional network by incorporating a symmetric contraction path for robust context capture, a symmetric expansion path for precise localization, and enhanced upsampling layers that utilize multiple feature channels to transmit contextual information to higher-resolution layers. This architecture enables more accurate segmentation with a reduced number of training images. U-Net improves resolution by adding a continuous layer to the traditional shrinking grid and replacing the pooling layer with an upsampling layer. The network does not have any fully connected layers and uses only the effective portion of each convolution, i.e., the segmentation map contains only the pixels available in the full context of the input image. Extensive data augmentation through the application of elastic deformations to the available training datasets, coupled with the utilization of a weighted loss function to distinguish objects in contact within the same class, renders the network well-suited for a diverse array of biomedical challenges. Figure 4 shows a generic U-Net architecture. Due to the simplicity and superior performance of the U-Net architecture, various U-Net-like methods continue to emerge, such as Res-UNet [24] and Dense-UNet [25], and it has been extended to other medical image analysis tasks, such as image denoising [26], image registration [27], and more.

2.1.3. Recurrent Neural Networks

RNNs are commonly used in natural language processing (NLP). In contrast to feed-forward networks, recurrent neural networks are suitable for processing temporal signals [28]. A typical RNN consists of three parts: sequential input data, a hidden state, and sequential output data. The RNN utilizes sequential information and executes the same operation for each element of the sequence, where the output is contingent upon prior computations. The RNN is capable of processing sequences of arbitrary length, and the internal state is used to model and “remember” the previously processed information. Long short term memory (LSTM) and gated recurrent units (GRUs) are two of the most prevalent architectures in recurrent neural network models, capable of effectively simulating long-term dependencies [29].
LSTM networks manage information flow through three gates—forget gate, input gate, and output gate—and a cell state. These components enable precise control over what information is retained, discarded, or updated, effectively solving the long-term dependency problem in traditional RNNs. The internal operations of an LSTM unit are described by the following equations:
f t = σ ( W f · [ h t 1 , x t ] + b f ) , / / Forget gate
i t = σ ( W i · [ h t 1 , x t ] + b i ) , / / Input gate
C ˜ t = tanh ( W C · [ h t 1 , x t ] + b C ) , / / Candidate cell state
C t = f t C t 1 + i t C ˜ t , / / Cell state update
o t = σ ( W o · [ h t 1 , x t ] + b o ) , / / Output gate
h t = o t tanh ( C t ) , / / Hidden state output
where σ denotes the sigmoid activation function, tanh is the hyperbolic tangent function, and ⊙ represents element-wise multiplication. The gates determine how much of the previous hidden state h t 1 and current input x t should influence the current step.
A GRU is a simplified variant of LSTM that merges the cell state and hidden state into a single hidden representation. GRU uses two gates—update gate and reset gate—to control the flow of information, reducing the number of parameters while preserving sequence modeling capabilities. The internal operations of a GRU unit are defined as follows:
r t = σ ( W r · [ h t 1 , x t ] + b r ) , / / Reset gate
z t = σ ( W z · [ h t 1 , x t ] + b z ) , / / Update gate
h ˜ t = tanh ( W h · [ r t h t 1 , x t ] + b h ) , / / Candidate hidden state
h t = ( 1 z t ) h t 1 + z t h ˜ t , / / Final hidden state
where σ is the sigmoid function, tanh is the hyperbolic tangent activation, and ⊙ denotes element-wise multiplication. The reset gate r t determines how much of the previous hidden state to forget, while the update gate z t controls how much of the candidate state h ˜ t should be included in the new hidden state h t .

2.2. Unsupervised Learning

Unsupervised methods do not rely on any human annotations. Instead, they extract segmentation-relevant information from intrinsic data patterns.

2.2.1. Generating Adversarial Networks

Goodfellow et al. proposed GANs in 2014, which have become a major breakthrough in generative models [30]. The core idea of GANs is to train two networks through an adversarial process: a generator and a discriminator. These two networks compete with each other during the training process to improve the quality of the generated data. GANs, which are typically employed in unsupervised learning scenarios, integrate both generator and discriminator models. By utilizing two neural networks with opposing objectives, GANs achieve joint optimization through adversarial training [31]. The generator network is considered as a set of pretenders and the discriminator is responsible for classifying samples as true or false. By leveraging adversarial attacks on the discriminators, the generators deceive the discriminators, leading them to compete inaccurately. During the training process, the generators and the discriminators fight against each other: the generators try to produce data that look real, while the discriminators try to distinguish between real data and generated data. This process can be viewed as a minimax game, where the ultimate objective is for the generator to be capable of producing high-quality synthetic data that are sufficient to deceive the discriminator. Figure 5 shows the basic framework of GAN.

2.2.2. Diffusion Model

A class of methods based on generative models, known as diffusion models (DMs), primarily generate data, including images, audio, and text. These models learn the distribution of data by modeling the inverse of the data generation process, i.e., the gradual generation of data from random noise. Diffusion models have received a significant deal of attention in recent years due to their excellent generation quality and versatility. We can divide the diffusion model’s working principle into two stages: the forward process and the reverse process. During the forward process, the model gradually introduces Gaussian noise into the data, ultimately transforming it into unstructured noise. This process usually consists of a series of predefined steps, each of which adds a certain amount of noise. In the reverse process, the model learns how to gradually recover the original data from the noise. This process essentially teaches the model the opposite of the forward process. Through training, the model learns to gradually remove the noise, eventually generating samples that match the distribution of the training data. Diffusion models have demonstrated extremely powerful generative capabilities in many domains, such as image generation, audio generation with super resolution, and so on. People widely recognize diffusion models for their high-quality and diverse generative capabilities. Compared to other generative models (e.g., GANs and VAEs), diffusion models are more stable during training and less prone to mode collapse. However, their main challenge is the high computational cost of the generative process, as multiple iterative steps are required to generate data from noise.

2.3. Reinforcement Learning

Reinforcement learning (RL) formulates segmentation as a sequential decision-making process. An agent learns to perform segmentation actions to maximize cumulative rewards, which are defined based on segmentation quality.

2.4. Deep Reinforcement Learning

Reinforcement learning is a theoretical framework that integrates psychological and engineering principles to investigate how intelligent agents interact with their environment to develop strategies that maximize expected rewards. Traditional reinforcement learning algorithms are predominantly designed for tabular scenarios, offering principled solutions for straightforward tasks. However, these algorithms encounter significant challenges when applied to highly complex domains, particularly those involving three-dimensional environments. Deep reinforcement learning emerges as a paradigm-shifting methodology that synergistically integrates reinforcement learning’s decision-making framework with deep neural networks’ representation capacity. This hybrid architecture capitalizes on deep learning’s hierarchical feature extraction to solve sequential decision problems in high-dimensional state spaces [32], effectively bridging the gap between perceptual understanding and strategic action planning in complex environments. Deep reinforcement learning provides a comprehensive framework for studying the interaction between learning, representation, and decision-making, offering a new set of research tools and a wide range of new hypotheses for neuroscience [33]. Most machine learning paradigms involve pre-collecting or constructing dataset labels and performing machine learning based on existing static data. In contrast, reinforcement learning is a typical representative of the closed-loop learning paradigm, using dynamic data and labels to bring feedback signals into the learning process (Figure 6). The emergent capability of deep reinforcement learning in constructing hierarchical state abstractions has positioned it as a principled methodology for sequential decision-making, where deep neural networks dynamically encode environmental observables while reinforcement learning orchestrates temporal credit assignment. The integration of these two advanced technologies is regarded as one of the most promising approaches to achieving artificial general intelligence [34].

2.5. Weakly Supervised Learning

Weakly supervised learning in deep learning is a method that utilizes limited or incompletely labeled data to train deep neural networks. This approach is particularly valuable when data labeling is costly or difficult. Weakly supervised learning operates on imperfectly annotated datasets [35], particularly in scenarios where merely image-level labels are accessible [36]. The related literature categorizes weakly supervised learning into three distinct categories.
More specifically, in incomplete supervision, common strategies include semi-supervised learning and active learning, where models leverage both labeled and unlabeled data. For instance, image classification tasks may use pseudo-labeling or consistency regularization techniques to exploit the large amount of unlabeled data. In the case of inaccurate supervision, label noise becomes a major challenge. To address this, methods such as robust loss functions, label correction algorithms, or noise-tolerant architectures have been developed. For example, in medical image analysis, misannotations are common due to subjective interpretation, and robust learning techniques help improve model reliability. Finally, uncertain supervision often arises in scenarios where labels are ambiguous or derived from weak sources, such as web-scraped data or user-generated tags. To handle this, researchers have explored probabilistic models or multiple instance learning (MIL), which allow learning from label distributions or ambiguous annotations.
The first type is incomplete supervision, also known as incomplete annotation. In this type of weak supervision, only a small portion of the training set is annotated [37]. The second type, known as inaccurate supervision, refers to weakly supervised object recognition using convolutional neural networks, where the labels in the dataset are not completely accurate. The third type is uncertain supervision, where only coarse labels exist [38]. Figure 7 shows how weakly supervised learning works.

Segment Anything Model

The segment anything model (SAM) is an advanced image segmentation model developed by Meta [39]. This model aims to automatically segment any object in any image without the need for pre-labeled datasets. The design goal of SAM is to quickly and accurately complete object segmentation tasks in images with minimal or even no labeled data, relying on its powerful learning ability and generalization capability. It employs efficient deep learning algorithms and large-scale pre-trained models, capable of handling various complex scenes and diverse image contents. SAM has a wide range of applications, including but not limited to autonomous driving, medical image analysis, intelligent surveillance, and augmented reality. By offering open-source code along with comprehensive technical documentation, SAM significantly enhances the development of image segmentation technology and provides robust tools and substantial support for associated research and applications.
It employs efficient deep learning algorithms and large-scale pre-trained models, capable of handling various complex scenes and diverse image contents.
Specifically, SAM is composed of three main components: a prompt encoder, an image encoder, and a mask decoder. The image encoder is based on a vision transformer (ViT), which processes the input image to produce a dense embedding E I . The prompt encoder transforms various user prompts (e.g., points, boxes, or masks) into embeddings E P . These two embeddings are fused within the mask decoder, which is a lightweight transformer decoder that predicts object masks in a single forward pass.
Let I be the input image and P be the prompt. The SAM pipeline can be summarized as follows:
E I = ViT ( I ) ,
E P = PromptEncoder ( P ) ,
M = MaskDecoder ( E I , E P ) ,
where M denotes the output segmentation mask. The model is trained with a combination of binary cross-entropy loss and Dice loss to improve both pixel-wise accuracy and mask quality.
A unique feature of SAM is its ability to handle multiple prompts in a unified framework, allowing segmentation to be guided by different types of user input. This promptable segmentation mechanism enables interactive segmentation and supports zero-shot generalization across various domains. Figure 8 shows the whole process of ultrasound image segmentation by SAM.

3. Deep Learning in Ultrasound Image and Video Segmentation

This paper categorizes deep learning-based methods for ultrasound image and video segmentation according to different learning paradigms, including (1) supervised learning-based methods, (2) unsupervised learning-based methods, (3) weakly supervised learning-based methods, and (4) deep reinforcement learning-based methods. A literature search was systematically conducted across multiple academic databases, including Google Scholar, PubMed, Web of Science, and Semantic Scholar. The search strategy involved a combination of keywords (e.g., “machine learning”, “deep learning”, “ultrasound image segmentation”, “ultrasound video segmentation”, “CNN”, “recurrent neural networks”, “deep reinforcement learning”), which were iteratively refined to capture a comprehensive set of relevant articles.
The search was limited to publications in English spanning the last 5–10 years to ensure that the review reflected the most recent advancements in the field. Papers were included based on the following labels: (1) studies that focus on ultrasound images or videos, including both conventional and deep learning-based segmentation methods; (2) articles that present segmentation techniques, evaluation methods (e.g., commonly used metrics, high-quality datasets), or comprehensive experimental evaluations specific to ultrasound imaging; (3) papers published in well-defined, peer-reviewed journals or conference proceedings with clear methodological details. Articles that did not meet the following criteria were excluded: (1) articles not directly related to ultrasound imaging segmentation or that focus solely on other imaging modalities; (2) non-English publications, reviews, commentaries, editorials, and conference abstracts without full-text availability; (3) duplicate records and studies that do not provide sufficient methodological details or experimental validation. Figure 9 illustrates the overall process of literature retrieval using a Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) flow diagram. Initially, about 25,000 documents were searched, and the search results were screened through several rounds of (1) removing duplicate papers to include the above criteria, (2) reading through the entire text to determine its relevance to the topic of this article, and excluding studies that were not pertinent to the research focus. Finally, about 97 articles were selected for use in this article, which systematically introduces the research on deep learning-based methods for ultrasound image and video segmentation and evaluates the model performance of individual studies, assessing the methodology and part of the high-quality dataset and generally ensuring the comprehensiveness of the deep learning-based methods for medical ultrasound images.

3.1. U-Net in Ultrasound Image and Video Segmentation

3.1.1. Overview of Works

Recently, Vahid et al. [40] addressed this problem with a new U-Net-based architecture called fast and accurate U-Net for the medical image segmentation task. The proposed fast and accurate U-Net model primarily consists of four tuned 2D convolutional layers, a 2D transposed convolutional layer, and a batch normalization layer. There are four blocks in the codec path. The architectural performance was rigorously validated through dual evaluation protocols: (1) a dedicated segmentation task dataset measuring fetal head and abdominal circumference and (2) the publicly benchmarked HC18-Grand Challenge dataset for fetal head circumference quantification. This rapid network significantly improved processing time compared to U-Net, dilated U-Net, R2U-Net, attention U-Net, and MFP U-Net. Amiri et al. [41] studied the impact of fine-tuning different layers of a pre-trained U-Net on breast ultrasound image segmentation. Compared to fixing the contraction part and adjusting the expansion part, adjusting the contraction part while fixing the expansion part produced significantly better results. In the same year, the authors proposed using one U-Net to detect lesions before performing breast ultrasound image segmentation and then using another U-Net to segment the detected areas. This approach improved the average Dice score by 1.8%. For images with an original Dice score below 70%, the average Dice score increased by 14.5% [42]. Cheng et al. [43] explored the best transfer learning techniques for accurate and fast image segmentation using U-Net. The study employed two transfer learning methods: first, they built U-Net using a pre-trained VGG16 model (V-Unet); second, they trained U-Net using a grayscale natural salient object dataset. The Dice coefficient of 0.8 achieved a satisfactory level. At the same time, Chen et al. [44] suggested a simple but powerful Nested U-net (NU-net) for accurately separating breast tumors. NU-net uses U-Nets with shared weights to create a strong representation of breast tumors. NU-net reduces the network’s sensitivity to input images of different scales and further enhances the representation capability of objects or regions. Xu et al. [45] suggested a multi-scale feature extraction and fusion network (MEF-UNet) for ultrasound image segmentation. It has three parts: a selective feature extraction module, a contextual information storage module, and a multi-scale feature fusion module. Neural Architecture Search (NAS) has made significant progress in improving image classification accuracy. Recently, some work has attempted to extend NAS to image segmentation. Weng et al. [46], inspired by the U-net architecture and its successful variants applied to various medical image segmentation tasks, proposed NAS-Unet. The proposed architecture employs symmetrical DownSC (Downsampling Separable Convolution) and UpSC (Upsampling Separable Convolution) blocks within a U-shaped topology, demonstrating superior segmentation accuracy with significantly reduced parameter complexity compared to standard U-Net architectures and their derivatives. Ultrasound image annotation heavily relies on skilled physicians, such as in transvaginal ultrasound (TVUS), commonly used for diagnosing infertility, where doctors manually annotate the size and shape of ovaries and follicles. Li et al. [47] proposed a new composite network, CR-Unet, to simultaneously segment ovaries and follicles in TVUS. The proposed architecture incorporates a recurrent neural network within a U-Net framework to capture multi-scale spatial dependencies and long-range contextual patterns. Experimental evaluations on a cohort of 219 patients (3204 TVUS images) demonstrated segmentation accuracy with mean DSC scores of 0.912 for ovarian structures and 0.858 for follicular regions. To aid diagnosis, the automatic segmentation of regions of interest (ROIs) in ultrasound images has become an essential component of computer-aided diagnosis (CAD). Wang et al. [48] proposed a Scale Attention Convolution module, which constructs MSAC-Unet based on U-net, using MSAC instead of standard convolutions in each encoder and decoder for segmentation. MSAC-Unet did the best job of segmenting three datasets: two thyroid nodule datasets (TND-PUH3 and DDTI) and one brachial plexus dataset (NSD), and all three had Dice coefficients of 0.822, 0.792, and 0.746. MSAC-Unet significantly improved segmentation accuracy, with more reliable ROI edges and boundaries, and reduced the number of incorrectly segmented ROIs in ultrasound images. Inan et al. [49] proposed a new AI decision support system for the automatic segmentation and classification of thyroid nodule types. This system is based on a hybrid deep learning approach, using ResUNet++ for automatic thyroid nodule segmentation and combining DenseNet and ResNet for classification. Table 1 summarizes the data sets, models used, areas of focus, and model performance of the literature reviewed above.

3.1.2. Assessments

U-Net has demonstrated significant performance improvements in medical ultrasound image and video segmentation. Fast and accurate architectures, transfer learning, nested U-Net, multi-scale feature extraction and fusion networks, and composite networks are among the advancements that have significantly enhanced the accuracy and efficiency of segmentation using U-Net. These improvements not only reduce processing time but also enhance robustness to input images of different scales, improving Dice coefficients, especially in specific tasks like breast ultrasound, fetal head circumference measurement, and thyroid nodule segmentation. The results show that U-Net and its variations are very good at dealing with different kinds of complex lesions. This makes them a strong technical support for computer-aided diagnosis and automated segmentation in ultrasound imaging.

3.2. Fully Convolutional Neural Network in Ultrasound Image and Video Segmentation

3.2.1. Overview of Works

Fully convolutional network-based methodologies continue to advance precision medicine in medical ultrasound analysis. For lymph node characterization, Zhang et al. [50] pioneered a cascaded Coarse-to-Fine Stacked FCN (CFS-FCN), utilizing multi-stage FCN modules to automate lymph node delineation, while Villa et al. [51] developed an FCN-driven bone interface localization algorithm, outperforming manual segmentation and Phase Symmetry Confidence (CPS) benchmarks. In breast cancer diagnostics, Xing et al. [52] innovatively integrated FCNs with GANs to propose the Semi-Pixel Cycle GAN (SPCGAN), elevating lesion segmentation performance from 0.90 to 0.93 DSC. Hu et al. [53] addressed speckle noise challenges through a dilated FCN (DFCN) coupled with Phase-Based Active Contour models, leveraging deep dilated convolutions (rates 6–24) and batch normalization to enhance tumor-background differentiation in breast ultrasound.
Expanding the FCN’s multitask capabilities, Li et al. [54] incorporated an attention-guided feature selection mechanism and an elliptical ROI pooling layer, enabling simultaneous fetal biometric measurement (OFD/BPD) and anatomical segmentation. Qian et al. [55] achieved state-of-the-art breast ultrasound segmentation via quad-branch dilated convolutions (rates 1/3/5/7) for multiscale context aggregation, refined by Laplacian edge correction. For prostate cancer analysis, Feng et al. [56] constructed a multi-stage FCN architecture encoding spatial–boundary correlations, attaining 94.90% DSC in gland segmentation. In dynamic imaging, Xu et al. [57] proposed the Local–Global Reciprocal Network (LGRNet), synergizing Cyclic Neighborhood Propagation (CNP) for spatiotemporal context modeling with the Hilbert Selective Scan (HilbertSS) for noise suppression, achieving 89.80% IoU and 96.49% recall in uterine fibroid video analysis through hybrid DenseNet-ResNet classification modules. Table 2 summarizes the data sets, models used, areas of focus, and model performance of the literature reviewed above.

3.2.2. Assessments

FCNs have significant advantages and some drawbacks in medical ultrasound image segmentation. The primary advantages include their ability to perform pixel-level predictions throughout the entire process, thereby simplifying segmentation and enhancing the efficiency of both training and inference. Additionally, they can effectively extract spatial features from images via convolution operations, making them well suited for handling large volumes of image data. Moreover, their high degree of flexibility allows them to adapt to a variety of segmentation tasks. However, FCNs also face some disadvantages, such as their limited capacity to capture fine-grained features, particularly in complex and boundary-blurred ultrasound images. Additionally, FCNs are weak in modeling long-range dependencies, which limits their performance in tasks requiring global contextual information. Future directions for FCNs in medical ultrasound image segmentation could encompass incorporating attention mechanisms to enhance detail capture, integrating multi-scale features to boost accuracy, and combining with other networks such as RNNs or transformers to address limitations in long-range dependency modeling. These advancements would collectively improve overall segmentation performance and application effectiveness.

3.3. Recurrent Neural Networks in Ultrasound Image and Video Segmentation

3.3.1. Overview of Works

Due to the need to not only remember relevant information but also actively forget irrelevant information, LSTM is one of the most widely used architectures in recurrent neural networks. In recent years, this architecture has been increasingly utilized to address challenges in automatic ultrasound image and video segmentation, such as the lack of distinct boundaries and significant shape variations in ultrasound images. Pan et al. [58] proposed an automatic breast tumor segmentation method based on bidirectional LSTM and spatial-channel attention modules integrated into a fully convolutional network. The Dice Similarity Coefficient (DSC), Recall, Precision, and Hausdorff Distance (HD) of this method were 0.8178, 0.8067, 0.8292, and 11.1367, respectively. Transperineal ultrasound (TPUS) serves as the primary imaging modality for evaluating pelvic floor disorders. Noort et al. [59] proposed a U-net-like neural network with several convolutional LSTM layers to automatically segment the levator ani muscle (LAM) in TPUS volumes, achieving human-level performance in the segmentation task. To address the issue of incomplete boundaries in prostate ultrasound images, Yang et al. [60] proposed using a recurrent neural network to learn shape priors, employing a multi-view fusion strategy to integrate shape predictions from different views. This method outperformed the state-of-the-art CNNs and FCNs. Meanwhile, Webb et al. [61] proposed a four-stage trained convolutional LSTM model based on DeepLabv3C for segmenting thyroid ultrasound images, achieving average intersection scores of 0.427, 0.533, and 0.739 for cysts, nodules, and the thyroid. To avoid the time-consuming and labor-intensive manual intervention in ultrasound image segmentation, Horng et al. [62] proposed a convolutional neural network framework called DeepNerve, which is based on U-Net and combines Masktrack and LSTM for locating and segmenting the median nerve. The experimental results showed that the generated average Dice measure, precision, recall, and F-score were 0.8975, 0.8912, 0.9119, and 0.9015, respectively. Birth defects and associated mortality are among the foremost causes of infant death. Devisri et al. [63] proposed a method for fetal growth analysis using head circumference biometrics from ultrasound images through optimal segmentation and hybrid classifiers. This method uses U-Net to gather important features, enhances the boosting algorithm to choose the best features, and combines bidirectional LSTM and a convolutional neural network (B-LSTM-CNN) for analyzing fetal growth. Compared to LSTM, gated recurrent units (GRUs) require less memory and exhibit similar performance. Anas et al. [64] developed a hybrid CNN-GRU architecture for prostate ultrasound segmentation, where convolutional layers extracted spatial features while gated recurrent units captured temporal dependencies. Their experimental protocol utilized 2238 labeled transrectal ultrasound images for training, with 637 and 1017 images allocated for validation and testing, respectively, achieving a Dice coefficient of 0.93. Notably, early breast cancer detection remains clinically critical due to its profound impact on therapeutic efficacy and patient survival rates. Khdhir et al. [65] came up with a new hybrid method for breast cancer segmentation that uses the Gray-Level Co-occurrence Matrix (GLCM) for feature extraction and gated recurrent units for classification after preprocessing. Table 3 summarizes the data sets, models used, areas of focus, and model performance of the literature reviewed above.

3.3.2. Assessments

Because they can pick up on temporal dependencies and contextual information across image sequences, RNNs have shown a lot of promise in the field of medical ultrasound image and video segmentation. These networks excel in scenarios where the spatial–temporal coherence of the data is crucial for accurate segmentation. One of the primary advantages of RNNs is their capacity to integrate temporal information, which is particularly beneficial in medical ultrasound imaging. Unlike static images, ultrasound data often comprise sequential frames that capture dynamic biological processes. By leveraging RNNs, segmentation algorithms can utilize the continuity and evolution of these frames, leading to more accurate and consistent segmentation results. However, the application of RNNs in medical ultrasound segmentation is not without challenges. The sparsity of manual annotations available for training poses a critical issue. Medical annotations are often labor-intensive to produce, and obtaining a comprehensive dataset with dense annotations can be impractical. To address this, advanced strategies such as label propagation and sophisticated loss functions have been employed to enhance the training process despite sparse annotations.

3.4. Generating Adversarial Networks and Diffusion Models in Ultrasound Image and Video Segmentation

3.4.1. Overview of Works

Liang et al. [67] looked into an autoencoder generative adversarial network that combined the best features of GANs and variational autoencoders (VAEs) to make realistic images for medical thyroid ultrasound image enhancement. They tackled the issue of mode collapse in GANs, which results in less varied images. Conversely, mode collapse does not affect VAEs, but they do tend to produce blurry images. The test results showed that the images created mimicked real ultrasound features and thyroid tissue, improving training and helping U-Net models perform better in segmenting images. Liang et al. [68] proposed an image synthesis framework based on GANs. Recent advances in synthetic medical imaging have introduced novel approaches for enhancing anatomical realism. Bargsten et al. [69] developed SpeckleGAN, a specialized GAN framework that synthesizes domain-specific speckle patterns through learnable noise layers, particularly valuable for ultrasound modality simulation. This architecture uniquely integrates trainable speckle generators as differentiable network components, enabling physics-informed texture synthesis. Meanwhile, Fatima et al. [70] proposed a new method for the automatic segmentation of two-dimensional echocardiographic images. This method uses a generative adversarial network (Pix2Pix GAN), which consists of the PatchGAN model as the discriminator and the U-Net model as the generator, to accurately segment cardiac structures. Simultaneously, based on the diffusion model, Tang et al. [71] proposed a multi-layer global-context cross-consistency (MGCC) framework, which uses images generated by a latent diffusion model (LDM) as unlabeled images for semi-supervised learning. This framework uses an LDM to generate synthetic medical images, reducing the workload of data annotation and addressing privacy issues associated with medical data collection. Stojanovski et al. [72] proposed a new denoising diffusion probabilistic model guided by cardiac semantic labels to address the challenge of limited labeled data in the medical imaging field, particularly in ultrasound image analysis and cardiac segmentation tasks. They demonstrated an innovative medical image synthesis and analysis process, showing that synthetic images can effectively replace real data in training deep learning models. Katakis et al. [73] proposed a method for generating musculoskeletal ultrasound images using a diffusion model, demonstrating that synthetic data play a crucial role in the final performance of models and can improve deep learning systems in musculoskeletal ultrasound. Stevens et al. [74] proposed a joint posterior sampling framework that combines two independent diffusion models to simulate the distribution of clean and hazy ultrasound images in a supervised manner. The proposed dehazing method works well to get rid of haze while keeping signals from tissues that are weakly reflective, as shown by experiments conducted with living things and in the lab. Yao et al. [75] proposed a semi-supervised ultrasound image segmentation method called the Dual-Frequency Cascade Graph (DFCG) model. This method uses a latent diffusion model to make fake medical images to make data annotation easier. It also combines Fourier frequency domain space with a multi-scale attention mechanism to lower the effect of scattering noise. Table 4 summarizes the data sets, models used, areas of focus, and model performance of the literature reviewed above.

3.4.2. Assessments

These methods contribute to improving the performance of segmentation models by generating high-resolution synthetic images with realistic features. GANs have demonstrated their powerful capabilities in image synthesis, producing realistic ultrasound images and enhancing structural details, while diffusion models have improved the performance of segmentation models on real images by generating large amounts of synthetic data. However, challenges remain, including GANs’ susceptibility to mode collapse during training and the limited diversity of generated images. Diffusion models, on the other hand, require substantial computational resources, and the quality of the generated images is highly dependent on the initial data quality. Future research trends may focus on improving the quality and diversity of generated images, reducing mode collapse during training, and exploring the generalizability of these models across different types of medical images and applications. Additionally, we expect to further enhance the accuracy and efficiency of medical image segmentation by integrating GANs and diffusion models with other advanced deep learning techniques.

3.5. Deep Reinforcement Learning in Ultrasound Image and Video Segmentation

3.5.1. Overview of Works

For the segmentation of the prostate in transrectal ultrasound images, Sahba et al. [76] introduced a novel method that aims to find suitable local values for sub-images and extract the prostate. It comprises an offline phase wherein a reinforcement learning agent is trained on a dataset consisting of multiple images along with their manually segmented counterparts. Wang et al. [77] proposed a medical image segmentation framework based on online reinforcement learning. The model adapts not only to the defined objective functions but also to user intentions and prior knowledge. Detailed validation results in the paper indicate that this model can significantly reduce user interaction while maintaining segmentation accuracy and consistency. Mathews et al. [78] created a new unsupervised reinforcement learning (RL) framework with new rewards to help unsupervised learning. This way, the manual labeling of ultrasound videos does not have to be as time-consuming and impractical, and the framework can be used more easily as a triage tool in emergency departments and for telemedicine. Ning et al. [79] presented a reinforcement learning (RL)-based autonomous robotic ultrasound imaging system. The proposed system and framework were designed to control the US probe, enabling it to perform fully autonomous imaging of soft, mobile, and unlabeled targets, using only a single RGB image of the scene.

3.5.2. Assessments

DRL for medical ultrasound image and video segmentation optimizes segmentation strategies through trial-and-error methods and reward mechanisms. DRL models adapt well to complex imaging scenarios and user-defined goals, improving segmentation accuracy and reducing user interaction. These methods excel at handling local value extraction, context adaptation, and reducing manual annotation efforts. However, challenges remain, including high computational costs, potential instability during training, and the need for large and diverse training datasets. In the future, people will probably focus on making DRL models more reliable and usable across a wider range of ultrasound types, combining them with other cutting-edge methods to improve segmentation performance, and reducing the need for human intervention even more. These advancements will help address current limitations and broaden the applicability of DRL in clinical settings.

3.6. Weakly Supervised Learning in Ultrasound Image and Video Segmentation

3.6.1. Overview of Works

Girum et al. [80] came up with a fast, interactive, weakly supervised deep learning method for medical ultrasound image segmentation to get around the problems with annotated datasets and the time-consuming nature of manual annotation. GANs were used to predict pseudo-contour landmark priors. Then, CNNs were used to improve the predicted priors (i.e., contour suggestions). The experimental results showed that this method achieved a Dice coefficient of 0.97 in colorectal ultrasound image segmentation. Additionally, Li et al. [81] proposed a two-step deep learning framework for breast ultrasound image segmentation. The first step is a semi-supervised semantic segmentation framework that decomposes four anatomical structures: the fat layer, breast layer, muscle layer, and chest cavity layer. The second step is to segment breast tumors at the image-level labels in a weakly supervised learning environment. The Dice coefficients for the four anatomical structures were 83.0 ± 11.8%, 84.3 ± 10.0%, 80.7 ± 15.4%, and 91.0 ± 11.4%, respectively. Regarding breast tumor segmentation, Li et al. [82] proposed a network based on a CNN backbone and spatial pyramid module, achieving a maximum Dice coefficient of 73.5 ± 18.0% in a weakly supervised learning scenario. Table 5 summarizes the data sets, models used, areas of focus, and model performance of the literature reviewed above.

3.6.2. Assessments

Weakly supervised learning has shown a lot of promise in medical ultrasound image and video segmentation by lowering the need for large datasets with lots of annotations while still maintaining strong segmentation performance. Achievements in this area include the development of novel frameworks that facilitate unsupervised learning and reduce the need for tedious manual labeling, thereby enhancing the utility of ultrasound imaging tools in emergency departments and telemedicine. These methods use advanced structures like attentional encoders and bi-directional long short-term memory networks to project high-dimensional images into lower-dimensional latent spaces more efficiently. This makes it easier to classify and divide images. However, challenges persist, including limited generalization to various imaging conditions and the potential need for high-quality initial data to train effective models. Future trends will likely focus on improving weakly supervised techniques to better manage different and changing imaging situations. This will include combining these methods with reinforcement learning to make imaging systems more independent and reduce the need for manual help. This technical convergence between attention-aware feature distillation and uncertainty-quantified pseudo-label generation paves the way for expanding weakly supervised applications into real-time clinical decision support systems, overcoming traditional barriers in annotation fidelity.

3.7. Segment Anything Model in Ultrasound Image and Video Segmentation

3.7.1. Overview of Works

The development of SAM has expanded its applications beyond natural images. Tu et al. [84] proposed the breast ultrasound segment anything model (BUSSAM), which migrates the SAM model to the field of breast ultrasound image segmentation through adapter technology. They achieved more accurate segmentation results on ultrasound image datasets by designing a lightweight CNN image encoder and introducing position and feature adapters to further fine-tune the branch. In the segmentation of tumors and cancerous regions in breast ultrasound images, Guo et al. [85] proposed the ClickSAM model, which fine-tunes the segment anything model using click prompts to improve the accuracy of ultrasound image segmentation. Pandey and others discovered that using YOLO + SAM for breaking down aortic images worked better than regular U-Net models, especially with POCUS images that had a lot of noise and small areas of interest. At the same time, research combining SAM with other models shows that the model is comparable to other deep learning models. Pandey et al. [86] compared the YOLOv8, YOLOv8 + SAM, and YOLOv8 + HQ-SAM models, demonstrating significant advantages in multimodal medical image segmentation. The Dice scores on the short-axis aorta ultrasound dataset were 0.5064, 0.769, and 0.7722, respectively. Zhao et al. [87] suggested a SAM-based weakly supervised framework that does a better job of segmenting nodules in clinical ultrasound images by creating fake labels and using a cross-teaching strategy that takes uncertainty into account. The application of SAM to ultrasound video The segmentation is equally surprising. Deng et al. [88] introduced MemSAM, adapting SAM to echocardiography videos via a spatiotemporal memory that encodes temporal cues and refines features with predicted masks to address speckle noise and ambiguous boundaries. It achieves state-of-the-art performance on two datasets with limited annotations, rivaling fully supervised methods. This work pioneers SAM’s medical video adaptation through noise-resilient, temporally consistent prompting. Kim et al. [89] developed MediViSTA, a parameter-efficient SAM adaptation for echocardiography videos, combining spatial frequency fusion and temporal adapters to enhance 3D medical data processing. It outperforms SOTA methods across three datasets, achieving superior generalization (2.15% Dice) without prompts, advancing cardiac segmentation accuracy and temporal consistency. Pang et al. [90] proposed BaS, a real-time breast lesion segmentation model for ultrasound videos, employing Stem modules and BaSBlock for inter-/intra-frame analysis. With variants BaS-S (accuracy) and BaS-L (speed), it outperforms SOTA in efficiency and precision on resource-limited devices, enabling practical clinical deployment.

3.7.2. Assessments

The adaptation of the SAM to ultrasound image and video segmentation has demonstrated notable progress but also highlighted critical challenges. Recent studies reveal SAM’s adaptability through lightweight architectural modifications, such as BUSSAM’s CNN encoder and position adapters and MediViSTA’s spatiotemporal adapters, enabling the efficient processing of noisy, low-contrast ultrasound data while maintaining parameter efficiency. Innovations like ClickSAM’s click-based prompts and MemSAM’s temporal-aware memory mechanisms further enhance segmentation accuracy by addressing speckle noise and ambiguous boundaries, achieving performance comparable to fully supervised methods with minimal annotations. Weakly supervised frameworks, such as Zhao et al.’s pseudo-labeling with uncertainty-aware cross-teaching, underscore SAM’s potential to reduce annotation dependency. However, challenges persist: SAM’s inherent 2D architecture necessitates complex 3D adaptations (e.g., MediViSTA’s temporal modules) for video analysis, increasing computational overhead. Multi-stage pipelines like YOLO + SAM, though effective for small regions in noisy POCUS images, suffer from high inference costs, limiting real-time deployment. Additionally, prompt-dependent methods (e.g., ClickSAM) hinder fully automated workflows. To bridge these gaps, recent efforts prioritize efficiency and scalability: BaS introduces inter-/intra-frame modules for real-time lesion segmentation on edge devices, while MemSAM leverages temporal consistency to minimize error propagation. Future directions should focus on 3D-native SAM extensions for volumetric ultrasound, self-supervised training to eliminate prompts, and lightweight architectures balancing accuracy and speed. Collectively, these advancements position SAM as a versatile foundation for medical imaging, yet its clinical translation demands further optimization in computational efficiency and dimensional adaptability.

3.8. Datasets

The implementation of deep learning relies on large-scale data, which necessitates a high quantity and quality of medical images in the field of medical imaging deep learning research. To ensure that data are suitable for target tasks, it is often required to collect medical images from multiple medical centers and different types of imaging equipment, thereby guaranteeing that the trained deep learning models possess excellent generalization performance. Additionally, obtaining large-scale medical images in clinical settings is challenging and usually requires manual annotation by skilled clinicians, which demands considerable time and effort. For image segmentation, high-quality datasets are crucial for achieving more precise segmentation. Therefore, high-quality public datasets are of paramount importance. This section will introduce some high-quality datasets of ultrasound images and videos. We integrated the statistical dataset into Table 6, which describes the corresponding parts of the dataset, the quantities, and the URLs.

4. Evaluation Metrics

This chapter will mainly introduce a summary of experimental results indexes based on ultrasonic image and video segmentation methods so that readers can understand the meaning of each index. The performance of various indicators clearly measure the experimental results of each research method. Therefore, it is very important to introduce the common evaluation indicators.

4.1. Medical Image Segmentation Evaluation Methods

Good segmentation results require effective evaluation methods to assess the segmentation results, enabling the differentiation of segmentation algorithms or the adjustment of algorithm parameters based on the evaluation results. More importantly, only after evaluating the algorithms to select the optimal segmented images can the next step of processing be carried out. With the development of deep learning models, clinicians, particularly those in radiology and pathology, strive to integrate deep learning-based medical image segmentation methods into their clinical routines as Clinical Decision Support (CDS) systems to aid in diagnosis, treatment, risk assessment, and the reduction of time-consuming examination processes. Given the direct impact on diagnostic and treatment decisions, the correct and robust evaluation of algorithms is crucial. Therefore, this section reviews common performance evaluation methods for ultrasound (BUSI) image segmentation, with the metrics used referenced in Table 7. Table 8 shows the indicator performance of different technical paths in the same Breast Ultrasound Images Dataset. We explain the indicators so that readers can understand the meaning of each indicator more clearly.

4.1.1. Dice Similarity Coefficient

A segmentation algorithm uses DSC, a symmetric similarity metric, to measure the overlap between its results and the ground truth. Most publications regard DSC as the primary metric for validation and performance interpretation in medical image segmentation.
D S C = 2 · T P 2 · T P + F P + F N

4.1.2. Jaccard Similarity Index

Medical ultrasound imaging uses the Jaccard similarity index (JSI) to assess the overlap between automated and manual segmentation results. It measures the accuracy and consistency of the segmentation algorithm by calculating the ratio of the intersection to the union of the two segmentation results. A high JSI value indicates a high degree of agreement between the automated and manual segmentation results, and vice versa.
J S I = T P T P + F P + F N

4.1.3. Hausdorff Distance

Medical ultrasound imaging uses the Hausdorff Distance to evaluate the shape similarity between two segmentation results. It quantifies the boundary differences of the segmentation results by measuring the distance between the farthest points of two point sets. This is crucial for evaluating the accuracy of automated segmentation algorithms, as it can detect extreme deviations at the boundaries.
H ( A , B ) = max sup a A inf b B d ( a , b ) , sup b B inf a A d ( b , a )

4.1.4. Precision

Since segmentation can be considered a classification problem for each pixel, Precision and Recall are also used as metrics. They can reflect the false positive and false negative pixels in the segmentation results.
P r e c i s i o n = T P T P + F P

4.1.5. Accuracy

Accuracy in ultrasound image segmentation refers to the percentage of pixels correctly classified as either the target region (e.g., lesion area) or the background.
A c c u r a c y = T P + T N T P + T N + F P + F N

4.1.6. Recall

Recall in ultrasound image segmentation refers to the percentage of pixels accurately predicted as part of the target region. A high Recall indicates that the model has a strong ability to recognize the target region, with few false negatives (target region pixels incorrectly identified as background).
R e c a l l l = T P T P + F N

4.1.7. F1-Score

In ultrasound image segmentation, the F1-score synthesizes the performance of both Precision and Recall, serving as a balanced metric between the two. It is particularly useful when there is a need to simultaneously consider the accuracy and completeness of the model. The F1-score provides a more comprehensive evaluation metric under these circumstances.
F 1 = 2 · P r e c i s i o n · R e c a l l P r e c i s i o n + R e c a l l

5. Discussion and Conclusions

5.1. Discussion

In recent years, medical imaging, encompassing ultrasound images and videos, has seen the widespread application of deep learning techniques. We categorized the typical methods into four major classes, providing an introduction and a brief summary of related work. Based on a summary of existing research achievements, this paper outlined the issues, challenges, and potential research directions in the field. In the segmentation of medical ultrasound images and videos, convolutional neural network models such as U-Net, RNN, and FCN each have their own advantages and disadvantages. U-Net, with its symmetric encoder–decoder structure and skip connections, effectively captures multi-scale features and performs well in handling complex medical images. However, its high computational resource demands and dependence on large amounts of high-quality annotated data are significant drawbacks. While RNNs demonstrate significant advantages in processing ultrasound image sequences and capturing temporal information through mechanisms like long short-term memory, their application in static image segmentation remains limited. FCNs employ a unique architecture that enables them to predict image details at the pixel level throughout the entire process. They work well with large images, but they struggle a bit with picking up small features. Future trends in the development of these models include the integration of multi-modal data, the introduction of attention mechanisms, and improvements in model lightweight design to enhance segmentation performance and computational efficiency, better meeting clinical application needs. In clinical medical imaging, there are various imaging modalities, such as MRI, X-rays, PET, and CT. We employ different diagnostic imaging methods based on different cases. Ultrasound imaging is considered the preferred method due to its low cost and lack of radiation. MRI or CT, while providing rich texture information, are time-consuming and expensive. To address the processing challenges posed by multiple resources, cross-channel transfer learning is a promising research direction.
In the field of medical ultrasound image processing, the quality and quantity of datasets are crucial. High-quality annotated datasets can significantly improve the training efficiency and segmentation accuracy of deep learning models. However, annotating ultrasound images often requires specialized knowledge, making the manual annotation process time-consuming and prone to bias. The sensitivity and privacy concerns of medical data also pose challenges in obtaining large annotated datasets. Therefore, efficiently utilizing limited annotated data and enhancing the generalization ability of models are current research focuses. Generative models, especially diffusion models, have shown enormous potential in ultrasound image segmentation. Diffusion models, by gradually adding noise and learning the reverse process to generate high-quality image data, can generate realistic medical images under unsupervised or weakly supervised conditions, aiding model training. The introduction of generative models, such as diffusion models, not only enriches the methods for ultrasound image processing but also provides new solutions for addressing the lack of annotated data and improving model performance. In medical ultrasound images, various pathological changes exist, which can be minute and difficult to detect. SAM needs to have high sensitivity and specificity to detect and segment these pathological areas. By introducing reinforcement learning and adaptive training methods, SAM can continuously learn and improve its segmentation performance in dynamic and real-time environments. This can enhance the model’s applicability in various clinical scenarios. Future research can further explore the performance and optimization strategies of generative models in different ultrasound imaging application scenarios, thereby promoting the development of medical image processing technology.

5.2. Application

Recent advancements in deep learning have spurred the translation of research algorithms into commercially viable segmentation solutions across diverse clinical imaging modalities. For instance, GE Healthcare’s innovative project, SonoSAMTrack, leverages NVIDIA-powered deep learning to accurately delineate anatomical structures and lesions in ultrasound images, demonstrating robust performance on datasets ranging from adult cardiac to fetal head imaging, as well as musculoskeletal pathologies. In parallel, GE’s LOGIQ E20 ultrasound system integrates a dual-engine design that couples advanced signal processing with AI-based lesion recognition, streamlining segmentation workflows and reducing the need for extensive manual annotation. Philips and Siemens have similarly incorporated deep learning into their platforms: Philips’ ISAI platform and Siemens’ AI-Rad Companion suite both exemplify how automated segmentation is being embedded into routine clinical practice to improve diagnostic accuracy and workflow efficiency. Moreover, several emerging startups (e.g., Koios Medical and DiA Imaging) have introduced segmentation solutions that are optimized for minimal training data and high adaptability across different scanning protocols and devices, further reinforcing the trend toward clinically integrated, AI-driven tools. Collectively, these initiatives not only underscore the technological maturity of deep learning in segmentation but also illustrate a paradigm shift from isolated research prototypes to integrated, multi-vendor solutions that are now being validated in real-world clinical environments and have gained regulatory clearances, thereby paving the way for broader adoption in precision diagnostics and treatment planning. Moreover, they facilitate the detection of small lesions in breast or thyroid ultrasound, reducing operator dependence and improving early detection rates. Alongside these commercial developments, open-source toolkits and large-scale public datasets (e.g., CAMUS for cardiac ultrasound, BUSI for breast lesions) empower researchers and innovators to refine algorithms, while international AI competitions (e.g., at MICCAI) continue to push the envelope for real-time and 3D/4D segmentation capabilities. Efforts to address data variability, privacy concerns, and regulatory standards—via federated learning, rigorous clinical validation, and adherence to guidelines—are further accelerating the transition of deep learning–driven ultrasound segmentation from research labs to everyday clinical practice.

5.3. Challenges and Opportunities

Ultrasound image and video segmentation using deep learning faces several significant challenges. First, the inherent characteristics of ultrasound data—such as speckle noise, low contrast, and blurry boundaries—make accurate segmentation difficult. Second, the variability in image quality across different devices and operators introduces inconsistencies that hinder model generalization. Additionally, obtaining high-quality, pixel-level annotations for training is time-consuming and requires expert knowledge, which limits the availability of large annotated datasets. Furthermore, real-time segmentation in video sequences demands that models that are not only accurate but also computationally efficient. Addressing these challenges is crucial for the successful deployment of deep learning models in clinical ultrasound applications. This section presents solutions to these challenges under different learning paradigms.

5.4. Conclusions

In our review, we systematically collected and analyzed nearly 90 articles focusing on deep learning-based methods for ultrasound image and video segmentation. The results from these studies indicate that deep learning models—especially those based on UNet and its variants—consistently outperform traditional segmentation methods. For instance, many of the selected studies reported an average Dice similarity coefficient of approximately 0.85, which is an improvement of over 15% compared to conventional approaches. Furthermore, evaluation metrics such as Intersection over Union (IoU) and sensitivity were generally higher in deep learning models, demonstrating their robustness in handling the inherent noise and variability in ultrasound imaging.
Additionally, our analysis revealed that while deep learning techniques show promising performance, challenges remain in terms of model generalization, especially when applied to multi-center datasets with varying imaging protocols. Several studies noted that the performance drop on external validation datasets highlights the need for standardized data acquisition and preprocessing methods. Moreover, the integration of these algorithms into clinical workflows—although promising—requires further validation to ensure that automated segmentation outputs are reliable enough for clinical decision-making.
These findings not only substantiate our conclusions regarding the significant performance advantages of deep learning in ultrasound segmentation but also underline critical areas for future research. Standardizing evaluation metrics and expanding the availability of high-quality annotated datasets will be essential steps toward improving model generalizability and clinical applicability.

Author Contributions

X.X.: Manuscript draft, search strategy, literature review, study selection, and data extraction; J.Z.: Literature review, study selection, and data extraction; Y.S., J.L., C.H. and K.S.: Literature review and figure design; D.K.: Conception and design of the study and review of final draft. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (grant nos. 12301676 and 12371429), Zhejiang Provincial Natural Science Foundation of China (grant no. LQ24A010018), the Jinhua Key Science and Technology Program Project (grant no. 2024-3-031), and the R&D project of Pazhou Lab (Huangpu) (grant 2023K0605).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Acknowledgments

The authors would like to thank the editors and anonymous reviewers for their valuable comments.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
SAMSegment Anything Model
FCNFully Convolutional Neural Network
RNNRecurrent Neural Network
GANGenerative Adversarial Network
LSTMLong Short Term Memory
GRUGated Recurrent Unit
VAEsVariational Autoencoders
DSCDice Similarity Coefficient
JSIJaccard Similarity Index
CADComputer-Aided Diagnosis
HDHausdorff Distance
DRLDeep Reinforcement Learning
WSLWeakly Supervised Learning
DMDiffusion Model
NLPNatural Language Processing

References

  1. Suganyadevi, S.; Seethalakshmi, V.; Balasamy, K. A review on deep learning in medical image analysis. Int. J. Multimed. Inf. Retr. 2022, 11, 19–38. [Google Scholar] [CrossRef] [PubMed]
  2. Chan, H.P.; Samala, R.K.; Hadjiiski, L.M.; Zhou, C. Deep Learning in Medical Image Analysis. In Deep Learning in Medical Image Analysis; Advances in Experimental Medicine and Biology; Lee, G., Fujita, H., Eds.; Springer International Publishing: Cham, Switzerland, 2020; Volume 1213, pp. 3–21. [Google Scholar] [CrossRef]
  3. Kareem, F.Q.; Abdulazeez, A.M. Ultrasound medical images classification based on deep learning algorithms: A review. Fusion Pract. Appl. 2021, 3, 29–42. [Google Scholar]
  4. Amjoud, A.B.; Amrouch, M. Object detection using deep learning, CNNs and vision transformers: A review. IEEE Access 2023, 11, 35479–35516. [Google Scholar] [CrossRef]
  5. Liu, X.; Gao, K.; Liu, B.; Pan, C.; Liang, K.; Yan, L.; Ma, J.; He, F.; Zhang, S.; Pan, S.; et al. Advances in Deep Learning-Based Medical Image Analysis. Health Data Sci. 2021, 2021, 8786793. [Google Scholar] [CrossRef]
  6. Wells, P.N. Ultrasound imaging. Phys. Med. Biol. 2006, 51, R83. [Google Scholar] [CrossRef]
  7. Almajalid, R.; Shan, J.; Du, Y.; Zhang, M. Development of a deep-learning-based method for breast ultrasound image segmentation. In Proceedings of the 2018 17th IEEE International Conference on Machine Learning and Applications (ICMLA), Orlando, FL, USA, 17–20 December 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1103–1108. [Google Scholar]
  8. Huang, Q.; Luo, Y.; Zhang, Q. Breast ultrasound image segmentation: A survey. Int. J. Comput. Assist. Radiol. Surg. 2017, 12, 493–507. [Google Scholar] [CrossRef]
  9. Ladak, H.M.; Mao, F.; Wang, Y.; Downey, D.B.; Steinman, D.A.; Fenster, A. Prostate boundary segmentation from 2D ultrasound images. Med. Phys. 2000, 27, 1777–1788. [Google Scholar] [CrossRef]
  10. Yu, H.; Ding, M.; Zhang, X.; Wu, J. PCANet based nonlocal means method for speckle noise removal in ultrasound images. PLoS ONE 2018, 13, e0205390. [Google Scholar] [CrossRef]
  11. Liu, B.; Cheng, H.D.; Huang, J.; Tian, J.; Tang, X.; Liu, J. Probability density difference-based active contour for ultrasound image segmentation. Pattern Recognit. 2010, 43, 2028–2042. [Google Scholar] [CrossRef]
  12. Zhang, M.; Dong, B.; Li, Q. Deep active contour network for medical image segmentation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention–MICCAI 2020: 23rd International Conference, Lima, Peru, 4–8 October 2020; Proceedings, Part IV 23. Springer: Cham, Switzerland, 2020; pp. 321–331. [Google Scholar]
  13. Archip, N.; Rohling, R.; Cooperberg, P.; Tahmasebpour, H. Ultrasound image segmentation using spectral clustering. Ultrasound Med. Biol. 2005, 31, 1485–1497. [Google Scholar] [CrossRef]
  14. Xie, J.; Jiang, Y.; Tsui, H.-t. Segmentation of kidney from ultrasound images based on texture and shape priors. IEEE Trans. Med. Imaging 2005, 24, 45–57. [Google Scholar] [PubMed]
  15. Shen, D.; Zhan, Y.; Davatzikos, C. Segmentation of prostate boundaries from ultrasound images using statistical shape model. IEEE Trans. Med. Imaging 2003, 22, 539–551. [Google Scholar] [CrossRef] [PubMed]
  16. Chang, C.Y.; Lei, Y.F.; Tseng, C.H.; Shih, S.R. Thyroid segmentation and volume estimation in ultrasound images. IEEE Trans. Biomed. Eng. 2010, 57, 1348–1357. [Google Scholar] [CrossRef] [PubMed]
  17. Sarti, A.; Corsi, C.; Mazzini, E.; Lamberti, C. Maximum likelihood segmentation of ultrasound images with Rayleigh distribution. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 2005, 52, 947–960. [Google Scholar] [CrossRef]
  18. Brusseau, E.; De Korte, C.L.; Mastik, F.; Schaar, J.; Van Der Steen, A.F. Fully automatic luminal contour segmentation in intracoronary ultrasound imaging—A statistical approach. IEEE Trans. Med. Imaging 2004, 23, 554–566. [Google Scholar] [CrossRef]
  19. Ker, J.; Wang, L.; Rao, J.; Lim, T. Deep learning applications in medical image analysis. IEEE Access 2017, 6, 9375–9389. [Google Scholar] [CrossRef]
  20. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  21. Pan, D.; Zhang, M.; Zhang, B. A generic FCN-based approach for the road-network extraction from VHR remote sensing images–using openstreetmap as benchmarks. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 2662–2673. [Google Scholar] [CrossRef]
  22. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III 18. Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  23. Guo, Y.; Duan, X.; Wang, C.; Guo, H. Segmentation and recognition of breast ultrasound images based on an expanded U-Net. PLoS ONE 2021, 16, e0253202. [Google Scholar] [CrossRef]
  24. Diakogiannis, F.I.; Waldner, F.; Caccetta, P.; Wu, C. ResUNet-a: A deep learning framework for semantic segmentation of remotely sensed data. ISPRS J. Photogramm. Remote Sens. 2020, 162, 94–114. [Google Scholar] [CrossRef]
  25. Cao, Y.; Liu, S.; Peng, Y.; Li, J. DenseUNet: Densely connected UNet for electron microscopy image segmentation. IET Image Process. 2020, 14, 2682–2689. [Google Scholar] [CrossRef]
  26. Li, J.; Guan, W. Patch merging refiner embedding UNet for image denoising. Inf. Sci. 2023, 641, 119123. [Google Scholar]
  27. Cao, C.; Cao, L.; Li, G.; Zhang, T.; Gao, X. BIRGU Net: Deformable brain magnetic resonance image registration using gyral-net map and 3D Res-Unet. Med. Biol. Eng. Comput. 2023, 61, 579–592. [Google Scholar] [PubMed]
  28. Zargar, S. Introduction to Sequence Learning Models: RNN, LSTM, GRU; Department of Mechanical and Aerospace Engineering, North Carolina State University: Raleigh, NC, USA, 2021. [Google Scholar]
  29. Basiri, M.E.; Nemati, S.; Abdar, M.; Cambria, E.; Acharya, U.R. ABCDM: An attention-based bidirectional CNN-RNN deep model for sentiment analysis. Future Gener. Comput. Syst. 2021, 115, 279–294. [Google Scholar]
  30. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; Volume 27. [Google Scholar]
  31. Liu, R.; Zhou, S.; Guo, Y.; Wang, Y.; Chang, C. U2F-GAN: Weakly supervised super-pixel segmentation in thyroid ultrasound images. Cogn. Comput. 2021, 13, 1099–1113. [Google Scholar] [CrossRef]
  32. Zhu, Z.; Lin, K.; Jain, A.K.; Zhou, J. Transfer learning in deep reinforcement learning: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 13344–13362. [Google Scholar]
  33. Botvinick, M.; Wang, J.X.; Dabney, W.; Miller, K.J.; Kurth-Nelson, Z. Deep reinforcement learning and its neuroscientific implications. Neuron 2020, 107, 603–616. [Google Scholar]
  34. Wang, H.; Liu, N.; Zhang, Y.; Feng, D.; Huang, F.; Li, D.; Zhang, Y. Deep reinforcement learning: A survey. Front. Inf. Technol. Electron. Eng. 2020, 21, 1726–1744. [Google Scholar]
  35. Shin, S.Y.; Lee, S.; Yun, I.D.; Kim, S.M.; Lee, K.M. Joint weakly and semi-supervised deep learning for localization and classification of masses in breast ultrasound images. IEEE Trans. Med. Imaging 2018, 38, 762–774. [Google Scholar]
  36. Kim, J.; Kim, H.J.; Kim, C.; Lee, J.H.; Kim, K.W.; Park, Y.M.; Kim, H.W.; Ki, S.Y.; Kim, Y.M.; Kim, W.H. Weakly-supervised deep learning for ultrasound diagnosis of breast cancer. Sci. Rep. 2021, 11, 24382. [Google Scholar]
  37. Kwak, S.; Hong, S.; Han, B. Weakly supervised semantic segmentation using superpixel pooling network. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar]
  38. Li, H.; Zhang, B.; Zhang, Y.; Liu, W.; Mao, Y.; Huang, J.; Wei, L. A semi-automated annotation algorithm based on weakly supervised learning for medical images. Biocybern. Biomed. Eng. 2020, 40, 787–802. [Google Scholar] [CrossRef]
  39. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.Y.; et al. Segment anything. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–3 October 2023; pp. 4015–4026. [Google Scholar]
  40. Ashkani Chenarlogh, V.; Ghelich Oghli, M.; Shabanzadeh, A.; Sirjani, N.; Akhavan, A.; Shiri, I.; Arabi, H.; Sanei Taheri, M.; Tarzamni, M.K. Fast and accurate U-net model for fetal ultrasound image segmentation. Ultrason. Imaging 2022, 44, 25–38. [Google Scholar] [PubMed]
  41. Amiri, M.; Brooks, R.; Rivaz, H. Fine-tuning U-Net for ultrasound image segmentation: Different layers, different outcomes. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 2020, 67, 2510–2518. [Google Scholar] [PubMed]
  42. Amiri, M.; Brooks, R.; Behboodi, B.; Rivaz, H. Two-stage ultrasound image segmentation using U-Net and test time augmentation. Int. J. Comput. Assist. Radiol. Surg. 2020, 15, 981–988. [Google Scholar]
  43. Cheng, D.; Lam, E.Y. Transfer learning U-Net deep learning for lung ultrasound segmentation. arXiv 2021, arXiv:2110.02196. [Google Scholar]
  44. Chen, G.; Li, L.; Zhang, J.; Dai, Y. Rethinking the unpretentious U-net for medical ultrasound image segmentation. Pattern Recognit. 2023, 142, 109728. [Google Scholar]
  45. Xu, M.; Ma, Q.; Zhang, H.; Kong, D.; Zeng, T. MEF-UNet: An end-to-end ultrasound image segmentation algorithm based on multi-scale feature extraction and fusion. Comput. Med. Imaging Graph. 2024, 114, 102370. [Google Scholar]
  46. Weng, Y.; Zhou, T.; Li, Y.; Qiu, X. Nas-unet: Neural architecture search for medical image segmentation. IEEE Access 2019, 7, 44247–44257. [Google Scholar]
  47. Li, H.; Fang, J.; Liu, S.; Liang, X.; Yang, X.; Mai, Z.; Van, M.T.; Wang, T.; Chen, Z.; Ni, D. Cr-unet: A composite network for ovary and follicle segmentation in ultrasound images. IEEE J. Biomed. Health Inform. 2019, 24, 974–983. [Google Scholar]
  48. Wang, R.; Zhou, H.; Fu, P.; Shen, H.; Bai, Y. A Multiscale Attentional Unet Model for Automatic Segmentation in Medical Ultrasound Images. Ultrason. Imaging 2023, 45, 159–174. [Google Scholar]
  49. Inan, N.G.; Kocadağlı, O.; Yıldırım, D.; Meşe, İ.; Kovan, Ö. Multi-class classification of thyroid nodules from automatic segmented ultrasound images: Hybrid ResNet based UNet convolutional neural network approach. Comput. Methods Programs Biomed. 2024, 243, 107921. [Google Scholar]
  50. Zhang, Y.; Ying, M.T.; Yang, L.; Ahuja, A.T.; Chen, D.Z. Coarse-to-fine stacked fully convolutional nets for lymph node segmentation in ultrasound images. In Proceedings of the 2016 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Shenzhen, China, 15–18 December 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 443–448. [Google Scholar]
  51. Villa, M.; Dardenne, G.; Nasan, M.; Letissier, H.; Hamitouche, C.; Stindel, E. FCN-based approach for the automatic segmentation of bone surfaces in ultrasound images. Int. J. Comput. Assist. Radiol. Surg. 2018, 13, 1707–1716. [Google Scholar] [PubMed]
  52. Xing, J.; Li, Z.; Wang, B.; Qi, Y.; Yu, B.; Zanjani, F.G.; Zheng, A.; Duits, R.; Tan, T. Lesion segmentation in ultrasound using semi-pixel-wise cycle generative adversarial nets. IEEE/ACM Trans. Comput. Biol. Bioinform. 2020, 18, 2555–2565. [Google Scholar]
  53. Hu, Y.; Guo, Y.; Wang, Y.; Yu, J.; Li, J.; Zhou, S.; Chang, C. Automatic tumor segmentation in breast ultrasound images using a dilated fully convolutional network combined with an active contour model. Med. Phys. 2019, 46, 215–228. [Google Scholar] [PubMed]
  54. Li, Y.; Zhao, H.; Qi, X.; Wang, L.; Li, Z.; Sun, J.; Jia, J. Fully convolutional networks for panoptic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 214–223. [Google Scholar]
  55. Qian, L.; Huang, H.; Xia, X.; Li, Y.; Zhou, X. Automatic segmentation method using FCN with multi-scale dilated convolution for medical ultrasound image. Vis. Comput. 2023, 39, 5953–5969. [Google Scholar]
  56. Feng, Y.; Atabansi, C.C.; Nie, J.; Liu, H.; Zhou, H.; Zhao, H.; Hong, R.; Li, F.; Zhou, X. Multi-stage fully convolutional network for precise prostate segmentation in ultrasound images. Biocybern. Biomed. Eng. 2023, 43, 586–602. [Google Scholar]
  57. Xu, H.; Yang, Y.; Aviles-Rivero, A.I.; Yang, G.; Qin, J.; Zhu, L. LGRNet: Local-Global Reciprocal Network for Uterine Fibroid Segmentation in Ultrasound Videos. arXiv 2024, arXiv:2407.05703. [Google Scholar]
  58. Pan, P.; Chen, H.; Li, Y.; Cai, N.; Cheng, L.; Wang, S. Tumor segmentation in automated whole breast ultrasound using bidirectional LSTM neural network and attention mechanism. Ultrasonics 2021, 110, 106271. [Google Scholar]
  59. van den Noort, F.; Sirmacek, B.; Slump, C.H. Recurrent U-net for automatic pelvic floor muscle segmentation on 3D ultrasound. arXiv 2021, arXiv:2107.13833. [Google Scholar]
  60. Yang, X.; Yu, L.; Wu, L.; Wang, Y.; Ni, D.; Qin, J.; Heng, P.A. Fine-grained recurrent neural networks for automatic prostate segmentation in ultrasound images. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar]
  61. Webb, J.M.; Meixner, D.D.; Adusei, S.A.; Polley, E.C.; Fatemi, M.; Alizad, A. Automatic deep learning semantic segmentation of ultrasound thyroid cineclips using recurrent fully convolutional networks. IEEE Access 2020, 9, 5119–5127. [Google Scholar] [CrossRef]
  62. Horng, M.H.; Yang, C.W.; Sun, Y.N.; Yang, T.H. DeepNerve: A new convolutional neural network for the localization and segmentation of the median nerve in ultrasound image sequences. Ultrasound Med. Biol. 2020, 46, 2439–2452. [Google Scholar]
  63. Devisri, B.; Kavitha, M. Fetal growth analysis from ultrasound videos based on different biometrics using optimal segmentation and hybrid classifier. Stat. Med. 2024, 43, 1019–1047. [Google Scholar] [CrossRef] [PubMed]
  64. Anas, E.M.A.; Mousavi, P.; Abolmaesumi, P. A deep learning approach for real time prostate segmentation in freehand ultrasound guided biopsy. Med. Image Anal. 2018, 48, 107–116. [Google Scholar] [CrossRef] [PubMed]
  65. Khdhir, R.; Othmen, S.; Belghith, A.; Alamri, J. Breast Cancer Segmentation in Mammograms using Antlion Optimization and CNN/GRU Architectures. In Proceedings of the 2024 International Wireless Communications and Mobile Computing (IWCMC), Ayia Napa, Cyprus, 27–31 May 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1030–1035. [Google Scholar]
  66. Van Veelen, G.; Schweitzer, K.; Van der Vaart, C. Reliability of pelvic floor measurements on three-and four-dimensional ultrasound during and after first pregnancy: Implications for training. Ultrasound Obstet. Gynecol. 2013, 42, 590–595. [Google Scholar] [CrossRef] [PubMed]
  67. Liang, J.; Chen, J. Data augmentation of thyroid ultrasound images using generative adversarial network. In Proceedings of the 2021 IEEE International Ultrasonics Symposium (IUS), Xi’an China, 11–16 September 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–4. [Google Scholar]
  68. Liang, J.; Yang, X.; Huang, Y.; Li, H.; He, S.; Hu, X.; Chen, Z.; Xue, W.; Cheng, J.; Ni, D. Sketch guided and progressive growing GAN for realistic and editable ultrasound image synthesis. Med. Image Anal. 2022, 79, 102461. [Google Scholar] [CrossRef]
  69. Bargsten, L.; Schlaefer, A. SpeckleGAN: A generative adversarial network with an adaptive speckle layer to augment limited training data for ultrasound image processing. Int. J. Comput. Assist. Radiol. Surg. 2020, 15, 1427–1436. [Google Scholar] [CrossRef]
  70. Fatima, N.; Afrakhteh, S.; Iacca, G.; Demi, L. Automatic segmentation of 2D echocardiography ultrasound images by means of generative adversarial network. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 2024, 71, 1552–1564. [Google Scholar] [CrossRef]
  71. Tang, F.; Ding, J.; Wang, L.; Xian, M.; Ning, C. Multi-level global context cross consistency model for semi-supervised ultrasound image segmentation with diffusion model. arXiv 2023, arXiv:2305.09447. [Google Scholar]
  72. Stojanovski, D.; Hermida, U.; Lamata, P.; Beqiri, A.; Gomez, A. Echo from noise: Synthetic ultrasound image generation using diffusion models for real image segmentation. In Proceedings of the International Workshop on Advances in Simplifying Medical Ultrasound, Vancouver, BC, Canada, 8 October 2023; Springer: Cham, Switzerland, 2023; pp. 34–43. [Google Scholar]
  73. Katakis, S.; Barotsis, N.; Kakotaritis, A.; Tsiganos, P.; Economou, G.; Panagiotopoulos, E.; Panayiotakis, G. Generation of Musculoskeletal Ultrasound Images with Diffusion Models. BioMedInformatics 2023, 3, 405–421. [Google Scholar] [CrossRef]
  74. Stevens, T.S.; Meral, F.C.; Yu, J.; Apostolakis, I.Z.; Robert, J.L.; Van Sloun, R.J. Dehazing ultrasound using diffusion models. IEEE Trans. Med. Imaging 2024, 43, 3546–3558. [Google Scholar] [CrossRef]
  75. Yao, Y.; Duan, X.; Qu, A.; Chen, M.; Chen, J.; Chen, L. DFCG: A Dual-Frequency Cascade Graph model for semi-supervised ultrasound image segmentation with diffusion model. Knowl.-Based Syst. 2024, 300, 112261. [Google Scholar] [CrossRef]
  76. Sahba, F.; Tizhoosh, H.R.; Salama, M.M. A reinforcement agent for object segmentation in ultrasound images. Expert Syst. Appl. 2008, 35, 772–780. [Google Scholar] [CrossRef]
  77. Wang, L.; Lekadir, K.; Lee, S.L.; Merrifield, R.; Yang, G.Z. A general framework for context-specific image segmentation using reinforcement learning. IEEE Trans. Med. Imaging 2013, 32, 943–956. [Google Scholar] [CrossRef] [PubMed]
  78. Mathews, R.P.; Panicker, M.R.; Hareendranathan, A.R.; Chen, Y.T.; Jaremko, J.L.; Buchanan, B.; Narayan, K.V.; Mathews, G. Unsupervised multi-latent space reinforcement learning framework for video summarization in ultrasound imaging. arXiv 2021, arXiv:2109.01309. [Google Scholar]
  79. Ning, G.; Zhang, X.; Liao, H. Autonomic robotic ultrasound imaging system based on reinforcement learning. IEEE Trans. Biomed. Eng. 2021, 68, 2787–2797. [Google Scholar] [CrossRef]
  80. Girum, K.B.; Créhange, G.; Hussain, R.; Lalande, A. Fast interactive medical image segmentation with weakly supervised deep learning method. Int. J. Comput. Assist. Radiol. Surg. 2020, 15, 1437–1444. [Google Scholar] [CrossRef]
  81. Li, Y.; Liu, Y.; Huang, L.; Wang, Z.; Luo, J. Deep weakly-supervised breast tumor segmentation in ultrasound images with explicit anatomical constraints. Med. Image Anal. 2022, 76, 102315. [Google Scholar] [CrossRef]
  82. Li, Y.; Liu, Y.; Wang, Z.; Luo, J. Weakly-supervised deep learning for breast tumor segmentation in ultrasound images. In Proceedings of the 2021 IEEE International Ultrasonics Symposium (IUS), Xi’an, China, 11–16 September 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–4. [Google Scholar]
  83. Leclerc, S.; Smistad, E.; Pedrosa, J.; Østvik, A.; Cervenansky, F.; Espinosa, F.; Espeland, T.; Berg, E.A.R.; Jodoin, P.M.; Grenier, T.; et al. Deep learning for segmentation using an open large-scale dataset in 2D echocardiography. IEEE Trans. Med. Imaging 2019, 38, 2198–2210. [Google Scholar] [CrossRef]
  84. Tu, Z.; Gu, L.; Wang, X.; Jiang, B. Ultrasound SAM Adapter: Adapting SAM for Breast Lesion Segmentation in Ultrasound Images. arXiv 2024, arXiv:2404.14837. [Google Scholar]
  85. Guo, A.; Fei, G.; Pasupuleti, H.; Wang, J. ClickSAM: Fine-tuning Segment Anything Model using click prompts for ultrasound image segmentation. In Proceedings of the Medical Imaging 2024: Ultrasonic Imaging and Tomography, San Diego, CA, USA, 18–22 February 2024; SPIE: Bellingham, WA, USA, 2024; Volume 12932, pp. 240–244. [Google Scholar]
  86. Pandey, S.; Chen, K.F.; Dam, E.B. Comprehensive multimodal segmentation in medical imaging: Combining yolov8 with sam and hq-sam models. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–3 October 2023; pp. 2592–2598. [Google Scholar]
  87. Paranjape, J.N.; Nair, N.G.; Sikder, S.; Vedula, S.S.; Patel, V.M. Adaptivesam: Towards efficient tuning of sam for surgical scene segmentation. In Proceedings of the Annual Conference on Medical Image Understanding and Analysis, Manchester, UK, 24–26 July 2024; Springer: Cham, Switzerland, 2024; pp. 187–201. [Google Scholar]
  88. Deng, X.; Wu, H.; Zeng, R.; Qin, J. MemSAM: Taming segment anything model for echocardiography video segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 9622–9631. [Google Scholar]
  89. Kim, S.; Jin, P.; Chen, C.; Kim, K.; Lyu, Z.; Ren, H.; Kim, S.; Liu, Z.; Zhong, A.; Liu, T.; et al. MediViSTA: Medical Video Segmentation via Temporal Fusion SAM Adaptation for Echocardiography. IEEE J. Biomed. Health Inform. 2025. [Google Scholar] [CrossRef]
  90. Pang, Y.; Li, Y.; Huang, T.; Liang, J.; Ding, Z.; Chen, H.; Zhao, B.; Hu, Y.; Zhang, Z.; Wang, Q. Efficient Breast Lesion Segmentation from Ultrasound Videos Across Multiple Source-limited Platforms. IEEE J. Biomed. Health Inform. 2025. [Google Scholar] [CrossRef]
  91. Badrinarayanan, V.; Kendall, A.; SegNet, R.C. A deep convolutional encoder-decoder architecture for image segmentation. arXiv 2022, 1511. [Google Scholar] [CrossRef] [PubMed]
  92. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  93. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support: 4th International Workshop, DLMIA 2018, and 8th International Workshop, ML-CDS 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, 20 September 2018; Proceedings 4. Springer: Cham, Switzerland, 2018; pp. 3–11. [Google Scholar]
  94. Fan, D.P.; Ji, G.P.; Zhou, T.; Chen, G.; Fu, H.; Shen, J.; Shao, L. Pranet: Parallel reverse attention network for polyp segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Lima, Peru, 4–8 October 2020; Springer: Cham, Switzerland, 2020; pp. 263–273. [Google Scholar]
  95. Wang, K.; Liang, S.; Zhang, Y. Residual feedback network for breast lesion segmentation in ultrasound image. In Proceedings of the Medical Image Computing and Computer Assisted Intervention–MICCAI 2021: 24th International Conference, Strasbourg, France, 27 September–1 October 2021; Proceedings, Part I 24. Springer: Cham, Switzerland, 2021; pp. 471–481. [Google Scholar]
  96. Tomar, N.K.; Shergill, A.; Rieders, B.; Bagci, U.; Jha, D. TransResU-Net: Transformer based ResU-Net for real-time colonoscopy polyp segmentation. arXiv 2022, arXiv:2206.08985. [Google Scholar]
  97. Lin, X.; Xiang, Y.; Zhang, L.; Yang, X.; Yan, Z.; Yu, L. SAMUS: Adapting segment anything model for clinically-friendly and generalizable ultrasound image segmentation. arXiv 2023, arXiv:2309.06824. [Google Scholar]
Figure 3. FCN’s resolution-preserving fully convolutional layers.
Figure 3. FCN’s resolution-preserving fully convolutional layers.
Sensors 25 02361 g003
Figure 4. U-Net’s skip-connected encoder–decoder for medical segmentation.
Figure 4. U-Net’s skip-connected encoder–decoder for medical segmentation.
Sensors 25 02361 g004
Figure 5. GAN’s adversarial generator–discriminator framework.
Figure 5. GAN’s adversarial generator–discriminator framework.
Sensors 25 02361 g005
Figure 6. RL’s reward-driven policy optimization mechanism.
Figure 6. RL’s reward-driven policy optimization mechanism.
Sensors 25 02361 g006
Figure 7. Common weakly supervised learning frameworks.
Figure 7. Common weakly supervised learning frameworks.
Sensors 25 02361 g007
Figure 8. SAM’s regular network framework.
Figure 8. SAM’s regular network framework.
Sensors 25 02361 g008
Figure 9. Study selection using PRISMA flowcharts.
Figure 9. Study selection using PRISMA flowcharts.
Sensors 25 02361 g009
Table 1. Overview of segmentation of ultrasound images and videos with U-Net.
Table 1. Overview of segmentation of ultrasound images and videos with U-Net.
ReferencesROIModelDatasetMetric
Ashkani et al. [40]Fetal abdomenFast and Accurate U-NetMFP datasetDice: 0.9714
AMiri et al. [41]BreastFine-tuning U-Net10,000 imagesDice: 0.8 ± 0.03
Cheng et al. [43]LungTL U-NetLUS datasetDice: 0.8
Amiri et al. [42]BreastTwo-stage U-Net163 B-mode US imagesDice: 0.805
Chen et al. [44]BreastNU-netBUSI and STU datasetsDice: 0.9405 ± 0.0066
Xu et al. [45]Breast and thyroidMEF-UNetBUSI, DDTL and BUSDice: 0.7276
Weng et al. [46]ProstateNAS-UnetPromise12Dice: 0.9737
Li et al. [47]TransvaginalCR-Unet3024 TVUSDice: 0.912
Wang et al. [48]ThyroidMSAC-UnetTND-PUH3Dice: 0.822
Inan et al. [49]Thyroid noduleResUNet++Data obtained from Acıbadem Hospital, Istanbul, TurkeyDice: 0.924
Table 2. Overview of segmentation of ultrasound images and videos with FCN.
Table 2. Overview of segmentation of ultrasound images and videos with FCN.
ReferencesROIModelDatasetMetric
Zhang et al. [50]Lymph nodesCFS-FCN80 ultrasound imagesF1 score: 0.858
Villa et al. [51]BoneFCN1738 US imagesRecall: 0.87
Xing et al. [52]BreastSPCGAN670 ultrasound imagesDice: 0.93
Hu et al. [53]BreastDFCN + PBAC570 BUS imagesDice: 0.8897
Li et al. [54]Fetal headFCNCOCO datasetDice: 96.96 ± 2.43%
Qian et al. [55]BreastFCNBreast ultrasound datasetDice: 0.9224
Feng et al. [56]ProstateMulti-stage FCNCCH-TRUSPSDice: 0.9490
Xu et al. [57]UterusLGRNet100 videos, each with 50 framesDice: 0.775
Table 3. Overview of segmentation of ultrasound images and videos with LSTM and GRUs.
Table 3. Overview of segmentation of ultrasound images and videos with LSTM and GRUs.
ReferencesROIModelDatasetMetric
Pan et al. [58]BreastBidirectional LSTMABUS image datasetDice: 0.8178
Noort et al. [59]Levator ani muscleCLSTM U-netAcquired by [66]Dice: 0.7
Yang et al. [60]ProstateRNN with shape priors300 prostate ultrasound imagesDSC: 0.9239
Webb et al. [61]ThyroidDeepLabv3 + CLSTM198 cineclipsRecall: 0.90
Horng et al. [62]Median nerveU-Net+ LSTM420 frames of imagesDice: 0.8975
Anas et al. [64]ProstateCNN + GRU29,394 imagesDice: 0.93
Devisri et al. [63]FetalLSTM + U-NetHC18Not provided
Khdhir et al. [65]BreastGLCM + GRUWisconsin Diagnostic Breast CancerDice: 0.69
Table 4. Overview of generating adversarial networks and diffusion models in ultrasound image and video segmentation.
Table 4. Overview of generating adversarial networks and diffusion models in ultrasound image and video segmentation.
ReferencesROIModelMetric
Liang et al. [67]ThyroidAutoencoder-GAN + VAEDice: 0.5111
Liang et al. [68]Lung, hip joint, and ovaryGAN + U-NetDice: 0.8876
Bargsten et al. [69]Artery layer and lumenSpeckleGAN + U-NetDice: 86.02 ± 0.44%
Fatima et al. [70]CardiologyPatchGAN + U-NetDice: 0.961
Tang et al. [71]BreastMGCC + LDMF1 score: 62.53 ± 2.99%
Stojanovski et al. [72]HeartDenoising diffusion modelDice: 88.6 ± 4.91%
Katakis et al. [73]MusculoskeletalDiffusion modelDice: 0.80
Stevens et al. [74]HeartJoint posterior sampling FrameworkNot provided
Yao et al. [75]BreastDFCGDice: 0.7920
Table 5. Overview of weakly supervised deep learning methods and reinforcement learning for ultrasound image and video segmentation.
Table 5. Overview of weakly supervised deep learning methods and reinforcement learning for ultrasound image and video segmentation.
ReferencesROIModelDatasetMetric
Girum et al. [80]ColorectalRL + CNNAcquired by [83]Dice: 0.97
Li et al. [81]BreastWeakly supervised learning1389 BUS imagesDice: 83.0 ± 11.8%
Li et al. [82]BreastCNN + spatial pyramid module2805 BUS imagesDice: 73.5 ± 18.0%
Sahba et al. [76]ProstateRL agent for sub-images30 TRUS imagesArea overlap: 90.65
Mathews et al. [78]LungUnsupervised RL frameworkLUS datasetF1 score: over 44 ± 1.7%
Table 6. Datasets of ultrasound images and videos.
Table 6. Datasets of ultrasound images and videos.
DatasetROINumberURL
Breast Ultrasound Images DatasetBreast780https://www.kaggle.com/datasets/aryashah2k/breast-ultrasound-images-dataset, accessed on 3 January 2025.
DDTI: Thyroid Ultrasound ImagesThyroid134https://www.kaggle.com/datasets/dasmehdixtr/ddti-thyroid-ultrasound-images, accessed on 3 January 2025.
US Simulation and SegmentationAbdomenNot providedhttps://www.kaggle.com/datasets/ignaciorlando/ussimandsegm, accessed on 3 January 2025.
CAMUS-Human Heart DataHeartNot providedhttps://www.kaggle.com/datasets/shoybhasan/camus-human-heart-data, accessed on 3 January 2025.
Fetal Health ClassificationFetal2126https://www.kaggle.com/datasets/andrewmvd/fetal-health-classification, accessed on 3 January 2025.
A New Dataset and A Baseline Model for Breast Lesion Detection in Ultrasound VideosBreast404 frameshttps://github.com/jhl-Det/CVA-Net, accessed on 3 January 2025.
Mus-V: Multimodal Ultrasound Vascular SegmentationNeck105 videoshttps://www.kaggle.com/datasets/among22/multimodal-ultrasound-vascular-segmentation, accessed on 3 January 2025.
Table 7. Definition of the indexes.
Table 7. Definition of the indexes.
CategoryPositiveNegative
Actual PositiveTrue Positive (TP)True Negative (TN)
Actual NegativeFalse Positive (FP)False Negative (FN)
Table 8. Quantitative evaluation results on BUSI dataset. ↑ means that higher is better and ↓ means that lower is better. The optimal outcomes are denoted using bold typeface.
Table 8. Quantitative evaluation results on BUSI dataset. ↑ means that higher is better and ↓ means that lower is better. The optimal outcomes are denoted using bold typeface.
MethodAcc (%) ↑Se (%) ↑Dice (%) ↑IoU (%) ↑HD (mm) ↓
U-Net (Ronneberger et al., 2015 [22])97.6786.1880.3870.806.68
SegNet (Badrinarayanan et al., 2017 [91])98.9387.3482.9072.716.73
DeepLabV3+ (Chen et al., 2018 [92])96.5284.5179.2568.907.17
U-Net++ (Zhou et al., 2018 [93])98.1987.6083.5673.366.43
PraNet (Fan et al., 2020 [94])98.9086.8283.0072.836.85
RF-Net (Wang et al., 2021 [95])98.1486.6383.2773.096.68
TransResUnet (Tomar et al., 2022 [96])98.9587.3583.7474.626.38
SAMUS (Lin et al., 2023 [97])99.2988.5285.8976.366.16
BUSSAM (Tu et al., 2024 [84])99.3289.1686.5977.216.14
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xiao, X.; Zhang, J.; Shao, Y.; Liu, J.; Shi, K.; He, C.; Kong, D. Deep Learning-Based Medical Ultrasound Image and Video Segmentation Methods: Overview, Frontiers, and Challenges. Sensors 2025, 25, 2361. https://doi.org/10.3390/s25082361

AMA Style

Xiao X, Zhang J, Shao Y, Liu J, Shi K, He C, Kong D. Deep Learning-Based Medical Ultrasound Image and Video Segmentation Methods: Overview, Frontiers, and Challenges. Sensors. 2025; 25(8):2361. https://doi.org/10.3390/s25082361

Chicago/Turabian Style

Xiao, Xiaolong, Jianfeng Zhang, Yuan Shao, Jialong Liu, Kaibing Shi, Chunlei He, and Dexing Kong. 2025. "Deep Learning-Based Medical Ultrasound Image and Video Segmentation Methods: Overview, Frontiers, and Challenges" Sensors 25, no. 8: 2361. https://doi.org/10.3390/s25082361

APA Style

Xiao, X., Zhang, J., Shao, Y., Liu, J., Shi, K., He, C., & Kong, D. (2025). Deep Learning-Based Medical Ultrasound Image and Video Segmentation Methods: Overview, Frontiers, and Challenges. Sensors, 25(8), 2361. https://doi.org/10.3390/s25082361

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop