Next Article in Journal
AI and IoT-Driven Monitoring and Visualisation for Optimising MSP Operations in Multi-Tenant Networks: A Modular Approach Using Sensor Data Integration
Previous Article in Journal
Real-Time Detection Sensor for Unmanned Aerial Vehicle Using an Improved YOLOv8s Algorithm
Previous Article in Special Issue
Grid Search and Genetic Algorithm Optimization of Neural Networks for Automotive Radar Object Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Images Versus Videos in Contrast-Enhanced Ultrasound for Computer-Aided Diagnosis

by
Marina Adriana Mercioni
1,2,*,
Cătălin Daniel Căleanu
1,* and
Mihai-Eronim-Octavian Ursan
1
1
Faculty of Electronics, Telecommunications and Information Technologies, Politehnica University Timisoara, Vasile Parvan Street, No. 2, 300223 Timisoara, Romania
2
Faculty of General Medicine, Victor Babes University of Medicine and Pharmacy Timisoara, Eftimie Murgu Square 2, 300041 Timisoara, Romania
*
Authors to whom correspondence should be addressed.
Sensors 2025, 25(19), 6247; https://doi.org/10.3390/s25196247
Submission received: 3 September 2025 / Revised: 3 October 2025 / Accepted: 7 October 2025 / Published: 9 October 2025

Abstract

The background of the article refers to the diagnosis of focal liver lesions (FLLs) through contrast-enhanced ultrasound (CEUS) based on the integration of spatial and temporal information. Traditional computer-aided diagnosis (CAD) systems predominantly rely on static images, which limits the characterization of lesion dynamics. This study aims to assess the effectiveness of Transformer-based architectures in enhancing CAD performance within the realm of liver pathology. The methodology involved a systematic comparison of deep learning models for the analysis of CEUS images and videos. For image-based classification, a Hybrid Transformer Neural Network (HTNN) was employed. It combines Vision Transformer (ViT) modules with lightweight convolutional features. For video-based tasks, we evaluated a custom spatio-temporal Convolutional Neural Network (CNN), a CNN with Long Short-Term Memory (LSTM), and a Video Vision Transformer (ViViT). The experimental results show that the HTNN achieved an outstanding accuracy of 97.77% in classifying various types of FLLs, although it required manual selection of the region of interest (ROI). The video-based models produced accuracies of 83%, 88%, and 88%, respectively, without the need for ROI selection. In conclusion, the findings indicate that Transformer-based models exhibit high accuracy in CEUS-based liver diagnosis. This study highlights the potential of attention mechanisms to identify subtle inter-class differences, thereby reducing the reliance on manual intervention.

1. Introduction

An imaging technique that provides real-time, non-invasive, and reasonably priced visualization of vascular structure and tissue perfusion is Contrast-Enhanced Ultrasound (CEUS).
Ultrasound has demonstrated broad utility across various fields of medicine, including hepatology [1], cardiology [2,3,4], gynecology [5,6], and dermatology [7], illustrating its versatility and motivating the adaptation of deep learning (DL) methods across anatomical and disease contexts.
Clinical results have been greatly improved by the extensive use of ultrasound in the identification of renal anomalies, FLLs, and other organ-specific diseases. However, effective analysis of dynamic contrast patterns over time requires professional expertise, making CEUS image and video interpretation a very difficult task [8], only available to very experienced medical personnel. Computer-Aided Diagnosis (CAD) technologies are now important in improving medical imaging diagnostic accuracy [9]. Developing efficient CAD systems for CEUS remains a difficult task due to a number of factors, including a limited number of available patients, a lack of expertise in data labeling, privacy concerns, and data storage [10,11,12].
The increasing usage of Artificial Intelligence (AI) approaches in medical imaging is a result of the need for reliable, efficient, and easily available diagnostic help. Among them, CAD systems are designed to help physicians automate processes such as differential diagnosis, characterization of lesions, and extraction of characteristics. However, because CEUS videos include nuanced enhancement patterns and temporal dynamics, creating reliable CAD solutions for CEUS is extremely challenging [13]. Moreover, one of the most significant advantages of AI in this industry is its ability to speed up medical image analysis. Traditional image interpretation methods are time-consuming and prone to human error. However, AI can analyze and interpret images considerably faster, drastically reducing the time it takes to diagnose a patient. This quickness is especially important in emergency circumstances [14,15].
Convolutional Neural Networks (CNNs) represent cutting edge technology for many medical imaging applications, and they can learn inductive bias [16]. Today, visual data analysis has been completely transformed by recent developments in the DL domain, especially by the introduction of Transformer-based neural networks. Transformers were first created for natural language processing (NLP) but have now been modified for vision applications. An important advantage over CNNs is the remarkable capacity to represent long-range relationships in data. This led us to the idea that Transformers could also be suitable for identifying the complex spatiotemporal correlation aspects of CEUS exploration by considering simultaneous local and global information [17,18].
The Vision Transformer (ViT) architecture is now used in some medical imaging applications [19,20,21]. The transformer excels in capturing long-range relationships through its self-attention mechanism, demonstrating outstanding results in complicated NLP tasks [22,23]. In identifying minor irregularities in anatomical imaging, the local connection among adjacent pixels is significant, along with long-range correlations [24]. On the other hand, CNN solutions are known for their ability to capture local dependencies. The ViT and CNNs have complementary strengths in the processing of medical images for various applications, such as segmentation, classification, and prediction. Many studies have attempted to merge these two designs to collect global and local visual content, resulting in hybrid concepts [25,26,27,28]. CNNs with spatial convolution filters may learn local characteristics; however, shallow CNNs with fewer layers may fail to grasp the image’s global context due to limited abstraction. The ViT uses self-attention to learn long-range dependencies between image patches, whereas patch-based positional encoding may miss local spatial information and fall short of CNN performance on small datasets. Recent investigations, notably in radiology, have emphasized the difficulty of the ViT in detecting modest results in confined spaces [29,30].

Motivation and Novelty

Liver cancer is the fourth most common kind of cancer worldwide [31]. Romania possesses the highest incidence rates in the EU for liver and nasopharyngeal malignancies among women [32]. Hepatocellular carcinoma, commonly referred to as HCC, is induced by cirrhosis and hepatitis B or C virus [33].
DL techniques have shown promise in pathology. Because manual tumor identification is prone to errors, liver cancers may be misdiagnosed, which might have deadly consequences. CAD methods, on the other hand, save time and need little human intervention [34]. The aim of this study is to develop a CAD system for automated diagnosis of FLLs from CEUS images and videos.
CEUS has emerged as a reliable imaging technology for assessing FLLs [35,36,37], and to address the challenges of CEUS analysis, this article firstly proposes a Hybrid Transformer Neural Network (HTNN) architecture for image-based CAD implementation. The proposed hybrid architecture achieves a balance between performance and computational economy by merging CNNs for efficient local feature extraction with transformers to capture temporal and spatial relationships. Particular focus is placed on creating a tiny, lightweight neural network architecture that can be implemented in real-world clinical situations, possibly on an embedded device, without requiring significant processing resources. Moreover, this approach addresses the limitations of typical medical/CEUS datasets, namely the small number of available patients, while being able to categorize a greater range of FLL types than most existing CEUS CAD systems, providing comparable accuracy to other ML techniques reported in the literature.
Secondly, this study expands our previous work [38] and advances the state-of-the-art in the field by proposing neural solutions that can perform an FLL diagnosis taking into account the CEUS video exploration. Taking into consideration the information provided by the video input data is a step forward in CEUS automatic diagnosis because it enables CAD to formulate an output based on spatio-temporal patterns, imitating the process performed by experienced practitioners, who not only observe separated frames (images and spatial patterns) but also the temporal characteristics of the entire exploration. For example, the “washout” feature describes a lesion becoming less bright (hypoenhancing) than the liver parenchyma, and its diagnostic significance depends on its timing.
In summary, the presented solutions offer (1) faster inference, making them appropriate in clinical workflows; (2) enhanced robustness across a range of imaging conditions; (3) less dependence on large, annotated datasets; and (4) accurate detection and classification of liver lesions.
This paper is structured as follows. Section 2 summarizes the state of the art. Section 3 introduces the materials and methods utilized throughout the study. Section 4 presents the experimental results using a deep neural architecture for CEUS images and video datasets. Finally, Section 5 concludes the paper.

2. State of the Art

2.1. Classical CNN-Based Approaches

Early applications of DL in CEUS and liver imaging have primarily relied on CNNs. Traditional computer vision and machine learning methods, such as multi-layer perceptrons radial basis function networks, and classical CNN architectures, were often used in place of hand-crafted features [39,40,41,42,43,44].
Both spatial and temporal information in CEUS examinations are clinically important: static contrast patterns are useful for diagnosing FLLs [45,46,47], while temporal dynamics extracted as time–intensity curves (TICs) provide complementary diagnostic clues [48,49].
For temporal modeling, CNNs have often been paired with recurrent architectures such as LSTM networks to capture enhancement dynamics across frames [50,51]. These CNN-based pipelines demonstrated strong performance but were limited by their local receptive fields and reliance on manual feature engineering in earlier stages.

2.2. Vision Transformers (ViTs)

With the success of the Transformer architecture in NLP, ViTs have been adapted for medical imaging. ViTs divide images into fixed-size patches and use multi-head self-attention to capture contextual relationships across the entire field of view [52,53,54,55,56,57,58].
This global attention mechanism makes them particularly suited for identifying fine-grained texture variations often present in liver lesions. Several studies [59,60,61,62] have reported ViTs surpassing CNNs in classification tasks when trained with sufficient data. In liver imaging, ViTs have been applied to tumor segmentation [63], microvascular invasion prediction (MVI-TR) [64], and cross-phase image fusion [65]. Adaptation strategies, such as AdaptFormer [66], aim to make ViTs more efficient for smaller medical datasets.

2.3. Hybrid CNN–Transformer Models

To overcome the limitations of pure CNN or pure Transformer designs, hybrid models have been proposed. These architectures leverage CNNs for efficient local feature extraction while employing Transformers to capture long-range dependencies [67,68,69,70,71]. Examples include TransLiver [65], which integrates convolutional encoders with Transformer blocks for multi-phase CEUS fusion, and CSWin [72], which introduces cross-shaped window self-attention for faster inference. Variants such as Swin Transformer, PVT, TNT, and DeiT [73,74,75,76] have also been designed with efficiency in mind, making them more suitable for clinical deployment. Hybrid methods extend further into generative paradigms, where Transformers are paired with Generative Adversarial Networks (GANs) to improve segmentation accuracy [77].

2.4. Video-Based Approaches

Because CEUS inherently provides dynamic video loops, video-based DL models have been investigated to jointly model spatial and temporal aspects. CNN–LSTM combinations [50,51,78], as well as 3D CNNs and Transformer-based video models, have shown promise in capturing both enhancement patterns and lesion morphology. These models remove the need for manual TIC extraction and better reflect the full diagnostic process used by radiologists. Video Transformers, in particular, extend ViTs to spatio-temporal contexts and are increasingly seen as a promising direction for CEUS analysis.

3. Materials and Methods

3.1. Materials

3.1.1. Image Data

For the first part of this study, the data input is represented just by CEUS images, and the utilized data came from the gastroenterology and hepatology department of the “Victor Babes” University of Medicine and Pharmacy of Timisoara (UMFVBT) dataset. Ultrasound equipment Siemens Acuson S2000 was used for all the examinations in Timisoara, Romania. The UMFVBT dataset includes a total of 273 CEUS video files that were processed, featuring five types of FLLs. The dataset includes videos from 91 patients, with 16 cases of focal nodular hyperplasia (FNH), 30 cases of HCC, 23 cases of Hemangioma (HMG), and 11 cases each of Hyperplasia (HYPERM) and Hypoplasia (HYPOM). This distribution highlights the diversity of lesions within the cohort and offers valuable examples for model training and evaluation. A total of 12,119 pictures were captured during the process. These ROI images come in different sizes [79] as depicted in Figure 1.
To prevent echogenic gas bubbles from the injected agent that is required for this kind of procedure from disappearing, the ultrasound device’s probe does not remain in constant contact with the patient during the study. In total, 273 videos were analyzed, covering five different types of liver lesions (FNH, HCC, HMG, METAHIPER, and METAHIPO). Each group has a varied number of patients (FNH, 16 patients; HCC, 30; HMG, 23; METAHIPER, 11; METAHIPO, 11; total: 91).
ROIs have been manually defined by experienced physicians. Based on the Romanian Society of Ultrasound in Medicine and Biology categorization, all exams were performed by skilled operators (Level II = advanced and III = expert). The 2012 CEUS guideline procedure from the European Federation of Societies for Ultrasound in Medicine and Biology (EFSUMB) was adhered to in every examination. The authors of [11] state that histology, contrast-enhanced CT, and MRI were available to corroborate the final diagnosis.
The primary limitation of the image-based detection approach lies in the fact that although the training and testing sets consist of distinct images, these images may still originate from the same patient. This overlap can lead to an overly simplified evaluation process, which may not accurately reflect the performance of CAD in real-world patient scenarios.

3.1.2. Video Data

For the second part of our study, the dataset includes dynamic video sequences acquired during clinical diagnostic procedures that covered a variety of pathological disorders, including hepatic lesions, e.g., HCC, HEM (HMG), and FNH. The SYSU-CEUS dataset includes 353 CEUS video sequences from three different types of FLLs (186 HCC, 109 HEM, and 58 FNH)—see Figure 2. The dataset, obtained at Sun Yat-sen University’s First Affiliated Hospital using an Aplio SSA-770A echograph from Toshiba Medical System, contains 186 malignant and 167 benign lesions. Videos with a resolution of 768 × 576 pixels were collected from different patients with large variations in the appearance and enhancement patterns (size, contrast, shape, and location) of focal lesions [81].

3.2. Methods

In this section, the deep neural network (DNN) architectures utilized in the current research are described as the design of a deep neural model represents a crucial aspect in CAD development, impacting its efficiency and performance, among other factors.

3.2.1. DNN Architecture for CEUS Image Analysis

We designed an HTNN architecture (Figure 3) for the experimental setup and tested it against previously published CNN designs, such as a custom CNN, EfficientNetB1, VGG, ResNet50, Inception, and a previous design of a ViT model; the findings are presented in [47,80] and shown in Figure 4.
The workflow shown in the diagram above begins with the available labeled (FNH, HCC, HMG, METAHIPER, and METAHIPO) dataset—see the Data block of the diagram. The data were then fed into two types of models for processing: CNN models and ViT models. The CNN models include designs known to be suitable for image classification tasks, including a custom CNN, EfficientNetB1, VGG, ResNet50, and Inception. In addition, the ViT and HTNN models are utilized as (better) alternatives to the CNN models. Figure 4 depicts the tested architectures.
The neural models’ outputs are integrated into an Evaluation module. This module produces two key performance outputs: a Classification Report, which summarizes performance metrics such as accuracy, precision, recall, and F1-score, and a Confusion Matrix, which provides a detailed breakdown of model predictions versus actual labels which are very useful in evaluating CAD classification performance. In addition to assessment, trained models can be used to predict outcomes based on fresh, previously unknown data. The prediction results contain a confidence score to assess the accuracy of each prediction, which is a crucial aspect for clinical decision assistance.
The hybrid DL architecture shown in Figure 5 is intended to categorize the five liver lesions (FNH, HCC, HMG, METAHIPER, and METAHIPO) from the image data. Multiple layers of convolution (Conv2D + Activation) are applied directly to the input data (ROIs containing the lesions). The max pooling layer is then used to form a more compact feature representation. Before the data goes through the concatenation, normalization, and multi-head attention layers, a class token is inserted in order to capture intricate relationships. To improve learning, many Transformer blocks—such as layer normalization, dense layers, dropout, and multi-head attention—are added. The model then produces predictions for every type of lesion using both local and global characteristics, combining the best characteristics of the CNN and Transformer. Using a CNN as a frontend improves sample efficiency and training stability by extracting low and mid-level local features, being most impactful in smaller datasets. The Vision Transformer layer performs global modeling, capturing long-range dependencies and context; this feature is the most useful when dealing with large-scale datasets. This hybrid interaction carefully mixes local CNN biases with global Transformer reasoning, thus leading to big performance gain.

3.2.2. DNN Architecture for CEUS Video Analysis

Our next proposal for this study is to use an end-to-end DNN-based solution to classify the focal lesion type. This approach will avoid the necessity of some hand-crafted feature engineering tasks, i.e., ROI selection, as DL architectures are capable of extracting meaningful features directly from the raw input video data. For this task, two different DNN architectures capable of directly processing video input data were used. Moreover, these recurrent DNN architectures can capture temporal features, a very important aspect in the context of a CEUS-based diagnosis process.
Firstly, we designed a custom, lightweight, frame-wise CNN for video classification—see Figure 6. This model is specifically designed to extract spatial features from individual frames while effectively aggregating these features over time. Each frame is processed using a compact MobileNetV2 backbone, which operates with a reduced width (alpha = 0.35) to maintain efficiency. This backbone is integrated within a TimeDistributed layer (32, 64, 64, 3) to ensure the sharing of weights across all frames. The temporal features obtained are summarized through GlobalAveragePooling1D (32, 1280), followed by a small fully connected layer that incorporates a dense rectified linear unit (ReLU) activation function (1280) [82], dropout (0.5) for regularization, and a softmax output layer (64) for classification purposes. The model is trained using the Adam optimizer in conjunction with a custom focal loss function that includes class weighting. This approach effectively addresses class imbalance by placing greater emphasis on instances that are more challenging to classify. The total number of parameters is ~492 k.
Secondly, as an additional baseline, we consider a compact depthwise separable CNN with a bidirectional LSTM (BiLSTM) [78] to capture both spatial and temporal dynamics in video sequences. Each frame passes through the SmallDepthwiseCNN block, which uses separable convolutions (32, 64, and 128 filters with 3 × 3 kernels) and pooling to efficiently extract hierarchical spatial features while reducing parameters. The frame-level features are then aggregated over time using a TimeDistributed wrapper, feeding into a bidirectional LSTM with 64 hidden units and dropout (0.3 recurrent + 0.3 standard) to learn temporal dependencies in both directions. The sequence-level representation is further refined with a fully connected layer of 64 ReLU units, followed by 50% dropout for regularization, and finally mapped to the target classes through a softmax output. Training uses the Adam optimizer with a learning rate of 10−4, a custom focal loss function, and accuracy as the evaluation metric, making it well-suited for multi-class sequence classification with relatively lightweight computation.
With a total parameter count of ~118 k, this model is lightweight and suitable for moderate-sized time-series tasks (Figure 7).
Thirdly, the Video Visual Transformer (ViViT) model integrates a lightweight depthwise separable CNN with a Transformer encoder to capture both spatial and temporal dependencies in video data. Each frame first passes through the SmallDepthwiseCNN block, which applies separable convolutions with 32, 64, and 128 filters; batch normalization; ReLU activations; pooling layers; and a global average pooling step to extract compact frame-level features efficiently. These frame features are then processed sequentially by a stack of Transformer encoder blocks, where multi-head self-attention (head size = 32, 4 heads) and feed-forward layers of 128 units learn temporal relationships and contextual interactions across frames. Afterward, global average pooling across the temporal dimension condenses the sequence representation, followed by a dense ReLU layer of 64 units and 50% dropout for regularization (Figure 8). The final softmax layer outputs class probabilities, while training uses the Adam optimizer with a learning rate of 1 × 10−4 and a focal loss function (with tunable α and γ = 2.0) to address class imbalance and focus learning on harder examples. The total number of parameters is extremely low, ~219 k.

4. Experimental Results

4.1. Image-Based Evaluation

The HTNN architecture behavior for CEUS FLLs classification, in comparison with other state-of-the-art methods, is presented in this section (see Table 1, Table 2 and Table 3 and Figure 9). Comparing SOTA methods is challenging for several reasons, primarily due to the varying number of lesions and, more importantly, the way images are divided into training and test sets. In all approaches analyzed, except for [47], the patient origin of the images was not taken into account when making this division. This oversight helps to explain the high accuracy rates reported in Table 2. However, in clinical practice, a trained model will be tasked with providing a diagnosis for a patient whose imaging data were not included in the model’s training. That is why a Leave-One-Patient-Out evaluation procedure more accurately represents the model’s performance.
For our case, 12,119 ROI images were organized into five folders, corresponding to the five lesions. We used 1211 (10%), 1212 (10%), and 9695 (80%) images for testing, validation, and training, respectively. To obtain the input shape of (512, 512, 3), all the images were resized and shuffled without any augmentation.
The experiments were performed using the following setup:
-
Hardware architecture: CPU: AMD RYZEN 7 5800X, RAM: 64 GB, GPU: NVIDIA GeForce RTX 3090, 24 GB RAM.
-
Software framework: TensorFlow 2.11.1, Python 3.8.0, Ubuntu 20.04 LTS 64-bit.
For the training process, we implemented the following configuration: categorical cross-entropy loss, the Adam optimizer, and a learning rate of 0.0001.
The results show that ResNet50 and Inception achieved the highest validation accuracy scores (99.17% and 99.34%), though Inception had a much higher loss, suggesting less stable learning. VGG also performed very well (97.61% accuracy with low loss) but required the longest training time and a large number of parameters. CNN and EfficientNetB1 reached moderate accuracy (~92%) with relatively low training times, while ViT underperformed with the lowest accuracy (86.54%) despite having the fewest parameters. Overall, ResNet50 provided the best balance between high accuracy and low loss, but at the cost of high model complexity (Table 1).
Achieving high accuracy in our models is promising, but clinical adoption requires computational efficiency as well. Larger architectures, such as ResNet50 (74.9 million parameters) and Inception (38.6 million parameters), deliver accuracy rates of 99.17% and 99.34%, respectively. However, their memory demands and longer inference times may render them unsuitable for resource-constrained environments.
In contrast, lightweight models like CNN (5.1 million parameters) and EfficientNetB1 (6.6 million parameters) maintain competitive accuracy of around 92% while offering smaller memory footprints and faster inference times, making them better suited for portable or point-of-care devices.
As main hyperparameters, we used 64 patches, 32 batch sizes, a 1 × 10−4 learning rate, the Adam optimizer, categorical cross-entropy loss, 100 training epochs, 5 classes, 8 layers, 12 hidden dimensions, 768 MLP dimensions, 12 heads, and a dropout rate of 0.4.

4.2. Video-Based Evaluation

In the video CNN setup, the dataset loader function prepares video data for training by extracting fixed-length frame sequences, applying preprocessing, and splitting into training, validation, and test sets while preserving patient-level grouping. Each video is read frame by frame, resized to a target image size (64, 64), normalized to the [0, 1] range, and optionally augmented with flips, small rotations, and brightness jitter for specified classes (FNH and HEM) to improve generalization. Frames are sampled at a defined interval (frame_skip = 5) and grouped into sequences of consistent length (frames_per_sequence = 32) to serve as model inputs. Labels are mapped to integer indices, converted into one-hot encoded vectors, and kept aligned with patient identifiers to ensure subject-independent evaluation. Finally, stratified group-aware splitting using GroupShuffleSplit ensures that frames from the same patient appear in only one subset, producing train, validation, and test sets with balanced class representation and no data leakage. First, GroupShuffleSplit divides the entire dataset (186 HCC patients, 109 HEM patients, and 58 FNH patients) into a training set (70%) and a temporary set (30%), using patient IDs as grouping keys so that all sequences from the same patient are assigned to only one split. Next, the temporary set is further split evenly into validation (15%) and test (15%) subsets using another GroupShuffleSplit, again grouping by patient IDs to avoid overlap. This approach guarantees that the training, validation, and test sets are disjointed at the patient level, preventing data leakage and providing a fair evaluation of the model’s generalization to unseen patients.
We used augmentation data that introduces variability into training frames to improve model generalization by applying random spatial and brightness transformations. With a 50% chance, the frame is flipped horizontally to simulate mirrored perspectives. Next, a random rotation between −15° and +15° is applied around the frame’s center using an affine transformation, mimicking slight camera angle changes. Finally, brightness jitter is introduced by multiplying pixel values with a random factor between 0.8 (darker) and 1.2 (brighter), with values clipped to the valid [0, 1] range. Together, these augmentations help the model become more robust to real-world variations in orientation and lighting conditions.
The callbacks are used to improve training stability and model performance by dynamically managing checkpoints, early stopping, and learning rates. The ModelCheckpoint callback saves only the model weights, keeping the best version based on maximum validation accuracy. The EarlyStopping callback stops training if validation accuracy does not improve for 30 consecutive epochs, restoring the best weights to prevent overfitting and wasted computation. Finally, the ReduceLROnPlateau callback monitors validation loss and reduces the learning rate by half whenever progress stalls for 5 epochs, with a minimum limit of 1 × 10−6, enabling finer convergence and helping the optimizer escape plateaus, training for 100 epochs with a batch size of 4, using ReLU and Softmax activation functions [82].
In the depthwise separable CNN model integrated with a bidirectional LSTM architecture, we ensured that the conditions for data preprocessing and evaluation settings were consistent with the video CNN setup.
In this Transformer setup, we integrated a small CNN model with a Transformer model that utilizes specifically tailored focal loss parameters and robust training strategies. To enhance generalization, data augmentation is selectively applied to the FNH and HEM classes. The model’s input shape is derived from the preprocessed training sequences, while the number of classes is based on one-hot encoded labels. We employ a custom focal loss with γ = 2.0 and class-specific weights of [1.4, 0.5, 1.0]. This approach emphasizes the FNH and HEM classes while down-weighting the more prevalent HCC class to improve precision in our predictions. The Transformer encoder comprises two blocks, each utilizing multi-head attention with four heads (head size = 32) and feed-forward layers containing 128 units. These elements are stacked following a lightweight CNN feature extractor. Training is conducted over a maximum of 100 epochs with a batch size of 4. We implement the ModelCheckpoint functionality to save the best weights based on validation accuracy. EarlyStopping is utilized to stop training after 30 epochs of stagnation, and ReduceLROnPlateau is applied to halve the learning rate when the validation loss plateaus, ensuring stable convergence throughout the training process.
Table 4 presents a comparative overview of classification performance across multiple architectures, highlighting trade-offs in handling three classes: FNH, HCC, and HEM.
The lightweight CNN achieves solid overall accuracy (~0.88) and perfect classification of HCC but struggles with FNH (low precision) and shows only moderate performance on HEM. This suggests a bias in favor of the dominant class.
By contrast, hybrid models such as CNN + BiLSTM (~0.75) and ViViT (~0.88) improve balance by capturing temporal dependencies and contextual relationships. These methods yield higher macro-averaged precision, recall, and F1-scores, reducing misclassification of FNH and HEM compared to the baseline CNN. Recurrent and attention-based mechanisms therefore provide more robust, equitable performance across all categories, especially for underrepresented or difficult cases.
The confusion matrices provide insight into how various architectures approach class separation within the test set—see Figure 10. All three models—(a) the lightweight CNN, (b) CNN + BiLSTM, and (c) ViViT—successfully achieve perfect classification for HCC, with 325 correct predictions and no errors, demonstrating the clear distinguishability of this class.
However, their performance in differentiating between FNH and HEM varies considerably. The lightweight CNN (a) correctly identifies 44 FNH cases and 63 HEM cases but shows substantial confusion, with 70 HEM instances misclassified as FNH and 16 FNH instances misclassified as HEM. The CNN + BiLSTM model (b) correctly classifies 49 FNH and only 16 HEM cases, while a large proportion of HEM cases (115) are misclassified as FNH, indicating poor separation of these two classes.
The ViViT (c) demonstrates the most balanced performance, accurately identifying 26 FNH and 105 HEM cases. Although some confusion between FNH and HEM remains (34 FNH predicted as HEM and 28 HEM predicted as FNH), this model achieves the best trade-off between the two.
In conclusion, while HCC is consistently identified with high accuracy across all models, the ViViT exhibits superior generalization capabilities when addressing the minority classes (FNH and HEM). This highlights the effectiveness of attention mechanisms in capturing subtle inter-class differences that conventional CNN-based architectures struggle with.
Figure 11 illustrates the ROC curves for three models evaluated on the test set: (a) a lightweight video CNN, (b) a video CNN integrated with BiLSTM, and (c) a ViViT. All three models exhibit exceptional performance in classifying HCC, consistently achieving an area under the curve (AUC) of 1.00.
For the remaining classes, performance varies. The lightweight CNN (a) demonstrates solid results, with AUCs of 0.87 for FNH and 0.93 for HEM. The CNN + BiLSTM model (b) shows slightly lower performance, achieving AUCs of 0.82 for FNH and 0.90 for HEM, suggesting that the addition of temporal modeling did not yield significant improvements in distinguishing between these classes.
In contrast, the ViViT (c) achieves the strongest overall performance, with the highest AUCs of 0.91 for FNH and 0.95 for HEM. This highlights its superior generalization capability and effectiveness in capturing subtle inter-class differences in multi-class video-based classification. In conclusion, while HCC is easily and perfectly separable across all models, the ViViT provides the most robust and balanced performance when addressing the more challenging distinction between FNH and HEM.
To optimize model training on a small and imbalanced video dataset, we selected a batch size of 4 and a learning rate of 1 × 10−4, along with relatively high dropout rates of 0.5 in the fully connected head and 0.3 in the Transformer/LSTM layers. This configuration balances hardware constraints, training stability, and regularization.
A small batch size fits GPU memory for long frame sequences and enhances generalization on limited data. The low learning rate supports stable weight updates, which is crucial for complex architectures like attention mechanisms and LSTMs. The strong dropout rates help mitigate overfitting, while four attention heads allow for the effective capture of temporal relationships without excessive computational requirements.

5. Conclusions and Future Work

The study presents a comprehensive analysis of DNN-based liver lesion classification methods, using images or video sequences obtained from CEUS exploration as input.
As elements of novelty, we proposed, for the case of image data input, a hybrid architecture that effectively combines CNN layers (to capture local spatial features) with Transformers that manage global context, thereby enhancing the differentiation of lesions, resulting in a slim Visual Transformer model called HTNN. The resulting architecture can address common medical dataset issues, such as small sample sizes. Compared to one of our previous works [80], we successfully streamlined the design by reducing the number of parameters from 57.6 million to 8.6 million. This optimization led to a notable reduction in training time to approximately half of the time previously required. Comparing our findings with those of the latest methodologies is challenging due to differences in datasets, experimental setups, and evaluation techniques. However, our ViT design for image tasks performs better than other approaches in terms of classification accuracy and the number of classes/FLLs. Medical imaging problems with limited information can benefit from the hybrid HTNN model’s ability to efficiently balance model complexity and performance. The potential for implementation in real medical scenarios, where computational resources and annotated data are frequently scarce, is highlighted by the ability to reduce model parameters without affecting accuracy.
The next contribution of our work goes into the diagnosis of FLLs from CEUS video data input. We firstly introduce a lightweight video 3D CNN architecture tailored for the analysis of medical video data for focal liver lesion classification. The second proposal presents a combination of a lightweight video 3D CNN with LSTM networks, yielding competitive results in the field. Our third proposal explores a lightweight hybrid Video Visual Transformer, called ViViT, which addresses the current gap in research regarding video detection, as most studies predominantly focus on image data for the identification of liver tumors. These setups demonstrated significant improvements in performance relative to prior studies [78].
Models use full or sampled video frames to implicitly identify lesion clues through video-level supervision, removing the need for manual regions of interest (ROIs) during inference. Video models have outperformed image-based baselines and have matched or exceeded some human readers in certain metrics [89].
Future initiatives will focus on evaluating image and video models in a range of clinical contexts, including the incorporation of additional CEUS datasets and expanding DL-based diagnostics to US imagery. Many datasets remain inaccessible due to privacy concerns.
Additionally, it is essential to benchmark inference times and memory usage across various hardware configurations to balance diagnostic accuracy with real-time clinical usability.
A future direction of our research involves the use of pre-trained models from Hugging Face, specifically a comparative analysis of three additional Transformer-based models. Subsequent studies will expand this comparison to include more models.
A possibility to elevate these concerns is given by a relatively new learning paradigm called Federated Learning (FL), a decentralized learning technique where multiple devices (clients) collaboratively train a model without sharing their raw data. Each device trains a local model on its own data, and then only the model updates (not the data) are sent to a central server. This learning paradigm offers some key advantages, e.g., those related to the privacy-preserving aspect (raw data never leave the device), distributed training (it happens across many devices), and efficiency (reduces the need for centralized data collection) [90].
Additionally, we will examine various validation methodologies to strengthen the generalization and robustness of these models. One potential approach for improvement involves addressing the unequal distribution of data across the classes to enhance overall performance. Additionally, we will aim to fine-tune the parameters and conduct a comparative analysis with the baseline models.
Testing alternative, more complex Transformer models is a valuable future approach to obtain better results, although their size and hunger for data might represent important hindrances. Options like Transformer-XL [91], Swin Transformer [74], or Aggregating Nested Transformers and Cross-Attention [92] can offer substantial advantages through efficiency-enhancing architectures, improved long-term relationships, or computational cost optimization. Therefore, investigating these models and contrasting them with the baseline Transformer may offer valuable insights for enhancing performance in the task under study. We will also consider pre-training models on general action recognition datasets such as Kinetics-400 before fine-tuning them on domain-specific medical datasets that include organs other than the liver (for example, the kidney, spleen, or multi-organ segmentation using datasets such as BTCV or LiTS [93] for CT/MR images). This method can increase model resilience and generalization, particularly when medical datasets are small or diverse [94,95].
Furthermore, we intend to investigate improved MLOps strategies to enable rapid model deployment, real-time inference, and effective lifecycle management in clinical contexts. We plan to add multimodal data sources and expand the dataset with larger patient cohorts to enhance classification accuracy and clinical value, ultimately increasing the model’s potential for general use in medical imaging diagnostics.

Author Contributions

Conceptualization: C.D.C.; methodology: M.A.M.; validation: M.-E.-O.U.; formal analysis: C.D.C.; investigation: M.A.M. and M.-E.-O.U.; resources and data curation M.A.M., M.-E.-O.U. and C.D.C.; writing—original draft preparation: M.A.M.; writing—review and editing: C.D.C. and M.A.M.; visualization: M.-E.-O.U.; supervision: M.A.M.; project administration: C.D.C. All authors have read and agreed to the published version of the manuscript.

Funding

We would like to acknowledge the “Victor Babeş” University of Medicine and Pharmacy Timișoara for their support in covering the costs of publication for this research paper. This research is partially supported by the project “Romanian Hub for Artificial Intelligence—HRIA”, Smart Growth, Digitization and Financial Instruments Program, 2021–2027, MySMIS no. 334906.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki and approved by the Institutional Ethics Committee of UPT, no. 26340/23.09.2025.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The CEUS SYSU video dataset is freely available, see [81,96]. The UMFVBT CEUS data are available upon request from the corresponding author of [11] and it is not publicly available.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Gerstenmaier, J.F.; Gibson, R.N. Ultrasound in Chronic Liver Disease. Insights Imaging 2014, 5, 441–455. [Google Scholar] [CrossRef] [PubMed]
  2. Joyner, C.R.; Reid, J.M. Applications of Ultrasound in Cardiology and Cardiovascular Physiology. Prog. Cardiovasc. Dis. 1963, 5, 482–497. [Google Scholar] [CrossRef]
  3. Kossoff, G. Diagnostic Applications of Ultrasound in Cardiology. Australas. Radiol. 1966, 10, 101–106. [Google Scholar] [CrossRef]
  4. Gramiak, R.; Shah, P.M. Cardiac ultrasonography. Radiol. Clin. N. Am. 1971, 9, 469–490. [Google Scholar] [CrossRef] [PubMed]
  5. Turkgeldi, E.; Urman, B.; Ata, B. Role of Three-Dimensional Ultrasound in Gynecology. J. Obstet. Gynecol. India 2015, 65, 146–154. [Google Scholar] [CrossRef] [PubMed]
  6. Derchi, L.E.; Serafini, G.; Gandolfo, N.; Gandolfo, N.G.; Martinoli, C. Ultrasound in Gynecology. Eur. Radiol. 2001, 11, 2137–2155. [Google Scholar] [CrossRef] [PubMed]
  7. Boostani, M.; Pellacani, G.; Wortsman, X.; Suppa, M.; Goldust, M.; Cantisani, C.; Pietkiewicz, P.; Lőrincz, K.; Bánvölgyi, A.; Wikonkál, N.M.; et al. FDA and EMA-Approved Noninvasive Imaging Techniques for Basal Cell Carcinoma Subtyping: A Systematic Review. JAAD Int. 2025, 21, 73–86. [Google Scholar] [CrossRef]
  8. Ying, X.; Dong, S.; Zhao, Y.; Chen, Z.; Jiang, J.; Shi, H. Research Progress on Contrast-Enhanced Ultrasound (CEUS) Assisted Diagnosis and Treatment in Liver-Related Diseases. Int. J. Med. Sci. 2025, 22, 1092–1108. [Google Scholar] [CrossRef]
  9. Yeasmin, M.N.; Al Amin, M.; Joti, T.J.; Aung, Z.; Azim, M.A. Advances of AI in Image-Based Computer-Aided Diagnosis: A Review. Array 2024, 23, 100357. [Google Scholar] [CrossRef]
  10. Sporea, I.; Badea, R.; Martie, A.; Sirli, R.; Socaciu, M.; Popescu, A.; Dănilă, M. Contrast Enhanced Ultrasound for the Characterization of Focal Liver Lesions. Med. Ultrason. 2011, 13, 38–44. [Google Scholar] [PubMed]
  11. Moga, T.V.; Popescu, A.; Sporea, I.; Danila, M.; David, C.; Gui, V.; Iacob, N.; Miclaus, G.; Sirli, R. Is Contrast Enhanced Ultrasonography a Useful Tool in a Beginner’s Hand? How Much Can a Computer Assisted Diagnosis Prototype Help in Characterizing the Malignancy of Focal Liver Lesions? Med. Ultrason. 2017, 19, 252. [Google Scholar] [CrossRef] [PubMed]
  12. Chammas, M.C.; Bordini, A.L. Contrast-Enhanced Ultrasonography for the Evaluation of Malignant Focal Liver Lesions. Ultrasonography 2022, 41, 4–24. [Google Scholar] [CrossRef] [PubMed]
  13. Khalifa, M.; Albadawy, M. AI in Diagnostic Imaging: Revolutionising Accuracy and Efficiency. Comput. Methods Programs Biomed. Update 2024, 5, 100146. [Google Scholar] [CrossRef]
  14. Srivastav, S.; Chandrakar, R.; Gupta, S.; Babhulkar, V.; Agrawal, S.; Jaiswal, A.; Prasad, R.; Wanjari, M.B. ChatGPT in Radiology: The Advantages and Limitations of Artificial Intelligence for Medical Imaging Diagnosis. Cureus 2023, 15, e41435. [Google Scholar] [CrossRef] [PubMed]
  15. Hameed, B.M.Z.; Prerepa, G.; Patil, V.; Shekhar, P.; Zahid Raza, S.; Karimi, H.; Paul, R.; Naik, N.; Modi, S.; Vigneswaran, G.; et al. Engineering and Clinical Use of Artificial Intelligence (AI) with Machine Learning and Data Science Advancements: Radiology Leading the Way for Future. Ther. Adv. Urol. 2021, 13, 17562872211044880. [Google Scholar] [CrossRef] [PubMed]
  16. Anwar, S.M.; Majid, M.; Qayyum, A.; Awais, M.; Alnowami, M.; Khan, M.K. Medical Image Analysis Using Convolutional Neural Networks: A Review. J. Med. Syst. 2018, 42, 226. [Google Scholar] [CrossRef]
  17. Supriyono; Wibawa, A.P.; Suyono; Kurniawan, F. Advancements in Natural Language Processing: Implications, Challenges, and Future Directions. Telemat. Inform. Rep. 2024, 16, 100173. [Google Scholar] [CrossRef]
  18. Islam, S.; Elmekki, H.; Elsebai, A.; Bentahar, J.; Drawel, N.; Rjoub, G.; Pedrycz, W. A Comprehensive Survey on Applications of Transformers for Deep Learning Tasks. Expert Syst. Appl. 2024, 241, 122666. [Google Scholar] [CrossRef]
  19. Shamshad, F.; Khan, S.; Zamir, S.W.; Khan, M.H.; Hayat, M.; Khan, F.S.; Fu, H. Transformers in Medical Imaging: A Survey. Med. Image Anal. 2023, 88, 102802. [Google Scholar] [CrossRef]
  20. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image Is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  21. Li, J.; Chen, J.; Tang, Y.; Wang, C.; Landman, B.A.; Zhou, S.K. Transforming Medical Imaging with Transformers? A Comparative Review of Key Properties, Current Progresses, and Future Perspectives. Med. Image Anal. 2023, 85, 102762. [Google Scholar] [CrossRef] [PubMed]
  22. Wolf, T.; Debut, L.; Sanh, V.; Chaumond, J.; Delangue, C.; Moi, A.; Cistac, P.; Rault, T.; Louf, R.; Funtowicz, M.; et al. Transformers: State-of-the-Art Natural Language Processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, Online, 16–20 November 2020; Association for Computational Linguistics: Kerrville, TX, USA; pp. 38–45. [Google Scholar]
  23. Chen, Y.; Song, X.; Lee, C.; Wang, Z.; Zhang, Q.; Dohan, D.; Kawakami, K.; Kochanski, G.; Doucet, A.; Ranzato, M.A.; et al. Towards learning universal hyperparameter optimizers with transformers. In Proceedings of the 36th International Conference on Neural Information Processing Systems (NIPS ’22), New Orleans, LA, USA, 28 November–9 December 2022; Curran Associates Inc.: Red Hook, NY, USA, 2022; pp. 32053–32068. [Google Scholar]
  24. Wang, W.; Chen, C.; Ding, M.; Yu, H.; Zha, S.; Li, J. TransBTS: Multimodal Brain Tumor Segmentation Using Transformer. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2021; De Bruijne, M., Cattin, P.C., Cotin, S., Padoy, N., Speidel, S., Zheng, Y., Essert, C., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2021; Volume 12901, pp. 109–119. ISBN 978-3-030-87192-5. [Google Scholar]
  25. Khan, A.; Rauf, Z.; Sohail, A.; Khan, A.R.; Asif, H.; Asif, A.; Farooq, U. A Survey of the Vision Transformers and Their CNN-Transformer Based Variants. Artif. Intell. Rev. 2023, 56, 2917–2970. [Google Scholar] [CrossRef]
  26. Gao, Y.; Zhou, M.; Metaxas, D.N. UTNet: A Hybrid Transformer Architecture for Medical Image Segmentation. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2021; De Bruijne, M., Cattin, P.C., Cotin, S., Padoy, N., Speidel, S., Zheng, Y., Essert, C., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2021; Volume 12903, pp. 61–71. ISBN 978-3-030-87198-7. [Google Scholar]
  27. Shin, H.-C.; Roth, H.R.; Gao, M.; Lu, L.; Xu, Z.; Nogues, I.; Yao, J.; Mollura, D.; Summers, R.M. Deep Convolutional Neural Networks for Computer-Aided Detection: CNN Architectures, Dataset Characteristics and Transfer Learning. IEEE Trans. Med. Imaging 2016, 35, 1285–1298. [Google Scholar] [CrossRef]
  28. Suzuki, K. Overview of Deep Learning in Medical Imaging. Radiol. Phys. Technol. 2017, 10, 257–273. [Google Scholar] [CrossRef]
  29. Murphy, Z.R.; Venkatesh, K.; Sulam, J.; Yi, P.H. Visual Transformers and Convolutional Neural Networks for Disease Classification on Radiographs: A Comparison of Performance, Sample Efficiency, and Hidden Stratification. Radiol. Artif. Intell. 2022, 4, e220012. [Google Scholar] [CrossRef]
  30. Li, C.; Zhang, C. Toward a Deeper Understanding: RetNet Viewed through Convolution. Pattern Recognit. 2024, 155, 110625. [Google Scholar] [CrossRef]
  31. El-Serag, H.B. Epidemiology of Hepatocellular Carcinoma. In The Liver; Arias, I.M., Alter, H.J., Boyer, J.L., Cohen, D.E., Shafritz, D.A., Thorgeirsson, S.S., Wolkoff, A.W., Eds.; Wiley: Hoboken, NJ, USA, 2020; pp. 758–772. ISBN 978-1-119-43682-9. [Google Scholar]
  32. Country Cancer Profile 2025. Available online: https://www.oecd.org/content/dam/oecd/en/publications/reports/2025/02/eu-country-cancer-profile-romania-2025_ef833241/8474a271-en.pdf (accessed on 29 September 2025).
  33. National Cancer Institute. Available online: https://www.cancer.gov/types/liver/research (accessed on 29 September 2025).
  34. Gul, S.; Khan, M.S.; Hossain, M.S.A.; Chowdhury, M.E.H.; Sumon, M.S.I. A Comparative Study of Decoders for Liver and Tumor Segmentation Using a Self-ONN-Based Cascaded Framework. Diagnostics 2024, 14, 2761. [Google Scholar] [CrossRef]
  35. Kohli, M.; Prevedello, L.M.; Filice, R.W.; Geis, J.R. Implementing Machine Learning in Radiology Practice and Research. Am. J. Roentgenol. 2017, 208, 754–760. [Google Scholar] [CrossRef]
  36. Turco, S.; Tiyarattanachai, T.; Ebrahimkheil, K.; Eisenbrey, J.; Kamaya, A.; Mischi, M.; Lyshchik, A.; Kaffas, A.E. Interpretable Machine Learning for Characterization of Focal Liver Lesions by Contrast-Enhanced Ultrasound. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 2022, 69, 1670–1681. [Google Scholar] [CrossRef] [PubMed]
  37. Zhang, X.-Y.; Wei, Q.; Wu, G.-G.; Tang, Q.; Pan, X.-F.; Chen, G.-Q.; Zhang, D.; Dietrich, C.F.; Cui, X.-W. Artificial Intelligence—Based Ultrasound Elastography for Disease Evaluation—A Narrative Review. Front. Oncol. 2023, 13, 1197447. [Google Scholar] [CrossRef] [PubMed]
  38. Adriana Mercioni, M.; Daniel Căleanu, C. Computer Aided Diagnosis for Contrast-Enhanced Ultrasound Using a Small Hybrid Transformer Neural Network. In Proceedings of the 2024 International Symposium on Electronics and Telecommunications (ISETC), Timisoara, Romania, 7–8 November 2024; pp. 1–4. [Google Scholar]
  39. Zakaria, R.; Abdelmajid, H.; Zitouni, D. Deep Learning in Medical Imaging: A Review. In Applications of Machine Intelligence in Engineering; CRC Press: New York, NY, USA, 2022; pp. 131–144. ISBN 978-1-00-326979-3. [Google Scholar]
  40. Timofte, E.M.; Ligia Balan, A.; Iftime, T. AI Driven Adaptive Security Mesh: Cloud Container Protection for Dynamic Threat Landscapes. In Proceedings of the 2024 International Conference on Development and Application Systems (DAS), Suceava, Romania, 23–25 May 2024; pp. 71–77. [Google Scholar]
  41. Albahar, M. A Survey on Deep Learning and Its Impact on Agriculture: Challenges and Opportunities. Agriculture 2023, 13, 540. [Google Scholar] [CrossRef]
  42. Taye, M.M. Understanding of Machine Learning with Deep Learning: Architectures, Workflow, Applications and Future Directions. Computers 2023, 12, 91. [Google Scholar] [CrossRef]
  43. Eugen, I.; Laurențiu, M.D. Monitoring Energy Losses of a Residential Building through Thermographic Assessments. In Proceedings of the 2024 International Conference on Development and Application Systems (DAS), Suceava, Romania, 23–25 May 2024; pp. 1–7. [Google Scholar]
  44. Timofte, E.M.; Ligia Balan, A.; Iftime, T. Designing an Authentication Methodology in IoT Using Energy Consumption Patterns. In Proceedings of the 2024 International Conference on Development and Application Systems (DAS), Suceava, Romania, 23–25 May 2024; pp. 64–70. [Google Scholar]
  45. Mitrea, D.; Badea, R.; Mitrea, P.; Brad, S.; Nedevschi, S. Hepatocellular Carcinoma Automatic Diagnosis within CEUS and B-Mode Ultrasound Images Using Advanced Machine Learning Methods. Sensors 2021, 21, 2202. [Google Scholar] [CrossRef]
  46. Sirbu, C.L.; Seiculescu, C.; Adrian Burdan, G.; Moga, T.; Daniel Caleanu, C. Evaluation of Tracking Algorithms for Contrast Enhanced Ultrasound Imaging Exploration. In Proceedings of the Australasian Computer Science Week 2022, Brisbane, Australia, 14–18 February 2022; pp. 161–167. [Google Scholar]
  47. Căleanu, C.D.; Sîrbu, C.L.; Simion, G. Deep Neural Architectures for Contrast Enhanced Ultrasound (CEUS) Focal Liver Lesions Automated Diagnosis. Sensors 2021, 21, 4126. [Google Scholar] [CrossRef] [PubMed]
  48. Caleanu, C.-D.; Simion, G.; David, C.; Gui, V.; Moga, T.; Popescu, A.; Sirli, R.; Sporea, I. A Study over the Importance of Arterial Phase Temporal Parameters in Focal Liver Lesions CEUS Based Diagnosis. In Proceedings of the 2014 11th International Symposium on Electronics and Telecommunications (ISETC), Timisoara, Romania, 14–15 November 2014; pp. 1–4. [Google Scholar]
  49. Shiraishi, J.; Sugimoto, K.; Moriyasu, F.; Kamiyama, N.; Doi, K. Computer-aided Diagnosis for the Classification of Focal Liver Lesions by Use of Contrast-enhanced Ultrasonography. Med. Phys. 2008, 35, 1734–1746. [Google Scholar] [CrossRef] [PubMed]
  50. Mercioni, M.A.; Holban, S. The Most Used Activation Functions: Classic Versus Current. In Proceedings of the 2020 International Conference on Development and Application Systems (DAS), Suceava, Romania, 21–23 May 2020; pp. 141–145. [Google Scholar]
  51. Maurício, J.; Domingues, I.; Bernardino, J. Comparing Vision Transformers and Convolutional Neural Networks for Image Classification: A Literature Review. Appl. Sci. 2023, 13, 5521. [Google Scholar] [CrossRef]
  52. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  53. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar] [CrossRef]
  54. Gašparović, B.; Mauša, G.; Rukavina, J.; Lerga, J. Evaluating YOLOV5, YOLOV6, YOLOV7, and YOLOV8 in Underwater Environment: Is There Real Improvement? In Proceedings of the 2023 8th International Conference on Smart and Sustainable Technologies (SpliTech), Split/Bol, Croatia, 20–23 June 2023; pp. 1–4. [Google Scholar]
  55. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.-Y.; et al. Segment Anything. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–6 October 2023. [Google Scholar]
  56. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  57. Playout, C.; Duval, R.; Boucher, M.C.; Cheriet, F. Focused Attention in Transformers for Interpretable Classification of Retinal Images. Med. Image Anal. 2022, 82, 102608. [Google Scholar] [CrossRef]
  58. Xiao, H.; Li, L.; Liu, Q.; Zhu, X.; Zhang, Q. Transformers in Medical Image Segmentation: A Review. Biomed. Signal Process. Control 2023, 84, 104791. [Google Scholar] [CrossRef]
  59. Vaswani, A.; Shazeer, N.M.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is All you Need. In Proceedings of the Neural Information Processing Systems, 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  60. Takahashi, S.; Sakaguchi, Y.; Kouno, N.; Takasawa, K.; Ishizu, K.; Akagi, Y.; Aoyama, R.; Teraya, N.; Bolatkan, A.; Shinkai, N.; et al. Comparison of Vision Transformers and Convolutional Neural Networks in Medical Image Analysis: A Systematic Review. J. Med. Syst. 2024, 48, 84. [Google Scholar] [CrossRef]
  61. Vernuccio, F.; Cannella, R.; Bartolotta, T.V.; Galia, M.; Tang, A.; Brancatelli, G. Advances in Liver US, CT, and MRI: Moving toward the Future. Eur. Radiol. Exp. 2021, 5, 52. [Google Scholar] [CrossRef] [PubMed]
  62. Cao, L.; Wang, Q.; Hong, J.; Han, Y.; Zhang, W.; Zhong, X.; Che, Y.; Ma, Y.; Du, K.; Wu, D.; et al. MVI-TR: A Transformer-Based Deep Learning Model with Contrast-Enhanced CT for Preoperative Prediction of Microvascular Invasion in Hepatocellular Carcinoma. Cancers 2023, 15, 1538. [Google Scholar] [CrossRef]
  63. Chen, S.; Ge, C.; Tong, Z.; Wang, J.; Song, Y.; Wang, J.; Luo, P. AdaptFormer: Adapting Vision Transformers for Scalable Visual Recognition. Adv. Neural Inf. Process. Syst. 2022, 35, 16664–16678. [Google Scholar]
  64. Basit, A.; Yaqoob, A.; Hannan, A. ViT-LiSeg: Vision Transformer-Based Liver Cancer Segmentation in High-Resolution Computed Tomography. In Second International Conference on Computing Technologies, Tools and Applications (ICTAPP-24), June 4–6, 2024 (Abstract Book); Pakistan Scientific and Technological Information Center: Islamabad, Pakistan, 2024. [Google Scholar]
  65. Wang, X.; Ying, H.; Xu, X.; Cai, X.; Zhang, M. TransLiver: A Hybrid Transformer Model for Multi-Phase Liver Lesion Classification. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2023; Greenspan, H., Madabhushi, A., Mousavi, P., Salcudean, S., Duncan, J., Syeda-Mahmood, T., Taylor, R., Eds.; Lecture Notes in Computer Science; Springer Nature Switzerland: Cham, Switzerland, 2023; Volume 14221, pp. 329–338. ISBN 978-3-031-43894-3. [Google Scholar]
  66. Zhou, H.; Ding, J.; Zhou, Y.; Wang, Y.; Zhao, L.; Shih, C.-C.; Xu, J.; Wang, J.; Tong, L.; Chen, Z.; et al. Malignancy Diagnosis of Liver Lesion in Contrast Enhanced Ultrasound Using an End-to-End Method Based on Deep Learning. BMC Med. Imaging 2024, 24, 68. [Google Scholar] [CrossRef]
  67. Sun, Q.; Fang, N.; Liu, Z.; Zhao, L.; Wen, Y.; Lin, H. HybridCTrm: Bridging CNN and Transformer for Multimodal Brain Image Segmentation. J. Healthc. Eng. 2021, 2021, 7467261. [Google Scholar] [CrossRef] [PubMed]
  68. Bougourzi, F.; Dornaika, F.; Distante, C.; Taleb-Ahmed, A. D-TrAttUnet: Toward Hybrid CNN-Transformer Architecture for Generic and Subtle Segmentation in Medical Images. Comput. Biol. Med. 2024, 176, 108590. [Google Scholar] [CrossRef] [PubMed]
  69. Wang, C.; Li, W. A Hybrid CNN-Transformer Architecture with Frequency Domain Contrastive Learning for Image Deraining. arXiv 2023, arXiv:2308.03340. [Google Scholar] [CrossRef]
  70. Wang, L.; Zhang, C.; Li, J. A Hybrid CNN-Transformer Model for Predicting N Staging and Survival in Non-Small Cell Lung Cancer Patients Based on CT-Scan. Tomography 2024, 10, 1676–1693. [Google Scholar] [CrossRef] [PubMed]
  71. Hu, B.; Jiang, W.; Zeng, J.; Cheng, C.; He, L. FOTCA: Hybrid Transformer-CNN Architecture Using AFNO for Accurate Plant Leaf Disease Image Recognition. Front. Plant Sci. 2023, 14, 1231903. [Google Scholar] [CrossRef] [PubMed]
  72. Liu, X.; Gao, P.; Yu, T.; Wang, F.; Yuan, R.-Y. CSWin-UNet: Transformer UNet with Cross-Shaped Windows for Medical Image Segmentation. Inf. Fusion 2025, 113, 102634. [Google Scholar] [CrossRef]
  73. Han, K.; Xiao, A.; Wu, E.; Guo, J.; Xu, C.; Wang, Y. Transformer in Transformer. Adv. Neural Inf. Process. Syst. 2021, 34, 15908–15919. [Google Scholar]
  74. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021. [Google Scholar]
  75. Touvron, H.; Cord, M.; Douze, M.; Massa, F.; Sablayrolles, A.; Jégou, H. Training Data-Efficient Image Transformers & Distillation through Attention. Proc. Mach. Learn. Res. 2020, 139, 10347–10357. [Google Scholar]
  76. Wang, W.; Xie, E.; Li, X.; Fan, D.-P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pyramid Vision Transformer: A Versatile Backbone for Dense Prediction without Convolutions. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021. [Google Scholar]
  77. Demir, U.; Zhang, Z.; Wang, B.; Antalek, M.; Keles, E.; Jha, D.; Borhani, A.; Ladner, D.; Bagci, U. Transformer Based Generative Adversarial Network for Liver Segmentation. In Image Analysis and Processing. ICIAP 2022 Workshops; Mazzeo, P.L., Frontoni, E., Sclaroff, S., Distante, C., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2022; Volume 13374, pp. 340–347. ISBN 978-3-031-13323-7. [Google Scholar]
  78. Schmiedt, K.; Simion, G.; Caleanu, C.D. Preliminary Results on Contrast Enhanced Ultrasound Video Stream Diagnosis Using Deep Neural Architectures. In Proceedings of the 2022 International Symposium on Electronics and Telecommunications (ISETC), Timisoara, Romania, 10–11 November 2022; pp. 1–4. [Google Scholar]
  79. Sirbu, C.L.; Simion, G.; Caleanu, C.D. Deep CNN for Contrast-Enhanced Ultrasound Focal Liver Lesions Diagnosis. In Proceedings of the 2020 International Symposium on Electronics and Telecommunications (ISETC), Timisoara, Romania, 5–6 November 2020; pp. 1–4. [Google Scholar]
  80. Mercioni, M.A.; Căleanu, C.D.; Sîrbu, C.L. Computer Aided Diagnosis for Contrast-Enhanced Ultrasound Using Transformer Neural Network. In Proceedings of the 2023 25th International Symposium on Symbolic and Numeric Algorithms for Scientific Computing (SYNASC), Nancy, France, 11–14 September 2023; pp. 256–259. [Google Scholar]
  81. Liang, X.; Cao, Q.; Huang, R.; Lin, L. Recognizing Focal Liver Lesions in Contrast-Enhanced Ultrasound with Discriminatively Trained Spatio-Temporal Model. In Proceedings of the 2014 IEEE 11th International Symposium on Biomedical Imaging (ISBI), Beijing, China, 29 April–2 May 2014; pp. 1184–1187. [Google Scholar]
  82. Mercioni, M.A.; Holban, S. A Brief Review of the Most Recent Activation Functions for Neural Networks. In Proceedings of the 2023 17th International Conference on Engineering of Modern Electric Systems (EMES), Oradea, Romania, 9–10 June 2023; pp. 1–4. [Google Scholar]
  83. Hassan, T.M.; Elmogy, M.; Sallam, E.-S. Diagnosis of Focal Liver Diseases Based on Deep Learning Technique for Ultrasound Images. Arab. J. Sci. Eng. 2017, 42, 3127–3140. [Google Scholar] [CrossRef]
  84. Pan, F.; Huang, Q.; Li, X. Classification of Liver Tumors with CEUS Based on 3D-CNN. In Proceedings of the 2019 IEEE 4th International Conference on Advanced Robotics and Mechatronics (ICARM), Toyonaka, Japan, 3–5 July 2019; pp. 845–849. [Google Scholar]
  85. Guo, L.; Wang, D.; Xu, H.; Qian, Y.; Wang, C.; Zheng, X.; Zhang, Q.; Shi, J. CEUS-Based Classification of Liver Tumors with Deep Canonical Correlation Analysis and Multi-Kernel Learning. In Proceedings of the 2017 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Jeju, Republic of Korea, 11–15 July 2017; pp. 1748–1751. [Google Scholar]
  86. Vancea, F.; Mitrea, D.; Nedevschi, S.; Rotaru, M.; Stefanescu, H.; Badea, R. Hepatocellular Carcinoma Segmentation within Ultrasound Images Using Convolutional Neural Networks. In Proceedings of the 2019 IEEE 15th International Conference on Intelligent Computer Communication and Processing (ICCP), Cluj-Napoca, Romania, 5–7 September 2019; pp. 483–490. [Google Scholar]
  87. Wu, K.; Chen, X.; Ding, M. Deep Learning Based Classification of Focal Liver Lesions with Contrast-Enhanced Ultrasound. Optik 2014, 125, 4057–4063. [Google Scholar] [CrossRef]
  88. Streba, C.T. Contrast-Enhanced Ultrasonography Parameters in Neural Network Diagnosis of Liver Tumors. World J. Gastroenterol. 2012, 18, 4427. [Google Scholar] [CrossRef]
  89. Chen, J.; Huang, Z.; Jiang, Y.; Wu, H.; Tian, H.; Cui, C.; Shi, S.; Tang, S.; Xu, J.; Xu, D.; et al. Diagnostic Performance of Deep Learning in Video-Based Ultrasonography for Breast Cancer: A Retrospective Multicentre Study. Ultrasound Med. Biol. 2024, 50, 722–728. [Google Scholar] [CrossRef]
  90. Rauniyar, A.; Hagos, D.H.; Jha, D.; Håkegård, J.E.; Bagci, U.; Rawat, D.B.; Vlassov, V. Federated Learning for Medical Applications: A Taxonomy, Current Trends, Challenges, and Future Research Directions. IEEE Internet Things J. 2024, 11, 7374–7398. [Google Scholar] [CrossRef]
  91. Dai, Z.; Yang, Z.; Yang, Y.; Carbonell, J.; Le, Q.; Salakhutdinov, R. Transformer-XL: Attentive Language Models beyond a Fixed-Length Context. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; Association for Computational Linguistics: Kerrville, TX, USA; pp. 2978–2988. [Google Scholar]
  92. Wang, Y.; Deng, Y.; Zheng, Y.; Chattopadhyay, P.; Wang, L. Vision Transformers for Image Classification: A Comparative Survey. Technologies 2025, 13, 32. [Google Scholar] [CrossRef]
  93. Bilic, P.; Christ, P.; Li, H.B.; Vorontsov, E.; Ben-Cohen, A.; Kaissis, G.; Szeskin, A.; Jacobs, C.; Mamani, G.E.H.; Chartrand, G.; et al. The Liver Tumor Segmentation Benchmark (LiTS). Med. Image Anal. 2023, 84, 102680. [Google Scholar] [CrossRef] [PubMed]
  94. Yoon, S.; Kim, T.H.; Jung, Y.K.; Kim, Y. Accelerated Muscle Mass Estimation from CT Images through Transfer Learning. BMC Med. Imaging 2024, 24, 271. [Google Scholar] [CrossRef] [PubMed]
  95. Fang, X.; Yan, P. Multi-Organ Segmentation Over Partially Labeled Datasets with Multi-Scale Feature Abstraction. IEEE Trans. Med. Imaging 2020, 39, 3619–3629. [Google Scholar] [CrossRef] [PubMed]
  96. Liang, X.; Lin, L.; Cao, Q.; Huang, R.; Wang, Y. Recognizing Focal Liver Lesions in CEUS With Dynamically Trained Latent Structured Models. IEEE Trans. Med. Imaging 2016, 35, 713–727. [Google Scholar] [CrossRef] [PubMed]
Figure 1. A total of 25 randomly selected ROI samples of different sizes from the UMFVBT CEUS image dataset. There are five pictures from each of the five available lesion types: FNH, HCC, HMG, Hypervascular Metastase (METAHIPER), and Hypovascular Metastase (METAHIPO) [80].
Figure 1. A total of 25 randomly selected ROI samples of different sizes from the UMFVBT CEUS image dataset. There are five pictures from each of the five available lesion types: FNH, HCC, HMG, Hypervascular Metastase (METAHIPER), and Hypovascular Metastase (METAHIPO) [80].
Sensors 25 06247 g001
Figure 2. Randomly selected frames from the SYSU CEUS video dataset illustrating the three possible FLLs: FNH, HCC, and HEM.
Figure 2. Randomly selected frames from the SYSU CEUS video dataset illustrating the three possible FLLs: FNH, HCC, and HEM.
Sensors 25 06247 g002
Figure 3. The experimental setup for the static CEUS image input data type [38] where we have newly introduced the HTNN architecture.
Figure 3. The experimental setup for the static CEUS image input data type [38] where we have newly introduced the HTNN architecture.
Sensors 25 06247 g003
Figure 4. The experimental setup for CEUS images as input data, where we have newly introduced the HTNN architecture, including CNN models.
Figure 4. The experimental setup for CEUS images as input data, where we have newly introduced the HTNN architecture, including CNN models.
Sensors 25 06247 g004
Figure 5. The proposed HTNN architecture for processing CEUS images [38].
Figure 5. The proposed HTNN architecture for processing CEUS images [38].
Sensors 25 06247 g005
Figure 6. The proposed lightweight CNN architecture for CEUS video data.
Figure 6. The proposed lightweight CNN architecture for CEUS video data.
Sensors 25 06247 g006
Figure 7. The proposed lightweight CNN + Bidirectional LSTM architecture for CEUS video data.
Figure 7. The proposed lightweight CNN + Bidirectional LSTM architecture for CEUS video data.
Sensors 25 06247 g007
Figure 8. The proposed ViViT architecture to process CEUS video data.
Figure 8. The proposed ViViT architecture to process CEUS video data.
Sensors 25 06247 g008
Figure 9. Confusion matrix for the HTNN model.
Figure 9. Confusion matrix for the HTNN model.
Sensors 25 06247 g009
Figure 10. Confusion matrix on the test set. (a) Video lightweight CNN model; (b) video CNN + LSTM model; (c) ViViT model.
Figure 10. Confusion matrix on the test set. (a) Video lightweight CNN model; (b) video CNN + LSTM model; (c) ViViT model.
Sensors 25 06247 g010
Figure 11. ROC curves on the test set. (a) Video lightweight CNN model; (b) video CNN + BiLSTM model; (c) ViViT model.
Figure 11. ROC curves on the test set. (a) Video lightweight CNN model; (b) video CNN + BiLSTM model; (c) ViViT model.
Sensors 25 06247 g011
Table 1. Comparison of architectures used.
Table 1. Comparison of architectures used.
ModelValidation
Accuracy
(Max)
Loss
(Min)
Total ParamsTotal Training
Time [min]
CNN91.99%0.32745,144,3578
EfficientNetB191.82%0.59926,647,2538
VGG97.61%0.117124,154,94953
ResNet5099.17%0.041574,971,01351
Inception99.34%1.236138,583,07734
ViT86.54%0.64312,584,51717
Table 2. Comparison of state-of-the-art methods.
Table 2. Comparison of state-of-the-art methods.
Ref.LesionsGeneral Accuracy [%]
Hassan et al. [83]Cyst, HEM, HCC97.2
Pan et al. [84]FNH, HCC93.1
Guo et al. [85]Malign, Benign90.4
Vancea et al. [86]HCC80.3
Wu et al. [87]HCC, CH, META, LFS86.3
Streba et al. [88]HCC, HYPERM, HYPOM, HEM, FFC87.1
Căleanu et al. [47]FNH, HCC, HMG, HYPERM HYPOM88.0
Mercioni et al. [80]FNH, HCC, HMG, METAHIPER, METAHIPO97.5
HTNN (ours)FNH, HCC, HMG, METAHIPER, METAHIPO97.77
Table 3. Classification report for the HTNN model in five FLLs: FNH, HCC, HMG, Hypervascular Metastase (METAHIPER), and Hypovascular Metastase (METAHIPO).
Table 3. Classification report for the HTNN model in five FLLs: FNH, HCC, HMG, Hypervascular Metastase (METAHIPER), and Hypovascular Metastase (METAHIPO).
Precision
[%]
Recall
[%]
F1-Score
[%]
Samples
FNH100100100222
HCC9910099394
HMG999999324
METAHIPER999999134
METAHIPO1009999138
Accuracy 991212
Macro avg10099991212
Weighted avg9999991212
Table 4. Classification report for the proposed neural video processing architectures.
Table 4. Classification report for the proposed neural video processing architectures.
ArchitectureLesion TypePrecisionRecallF1-ScoreSamples
Video lightweight CNNFNH0.390.730.5160
HCC1.001.001.00325
HEM (HMG)0.800.470.59133
macro avg0.730.740.70518
weighted avg0.880.830.84518
accuracy 0.83 518
Video CNN + LSTMFNH0.300.820.4460
HCC0.991.001.00325
HEM (HMG)0.620.120.20133
macro avg0.640.650.54518
weighted avg0.810.750.73518
accuracy 0.75 518
ViVitFNH0.480.430.4660
HCC1.001.001.00325
HEM (HMG)0.760.790.77133
macro avg0.750.740.74518
weighted avg0.880.880.88518
accuracy 0.88 518
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mercioni, M.A.; Căleanu, C.D.; Ursan, M.-E.-O. Images Versus Videos in Contrast-Enhanced Ultrasound for Computer-Aided Diagnosis. Sensors 2025, 25, 6247. https://doi.org/10.3390/s25196247

AMA Style

Mercioni MA, Căleanu CD, Ursan M-E-O. Images Versus Videos in Contrast-Enhanced Ultrasound for Computer-Aided Diagnosis. Sensors. 2025; 25(19):6247. https://doi.org/10.3390/s25196247

Chicago/Turabian Style

Mercioni, Marina Adriana, Cătălin Daniel Căleanu, and Mihai-Eronim-Octavian Ursan. 2025. "Images Versus Videos in Contrast-Enhanced Ultrasound for Computer-Aided Diagnosis" Sensors 25, no. 19: 6247. https://doi.org/10.3390/s25196247

APA Style

Mercioni, M. A., Căleanu, C. D., & Ursan, M.-E.-O. (2025). Images Versus Videos in Contrast-Enhanced Ultrasound for Computer-Aided Diagnosis. Sensors, 25(19), 6247. https://doi.org/10.3390/s25196247

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop