Next Article in Journal
Molecular Testing as Triage in Cervical Cancer Screening: Economic Evaluation Using Headroom Analysis
Next Article in Special Issue
Immersive Virtual Reality for Enabling Patient Experience and Enrollment in Oncology Clinical Trials: A Feasibility Study
Previous Article in Journal
IER5 Promotes Ovarian Cancer Cell Proliferation and Peritoneal Dissemination
Previous Article in Special Issue
Digital Rehabilitation Program for Breast Cancer Survivors on Adjuvant Hormonal Therapy: A Feasibility Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Transfer Learning-Based Integration of Dual Imaging Modalities for Enhanced Classification Accuracy in Confocal Laser Endomicroscopy of Lung Cancer

by
Mircea-Sebastian Șerbănescu
1,†,
Liliana Streba
2,*,
Alin Dragoș Demetrian
3,†,
Andreea-Georgiana Gheorghe
4,*,
Mădălin Mămuleanu
5,
Daniel-Nicolae Pirici
6 and
Costin-Teodor Streba
7
1
Department of Medical Informatics and Statistics, University of Medicine and Pharmacy of Craiova, 200349 Craiova, Romania
2
Department of Oncology and Palliative Care, University of Medicine and Pharmacy of Craiova, 200349 Craiova, Romania
3
Department of Thoracic Surgery, University of Medicine and Pharmacy of Craiova, 200349 Craiova, Romania
4
Doctoral School, University of Medicine and Pharmacy of Craiova, 200349 Craiova, Romania
5
Department of Automatic Control and Electronics, University of Craiova, 200585 Craiova, Romania
6
Department of Histology, University of Medicine and Pharmacy of Craiova, 200349 Craiova, Romania
7
Department of Pulmonology, University of Medicine and Pharmacy of Craiova, 200349 Craiova, Romania
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Cancers 2025, 17(4), 611; https://doi.org/10.3390/cancers17040611
Submission received: 31 December 2024 / Revised: 6 February 2025 / Accepted: 7 February 2025 / Published: 11 February 2025
(This article belongs to the Special Issue Digital Health Technologies in Oncology)

Simple Summary

Lung cancer is the most common and deadly cancer worldwide, making early and accurate diagnosis essential for improving patient outcomes. Traditional diagnostic methods often require invasive biopsies, which can be uncomfortable and carry risks. This study investigates a new approach that combines two imaging techniques, namely confocal laser endomicroscopy (pCLE), which provides real-time microscopic views of lung tissue, and standard histological analysis of stained tissue samples. By using a machine learning method called dual transfer learning, we trained computer models to better differentiate between cancerous and non-cancerous tissues in pCLE. Our findings demonstrate that integrating these two imaging modalities results in a statistically significant enhancement in the accuracy of lung cancer detection. This innovative approach has the potential to make diagnoses faster and less invasive, ultimately leading to better patient care and more effective treatment strategies.

Abstract

Background/Objectives: Lung cancer remains the leading cause of cancer-related mortality, underscoring the need for improved diagnostic methods. This study seeks to enhance the classification accuracy of confocal laser endomicroscopy (pCLE) images for lung cancer by applying a dual transfer learning (TL) approach that incorporates histological imaging data. Methods: Histological samples and pCLE images, collected from 40 patients undergoing curative lung cancer surgeries, were selected to create 2 balanced datasets (800 benign and 800 malignant images each). Three CNN architectures—AlexNet, GoogLeNet, and ResNet—were pre-trained on ImageNet and re-trained on pCLE images (confocal TL) or using dual TL (first re-trained on histological images, then pCLE). Model performance was evaluated using accuracy and AUC across 50 independent runs with 10-fold cross-validation. Results: The dual TL approach statistically significant outperformed confocal TL, with AlexNet achieving a mean accuracy of 94.97% and an AUC of 0.98, surpassing GoogLeNet (91.43% accuracy, 0.97 AUC) and ResNet (89.87% accuracy, 0.96 AUC). All networks demonstrated statistically significant (p < 0.001) improvements in performance with dual TL. Additionally, dual TL models showed reductions in both false positives and false negatives, with class activation mappings highlighting enhanced focus on diagnostically relevant regions. Conclusions: Dual TL, integrating histological and pCLE imaging, results in a statistically significant improvement in lung cancer classification. This approach offers a promising framework for enhanced tissue classification. and with future development and testing, iy has the potential to improve patient outcomes.

1. Introduction

Lung cancer remains one of the most significant medical challenges worldwide, both in terms of prevalence and associated mortality. Epidemiological studies indicate that this pathology ranks first among cancer-related deaths globally, with a markedly increased incidence in countries with high Human Development Index scores, where smoking and prolonged exposure to risk factors result in alarmingly high mortality rates [1]. Furthermore, a notable sex disparity persists, with men being more affected, although recent years have shown a rising incidence among women [1]. Countries undergoing economic transition are also experiencing an increase in lung cancer incidence due to air pollution and occupational exposure to toxic agents [2]. Unfortunately, the majority of cases are diagnosed at advanced stages, which is reflected in a 5-year survival rate below 20% [3].
The risk factors for lung cancer are well documented, with smoking directly associated with the largest proportion of cases [4]. Air pollution, particularly through the inhalation of fine particles (PM2.5) and chemical agents, such as nitrogen dioxide, is another critical contributor to incidence [5]. Additionally, occupational exposure to hazardous substances, such as asbestos, significantly increases oncological risk [6]. The socioeconomic consequences are considerable, encompassing both direct costs (treatment, hospitalization, advanced therapies) and indirect costs (productivity loss, family impact), while early screening and diagnostic methods, such as low-dose computed tomography (LDCT), generate additional financial burdens [7]. In resource-limited economies, restricted access to early detection programs exacerbates the disease burden [8].
In this context, improving diagnostic methods becomes essential. LDCT screening has proven effective for early detection, particularly in individuals with elevated risk factors [9]. Moreover, the introduction of molecular biomarkers and complementary analyses offers a promising perspective for earlier diagnosis [10], especially when targeted at high-risk populations [11]. However, challenges related to the specificity and sensitivity of conventional methods remain.
A complementary and increasingly utilized approach in bronchoscopy is probe-based confocal laser endomicroscopy (pCLE). This technique enables real-time in vivo acquisition of microscopic images of pulmonary tissue, providing valuable insights into cellular architecture and structural abnormalities [12]. Explorations, like pCLE, can differentiate neoplastic from normal tissue, optimizing biopsy sampling by guiding toward highly suspicious areas [13,14]. While the technique does not entirely replace biopsy—as tumors lack autofluorescence and require the use of contrast agents—pCLE significantly accelerates the diagnostic process [15]. Features, such as cellular atypia, disorganized elastic fibers, and alterations in alveolar structure, can be rapidly identified [16]. The sensitivity and specificity of pCLE for certain pathologies, such as Pneumocystis jirovecii pneumonia, are notably high, surpassing conventional methods [17]. Additionally, this technique is valuable in monitoring post-lung transplant patients by detecting signs of acute rejection [18]. The development of standardized interpretation guidelines and specialist training ensures high diagnostic accuracy [19,20].
Alongside these imaging advances, artificial intelligence (AI), particularly deep learning (DL), has gained prominence in medical image diagnostics. Deep convolutional neural network (CNN) architectures, such as ResNet, VGG, and Inception, have demonstrated remarkable performance in classifying pulmonary nodules on radiographs and CT scans [21,22]. In histopathology, DL can uncover subtle patterns and microscopic features that are challenging for human experts to detect [23]. Nevertheless, a major limitation in the medical field remains the lack of sufficiently labeled data. To overcome this challenge, transfer learning (TL) allows for the reuse of knowledge acquired by networks initially trained on massive datasets (e.g., ImageNet), adapting them to rarer and more specialized medical images [24,25]. It has been proven that TL thus reduces the time and costs associated with annotation while improving the models’ ability to generalize on limited datasets [26].
Pre-trained networks, such as AlexNet, GoogLeNet, and ResNet, have been successfully employed in classifying pCLE and histopathological images, achieving very high accuracy [27,28]. Furthermore, dual TL—integrating knowledge from two domains—can further enhance the performance of these models [29,30]. Multi-modal data integration (e.g., CT imaging, histopathology, biomarkers) represents another frontier in lung cancer diagnostics. Recent methods propose the fusion of information extracted from various sources, increasing the accuracy and robustness of diagnoses [31].
A notable example highlights the potential of combining pCLE images with other modalities to achieve superior detection and classification of pulmonary lesions [30]. The integration of histopathological data with pCLE imaging, coupled with TL techniques, can substantially improve the likelihood of precise and early diagnosis. This multi-modal approach addresses challenges related to incomplete, misaligned, or heterogeneous data, bringing us closer to an integrative framework for computer-assisted diagnostics.
This highlights the pressing need for substantial improvements in early screening and diagnostic methods to address the high incidence and mortality rates associated with lung cancer. Techniques, such as pCLE, have demonstrated utility in rapidly obtaining in vivo microscopic information, while AI, particularly DL and TL, opens new avenues by reducing reliance on large datasets and diversifying information sources. The multi-modal integration of imaging and clinical data, as illustrated by recent studies, represents a crucial step toward more accurate, personalized, and cost-effective diagnostics. Continuous development of these technologies and associated methodologies will play an essential role in improving the prognosis and quality of life for patients diagnosed with lung cancer.
Our research focuses on exploring the potential of dual TL by integrating two imaging modalities. The objective is to enhance the classification accuracy of pCLE for lung cancer images by introducing knowledge derived from histology techniques, thereby addressing the challenges associated with the clinical use of pCLE. This approach demonstrated the efficacy of leveraging multiple sources of information to improve diagnostic outcomes in lung cancer.

2. Materials and Methods

This study was conducted between January 2019 and January 2021 on a cohort of 40 patients from the Thoracic Surgery Clinic at the Emergency County Hospital of Craiova, Romania, who had undergone curative lung cancer surgeries. The target population comprised residents of Dolj County, Romania, with the only inclusion criterion being a recommendation for curative lung cancer surgery. Ethical approval was granted by the Ethics Committee of the University of Medicine and Pharmacy of Craiova, and informed consent was obtained from all participants.
Histological samples were obtained intraoperatively, labeled with acriflavine, and analyzed ex vivo using a pCLE probe (pCLE, CellVizio®, Mauna Kea Technologies, Paris, France). The resulting images were saved for subsequent analysis. The same tissue samples were processed histopathologically and stained with hematoxylin–eosin for tumor localization and silver impregnation, which followed standardized clinical staining protocols. Whole-slide scanning was performed to digitize the histology slides, and the images were stored electronically. The scanning process utilized a Nikon Eclipse 90i (Nikon Instruments Inc., Melville, NY, USA) motorized microscope equipped with a Prior OptiScan ES111 (Prior Scientific Inc., Rockland, MA, USA) motorized stage, a 16-megapixel Nikon DS-Ri-2 CMOS cooled camera (Nikon Instruments Inc., Melville, NY, USA), and Nikon NIS-Elements AR software (Nikon Instruments Inc., Melville, NY, USA) for image acquisition and analysis.
Two distinct datasets were obtained: one derived from the pCLE images (confocal dataset) and one from the histology slides (histology dataset).
Both datasets were independently reviewed by two pathologists, who annotated malignant and benign regions without access to any clinical information or diagnostic data. Hematoxylin–eosin-stained images were registered onto silver impregnation images to facilitate clear tumor identification. The resulting digital slides were divided into tiles of 512 × 512 pixels. These tiles were presented to the pathologists for annotation, and the labeled silver impregnation images were subsequently saved. Confocal images (512 × 512 pixels) were obtained from confocal video recordings, and these images were also labeled by the pathologists. From each patient, 10 image samples were selected for each class (malignant and benign) and for each imaging technique. Complete inter-observer agreement was required for all images. Images with discrepancies in labeling were replaced to ensure consistency. This resulted in a total of 800 confocal images and 800 histology images, equally balanced between malignant and benign labels. Representative examples of images from both datasets are illustrated in Figure 1.
The original dataset did not undergo normalization or noise reduction prior to analysis. However, data augmentation was applied during the training process using Matlab’s (The MathWorks, Inc., Natick, MA, USA) built-in image data augmenter. The following augmentation parameters were used: reflection along the X and Y axes, random rotations ranging from 0° to 360°, and scaling factors between 0.9 and 1.1.
The datasets were divided to facilitate a 10-fold cross-validation evaluation. Specifically, 70% of the data was allocated for training, 20% for validation, and 10% for testing. This split was determined prior to data augmentation, and the images were randomly selected before each network training iteration.
Given the inherent challenges in precise spatial registration between the two imaging modalities, we explored a dual TL approach to enhance the performance of the confocal image classifier. This approach leverages features learned from the histology dataset to improve classification accuracy on the confocal dataset.
Three DL networks—AlexNet [32], GoogLeNet [33], and ResNet [34]—were employed for this purpose. Several factors influenced the selection of these architectures. First, all three networks have pre-trained versions on ImageNet [35] available for Matlab, providing a solid foundation for TL. Second, given the relatively small size of our dataset, choosing more complex architectures with a greater number of trainable parameters could have led to overfitting, while these networks strike a balance between complexity and generalizability. Lastly, the team’s experience with these architectures and their proven success in other medical applications [27,36] further justified their selection for this task.
AlexNet, introduced by Krizhevsky et al. [37] in 2012, represents a seminal work in the field of DL, particularly in image classification tasks. It is a CNN architecture consisting of five convolutional layers, each followed by rectified linear unit activations to introduce non-linearity. These layers are interspersed with max-pooling layers to reduce spatial dimensions and improve computational efficiency. The network transitions to the fully connected layers with two fully connected layers, ultimately concluding with a 1000-way softmax layer to classify images into 1000 categories from the ImageNet dataset. AlexNet was among the first architectures to leverage dropout regularization to mitigate overfitting and data augmentation techniques, such as random cropping and flipping, to enhance model robustness. The model is freely available for download and compatible with Matlab.
GoogLeNet, introduced in 2014 by Szegedy et al. [38], marked a departure from traditional deep CNNs by employing a more sophisticated, non-linear architecture through the introduction of the Inception module. The network consists of 22 layers but is computationally efficient due to its modular design. The Inception module enables the simultaneous application of convolutions with multiple filter sizes (e.g., 1 × 1, 3 × 3, and 5 × 5) and max-pooling, with the results concatenated into a single output. This architecture reduces the number of parameters compared to earlier CNNs while preserving representational capacity. Key innovations in GoogLeNet include the use of global average pooling instead of fully connected layers at the network’s end, significantly reducing overfitting. Additionally, GoogLeNet incorporates auxiliary classifiers at intermediate layers to facilitate gradient flow during backpropagation and improve training stability. GoogLeNet’s model is freely available for download and compatible with Matlab.
The ResNet-18 architecture, introduced by He et al. [39] in 2016, is a part of the Residual Network (ResNet) family, which revolutionized DL by addressing the vanishing gradient problem in very deep networks. ResNet-18 features 18 layers organized into a series of residual blocks, where each block introduces skip connections (also called shortcuts) that bypass one or more layers. These skip connections allow the network to learn residual mappings rather than full transformations, facilitating the efficient training of deeper models. Each residual block typically consists of two convolutional layers with batch normalization and ReLU activations, followed by the addition of the input (identity) to the output. This approach significantly improves gradient flow during training and mitigates the degradation problem seen in deeper networks. The model is also freely available for download and compatible with Matlab.
Initially, the networks were pre-trained on the ImageNet dataset and re-trained for binary classification (malignant vs. benign) using the histology dataset. This constituted the first phase of TL. Subsequently, the resulting networks were re-trained on the confocal dataset without modifying the output layer, representing the second TL phase. The final models, therefore, integrated knowledge from both imaging techniques, creating a dual TL framework.
Each of these pre-trained network architectures was adapted to accommodate a new classification task distinct from their original purpose of classifying real-world images, such as those from the ImageNet dataset. Specifically, the original networks were modified to suit the requirements of a binary classification problem. The final fully connected layer and the output layer of each model were replaced to reflect the new dataset’s structure. The last fully connected layer was redesigned to output features corresponding to the binary classification problem, while the softmax output layer was modified to produce probabilities for two classes, aligning with the label structure of the dataset.
The training process for all three networks was conducted under identical conditions to ensure a fair comparison of their performance. The hyperparameters were set empirically based on preliminary experimentation. A mini-batch size of 20 was employed to achieve a balance between computational efficiency and the stability of gradient estimates. The initial learning rate was set to 0.0001 to ensure gradual convergence and to minimize the risk of overshooting during optimization. Furthermore, validation patience was configured to 4, meaning the training would halt early if the validation loss did not improve over four consecutive epochs, helping to prevent overfitting.
All networks were trained using stochastic gradient descent with momentum as the optimization algorithm. The momentum mechanism was incorporated to accelerate convergence by dampening oscillations in the optimization trajectory and to help escape shallow local minima. This consistent setup across the architectures ensured a standardized environment for evaluating the models on the new classification task, providing a reliable basis for comparative analysis.
The decision to utilize pre-trained DL networks rather than developing a custom mode was based on the size of the dataset, which is relatively small and insufficient for training a new network from scratch without risking overfitting. Pre-trained networks leverage TL, where pre-existing trained weights from large datasets are used to improve model performance on smaller datasets. This approach ensures more reliable results and mitigates the risk of overfitting, thereby making it a suitable choice for the scope of our study.
To benchmark the dual TL scenario, we conducted a control experiment wherein the same networks, initialized with ImageNet weights, were re-trained exclusively on the confocal dataset. This control experiment provided the confocal TL scenario. A graphical representation of the study design is presented in Figure 2.
To ensure robust and statistically reliable results, all DL models were executed independently over 50 iterations. The stochastic nature of DL algorithms necessitates multiple independent runs to adequately sample decision performance. For each run, a 10-fold cross-validation strategy was employed. At each iteration, 40 images of each class/scenario were randomly selected for testing prior to data augmentation.
Statistical analyses were conducted to confirm adequate power, with a two-tailed null hypothesis, a Type I error rate (α) of 0.05, and a target statistical power exceeding 95%.
To further elucidate the classification performance, confusion matrices were generated for two randomly selected networks from each scenario. These matrices illustrate the true versus predicted classifications for both benign and malignant classes.
To visualize and interpret the regions of the images that the networks focus on during classification, class activation mapping (CAM) was employed.
All DL model implementations, data processing, and statistical evaluations were performed using Matlab).
Computations were performed on an Intel® Xeon® Silver 4216 processor (Intel Corp., Santa Clara, CA, USA), equipped with 128 GB of RAM, and an NVIDIA® Quadro® RTX 6000 graphics processing unit (NVIDIA Corp., Santa Clara, CA, USA) with 24 GB of memory. The average training time for one network was 53 ± 13 min.

3. Results

A total of 50 DL networks were trained using the dual TL approach, integrating both histology and confocal datasets, and an additional 50 networks were trained exclusively on the confocal dataset for each of the 3 network architectures: AlexNet, GoogLeNet, and ResNet. The performance metrics, including accuracy and area under the curve (AUC), for both scenarios across all network architectures are summarized in Table 1 and Table 2, respectively.

3.1. Network Performance

3.1.1. Accuracy Assessment

The dual TL scenario consistently outperformed the confocal TL scenario across all network architectures. Specifically, AlexNet achieved a mean accuracy of 94.97% (±1.76), GoogLeNet attained 91.43% (±2.17), and ResNet reached 89.87% (±2.15) in the dual TL scenario. In contrast, the confocal TL scenario yielded lower accuracies of 90.14% (±2.13) for AlexNet, 85.71% (±2.55) for GoogLeNet, and 84.65% (±1.84) for ResNet. An ANOVA test confirmed that these differences were statistically significant (p < 0.001). Furthermore, Student’s t-test revealed that the improvements in accuracy for the dual TL scenario were statistically significant across all network architectures (p < 0.001).

3.1.2. AUC Assessment

Similarly, the AUC values demonstrated superior performance for the dual TL approach. AlexNet achieved an AUC of 0.98 (±0.01), GoogLeNet obtained 0.97 (±0.01), and ResNet reached 0.96 (±0.01) in the dual TL scenario. Comparatively, the confocal TL scenario resulted in AUCs of 0.97 (±0.01) for AlexNet, 0.93 (±0.02) for GoogLeNet, and 0.94 (±0.01) for ResNet (Table 1).
Table 1. Network performance accuracy (%) assessment.
Table 1. Network performance accuracy (%) assessment.
Accuracy, Mean ± Standard Deviation (SD)AlexNetGoogLeNetResNetANOVA, p
Dual TL scenario94.97 ± 1.7691.43 ± 2.1789.87 ± 2.15<0.001
Confocal TL scenario90.14 ± 2.1385.71 ± 2.5584.65 ± 1.84<0.001
Student’s t-test, p<0.001<0.001<0.001
The ANOVA results indicated statistically significant differences between the two scenarios (p < 0.001), and Student’s t-test confirmed the statistical significance of these differences across all network architectures (p < 0.001) (Table 2).
Table 2. Network Performance AUC Assessment.
Table 2. Network Performance AUC Assessment.
AUC, Mean ± Standard Deviation (SD)AlexNetGoogLeNetResNetANOVA, p
Dual TL scenario0.98 ± 0.010.97 ± 0.010.96 ± 0.01<0.001
Confocal TL scenario0.97 ± 0.010.93 ± 0.020.94 ± 0.01<0.001
Student’s t-test, p<0.001<0.001<0.001

3.2. Confusion Matrix Analysis

Table 3 presents the confusion matrices for AlexNet, GoogLeNet, and ResNet under both the confocal TL and dual TL scenarios.
In the confocal TL scenario, AlexNet correctly classified 98.75% of the benign and 89.25% of the malignant cases. GoogLeNet correctly identified 91.5% of the benign and 88.5% of the malignant cases. ResNet achieved correct classifications for 85.25% of the benign and 92% of the malignant cases.
Conversely, in the dual TL scenario, AlexNet demonstrated enhanced performance by correctly classifying 98.5% of the benign (a 0.25% decrease) and 94% of the malignant cases. GoogLeNet improved to 96.75% of the benign and 85.5% (a 3% decrease) of the malignant correctly classified cases. ResNet maintained a rate of 85.25% of the benign and improved to a rate of 93.25% for the malignant correctly classified cases.
These results indicate that the dual TL approach results in a statistically significant reduction in the number of misclassifications, particularly for malignant cases, thereby enhancing the diagnostic accuracy and reliability of the models (Table 3).

3.3. Class Activation Mapping

Figure 3 illustrates the CAMs for both scenarios using the best-performing AlexNet architecture.

3.4. Summary of Performance Metrics

The comparative analysis of the network performances underscores the efficacy of the dual TL approach. Across all architectures, dual TL not only improved mean accuracy and AUC but also demonstrated greater consistency and reduced variability in performance metrics. The statistical analyses affirm that these improvements are not due to random chance, thereby validating the robustness and superiority of integrating dual imaging modalities through TL.

3.5. Statistical Significance

The comparative analysis between dual TL and confocal TL scenarios was rigorously evaluated using ANOVA and Student’s t-test, with all p-values reported as <0.001. This indicates a statistically significant improvement in both accuracy and AUC metrics when utilizing the dual TL approach across all network architectures. The substantial reduction in misclassifications, particularly in the malignant class, underscores the clinical relevance and potential of dual TL in enhancing diagnostic precision in lung cancer classification.

3.6. Overall Performance

The dual TL framework not only demonstrated higher accuracy and AUC but also exhibited greater consistency and reliability across multiple iterations. The reduced standard deviations in performance metrics suggest that dual TL mitigates the variability inherent in stochastic DL algorithms, thereby providing more stable and dependable diagnostic outputs.
The results highlight the superior performance of the dual TL approach in classifying lung cancer using pCLE and histological imaging. By leveraging complementary information from both imaging modalities, dual TL statistically significant enhances classification accuracy and AUC, offering a robust tool for early and precise lung cancer diagnosis.

4. Discussion

All resultant networks achieved accuracies and AUC values surpassing 80% and 0.8, respectively, indicating strong diagnostic capabilities. These results are broadly consistent with findings from other studies that reported comparable improvements in classification performance when applying TL to medical imaging tasks [40,41,42,43]. The influence of the chosen architecture remained statistically significant, with AlexNet yielding better outcomes than GoogLeNet and ResNet. This somewhat counterintuitive result, where a simpler architecture outperforms more complex ones, emphasizes the importance of careful design and optimization in TL workflows and resonates with observations made in other fields of medical image classification where architectural complexity does not always translate into superior results [44].
More importantly, regardless of the network architecture, the application of dual TL enhanced the classification accuracy by about 5%—a statistically significant improvement (p < 0.001)—and also increased the AUC by approximately 0.02. Similar improvements have been noted in other medical imaging contexts, where incorporating data from different sources or modalities contributed to better diagnostic metrics, as has been reported in various studies adopting TL to supplement limited domain-specific data [27]. Examination of the confusion matrices further underscored the benefits of dual TL, revealing fewer misclassifications for both benign and malignant categories. In particular, AlexNet exhibited a marked decrease in false benign predictions, while GoogLeNet, although showing a slight increase in false malignant predictions, reduced the false negative rate. These nuanced differences suggest that individual architectures may internalize the morphological cues provided by histology and pCLE to varying degrees, highlighting the value of combining complementary imaging data to fine-tune network decision boundaries.
The superior performance observed with dual TL was further supported by CAMs, which showed enlarged activation areas for both benign and malignant classes. By integrating histology-derived features into the learning process, the CNNs gained a more robust morphological foundation, enabling them to focus more accurately on diagnostically relevant regions within the pCLE images. The CAMs provide a clear view of how the two imaging modalities overlap, with class activations better highlighting fiber size/intersections. Such visualizations offer insight into the networks’ decision-making processes and illustrate how combining two imaging modalities enhances interpretability and reliability.
Our earlier research efforts have established that CNNs can accurately predict image diagnoses in various contexts, providing a strong basis for integrating multi-modal data [45,46,47,48]. While previous approaches often examined separate imaging modalities or made comparisons rather than integrations, the present dual TL approach leverages the synergy of histology and pCLE data to produce more refined diagnostic predictions. This integration aligns with prior explorations of tumor architecture, vascularization, and the interstitial fibrillary network using conventional and fractal dimension analyses [49,50,51,52,53]. The current findings reinforce those earlier conclusions, demonstrating that advanced DL frameworks can capture and extend these morphological correlations by combining the complementary strengths of histology and pCLE.
In the broader landscape of medical imaging, other researchers have consistently shown that carefully selected CNN architectures, like AlexNet, GoogLeNet, and ResNet, can effectively classify a range of pathologies once guided by TL [54,55,56,57,58]. More recent approaches have taken this further, using dual TL to integrate data from diverse sources and imaging modalities, thereby achieving remarkable accuracy and AUC values [59,60]. The improved metrics observed in this study reflect similar trends noted in multi-parametric imaging and radiomics-based analyses, which have successfully enhanced classification performance and mitigated the risk of overfitting [61,62,63,64,65,66].
The enhanced pCLE classifier could have direct clinical applications, allowing surgeons to apply the model’s output to resection specimens in real time. This would enable more accurate identification of clear resection margins, offering potential benefits in surgical outcomes.

4.1. Limitations

A primary limitation of this study lies in the reliance on relatively small and localized datasets. Although the dual TL approach substantially improved classification accuracy and AUC, these findings would benefit from confirmation on larger, more heterogeneous patient cohorts. Additionally, while the histology and pCLE imaging modalities provided complementary morphological cues, other data sources—such as molecular markers, clinical parameters, and advanced imaging sequences—could potentially further enhance diagnostic precision. The computational cost associated with repeatedly re-training CNNs remains another consideration, as does the inherent variability in histological interpretation and pCLE image acquisition. Future studies should also consider prospective, multi-center validations and incorporate interpretability strategies beyond CAM to better understand the underlying decision processes.
The results presented in Table 3 represent a randomly selected network from the 50 trained models. It is important to note that no specific constraints were imposed on the true positive or true negative rates for malignant cases during the training process. Consequently, the slight increase in false negatives observed in malignant classifications does not pose a significant concern. This is because the primary performance metric optimized during training was overall accuracy, which evaluates the model’s general classification performance across all classes. While false negatives are clinically relevant, their impact in this context is mitigated by the study’s focus on achieving high overall accuracy rather than specifically optimizing sensitivity or specificity for malignant cases. Future work could explore strategies, such as threshold adjustments or cost-sensitive learning, to further reduce false negatives without compromising accuracy.
Finally, the lack of standardized staining in the histological preparation may have an unknown impact on classification performance. It can be argued that the performance could improve with the use of standardized stain normalization. We previously proposed and validated a methodology for stain normalization [67], though it has not yet been applied clinically. Consequently, varying tissue preparation methods and staining techniques may potentially enhance classification performance.

4.2. Wrap-Up and Future Work

Worldwide, lung cancer remains the most common malignancy, with increasing incidence and mortality rates. Utilizing modern early diagnosis techniques is crucial as they may statistically significantly shorten the waiting time for histopathological results. The novel imaging technique—pCLE—has demonstrated its ability to accurately diagnose malignant lung lesions and with larger studies holds the potential to transform the role of the biopsy in lung endoscopy in the near future.
Future work should aim to address the current limitations by developing objective control mechanisms for histological staining processes to minimize variability. Additionally, efforts should be made to reduce the financial barriers associated with pCLE to facilitate its broader application and integration into routine diagnostic workflows. Expanding the dataset to include a larger and more diverse patient population will enhance the generalizability and robustness of the DL models. Furthermore, incorporating additional data modalities, such as molecular biomarkers and clinical parameters, could further improve diagnostic accuracy and model performance. Prospective, multi-center studies are also recommended to validate the TL approach across different clinical settings and populations, ensuring its applicability and scalability in real-world scenarios.

5. Conclusions

This study demonstrates that integrating histological image features through dual TL can enhance the diagnostic performance of CNN-based classification models applied to pCLE images in lung cancer. By incorporating complementary imaging modalities, the CNN models showed improvements in accuracy and AUC, with AlexNet achieving an increase in accuracy from 90.14 ± 2.13 to 94.97 ± 1.76, GoogLeNet from 85.71 ± 2.55 to 91.43 ± 2.17, and ResNet from 84.65 ± 1.84 to 89.87 ± 2.15. These results suggest that a multi-modal, TL-based approach may contribute to more reliable and interpretable diagnostic outputs, providing potential benefits for early and precise lung cancer detection.
However, the study has certain limitations that warrant further investigation. Specifically, the small dataset size, computational demands, and the lack of standardized histological staining protocols may affect the generalizability of the findings. Future work should address these limitations by incorporating larger, more diverse datasets and exploring methods to standardize preprocessing and staining. Ultimately, further studies are necessary to validate the proposed approach and assess its clinical utility in improving diagnostic outcomes.

Author Contributions

Conceptualization, C.-T.S., M.-S.Ș. and A.D.D.; methodology, L.S. and A.-G.G.; software, M.-S.Ș. and M.M.; validation, A.-G.G., L.S., D.-N.P. and M.-S.Ș.; formal analysis, M.-S.Ș.; investigation, A.D.D. and D.-N.P.; resources, C.-T.S.; data curation, M.-S.Ș. and A.-G.G.; writing—original draft preparation, M.-S.Ș., L.S., C.-T.S. and A.-G.G.; writing—review and editing, C.-T.S. and M.-S.Ș.; visualization, M.-S.Ș.; supervision, C.-T.S. and A.D.D.; project administration, L.S. and C.-T.S.; funding acquisition, L.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the MEN, Romania, grant number 26/23C/13.07.2021 titled Complex diagnostic system for lung, hepatic and colorectal malignant tumors. The article processing charges were funded by the University of Medicine and Pharmacy of Craiova, Romania.

Institutional Review Board Statement

This observational prospective study was conducted according to the guidelines of the Declaration of Helsinki, and ethical approval was obtained from both the County Clinical Hospital of Craiova (approval 26659/08.07.2020) and from the University of Medicine and Pharmacy of Craiova (42/17.06.2020).

Informed Consent Statement

All patients provided informed consent prior to inclusion.

Data Availability Statement

Data is available from the corresponding authors, upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIartificial intelligence
AUCarea under the curve
CAMclass activation mapping
CNNconvolutional neural network
DLdeep learning
LDCTlow-dose computed tomography
pCLEconfocal laser endomicroscopy
TLtransfer learning

References

  1. Cai, W.; Zhu, X.; Li, Y.; Xu, Y. Interpretation of global lung cancer statistics. J. Thorac. Oncol. 2024, 19, 562–569. [Google Scholar]
  2. Barta, J.A.; Powell, C.A.; Wisnivesky, J.P. Global Epidemiology of Lung Cancer. Ann. Glob. Health 2019, 85, 8. [Google Scholar] [CrossRef]
  3. Collins, L.G.; Haines, C.; Perkel, R.; Enck, R.E. Lung cancer: Diagnosis and management. Am. Fam. Physician 2007, 75, 56–63. [Google Scholar]
  4. Malhotra, J.; Malvezzi, M.; Negri, E.; La Vecchia, C.; Boffetta, P. Risk factors for lung cancer worldwide. Eur. Respir. J. 2016, 48, 889–902. [Google Scholar] [CrossRef] [PubMed]
  5. Lipfert, F.W.; Wyzga, R.E. Longitudinal relationships between lung cancer mortality and air pollution. Risk Anal. 2019, 39, 1646–1664. [Google Scholar]
  6. Maisonneuve, P.; Rampinelli, C.; Bertolotti, R.; Melloni, G.; Pelosi, G.; de Braud, F.; Spaggiari, L.; Pastorino, U. Low-dose computed tomography screening for lung cancer in smokers. JAMA 2019, 322, 584–594. [Google Scholar]
  7. Gouvinhas, C.; De Mello, R.A.; Oliveira, D.; Castro-Lopes, J.M.; Castelo-Branco, P.; Dos Santos, R.S.; Hespanhol, V.; Pozza, D.H. Lung cancer: A brief review of epidemiology and screening strategies. Future Oncol. 2018, 14, 567–575. [Google Scholar] [CrossRef] [PubMed]
  8. Shankar, A.; Dubey, A.; Saini, D.; Singh, M.; Prasad, C.P.; Roy, S.; Mandal, C.; Bhandari, S.; Kumar, S.; Lathwal, A.; et al. Environ-mental and occupational determinants of lung cancer. Transl. Lung Cancer Res. 2019, 8 (Suppl. S1), S31–S49. [Google Scholar] [CrossRef]
  9. Hocking, W.G. Integrating prevention and screening for lung cancer into clinical practice. Chest 2013, 143, 1216–1224. [Google Scholar]
  10. D’Urso, D.; Doneddu, G.; Lojacono, M.; Tanda, F.; Serra, M.; Marras, V.; Cossu, A.; Corda, S.; Piredda, S.; Deidda, M.; et al. Sputum analysis for non-invasive early lung cancer detection. Respir. Med. 2013, 107, 853–858. [Google Scholar] [CrossRef]
  11. Lebrett, M.B.; Crosbie, P.A. Targeting lung cancer screening to individuals at highest risk. Lancet Respir. Med. 2020, 8, 226–228. [Google Scholar]
  12. Takemura, S.; Kurimoto, N.; Miyazawa, T.; Ohta, K.; Hino, H.; Murata, Y.; Takai, Y.; Kaneko, Y.; Maekura, R. Probe-based confocal laser endomicroscopy for rapid diagnosis of pulmonary nodules. Respir. Res. 2019, 20, 162. [Google Scholar]
  13. Chaudoir, B.R.; Brandi, C.; McGill, S.; Smith, R.A.; Wilson, J.; Park, G.; Klein, C. Frontier in pathology: Comparison of pCLE and biopsy. J. Thorac. Dis. 2014, 6, 523–530. [Google Scholar]
  14. Wellikoff, A.S.; Holladay, R.C.; Downie, G.H. Comparison of in vivo probe-based confocal laser endomicroscopy with biopsy. Respiration 2015, 90, 205–212. [Google Scholar]
  15. Yserbyt, J.; Dooms, C.; Ninane, V. Perspectives using probe-based confocal laser endomicroscopy. Respiration 2013, 85, 304–310. [Google Scholar]
  16. Danilevskaya, O.V.; Sazonov, D.V.; Zabozlaev, F.G.; Averyanov, A.V.; Sorokina, A.; Sotnikova, A.G.; Urazovsky, N.; Kuzovlev, O.P.; Shablovsky, O.R. Confocal laser endomicroscopy in diagnosis of solitary and multiple pulmonary nodular infiltrates. Eur. Respir. J. 2012, 40, 660. [Google Scholar]
  17. Shafiek, H.; Fiorentino, F.; Larici, A.; Ravaglia, C.; Poletti, V.; Tognini, G.; Roviaro, G.; Fanti, S.; Neri, E.; Fiume, D. Usefulness of bron-choscopic probe-based confocal laser endomicroscopy in pneumonia. Respir. Res. 2016, 17, 51. [Google Scholar]
  18. Yserbyt, J.; Dooms, C.; Verleden, G.M.; Vanaudenaerde, B.M.; Vos, R.; Weynand, B.; Decramer, M.; Verleden, S.E. Probe-based confocal laser endomicroscopy in acute lung rejection. Am. J. Transplant. 2011, 11, 2466–2472. [Google Scholar]
  19. Salaün, M.; Guisier, F.; Lena, H.; Thiberville, L. In vivo probe-based confocal laser endomicroscopy in pulmonary pathology. Respir. Res. 2019, 20, 225. [Google Scholar]
  20. Buchner, A.M.; Gómez, V.; Heckman, M.G.; Shah, R.J.; Schueler, B.A.; Ghabril, M.S.; Raimondo, M.; Krishna, M.; Wallace, M.B. The learning curve of in vivo probe-based confocal laser endomicroscopy. Endoscopy 2009, 41, 902–908. [Google Scholar]
  21. Nóbrega, R.V.M.; Peixoto, S.A.; Silva, S.P.; Filho, P.P. Lung nodule classification via deep transfer learning in CT Lung Images. In Proceedings of the 2018 IEEE 31st International Symposium on Computer-Based Medical Systems (CBMS), Karlstad, Sweden, 18–21 June 2018; pp. 244–249. [Google Scholar]
  22. Lakhani, P.; Sundaram, B. Deep learning at chest radiography: Automated classification of tuberculosis. Radiology 2017, 284, 574–582. [Google Scholar] [CrossRef] [PubMed]
  23. Lin, Z. Interpretability study of pretrained models via transfer learning for lung cancer prediction. Proc. SPIE Med. Imaging 2022, 12451, 124514X. [Google Scholar]
  24. Zhang, Z. The transferability of transfer learning model based on ImageNet for medical image classification tasks. Appl. Comput. Eng. 2023, 18, 143–151. [Google Scholar] [CrossRef]
  25. Osmani, N.; Rezayi, S.; Esmaeeli, E.; Karimi, A. Transfer learning from non-medical images to medical images. Front. Health Inform. 2024, 13, 177. [Google Scholar] [CrossRef]
  26. Romero, M.; Interian, Y.; Solberg, T.; Valdes, G. Targeted transfer learning for small medical datasets. Med. Phys. 2019, 46, 2328–2337. [Google Scholar]
  27. Bungărdean, R.M.; Şerbănescu, M.-S.; Streba, C.T.; Crişan, M. Deep learning with transfer learning in pathology. Case study: Classification of basal cell carcinoma. Rom. J. Morphol. Embryol. 2022, 62, 1017–1028. [Google Scholar] [CrossRef]
  28. Celik, Y.; Talo, M.; Yıldırım, Ö.; Karabatak, M.; Acharya, U.R. Automated invasive ductal carcinoma detection using deep transfer learning. Pattern Recognit. Lett. 2020, 133, 232–239. [Google Scholar] [CrossRef]
  29. Anusha, M.; Reddy, D.S. Enhancing lung and colon cancer diagnosis using ImageNet-trained transfer learning. In Proceedings of the 2024 Tenth International Conference on Bio Signals, Images, and Instrumentation (ICBSII), Chennai, India, 20–22 March 2024; pp. 1–4. [Google Scholar]
  30. Said, M.M.; Islam, M.S.; Sumon, M.S.I.; Vranić, S.; Al Saady, R.M.; Alqahtani, A.; Chowdhury, M.; Pedersen, S. Innovative deep learning for lung and colon cancer classification. Appl. Comput. Intell. Soft. Comput. 2024, 2024, 5562890. [Google Scholar]
  31. Gao, R.; Tang, Y.; Xu, K.; Kammer, M.; Antic, S.; Deppen, S.; Sandler, K.; Massion, P.; Huo, Y.; Landman, B.A. Deep multi-path network integrating incomplete biomarker and chest CT data. Proc. SPIE Med. Imaging 2020, 11596, 115961E. [Google Scholar]
  32. MathWorks. AlexNet. Available online: https://www.mathworks.com/help/deeplearning/ref/alexnet.html (accessed on 28 January 2025).
  33. MathWorks. GoogLeNet. Available online: https://www.mathworks.com/help/deeplearning/ref/googlenet.html (accessed on 28 January 2025).
  34. MathWorks. ResNet-18. Available online: https://www.mathworks.com/help/deeplearning/ref/resnet18.html (accessed on 28 January 2025).
  35. ImageNet. Available online: http://www.image-net.org/ (accessed on 28 January 2025).
  36. Nica, R.-E.; Șerbănescu, M.-S.; Florescu, L.-M.; Camen, G.-C.; Streba, C.T.; Gheonea, I.-A. Deep Learning: A Promising Method for Histological Class Prediction of Breast Tumors in Mammography. J. Digit. Imaging 2021, 34, 1190–1198. [Google Scholar] [CrossRef]
  37. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  38. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convo-lutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 8–12 June 2015; pp. 1–9. [Google Scholar]
  39. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  40. Tortora, M.; Cordelli, E.; Sicilia, R.; Nibid, L.; Ippolito, E.; Perrone, G.; Ramella, S.; Soda, P. RadioPathomics: Multimodal Learning in Non-Small Cell Lung Cancer. arXiv 2022, arXiv:2204.12423. [Google Scholar]
  41. Mercuţ, R.; Ciurea, M.E.; Traşcă, E.T.; Ionescu, M.; Mercuţ, M.F.; Rădulescu, P.M.; Călăraşu, C.; Streba, L.; Ionescu, A.G.; Rădulescu, D. Applying Neural Networks to Analyse Inflammatory, Sociodemographic, and Psychological Factors in Non-Melanoma Skin Cancer and Colon Cancer: A Statistical and Artificial Intelligence Approach. Diagnostics 2024, 14, 2759. [Google Scholar] [CrossRef]
  42. Radulescu, D.; Calafeteanu, D.M.; Radulescu, P.-M.; Boldea, G.-J.; Mercut, R.; Ciupeanu-Calugaru, E.D.; Georgescu, E.-F.; Boldea, A.M.; Georgescu, I.; Caluianu, E.-I.; et al. Enhancing the Understanding of Abdominal Trauma During the COVID-19 Pandemic Through Co-Occurrence Analysis and Machine Learning. Diagnostics 2024, 14, 2444. [Google Scholar] [CrossRef] [PubMed]
  43. Şerbănescu, M.-S.; Manea, N.C.; Streba, L.; Belciug, S.; Pleşea, I.E.; Pirici, I.; Bungărdean, R.M.; Pleşea, R.M. Automated Gleason grading of prostate cancer using transfer learning from general-purpose deep-learning networks. Rom. J. Morphol. Embryol. 2020, 61, 149–155. [Google Scholar] [CrossRef]
  44. Şerbănescu, M.-S.; Oancea, C.-N.; Streba, C.T.; Pleşea, I.E.; Pirici, D.; Streba, L.; Pleşea, R.M. Agreement of two pre-trained deep-learning neural networks built with transfer learning with six pathologists on 6000 patches of prostate cancer from Gleason2019 Challenge. Rom. J. Morphol. Embryol. 2020, 61, 513–519. [Google Scholar] [CrossRef]
  45. Șerbănescu, M.-S.; Bungărdean, R.M.; Georgiu, C.; Crișan, M. Nodular and Micronodular Basal Cell Carcinoma Subtypes Are Different Tumors Based on Their Morphological Architecture and Their Interaction with the Surrounding Stroma. Diagnostics 2022, 12, 1636. [Google Scholar] [CrossRef] [PubMed]
  46. Bungărdean, R.-M.; Şerbănescu, M.-S.; Colosi, H.A.; Crişan, M. High-frequency ultrasound: An essential non-invasive tool for the pre-therapeutic assessment of basal cell carcinoma. Rom. J. Morphol. Embryol. 2022, 62, 545–551. [Google Scholar] [CrossRef] [PubMed]
  47. Mitroi, G.; Pleşea, R.M.; Pop, O.T.; Ciovică, D.V.; Şerbănescu, M.S.; Alexandru, D.O.; Stoiculescu, A.; Pleşea, I.E. Correlations between intratumoral interstitial fibrillary network and vascular network in Srigley patterns of prostate adenocarcinoma. Rom. J. Morphol. Embryol. 2015, 56, 1319–1328. [Google Scholar] [PubMed]
  48. Pleşea, I.E.; Stoiculescu, A.; Serbănescu, M.; Alexandru, D.O.; Man, M.; Pop, O.T.; Pleşea, R.M. Correlations between intratumoral vascular network and tumoral architecture in prostatic adenocarcinoma. Rom. J. Morphol. Embryol. 2013, 54, 299–308. [Google Scholar]
  49. Stoiculescu, A.; Plesea, I.E.; Pop, O.T.; Alexandru, D.O.; Man, M.; Serbanescu, M.; Plesea, R.M. Correlations between intratumoral in-terstitial fibrillary network and tumoral architecture in prostatic adenocarcinoma. Rom. J. Morphol. Embryol. 2012, 53, 941–950. [Google Scholar] [PubMed]
  50. Plesea, R.M.; Serbanescu, M.-S.; Ciovica, D.V.; Rosu, G.-C.; Moldovan, V.T.; Bungardean, R.M.; Popescu, N.A.; Plesea, I.E. The study of tumor architecture components in prostate adenocarcinoma using fractal dimension analysis. Rom. J. Morphol. Embryol. 2019, 60, 501–519. [Google Scholar] [PubMed]
  51. Hassan, S.A.A.; Sayed, M.S.; Abdalla, M.I.; Rashwan, M.A. Breast cancer masses classification using deep convolutional neural networks and transfer learning. Multimed. Tools Appl. 2020, 79, 30735–30768. [Google Scholar] [CrossRef]
  52. Jiang, Y.; Chen, L.; Zhang, H.; Xiao, X. Breast cancer histopathological image classification using convolutional neural networks with small SE-ResNet modules. Int. J. Comput. Assist. Radiol. Surg. 2017, 12, 1479–1486. [Google Scholar] [CrossRef]
  53. Mazo, C.; Bernal, J.; Trujillo, M.; Alegre, E. Transfer learning for classification of cardiovascular tissues in histological images. Comput. Methods Programs Biomed. 2018, 165, 69–76. [Google Scholar] [CrossRef]
  54. Neamah, K.; Mohamed, F.; Waheed, S.R.; Kurdi, W.H.M.; Taha, A.Y.; Kadhim, K.A. Utilizing Deep Improved ResNet-50 for Brain Tumor Classification Based on MRI. IEEE Open J. Comput. Soc. 2024, 5, 446–456. [Google Scholar] [CrossRef]
  55. Saleh, L.; Zhang, L. Medical Image Classification Using Transfer Learning and Network Pruning Algorithms. In Proceedings of the 2023 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Honolulu, HI, USA, 1–4 October 2023; pp. 1932–1938. [Google Scholar]
  56. Anusha, M.; Reddy, D.S. Enhancing Lung and Colon Cancer Diagnosis: An ImageNet-Trained Transfer Learning Approach for Histopathological Image Analysis. In Proceedings of the 2024 Tenth International Conference on Bio Signals, Images, and Instrumentation (ICBSII), Chennai, India, 20–22 March 2024; pp. 1–4. [Google Scholar]
  57. Boumaraf, S.; Liu, X.; Zheng, Z.; Ma, X.; Ferkous, C. A new transfer learning-based approach to magnification dependent and independent classification of breast cancer in histopathological images. Biomed. Signal Process. Control 2021, 63, 102192. [Google Scholar] [CrossRef]
  58. Leung, K.H.; Rowe, S.P.; Sadaghiani, M.S.; Leal, J.P.; Mena, E.; Choyke, P.L.; Du, Y.; Pomper, M.G. Deep Semisupervised Transfer Learning for Fully Automated Whole-Body Tumor Quantification and Prognosis of Cancer on PET/CT. J. Nucl. Med. 2024, 65, 643–650. [Google Scholar] [CrossRef] [PubMed]
  59. Hu, Q.; Whitney, H.M.; Giger, M.L. A deep learning methodology for improved breast cancer diagnosis using multiparametric MRI. Sci. Rep. 2020, 10, 10536. [Google Scholar] [CrossRef]
  60. Xu, Y.; Jia, Z.; Wang, L.-B.; Ai, Y.; Zhang, F.; Lai, M.; Chang, E.I.-C. Large scale tissue histopathology image classification, segmentation, and visualization via deep convolutional activation features. BMC Bioinform. 2019, 19, 281. [Google Scholar] [CrossRef]
  61. Syed, A.H.; Khan, T.; Khan, S.A. Deep Transfer Learning Techniques-Based Automated Classification and Detection of Pulmonary Fibrosis from Chest CT Images. Processes 2023, 11, 443. [Google Scholar] [CrossRef]
  62. Shi, Z.; Hao, H.; Zhao, M.; Feng, Y.; He, L.; Wang, Y.; Suzuki, K. A deep CNN-based transfer learning method for false positive reduction. Multimed. Tools Appl. 2019, 78, 1017–1033. [Google Scholar] [CrossRef]
  63. Jin, H.; Li, Z.; Tong, R.; Lin, L. A deep 3D residual CNN for false-positive reduction in pulmonary nodule detection. Med. Phys. 2018, 45, 2097–2107. [Google Scholar] [CrossRef]
  64. Noaman, N.F.; Kanber, B.M.; Al Smadi, A.; Jiao, L.; Alsmadi, M.K. Advancing Oncology Diagnostics: AI-Enabled Early Detection of Lung Cancer Through Hybrid Histological Image Analysis. IEEE Access 2024, 12, 64396–64415. [Google Scholar] [CrossRef]
  65. Setio, A.A.; Ciompi, F.; Litjens, G.J.; Gerke, P.K.; Jacobs, C.; Riel, S.J.; Wille, M.M.; Naqibullah, M.; Sánchez, C.I.; Ginneken, B.V. Pulmonary Nodule Detection in CT Images: False Positive Reduction Using Multi-View Convolutional Networks. IEEE Trans. Med. Imaging 2016, 35, 1160–1169. [Google Scholar] [CrossRef] [PubMed]
  66. Chaunzwa, T.L.; Hosny, A.; Xu, Y.; Shafer, A.; Diao, N.; Lanuti, M.; Christiani, D.C.; Mak, R.H.; Aerts, H.J.W.L. Deep learning classification of lung cancer histology using CT images. Eur. J. Radiol. 2021, 145, 110013. [Google Scholar] [CrossRef]
  67. Şerbănescu, M.S.; Pleşea, I.E. A hardware approach for histological and histopathological digital image stain normalization. Rom. J. Morphol. Embryol. 2015, 56 (Suppl. S2), 735–741. [Google Scholar]
Figure 1. Representative samples from the dataset: (AD) Images from the histology dataset, where (A,B) represent benign samples and (C,D) represent malignant samples. (EH) Images from the confocal dataset, where (E,F) correspond to benign samples and (G,H) correspond to malignant samples.
Figure 1. Representative samples from the dataset: (AD) Images from the histology dataset, where (A,B) represent benign samples and (C,D) represent malignant samples. (EH) Images from the confocal dataset, where (E,F) correspond to benign samples and (G,H) correspond to malignant samples.
Cancers 17 00611 g001
Figure 2. Study design schematic: The dual TL approach integrates features from histological and pCLE image datasets to improve classification accuracy. Path A (yellow) applies a single TL step directly on the pCLE image dataset. Path B (green) involves two sequential TL steps: the first is performed on the histological image dataset, followed by the second TL on the pCLE image dataset. The performance of both approaches is evaluated using a 10-fold cross-validation framework.
Figure 2. Study design schematic: The dual TL approach integrates features from histological and pCLE image datasets to improve classification accuracy. Path A (yellow) applies a single TL step directly on the pCLE image dataset. Path B (green) involves two sequential TL steps: the first is performed on the histological image dataset, followed by the second TL on the pCLE image dataset. The performance of both approaches is evaluated using a 10-fold cross-validation framework.
Cancers 17 00611 g002
Figure 3. CAM visualization: Panels (A,B) depict CAMs for the confocal TL scenario, while panels (C,D) correspond to the dual TL scenario, both utilizing the best-performing AlexNet architecture. The highlighted regions indicate image areas that contribute to the classification decision, particularly fiber size and overlapping structures, offering insights into the model’s decision-making process.
Figure 3. CAM visualization: Panels (A,B) depict CAMs for the confocal TL scenario, while panels (C,D) correspond to the dual TL scenario, both utilizing the best-performing AlexNet architecture. The highlighted regions indicate image areas that contribute to the classification decision, particularly fiber size and overlapping structures, offering insights into the model’s decision-making process.
Cancers 17 00611 g003
Table 3. Confusion matrices of selected networks in both scenarios.
Table 3. Confusion matrices of selected networks in both scenarios.
AlexNetGoogLeNetResNet
BenignMalignantBenignMalignantBenignMalignant
Actual classConfocal TL scenarioBenign395433664634132
Malignant53573435459368
Dual TL scenarioBenign394243875834127
Malignant63761334259373
Predicted class
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Șerbănescu, M.-S.; Streba, L.; Demetrian, A.D.; Gheorghe, A.-G.; Mămuleanu, M.; Pirici, D.-N.; Streba, C.-T. Transfer Learning-Based Integration of Dual Imaging Modalities for Enhanced Classification Accuracy in Confocal Laser Endomicroscopy of Lung Cancer. Cancers 2025, 17, 611. https://doi.org/10.3390/cancers17040611

AMA Style

Șerbănescu M-S, Streba L, Demetrian AD, Gheorghe A-G, Mămuleanu M, Pirici D-N, Streba C-T. Transfer Learning-Based Integration of Dual Imaging Modalities for Enhanced Classification Accuracy in Confocal Laser Endomicroscopy of Lung Cancer. Cancers. 2025; 17(4):611. https://doi.org/10.3390/cancers17040611

Chicago/Turabian Style

Șerbănescu, Mircea-Sebastian, Liliana Streba, Alin Dragoș Demetrian, Andreea-Georgiana Gheorghe, Mădălin Mămuleanu, Daniel-Nicolae Pirici, and Costin-Teodor Streba. 2025. "Transfer Learning-Based Integration of Dual Imaging Modalities for Enhanced Classification Accuracy in Confocal Laser Endomicroscopy of Lung Cancer" Cancers 17, no. 4: 611. https://doi.org/10.3390/cancers17040611

APA Style

Șerbănescu, M.-S., Streba, L., Demetrian, A. D., Gheorghe, A.-G., Mămuleanu, M., Pirici, D.-N., & Streba, C.-T. (2025). Transfer Learning-Based Integration of Dual Imaging Modalities for Enhanced Classification Accuracy in Confocal Laser Endomicroscopy of Lung Cancer. Cancers, 17(4), 611. https://doi.org/10.3390/cancers17040611

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop