Next Article in Journal
The Role of Ultrasound Examination in the Assessment of Suitability of Calcified Arteries for Vascular Access Creation—Mini Review
Previous Article in Journal
An Automated Deep Learning Approach for Spine Segmentation and Vertebrae Recognition Using Computed Tomography Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep-Learning-Based Visualization and Volumetric Analysis of Fluid Regions in Optical Coherence Tomography Scans

by
Harishwar Reddy Kasireddy
1,
Udaykanth Reddy Kallam
1,
Sowmitri Karthikeya Siddhartha Mantrala
1,
Hemanth Kongara
1,
Anshul Shivhare
1,
Jayesh Saita
2,
Sharanya Vijay
2,
Raghu Prasad
2,
Rajiv Raman
3 and
Chandra Sekhar Seelamantula
1,*
1
Department of Electrical Engineering, Indian Institute of Science, Bengaluru 560012, India
2
Carl Zeiss India Pvt. Ltd., Bengaluru 560099, India
3
Shri Bhagwan Mahavir Vitreoretinal Services, Sankara Nethralaya, Chennai 600006, India
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(16), 2659; https://doi.org/10.3390/diagnostics13162659
Submission received: 31 May 2023 / Revised: 27 July 2023 / Accepted: 28 July 2023 / Published: 12 August 2023
(This article belongs to the Section Pathology and Molecular Diagnostics)

Abstract

:
Retinal volume computation is one of the critical steps in grading pathologies and evaluating the response to a treatment. We propose a deep-learning-based visualization tool to calculate the fluid volume in retinal optical coherence tomography (OCT) images. The pathologies under consideration are Intraretinal Fluid (IRF), Subretinal Fluid (SRF), and Pigmented Epithelial Detachment (PED). We develop a binary classification model for each of these pathologies using the Inception-ResNet-v2 and the small Inception-ResNet-v2 models. For visualization, we use several standard Class Activation Mapping (CAM) techniques, namely Grad-CAM, Grad-CAM++, Score-CAM, Ablation-CAM, and Self-Matching CAM, to visualize the pathology-specific regions in the image and develop a novel Ensemble-CAM visualization technique for robust visualization of OCT images. In addition, we demonstrate a Graphical User Interface that takes the visualization heat maps as the input and calculates the fluid volume in the OCT C-scans. The volume is computed using both the region-growing algorithm and selective thresholding technique and compared with the ground-truth volume based on expert annotation. We compare the results obtained using the standard Inception-ResNet-v2 model with a small Inception-ResNet-v2 model, which has half the number of trainable parameters compared with the original model. This study shows the relevance and usefulness of deep-learning-based visualization techniques for reliable volumetric analysis.

1. Introduction

Intraretinal Fluid (IRF), Subretinal Fluid (SRF), and Pigmented Epithelial Detachment (PED) are three of the most prevalent retinal pathologies [1,2,3]. Quantitative measures of retinal fluids are important biomarkers, and fluid volume computation from optical coherence tomography (OCT) scans by artificial intelligence (AI) algorithms can guide the treatment of exudative retinal conditions, such as Neovascular Age-related Macular Degeneration (NAMD). In NAMD, pathological fluids accumulate in different compartments of the retina—IRF inside the retina, SRF beneath the neurosensory retina, and PED between the Retinal Pigment Epithelium (RPE) and Bruch’s membrane (BM). Different types of OCT technologies are used to capture the retina of the eye, such as time-domain, spectral-domain, and swept-source OCT [4]. In this paper, we use data generated from commercially available CIRRUS and PRIMUS OCT machines, which are spectral-domain-based OCT machines.
OCT captures structural information of the retina using the principle of coherent detection. Coherent detection produces multiplicative noise, known as speckle in the acquired images. Speckle lowers image quality and makes visual and automated analysis of OCT images challenging [5]. The noise statistics in OCT images play a significant role in developing suitable despeckling algorithms. Various data distributions have been proposed in the literature for modeling speckles in OCT images. According to Bashkansky and Reintjes [6], speckle noise follows a Gaussian distribution. Pircher et al. [7] and Karamata et al. [8] suggested that speckle noise follows a Rayleigh distribution. Schmitt et al. [9] proposed an exponentially decaying distribution for speckle. Sudeep et al. [10] proposed a denoising method based on the Gamma distribution for speckle reduction in OCT images. In this paper, we model the speckle noise as gamma distribution and perform denoising of OCT images as the first step.
Deep learning has produced remarkable advancements in the field of healthcare [11,12,13,14], especially in the automatic detection of retinal abnormalities such as glaucoma, diabetic retinopathy, Diabetic Macular Edema (DME), Age-related Macular Degeneration (AMD), and Retinal Vein Occlusion (RVO). Deep-learning-based classifiers are used to detect the relationship between the multiple input features in the image and analyze the relationships between these features to make predictions or classifications. By leveraging the complex patterns and representations learned by the model, it can provide an overall prediction based on these detected features. However, for deploying these classifiers in the real world, there is a need to provide an explanation of the relationships captured among the features of the image. This problem of explainable artificial intelligence (XAI) can be addressed to a certain extent by using appropriate visualization tools. Techniques such as Grad-CAM [15], Grad-CAM++ [16], Score-CAM [17], Ablation-CAM [18], and Selfmatching- CAM [19] are used to explain which features of the image are significant contributors to the prediction score. Explainability enhances the confidence of doctors in deploying AI models in real-world applications. The models are otherwise considered opaque or black-box type. Explainability in AI has now become an indispensable part of the deployment pipeline, because when deploying a deep learning model in real-world scenarios, it is crucial to be able to explain and interpret the model’s decisions or predictions, particularly when dealing with medical applications. Explainability in AI is necessary to provide insights and justifications to doctors or medical professionals, enabling them to understand why a certain deep learning model is classifying an image into a specific pathology. Through explainability, doctors can gain trust and confidence in the model’s capabilities, facilitating its adoption and integration into clinical practice.
There are several studies available in the literature that proposed automatic volume computation of retinal fluid. Typically, one performs segmentation for calculating the fluid volume, for instance, Yukun et al.’s Retinal Fluid Segmentation Network (ReF-Net) [20], Thomas et al.’s encoder–decoder-based segmentation network [21], Wilson et al.’s U-Net-based architecture [22], and Lu et al.’s automated segmentation technique [23]. To the best of our knowledge, ours is the first attempt at adopting the visualization outcome of a deep learning model for retinal fluid volume computation instead of a segmentation output. Typically, one employs a segmentation-based approach for calculating the retinal fluid volume. However, our methodology is significantly different, as we utilize the results obtained from visualization techniques. We also present a Graphical User Interface (GUI) for automatic fluid volume computation. We consider GradCAM, GradCAM++, Score-CAM, Ablation-CAM, and Self-Matching-CAM as visualization techniques for the binary classification models, which detect IRF, SRF, and PED pathologies in the retinal OCT images. The volume is computed separately for IRF, SRF, and PED. In addition, we also propose a novel fusion technique, namely Ensemble-CAM, which is a robust visualization technique.
A combination of IRF, SRF, and PED, as well as fluid volume variations, are used to categorize Age-related Macular Degeneration (AMD) patient response to antivascular endothelial growth factor (Anti-VEGF) therapy. The AMD patient is categorized as a Responder if the post-treatment fluid volume has reduced by more than 10% compared with the pretreatment fluid volume and as a Nonresponder if the post-treatment fluid volume has increased or remained the same or reduced by less than 10% compared with the pretreatment fluid volume. The automatic quantification of fluid response may be a better means of monitoring than the current qualitative evaluation used in clinical practice. The computed fluid volume could be a potential input to an AI model to predict the AMD patient’s response to an Anti-VEGF injection [24]. The AI response prediction can help clinicians in deciding the form of the treatment and could be a beneficial tool in the clinical armamentarium for the therapeutic management of AMD.
Section 2 contains a description of the dataset used for classification and visualization tasks and the methodology used in the paper. Section 3 presents the results obtained for classification, visualization, and fluid volume computation, and Section 4 contains a discussion of the results. Concluding remarks are given in Section 5.
Figure 1 shows an OCT image in which all three pathologies, namely IRF, SRF, and PED, are present.

2. Materials and Methods

2.1. Dataset

In this section, we present details of the datasets used for classification and visualization.

2.1.1. Classification Dataset

The dataset used in this study is proprietary to Carl Zeiss and is the same as the one used in our recent publication [25]. It comprises OCT B-scan images stacked as cubes. The images are obtained using two brands of Carl Zeiss machines—CIRRUS and PRIMUS. There are, in turn, two variants of CIRRUS, one that captures 200 images per cube and an other that captures 128 images per cube. There are two types of PRIMUS machines, of which one captures 128 images per cube and an other that captures 32 images per cube. The training and test data for classification consist of 207,535 and 52,650 images, respectively. The B-scan images from CIRRUS and PRIMUS are of size 1024 × 512 and 1024 × 200, respectively.

2.1.2. Segmentation Dataset

The ground-truth segmentation data required for visualization is available for 73 cubes, each, in turn, containing 128 B-scans, making up a total of 9344 (73 × 128) images. Of these, 1235, 1361, and 3686 images have IRF, SRF, and PED pathologies, respectively.

2.1.3. Performance Measures

The segmentation ground truth is given by an expert and is used for the computation of the Jaccard Index/Intersection over Union (IoU) [26] and Dice score [27,28] given as follows:
J a c c a r d ( A , B ) = | A B | | A B | ,
and
D i c e ( A , B ) = 2 | A B | | A | + | B | ,
where A is the generated mask, B is the ground truth mask, and | · | represents the cardinality of the set.

2.2. Architecture of the Model

In this work, we use the Inception-ResNet-v2 [29] to develop the binary classification models for each pathology. Figure 2a shows the standard Inception-ResNet-v2 architecture divided into several parts, such as stem, Inception-A, Inception-B, Inception-C, Reduction-A, and Reduction-B, each explained in the following:
  • Stem: The stem of the Inception-Resnet-v2 architecture consists of a convolutional layer followed by a max pooling layer, and several convolutional layers with increasing dilation rates.
  • Inception-ResNet-A: The Inception-A blocks consist of multiple branches, each of which applies a different type of operation (such as convolution or pooling) to the input. These branches are then concatenated together and passed through a final convolutional layer.
  • Inception-ResNet-B: The Inception-B blocks are similar to Inception-A blocks, but with a different set of operations. It uses an average pooling operation followed by a convolutional layer.
  • Inception-ResNet-C: Inception-C blocks are also similar to Inception-B blocks but with a different set of operations and a larger number of filters.
  • Reduction: Reduction-A and Reduction-B blocks reduce the spatial dimensions of the feature maps. They consist of several convolutional layers and max pooling layers.
Finally, the output of the last block is fed into an average pooling layer, followed by a dropout layer and a fully connected layer, which produces the final output. The Inception-ResNet-v2 is trained on 1 million images from ImageNet [30] dataset with 1000 classes. We fine-tuned the model using the training dataset with two classes.

2.3. Comparison with Small Inception-ResNet-v2

We developed a novel small Inception-ResNet-v2 architecture, which has a similar skeleton to the original Inception-ResNet-v2 but with half the number of trainable parameters (28 million parameters as opposed to 56 million trainable parameters). Ten Inception-ResNet-A, twenty Inception-ResNet-B, and ten Inception-ResNet-C layers in Inception-ResNet-v2 architecture are replaced by eight Inception-ResNet-A, seven Inception-ResNet-B, and four Inception-ResNet-C layers, respectively, to obtain the small Inception-ResNet-v2 architecture shown in Figure 2b. Figure 2c shows the stem.

2.4. Denoising

The OCT images obtained from the OCT scanners are typically noisy. Hence, there is a need to perform mild denoising of the images. If the models are trained directly on noisy images, there is a chance that the model might fit the noise in the image rather than the pathological features of interest. After experimenting with several noise models, we found the gamma distribution to be an appropriate model for the noise in OCT images. We carry out denoising using the Gated Convolution and Deconvolution Structure (GCDS) model [5]. Figure 3a,b shows the noisy OCT image and the denoised OCT image, respectively, obtained from the GCDS model. The advantages of the GCDS model are as follows:
  • The GCDS model consists of two phases—an encoding phase and a decoding phase. The encoding phase consists of 5 convolution layers that create a representation that encapsulates all fundamental features but leaves out the noise.
  • The model also contains skip connections between corresponding convolution and deconvolution layers. These skip connections reduce the number of weights to be trained in the neural network, which leads to quicker convergence of the model.
  • Each skip connection is associated with a gating factor, which determines the ratio of split of the information between the next convolution layer and the corresponding deconvolution layer.

2.5. Classification of the OCT Images

We now present the details involved in training the classification model.
Figure 4 shows the pipeline for training a classification model. From here on, we refer to the model trained on noisy images as the noisy model; the model trained on denoised images obtained from the GCDS model using original Inception-ResNet-v2 architecture as the Inception-ResNet-v2 GCDS model; and the model trained on denoised images obtained from the GCDS model using small Inception-ResNet-v2 architecture as the small Inception-ResNet-v2 GCDS model. For the Inception-ResNet-v2 GCDS model and small Inception-ResNet-v2 GCDS model training, noisy OCT images are denoised and used for training the classification model. Post training, we evaluate the performance of the model on the test dataset and compute the classification metrics. For training the noisy model, the denoising step is bypassed, and all the other steps remain the same.

Training Phase

We trained the classification model on 207,535 images. Each image carries labels for IRF, SRF, and PED. For training the classifier, we resize the images, which are of size (1024, 512, 3) to (450, 450, 3) for compatibility with the Inception-ResNet-v2 architecture. The model is tested using data consisting of 52,650 images. We used an Inception-ResNet-v2 architecture pretrained on the ImageNet dataset. We fine-tuned the pretrained model using OCT training images. We used the Adam optimization algorithm [31], which combines adaptive gradient methods and stochastic optimization techniques. Adam has been shown to be effective in various deep learning tasks, as it dynamically adjusts the learning rate based on the gradient magnitudes of individual parameters.
We set the learning rate to 10 4 , which was chosen based on empirical observations and previous research in the field. This learning rate provides a good balance between convergence speed and stability, ensuring that the model converges to a suitable solution while avoiding overshooting or instability issues.
The training phase comprises iterative updating of the model parameters using backpropagation and stochastic gradient descent. The Adam optimizer adjusted the learning rate for each parameter separately, allowing the model to effectively adapt to different features and gradients present in the training data. The loss function employed is Categorical Cross-Entropy (CCE) [32] given by
C C E L o s s = i = 1 o u t p u t s i z e y i · log y ^ i ,
where y ^ i is the ith entry (scalar value) in the model output, y i is the corresponding target value, and the output size is the number of scalar values in the model output.

2.6. Visualization of the OCT Images

Due to the black-box nature of the deep-learning-based classification models, there is a need for explainability of the model. We consider standard visualization-based techniques such as GradCAM, GradCAM++, Score-CAM, Ablation-CAM, and Self-Matching-CAM on the binary classification models of IRF, SRF, and PED. These visualization techniques help doctors understand why the model made a certain decision for an image by highlighting relevant parts of the image. Figure 5b shows the Grad-CAM output of the GCDS denoised OCT image shown in Figure 5a. In the heat map, the yellow region indicates the most relevant region, and the relevance decreases as the color changes from yellow to blue.

Ensemble CAM

We applied standard visualization techniques such as Grad-CAM (GD), Grad-CAM++ (GD++), Score-CAM (SC), Ablation-CAM (AC), and Self-Matching-CAM (SM) on the nine binary classification models developed for each of the three pathologies using the noisy OCT images and denoised OCT images. The heat maps obtained from the visualization techniques are converted to binary maps using the Otsu thresholding [33] technique. The procedure is applied to expert-annotated ground-truth segmentation comprising 128 B-scans, making up a total of 9344 (73 × 128) images to calculate the Intersection over Union (IoU) and Dice scores.
We observed that for several OCT images, the heat maps given by various visualization techniques are not always in agreement, as the techniques emphasize different regions in the OCT image. The Ensemble CAM output is a binary map with the value at a certain pixel equal to one, where three or more than three visualization techniques agree with each other. Computationally, the Ensemble CAM heat map can be obtained by simply adding the five binarized class activation maps and using binary thresholding with a threshold value of three. Figure 6 shows the binary maps obtained from various visualization techniques, including the Ensemble-CAM that we developed.

2.7. Retinal Fluid Volume Computation

AI-based fluid volume computation compared with manual-annotations-based volume computation saves an enormous amount of time and removes intraobserver and interobserver variability. Currently, the device does not provide clinicians the ability to measure fluid volumes automatically, and the assessment is performed by visual inspection. An automatic fluid volume computation tool would greatly enable the clinicians in providing reliable diagnosis.
The fluid volume is computed using two different techniques, namely the region-growing algorithm [34,35] and the thresholding technique. We computed the volumes of 73 C-scans for IRF, SRF, and PED pathologies separately using the region-growing and thresholding techniques.
Figure 7 shows the pipeline for computing the volume of IRF, SRF, and PED in a C-scan. First, forthe Inception-ResNet-v2 GCDS model and the small Inception-ResNet-v2 GCDS model, we pass the noisy OCT images through the denoising model and preprocess them before passing them to the classification model. If the prediction score is more than 0.5, we pass the preprocessed image through the visualization technique to obtain the heat map, which is further binarized using the Otsu thresholding technique [33]. Figure 8 shows the snapshots of the Graphical User Interface (GUI) created for volume computation. The doctors select a C-scan and press the volume computation button to display the results. Four images are displayed in the GUI—the input image, the CAM output, the binary map obtained from the CAM heat map, and the ground truth corresponding to the input image.

2.8. Post Processing

We used a classification model to gain insights into the explainability of the model’s predictions. To visualize the important regions contributing to the model’s decision-making process, we employed several popular techniques, such as GradCAM, GradCAM++, ScoreCAM, LayerCAM, and AblationCAM. These techniques generate heat maps that highlight the regions of interest in the input image. However, we observed that these heat maps only provided an indication of the regions without precisely outlining the boundaries of the specific regions of interest. This limitation became apparent when dealing with conditions such as IRF (Intraretinal Fluid) and SRF (Subretinal Fluid), where it is crucial to localize the presence of fluid in the retina precisely.
We explored two additional techniques to address this issue, selective thresholding and region-growing, with the aim of refining the heat maps generated by the CAM techniques by emphasizing the regions corresponding to fluid-containing areas. We noticed that these regions appeared darker in contrast to the other areas. By applying selective thresholding and region-growing techniques, we improved the localization of the fluid-contained regions and obtained more accurate outlines of these regions. These techniques allowed us to enhance the interpretability of the classification model for identifying and precisely delineating the presence of fluid in the retina, specifically for IRF and SRF.

2.8.1. Region-Growing Technique

To further enhance the identification and delineation of regions of interest, we employed the region-growing technique [36,37], leveraging the CAM output from the classification model. The purpose was to identify potential regions within the image that corresponded to specific features, particularly in the case of Intraretinal Fluid (IRF) and Subretinal Fluid (SRF).
The region-growing approach comprises the following steps:
  • Thresholding and contour extraction: The CAM output was thresholded to convert it into a binary image. Contours were then extracted from the binary image. The centroids of these contours were considered as seed points, representing potential regions of interest within the original input image.
  • Seed point selection: The centroids of the extracted contours served as the seed points for region expansion. These seed points acted as starting points for the iterative growth of the regions.
  • Region-growing algorithm: Starting from the seed points, a region-growing algorithm was applied to expand the regions of interest iteratively. This algorithm examined the neighboring pixels of each seed point and determined their inclusion in the growing region based on predefined criteria.
  • Inclusion criteria: The decision to include neighboring pixels in the growing region was based on a specific criterion, which involved evaluating the pixel intensity difference between the seed pixel and the corresponding neighboring pixel. If the difference was below a predefined threshold, such as 15, the neighboring pixel was considered similar and included in the region. Otherwise, it was excluded.
By iteratively growing the regions based on the inclusion criteria, we could accurately outline the desired regions of interest, specifically the fluid-contained areas in the case of IRF and SRF. This enabled precise visualization and analysis of the extent and location of these regions within the retinal images.
It is important to note that while the region-growing technique demonstrated effectiveness in identifying and delineating fluid regions, its applicability may vary for different conditions. For example, in the case of Pigment Epithelial Detachment (PED), where the objective is to highlight layer detachment areas rather than fluid regions, alternative techniques may be necessary to represent the specific features of interest accurately.
In summary, the region-growing technique, utilizing the CAM output and contour extraction, allowed us to iteratively expand regions of interest and achieve accurate outlines of fluid-contained areas. However, its suitability should be carefully considered in the context of the specific condition under analysis. Alternative techniques may be required to address different types of features and meet the specific requirements of the analysis.

2.8.2. Selective Thresholding

In order to enhance the visualization and outline of regions of interest, we employed a technique known as selective thresholding. This approach aimed to improve upon the existing CAM-based visualization methods, such as GradCAM, GradCAM++, ScoreCAM, LayerCAM, and AblationCAM, which provided highlight regions but lacked precise outlines. The specific goal was to highlight fluid-containing regions in the case of IRF and SRF, while recognizing that a different approach was necessary for PED.
The selective thresholding process involved the following steps:
  • Obtaining the CAM output: We obtained the CAM output from the classification model, which provided an indication of the regions of interest within the image.
  • Applying Otsu thresholding: To convert the CAM output into a binary image, we applied the Otsu thresholding technique. This process assigns a value of 1 to pixels that exceed a certain threshold and 0 to pixels that are below the threshold. The result is a binary mask that highlights potential regions of interest.
  • Pixel-wise product: We then applied a pixel-wise product operation between the original image and the binary mask. This operation allowed us to select and isolate the regions in the image where the mask pixel values were 1. By doing so, we focused only on the areas indicated by the CAM as regions of interest.
  • Histogram analysis: To further refine the selection, we analyzed the histogram of the image. This analysis revealed that pixel values in the range of 25 to 50 exhibited a significant spike, indicating the presence of the desired regions.
  • Thresholding the image: Based on the observation from the histogram analysis, we applied a thresholding operation to the image. Pixels with values in the range of 20 to 50 were considered as 1, representing the regions of interest, while the remaining pixels were set to 0. This thresholding step allowed us to achieve a more precise outline of the desired regions.
  • Performance evaluation: We evaluated the effectiveness of selective thresholding by assessing metrics such as Intersection over Union (IoU) and Dice score. The results indicate a significant improvement in both metrics for the detection of fluid-containing regions in IRF and SRF cases. However, it should be noted that selective thresholding was not suitable for PED, as the objective in the case of PED was to highlight layer detachment rather than fluid regions.
In summary, selective thresholding provided a refined visualization approach by emphasizing the outline and extent of fluid-containing regions in the retinal images. While it demonstrated promising results for specific conditions, its applicability varied depending on the nature of the features of interest. For PED, alternative techniques were required to accurately highlight the layer detachment areas. Figure 9 depicts the procedure adopted for calculating the volume using the selective thresholding technique.
The volume computation process involves employing the region growing or selective thresholding techniques for all the images in a C-scan to derive masks for each True Positive image. Each pixel within the mask is considered as a cuboid, and the volume of the entire scan is computed by summing up all the individual pixel volumes given in Equation (4) across all the identified pixels in the C-scan. The volume occupied by each pixel v p i x e l [38] is scanner-specific, and for the OCT scanners under consideration, it is given by
v p i x e l = 11.7 × 47.2 × 2.0 μ m 3 .

3. Results

3.1. Classification Results

Metrics such as True Positive, False Positive, True Negative, False Negative, Accuracy, Precision, Sensitivity, Specificity, and F1 score are used to evaluate the classification performance. We refer to the positive labels as images in which the pathology is present and negative labels in which the pathology is absent. True Positives (TP) are the images for which prediction is abnormal, and the ground truth is also a positive label. In the case of True Negative (TN) images, the prediction is normal, and the ground truth is also a negative label. For False Positive (FP) cases, the prediction is abnormal, but the ground truth is a negative label, whereas for False Negative (FN) cases, the prediction is normal, but the ground truth is a positive label.
Accuracy is the ratio of total number of correct predictions to the total number of images and is expressed as
A c c u r a c y = T P + T N T P + F P + T N + F N .
Precision is the ratio of the total number of predictions that are classified correctly as positive to the total number of positive predictions:
P r e c i s i o n = T P T P + F P .
Sensitivity or Recall is the ratio of the number of images that are correctly classified as positive to the total number of positive images:
S e n s i t i v i t y / R e c a l l = T P T P + F N .
Specificity is the ratio of the number of images that are correctly classified as negative to the total number of negative images, given by
S p e c i f i c i t y = T N T N + F P .
F1 score is the harmonic mean between precision and recall and is given by
F 1 s c o r e = 2 1 P r e c i s i o n + 1 R e c a l l = 2 T P 2 T P + F P + F N .
F1 score is mainly used when the dataset has a class imbalance.
Table 1, Table 2 and Table 3 present the classification metrics for the IRF, SRF, and PED pathologies.

3.2. Visualization Results

In this section, we present the quantitative visualization results of the Grad-CAM, Grad-CAM++, Score-CAM, Ablation-CAM, Self-Matching CAM, and Ensemble-CAM visualization techniques on the noisy model, Inception-ResNet-v2 GCDS model, and small Inception-ResNet-v2 GCDS model. We used IoU and Dice metrics for evaluating the visualization results. Figure 10, Figure 11 and Figure 12 show the visualization outputs obtained using the Grad-CAM technique, binary maps, and ground truths for IRF, SRF, and PED, respectively.

3.3. Volume Computation Results

We computed the volume over 73 C-scans obtained from the CIRRUS machine for all three kinds of models for IRF, SRF, and PED and calculated the IoU mean and standard deviation, Dice mean and standard deviation, predicted volume mean and standard deviation, and ground truth volume mean and standard deviation.

4. Discussion

The evaluation of the model’s performance includes various classification metrics, such as accuracy, precision, sensitivity, specificity, and F1 score. The classification results for each pathology are presented in Table 1, Table 2 and Table 3. In the case of IRF, the GCDS model of Inception-ResNet-v2 outperformed the other two models. Despite reducing the number of trainable parameters to 28 million from the original 56 million, the model’s performance metric in terms of F1 score only marginally dropped from 0.8353 to 0.8201. The noisy model’s performance lies between that of the Inception-ResNet-v2 GCDS model and the small Inception-ResNet-v2 GCDS model.
For SRF, the small Inception-ResNet-v2 GCDS model exhibited significant improvement over the original Inception-ResNet-v2 GCDS model, with the model’s performance increasing from 0.7835 to 0.8443. The noisy model’s performance metric in terms of F1 score was comparable to that of the small Inception-ResNet-v2 GCDS model. Regarding PED, the small Inception-ResNet-v2 GCDS model outperformed both the Inception-ResNet-v2 GCDS model and the noisy model.
Table 4, Table 5 and Table 6 present the visualization metrics at a B-scan level, where the Intersection over Union (IoU) and Dice coefficient are calculated between the binary map obtained from Otsu thresholding of the Grad-CAM, Grad-CAM++, Score-CAM, Ablation-CAM, Self-Matching-CAM, and Ensemble-CAM heat maps, and the segmentation data masks. For IRF, the small Inception-ResNet-v2 GCDS model achieved the best IoU using Grad-CAM++. In the case of SRF, the noisy model achieved the best IoU using the newly proposed Ensemble-CAM visualization technique. For PED, the small Inception-ResNet-v2 GCDS model obtained the best IoU using the Self-Matching-CAM visualization technique.
Table 7, Table 8, Table 9, Table 10, Table 11, Table 12, Table 13, Table 14, Table 15, Table 16, Table 17, Table 18, Table 19, Table 20, Table 21, Table 22 and Table 23 present the Intersection over Union (IoU) and Dice coefficient on a volumetric level for the region-growing and selective thresholding technique. The mean and standard deviation of these metrics are calculated for both the predicted and ground truth volume. The results are presented for all three types of models: the noisy model, the Inception-ResNet-v2 GCDS model, and the Small Inception-ResNet-v2 GCDS model. While all the models performed reasonably well on IRF and SRF, their performance in terms of volume computation for PED was below par.
IoU and Dice results: In certain cases of performance evaluation, such as in PED, we observed that the standard deviation is greater than the mean. This indicates that the model’s performance is inconsistent, and there is a large variability in the results obtained across different evaluations or samples. It also suggests that the model’s performance varies widely from one test instance to another, indicating outliers, instability, and unpredictability. On the other hand, for IRF and SRF, this anomaly was not observed, indicating that the model’s performance is relatively consistent, with the results clustering closer to the mean. A lower standard deviation indicates that the model’s performance is more stable and less prone to significant variations across different evaluations. In summary, when the standard deviation is greater than the mean, it implies inconsistency and variability in the performance of the model, whereas when the mean is greater than the standard deviation, it suggests stabler and consistent model performance across different evaluations. The specific interpretation may vary based on the context of the evaluation and the performance metric used.
Fluid volume computation: If the data points in the dataset have a wide range and are spread out, it results in a large standard deviation. This can happen when there are extreme values (outliers) that are significantly different from the rest of the data. In our case, for some patients, the fluid volume calculated was pretty high, making them outliers. Hence, the standard deviation is more than the mean in some cases.

5. Conclusions

In this paper, we presented a technique to compute the retinal fluid volume starting from deep-learning-based visualization modules such as Grad-CAM, Grad-CAM++, Score-CAM, Ablation-CAM, Self-Matching-CAM, and Ensemble CAM optimized on 73 C-scans obtained from CIRRUS and PRIMUS OCT machines. The visualization techniques are applied to the models trained on the Inception-ResNet-v2 and small Inception-ResNet-v2 architectures. The calculated volume is obtained using two techniques based on the region-growing algorithm and selective thresholding scheme and compared. Automatic volume computation using AI helps in reducing the effort put in by the annotators and reduces intraobserver and interobserver variability.

Author Contributions

H.R.K., U.R.K., S.K.S.M., H.K. and A.S. performed the experiments and calculated the results. C.S.S., H.R.K., and U.R.K. wrote the manuscript. C.S.S., J.S., S.V., R.P. and R.R. supervised the project and provided constant feedback. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Carl Zeiss India Private Limited.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data were obtained from a Sankara Nethralaya eye hospital, and the patient data will be maintained with confidentiality. The data will be shared upon request and can be provided only after approval from the institutional review board.

Acknowledgments

We would like to thank Ganesh for his help with data curation and analysis of denoised images.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shu, Y.; Ye, F.; Liu, H.; Wei, J.; Sun, X. Predictive value of pigment epithelial detachment markers for visual acuity outcomes in Neovascular Age-related Macular Degeneration. BMC Ophthalmol. 2023, 23, 83. [Google Scholar] [CrossRef] [PubMed]
  2. Reiter, G.S.; Schmidt-Erfurth, U. Quantitative assessment of retinal fluid in Neovascular Age-related Macular Degeneration under anti-VEGF therapy. Ther. Adv. Ophthalmol. 2022, 14, 25158414221083363. [Google Scholar] [CrossRef]
  3. Song, X.; Xu, Q.; Li, H.; Fan, Q.; Zheng, Y.; Zhang, Q.; Chu, C.; Zhang, Z.; Yuan, C.; Ning, M.; et al. Automatic quantification of retinal photoreceptor integrity to predict persistent disease activity in Neovascular Age-related Macular Degeneration using deep learning. Front. Neurosci. 2022, 16, 952735. [Google Scholar] [CrossRef] [PubMed]
  4. What’s the Right OCT for You? Available online: https://eyesoneyecare.com/resources/whats-the-right-oct-for-you/ (accessed on 14 May 2023).
  5. Anoop, B.; Kalmady, K.S.; Udathu, A.; Siddharth, V.; Girish, G.; Kothari, A.R.; Rajan, J. A cascaded convolutional neural network architecture for despeckling oct images. Biomed. Signal Process. Control 2021, 66, 102463. [Google Scholar] [CrossRef]
  6. Bashkansky, M.; Reintjes, J. Statistics and reduction of speckle in optical coherence tomography. Opt. Lett. 2000, 25, 545–547. [Google Scholar] [CrossRef] [PubMed]
  7. Pircher, M.; Gotzinger, E.; Leitgeb, R.A.; Fercher, A.F.; Hitzenberger, C.K. Speckle reduction in optical coherence tomography by frequency compounding. J. Biomed. Opt. 2003, 8, 565–570. [Google Scholar] [CrossRef]
  8. Karamata, B.; Hassler, K.; Laubscher, M.; Lasser, T. Speckle statistics in optical coherence tomography. JOSA A 2005, 22, 593–596. [Google Scholar] [CrossRef]
  9. Schmitt, J.M.; Xiang, S.; Yung, K.M. Speckle in Optical Coherence Tomography. J. Biomed. Opt. 1999, 4, 95–106. [Google Scholar] [CrossRef]
  10. Sudeep, P.; Niwas, S.I.; Palanisamy, P.; Rajan, J.; Xiaojun, Y.; Wang, X.; Luo, Y.; Liu, L. Enhancement and bias removal of optical coherence tomography images: An iterative approach with adaptive bilateral filtering. Comput. Biol. Med. 2016, 71, 97–107. [Google Scholar] [CrossRef]
  11. Shamshirband, S.; Fathi, M.; Dehzangi, A.; Chronopoulos, A.T.; Alinejad-Rokny, H. A review on deep learning approaches in healthcare systems: Taxonomies, challenges, and open issues. J. Biomed. Inform. 2021, 113, 103627. [Google Scholar] [CrossRef]
  12. Amin, R.; Al Ghamdi, M.A.; Almotiri, S.H.; Alruily, M. Healthcare Techniques Through Deep Learning: Issues, Challenges and Opportunities. IEEE Access 2021, 9, 98523–98541. [Google Scholar]
  13. Miotto, R.; Wang, F.; Wang, S.; Jiang, X.; Dudley, J.T. Deep learning for healthcare: Review, opportunities and challenges. Brief. Bioinform. 2018, 19, 1236–1246. [Google Scholar] [CrossRef] [PubMed]
  14. Esteva, A.; Robicquet, A.; Ramsundar, B.; Kuleshov, V.; DePristo, M.; Chou, K.; Cui, C.; Corrado, G.; Thrun, S.; Dean, J. A guide to deep learning in healthcare. Nat. Med. 2019, 25, 24–29. [Google Scholar] [CrossRef] [PubMed]
  15. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  16. Chattopadhay, A.; Sarkar, A.; Howlader, P.; Balasubramanian, V.N. Grad-CAM++: Generalized Gradient-based Visual Explanations for Deep Convolutional Networks. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 839–847. [Google Scholar]
  17. Wang, H.; Wang, Z.; Du, M.; Yang, F.; Zhang, Z.; Ding, S.; Mardziel, P.; Hu, X. Score-cam: Score-weighted visual explanations for convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 24–25. [Google Scholar]
  18. Saurabh, D.; Guruprasad, R.H. Ablation-CAM: Visual Explanations for Deep Convolutional Network via Gradient-free Localization. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision (WACV), Snowmass Village, CO, USA, 1–5 March 2020; pp. 972–980. [Google Scholar]
  19. Zhenpeng, F.; Mingzhe, Z.; Ljubiša, S.; Hongbing, J. Self-Matching CAM: A Novel Accurate Visual Explanation of CNNs for SAR Image Interpretation. Remote Sens. 2021, 13, 1772. [Google Scholar]
  20. Guo, Y.; Hormel, T.T.; Xiong, H.; Wang, J.; Hwang, T.S.; Jia, Y. Automated Segmentation of Retinal Fluid Volumes From Structural and Angiographic Optical Coherence Tomography Using Deep Learning. Transl. Vis. Sci. Technol. 2020, 9, 54. [Google Scholar] [CrossRef]
  21. Schlegl, T.; Waldstein, S.M.; Bogunovic, H.; Endstraßer, F.; Sadeghipour, A.; Philip, A.M.; Podkowinski, D.; Gerendas, B.S.; Langs, G.; Schmidt-Erfurth, U. Fully Automated Detection and Quantification of Macular Fluid in OCT Using Deep Learning. Am. Acad. Ophthalmol. 2018, 125, 549–558. [Google Scholar] [CrossRef] [Green Version]
  22. Wilson, M.; Chopra, R.; Wilson, M.Z.; Cooper, C.; MacWilliams, P.; Liu, Y.; Wulczyn, E.; Florea, D.; Hughes, C.O.; Karthikesalingam, A.; et al. Validation and Clinical Applicability of Whole-Volume Automated Segmentation of Optical Coherence Tomography in Retinal Disease Using Deep Learning. JAMA Ophthalmol. 2021, 139, 964–973. [Google Scholar] [CrossRef]
  23. Lu, D.; Heisler, M.; Lee, S.; Ding, G.; Sarunic, M.V.; Beg, M.F. Retinal Fluid Segmentation and Detection in Optical Coherence Tomography Images using Fully Convolutional Neural Network. arXiv 2017, arXiv:1710.04778. [Google Scholar]
  24. Chakravarthy, U.; Havilio, M.; Syntosi, A.; Pillai, N.; Wilkes, E.; Benyamini, G.; Best, C.; Sagkriotis, A. Impact of macular fluid volume fluctuations on visual acuity during anti-VEGF therapy in eyes with nAMD. Eye 2021, 35, 2983–2990. [Google Scholar] [CrossRef]
  25. Reddy, K.H.; Shivhare, A.; Kongara, H.; Saita, J.; Prasad, R.; Seelamantula, C.S. Denoising Enhances Visualization of Optical Coherence Tomography Images. In Proceedings of the Sixth NeurIPS workshop on Medical Imaging meets NeurIPS (Med-NeurIPS), New Orleans, LA, USA, 3 December 2022. [Google Scholar]
  26. Jaccard, P. The Distribution of the Flora in the Alpine Zone.1. New Phytol. 1912, 11, 37–50. [Google Scholar] [CrossRef]
  27. Sørensen, T. A method of establishing groups of equal amplitude in plant sociology based on similarity of species and its application to analyses of the vegetation on Danish commons. K. Dan. Vidensk. Selsk. 1945, 5, 1–34. [Google Scholar]
  28. Dice, L.R. Measures of the Amount of Ecologic Association Between Species. Ecology 1948, 26, 297–302. [Google Scholar] [CrossRef]
  29. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. In Proceedings of the AAAI Conference on Artificial Intelligence, San Fransico, CA, USA, 4–9 February 2017; pp. 2950–2958. [Google Scholar]
  30. Jia, D.; Wei, D.; Richard, S.; Li-Jia, L.; Kai, L.; Li, F. Imagenet: A large-scale hierarchical image database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  31. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  32. Zhilu, Z.; Sabuncu, M. Generalized cross entropy loss for training deep neural networks with noisy labels. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; p. 31. [Google Scholar]
  33. Otsu, N. A Threshold Selection Method from Gray-Level Histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 1079–1084. [Google Scholar] [CrossRef] [Green Version]
  34. Adams, R.; Bischof, L. Seeded region-growing. IEEE Trans. Pattern Anal. Mach. Intell. 1994, 16, 641–647. [Google Scholar] [CrossRef] [Green Version]
  35. Hojjatoleslami, S.A.; Kittler, J. Region-growing: A new approach. IEEE Trans. Image Process. 1998, 7, 1079–1084. [Google Scholar] [CrossRef] [Green Version]
  36. Pal, N.R.; Pal, S.K. A review on image segmentation techniques. Pattern Recognit. 1993, 26, 1277–1278. [Google Scholar] [CrossRef]
  37. Gonzalez, R.C.; Woods, R.E. Digital Image Processing, 2nd ed.; Prentice Hall: Englewood Cliffs, NJ, USA; New Delhi, India, 2002; pp. 32–58. [Google Scholar]
  38. Schmidt-Erfurth, U.; Vogl, W.D.; Jampol, L.M.; Bogunović, H. Application of Automated Quantification of Fluid Volumes to Anti–VEGF Therapy of Neovascular Age-related Macular Degeneration. Ophthalmology 2022, 127, 1211–1219. [Google Scholar] [CrossRef]
Figure 1. Various pathologies of interest that are observed in OCT images: IRF: Intraretinal Fluid; SRF: Subretinal Fluid; and PED: Pigmented Epithelial Detachment.
Figure 1. Various pathologies of interest that are observed in OCT images: IRF: Intraretinal Fluid; SRF: Subretinal Fluid; and PED: Pigmented Epithelial Detachment.
Diagnostics 13 02659 g001
Figure 2. Block diagram of (a) Inception-ResNet-v2 architecture; (b) Small Inception-ResNet-v2 architecture; and (c) Stem.
Figure 2. Block diagram of (a) Inception-ResNet-v2 architecture; (b) Small Inception-ResNet-v2 architecture; and (c) Stem.
Diagnostics 13 02659 g002
Figure 3. (a) Noisy OCT image and (b) Denoised OCT image using GCDS.
Figure 3. (a) Noisy OCT image and (b) Denoised OCT image using GCDS.
Diagnostics 13 02659 g003
Figure 4. The pipeline for the training classification model: Path-1 is the pipeline for the Inception-ResNet-v2 GCDS model and small Inception-ResNet-v2 GCDS model, whereas Path-2 is the pipeline for the noisy model.
Figure 4. The pipeline for the training classification model: Path-1 is the pipeline for the Inception-ResNet-v2 GCDS model and small Inception-ResNet-v2 GCDS model, whereas Path-2 is the pipeline for the noisy model.
Diagnostics 13 02659 g004
Figure 5. (a) GCDS denoised OCT image and (b) GradCAM output.
Figure 5. (a) GCDS denoised OCT image and (b) GradCAM output.
Diagnostics 13 02659 g005
Figure 6. This figure shows the outputs from the Inception-ResNet-v2 GCDS model: (a) OCT image; (b) ground truth; (c) Grad-CAM binary map; (d) Grad-CAM++ binary map; (e) Score-CAM binary map; (f) Ablation-CAM binary map; (g) Self-Matching-CAM binary map; and (h) Ensemble-CAM binary map.
Figure 6. This figure shows the outputs from the Inception-ResNet-v2 GCDS model: (a) OCT image; (b) ground truth; (c) Grad-CAM binary map; (d) Grad-CAM++ binary map; (e) Score-CAM binary map; (f) Ablation-CAM binary map; (g) Self-Matching-CAM binary map; and (h) Ensemble-CAM binary map.
Diagnostics 13 02659 g006
Figure 7. This figure shows the pipeline for volume computation: Path-1 is the pipeline for the Inception-ResNet-v2 GCDS model and small Inception-ResNet-v2 GCDS model, while Path-2 is the pipeline for the noisy model.
Figure 7. This figure shows the pipeline for volume computation: Path-1 is the pipeline for the Inception-ResNet-v2 GCDS model and small Inception-ResNet-v2 GCDS model, while Path-2 is the pipeline for the noisy model.
Diagnostics 13 02659 g007
Figure 8. (a) Visualization output and (b) Graphical User Interface displaying the computed volume.
Figure 8. (a) Visualization output and (b) Graphical User Interface displaying the computed volume.
Diagnostics 13 02659 g008
Figure 9. (a) OCT image; (b) binary mask obtained from visualization; (c) region of interest; (d) histogram of the region of interest; (e) predicted region of interest from the selective thresholding technique; and (f) ground-truth of the region of interest.
Figure 9. (a) OCT image; (b) binary mask obtained from visualization; (c) region of interest; (d) histogram of the region of interest; (e) predicted region of interest from the selective thresholding technique; and (f) ground-truth of the region of interest.
Diagnostics 13 02659 g009
Figure 10. This figure shows the outputs from the Inception-ResNet-v2 GCDS model: (a) OCT image; (b) Grad-CAM heat map of OCT image for IRF; (c) binary map of Grad-CAM heat map; and (d) corresponding ground truth.
Figure 10. This figure shows the outputs from the Inception-ResNet-v2 GCDS model: (a) OCT image; (b) Grad-CAM heat map of OCT image for IRF; (c) binary map of Grad-CAM heat map; and (d) corresponding ground truth.
Diagnostics 13 02659 g010
Figure 11. This figure shows the outputs from the Inception-ResNet-v2 GCDS model: (a) OCT image; (b) Grad-CAM heat map of OCT image for SRF; (c) binary map of Grad-CAM heat map; and (d) corresponding ground truth..
Figure 11. This figure shows the outputs from the Inception-ResNet-v2 GCDS model: (a) OCT image; (b) Grad-CAM heat map of OCT image for SRF; (c) binary map of Grad-CAM heat map; and (d) corresponding ground truth..
Diagnostics 13 02659 g011
Figure 12. This figure shows the outputs from Inception-ResNet-v2 GCDS model: (a) OCT image; (b) Grad-CAM heat map of OCT image for PED; (c) binary map of Grad-CAM heat map; and (d) corresponding ground truth.
Figure 12. This figure shows the outputs from Inception-ResNet-v2 GCDS model: (a) OCT image; (b) Grad-CAM heat map of OCT image for PED; (c) binary map of Grad-CAM heat map; and (d) corresponding ground truth.
Diagnostics 13 02659 g012
Table 1. Classification metrics of various models on Intraretinal Fluid pathology.
Table 1. Classification metrics of various models on Intraretinal Fluid pathology.
Performance MeasureNoisy ModelInception-ResNet-v2 GCDS ModelSmall Inception-ResNet-v2 GCDS Model
True Positive879784078556
False Positive291922132803
True Negative40,22440,93040,340
False Negative7101100951
Accuracy0.93100.93700.9284
Precision0.75080.79160.7532
Sensitivity0.92530.88420.9000
Specificity0.93230.94870.9350
F1 score0.82900.83530.8201
Table 2. Classification metrics of various models on Subretinal Fluid pathology.
Table 2. Classification metrics of various models on Subretinal Fluid pathology.
Performance MeasureNoisy ModelInception-ResNet-v2 GCDS ModelSmall Inception-ResNet-v2 GCDS Model
True Positive218222912017
False Positive6131161365
True Negative49,64149,09349,889
False Negative214105379
Accuracy0.98420.97590.9859
Precision0.78060.66360.8468
Sensitivity0.91060.95610.8418
Specificity0.98780.97680.9927
F1 score0.84060.78350.8443
Table 3. Classification metrics of various models on Pigmented Epithelial Detachment pathology.
Table 3. Classification metrics of various models on Pigmented Epithelial Detachment pathology.
Performance MeasureNoisy ModelInception-ResNet-v2 GCDS ModelSmall Inception-ResNet-v2 GCDS Model
True Positive271426992531
False Positive251127951464
True Negative46,43446,15047,481
False Negative99110061174
Accuracy0.93340.92780.9499
Precision0.51940.49130.6335
Sensitivity0.73250.72850.6831
Specificity0.94860.94290.9701
F1 score0.60780.58680.6574
Table 4. Visualization results of various models on Intraretinal Fluid pathology.
Table 4. Visualization results of various models on Intraretinal Fluid pathology.
Noisy ModelInception-ResNet-v2 GCDS ModelSmall Inception-ResNet-v2 GCDS Model
IoUDiceIoUDiceIoUDice
Grad-CAM0.09090.15560.11620.19120.14380.2325
Grad-CAM++0.12950.21050.14980.23500.16150.2570
Score-CAM0.12800.20830.13900.22060.14000.2274
Ablation-CAM0.12930.21020.14970.23480.06880.1165
Self-MCAM0.06640.11950.04100.07700.03450.0653
Ensemble-CAM0.12780.20810.13930.22150.13650.2238
Table 5. Visualization results of various models on Subretinal Fluid pathology.
Table 5. Visualization results of various models on Subretinal Fluid pathology.
Performance MeasureNoisy ModelInception-ResNet-v2 GCDS ModelSmall Inception-ResNet-v2 GCDS Model
IoUDiceIoUDiceIoUDice
Grad-CAM0.09820.17110.08110.14490.04130.0758
Grad-CAM++0.09920.17250.08340.14850.06680.1207
Score-CAM0.05170.08940.08750.15520.09650.1698
Ablation-CAM0.08690.15200.06810.12180.01480.0272
Self-Matching-CAM0.07010.12800.07580.13790.03730.0707
Ensemble-CAM0.09950.17300.08590.15240.06730.1211
Table 6. Visualization results of various models on Pigmented Epithelial Detachment pathology.
Table 6. Visualization results of various models on Pigmented Epithelial Detachment pathology.
Performance MeasureNoisy ModelInception-ResNet-v2 GCDS ModelSmall Inception-ResNet-v2 GCDS Model
IoUDiceIoUDiceIoUDice
Grad-CAM0.11000.19240.10330.18190.07400.1289
Grad-CAM++0.01740.03110.11870.20550.10280.1785
Score-CAM0.01750.03130.11700.20330.11460.1984
Ablation-CAM0.01760.03150.11860.20540.10390.1806
Self-Matching-CAM0.07870.13900.12390.21190.13100.2145
Ensemble-CAM0.01770.03160.12030.20820.12480.2142
Table 7. Volume computation results of various models on Intraretinal Fluid pathology for various visualization techniques using selective thresholding technique for the noisy model.
Table 7. Volume computation results of various models on Intraretinal Fluid pathology for various visualization techniques using selective thresholding technique for the noisy model.
ModelGrad-CAMGrad-CAM++Score-CAMAblation-CAMSelf-Matching-CAMEnsemble-CAM
IoU mean0.190.240.240.240.210.24
IoU std0.130.140.140.140.100.14
Dice mean0.280.350.350.350.320.35
Dice std0.170.180.180.180.130.18
Predicted vol. mean0.620.350.310.350.270.35
Predicted vol. std0.680.230.340.380.230.38
Ground truth vol. mean0.150.150.150.150.150.15
Ground truth vol. std0.230.230.230.230.230.23
Table 8. Volume computation results of various models on Intraretinal Fluid pathology for various visualization techniques using selective thresholding technique for the Inception-ResNet-v2 GCDS model.
Table 8. Volume computation results of various models on Intraretinal Fluid pathology for various visualization techniques using selective thresholding technique for the Inception-ResNet-v2 GCDS model.
ModelGrad-CAMGrad-CAM++Score-CAMAblation-CAMSelf-Matching-CAMEnsemble-CAM
IoU mean0.200.250.220.250.160.24
IoU std0.170.200.200.200.080.20
Dice mean0.290.340.300.340.240.33
Dice std0.210.240.240.240.120.23
Predicted vol. mean0.350.260.370.260.110.25
Predicted vol. std0.490.350.450.360.130.35
Ground truth vol. mean0.150.150.150.150.150.15
Ground truth vol. std0.230.230.230.230.230.23
Table 9. Volume computation results of various models on Intraretinal Fluid pathology for various visualization techniques using selective thresholding technique for the small Inception-ResNet-v2 GCDS model.
Table 9. Volume computation results of various models on Intraretinal Fluid pathology for various visualization techniques using selective thresholding technique for the small Inception-ResNet-v2 GCDS model.
ModelGrad-CAMGrad-CAM++Score-CAMAblation-CAMSelf-Matching-CAMEnsemble-CAM
IoU mean0.250.290.230.120.150.27
IoU std0.170.200.200.100.090.18
Dice mean0.340.400.310.170.230.37
Dice std0.210.240.250.140.130.22
Predicted vol. mean0.180.220.390.090.140.14
Predicted vol. std0.180.210.340.090.160.14
Ground truth vol. mean0.150.150.150.150.150.15
Ground truth vol. std0.230.230.230.230.230.23
Table 10. Volume computation results of various models on Intraretinal Fluid pathology for various visualization techniques using region-growing technique for the noisy model.
Table 10. Volume computation results of various models on Intraretinal Fluid pathology for various visualization techniques using region-growing technique for the noisy model.
ModelGrad-CAMGrad-CAM++Score-CAMAblation-CAMSelf-Matching-CAMEnsemble-CAM
IoU mean0.220.250.240.250.220.26
IoU std0.160.170.170.170.130.17
Dice mean0.290.320.320.320.300.33
Dice std0.200.210.210.210.160.21
Predicted vol. mean1.440.750.470.770.540.72
Predicted vol. std1.670.900.580.920.700.90
Ground truth vol. mean0.150.150.150.150.150.15
Ground truth vol. std0.230.230.230.230.230.23
Table 11. Volume computation results of various models on Intraretinal Fluid pathology for various visualization techniques using region-growing technique for the Inception-ResNet-v2 GCDS model.
Table 11. Volume computation results of various models on Intraretinal Fluid pathology for various visualization techniques using region-growing technique for the Inception-ResNet-v2 GCDS model.
ModelGrad-CAMGrad-CAM++Score-CAMAblation-CAMSelf-Matching-CAMEnsemble-CAM
IoU mean0.230.220.220.220.100.23
IoU std0.180.190.200.190.080.19
Dice mean0.290.280.290.280.150.30
Dice std0.220.220.230.220.120.22
Predicted vol. mean0.820.760.850.770.780.66
Predicted vol. std1.080.840.990.870.790.85
Ground truth vol. mean0.150.150.150.150.150.15
Ground truth vol. std0.230.230.230.230.230.23
Table 12. Volume computation results of various models on Intraretinal Fluid pathology for various visualization techniques using region-growing technique for the small Inception-ResNet-v2 GCDS model.
Table 12. Volume computation results of various models on Intraretinal Fluid pathology for various visualization techniques using region-growing technique for the small Inception-ResNet-v2 GCDS model.
ModelGrad-CAMGrad-CAM++Score-CAMAblation-CAMSelf-Matching-CAMEnsemble-CAM
IoU mean0.230.230.150.140.150.24
IoU std0.180.190.150.140.100.19
Dice mean0.300.300.200.180.190.31
Dice std0.230.240.200.180.150.23
Predicted vol. mean0.870.620.940.940.650.60
Predicted vol. std1.160.660.850.970.850.87
Ground truth vol. mean0.150.150.150.150.150.15
Ground truth vol. std0.230.230.230.230.230.23
Table 13. Volume computation results of various models on Subretinal Fluid pathology for various visualization techniques using selective thresholding technique for the noisy model.
Table 13. Volume computation results of various models on Subretinal Fluid pathology for various visualization techniques using selective thresholding technique for the noisy model.
ModelGrad-CAMGrad-CAM++Score-CAMAblation-CAMSelf-Matching-CAMEnsemble-CAM
IoU mean0.160.170.060.150.160.17
IoU std0.110.110.100.100.090.11
Dice mean0.250.260.090.240.260.26
Dice std0.150.150.140.130.120.15
Predicted vol. mean0.660.625.980.550.360.61
Predicted vol. std0.960.876.810.720.520.87
Ground truth vol. mean0.370.370.370.370.370.37
Ground truth vol. std0.710.710.710.710.710.71
Table 14. Volume computation results of various models on Subretinal Fluid pathology for various visualization techniques using selective thresholding technique for the Inception-ResNet-v2 GCDS model.
Table 14. Volume computation results of various models on Subretinal Fluid pathology for various visualization techniques using selective thresholding technique for the Inception-ResNet-v2 GCDS model.
ModelGrad-CAMGrad-CAM++Score-CAMAblation-CAMSelf-Matching-CAMEnsemble-CAM
IoU mean0.090.110.130.090.130.11
IoU std0.090.110.120.080.100.12
Dice mean0.150.160.190.140.210.17
Dice std0.130.150.170.110.140.16
Predicted vol. mean0.630.560.460.480.120.52
Predicted vol. std0.850.710.610.570.210.67
Ground truth vol. mean0.370.370.370.370.370.37
Ground truth vol. std0.710.710.710.710.710.71
Table 15. Volume computation results of various models on Subretinal Fluid pathology for various visualization techniques using selective thresholding technique for the small Inception-ResNet-v2 GCDS model.
Table 15. Volume computation results of various models on Subretinal Fluid pathology for various visualization techniques using selective thresholding technique for the small Inception-ResNet-v2 GCDS model.
ModelGrad-CAMGrad-CAM++Score-CAMAblation-CAMSelf-Matching-CAMEnsemble-CAM
IoU mean0.070.130.180.010.140.10
IoU std0.070.120.140.020.090.11
Dice mean0.110.210.270.020.210.15
Dice std0.110.160.190.030.150.15
Predicted vol. mean0.220.170.190.130.190.69
Predicted vol. std0.360.300.330.180.250.66
Ground truth vol. mean0.370.370.370.370.370.37
Ground truth vol. std0.710.710.710.710.710.71
Table 16. Volume computation results of various models on Subretinal Fluid pathology for various visualization techniques using region-growing technique for the noisy model.
Table 16. Volume computation results of various models on Subretinal Fluid pathology for various visualization techniques using region-growing technique for the noisy model.
ModelGrad-CAMGrad-CAM++Score-CAMAblation-CAMSelf-Matching-CAMEnsemble-CAM
IoU mean0.310.300.120.280.280.31
IoU std0.190.180.160.170.160.19
Dice mean0.390.380.140.360.360.39
Dice std0.220.210.190.200.190.22
Predicted vol. mean1.161.109.081.011.141.08
Predicted vol. std1.161.158.781.401.481.54
Ground truth vol. mean0.370.370.370.370.370.37
Ground truth vol. std0.710.710.710.710.710.71
Table 17. Volume computation results of various models on Subretinal Fluid pathology for various visualization techniques using region-growing technique for the Inception-ResNet-v2 GCDS model.
Table 17. Volume computation results of various models on Subretinal Fluid pathology for various visualization techniques using region-growing technique for the Inception-ResNet-v2 GCDS model.
ModelGrad-CAMGrad-CAM++Score-CAMAblation-CAMSelf-Matching-CAMEnsemble-CAM
IoU mean0.120.170.220.140.200.17
IoU std0.130.150.160.130.150.14
Dice mean0.160.220.290.190.260.22
Dice std0.160.180.190.160.190.18
Predicted vol. mean0.840.881.161.580.894.39
Predicted vol. std1.050.871.051.161.174.21
Ground truth vol. mean0.370.370.370.370.370.37
Ground truth vol. std0.710.710.710.710.710.71
Table 18. Volume computation results of various models on Subretinal Fluid pathology for various visualization techniques using region-growing technique for the small Inception-ResNet-v2 GCDS model.
Table 18. Volume computation results of various models on Subretinal Fluid pathology for various visualization techniques using region-growing technique for the small Inception-ResNet-v2 GCDS model.
ModelGrad-CAMGrad-CAM++Score-CAMAblation-CAMSelf-Matching-CAMEnsemble-CAM
IoU mean0.130.190.270.130.210.14
IoU std0.180.180.190.110.170.16
Dice mean0.170.250.350.180.270.19
Dice std0.220.220.230.150.200.20
Predicted vol. mean0.930.660.511.750.960.62
Predicted vol. std1.511.020.781.201.550.86
Ground truth vol. mean0.370.370.370.370.370.37
Ground truth vol. std0.710.710.710.710.710.71
Table 19. Volume computation results of various models on Pigmented Epithelial Detachment pathology for various visualization techniques using selective thresholding technique for the noisy model.
Table 19. Volume computation results of various models on Pigmented Epithelial Detachment pathology for various visualization techniques using selective thresholding technique for the noisy model.
ModelGrad-CAMGrad-CAM++Score-CAMAblation-CAMSelf-Matching-CAMEnsemble-CAM
IoU mean0.020.000.000.000.010.00
IoU std0.030.010.010.010.020.01
Dice mean0.040.010.010.010.020.01
Dice std0.050.020.020.020.030.02
Predicted vol. mean1.520.610.610.613.670.61
Predicted vol. std1.280.670.670.673.310.67
Ground truth vol. mean0.620.620.620.620.620.62
Ground truth vol. std0.480.480.480.480.480.48
Table 20. Volume computation results of various models on Pigmented Epithelial Detachment pathology for various visualization techniques using selective thresholding technique for the Inception-ResNet-v2 GCDS model.
Table 20. Volume computation results of various models on Pigmented Epithelial Detachment pathology for various visualization techniques using selective thresholding technique for the Inception-ResNet-v2 GCDS model.
ModelGrad-CAMGrad-CAM++Score-CAMAblation-CAMSelf-Matching-CAMEnsemble-CAM
IoU mean0.010.020.020.020.010.02
IoU std0.020.020.020.020.010.02
Dice mean0.030.030.030.030.010.03
Dice std0.040.040.040.040.020.04
Predicted vol. mean0.840.520.550.520.200.49
Predicted vol. std0.850.530.560.530.280.54
Ground truth vol. mean0.620.620.620.620.620.62
Ground truth vol. std0.480.480.480.480.480.48
Table 21. Volume computation results of various models on Pigmented Epithelial Detachment pathology for various visualization techniques using selective thresholding technique for the small Inception-ResNet-v2 GCDS model.
Table 21. Volume computation results of various models on Pigmented Epithelial Detachment pathology for various visualization techniques using selective thresholding technique for the small Inception-ResNet-v2 GCDS model.
ModelGrad-CAMGrad-CAM++Score-CAMAblation-CAMSelf-Matching-CAMEnsemble-CAM
IoU mean0.020.020.010.010.010.01
IoU std0.030.020.020.020.010.02
Dice mean0.030.030.020.020.020.02
Dice std0.050.040.040.040.020.03
Predicted vol. mean0.990.760.430.350.190.28
Predicted vol. std0.900.700.400.330.250.28
Ground truth vol. mean0.620.620.620.620.620.62
Ground truth vol. std0.480.480.480.480.480.48
Table 22. Volume computation results of various models on Pigmented Epithelial Detachment pathology for various visualization techniques using region-growing technique for the noisy model. GT stands for ground truth provided by an expert.
Table 22. Volume computation results of various models on Pigmented Epithelial Detachment pathology for various visualization techniques using region-growing technique for the noisy model. GT stands for ground truth provided by an expert.
ModelGrad-CAMGrad-CAM++Score-CAMAblation-CAMSelf-Matching-CAMEnsemble-CAM
IoU mean0.040.040.010.010.040.05
IoU std0.030.040.010.020.050.06
Dice mean0.060.070.010.010.070.08
Dice std0.050.060.020.020.070.05
Predicted vol. mean4.994.852.2612.315.852.36
Predicted vol. std4.054.461.468.344.571.99
GT vol. mean0.620.620.620.620.620.62
GT vol. std0.480.480.480.480.480.48
Table 23. Volume computation results of various models on Pigmented Epithelial Detachment pathology for various visualization techniques using region-growing technique for the Inception-ResNet-v2 GCDS model.
Table 23. Volume computation results of various models on Pigmented Epithelial Detachment pathology for various visualization techniques using region-growing technique for the Inception-ResNet-v2 GCDS model.
ModelGrad-CAMGrad-CAM++Score-CAMAblation-CAMSelf-Matching-CAMEnsemble-CAM
IoU mean0.060.040.030.050.050.05
IoU std0.040.040.030.040.040.04
Dice mean0.090.070.050.090.090.09
Dice std0.060.060.040.060.060.06
Predicted vol. mean2.244.851.522.023.101.76
Predicted vol. std1.894.461.561.073.221.49
Ground truth vol. mean0.620.620.620.620.620.62
Ground truth vol. std0.480.480.480.480.480.48
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kasireddy, H.R.; Kallam, U.R.; Mantrala, S.K.S.; Kongara, H.; Shivhare, A.; Saita, J.; Vijay, S.; Prasad, R.; Raman, R.; Seelamantula, C.S. Deep-Learning-Based Visualization and Volumetric Analysis of Fluid Regions in Optical Coherence Tomography Scans. Diagnostics 2023, 13, 2659. https://doi.org/10.3390/diagnostics13162659

AMA Style

Kasireddy HR, Kallam UR, Mantrala SKS, Kongara H, Shivhare A, Saita J, Vijay S, Prasad R, Raman R, Seelamantula CS. Deep-Learning-Based Visualization and Volumetric Analysis of Fluid Regions in Optical Coherence Tomography Scans. Diagnostics. 2023; 13(16):2659. https://doi.org/10.3390/diagnostics13162659

Chicago/Turabian Style

Kasireddy, Harishwar Reddy, Udaykanth Reddy Kallam, Sowmitri Karthikeya Siddhartha Mantrala, Hemanth Kongara, Anshul Shivhare, Jayesh Saita, Sharanya Vijay, Raghu Prasad, Rajiv Raman, and Chandra Sekhar Seelamantula. 2023. "Deep-Learning-Based Visualization and Volumetric Analysis of Fluid Regions in Optical Coherence Tomography Scans" Diagnostics 13, no. 16: 2659. https://doi.org/10.3390/diagnostics13162659

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop