Next Article in Journal
Isolation and Characterization of Exosomes from Cancer Cells Using Antibody-Functionalized Paddle Screw-Type Devices and Detection of Exosomal miRNA Using Piezoelectric Biosensor
Previous Article in Journal
Early Bolt Loosening Detection Method Based on Digital Image Correlation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dense Convolutional Neural Network-Based Deep Learning Pipeline for Pre-Identification of Circular Leaf Spot Disease of Diospyros kaki Leaves Using Optical Coherence Tomography

by
Deshan Kalupahana
1,
Nipun Shantha Kahatapitiya
1,
Bhagya Nathali Silva
2,3,
Jeehyun Kim
4,
Mansik Jeon
4,
Udaya Wijenayake
1,* and
Ruchire Eranga Wijesinghe
3,5,*
1
Department of Computer Engineering, Faculty of Engineering, University of Sri Jayewardenepura, Nugegoda 10250, Sri Lanka
2
Department of Information Technology, Faculty of Computing, Sri Lanka Institute of Information Technology, Malabe 10115, Sri Lanka
3
Center for Excellence in Informatics, Electronics & Transmission (CIET), Sri Lanka Institute of Information Technology, Malabe 10115, Sri Lanka
4
School of Electronic and Electrical Engineering, College of IT Engineering, Kyungpook National University, 80 Daehak-ro, Buk-gu, Daegu 41566, Republic of Korea
5
Department of Electrical and Electronic Engineering, Faculty of Engineering, Sri Lanka Institute of Information Technology, Malabe 10115, Sri Lanka
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(16), 5398; https://doi.org/10.3390/s24165398 (registering DOI)
Submission received: 21 June 2024 / Revised: 30 July 2024 / Accepted: 5 August 2024 / Published: 21 August 2024

Abstract

:
Circular leaf spot (CLS) disease poses a significant threat to persimmon cultivation, leading to substantial harvest reductions. Existing visual and destructive inspection methods suffer from subjectivity, limited accuracy, and considerable time consumption. This study presents an automated pre-identification method of the disease through a deep learning (DL) based pipeline integrated with optical coherence tomography (OCT), thereby addressing the highlighted issues with the existing methods. The investigation yielded promising outcomes by employing transfer learning with pre-trained DL models, specifically DenseNet-121 and VGG-16. The DenseNet-121 model excels in differentiating among three stages of CLS disease (healthy (H), apparently healthy (or healthy-infected (HI)), and infected (I)). The model achieved precision values of 0.7823 for class-H, 0.9005 for class-HI, and 0.7027 for class-I, supported by recall values of 0.8953 for class-HI and 0.8387 for class-I. Moreover, the performance of CLS detection was enhanced by a supplemental quality inspection model utilizing VGG-16, which attained an accuracy of 98.99% in discriminating between low-detail and high-detail images. Moreover, this study employed a combination of LAMP and A-scan for the dataset labeling process, significantly enhancing the accuracy of the models. Overall, this study underscores the potential of DL techniques integrated with OCT to enhance disease identification processes in agricultural settings, particularly in persimmon cultivation, by offering efficient and objective pre-identification of CLS and enabling early intervention and management strategies.

1. Introduction

Diospyros kaki, more generally referred to as the Japanese persimmon, is a fruit that is native to a number of countries and regions, including Spain, China, South Korea, and Japan. This plant possesses considerable importance due to its status as one of the earliest cultivated plants on a worldwide scale, being highly regarded for its nutritional content and therapeutic properties. Among persimmon-related diseases, Mycosphaerella nawae, which causes circular leaf spot (CLS) disease, stands out as one of the most devastating diseases [1]. CLS is characterized by the discoloration and defoliation of persimmon leaves, resulting in the premature abscission of the fruit [2]. The widespread prevalence of CLS poses significant economic challenges in regions where persimmons are cultivated. Detecting and controlling CLS disease in its nascent stages proves daunting due to the complexities involved in its identification based on external characteristics.
The extended incubation period of CLS disease can last anywhere from 90 to 120 days. Thus, earlier detection of infection will help in limiting the harm that the disease can cause [3]. Farmers traditionally rely on manual inspection of leaf thickness as an initial method of identifying CLS disease before visible symptoms manifest. However, such identification methods necessitate significant experiential expertise and remain inherently subjective. Polymerase chain reaction (PCR) serves as the gold standard for pathogen identification, offering direct detection capabilities [4]. Nonetheless, its implementation in the field is hindered by the destructive nature of sample preparation and time-consuming procedures. Visual inspection techniques, including scanning electron microscopy (SEM), transmission electron microscopy (TEM), light microscopy, and magnetic resonance imaging (MRI), have been utilized for plant disease identification by observing morphological changes [5]. However, microscopy techniques require destructive sampling, while the resolution of MRI has challenges in distinguishing diseased and healthy tissues [6,7].
To address challenges associated with both direct identification techniques, such as PCR and indirect visual inspection methods, optical coherence tomography (OCT) emerges as a promising alternative by offering several advantageous features, such as real-time operation and non-invasive inspection and visualization with a micrometer resolution, which are highly beneficial for agricultural inspections [8,9,10,11]. This non-invasive OCT technology offers high-resolution images at the micrometer scale in real-time. Numerous studies have been directed towards leveraging OCT for CLS identification, employing it as an indirect method to analyze leaf morphology for disease detection. However, accurate interpretation of OCT data often necessitates specialized knowledge and expertise in plant pathology, which may not be widely available.
In response, the application of deep learning (DL) techniques has gained traction for automated analysis of optical imaging across various domains [12]. This approach enables efficient analysis while circumventing the limitations associated with expertise in image interpretation, thereby enhancing objectivity in detection.
Building on this trend, this study aimed to enable early detection of CLS disease in persimmons while addressing the limitations of current CLS identification methods in agricultural settings. It involved creating a specific dataset and developing optimized DL models for efficient, advanced identification of CLS. The significance of this study can be summarized as follows:
  • Developed a DL-based framework using optimized transfer learning models for the automated pre-identification of CLS disease in persimmons, addressing the time-consuming and subjective nature of traditional detection methods.
  • Developed a highly curated dataset for CLS classification using manual analysis of OCT amplitude scans (A-scan) and loop-mediated isothermal amplification (LAMP), focusing on pre-identification of CLS.
  • Presented the inaugural application of OCT combined with DL for disease detection in plant leaves within agricultural contexts.
  • Utilized a novel combination of LAMP with A-scan for detecting CLS disease in persimmon.

2. Related Works

OCT gained prominence following the publication of a seminal paper by Huang et al. in 1991, which focused on the non-invasive cross-sectional imaging of biological systems [13]. This work demonstrated the application of OCT for imaging the retina within the human eye. Subsequent to this, advancements and widespread applications of OCT in acquiring images of the human body have been reported [14,15,16,17,18,19]. In the OCT scanning system illustrated in Figure 1, the optical light beam from the broadband light source splits into the reference and sample arms using an optical fiber coupler. A galvanometer-based optical scanner performs the sample scanning. The reflected signals from the sample and reference arm get interfered with by the beam coupler and forwarded to a spectrometer. The attached computer system will create the image of the sample from the data acquired through the spectrometer. This mechanism yields high-resolution cross-sectional images of biological tissues [20].
The OCT images can be assessed quantitatively and qualitatively using amplitude scans (A-scans), two-dimensional OCT (2D-OCT) images, and volumetrics. A-scans provide insights into the reflectivity or amplitude of light as a function of depth within the imaged tissue. B-scans generate a 2D cross-sectional image of the imaged tissue, while volumetric scanning produces a three-dimensional (3D) representation of the scanned sample.
In addition to being non-invasive and non-destructive, OCT is able to produce high-resolution images at the micrometer scale, making it an ideal diagnostic tool for a wide range of plant-related scientific endeavors. Wijesinghe et al. conducted a study on plant disease identification, focusing on analyzing morphological changes in infected tissues throughout the incubation period [3]. The research team developed a wearable backpack-type OCT system equipped with a customized spectral domain OCT (SD-OCT) for on-site inspections for grey leaf spot disease in Capsicum annuum and apple blotch in Marssonina coronaria, while a swept source OCT (SS-OCT) system was employed for identifying CLS disease in persimmon [11,21,22]. Also, a method was proposed by the same group for the early detection of CLS disease utilizing OCT imaging [3]. The primary aim of their study was to introduce a protocol for detecting CLS disease by assessing the reduction in thickness of the palisade parenchyma cell layer in leaf cross-sections, thereby curtailing disease spread. Furthermore, a histological examination was carried out to better understand the connection that exists between OCT images and the intended biological region of interest.
DL is a subset of machine learning (ML) that employs artificial neural networks (ANN) with multiple layers of interconnected nodes, allowing the system to learn intricate patterns and representations directly from raw data. It has revolutionized various fields, including computer vision, natural language processing, and speech recognition, by enabling the development of highly accurate and efficient models for complex tasks [23,24,25]. DL algorithms excel at automatically extracting features from large datasets, making them particularly well-suited for tasks such as image classification, object detection, and language translation. With advancements in hardware and algorithms, DL has become increasingly prevalent in industries ranging from healthcare to autonomous vehicles and robotics, driving innovations and transformative solutions across diverse domains [26,27].
The utilization of DL methods has witnessed a growing trend in the field of medical imaging, specifically for the automated analysis of medical images [28]. OCT has been shown to have a wide range of applications in the diagnosis of ocular diseases. There have been numerous studies that have been conducted to investigate these diseases by employing OCT in conjunction with DL approaches [29,30,31,32,33,34,35]. For instance, Karry et al. utilized a publicly available OCT dataset in 2017 to classify and recognize age-related macular degeneration (AMD) and diabetic macular edema (DME) from OCT images, achieving satisfactory results by employing GoogleLeNet a Convolutional Neural Network (CNN) developed by Google for the ImageNet competition [36]. In 2019, Wang et al. applied DL techniques to diagnose AMD and DME using two publicly available datasets. They trained a novel CNN model named CliqueNet and attained a high classification accuracy exceeding 98% for both datasets, comparing favorably with transfer learning methods utilizing architectures such as VGG16, VGG19, and ResNet50 [37]. Furthermore, Kugelman et al. employed a DL model combining CNN and a Recurrent Neural Network (RNN) for automatic choroidal segmentation in OCT images [38]. Despite the relatively limited use of OCT in agriculture, there have been notable attempts to apply DL in agricultural contexts for OCT analysis. Joshi et al. explored a novel approach for classifying rice seeds using OCT in conjunction with a Deep Neural Network (DNN), eliminating the need for labels and invasive procedures [39]. Likewise, Manhando et al. introduced an innovative framework aimed at early identification of mold-contaminated peanut kernels in the post-harvest phase, utilizing OCT images [40].

3. Materials and Methodology

The proposed pipeline, based on DL, aims to automatically pre-identify CLS disease in persimmon leaves, as illustrated in Figure 2. Initially, OCT B-scan images of leaf specimens are captured using a scanning device. Preprocessing is performed on these images in order to prepare them for subsequent analysis. Subsequently, the preprocessed images are fed into the first DL model, which determines the suitability of the image samples for CLS disease prediction. This model works as a binary classifier that outputs true for suitable images and false for images that lack quality. When images lack sufficient details, processing will be paused to facilitate either rescanning or moving to the next sample. Conversely, images with adequate detail progress to the next stage, where the second DL model is employed for classification. This model categorizes samples as infected (class-I), apparently healthy or apparently infected (class-HI-healthy infected), or non-infected (class-H-healthy).

3.1. Experimental Setup and Data Acquisition

3.1.1. Collection of Plant Materials

All of the persimmon leaf specimens were collected from a persimmon orchard located in Sangju in Gyeongbuk province, Korea. The entire persimmon orchard with an annual yield is treated with fungicides and insecticide during cultivation. A total number of ten trees were selected and the sample collection was conducted in a random process. The entire experiment was carried out once a week from early May to late September. During the experimental period, 5 leaf specimens were selected from each tree on experimental day, while examining a total of 1000 leaf specimens at 20 attempts.

3.1.2. System Configuration

The OCT image acquisition was conducted using a laboratory customized backpack type wearable OCT system. The OCT engine consists of a broadband light source consisting of a superluminescent diode (EXS210068-01, Exalos, Schlieren, Switzerland) with a center wavelength of 850 nm and a bandwidth of 55 nm. A line scan camera was used as the detector (spL2048-140 km, Basler, Ahrensburg, Germany). All of the components were precisely calibrated to enhance the axial resolution. The scanning depth of the system was measured as 3 mm, while the axial and lateral resolutions were measured as 8 μm and 12 μm. Detailed technical specifications can be found in [10,21].
The OCT signal fluctuation along the axial direction as well as the quantitative evaluation procedure fundamentally depends on the refractive index of the structure, since the refractive index of each deep layer differs from one to another. Although each layer in each depth level has a unique refractive index, a refractive index of 1.42, which is the fundamental refractive index of plant cells, was used herein. Therefore, once the particular refractive index of each layer is applied to the acquired 2D-OCT images, the corresponding most accurate axial direction depth profile can be obtained. Since the employed refractive index 1.42 provides sufficient visualization of the inner morphology, this value was used for the entire depth range.

3.1.3. LAMP Technique

To identify a clear physiological stage, all of the collected leaf specimens underwent LAMP examination followed by the OCT inspection. During LAMP examination, the collected leaf surface was sterilized. Next, these leaves were extracted using a mesh sample bag (Agdia, Elkhart, IN, USA) and tissue homogenizer (ACC00900, Agdia) with extraction buffer. Then, 1 µL of extraction buffer was added to a LAMP mixture, where the mixture content can be found in [22]. Afterwards, the mixture was incubated at the optimized temperature using a digital water bath (JSWB-22T, JSR, Gongju-City, Korea) for 5 min. Next, 1 µL of Bst polymerase large fragment (8 unit/µL, New England Biolabs, Ipswich, MA, USA) was added, and the samples were incubated at the optimized temperature for 1 h. A description of the overall LAMP process can be found in [22].

3.1.4. Dataset

The dataset comprised 1500 2D-OCT images obtained from the OCT system, with the scanning area predominantly centered on regions near the midrib of the leaf samples. Each image possessed a resolution of 500 × 200 pixels.

3.2. Pre-Processing

The original dataset comprises 1500 OCT B-scan images of leaf samples. DL models necessitate a larger number of training samples to effectively learn and generalize patterns from data [41]. Additionally, DL models typically operate optimally with square matrices due to computational considerations and network architecture constraints. To augment the dataset, a single image measuring 500 × 200 pixels was cropped at three distinct horizontal pixel locations, specifically at the 50th, 100th, and 150th positions, resulting in cropped images of dimensions 200 × 200 pixels each. The choice of cropping positions ensured that the cropping window remained below the 400th horizontal pixel to mitigate the presence of noise, commonly observed in all images beyond this threshold, which could adversely impact the learning process of the models. This augmentation process yielded a total of 4500 images. Figure 3 depicts the original OCT image obtained from a persimmon, as well as a cropped image derived from the original image.
Two distinct datasets were created to facilitate the demands of the two separate DL models of the pipeline. Due to the presence of low-detail images within the dataset, a new balanced dataset was constructed using all low-detail images and randomly selected high-detail images. This refined dataset was employed to train the initial DL model responsible for identifying high-quality images to be fed into the second DL model in the fourth stage of CLS identification, as depicted in Figure 2. For training, the second dataset was utilized, which contained 3747 images after removing low-detail images and images with label ambiguities. Since horizontal flipping does not affect OCT images in this use case, all selected images were horizontally flipped to augment the dataset, resulting in a total of 7494 samples. This augmented dataset encompassed 1872 healthy (class-H) images, 5160 healthy-infected (class-HI) images, and 462 infected (class-I) images. Figure 4 shows example persimmon leaf specimens falling into each category.
Speckle noise, characterized by random patterns of bright and dark speckles overlaying images, is a common occurrence in OCT imaging. This noise originates primarily from the scattering of light in multiple directions and the resulting constructive and destructive interference during the scanning process [42]. To mitigate the effects of speckle noise, the Block-matching and 3D filtering (BM3D) algorithm was employed to effectively reduce its presence to an acceptable level.

3.3. Labeling of Datasets

The initial dataset, comprising high-detail and low-detail images, was curated manually by meticulously examining all 4500 images. In the low-detail images, the tissue layers of the B-scan images of leaf samples were notably less discernible, rendering the identification of CLS disease more challenging. The dataset for the quality inspection model was generated based on the aforementioned dataset. All low-detail images were manually filtered out to ensure a balanced dataset, with an equal number of samples selected from the high-detail images. The resulting dataset comprised a total of 580 OCT B-scans (290 low-detail and 290 high-detail images). Figure 5 illustrates the disparities between a high-detail B-scan image and a low-detail B-scan image.
Several procedures were undertaken to label the dataset for the second model. LAMP represents a gold standard method for identifying plant pathogens. The original dataset was labeled based on LAMP tests that were carried out on leaf samples that were gathered from persimmon fields. These labeled leaf samples served as the ground truth for the labeling process. However, this can only detect whether the leaf sample is affected with the pathogen or not, dividing the samples into two classes. Pre-identification necessitates the detection of intermediary stages in order to allow for prompt treatment of plants prior to CLS spreading widely across the leaf.
The study undertaken by Wijesinghe et al. presents a methodology for differentiating between leaves that are in a healthy state, leaves that are looking healthy but infected, and leaves that are infected [3,43]. According to their analysis, infected leaves exhibit a thinner profile compared to healthy leaves. This differentiation is achieved by generating an A-scan profile from a corresponding B-scan image. Figure 6 illustrates the respective B-scan image alongside the generated A-scan image. The A-scan image depicts the mean intensity levels along the leaf’s surface as the scanner moves from the uppermost layer to the lowermost layer. The A-scan profile provides graphical representation of the different sub-layers present in the scanned leaf. The presence of CLS leads to a decrease in the separation between these sub-layers. This study suggests a strategy that utilizes the distance between two important layers, namely the epidermal layer and the spongy parenchyma layer, to simplify the process of labeling. In the A-profile, these layers exhibit the highest peaks, and the distance between these peaks serves as the basis for labeling class H, HI, and I samples. The labeling procedure is detailed in Table 1.
A MATLAB program was created to produce cropped images. Additionally, it was enhanced with a feature to generate A-scan profiles from these cropped images. Subsequently, it calculates the distance between the top two peaks in micrometers of the A-scan profiles. However, certain images presented challenges, such as unclear identification of peaks and the presence of multiple peaks unrelated to the calculation of interest. Consequently, all automatically labeled images underwent manual inspection using a dedicated MATLAB Graphical User Interface (GUI) program designed specifically for this purpose. This GUI comprises the label assigned by the automated labeling algorithm and respective LAMP of the sample. Manual labeling occurs through a comparison between these two values and the calculated peak difference. The label for a sample was manually assigned in cases of mismatch between the ground truth and the calculated label. If ambiguity persists between the ground truth and the label derived from the peak difference calculation, the respective images were excluded. The manual labeling procedure is outlined in Table 2.

3.4. Deep Learning Models

DL, which operates through ANN, has significantly revolutionized image identification and classification compared to alternative methods. Its capacity to automatically discern features from unstructured data by adjusting parameters and leveraging previously trained layers has enabled its widespread utilization in various analyses, particularly in morphological image studies [44]. Although the process of training DL models is time-consuming, the resulting trained models can be efficiently utilized in real-time to detect diseases. The pipeline’s process, which incorporates two DL models, is illustrated in Figure 7.
In this study, VGG-16 [45] was chosen as the optimal base model for the quality inspection model, which aims to classify high-detailed and low-detailed images. DenseNet-121 [46] was similarly selected as the most suitable base model for classifying samples into the categories of healthy (class-H), healthy-infected (class-HI), and infected (class-I). These models were preferred over a selection of other pre-trained models, including VGG-19, ResNet-50 [47], Inception3 [48], and Inception-ResNet-v2 [49].
DL models, such as VGG-16 and DenseNet-121, typically require a substantial number of training samples to reach an acceptable level of accuracy. However, the dataset collected for this study was insufficient to train all of the parameters of these models adequately. Consequently, transfer learning was employed to mitigate the impact of the limited dataset size. This approach has been widely utilized across various domains where large datasets are not readily available for training DL models [44,50]. The two pre-trained models employed in this study, namely VGG-16 and DenseNet-121, underwent pre-training using the ImageNet dataset, which encompasses 1.2 million color images spanning 1000 categories [51]. Subsequently, the final layer of each model was omitted, and a set of additional convolutional layers was introduced, followed by a set of fully connected layers for classification purposes. During the training process, the learning parameters of the original base model (VGG-16 or DenseNet-121) remained fixed, while the introduced parameters were permitted to be learned through the iterative training.
Both models were created using the Python programming language and the Google TensorFlow (v2.10.0) framework. These models share a similar code base, with variations made to accommodate different datasets, model types, and evaluation parameters. The models underwent training within a system equipped with specific hardware specifications, including a processor (Core i9-14900K, Intel, Santa Clara, CA, USA) having 24 physical cores and 32 threads, accompanied by 32 GB of memory. Additionally, the system was enhanced with a graphical processing unit (GPU) (GeForce RTX 4070, NVIDIA, Santa Clara, CA, USA) featuring 12 GB of memory, 5888 CUDA cores, and capable of delivering 568 Artificial Intelligence Trillions or Tera Operations per Second (AI TOPS).

3.5. Deep Learning Models for Quality Inspection

The quality inspection model was developed as a supplementary tool to enhance the detection accuracy of the primary objective of the study, which is the pre-identification of CLS. The dataset utilized for this model comprised 590 images, divided into training, validation, and testing subsets in a 3:1:1 ratio. Although the original image size was 200 × 200, it was scaled to 224 × 224, the default input dimension for most pre-trained models. The decision to use pre-trained models was driven by the limited number of samples.
The DL model is a customizable approach, allowing for easy adjustment of the backbone model, Fully Connected Network (FCN) layers, and other hyperparameters. In the best-performing model, all convolutional layers of the backbone model were frozen, and the FCN was connected by removing the initial FCN. The learning parameters of the FCN were randomly initialized for training purposes. This model performs binary classification to distinguish between high and low detailed images, using sigmoid activation in the final layer. Categorical cross-entropy loss was chosen as the loss function due to its efficacy with categorical data, and the optimizer used was stochastic gradient descent (SGD).
The model exhibited a tendency to overfit with a low number of iterations. Therefore, the number of layers in the FCN, dropout values in the layers, and L1 and L2 regularization were primarily tested to develop the best-performing model. The final optimal model included two layers for the FCN, with a hidden layer dropout value of 0.2 and an output layer configured with a kernel regularizer value of 0.04. This configuration resulted in the lowest bias and the best variance during training and testing. The quality inspection model was evaluated based on its overall accuracy and loss, given the balanced nature of the dataset.

3.6. Deep Learning Model for Identification of Circular Leaf Spot Disease

The CLS detection model is a crucial component of this study, inheriting similar development properties from the quality inspection model. Transfer learning was selected for this model due to its superior accuracy and generalization in classification compared to custom-built models. Nine pre-trained models were employed in this investigation: VGG-16, VGG-19, Inception-V3, Inception-ResNet-V3, ResNet-101, ResNet-50, EfficientNetB0, Xception, and DenseNet-121.
This model also utilizes input images of size 224 × 224, as square matrices offer benefits to CNNs by simplifying the application of filters in convolutions, efficiently sharing parameters across the image through the grid structure and standardizing the training data by establishing a uniform input format from images with different aspect ratios. The original backbone’s FCN was removed, and a new FCN was appended for CLS identification. Various foundational configurations of the FCN were explored in this study, with ReLU activation in the hidden layers and SoftMax activation in the output layer. Throughout the different phases of the study, various model variations were developed based on these foundational architectures by altering hyperparameters such as learning rate, batch size, and regularization parameters. The impact of each hyperparameter and model variation is detailed in the results section for selection of the best performing model.
A program was developed to facilitate the rapid modification of the pre-trained model and the incorporation of new layers. During the training of the DL models, initial hyperparameters were configured, and training commenced. Following the training phase, if evaluation results indicated overfitting, two adjustments were made. First, efforts were made to increase the dataset size and improve its quality through data cleaning procedures. Second, hyperparameters related to regularization, such as dropout and L1 and L2 regularization methods, were modified. If overfitting was not observed, the model’s accuracy was assessed. If the accuracy was deemed unsatisfactory, adjustments were made to the hyperparameters. If this proved ineffective, modifications were made to the architecture of the added layers. This iterative process was conducted across multiple pre-trained models to identify the optimal model architecture.
The last layer of the best performing model (DenseNet 121) was eliminated for the CLF identification model. Following this, a pooling layer was inserted after the final convolution layer, which was succeeded by the addition of five dense layers. These dense layers incorporated dropout values of 0.5, 0.7, 0.5, 0.7, and 0.7, along with ReLU activation functions. The last layer comprised three outputs with a SoftMax activation function. The Adam optimizer was utilized to optimize the weights and biases of the DenseNet 121, employing a learning rate of 0.000125. The dataset was partitioned into training, validation, and testing sets in a ratio of 3:1:1.
During the training phase of each model, configurations were implemented to cease the training process if the gradient descent algorithm failed to converge after a specified number of epochs. This prevents the models from undergoing unnecessary epochs due to a lack of convergence. Additionally, each model was programmed to save its optimal version. Therefore, even if the model’s testing accuracy decreases at the cessation of training, the model has already stored the best instance with the highest accuracy.
The model was evaluated using Area Under the Curve–Receiver Operating Characteristics (AUC–ROC), precision, and recall, along with overall accuracy and loss, to ensure comprehensive assessment despite the imbalanced dataset (refer to the Supplementary Materials).

4. Experimental Results and Analysis

4.1. Quality Inspection Model

Table 3 compares the models utilized for quality inspection, which involves classifying the input data as either low-detailed or high-detailed images. Accuracy was selected as the primary metric to assess the model performance because the training dataset was balanced. Due to its superior accuracy, the VGG-16 model was the most effective of the three models tested (VGG-16, InceptionV3, and Inception-ResNet-V2). The model reached convergence after approximately 20 epochs, resulting in a loss of 0.1608. While the other three models also demonstrated accuracy levels close to VGG-16, some base models like DenseNet-121 and ResNet-50 exhibited over fitting, converging at fewer epochs. L1 regularization was implemented on the final layer (output sigmoid layer) with a value of 0.01 to address over fitting. Although dropouts can also be employed to address over fitting, increasing other learning parameters and adding new layers with dropouts may exacerbate over fitting in this particular use case. Figure 8 illustrates the model convergence during the training of the VGG-16 model.
The model architecture was tested with other pre-trained models used in the CLS detection model. However, these models tended to overfit quickly compared to the selected three backbone models. Despite the model performing exceptionally well with the current dataset, its effectiveness is limited by the lack of diverse data for accurately identifying low-detail or noisy images. The dataset lacks different categories of noisy images that were taken while scanning the samples, in addition to several augmentations being made to it, such as flipping and adding synthetic noise. Incorporating a variety of noisy images will aid in mitigating over fitting during the training phase. Transfer learning leverages pre-trained models, which have already been exposed to a wide range of features and patterns, making them more robust and effective in handling unseen datasets.

4.2. Circular Leaf Spot Disease Detection Model

The comparative testing results of different baseline models employed in the development of the CLS disease stage identification model are presented in Table 4. The evaluation metrics used to assess the effectiveness of these models for CLS stage classification include OA, AUC–ROC for each class compared to others, micro-average AUC–ROC, precision, and recall.
Figure 9 depicts the comparison of OA among the models. As indicated in the chart, all models achieved an accuracy of 65%, with DenseNet-121 and the Xception base models exhibiting the highest accuracies, surpassing the 85% threshold. However, due to the imbalance in the dataset, a definitive decision regarding the best model cannot be made, similar to the findings observed in the quality inspection model.
An AUC–ROC is used to assess the adequate performance of a classification model, specifically in the context of binary classification tasks. Hence, the AUC–ROC was calculated for each class compared to the other two classes in the testing dataset, aiming to demonstrate the model’s performance for each class explicitly. Figure 10 displays the plot extracted from Table 4. A value of 0.5 signifies the poorest model performance, while 1 indicates the best performance in classifying each class. Notably, EfficientNetB0 exhibits the lowest performance, followed by ResNet-50 and ResNet-101. Conversely, DenseNet-121, VGG-16, VGG-19 models, InceptionV3, and Xception-based models demonstrate AUC–ROC values exceeding 0.75. The highest AUC for “H vs. Others” is observed in the Xception model at 0.85, followed by DenseNet-121 at 0.84. In addition, DenseNet-121 achieves the highest AUC–ROC values of 0.84 and 0.91 for the “HI vs. Others” and “I vs. Others” scenarios, respectively. The micro-average AUC–ROC for each model is also depicted in Figure 10. According to the chart, all of the values surpass 0.75, with the DenseNet-121 and Xception base models exhibiting the highest values.
The precision and recall metrics complement the insights provided by the AUC–ROC analysis for one class versus others. As depicted in Figure 9, ResNet-50 achieves the highest precision, close to 0.8, for class-H. However, the recall value for the ResNet-50 model is notably lower, at less than 0.25. This indicates that while the model correctly identifies a substantial proportion of samples belonging to class-H out of all of its predictions, it also overlooks a significant number of actual class-H samples that should have been identified as such. Similarly, InceptionResNetV2 exhibits the highest precision for class-I, but it faces a similar challenge as ResNet-50 with class-H, as evidenced by its very low recall value of 0.0967. On the other hand, ResNet-50, ResNet-101, and EfficientNetB0 struggle to accurately identify class-I. EfficientNetB0 demonstrates a precision of 0 for all classes and a recall value of 1 for class-HI, indicating an inability to correctly classify any class, leading to an output of HI for any input.
Although the OA and micro-averaged AUC–ROC suggest relatively similar performances across all models, it is imperative to consider additional metrics when dealing with imbalanced datasets to identify the best-performing model accurately. DenseNet-121, VGG16, and Xception-based models exhibit higher precision and recall than the others. Notably, DenseNet-121 stands out with the highest precision values of 0.7823 for class-H, 0.9005 for class-HI, and 0.7027 for class-I, supported by the highest recall values of 0.8953 for class-HI and 0.8387 for class-I, as detailed in Table 4. Consequently, DenseNet-121 has been identified as the optimal model for CLS stage identification.
The generated confusion matrix depicted in Figure 11 illustrates the model’s proficient identification of a substantial number of testing samples across the H, HI, and I classes. Notably, the model successfully avoided misclassifying any infected leaf samples (class-I) as healthy (class-H), a critical aspect given the imperative need for accurate disease diagnosis. However, approximately 10% of class-HI samples were erroneously classified as class-H. This misclassification is attributed to instances where there exists minimal discernible difference between the H and HI stages of the disease. An example of such instances is presented in Figure 12, showcasing the subtle disparities between H and HI classifications. Conversely, misclassifying H as HI can similarly occur due to the aforementioned minimal differentiations between these stages.

5. Discussion

The primary focus of this investigation was to develop a method for identifying three stages of CLS, namely healthy (class-H), healthy infected (class-HI), and infected (class-I), in persimmon leaves, with a subsequent emphasis on pre-identification of the disease. To achieve this objective, a detection pipeline was formulated utilizing DL techniques integrated with SD-OCT. While prior studies have examined CLS disease diagnosis [2,4], there is a scarcity of research that focuses on using OCT for this purpose [3]. The proposed DL-based pipeline demonstrates promising diagnostic performance for automated identification of CLS disease and exhibits notable proficiency in distinguishing between its different stages. It is worthy to note that the results acquired through our previously published work [3] verified the thickness threshold ranges to confirm the physical state of the leaf by measuring the thickness difference between the first two peaks of the A-scan profiles, which represent the epidermal and spongy parenchyma cell layers. Although the statistical values and quantifications such as the average, standard deviations, and minimum and maximum fluctuation range were approximated, confirming the accuracy of the study was a challenge due to the limited data analysis of a small number of amplitude scan analyses and the subjectivity of the A-scan selection. To address this major shortcoming, the proposed artificial intelligence (AI) incorporated fully automated diagnosis of CLS aims to reduce the subjectivity and time-consuming process involved in interpreting A-scan profiles generated from OCT images for CLS identification by employing deep learning techniques, which eventually confirmed a higher number accuracy that the previously conducted study.
Although DL-based methods are used to discern between infected and non-infected cases, which is highly critical in most practical agricultural contexts, circular leaf spot disease (CLS) has a long incubation period of 90–120 days, during which leaves show no symptoms until the disease is widespread. Since the apparently healthy (HI) stage is indistinguishable from the healthy stage (H) by the naked eye, identifying the disease at the HI stage is crucial for effective disease management before it becomes severe. To overcome the limitations of histological analysis, PCR, and LAMP, OCT was incorporated as a non-destructive imaging method for the early identification of CLS. Although (H) stage and (I) stage are visually distinguishable, the differentiation between (H) and (HI) is highly challenging. Also, the reliable detection of the ‘H’ class is important to avoid unnecessary exposure to pesticides, and ‘HI’ class is important for the pre-identification of the disease spread. Therefore, the proposed DL approach enables the differentiation among all three classes (‘H’, ‘HI’, and ‘I’) to ensure accurate and efficient disease management. The incorporation of both manual A-scan profile analysis and LAMP analysis into the DL approach resulted in higher accuracy compared to using manual A-scan analysis alone. Moreover, this study also marks the first use of a combination of LAMP with A-scan for the early detection of CLS disease.
The integration of OCT and DL for automated classification within agricultural contexts remains uncommon. Prior work by Josh et al. [39] focused on rice seed classification employing a modified Inception-ResNet-V2 model, while Manhando et al. investigated early detection of mold-contaminated peanuts using a modified VGG-16 model, yielding promising outcomes. In contrast, this study is the first attempt to use OCT with DL to automatically identify diseases that spread across plant leaves. However, extensive research has been undertaken regarding the automated diagnosis of medical conditions using OCT and DL. The utilization of the DenseNet-121 and VGG-16 models used in this study resulted in positive outcomes within the medical domain [52,53,54,55,56].
In this study, two CNN models were constructed, leveraging pre-trained DL models renowned for their efficacy in handling small datasets. Although the primary focus of the study pertains to the identification of CLS stages, a supplementary quality inspection model was introduced following pre-processing to mitigate the impact of low-detail images on CLS disease identification. This addition notably enhanced the performance of the CLS detection model. The modified DenseNet-121 model developed for CLS detection demonstrated promising performance, particularly in precision across all classes and recall values for specific classes, resulting in an impressive overall AUC–ROC of 0.89. The evaluation of quality inspection models revealed that the modified VGG-16 model achieved the highest accuracy of 98.99% among various pre-trained models, specifically tailored for discerning low and high detailed images.
Overall, the proposed models demonstrate a robust capacity to effectively discern classes HI and I with elevated recall (sensitivity) according to Figure 11, thereby reducing the risk of erroneously categorizing diseased leaves as healthy. This aspect holds significant importance for farmers, as it enables prompt initiation of treatments for afflicted crops.
The predictive accuracy of the employed DL models in the study could be enhanced through various means. A primary limitation lies in the scarcity of the dataset, as DL models typically require extensive datasets to effectively train their parameters [41]. Furthermore, not all of the images in the dataset could be taken into consideration due to their poor level of detail. Additionally, inconsistencies in the labeling procedure were observed when comparing results obtained from LAMP tests and A-scan profile analysis. There were several instances that were not taken into consideration since the LAMP result suggested class-H, but the A-scan result indicated class-I. These inconsistencies can be attributed to differences in the physical thickness of leaves across different scanning locations. The model is customized for the OCT scanning data used in this study but faces challenges in generalizing to accommodate diverse images from different devices and conditions, similar to issues in other imaging modalities [57]. As a result, efforts are needed to enhance the model’s ability to generalize across various datasets with distinct properties and acquisition conditions.
The suggested framework for the early detection of CLS holds promise for integration into an OCT scanning system, particularly one utilized in prior investigations by Wijesinghe et al. [21]. By incorporating this framework into a handheld scanner, the capability for real-time, in situ disease identification in field settings would be enhanced, facilitating the development of a farmer-centric product for automated disease detection with high accuracy, eliminating the necessity for expert knowledge in plant diseases.

6. Conclusions

In conclusion, this study developed a detection pipeline for the pre-identification of CLS in persimmon leaves by integrating DL techniques with SD-OCT. While previous studies have explored manual CLS disease diagnosis, this investigation represents a pioneering effort to automate CLS identification, addressing the issues of time-consuming analysis and subjectivity found in earlier methods. The image labeling process, incorporating both A-scan analysis and LAMP results, significantly enhanced the accuracy for efficient pre-identification of CLS compared to manual analysis relying solely on A-scan profiles.
The application of CNNs, particularly the adapted DenseNet-121 and VGG-16 models, displayed promising diagnostic capabilities. The adapted DenseNet-121 model excels in differentiating between three stages of CLS disease in persimmon leaves, including healthy (class-H), healthy-looking (class-HI), and infected (class-I), specifically focusing on the pre-identification of CLS through the “healthy-looking” class. This adapted DenseNet-121 model achieved commendable precision values of 0.782 for class-H, 0.9 for class-HI, and 0.702 for class-I, supported by the highest recall values of 0.8953 for class-HI and 0.838 for class-I. The performance of CLS detection was further boosted by the introduction of a supplemental quality inspection model adapted using VGG-16, which achieved an accuracy of 98.99% in discriminating between low-detailed images and high-detailed images.
This research highlights the capacity of DL techniques combined with OCT to improve disease identification procedures, particularly within agricultural contexts. By offering effective and objective pre-identification of CLS, this study enables early intervention and management approaches in persimmon cultivation.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s24165398/s1.

Author Contributions

Conceptualization, U.W., R.E.W., D.K., J.K. and M.J.; methodology, D.K.; software, D.K.; validation, U.W., B.N.S. and D.K.; formal analysis, D.K. and B.N.S.; investigation, D.K., U.W. and R.E.W.; resources, R.E.W., U.W. and D.K.; data curation, D.K. and R.E.W.; writing—original draft preparation, N.S.K. and D.K.; writing—review and editing, N.S.K., B.N.S. and D.K.; visualization, N.S.K., B.N.S. and D.K.; supervision, U.W. and R.E.W.; project administration, U.W. and R.E.W.; funding acquisition, U.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Science and Technology Human Resource Development Project, Ministry of Education, Sri Lanka, funded by the Asian Development Bank (Grant No. STHRD/CRG/R3/SJ/07), Research grant funded by the University of Sri Jayewardenepura, Sri Lanka (Grant No: ASP/01/RE/ENG/2022/86), and research grants funded by the Sri Lanka Institute of Information Technology, Sri Lanka (Grant Nos. PVC (R&D) RG/2024/13 and PVC (R&D) RG/2024/16).

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

AMDAge-related macular degeneration
ANNArtificial neural networks
AUCArea under the curve
BNBatch normalization
CLSCircular leaf spot
CNNConvolutional neural network
CUDACompute unified device architecture
DLDeep learning
DMEDiabetic macular edema
DNNDeep neural network
FPRFalse positive rate
GPUGraphical processing unit
GUIGraphical user interface
LAMPLoop-mediated isothermal amplification
MLMachine learning
MRIMagnetic resonance imaging
OAOverall accuracy
OCTOptical coherence tomography
PCRPolymerase chain reaction
ReLURectified linear unit
RNNRecurrent neural network
ROCReceiver operating characteristics
SD-OCTSpectral domain OCT
SGDStochastic gradient descent
TPRTrue positive rate

References

  1. Hassan, O.; Chang, T. Phylogenetic and Morphological Reassessment of Mycosphaerella nawae, the Causal Agent of Circular Leaf Spot in Persimmon. Plant Dis. 2019, 103, 200–213. [Google Scholar] [CrossRef] [PubMed]
  2. Choi, O.; Park, J.-J.; Kang, B.; Lee, Y.; Park, J.; Kwon, J.-H.; Kim, J. Evaluation of Ascospore Prediction Model for Circular Leaf Spot Caused by Mycosphaerella nawae of Persimmon. J. Agric. Life Sci. 2018, 52, 13–19. [Google Scholar] [CrossRef]
  3. Wijesinghe, R.E.; Lee, S.-Y.; Kim, P.; Jung, H.-Y.; Jeon, M.; Kim, J. Optical Inspection and Morphological Analysis of Diospyros Kaki Plant Leaves for the Detection of Circular Leaf Spot Disease. Sensors 2016, 16, 1282. [Google Scholar] [CrossRef] [PubMed]
  4. Berbegal, M.; Mora-Sala, B.; García-Jiménez, J. A Nested-Polymerase Chain Reaction Protocol for the Detection of Mycosphaerella Nawae in Persimmon. Eur. J. Plant Pathol. 2013, 137, 273–281. [Google Scholar] [CrossRef]
  5. Singh, V.; Sharma, N.; Singh, S. A Review of Imaging Techniques for Plant Disease Detection. Artif. Intell. Agric. 2020, 4, 229–242. [Google Scholar] [CrossRef]
  6. Ayache, J.; Beaunier, L.; Boumendil, J.; Ehret, G.; Laub, D. Sample Preparation Handbook for Transmission Electron Microscopy; Springer: New York, NY, USA, 2010; ISBN 978-1-4419-5974-4. [Google Scholar]
  7. Aniq, H.; Campbell, R. Magnetic Resonance Imaging. In Pain Management; Elsevier: Amsterdam, The Netherlands, 2011; pp. 106–116. ISBN 978-1-4377-0721-2. [Google Scholar]
  8. Abdhul Rahuman, M.A.; Kahatapitiya, N.S.; Amarakoon, V.N.; Wijenayake, U.; Silva, B.N.; Jeon, M.; Kim, J.; Ravichandran, N.K.; Wijesinghe, R.E. Recent Technological Progress of Fiber-Optical Sensors for Bio-Mechatronics Applications. Technologies 2023, 11, 157. [Google Scholar] [CrossRef]
  9. Manattayil, J.K.; Ravichandran, N.K.; Wijesinghe, R.E.; Shirazi, M.F.; Lee, S.-Y.; Kim, P.; Jung, H.-Y.; Jeon, M.; Kim, J. Non-Destructive Classification of Diversely Stained Capsicum Annuum Seed Specimens of Different Cultivars Using Near-Infrared Imaging Based Optical Intensity Detection. Sensors 2018, 18, 2500. [Google Scholar] [CrossRef] [PubMed]
  10. Wijesinghe, R.E.; Lee, S.-Y.; Ravichandran, N.K.; Shirazi, M.F.; Moon, B.; Jung, H.-Y.; Jeon, M.; Kim, J. Bio-Photonic Detection Method for Morphological Analysis of Anthracnose Disease and Physiological Disorders of Diospyros Kaki. Opt. Rev. 2017, 24, 199–205. [Google Scholar] [CrossRef]
  11. Ravichandran, N.K.; Wijesinghe, R.E.; Shirazi, M.F.; Kim, J.; Jung, H.-Y.; Jeon, M.; Lee, S.-Y. In Vivo Non-Destructive Monitoring of Capsicum Annuum Seed Growth with Diverse NaCl Concentrations Using Optical Detection Technique. Sensors 2017, 17, 2887. [Google Scholar] [CrossRef]
  12. Yang, D.; Ran, A.R.; Nguyen, T.X.; Lin, T.P.H.; Chen, H.; Lai, T.Y.Y.; Tham, C.C.; Cheung, C.Y. Deep Learning in Optical Coherence Tomography Angiography: Current Progress, Challenges, and Future Directions. Diagnostics 2023, 13, 326. [Google Scholar] [CrossRef]
  13. Huang, D.; Swanson, E.A.; Lin, C.P.; Schuman, J.S.; Stinson, W.G.; Chang, W.; Hee, M.R.; Flotte, T.; Gregory, K.; Puliafito, C.A.; et al. Optical Coherence Tomography. Science 1991, 254, 1178–1181. [Google Scholar] [CrossRef] [PubMed]
  14. Bhende, M.; Shetty, S.; Parthasarathy, M.; Ramya, S. Optical Coherence Tomography: A Guide to Interpretation of Common Macular Diseases. Indian J. Ophthalmol. 2018, 66, 20. [Google Scholar] [CrossRef] [PubMed]
  15. Cho, N.H.; Lee, S.H.; Jung, W.; Jang, J.H.; Kim, J. Optical Coherence Tomography for the Diagnosis and Evaluation of Human Otitis Media. J. Korean Med. Sci. 2015, 30, 328. [Google Scholar] [CrossRef] [PubMed]
  16. Burwood, G.W.S.; Fridberger, A.; Wang, R.K.; Nuttall, A.L. Revealing the Morphology and Function of the Cochlea and Middle Ear with Optical Coherence Tomography. Quant. Imaging Med. Surg. 2019, 9, 858–881. [Google Scholar] [CrossRef] [PubMed]
  17. Zeppieri, M.; Marsili, S.; Enaholo, E.S.; Shuaibu, A.O.; Uwagboe, N.; Salati, C.; Spadea, L.; Musa, M. Optical Coherence Tomography (OCT): A Brief Look at the Uses and Technological Evolution of Ophthalmology. Medicina 2023, 59, 2114. [Google Scholar] [CrossRef] [PubMed]
  18. Hsieh, Y.-S.; Ho, Y.-C.; Lee, S.-Y.; Chuang, C.-C.; Tsai, J.; Lin, K.-F.; Sun, C.-W. Dental Optical Coherence Tomography. Sensors 2013, 13, 8928–8949. [Google Scholar] [CrossRef] [PubMed]
  19. Sattler, E.; Kästle, R.; Welzel, J. Optical Coherence Tomography in Dermatology. J. Biomed. Opt. 2013, 18, 061224. [Google Scholar] [CrossRef] [PubMed]
  20. Aumann, S.; Donner, S.; Fischer, J.; Müller, F. Optical Coherence Tomography (OCT): Principle and Technical Realization. In High Resolution Imaging in Microscopy and Ophthalmology: New Frontiers in Biomedical Optics; Bille, J.F., Ed.; Springer International Publishing: Cham, Switzerland, 2019; pp. 59–85. ISBN 978-3-030-16638-0. [Google Scholar]
  21. Wijesinghe, R.; Lee, S.-Y.; Ravichandran, N.K.; Han, S.; Jeong, H.; Han, Y.; Jung, H.-Y.; Kim, P.; Jeon, M. Optical Coherence Tomography-Integrated, Wearable (Backpack-Type), Compact Diagnostic Imaging Modality for in Situ Leaf Quality Assessment. Appl. Opt. 2017, 56, D108. [Google Scholar] [CrossRef]
  22. Lee, J.; Lee, S.-Y.; Wijesinghe, R.E.; Ravichandran, N.K.; Han, S.; Kim, P.; Jeon, M.; Jung, H.-Y.; Kim, J. On-Field In Situ Inspection for Marssonina Coronaria Infected Apple Blotch Based on Non-Invasive Bio-Photonic Imaging Module. IEEE Access 2019, 7, 148684–148691. [Google Scholar] [CrossRef]
  23. Chen, L.; Li, S.; Bai, Q.; Yang, J.; Jiang, S.; Miao, Y. Review of Image Classification Algorithms Based on Convolutional Neural Networks. Remote Sens. 2021, 13, 4712. [Google Scholar] [CrossRef]
  24. Khurana, D.; Koli, A.; Khatter, K.; Singh, S. Natural Language Processing: State of the Art, Current Trends and Challenges. Multimed. Tools Appl. 2023, 82, 3713–3744. [Google Scholar] [CrossRef]
  25. Mehrish, A.; Majumder, N.; Bhardwaj, R.; Mihalcea, R.; Poria, S. A Review of Deep Learning Techniques for Speech Processing. arXiv 2023, arXiv:2305.00359. [Google Scholar] [CrossRef]
  26. Bhattacharya, G. From DNNs to GANs: Review of Efficient Hardware Architectures for Deep Learning. arXiv 2021, arXiv:2107.00092. [Google Scholar] [CrossRef]
  27. Alzubaidi, L.; Zhang, J.; Humaidi, A.J.; Al-Dujaili, A.; Duan, Y.; Al-Shamma, O.; Santamaría, J.; Fadhel, M.A.; Al-Amidie, M.; Farhan, L. Review of Deep Learning: Concepts, CNN Architectures, Challenges, Applications, Future Directions. J. Big Data 2021, 8, 53. [Google Scholar] [CrossRef]
  28. Shen, D.; Wu, G.; Suk, H.-I. Deep Learning in Medical Image Analysis. Annu. Rev. Biomed. Eng. 2017, 19, 221–248. [Google Scholar] [CrossRef] [PubMed]
  29. Tasnim, N.; Hasan, M.; Islam, I. Comparisonal Study of Deep Learning Approaches on Retinal OCT Image. arXiv 2019, arXiv:1912.07783. [Google Scholar]
  30. Fang, L.; Wang, C.; Li, S.; Rabbani, H.; Chen, X.; Liu, Z. Attention to Lesion: Lesion-Aware Convolutional Neural Network for Retinal Optical Coherence Tomography Image Classification. IEEE Trans. Med. Imaging 2019, 38, 1959–1970. [Google Scholar] [CrossRef]
  31. Fu, D.J.; Glinton, S.; Lipkova, V.; Faes, L.; Liefers, B.; Zhang, G.; Pontikos, N.; McKeown, A.; Scheibler, L.; Patel, P.J.; et al. Deep-Learning Automated Quantification of Longitudinal OCT Scans Demonstrates Reduced RPE Loss Rate, Preservation of Intact Macular Area and Predictive Value of Isolated Photoreceptor Degeneration in Geographic Atrophy Patients Receiving C3 Inhibition Treatment. Br. J. Ophthalmol. 2024, 108, 536–545. [Google Scholar] [CrossRef] [PubMed]
  32. Mariottoni, E.B.; Datta, S.; Shigueoka, L.S.; Jammal, A.A.; Tavares, I.M.; Henao, R.; Carin, L.; Medeiros, F.A. Deep Learning–Assisted Detection of Glaucoma Progression in Spectral-Domain OCT. Ophthalmol. Glaucoma 2023, 6, 228–238. [Google Scholar] [CrossRef]
  33. Marvdashti, T.; Duan, L.; Aasi, S.Z.; Tang, J.Y.; Ellerbee Bowden, A.K. Classification of Basal Cell Carcinoma in Human Skin Using Machine Learning and Quantitative Features Captured by Polarization Sensitive Optical Coherence Tomography. Biomed. Opt. Express 2016, 7, 3721. [Google Scholar] [CrossRef]
  34. Butola, A.; Prasad, D.K.; Ahmad, A.; Dubey, V.; Qaiser, D.; Srivastava, A.; Senthilkumaran, P.; Ahluwalia, B.S.; Mehta, D.S. Deep Learning Architecture LightOCT for Diagnostic Decision Support Using Optical Coherence Tomography Images of Biological Samples. arXiv 2020, arXiv:1812.02487. [Google Scholar]
  35. Liu, X.; Ouellette, S.; Jamgochian, M.; Liu, Y.; Rao, B. One-Class Machine Learning Classification of Skin Tissue Based on Manually Scanned Optical Coherence Tomography Imaging. Sci. Rep. 2023, 13, 867. [Google Scholar] [CrossRef] [PubMed]
  36. Karri, S.P.K.; Chakraborty, D.; Chatterjee, J. Transfer Learning Based Classification of Optical Coherence Tomography Images with Diabetic Macular Edema and Dry Age-Related Macular Degeneration. Biomed. Opt. Express 2017, 8, 579–592. [Google Scholar] [CrossRef] [PubMed]
  37. Wang, D.; Wang, L. On OCT Image Classification via Deep Learning. IEEE Photonics J. 2019, 11, 3900714. [Google Scholar] [CrossRef]
  38. Kugelman, J.; Alonso-Caneiro, D.; Read, S.A.; Hamwood, J.; Vincent, S.J.; Chen, F.K.; Collins, M.J. Automatic Choroidal Segmentation in OCT Images Using Supervised Deep Learning Methods. Sci. Rep. 2019, 9, 13298. [Google Scholar] [CrossRef] [PubMed]
  39. Joshi, D.; Butola, A.; Kanade, S.R.; Prasad, D.K.; Amitha Mithra, S.V.; Singh, N.K.; Bisht, D.S.; Mehta, D.S. Label-Free Non-Invasive Classification of Rice Seeds Using Optical Coherence Tomography Assisted with Deep Neural Network. Opt. Laser Technol. 2021, 137, 106861. [Google Scholar] [CrossRef]
  40. Manhando, E.; Zhou, Y.; Wang, F. Early Detection of Mold-Contaminated Peanuts Using Machine Learning and Deep Features Based on Optical Coherence Tomography. AgriEngineering 2021, 3, 703–715. [Google Scholar] [CrossRef]
  41. Benkendorf, D.J.; Hawkins, C.P. Effects of Sample Size and Network Depth on a Deep Learning Approach to Species Distribution Modeling. Ecol. Inform. 2020, 60, 101137. [Google Scholar] [CrossRef]
  42. Kompanets, I.; Zalyapin, N. Methods and Devices of Speckle-Noise Suppression (Review). Opt. Photonics J. 2020, 10, 219–250. [Google Scholar] [CrossRef]
  43. Saleah, S.A.; Lee, S.-Y.; Wijesinghe, R.E.; Lee, J.; Seong, D.; Ravichandran, N.K.; Jung, H.-Y.; Jeon, M.; Kim, J. Optical Signal Intensity Incorporated Rice Seed Cultivar Classification Using Optical Coherence Tomography. Comput. Electron. Agric. 2022, 198, 107014. [Google Scholar] [CrossRef]
  44. Kim, H.E.; Cosa-Linan, A.; Santhanam, N.; Jannesari, M.; Maros, M.E.; Ganslandt, T. Transfer Learning for Medical Image Classification: A Literature Review. BMC Med. Imaging 2022, 22, 69. [Google Scholar] [CrossRef] [PubMed]
  45. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar] [CrossRef]
  46. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  47. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  48. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  49. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, Inception-Resnet and the Impact of Residual Connections on Learning. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar]
  50. Gupta, J.; Pathak, S.; Kumar, G. Deep Learning (CNN) and Transfer Learning: A Review. J. Phys. Conf. Ser. 2022, 2273, 012029. [Google Scholar] [CrossRef]
  51. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Li, F.-F. ImageNet: A Large-Scale Hierarchical Image Database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  52. Aoyama, Y.; Maruko, I.; Kawano, T.; Yokoyama, T.; Ogawa, Y.; Maruko, R.; Iida, T. Diagnosis of Central Serous Chorioretinopathy by Deep Learning Analysis of En Face Images of Choroidal Vasculature: A Pilot Study. PLoS ONE 2021, 16, e0244469. [Google Scholar] [CrossRef] [PubMed]
  53. An, G.; Akiba, M.; Yokota, H.; Motozawa, N.; Takagi, S.; Mandai, M.; Kitahata, S.; Hirami, Y.; Takahashi, M.; Kurimoto, Y. Deep Learning Classification Models Built with Two-Step Transfer Learning for Age Related Macular Degeneration Diagnosis. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; pp. 2049–2052. [Google Scholar]
  54. Lee, C.S.; Baughman, D.M.; Lee, A.Y. Deep Learning Is Effective for the Classification of OCT Images of Normal versus Age-Related Macular Degeneration. Ophthalmol. Retin. 2017, 1, 322–327. [Google Scholar] [CrossRef] [PubMed]
  55. Sotoudeh-Paima, S.; Jodeiri, A.; Hajizadeh, F.; Soltanian-Zadeh, H. Multi-Scale Convolutional Neural Network for Automated AMD Classification Using Retinal OCT Images. Comput. Biol. Med. 2022, 144, 105368. [Google Scholar] [CrossRef] [PubMed]
  56. Khan, A.; Pin, K.; Aziz, A.; Han, J.W.; Nam, Y. Optical Coherence Tomography Image Classification Using Hybrid Deep Learning and Ant Colony Optimization. Sensors 2023, 23, 6706. [Google Scholar] [CrossRef]
  57. Sarah, M.; Mathieu, L.; Philippe, Z.; Guilcher, A.L.; Borderie, L.; Cochener, B.; Quellec, G. Generalizing Deep Learning Models for Medical Image Classification. arXiv 2024, arXiv:2403.12167. [Google Scholar] [CrossRef]
Figure 1. Schematic of the spectral domain optical coherence tomography system (SD-OCT), highlighting the reference arm, sample arm, spectrometer, computer and key components such as the broadband light source (BLS), fiber coupler (FC), collimator (C), lenses (L), mirror (M), sample (S) and galvanometer scanner (GS).
Figure 1. Schematic of the spectral domain optical coherence tomography system (SD-OCT), highlighting the reference arm, sample arm, spectrometer, computer and key components such as the broadband light source (BLS), fiber coupler (FC), collimator (C), lenses (L), mirror (M), sample (S) and galvanometer scanner (GS).
Sensors 24 05398 g001
Figure 2. Deep learning (DL) based pipeline for pre-identification of circular leaf spot (CLS) disease of persimmon. Class H, HI and I denote healthy, healthy-infected and infected, respectively.
Figure 2. Deep learning (DL) based pipeline for pre-identification of circular leaf spot (CLS) disease of persimmon. Class H, HI and I denote healthy, healthy-infected and infected, respectively.
Sensors 24 05398 g002
Figure 3. OCT images acquired from scanning the leaf samples of persimmon. (a) Original OCT image acquired with 500 × 200 pixels of size. (b) The image cropped from the original OCT image. Respective cropped position is shown in the original image.
Figure 3. OCT images acquired from scanning the leaf samples of persimmon. (a) Original OCT image acquired with 500 × 200 pixels of size. (b) The image cropped from the original OCT image. Respective cropped position is shown in the original image.
Sensors 24 05398 g003
Figure 4. Photographs of the persimmon leaves from the infected and healthy trees. (a) Leaf from a healthy tree. (b) Healthy-looking leaf from an infected tree. (c) Infected leaf.
Figure 4. Photographs of the persimmon leaves from the infected and healthy trees. (a) Leaf from a healthy tree. (b) Healthy-looking leaf from an infected tree. (c) Infected leaf.
Sensors 24 05398 g004
Figure 5. Sample OCT-B scan images utilized for training the quality inspection model. (a) High-detail image. (b) Low-detail image.
Figure 5. Sample OCT-B scan images utilized for training the quality inspection model. (a) High-detail image. (b) Low-detail image.
Sensors 24 05398 g005
Figure 6. A-scan profile generated for the OCT B-scan image. The peaks denote the epidermal layer and parenchyma layer. The distance between the two peaks (∆t) was used as the factor to determine the class label according to Wijesinghe et al. [3].
Figure 6. A-scan profile generated for the OCT B-scan image. The peaks denote the epidermal layer and parenchyma layer. The distance between the two peaks (∆t) was used as the factor to determine the class label according to Wijesinghe et al. [3].
Sensors 24 05398 g006
Figure 7. The process of training two deep learning (DL) models using transfer learning.
Figure 7. The process of training two deep learning (DL) models using transfer learning.
Sensors 24 05398 g007
Figure 8. Accuracy and loss of training and validation of the quality inspection model. (a) Variation of accuracy with epochs. (b) Variation of loss with epochs.
Figure 8. Accuracy and loss of training and validation of the quality inspection model. (a) Variation of accuracy with epochs. (b) Variation of loss with epochs.
Sensors 24 05398 g008
Figure 9. Overall accuracy (OA) with precision and recall for CLS detection model. (a) Precision of each class for adapted pretrained models. (b) Recall of each class for adapted pretrained models.
Figure 9. Overall accuracy (OA) with precision and recall for CLS detection model. (a) Precision of each class for adapted pretrained models. (b) Recall of each class for adapted pretrained models.
Sensors 24 05398 g009
Figure 10. AUC–ROC calculated for predictions given by the analyzed deep learning (DL) models developed for classification of the stages of CLS disease.
Figure 10. AUC–ROC calculated for predictions given by the analyzed deep learning (DL) models developed for classification of the stages of CLS disease.
Sensors 24 05398 g010
Figure 11. Confusion matrix of best deep learning (DL) model trained for circular leaf spot (CLS) stage identification.
Figure 11. Confusion matrix of best deep learning (DL) model trained for circular leaf spot (CLS) stage identification.
Sensors 24 05398 g011
Figure 12. B-scan images and respective A-scan profiles of scanned leaf samples. (a) B-scan image of a sample labeled as H. (b) A-scan profile of the B-scan labeled as H because the peak difference (∆t1) is 141 µm and labeled as H by LAMP analysis. (c) B-scan image of a sample labeled as HI. (d) A-scan profile of the B-scan is labeled as HI because the peak difference (∆t2) is 126 µm and labeled as I by LAMP analysis.
Figure 12. B-scan images and respective A-scan profiles of scanned leaf samples. (a) B-scan image of a sample labeled as H. (b) A-scan profile of the B-scan labeled as H because the peak difference (∆t1) is 141 µm and labeled as H by LAMP analysis. (c) B-scan image of a sample labeled as HI. (d) A-scan profile of the B-scan is labeled as HI because the peak difference (∆t2) is 126 µm and labeled as I by LAMP analysis.
Sensors 24 05398 g012
Table 1. Difference between peaks of A-scan profiles for labeling the CLS stages.
Table 1. Difference between peaks of A-scan profiles for labeling the CLS stages.
Class LabelRanges in µm (∆t)
H-healthy140+
HI-apparently healthy90–130
I-infected0–89
Table 2. Labeling procedure of H, HI, and I based on LAMP and A-scan analysis.
Table 2. Labeling procedure of H, HI, and I based on LAMP and A-scan analysis.
A-Scan LabelLAMP LabelClass Label
HHH
HIIHI
IHH
HIHI
HIHH
III
Table 3. Testing results of quality inspection models trained over multiple pre-trained models.
Table 3. Testing results of quality inspection models trained over multiple pre-trained models.
Base ModelEpochsLossAccuracy
VGG16200.06080.9899
InceptionResNetV2200.10880.9828
InceptionV3500.10220.9885
Table 4. Testing results of CLS classification models trained over multiple pre-trained models.
Table 4. Testing results of CLS classification models trained over multiple pre-trained models.
Base ModelEpochsLearning RateOLOAAUC–ROCPrecisionRecall
H vs. OHI vs. OI vs. OMicro
Averaged
HHIIHHII
DenseNet1211100.0001250.32490.85730.840.840.910.890.78230.90050.70270.75730.89530.8387
VGG161140.00010.3690.84870.840.830.860.890.75530.89150.74720.75730.89240.7311
VGG191110.00010.38610.8260.80.790.780.870.7190.85960.86880.6960.89630.5698
InceptionV3820.00010.43910.810.830.80.810.860.67270.88750.64830.78930.8330.6344
InceptionResNetV21090.00010.43170.80730.810.750.550.860.72980.830310.69860.91080.0967
ResNet101680.000010.58150.74530.690.650.50.810.65780.764100.46660.91370
ResNet50860.000010.58160.73470.610.590.50.80.79820.729400.24260.97960
EfficientNetB01220.00010.77630.6880.50.50.50.7700.6880010
Xception860.0001250.35140.85270.850.830.850.890.77950.89670.67320.77330.89240.7311
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kalupahana, D.; Kahatapitiya, N.S.; Silva, B.N.; Kim, J.; Jeon, M.; Wijenayake, U.; Wijesinghe, R.E. Dense Convolutional Neural Network-Based Deep Learning Pipeline for Pre-Identification of Circular Leaf Spot Disease of Diospyros kaki Leaves Using Optical Coherence Tomography. Sensors 2024, 24, 5398. https://doi.org/10.3390/s24165398

AMA Style

Kalupahana D, Kahatapitiya NS, Silva BN, Kim J, Jeon M, Wijenayake U, Wijesinghe RE. Dense Convolutional Neural Network-Based Deep Learning Pipeline for Pre-Identification of Circular Leaf Spot Disease of Diospyros kaki Leaves Using Optical Coherence Tomography. Sensors. 2024; 24(16):5398. https://doi.org/10.3390/s24165398

Chicago/Turabian Style

Kalupahana, Deshan, Nipun Shantha Kahatapitiya, Bhagya Nathali Silva, Jeehyun Kim, Mansik Jeon, Udaya Wijenayake, and Ruchire Eranga Wijesinghe. 2024. "Dense Convolutional Neural Network-Based Deep Learning Pipeline for Pre-Identification of Circular Leaf Spot Disease of Diospyros kaki Leaves Using Optical Coherence Tomography" Sensors 24, no. 16: 5398. https://doi.org/10.3390/s24165398

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop