Next Article in Journal
Strategic Entrepreneurship Mindset, Strategic Entrepreneurship Leadership, and Entrepreneurial Value Creation of SMEs in East Java, Indonesia: A Strategic Entrepreneurship Perspective
Previous Article in Journal
Response of Antioxidant Enzyme Activities of the Green Microalga Chlorococcum sp. AZHB to Cu2+ and Cd2+ Stress
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On Using Deep Artificial Intelligence to Automatically Detect Apple Diseases from Leaf Images

1
Department of Computer Engineering, Jordan University of Science and Technology, Ar-Ramtha 3030, Jordan
2
Department of Software Engineering, Jordan University of Science and Technology, Ar-Ramtha 3030, Jordan
*
Author to whom correspondence should be addressed.
Sustainability 2022, 14(16), 10322; https://doi.org/10.3390/su141610322
Submission received: 13 July 2022 / Revised: 6 August 2022 / Accepted: 14 August 2022 / Published: 19 August 2022

Abstract

:
Plant diseases, if misidentified or ignored, can drastically reduce production levels and harvest quality. Technology in the form of artificial intelligence applications has the potential to facilitate and improve the disease identification process, which in turn will empower prompt control. More specifically, the work in this paper addressed the identification of three common apple leaf diseases—rust, scab, and black rot. Twelve deep transfer learning artificial intelligence models were customized, trained, and tested with the goal of categorizing leaf images into one of the aforementioned three diseases or a healthy state. A dataset of 3171 leaf images (621 black rot, 275 rust, 630 scab, and 1645 healthy) was used. Extensive performance evaluation revealed the excellent ability of the transfer learning models to achieve high values (i.e., >99%) for F1 score, precision, recall, specificity, and accuracy. Hence, it is possible to design smartphone applications that enable farmers with poor knowledge or limited access to professional care to easily identify suspected infected plants.

1. Introduction

Global circumstances in terms of climate change (e.g., drought, extreme weather conditions, wildfires, and unseasonable temperatures), ramifications of political unrest (e.g., reduction of fertilizer exports), and supply chain disruptions are straining food production. It is estimated that by the year 2050, the demand for food will increase by 70% as the global population reaches nearly 9.1 billion [1]. These conditions necessitate the optimal usage and conservation of existing resources. To this end, protecting plants from diseases is of paramount importance. It is one of the major factors that help reduce waste in agricultural production [2]. Although popular and easy, traditional methods of detecting diseases by naked eye observations are insufficient to determine the disease [3]. Generally, it takes time, effort, experience, and knowledge of plant diseases to diagnose these ailments. In addition, a large number of workers and material costs may be required in large farms [4,5].
Expert identification of plant diseases goes through several methods that relate to the shape of the leaf, the trunk of the tree, and the shape of the fruit [6]. In addition, possible diseases could manifest themselves through changes in the color of the leaf or fruit, and/or the appearance of spots [7]. Thus, leaf images typically carry significant clues into the sickness state of the tree. Hence, it is possible to use these images to determine the type of disease and treat the plant before the onset of irreversible damage or yield loss [8]. Artificial intelligence (AI) has come to play an increasingly diverse role in the development of smart practical applications of great benefit. More specifically, recent advances in deep learning have empowered complex image classification and analysis models [9,10]. Convolutional neural networks (CNNs) are a subclass of deep learning algorithms that is suitable to discover features in images without the need for explicit feature extraction or complex image processing algorithms [11].
Apple is one of the most productive fruits in the world [12], but it is greatly affected by fungi that appear clearly on its leaves. Several diseases affect apples, reduce their yield, or destroy entire fields [13]. Some of these include: black rot, apple scab, and rice apple rust [14]. Cedar-Apple rust is one of the common apple diseases. It is transmitted to apples from the surrounding cedar trees, and it can weaken the apple tree and reduce the production of fruits over time [15]. Hence, a large-scale infection can lead to adverse effects. One of the most used treatment methods is to cut the affected branch or leaves or to remove cedar trees that are within a radius of 1 to 3 miles. Apple scab disease is a seasonal infection that affects apples during the later stages of winter and the beginning of spring. It is caused by a fungus, which makes apparent changes to the fruits and leaves. In this case, crop loss may reach 70% [16]. Experts use pesticides or biological control treatments to stop its spread. Thus, early detection and control are highly beneficial. Black rot disease is another fungal infection that affects the stems, leaves, and fruits [17]. It weakens the infected tree and can spread to other trees, which causes a significant crop loss.
Several AI-based studies were conducted to detect and classify apple diseases, see Table 1. Some of these studies utilized fruit images while others rely upon leaf images as input. Dubey and Jalal [18] combined image processing and AI techniques to classify fruit images into four possible classes (i.e., blotch, rot, scab, and healthy) with a 93% accuracy. K-means clustering was used for image segmentation and four features (i.e., global color histogram, color coherence vector, local binary pattern, and complete local binary pattern) were extracted from the images after segmentation. The classification was performed using multiclass support vector machine (SVM). In a later study by the same authors [18], Zernike moments were combined with the set of previous features, which resulted in a 95.94% accuracy. Li et al. [19] developed a model based on a CNN with the purpose of identifying the quality of apples (i.e., premium, middle, or poor) using fruit images as input. They reported a 95.33% accuracy.
Geetharamani and Pandian [20] proposed a nine-layer CNN to classify leaf diseases into one of 38 disease categories that can infect 13 plant types. An accuracy of 96.46% was achieved. However, image augmentation was used in a fashion that may have led to data leaking as the total number of images was artificially increased. A more concrete approach would have discarded the originals and kept the augmented images. Fang et al. [13] developed an improved CNN model based on VGG16. Using leaf images of healthy apples as well as seven diseases (i.e., leaf blight, cedar rust, gray spot, leaf rust, black rot, and black spot), they achieved an accuracy range of 95.0% to 99.7%. However, their approach is based mainly on tuning model hyperparameters rather than major design modifications. Yu et al. [21] proposed a ResNet50-based model [22], which they called multiple-step optimization ResNet (MSO-ResNet). They reported an average precision, recall, and F1 score of 95.7%,95.8%, and 95.7%, respectively, in identifying five leaf disease states and one healthy. Chakraborty et al. [23] differentiated between infected (black rot and cedar apple rust) and non-infected apple leaves. Using leaf images as input, they reduced the noise using a Gaussian filter, and segmented infected regions of the leaves using histogram equalization and the Otsu thresholding algorithm. Ten features were extracted from the segmented images, which fed a multiclass support vector machine (SVM) classifier. Similarly, Alqethami et al. [1] proposed a new methodology with 3 prediction models (SVM, K-nearest neighbors (KNN) and CNN), The researchers categorized the leaves into four labels according to the type of disease as Black Rot, Cedar Apple Rust, Apple scab and Healthy, using features that pertain to color, shape, and texture.
In this work, the application of deep transfer learning in identifying apple diseases was investigated. Such an approach renders explicit feature extraction and complex image processing unnecessary. In addition, the robustness and power of well-known deep CNN models were harnessed toward the specific goal. The contributions of this paper are as follows:
  • Implement a deep transfer learning-based diagnosis system using leaf images for three apple diseases—scab, black rot, and rust. In addition, a fourth, healthy class was part of the classification;
  • Implement transfer learning of 12 deep convolutional neural networks models for the classification of leaf images into four classes. Such a system can aid farmers in quick disease identification and control;
  • Compare the performance using multiple indices that reflect the true behavior of the algorithms. In addition, the training behavior and time requirements were also included.
The remainder of this paper is organized as follows: the dataset, deep learning models, and performance evaluation metrics and setup are explained in detail in Section 2, Section 3 presents the performance evaluation results along with comparison to the related literature and discussion of the models, and we conclude in Section 4.

2. Materials and Methods

Figure 1 shows an architecture diagram of the diagnosis system. The use of CNNs renders any feature extraction or segmentation (i.e., separation of relevant image parts) needless. This is because such steps are performed inherently by the intricacies of the models’ layers and mathematical operations. The next few subsections discuss each part of the system in detail.

2.1. Dataset

The dataset used in this study consists of 3171 images divided into four categories: 621 black rot, 275 rust, 630 scab, and 1645 healthy [20,25]. Each image is a photo of an individual leaf from one of the four classes, which was taken under unified background and light conditions. The images are publicly available in JPEG format and a resolution of 256 × 256 pixels. They were later resized to match the requirements of the specific deep learning model. No cropping or other preprocessing operations were conducted on the images. Samples of the images of the three diseases and healthy leaves are shown in Figure 2.

2.2. Deep Transfer Learning

Transfer learning, as the name suggests, allows the customization of a generically trained model to be used for a specific application. Such an approach has the advantage of utilizing well-established and well-designed models without repeating the model-building efforts [26]. Lower-numbered layers (i.e., closer to the input) typically learn common features (e.g., colors). On the other hand, layers closer to the output need to be refitted and retrained to serve the targeted output based on the application. This approach has been successfully used for many deep learning applications in the literature [9,10].
In this paper, 12 pre-trained convolutional neural network models were used separately to classify apple leaf images. These models were: DarkNet-53 (53 layers) [27], DenseNet-201 (201 layers) [28], EfficientNet-b0 (237 layers) [29], GoogLeNet (22 layers) [30], Inceptionv3 (48 layers) [31], Inception-ResNet (164 layers), MobileNetv2 (53 layers), ResNet-18 (18 layers), ResNet-101 (101 layers) [22], ShuffleNet (50 layers) [32], SqueezeNet (18 layers) [33], and Xception (71 layers) [34]. The models have common features in the type of operations being conducted on the input (e.g., convolutions and pooling). However, they differ in their depth (i.e., number of layers), width (i.e., input size and intermediate fields), and network structure. Moreover, their computational requirements may be different based on the type of innovation introduced in the model to reduce the number of required mathematical operations (e.g., residual networks). All of the models were used based on the same hyperparameters (e.g., number of epochs) and were pre-trained using the ImageNet database of one million images [35].
The CNN design literature has witnessed great efforts in improving the performance and efficiency of these networks. The modifications range from optimizing certain parameters to reforming the network architecture with new processing elements, connection types, block designs, and depths [36]. GoogLeNet is a type of small CNN that considers the spatial locality of image pixels at various granularity levels. Moreover, it was the first to introduce the inception block concept. ShuffleNet reduced the number of required operations and improved the efficiency of GoogLeNet by introducing the shuffle process. The ResNet, Incpetionv3, and Inception-ResNet models are depth-based CNNs, which rely on a larger number of network layers in an effort to efficiently capture the input features and characteristics. The ResNet family of models introduced residual learning and used bypass pathways (i.e., skip connections) in a data-independent and parameter-free manner. The Inceptionv3 and Inception-ResNet models used small asymmetric filers to improve the computational cost. The DenseNet-201 is a multipath design that aims to solve the vanishing gradient problem. This was accomplished using dense connectivity (i.e., feed-forward cross-layer connections). The Xception model is a width-based multi connection CNN that emphasized the importance of the width of the architecture through parallel processing within the network layers. SqueezeNet uses feature maps and suppresses the less important features using the SE-block, which performs two operations: squeeze (i.e., suppress) and excitation. The EfficientNet-b0 is the baseline EfficientNet model with further versions reaching level b7. The MobileNetv2 model, as the name suggests, was intended for resource-limited mobile devices. This was accomplished by reducing the number of trainable parameters using the depth-wise separable convolution operations.

2.3. Experimental Setup

The training parameters were unified for all models and set as follows: The minimum batch size, which is the number of input instances to be processed at the same time, was set to 16. This allows faster training, but at the expense of more memory requirements. The maximum number of epochs was set to 10, which was enough for the learning process to reach maximum accuracy and the loss curve to flatten without unnecessary extra training. The learning rate was set to 3 × 10 4 . The solver optimization algorithm for network training was the stochastic gradient descent with momentum (SGDM), which is commonly used for training due to its fast convergence [37].
In order to check the models’ ability to learn with more data and their overfitting/underfitting behavior, the data were split for training/testing using two methods 50/50 and 70/30. After that, input images were augmented to improve the training process by introducing variety to the dataset [38]. This was performed by randomly shifting the images along the x-axis and y-axis using the range [−30, 30], and random scaling in the range [0.9, 1.1]. Moreover, the images were randomly reflected along the x-axis.
The models were implemented and evaluated using MATLAB R2021a software running on an HP OMEN 30L desktop GT13 with 64 GB RAM, NVIDIA® GeForce RTXTM 3080 GPU, Intel® CoreTM i7-10700K CPU @ 3.80 GHz, and 1 TB SSD.

2.4. Performance Evaluation

Several metrics were used to evaluate the performance of the models, see Equations (1)–(5), where T P is the true positive, F N is the false negative, F P is the false positive, and T N is the true negative. The ability to identify a diseased leaf image in the positive correct class is measured by the true positive rate (i.e., R e c a l l or sensitivity). Moreover, the true negative rate (i.e., S p e c i f i c i t y ) measures the number of correctly classified normal leaf images as such. A highly sensitive model will recognize a large proportion of positive cases, but this may have been accomplished at the cost of a large number of false positives. P r e c i s i o n provides more insight into the results by measuring the percentage of true positive images (i.e., image of leaves with a disease) as a percentage of all reported positive images, including false ones. A c c u r a c y measures the ratio of correctly identified positive and negative cases to the total number of images. In the case of imbalanced datasets with big differences in the number of images per class, the F1 score provides a good indicator of accuracy [39]. In addition, the training and validation time was reported.
A c c u r a c y = T P + T N P + N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
S p e c i f i c i t y = T N T N + F P
F 1 = 2 × T P 2 × T P + F P + F N
where T P : The number of correctly classified images. F P : The number of wrongly classified images. F N : The number of images missed by the classifier. T N : The number of correctly classified images as the negative class. P: The number of all images considered as the positive class. N: The number of all images other than the positive class.

3. Results and Discussion

The experiments aimed at evaluating the effectiveness of the CNN models in identifying the correct health category of apple leaves. The performance evaluation metrics were recorded for all models along with the required training and validation times.
Starting with a data split of 50/50, the mean overall F1 score, precision, recall, specificity, and accuracy for each deep learning model are shown in Table 2. All models achieved exceptionally high F1 score in addition to the other metrics. The Incpetionv3 achieved the highest value (F1 score = 99.9%). The lowest-performing model was InceptinResNet-v2 with an F1 score of 98.1% and a recall value of 97.4%. However, such performance is extremely excellent. This performance is confirmed and further details are revealed by the confusion matrices in Figure 3. They show that in most cases, different classes were classified perfectly. However, some images from the scab class caused the most error by being misclassified as healthy in several models.
Although the reported performance is very high, deep learning models require a large number of images in order to realize their full potential. The ramifications of an increased size of the training dataset were evaluated by setting the training/validation data split as 70/30. The mean overall F1 score, precision, recall, specificity, and accuracy for all models and 70/30 data split are shown in Table 3. Nearly all models achieved near-perfect specificity (i.e., true negative rate). Moreover, most models improved on their smaller training dataset performance (i.e., 50/50 split) with Inceptionv3 and MobileNetv2 achieving perfect results (i.e., all images were correctly classified). However, the Incpetion-ResNet-v2 model seems to have suffered from overfitting of the training data and actually produced a slightly worse performance compared with that of the 50/50 data split. The corresponding confusion matrices are shown in Figure 4. The leaf images showing the scab disease caused the most errors in all non-perfect models.
Machine learning models are typically susceptible to overfitting or underfitting. Further insight into the training and validation behavior of the deep learning models is necessary to expose any deficiencies. A sample of the training and validation progress and the cross entropy loss during training for the Darknet53 model using a 50/50 data split is shown in Figure 5. The figure shows a stable learning process and a steady loss decline for both training and validation. Moreover, Figure 6 and Figure 7 show the training and validation progress curve using 70/30 data split for the SqueezeNet and MobileNetv2 models, respectively. Similarly, both figures show no overfitting or underfitting visible. In addition, the learning process seems stable learning process with decreasing loss.
The mean training and validation times for all models and data split strategies are shown in Table 4. Two observations can be drawn from the table. First, although the classification performance of the models is comparably close, there is a big speed gap between the fastest and slowest models (i.e., 118.9 s vs. 2249.4 s using 50/50 split, and 130.6 s vs. 4240 s using 70/30 split). This makes the faster models more favorable for smartphone deployment. Second, some models scaled better than others with more training data (e.g., SqueezeNet vs. ShuffleNet).
The application of artificial intelligence in the agricultural sector is receiving increasing attention in the literature. Table 5 shows a comparison of the latest results in classification of apple diseases. Li et al. [19] and the two studies by Dubey and Jalal [18,24] relied upon fruit images as input. Such an approach hinders large-scale automated measurements as it may require 3D images from all sides of the fruit to truly capture the disease state of the fruit. A single 2D image may capture one side of the fruit but miss the infected side. Such a problem is not present when dealing with leaf images. The other studies included in the table use leaf images as input. Chakraborty et al. [23] used image segmentation to extract features of the disease, which were fed to a support vector machine classifier to achieve a 96% F1 score. A similar approach was used by Alqethami et al. [1], but failed to achieve better performance. The other studies either designed their own CNN or modified existing CNN models without transfer learning. Yu et al. [21] modified a res50 model to achieve an F1 score of 95.7%. Whereas, Fang et al. [13] fine-tuned a VGG16 model to achieve an accuracy range of 95–97%. Geetharamani and Pandian [20] built their own nine-layer model and achieved an accuracy of 96.46%. The work in this paper is different in that it does not rely upon explicit image processing algorithms (e.g., segmentation), nor on the quality of a proposed feature. In addition, new CNN designs may need extensive evaluation to establish their quality. Moreover, some of their results used augmented data that replicated, not replaced, the original data, which may skew their results. Using deep transfer learning has the credibility of established models while avoiding any overhead in terms of image processing and feature extraction. Moreover, superior performance in comparison to the literature was achieved over all metrics.
The present study has some limitations. First, the dataset may need to include other possible apple diseases. Moreover, other plant diseases may be combined in the dataset. However, we do not see great value in terms of field deployment, by complicating the classification problem. For example, although it is necessary to include other plant diseases (e.g., corn rust), jumbling all the diseases in one model will hurt the classification performance and speed of the model. A user (i.e., farmer) should be able to choose the correct plant type when checking for a disease. Second, more diverse leaf images in terms of the picture background may be desirable to simulate actual deployment scenarios. Third, it is highly worthwhile to embed the model and other future models in smartphone applications that can be made available on smartphone application stores. This will allow the system to grow, improve, and identify any shortcomings.

4. Conclusions

Climate change due to global warming from green gasses and the resulting conditions (e.g., drought, wildfire, and extreme weather conditions, and unseasonable temperatures) necessitates optimal use of available agricultural resources and the implementation of conservation strategies. Moreover, geopolitical unrest and its ramifications on food prices, availability of fertilizers, and supply chain problems exacerbate the shortages in available food. Thus, protecting plants from disease is essential to mitigate existing circumstances and reduce waste in agricultural production. To this end, the work in this paper investigated the application of deep learning artificial intelligence in the identification of apple diseases from leaf images.
Recent advances in deep convolutional neural networks and transfer learning were applied to the problem of apple disease identification. Such an approach has the potential to achieve high classification accuracy with little overhead to the specialists and great benefit to the farmers. Moreover, it does not require specialist knowledge, elaborate preprocessing of images, nor does it incur errors from image processing techniques. Future work will focus on improving the system by applying incremental learning approaches to evolve the application during deployment. Moreover, we will consider 3D images for disease identification from fruit images. In addition, the development of ensemble deep learning models for plant disease classification is another avenue for research. Regarding deployment, the extra mile of wrapping the deep learning models in smartphone applications should be a worthwhile effort.

Author Contributions

Conceptualization, M.F.; methodology, M.F. and E.F.; software, M.F. and E.F.; validation, M.F. and E.F.; formal analysis, M.F.; investigation, M.F., E.F. and N.K.; resources, M.F., and N.K.; data curation, M.F. and E.F.; writing—original draft preparation, M.F. and E.F.; writing—review and editing, M.F. and N.K.; supervision, M.F. and N.K.; project administration, M.F.; funding acquisition, M.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by the Deanship of Scientific Research at Jordan University of Science and Technology, Jordan, grants No. 20210047.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIartificial intelligence
CNNConvolutional neural networks
SVMSupport vector machine
KNNK-nearest neighbors
T P True positive
T N True negative
F N False negative
F P False positive
MSO-ResNetmultiple-step optimization ResNet
NNegatives
PPositives
SGDMStochastic gradient descent with momentum

References

  1. Alqethami, S.; Almtanni, B.; Alzhrani, W.; Alghamdi, M. Disease Detection in Apple Leaves Using Image Processing Techniques. Eng. Technol. Appl. Sci. Res. 2022, 12, 8335–8341. [Google Scholar] [CrossRef]
  2. Singh, V.; Varsha; Misra, A.K. Detection of unhealthy region of plant leaves using image processing and genetic algorithm. In Proceedings of the 2015 International Conference on Advances in Computer Engineering and Applications, Ghaziabad, India, 19–20 March 2015; IEEE: Piscataway, NJ, USA, 2015. [Google Scholar] [CrossRef]
  3. Sharath, D.M.; Kumar, S.A.; Rohan, M.G.; Prathap, C. Image based Plant Disease Detection in Pomegranate Plant for Bacterial Blight. In Proceedings of the 2019 International Conference on Communication and Signal Processing (ICCSP), Melmaruvathur, India, 4–6 April 2019; IEEE: Piscataway, NJ, USA, 2019. [Google Scholar] [CrossRef]
  4. Sivasangari, A.; Kishore, M.S.; Poornesh, M.; Gomathi, R.M.; Deepa, D. Plant Disease Detection and Classification Using Image Processing and Neural Networks. J. Comput. Theor. Nanosci. 2020, 17, 4920–4924. [Google Scholar] [CrossRef]
  5. Mojjada, R.K.; Kumar, K.K.; Yadav, A.; Prasad, B.S.V. Detection of plant leaf disease using digital image processing. Mater. Today Proc. 2020. [Google Scholar] [CrossRef]
  6. Petrellis, N. A smart phone image processing application for plant disease diagnosis. In Proceedings of the 2017 6th International Conference on Modern Circuits and Systems Technologies (MOCAST), Thessaloniki, Greece, 4–6 May 2017; IEEE: Piscataway, NJ, USA, 2017. [Google Scholar] [CrossRef]
  7. Samajpati, B.J.; Degadwala, S.D. Hybrid approach for apple fruit diseases detection and classification using random forest classifier. In Proceedings of the 2016 International Conference on Communication and Signal Processing (ICCSP), Melmaruvathur, India, 6–8 April 2016; IEEE: Piscataway, NJ, USA, 2016. [Google Scholar] [CrossRef]
  8. Sakhamuri, S.; Kompalli, V.S. An Overview on Prediction of Plant Leaves Disease using Image Processing Techniques. IOP Conf. Ser. Mater. Sci. Eng. 2020, 981, 022024. [Google Scholar] [CrossRef]
  9. Fraiwan, M.; Audat, Z.; Fraiwan, L.; Manasreh, T. Using deep transfer learning to detect scoliosis and spondylolisthesis from X-ray images. PLoS ONE 2022, 17, e0267851. [Google Scholar] [CrossRef]
  10. Khasawneh, N.; Fraiwan, M.; Fraiwan, L.; Khassawneh, B.; Ibnian, A. Detection of COVID-19 from Chest X-ray Images Using Deep Convolutional Neural Networks. Sensors 2021, 21, 5940. [Google Scholar] [CrossRef] [PubMed]
  11. Sharma, N.; Jain, V.; Mishra, A. An Analysis of Convolutional Neural Networks for Image Classification. Procedia Comput. Sci. 2018, 132, 377–384. [Google Scholar] [CrossRef]
  12. Jiang, P.; Chen, Y.; Liu, B.; He, D.; Liang, C. Real-Time Detection of Apple Leaf Diseases Using Deep Learning Approach Based on Improved Convolutional Neural Networks. IEEE Access 2019, 7, 59069–59080. [Google Scholar] [CrossRef]
  13. Fang, T.; Chen, P.; Zhang, J.; Wang, B. Identification of Apple Leaf Diseases Based on Convolutional Neural Network. In Intelligent Computing Theories and Application; Springer International Publishing: Cham, Switzerland, 2019; pp. 553–564. [Google Scholar] [CrossRef]
  14. Kumar, M.; Pal, Y.; Gangadharan, S.M.P.; Chakraborty, K.; Yadav, C.S.; Kumar, H.; Tiwari, B. Apple Sweetness Measurement and Fruit Disease Prediction Using Image Processing Techniques Based on Human-Computer Interaction for Industry 4.0. Wirel. Commun. Mob. Comput. 2022, 2022, 5760595. [Google Scholar] [CrossRef]
  15. Bowers, S. Cedar Apple Rust. Available online: https://hgic.clemson.edu/factsheet/cedar-apple-rust/ (accessed on 26 June 2022).
  16. Ontario Ministry of Agriculture, Food And Rural Affairs. Apple Scab. Available online: http://omafra.gov.on.ca/english/crops/facts/apscab.htm#periods (accessed on 26 June 2022).
  17. Grabowski, M. Black Rot of Apple. Available online: https://extension.umn.edu/plant-diseases/black-rot-apple (accessed on 26 June 2022).
  18. Dubey, S.R.; Jalal, A.S. Detection and Classification of Apple Fruit Diseases Using Complete Local Binary Patterns. In Proceedings of the 2012 Third International Conference on Computer and Communication Technology, Allahabad, India, 23–25 November 2012; IEEE: Piscataway, NJ, USA, 2012. [Google Scholar] [CrossRef]
  19. Li, Y.; Feng, X.; Liu, Y.; Han, X. Apple quality identification and classification by image processing based on convolutional neural networks. Sci. Rep. 2021, 11, 16618. [Google Scholar] [CrossRef] [PubMed]
  20. Geetharamani, G.; Pandian, J.A. Identification of plant leaf diseases using a nine-layer deep convolutional neural network. Comput. Electr. Eng. 2019, 76, 323–338. [Google Scholar] [CrossRef]
  21. Yu, H.; Cheng, X.; Chen, C.; Heidari, A.A.; Liu, J.; Cai, Z.; Chen, H. Apple leaf disease recognition method with improved residual network. Multimed. Tools Appl. 2022, 81, 7759–7782. [Google Scholar] [CrossRef]
  22. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  23. Chakraborty, S.; Paul, S.; Rahat-uz Zaman, M. Prediction of Apple Leaf Diseases Using Multiclass Support Vector Machine. In Proceedings of the 2021 2nd International Conference on Robotics, Electrical and Signal Processing Techniques (ICREST), Dhaka, Bangladesh, 5–7 January 2021; pp. 147–151. [Google Scholar] [CrossRef]
  24. Dubey, S.R.; Jalal, A.S. Apple disease classification using color, texture and shape features from images. Signal Image Video Process. 2015, 10, 819–826. [Google Scholar] [CrossRef]
  25. Pandian, J.A.; Geetharamani, G. Data for: Identification of Plant Leaf Diseases Using a 9-Layer Deep Convolutional Neural Network. Version 1. 2019. Available online: https://data.mendeley.com/datasets/tywbtsjrjv/1 (accessed on 1 August 2022). [CrossRef]
  26. Kim, H.E.; Cosa-Linan, A.; Santhanam, N.; Jannesari, M.; Maros, M.E.; Ganslandt, T. Transfer learning for medical image classification: A literature review. BMC Med. Imaging 2022, 22, 69. [Google Scholar] [CrossRef] [PubMed]
  27. Redmon, J. Darknet: Open Source Neural Networks in C, 2013–2016. Available online: http://pjreddie.com/darknet/ (accessed on 1 August 2022).
  28. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar] [CrossRef]
  29. Tan, M.; Le, Q. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; Volume 97, pp. 6105–6114. [Google Scholar]
  30. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar] [CrossRef]
  31. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A.A. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, Francisco, CA, USA, 4–9 February 2017; pp. 4278–4284. [Google Scholar]
  32. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6848–6856. [Google Scholar] [CrossRef]
  33. Iandola, F.N.; Moskewicz, M.W.; Ashraf, K.; Han, S.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50× fewer parameters and <1 MB model size. arXiv 2016, arXiv:1602.07360. [Google Scholar]
  34. Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1800–1807. [Google Scholar] [CrossRef]
  35. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar] [CrossRef]
  36. Khan, A.; Sohail, A.; Zahoora, U.; Qureshi, A.S. A survey of the recent architectures of deep convolutional neural networks. Artif. Intell. Rev. 2020, 53, 5455–5516. [Google Scholar] [CrossRef]
  37. Qian, N. On the momentum term in gradient descent learning algorithms. Neural Netw. 1999, 12, 145–151. [Google Scholar] [CrossRef]
  38. Shorten, C.; Khoshgoftaar, T.M. A survey on Image Data Augmentation for Deep Learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  39. Tharwat, A. Classification assessment methods. Appl. Comput. Inform. 2020, 17, 168–192. [Google Scholar] [CrossRef]
Figure 1. A graphical representation of the system architecture.
Figure 1. A graphical representation of the system architecture.
Sustainability 14 10322 g001
Figure 2. Sample images from the three disease classes and the healthy one.
Figure 2. Sample images from the three disease classes and the healthy one.
Sustainability 14 10322 g002
Figure 3. The confusion matrices for all algorithms using 50/50 data split.
Figure 3. The confusion matrices for all algorithms using 50/50 data split.
Sustainability 14 10322 g003
Figure 4. The confusion matrices for all algorithms using 70/30 data split.
Figure 4. The confusion matrices for all algorithms using 70/30 data split.
Sustainability 14 10322 g004
Figure 5. Sample training/validation progress curve for DarkNet-53 and 50/50 data split. The training curves are in blue and orange, while the validation curves are in black.
Figure 5. Sample training/validation progress curve for DarkNet-53 and 50/50 data split. The training curves are in blue and orange, while the validation curves are in black.
Sustainability 14 10322 g005
Figure 6. Sample training/validation progress curve for SqueezeNet and 70/30 data split. The training curves are in blue and orange, while the validation curves are in black.
Figure 6. Sample training/validation progress curve for SqueezeNet and 70/30 data split. The training curves are in blue and orange, while the validation curves are in black.
Sustainability 14 10322 g006
Figure 7. Sample training/validation progress curve for MobileNetv2 and 70/30 data split. The training curves are in blue and orange, while the validation curves are in black.
Figure 7. Sample training/validation progress curve for MobileNetv2 and 70/30 data split. The training curves are in blue and orange, while the validation curves are in black.
Sustainability 14 10322 g007
Table 1. A summary of the latest literature on the classification of apple diseases.
Table 1. A summary of the latest literature on the classification of apple diseases.
Study and YearObjectiveDatasetApproach
Yu et al., 2022 [21]Classification of five diseases + one healthy11,397 leaf imagesModified res50 model.
Chakraborty et al., 2021 [23]Classification of infected (two classes) and non-infected apple leaves500 leaf imagesImage segmentation and SVM.
Alqethami et al., 2022 [1]Binary classification (disease vs. healthy)240 leaf imagesSegmentation, feature extraction, and machine learning.
Li et al., 2021 [19]Classification of apple images into premium, middle, and poor grade3600 fruit imagesCustom CNN, Inception-v3, and SVM.
Fang et al., 2019 [13]Seven-way classification of apple diseases7056 leaf imagesFine tuned VGG16.
Geetharamani and Pandian [20]39-way classification (38 diseases + 1 background)54,305 leaf images (61,486 after augmentation)Nine-layer CNN.
Dubey and Jalal, 2016 [24]Four-way classification of apple diseases320 fruit imagesDefect segmentation using K-means, color and shape feature extraction, and classification using SVM.
Dubey and Jalal, 2012 [18]Four-way classification of apple diseases431 fruit imagesSegmentation, feature extraction, and multi-class SVM.
This workFour-way classification of apple diseases3171 leaf imagesDeep transfer learning (12 models).
Table 2. The mean overall F1 score, Precision, Recall, Specificity, and Accuracy for each deep learning model and 50/50 data split.
Table 2. The mean overall F1 score, Precision, Recall, Specificity, and Accuracy for each deep learning model and 50/50 data split.
ModelF1 ScorePrecisionRecallSpecificityAccuracy
SqueezeNet98.2%99.0%97.6%99.0%98.1%
GoogLeNet98.2%98.3%98.1%99.4%98.4%
Inceptionv399.9%99.9%99.9%99.9%99.9%
DenseNet-20199.9%100.0%99.9%100.0%99.9%
MobileNetv299.6%99.7%99.6%99.8%99.6%
Resnet10199.1%99.2%99.0%99.7%99.2%
Resnet1899.4%99.6%99.3%99.7%99.4%
Xception98.8%99.0%98.6%99.6%99.0%
Inception-ResNet-v298.1%98.9%97.4%99.1%98.2%
ShuffleNet99.7%99.8%99.7%99.8%99.7%
DarkNet-5399.8%99.9%99.8%99.9%99.8%
EfficientNet-b099.3%99.5%99.2%99.8%99.5%
Table 3. The mean overall F1 score, Precision, Recall, Specificity, and Accuracy for each deep learning model and 70/30 data split.
Table 3. The mean overall F1 score, Precision, Recall, Specificity, and Accuracy for each deep learning model and 70/30 data split.
ModelF1 ScorePrecisionRecallSpecificityAccuracy
SqueezeNet98.4%98.8%98.1%99.3%98.5%
GoogLeNet98.1%98.8%97.6%99.1%98.1%
Inceptionv3100.0%100.0%100.0%100.0%100.0%
DenseNet-20199.7%99.8%99.6%99.9%99.8%
MobileNetv2100.0%100.0%100.0%100.0%100.0%
Resnet10199.5%99.5%99.5%99.8%99.6%
Resnet1899.8%99.8%99.8%99.9%99.8%
Xception99.4%99.7%99.2%99.7%99.5%
Inception-ResNet-v296.5%97.9%95.3%98.4%96.7%
ShuffleNet99.9%99.9%99.9%99.9%99.9%
DarkNet-5399.7%99.8%99.6%99.8%99.7%
EfficientNet-b099.3%99.6%99.0%99.7%99.4%
Table 4. The mean training and validation times for all algorithms and data split strategies. All times are in seconds.
Table 4. The mean training and validation times for all algorithms and data split strategies. All times are in seconds.
Data Split50/5070/30
Model
SqueezeNet118.9130.6
GoogLeNet260.1293.7
Inceptionv3571.4703.1
DenseNet-2011891.12298.2
MobileNetv28241091.2
Resnet101587866
Resnet18164.6238.12
Xception2249.44240
Inception-ResNet-v21784.93814.0
ShuffleNet637.21657.0
DarkNet-53439.61543.6
EfficientNet-b01347.23846.8
Table 5. A comparison to the latest results in classification of apple diseases from leaf images.
Table 5. A comparison to the latest results in classification of apple diseases from leaf images.
Study and YearPerformance
Yu et al., 2022 [21]Average precision, recall, and F1 score of 95.7%, 95.8%, and 95.7%, respectively.
Chakraborty et al., 2021 [23]Accuracy 96%, Recall 96%, Precision 96%, F1 score 96%.
Alqethami et al., 2022 [1]Accuracy 82.25% (SVM), 70.3% (KNN), and 98.5% (GoogleNet).
Li et al., 2021 [19]Best accuracy 95.33% (custom CNN).
Fang et al., 2019 [13]Accuracy 95–97%.
Geetharamani and Pandian [20]Accuracy 96.46%.
Dubey and Jalal, 2016 [24]Accuracy 95.94%.
Dubey and Jalal, 2012 [18]Accuracy 93%.
This work(>99%) for all measures.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Fraiwan, M.; Faouri, E.; Khasawneh, N. On Using Deep Artificial Intelligence to Automatically Detect Apple Diseases from Leaf Images. Sustainability 2022, 14, 10322. https://doi.org/10.3390/su141610322

AMA Style

Fraiwan M, Faouri E, Khasawneh N. On Using Deep Artificial Intelligence to Automatically Detect Apple Diseases from Leaf Images. Sustainability. 2022; 14(16):10322. https://doi.org/10.3390/su141610322

Chicago/Turabian Style

Fraiwan, Mohammad, Esraa Faouri, and Natheer Khasawneh. 2022. "On Using Deep Artificial Intelligence to Automatically Detect Apple Diseases from Leaf Images" Sustainability 14, no. 16: 10322. https://doi.org/10.3390/su141610322

APA Style

Fraiwan, M., Faouri, E., & Khasawneh, N. (2022). On Using Deep Artificial Intelligence to Automatically Detect Apple Diseases from Leaf Images. Sustainability, 14(16), 10322. https://doi.org/10.3390/su141610322

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop