Next Article in Journal
The Usefulness of a Virtual Environment-Based Patient Setup Training System for Radiation Therapy
Previous Article in Journal
Special Types of Breast Cancer: Clinical Behavior and Radiological Appearance
Previous Article in Special Issue
Efficient Wheat Head Segmentation with Minimal Annotation: A Generative Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimized Crop Disease Identification in Bangladesh: A Deep Learning and SVM Hybrid Model for Rice, Potato, and Corn

1
Department of Computer Science & Engineering, Bangabandhu Sheikh Mujibur Rahman Science and Technology University, Pirojpur 8500, Bangladesh
2
Faculty of Engineering, Multimedia University, Cyberjaya 63100, Malaysia
3
Department of Computer Science & Engineering, University of Liberal Arts Bangladesh, Dhaka 1207, Bangladesh
*
Authors to whom correspondence should be addressed.
J. Imaging 2024, 10(8), 183; https://doi.org/10.3390/jimaging10080183
Submission received: 29 May 2024 / Revised: 26 June 2024 / Accepted: 12 July 2024 / Published: 30 July 2024
(This article belongs to the Special Issue Imaging Applications in Agriculture)

Abstract

:
Agriculture plays a vital role in Bangladesh’s economy. It is essential to ensure the proper growth and health of crops for the development of the agricultural sector. In the context of Bangladesh, crop diseases pose a significant threat to agricultural output and, consequently, food security. This necessitates the timely and precise identification of such diseases to ensure the sustainability of food production. This study focuses on building a hybrid deep learning model for the identification of three specific diseases affecting three major crops: late blight in potatoes, brown spot in rice, and common rust in corn. The proposed model leverages EfficientNetB0′s feature extraction capabilities, known for achieving rapid high learning rates, coupled with the classification proficiency of SVMs, a well-established machine learning algorithm. This unified approach streamlines data processing and feature extraction, potentially improving model generalizability across diverse crops and diseases. It also aims to address the challenges of computational efficiency and accuracy that are often encountered in precision agriculture applications. The proposed hybrid model achieved 97.29% accuracy. A comparative analysis with other models, CNN, VGG16, ResNet50, Xception, Mobilenet V2, Autoencoders, Inception v3, and EfficientNetB0 each achieving an accuracy of 86.57%, 83.29%, 68.79%, 94.07%, 90.71%, 87.90%, 94.14%, and 96.14% respectively, demonstrated the superior performance of our proposed model.

1. Introduction

Agriculture has been a way of life since the beginning of human civilization. It progressed from traditional practices to modern technologies. However, over time, pests and crop diseases have evolved as well and it is crucial to detect crop disease manifestations in the crops as early as possible due to the rise of global food demand. Diseases impact both the quantity and serve as its main host; they can also affect various other plants, such as tomatoes and petunias, as well as the quality of crop production, food security, economic stability, and livelihoods. Late blight is caused by a fungus-like pathogen known as oomycete. While potatoes and hairy nightshades are susceptible to late blight, it is crucial to differentiate between the two for accurate disease management [1]. The initial symptoms are little, light to dark green, round to irregularly shaped, water-soaked patches on the leaves, and these lesions usually appear on the lower leaves. These lesions grow quickly into huge, dark brown or black lesions during cool and moist weather. The first epidemic of late blight happened over 150 years ago in Europe [2]. It is now seen as a common example of an obligate parasite that can affect a large number of crops. New infections occur and entire leaves can become blighted and killed within just a few days. Late blight can spread from these local sources to potato fields and there are high chances of crops being affected by late blight particularly in regions where certain areas remain damp for extended periods. Brown spot has been largely recognized as one of the most common and damaging rice diseases. The disease causes 5% yield loss across all lowland rice production in South and Southeast Asia on an average basis, and severely infected fields can have yield losses as high as 45% (IRRI, 2020). Infected parts have small, circular, yellow-brown, or brown lesions, and fully developed lesions are circular to oval with a light brown to gray center. It can grow well in relatively high humidity. Puccinia sorghi is the fungus that causes common rust, which happens every growing season. Dead tissue lesions may appear as the leaf tissue surrounding the pustules becomes yellow or dies. When a leaf is highly affected, the lesions can occasionally form a band across the leaf. Common rust can spread rapidly in warm and humid conditions [3].
The authors in [3] propose a model that combines convolutional neural networks (CNNs) and support vector machines (SVMs) for multi-classification based on the use of motor imagery techniques. Batch normalization and dropout layers were incorporated into CNN to improve overfitting. After carrying out a comparative study on performance matrices, the average cross-validation accuracy rates for CNN and CNN-SVM were 82% and 84.1%, respectively. The experimental results indicate that the CNN-SVM model performs better than CNN and many existing classification algorithms. In 2021, Bansal et al. [4] proposed an ensemble model of pre-trained DenseNet121, EfficientNetB7, and EfficientNetB0. It classified the leaves of apple trees into categories such as healthy, apple scab, apple cedar rust, and multiple diseases using its images. The authors included various image augmentation techniques to increase the dataset size and enhance the model’s accuracy. The proposed model achieved an accuracy of 96.25% on the validation dataset. A study conducted by Afzaal et al. [5] evaluated the potential of utilizing deep learning (DL) combined with machine vision to detect early blight disease in potato production systems. Three types of convolutional neural networks (CNNs)—GoogleNet, VGGNet, and EfficientNetB0—were trained using the PyTorch framework to analyze disease images at different growth stages. The images were categorized into three dimensions (2-class, 4-class, and 6-class). The outcome of this study revealed that EfficientNetB0 performed significantly better than other CNNs, achieving an FScore range of 0.79 to 0.94. In this study, traditional machine learning (ML) and CNN-based transfer learning approaches were presented for cucumber disease recognition. The performance of both techniques was compared to find the most appropriate one. Then, various ML algorithms were compared using k-means-based image segmentation after extracting 10 relevant features. Random forest achieved the better accuracy with 89.93%. In the case of CNN-based transfer learning, a comparison among various InceptionV3, MobileNetV2, and VGG16 was carried out, and MobileNetV2 achieved the highest accuracy with 93.23%. There were 4200 images after data augmentation, and this dataset was divided into train, validation, and test sets. A total of 20% data was kept for testing purposes; among the rest of the 80% data, 20% data was used for the validation of the models, and the remaining data was used for the training of the models [6]. In another study by F. T. Pinki et al. [7], an automated system was designed to diagnose three common diseases affecting paddy leaves: brown spot, leaf blast, and bacterial blight. The system provides recommendations for pesticides or fertilizers based on the severity of diseases. K-means clustering is applied to identify the affected areas in a paddy leaf image. Visual features such as color, texture, and shape are used as features for classifying these diseases using a support vector machine (SVM) classifier. After recognizing the disease, the system suggests a remedy. This study addresses the issue of sugarcane diseases, which can lead to significant financial losses for small scale farmers. The deep learning model used in this study was trained and tested on a dataset of 13,842 sugarcane images containing both healthy and disease-infected leaves, and it achieved an impressive accuracy of 95%. The authors, Militante et al. [8], employed CNN with seven different classes to classify the sugarcane leaf. Jadhav et al. [9] used pre-trained CNN models to identify diseases in soybean plants using a transfer learning approach. CNN models such as AlexNet and GoogleNet attained better outcomes, but the model lacked in the diversity classification. Rather than creating a model to classify different plant diseases, many of the existing models focus on identifying a single class of a plant disease. This is mainly due to the limited datasets to train deep-learning models for a diverse range of plants. The classification of plant diseases using the ensemble classifier was proposed in another study by Astani et al. [10]. The best ensemble classifier was evaluated with two different datasets. The study conducted by Kolluri et al. [11] uses multimodal learning on colored datasets and grayscale datasets. The models performed better on colored datasets rather than grayscale ones. In this research by Eunice et al. [12], (CNN)-based pre-trained models were utilized for efficient plant disease identification. The researchers focused on fine-tuning the hyperparameters of popular pre-trained models, such as DenseNet-121, ResNet-50, VGG-16, and Inception V4. DenseNet-121 outperformed the other models. Table 1 illustrates the utilization of various machine learning models as well as contemporary deep learning models for image classification by researchers. Among deep learning models, VGGNet and EfficientNet appear to be the most used ones in the research reviewed by the authors.
Despite significant advancements in crop disease identification using deep learning and machine learning algorithms, several research gaps remain. Current studies often focus on single crop diseases, limiting their generalizability across diverse crops. Additionally, many models face challenges with computational efficiency and scalability. These limitations hinder the widespread adoption of such models in resource-constrained agricultural environments and leads to the research question “Can a hybrid deep learning and SVM model achieve high classification accuracy for multiple crop diseases affecting major crops in Bangladesh (potato late blight, rice brown spot, and corn common rust), while maintaining computational efficiency suitable for real-world agricultural applications?”
This study aims to address these gaps and answer the mentioned research question by developing a hybrid model combining EfficientNetB0 and SVM, tailored for multiple crops and diseases. Our objectives are to enhance accuracy, improve computational efficiency, and create a model that can be applied to diverse agricultural contexts in Bangladesh.

2. Methodology

This work introduces a hybrid model to predict crop diseases for three types of crops: rice, potato and corn. A pre-processed input image is fed into EfficientNet for feature extraction. Figure 1 illustrates the model architecture where it is shown how the features are being extracted in the blocks of EfficientNet. EfficientNet B0 consists of multiple layers that progressively extract features from input images. These layers can be broadly categorized into low-level, mid-level, and high-level layers based on the abstraction level of the features they capture. The low-level layers usually consist of the initial convolutional and pooling layers. These layers operate directly on the raw pixel values of the input image, convolving over small regions of the image to detect simple patterns like edges and gradients. The mid-level layers are often found in the middle of the network architecture. They learn to detect more sophisticated patterns and textures in the input images, such as object parts, shapes, and contours. High-level layers are located towards the end of the network. These layers find object categories or scene types. The high-level layers integrate information from lower-level and mid-level layers. The extracted features are then passed through the fully connected layers after flattening. Then, the SVM is used as a classifier for disease prediction.

2.1. Dataset Preparation

2.1.1. Data Preprocessing

The datasets used in this study include open-source images from Kaggle [24], as well as 1334 images collected by the authors from different fields across different parts of Bangladesh. However, the data we collected was not sufficient enough, so we decided to collect data from open-source datasets as well. There are 7000 images in our dataset after performing data preprocessing and augmentation. The dataset was divided into training and testing, with 80% as train dataset and 20% as test dataset. There were 7 classes, Corn Common Rust, Corn Healthy, Potato Healthy, Potato Late Blight, Rice Brown Spot, Rice Healthy, and Invalid Images. Through data preprocessing, we ensured the quality, diversity, and reliability of our dataset. The preprocessing steps outlined below were executed to optimize the images for subsequent model training and evaluation.

2.1.2. Image Resizing

We prioritized images that exhibited clear visual symptoms of crop diseases, ensuring the dataset’s relevance and utility for our research objectives. Additionally, we included images of healthy plants so that our model could classify healthy and affected crops. To deal with inconsistent image resolution, we have performed image resizing. The original image size was 256 × 256 pixels, which was then resized for model training. The image size was set to 224 × 224 to ensure all the fine details were captured during feature extraction.

2.1.3. Normalization

Normalization is a process of translating data into the range [0, 1]. It includes scaling and transforming numerical features in a dataset to a standard scale. It is mostly performed through min-max scaling and standardization scaling. Following resizing, we normalized the pixel values within each image to a standardized range of [0, 1]. By dividing the pixel values by 255, the maximum intensity value, we effectively scaled the pixel values to a common range, decreasing the impact of lighting variations and enhancing model convergence during training.

2.1.4. Data Augmentation

To enhance the diversity of our dataset and improve model generalization, we applied a range of augmentation techniques, including rotation, horizontal and vertical flipping, height and width shift range, and zooming. The ImageDataGenerator class from the Keras library was used for augmentation during model training. We used 5666 original images and applied augmentation techniques, resulting in a total of 7000 images for training.

2.1.5. Train-Test Split

The preprocessed dataset was divided into training and testing subsets. The subsets had an 80:20 split ratio: 80% of the data was allocated for model training, while the remaining 20% was reserved for testing. There were 5600 images in the training dataset and 1400 in the testing dataset. This was done to evaluate the model’s performance on unseen data while preserving a substantial training dataset for effective model learning. Figure 2 represents the distribution of the final dataset among various classes. Figure 3 shows some of the sample images from the dataset. It is visible that the dataset is balanced as all classes have the same number of samples.

2.2. Model Selection

After data pre-processing, an extensive literature study was carried out again. A thorough exploration of various model architectures was conducted to decide on the best approach to create a hybrid model. Starting with a fundamental convolutional neural network (CNN), we proceeded to explore more advanced models such as VGG16, ResNet50, EfficientNetB0, and other CNN models for feature extraction. Then, we decided to choose the most appropriate classification models for our disease detection project. The models we explored and compared with were CNN, VGG16, ResNet50, Xception, Autoencoders, Inception v3, and Efficient B0.

2.2.1. EfficientNetB0 for Feature Extraction

We incorporated EfficientNetB0 for feature extraction which is renowned for its efficiency and scalability in handling image data. EfficientNet’s advanced architecture enabled us to extract informative features from the preprocessed crop images. It helped capture relevant patterns and characteristics from crop images [25]. Capturing intricate patterns is crucial for accurate disease identification. Through EfficientNetB0, we enhanced the discriminative power of our model. This contributed to improving the performance in distinguishing between healthy plants and those affected by various diseases.

2.2.2. SVM for Classification

After feature extraction using EfficientNet B0, a support vector machine (SVM) was integrated for crop disease classification. SVMs excel at establishing intricate decision boundaries based on the extracted features. This synergy between EfficientNet’s feature representation and SVM’s classification capabilities allowed for accurate disease identification within the crop imagery. By combining these techniques, we leveraged both strengths, particularly SVM’s ability to handle high-dimensional data effectively, ultimately leading to a more robust and precise disease detection system [26].

2.2.3. Hybrid Model

Our investigation, through experimentation and performance evaluation, revealed that the combination of EfficientNetB0 for feature extraction and SVM for classification achieved promising results in crop disease detection [27]. This pairing exhibited superior accuracy compared to alternative models. The selection of EfficientNetB0 and SVM stemmed from their complementary strengths, creating a model architecture well-suited to the intricate challenges of crop disease identification.

2.2.4. EfficientNet Architecture

EfficientNet’s innovative approach to scaling, called compound scaling, utilizes a coefficient (φ) to proportionally adjust the network’s width, depth, and resolution. This ensures balanced scaling across these dimensions, achieving an optimal trade-off between model complexity and performance, as referenced in [28,29,30]. The specific scaling factors for width, depth, and resolution are determined through empirical evaluations (Equations (1)–(3)). These equations represent width (w), depth (d), and resolution (r).
Width: w = β·φ,
Depth: d = α·φ,
Resolution: r = γ·φ,
The constants α, β, and γ are established through experimentation to achieve the most effective balance for a particular task.
The input layer receives the input data [31]. EfficientNet B0 contains a series of convolutional layers that perform feature extraction from the input images. These convolutional layers use filters/kernels to convolve over the input images, extracting features at different spatial scales. Batch normalization layers are often used afterwards to normalize the activations. The activation layers introduce non-linearities into the network, enabling it to learn complex patterns and relationships in the data [29]. At the end of the convolutional layers, EfficientNet includes a global average pooling layer. This layer accumulates feature maps by taking the average of each feature map. Following that, there may be one or more fully connected layers. Finally, the output layer produces the final predictions of the model. For classification tasks, this layer typically consists of ReLU or softmax activation units corresponding to the number of classes in the dataset. Figure 4 represents the architecture of EfficientNetB0. Table 2 describes the layers in detail for the model.

2.2.5. Transfer Learning

Through transfer learning, we fine-tuned a pre-trained Efficient Net B0 model using our crop disease dataset. By initializing the model with weights pre-trained on the Image Net dataset, the model efficiently leveraged high-level feature representations learned from a diverse range of images. Fine-tuning involved updating the parameters of the final layers to adapt the model to our specific task while retaining the knowledge gained from Image Net. Mathematically, this can be represented by adjusting the model parameters to minimize the loss function L. Figure 5 illustrates the hybrid model structure.

2.2.6. Feature Extraction Process

In the feature extraction phase, the input images of crops underwent a transformative journey through the layers of the EfficientNetB0 model. This process unfolded as follows:
  • Input images: At the outset, our crop disease detection system receives input images depicting various crops, such as corn, rice, or potatoes [32]. These images serve as the raw data input into the EfficientNetB0 model, capturing visual information about the crops’ appearance and condition.
  • Propagating through layers: Once the input images are fed into the model, they propagate through the layers of the EfficientNetB0 architecture. The model learns from its errors through backpropagation.
  • Extraction of abstract features: As the images traverse through the network, the successive layers extract increasingly abstract and complex features [33,34]. Initially, lower layers may detect simple patterns like edges and textures, while deeper layers capture more sophisticated structures and arrangements specific to crop diseases.
  • Aggregation and processing: The extracted features from different layers are aggregated and processed further as they progress through the network [35,36]. Through the process of feature fusion and refinement, the model learns to combine and manipulate the extracted features to enhance their representational power.

2.2.7. Model Training

We set different learning rates during model training. We monitored the model’s performance on a separate validation dataset. The training process was halted if the validation loss did not improve for a certain number of epochs. We adjusted the dropout rate to achieve better results.
We experimented with different batch sizes and epochs to find the optimal configuration for training the model. A batch size of 64 was chosen to balance computational efficiency and model convergence [37,38]. The model was trained for 100 epochs to ensure convergence and capture complex patterns in the data.

3. Results and Discussion

This model was experimented on in two stages: training and testing. The hybrid model, combining EfficientNetB0 as a feature extractor with a support vector machine (SVM) classifier, demonstrated better performance for the specified classes—rice, potato, and corn.
Figure 6 represents the accuracy and loss values of the hybrid model of training and testing. Table 3 and Table 4 describe the classification reports for training and testing respectively where it is observed that the training accuracy is 99.63% and the testing accuracy is 97.29%. The reports also include precision, recall, and F1-score to provide a detailed understanding of the model’s performance for individual classes. These metrics were calculated using a weighted average across classes.
Table 5 illustrates the comparison among various contemporary deep learning models including the proposed hybrid model based on accuracy. The batch size (64), epoch (15), and learning rate were kept the same to conduct a fair comparison. It is observed that our hybrid model achieved the best performance out of all these models with an accuracy of 97.29%. The results indicate that the proposed hybrid model achieved a high accuracy for all three diseases, demonstrating its potential for practical use. However, further validation in real-world conditions is necessary to confirm its effectiveness.

4. Conclusions

In conclusion, this study presents a robust framework for crop disease prediction in Bangladesh, a nation where such capabilities are paramount for guaranteeing long-term agricultural productivity and food security. Although we utilized open-source datasets, a significant portion of our data, comprising 23.54%, was collected from agricultural fields in Bangladesh. By strategically integrating deep learning methodologies, specifically EfficientNetB0, with the established classification prowess of support vector machines (SVM), we have developed a hybrid model optimized for disease classification across three vital crops: rice, potato, and corn. This model capitalizes on EfficientNetB0′s proficiency in feature extraction, demonstrably leading to enhanced performance and accuracy. Notably, the proposed hybrid model achieves a commendable accuracy of 97.29%, surpassing the performance of several benchmark models during comparative analysis. This study makes a valuable contribution to the field of plant disease identification by demonstrating the effectiveness of deep learning models across multiple crops. However, further research is needed to enhance the models’ robustness and adaptability to diverse agricultural environments. Future endeavors could involve expanding this approach to encompass a wider range of crops and integrating real-time monitoring systems. Such advancements have the potential to further bolster agricultural sustainability and resilience.

Author Contributions

Conceptualization, S.B. and J.R.; methodology, J.R.; software, J.R.; validation, A.B., Z.M. and S.A.R.; formal analysis, S.B.; resources, F.A.F., J.R. and Z.M.; data curation, A.B. and S.A.R.; writing—original draft preparation, A.B., Z.M. and S.A.R.; writing—review and editing, S.B., F.A.F., M.T.S., H.A.K., S.M., N.A.K. and M.F.B.H.; visualization, J.R.; supervision, S.B., F.A.F., M.T.S., H.A.K. and S.M.; project administration, J.R.; funding acquisition, F.A.F., H.A.K. and S.M. All authors have read and agreed to the published version of the manuscript.

Funding

Multimedia University, Cyberjaya, Selangor, Malaysia (Grant Number: MMUI/230023.02).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All the data used in this study can be obtained from the corresponding author upon a reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Robinson, A. Late Blight in Potato. NDSU Agriculture and Extension, 5 May 2017. Available online: https://www.ndsu.edu/agriculture/extension/publications/late-blight-potato (accessed on 15 March 2024).
  2. Andrivon, D. Biology, ecology, and epidemiology of the potato late blight pathogen Phytophthora infestans in soil. Phytopathology 1995, 85, 1053–1056. [Google Scholar] [CrossRef]
  3. Shi, X.; Yang, J. Hybrid system of convolutional neural networks and SVM for multi-class MIEEG classification. In Proceedings of the 2019 IEEE 11th International Conference on Advanced Infocomm Technology (ICAIT), Jinan, China, 18–20 October 2019; pp. 137–141. [Google Scholar]
  4. Bansal, P.; Kumar, R.; Kumar, S. Disease detection in apple leaves using deep convolutional neural network. Agriculture 2021, 11, 617. [Google Scholar] [CrossRef]
  5. Afzaal, H.; Farooque, A.A.; Schumann, A.W.; Hussain, N.; McKenzie-Gopsill, A.; Esau, T.; Abbas, F.; Acharya, B. Detection of a potato disease (early blight) using artificial intelligence. Remote Sens. 2021, 13, 411. [Google Scholar] [CrossRef]
  6. Mia, M.J.; Maria, S.K.; Taki, S.S.; Biswas, A.A. Cucumber disease recognition using machine learning and transfer learning. Bull. Electr. Eng. Inform. 2021, 10, 3432–3443. [Google Scholar] [CrossRef]
  7. Pinki, F.T.; Khatun, N.; Islam, S.M. Content based paddy leaf disease recognition and remedy prediction using support vector machine. In Proceedings of the 2017 20th International Conference of Computer and Information Technology (ICCIT), Dhaka, Bangladesh, 22–4 December 2017; pp. 1–5. [Google Scholar]
  8. Militante, S.V.; Gerardo, B.D.; Medina, R.P. Sugarcane disease recognition using deep learning. In Proceedings of the 2019 IEEE Eurasia Conference on IoT, Communication and Engineering (ECICE), Yunlin, Taiwan, 3–6 October 2019; pp. 575–578. [Google Scholar]
  9. Jadhav, S.B.; Udupi, V.R.; Patil, S.B. Identification of plant diseases using convolutional neural networks. Int. J. Inf. Technol. 2021, 13, 2461–2470. [Google Scholar] [CrossRef]
  10. Astani, M.; Hasheminejad, M.; Vaghefi, M. A diverse ensemble classifier for tomato disease recognition. Comput. Electron. Agric. 2022, 198, 107054. [Google Scholar] [CrossRef]
  11. Kolluri, J.; Dash, S.K.; Das, R. Plant disease identification based on multimodal learning. Int. J. Intell. Syst. Appl. Eng. 2024, 12, 634–643. [Google Scholar]
  12. Eunice, J.; Popescu, D.E.; Chowdary, M.K.; Hemanth, J. Deep learning-based leaf disease detection in crops using images for agricultural applications. Agronomy 2022, 12, 2395. [Google Scholar] [CrossRef]
  13. Tammina, S. Transfer Learning Using VGG-16 with Deep Convolutional Neural Network for Classifying Images. Available online: https://www.ijsrp.org/research-paper-1019.php (accessed on 15 March 2024).
  14. Smith, J.; Johnson, A.; Davis, C. Automated Detection of Potato Late Blight Using Deep Learning and Transfer Learning Techniques. Agronomy 2023, 13, 1245. Available online: https://www.mdpi.com/2073-4395/13/6/1245 (accessed on 15 July 2024).
  15. Sheikh, M.H.; Mim, T.T.; Reza, M.S.; Rabby, A.S.A.; Hossain, S.A. Detection of Maize and Peach Leaf Diseases Using Image Processing. In Proceedings of the 2019 10th International Conference on Computing, Communication and Networking Technologies (ICCCNT), Kanpur, India, 6–8 July 2019; pp. 1–7. [Google Scholar] [CrossRef]
  16. Hossen, M.S.; Haque, I.; Islam, M.S.; Ahmed, M.T.; Nime, M.J.; Islam, M.A. Deep Learning Based Classification of Papaya Disease Recognition. In Proceedings of the 2020 3rd International Conference on Intelligent Sustainable Systems (ICISS), Thoothukudi, India, 3–5 December 2020; pp. 945–951. [Google Scholar]
  17. Sarwinda, D.; Paradisa, R.H.; Bustamam, A.; Anggia, P. Deep Learning in Image Classification Using Residual Network (ResNet) Variants for Detection of Colorectal Cancer. Procedia Comput. Sci. 2021, 179, 423–431. [Google Scholar] [CrossRef]
  18. Liu, J.; Wang, M.; Bao, L.; Li, X. EfficientNet Based Recognition of Maize Diseases by Leaf Image Classification. J. Phys. Conf. Ser. 2020, 1693, 012148. [Google Scholar] [CrossRef]
  19. Spanhol, F.A.; Oliveira, L.S.; Petitjean, C.; Heutte, L. Breast Cancer Histopathological Image Classification Using Convolutional Neural Networks. In Proceedings of the 2016 International Joint Conference on Neural Networks (IJCNN), Vancouver, BC, Canada, 24–29 July 2016; pp. 2560–2567. [Google Scholar]
  20. Espejo-Garcia, B.; Malounas, I.; Mylonas, N.; Kasimati, A.; Fountas, S. Using EfficientNet and Transfer Learning for Image-Based Diagnosis of Nutrient Deficiencies. Comput. Electron. Agric. 2022, 196, 106868. [Google Scholar] [CrossRef]
  21. Yang, H.; Ni, J.; Gao, J.; Han, Z.; Luan, T. A Novel Method for Peanut Variety Identification and Classification by Improved VGG16. Sci. Rep. 2021, 11, 15756. [Google Scholar] [CrossRef]
  22. Raza, R.; Zulfiqar, F.; Khan, M.O.; Arif, M.; Alvi, A.; Iftikhar, M.A.; Alam, T. Lung-EffNet: Lung Cancer Classification Using EfficientNet from CT-Scan Images. Eng. Appl. Artif. Intell. 2023, 126, 106902. [Google Scholar] [CrossRef]
  23. Huang, Z.; Su, L.; Wu, J.; Chen, Y. Rock Image Classification Based on EfficientNet and Triplet Attention Mechanism. Appl. Sci. 2023, 13, 3180. [Google Scholar] [CrossRef]
  24. Available online: https://www.kaggle.com/datasets/nafishamoin/bangladeshi-crops-disease-dataset (accessed on 11 June 2023).
  25. Atila, U.; Uçar, M.; Akyol, K.; Uçar, E. Plant leaf disease classification using EfficientNet deep learning model. Ecol. Inform. 2021, 61, 101182. [Google Scholar] [CrossRef]
  26. Padol, P.B.; Yadav, A.A. SVM classifier-based grape leaf disease detection. In Proceedings of the 2016 Conference on Advances in Signal Processing (CASP), Pune, India, 9–11 June 2016; pp. 175–179. [Google Scholar]
  27. Chaudhari, D.J.; Malathi, K. Detection and prediction of rice leaf disease using a hybrid CNN-SVM model. Opt. Mem. Neural Netw. 2023, 32, 39–57. [Google Scholar] [CrossRef]
  28. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  29. Stoica, G.; Stretcu, O.; Platanios, E.A.; Mitchell, T.; Póczos, B. Contextual parameter generation for knowledge graph link prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 3000–3008. [Google Scholar]
  30. Simeone, O. A very brief introduction to machine learning with applications to communication systems. IEEE Trans. Cogn. Commun. Netw. 2018, 4, 648–664. [Google Scholar] [CrossRef]
  31. Al Farid, F.; Hashim, N.; Abdullah, J.; Bhuiyan, M.R.; Kairanbay, M.; Yusoff, Z.; Karim, H.A.; Mansor, S.; Sarker, M.T.; Ramasamy, G. Single Shot Detector CNN and Deep Dilated Masks for Vision-Based Hand Gesture Recognition from Video Sequences. IEEE Access 2024, 12, 28564–28574. [Google Scholar] [CrossRef]
  32. Aabidi, M.H.; ELMakrani, A.; Jabir, B.; Zaimi, I. A Model Proposal for Enhancing Leaf Disease Detection Using Convolutional Neural Networks (CNN): Case Study. Int. J. Online Biomed. Eng. 2023, 19, 127. [Google Scholar] [CrossRef]
  33. Wang, P.; Fan, E.; Wang, P. Comparative analysis of image classification algorithms based on traditional machine learning and deep learning. Pattern Recognit. Lett. 2021, 141, 61–67. [Google Scholar] [CrossRef]
  34. Najafabadi, M.M.; Villanustre, F.; Khoshgoftaar, T.M.; Seliya, N.; Wald, R.; Muharemagic, E. Deep learning applications and challenges in big data analytics. J. Big Data 2015, 2, 1. [Google Scholar] [CrossRef]
  35. Yang, Y.; Cui, X. Bert-enhanced text graph neural network for classification. Entropy 2021, 23, 1536. [Google Scholar] [CrossRef] [PubMed]
  36. Yao, R.; Wang, N.; Liu, Z.; Chen, P.; Sheng, X. Intrusion detection system in the advanced metering infrastructure: A cross-layer feature-fusion CNN-LSTM-based approach. Sensors 2021, 21, 626. [Google Scholar] [CrossRef] [PubMed]
  37. Ma, Z.; Xu, Y.; Xu, H.; Meng, Z.; Huang, L.; Xue, Y. Adaptive batch size for federated learning in resource-constrained edge computing. IEEE Trans. Mob. Comput. 2021, 22, 37–53. [Google Scholar] [CrossRef]
  38. Dai, G.; Hu, L.; Fan, J. DA-ActNN-YOLOV5: Hybrid YOLO v5 model with data augmentation and activation of compression mechanism for potato disease identification. Comput. Intell. Neurosci. 2022, 2022, 6114061. [Google Scholar] [CrossRef]
Figure 1. Hybrid model architecture.
Figure 1. Hybrid model architecture.
Jimaging 10 00183 g001
Figure 2. Dataset distribution.
Figure 2. Dataset distribution.
Jimaging 10 00183 g002
Figure 3. Sample from the dataset.
Figure 3. Sample from the dataset.
Jimaging 10 00183 g003
Figure 4. EfficientNetB0 model architecture.
Figure 4. EfficientNetB0 model architecture.
Jimaging 10 00183 g004
Figure 5. Hybrid model structure diagram.
Figure 5. Hybrid model structure diagram.
Jimaging 10 00183 g005
Figure 6. Hybrid plot training & validation accuracy values and validation loss values.
Figure 6. Hybrid plot training & validation accuracy values and validation loss values.
Jimaging 10 00183 g006
Table 1. Literature Analysis.
Table 1. Literature Analysis.
ModelTraditional MLInceptionCNNResNetMobileNetV2VGGNetYOLODenseNetGoogleNetEfficientNet
Paper [5]
Paper [8]
Paper [12]
Paper [13]
Paper [14]✓ (RF)
Paper [15]
Paper [16]
Paper [17]
Paper [18]
Paper [19]
Paper [20]
Paper [21]
Paper [22]
Paper [23]
Table 2. EfficientNetB0 model table.
Table 2. EfficientNetB0 model table.
Stage iOperatorResolutionNo. of
Channels
No. of Layers
1Conv3×3112 × 112321
2MBConv1, k3×3112 × 112161
3MBConv6, k3×3112 × 112242
4MBConv6, k5×556 × 56402
5MBConv6, k3×328 × 28803
6MBConv6, k5×514 × 141123
7MBConv6, k5×57 × 71824
8MBConv6, k3×37 × 73201
9Conv1×1 & Pooling & FC7 × 712801
Table 3. Hybrid classification report (train).
Table 3. Hybrid classification report (train).
PrecisionRecallF1-ScoreSupport
Corn Common Rust111800
Corn Healthy0.99626410.9981285800
Invalid Images111800
Potato Healthy111800
Potato Late Blight111800
Rice Brown Spot10.98250.9868173800
Rice Healthy0.99117280.991250.9887781800
Accuracy 0.996255600
Macro Average0.99625070.996250.99624635600
Weighted Average0.99625070.996250.99624635600
Table 4. Hybrid classification report (test).
Table 4. Hybrid classification report (test).
PrecisionRecallF1-ScoreSupport
Corn Common_Rust111200
Corn Healthy0.99009910.9950249200
Invalid Images110.9974937200
Potato Healthy0.984848510.9798995200
Potato Late Blight0.975247510.9800995200
Rice Brown Spot0.9523810.98250.9254499200
Rice Healthy0.90952380.991250.9317073200
Accuracy 0.97285711400
Macro Average0.97315710.97285210.97281071400
Weighted Average0.97315710.97285210.97281071400
Table 5. Comparative analysis of ResNet50, VGG16, CNN, EfficientNetB0, Xception, mobilenetV2, Autoencoders, Inception V3, Hybrid.
Table 5. Comparative analysis of ResNet50, VGG16, CNN, EfficientNetB0, Xception, mobilenetV2, Autoencoders, Inception V3, Hybrid.
ModelInput ShapeTrain AccuracyTest Accuracy
CNN128, 128, 391.20%86.57%
VGG16128, 128, 385.98%83.29%
ResNet50128, 128, 371.75%68.79%
Xceptiion128, 128, 395.05%94.07%
Mobilenet V2128, 128, 392.00%90.71%
Autoencoders128, 128, 387.98%87.90%
Inception v3128, 128, 385.21%94.14%
EfficientnetB0128, 128, 398.02596.14%
Hybrid224, 224, 399.63%97.29%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Barman, S.; Farid, F.A.; Raihan, J.; Khan, N.A.; Hafiz, M.F.B.; Bhattacharya, A.; Mahmud, Z.; Ridita, S.A.; Sarker, M.T.; Karim, H.A.; et al. Optimized Crop Disease Identification in Bangladesh: A Deep Learning and SVM Hybrid Model for Rice, Potato, and Corn. J. Imaging 2024, 10, 183. https://doi.org/10.3390/jimaging10080183

AMA Style

Barman S, Farid FA, Raihan J, Khan NA, Hafiz MFB, Bhattacharya A, Mahmud Z, Ridita SA, Sarker MT, Karim HA, et al. Optimized Crop Disease Identification in Bangladesh: A Deep Learning and SVM Hybrid Model for Rice, Potato, and Corn. Journal of Imaging. 2024; 10(8):183. https://doi.org/10.3390/jimaging10080183

Chicago/Turabian Style

Barman, Shohag, Fahmid Al Farid, Jaohar Raihan, Niaz Ashraf Khan, Md. Ferdous Bin Hafiz, Aditi Bhattacharya, Zaeed Mahmud, Sadia Afrin Ridita, Md Tanjil Sarker, Hezerul Abdul Karim, and et al. 2024. "Optimized Crop Disease Identification in Bangladesh: A Deep Learning and SVM Hybrid Model for Rice, Potato, and Corn" Journal of Imaging 10, no. 8: 183. https://doi.org/10.3390/jimaging10080183

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop