Deep Convolutional Neural Networks for Tea Tree Pest Recognition and Diagnosis
Abstract
:1. Introduction
- Take CNN (also recognized as the pre-training model) as the feature extractor, remove the last layer of the network, take the rest of CNN as the feature extractor to get the vectors, and then use these feature vectors to train a classifier (such as an SVM).
- Fine-tune the weight parameters of the target CNN on the new dataset. In the fine-tuning process, all the layer parameters of the model can be adjusted, or the parameters of the first several layers of the model can be fixed. Then we train the parameters of the last layers or just the softmax layer. We fine-tune the last layer because the features extracted from the first few layers of the CNN are universal features, while the features extracted from the last few layers of the model are related to datasets and classification tasks. Therefore, only the latter layers can be adjusted in the new datasets, and the training time can be greatly shortened.
2. Materials and Methods
2.1. Insect Image Dataset
2.2. Performance Measurements
2.3. Pre-Trained Convolutional Neural Networks
2.4. Fine-Tune the CNN Model
2.5. Dense SIFT-Based BoVW Model
- Extraction and description the image features: Detect interest points through random sampling and obtain local features of the image. Common local features descriptors include the SIFT, SURF, and HOG. In our study, we employ the DSIFT descriptor. The main difference between the DSIFT and SIFT is their different sampling methods. SIFT descriptor detects and screens the feature point by building scale space. Conversely, the DSIFT descriptor divides the image into rectangular areas of the same size and then uses a fixed-size window to sample from left to right and from top to bottom with a certain step length to extract the SIFT features. Finally, each feature is represented by a 128-dimension vector. The features extracted by this method are evenly distributed with the same specifications, and have stable illumination, translation, and rotation.
- Construction the visual vocabulary: K-means algorithm is used to cluster the local feature vectors of all sample images. The size of the visual vocabulary is N, if there are N cluster centers (defined as a visual word). In this manuscript, the size of the visual vocabulary is defined as 1000.
- Image representation: We measure distance each local features to the visual word of the vocabulary and map the local features to the nearest visual word. We then compute the occurrences of each visual word in the image, which becomes a N-dimensional numerical vector. In this paper, each tea pests image represented by a 1000-dimensional numerical vector.
- Training classifier: In our study, we apply the SVM and MLP to classify and identify the 1000-dimensional numerical vector and the label of the input image is determined by the classifier. The workflow based on the BoVW model is shown in Figure 4.
3. Results
4. Discussion
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Conflicts of Interest
References
- Yang, N.; Yuan, M.F.; Wang, P.; Zhang, R.B.; Sun, J.; Mao, H.P. Tea Diseases Detection Based on Fast Infrared Thermal Image Processing Technology. J. Sci. Food Agric. 2019, 99, 3459–3466. [Google Scholar] [CrossRef] [PubMed]
- Hazarika, L.K.; Bhuyan, M.; Hazarika, B.N. Insect Pests of Tea and Their Management. Annu. Rev. Entomol. 2009, 54, 267–284. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Martineau, M.; Conte, D.; Raveaux, R.; Arnault, I.; Munier, D.; Venturini, G. A survey on image-based insect classification. Pattern Recognit. 2017, 65, 273–284. [Google Scholar] [CrossRef] [Green Version]
- Liu, Z.; Gao, J.; Yang, G.; Zhang, H.; He, Y. Localization and Classification of Paddy Field Pests using a Saliency Map and Deep Convolutional Neural Network. Sci. Rep. 2016, 6, 20410. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
- Hughes, D.; Salathe, M. An open access repository of images on plant health to enable the development of mobile disease diagnostics. arXiv 2015, arXiv:1511.08060v2. [Google Scholar]
- Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Pichayoot, O. Corn Disease identification from leaf images using convolutional neural networks. In Proceedings of the 2017 21st International Computer Science and Engineering Conference (ICSEC), Bangkok, Thailand, 15–18 November 2017; pp. 233–238. [Google Scholar]
- Ma, J.; Du, K.; Zheng, F.; Zhang, L.; Gong, Z.; Sun, Z. A recognition method for cucumber diseases using leaf symptom images based on deep convolutional neural network. Comput. Electron. Agric. 2018, 154, 18–24. [Google Scholar] [CrossRef]
- Chen, J.; Liu, Q.; Gao, L. Visual Tea Leaf Disease Recognition Using a Convolutional Neural Network Model. Symmetry 2019, 11, 343. [Google Scholar] [CrossRef] [Green Version]
- Zhang, X.; Qiao, Y.; Meng, F.; Fan, C.; Zhang, M. Identification of Maize Leaf Diseases Using Improved Deep Convolutional Neural Networks. IEEE Access 2018, 6, 30370–30377. [Google Scholar] [CrossRef]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. In Proceedings of the 25th International Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2015, arXiv:1409.1556. [Google Scholar]
- Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. arXiv 2015, arXiv:1409.4842. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
- Tan, C.; Sun, F.; Kong, T.; Zhang, W.; Yang, C.; Liu, C. A Survey on deep transfer learning. In Proceedings of the 27th International Conference on Artificial Neural Networks, Rhodes, Greece, 4–7 October 2018; pp. 270–279. [Google Scholar] [CrossRef] [Green Version]
- Souza, W.; Alves, A.N.; Borges, D.L. A deep learning model for recognition of pest insects in maize plantations. In Proceedings of the 2019 IEEE International Conference on Systems, Man and Cybernetics (SMC), Bari, Italy, 6–9 October 2019. [Google Scholar]
- Cao, X.; Wei, Z.; Gao, Y.; Huo, Y. Recognition of Common Insect in Field Based on Deep Learning. J. Phys. Conf. Ser. 2020, 1634, 012034. [Google Scholar] [CrossRef]
- Pan, S.J.; Yang, Q. A Survey on Transfer Learning. IEEE Trans. Knowl. Data Eng. 2009, 22, 1345–1359. [Google Scholar] [CrossRef]
- Yosinski, J.; Clune, J.; Bengio, Y.; Lipson, H. How transferable are features in deep neural networks? Adv. Neural Inf. Process. Syst. 2014, 27, 3320–3328. [Google Scholar]
- Sladojevic, S.; Arsenovic, M.; Anderla, A.; Culibrk, D.; Stefanovic, D. Deep Neural Networks Based Recognition of Plant Diseases by Leaf Image Classification. Comput. Intell. Neurosci. 2016, 2016, 3289801. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Liu, B.; Zhang, Y.; He, D.; Li, Y. Identification of Apple Leaf Diseases Based on Deep Convolutional Neural Networks. Symmetry 2017, 10, 11. [Google Scholar] [CrossRef] [Green Version]
- Ferentinos, K. Deep learning models for plant disease detection and diagnosis. Comput. Electron. Agric. 2018, 145, 311–318. [Google Scholar] [CrossRef]
- Ramcharan, A.; Baranowski, K.; McCloskey, P.; Ahmed, B.; Legg, J.; Hughes, D. Using Transfer Learning for Image-Based Cassava Disease Detection. arXiv 2017, arXiv:1707.03717v2. [Google Scholar]
- Korzh, O.; Cook, G.; Andersen, T.; Serra, E. Stacking approach for CNN transfer learning ensemble for remote sensing imagery. In Proceedings of the 2017 Intelligent Systems Conference, London, UK, 7–8 September 2017; pp. 599–608. [Google Scholar]
- Jia, Y.; Shelhamer, E.; Donahue, J.; Karayev, S.; Long, J.; Girshick, R.; Guadarrama, S.; Darrell, T. Caffe: Convolutional architecture for fast feature embedding. In Proceedings of the ACM International Conference on Multimedia, Orlando, FL, USA, 3–7 November 2014; pp. 675–678. [Google Scholar]
- Csurka, G.; Dance, C.R.; Fan, L.X.; Willamowski, J.; Bray, C. Visual categorization with bags of keypoints. In Proceedings of the 2004 Workshop on Statistical Learning in Computer Vision, Prague, Czech Republic, 10–16 May 2004; pp. 1–2. [Google Scholar]
Class | Number of Images from the Dataset Used for Train | Number of Images from the Dataset Used for Validate | Number of Images from the Dataset Used for Test |
---|---|---|---|
(1) Ricania speculum (Walker) | 768 | 96 | 96 |
(2) Euricania ocellus (Walker) | 730 | 91 | 91 |
(3) Ricania sublimbata (Jacobi) | 722 | 90 | 90 |
(4) Ceratonoros transiens (Walker) | 768 | 96 | 96 |
(5) Spilarctia subcarnea (Walker) | 804 | 101 | 100 |
(6) Homona coffearia (Meyrick) | 710 | 89 | 89 |
(7) Eterusia aedea (Linnaeus) | 774 | 97 | 97 |
(8) Culcula panterinaria (Bremer et Gray) | 770 | 96 | 96 |
(9) Euproctis pseudoconspersa (Strand) | 802 | 100 | 100 |
(10) Arctornis alba (Bremer) | 760 | 95 | 95 |
(11) Rikiosatoa vandervoordeni (Prout) | 843 | 105 | 105 |
(12) Scopula subpunctaria (Herrich-Schaeffer) | 794 | 99 | 99 |
(13) Amata germana (Felder) | 790 | 99 | 99 |
(14) Spilosoma menthastri (Esper) | 756 | 95 | 94 |
Total | 10,791 | 1349 | 1347 |
Layer | Parameter | Output |
---|---|---|
Input | 224 × 224 × 3 | - |
Conv1-1 | 64 convolution filters (3 × 3), 1 stride, 1 pad | 224 × 224 × 64 |
Conv1-2 | 64 convolution filters (3 × 3), 1 stride, 1 pad | 224 × 224 × 64 |
Max pool 1 | Max pooling (2 × 2), 2 stride | 112 × 112 × 64 |
Conv2-1 | 128 convolution filters (3 × 3), 1 stride, 1 pad | 112 × 112 × 128 |
Conv2-2 | 128 convolution filters (3 × 3), 1 stride, 1 pad | 112 × 112 × 128 |
Max pool 2 | Max pooling (2 × 2), 2 stride | 56 × 56 × 128 |
Conv3-1 | 256 convolution filters (3 × 3), 1 stride, 1 pad | 56 × 56 × 256 |
Conv3-2 | 256 convolution filters (3 × 3), 1 stride, 1 pad | 56 × 56 × 256 |
Conv3-3 | 256 convolution filters (3 × 3), 1 stride, 1 pad | 56 × 56 × 256 |
Max pool 3 | Max pooling (2 × 2), 2 stride | 28× 28 × 256 |
Conv4-1 | 512 convolution filters (3 × 3), 1 stride, 1 pad | 28× 28 × 512 |
Conv4-2 | 512 convolution filters (3 × 3), 1 stride, 1 pad | 28× 28 × 512 |
Conv4-3 | 512 convolution filters (3 × 3), 1 stride, 1 pad | 28× 28 × 512 |
Max pool 4 | Max pooling (2 × 2), 2 stride | 14× 14 × 512 |
Conv5-1 | 512 convolution filters (3 × 3), 1 stride, 1 pad | 14× 14 × 512 |
Conv5-2 | 512 convolution filters (3 × 3), 1 stride, 1 pad | 14× 14 × 512 |
Conv5-3 | 512 convolution filters (3 × 3), 1 stride, 1 pad | 14× 14 × 512 |
Max pool 5 | Max pooling (2 × 2), 2 stride | 7× 7 × 512 |
Full Connect-6 | 4096 × 1 × 1, 1 stride | 4069 |
Full Connect-7 | 4096 × 1 × 1, 1 stride | 4069 |
Full Connect-8 | 14 × 1 × 1, 1 stride | 14 |
Output | - | 1 |
Method | Pest Category | Accuracy (%) |
---|---|---|
VGGNet-16 | 14 | 97.75% |
VGGNet-16 | 9 | 92.13% |
VGGNet-19 | 9 | 97.39% |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Chen, J.; Liu, Q.; Gao, L. Deep Convolutional Neural Networks for Tea Tree Pest Recognition and Diagnosis. Symmetry 2021, 13, 2140. https://doi.org/10.3390/sym13112140
Chen J, Liu Q, Gao L. Deep Convolutional Neural Networks for Tea Tree Pest Recognition and Diagnosis. Symmetry. 2021; 13(11):2140. https://doi.org/10.3390/sym13112140
Chicago/Turabian StyleChen, Jing, Qi Liu, and Lingwang Gao. 2021. "Deep Convolutional Neural Networks for Tea Tree Pest Recognition and Diagnosis" Symmetry 13, no. 11: 2140. https://doi.org/10.3390/sym13112140
APA StyleChen, J., Liu, Q., & Gao, L. (2021). Deep Convolutional Neural Networks for Tea Tree Pest Recognition and Diagnosis. Symmetry, 13(11), 2140. https://doi.org/10.3390/sym13112140