Using a Hybrid Convolutional Neural Network with a Transformer Model for Tomato Leaf Disease Detection
Abstract
:1. Introduction
- The categories in the three datasets, which are unbalanced and have inadequate sample amounts, are used to augment the tomato dataset by synthesizing new tomato disease images based on an improved cyclegan network (CyTrGan) to improve the robustness and generalization of the subsequent classification models.
- A hybrid model with high classification accuracy and small parameters composed of a Transformer and densely connected networks is proposed.
- This paper compares the recognition performance of the proposed Dual Vision Transformer model (DVT) with seven classical models and recent works in the literature on the tomato disease image classification task. The results show that the proposed classification model can achieve better classification performance. In addition, through visualization experiments, we also found that the model can better learn the detailed features of tomato disease images.
2. Materials and Methods
2.1. CyTrGan as Data Augmentation
Loss Functions
- (1)
- Adversarial loss: Adversarial loss allows the generated disease pictures to match the distribution of the target area as much as possible and could enhance the quality of the disease pictures generated. For : , the generator expects the generated image () to be as similar as possible to the real data, y, in the y-domain. The discriminator D needs to decide whether the input image of a disease is a real image or a synthetic image of disease, and the goal of the generator is to minimize , and that of the discriminator is to maximize it. Both of them are playing games with each other; similarly, : is opposite to the above.
- (2)
- Cycle-consistent loss: The cycle-consistency loss enables the generator to learn the distribution characteristics of diseased images. However, when an image is mapped from the source domain (X) to the target domain (Y) and back to the source, it should preserve similarity with the original image. The same applies when mapping between two diseased images. The cycle consistency loss ensures the cyclic transformations could recover the image to its original state. It acts as a regularization, with the regularization strength being controlled by the coefficient.
- (3)
- Perceptual loss [27]: Perceptual loss has been widely applied as an effective loss in image generation tasks, such as super-resolution and style transfer. It acts as a loss function that can approximate human image similarity evaluation. It compares two different but similar-looking images and extracts high-level semantic features. Perceptual loss utilizes a pre-trained VGG-16 network on the Imagenet dataset to extract texture and structural information, while the fully connected layers extract semantic information. This study regards perceptual loss as a supplement to cycle consistency loss to improve the quality of the generated diseased images. It is typically calculated by passing the input and target images through a pre-trained neural network to obtain their feature representations. These are input to the loss function to compute the Euclidean or Manhattan distance between them.
- (4)
- Total loss:
2.2. Tomato Disease Classification with Dual Vision Transformer Model
Design of Convolutional Blocks
- (1)
- Dense network (local feature extraction module): As shown in Figure 2a, convolutional structures are utilized to densely connect and extract local features from the input image, with a stride of 2 and a channel number of 14 or 16 to lower the image size and computational cost. To better represent the model hierarchically, a supplementary layer comprising a convolutional layer and normalization are added before each stage to reduce the model size for projection to higher dimensions.
- (2)
- ConvNext Block [28]: Compared to ResNet residual blocks, this bottleneck structure has fewer parameters and is more efficient, enhancing the model’s perception of global contexts. Using depthwise convolutions instead of regular convolutions reduces parameters. Applying LayerNorm to the channel dimension (N, H, W, C) is more efficient. Two linear layers are used instead of 1 × 1 convolutions to implement pointwise convolution, improving feature competition and classification performance.
- (3)
- Lightweight Multi-Head Attention [29]: In the original self-attention module, the input linear transformation is for (query), (key), and (value), where is the number of patches. To reduce the computational complexity, depthwise separable convolutions (DWCONVs) are used to extract the key and value instead of directly using the input feature maps, significantly reducing computations. As shown in Figure 2b, the step of inputting the tensor from to n × d is omitted. And , , and are the dimensions of the input, keys (queries), and values, respectively. In calculating attention, a learnable bias parameter, B, is added, which helps the module learn positional information. Finally, the output is not the concatenation of multi-head attention, but a linear mapping added to the input, which reduces parameters. Layer normalization is applied to queries before attention calculation, helping to stabilize training. Overall, compared to standard multi-head self-attention, this module has specific improvement while retaining the merits of self-attention.
- (4)
- FFN (Residual Feed-Forward Network): An FFN consists of two linear layers with GELU activation in between, where the first layer expands the dimension by four times and the second layer reduces it correspondingly. The advantage of FFN modules lies in their small number of parameters. An FFN uses the 1 × 1 convolution to lower and raise the dimension, which is the middle part of using DWCONV, and the overall number of parameters will be less than the large kernel of the standard convolution. The residual connection can effectively prevent the gradient from disappearing in the deep network and play the role of regularization. FFN facilitates training deeper architectures through residual connections, which mitigate gradient vanishing, and provides a regularization effect. Standard convolutions can be replaced with an FFN for hybrid models to reduce parameters and improve computational efficiency while enhancing representational capacity.
3. Implementation and Model Assessment
3.1. Tomato Dataset and Experimental Setup
3.2. Evaluation Metrics
3.3. Experimental Results and Analysis
3.3.1. Performance of CyTrGAN
3.3.2. Turing Study
- (1)
- Confirming whether the images are real or synthetic (Accuracy1);
- (2)
- Determining the category to which the diseased tomato images belong (Accuracy2).
3.3.3. DVT as Classification Model
3.3.4. Performance Comparison
3.3.5. Visualization of Results
4. Conclusions
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Panno, S.; Davino, S.; Caruso, A.G.; Bertacca, S.; Crnogorac, A.; Mandić, A.; Noris, E.; Matić, S. A review of the most common and economically important diseases that undermine the cultivation of tomato crop in the mediterranean basin. Agronomy 2021, 11, 2188. [Google Scholar] [CrossRef]
- Saleem, M.H.; Potgieter, J.; Arif, K.M. Plant disease detection and classification by deep learning. Plants 2019, 8, 468. [Google Scholar] [CrossRef] [PubMed]
- Tanner, M.A.; Wong, W.H. The calculation of posterior distributions by data augmentation. J. Am. Stat. Assoc. 1987, 82, 528–540. [Google Scholar] [CrossRef]
- Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27. [Google Scholar]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:14091556. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
- Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:170404861. [Google Scholar]
- Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:201011929. [Google Scholar]
- Hassan, S.M.; Jasinski, M.; Leonowicz, Z.; Jasinska, E.; Maji, A.K. Plant disease identification using shallow convolutional neural network. Agronomy 2021, 11, 2388. [Google Scholar] [CrossRef]
- Liu, B.; Tan, C.; Li, S.; He, J.; Wang, H. A data augmentation method based on generative adversarial networks for grape leaf disease identification. IEEE Access 2020, 8, 102188–102198. [Google Scholar] [CrossRef]
- Douarre, C.; Crispim-Junior, C.F.; Gelibert, A.; Tougne, L.; Rousseau, D. Novel data augmentation strategies to boost supervised segmentation of plant disease. Comput. Electron. Agric. 2019, 165, 104967. [Google Scholar] [CrossRef]
- Zhu, F.; He, M.; Zheng, Z. Data augmentation using improved cDCGAN for plant vigor rating. Comput. Electron. Agric. 2020, 175, 105603. [Google Scholar] [CrossRef]
- Tian, Y.; Yang, G.; Wang, Z.; Li, E.; Liang, Z. Detection of apple lesions in orchards based on deep learning methods of cyclegan and yolov3-dense. J. Sens. 2019, 2019, 7630926. [Google Scholar] [CrossRef]
- Cap, Q.H.; Uga, H.; Kagiwada, S.; Iyatomi, H. Leafgan: An effective data augmentation method for practical plant disease diagnosis. IEEE Trans. Autom. Sci. Eng. 2020, 19, 1258–1267. [Google Scholar] [CrossRef]
- Abbas, A.; Jain, S.; Gour, M.; Vankudothu, S. Tomato plant disease detection using transfer learning with C-GAN synthetic images. Comput. Electron. Agric. 2021, 187, 106279. [Google Scholar] [CrossRef]
- Sagar, A.; Dheeba, J. On using transfer learning for plant disease detection. bioRxiv 2020. [Google Scholar] [CrossRef]
- Widiyanto, S.; Fitrianto, R.; Wardani, D.T. Implementation of convolutional neural network method for classification of diseases in tomato leaves. In Proceedings of the 2019 4th International Conference on Informatics and Computing (ICIC), Semarang, Indonesia, 16–17 October 2019. [Google Scholar]
- Albogamy, F.R. A Deep Convolutional Neural Network with Batch Normalization Approach for Plant Disease Detection. Int. J. Comput. Sci. Netw. Secur. 2021, 21, 51–62. [Google Scholar]
- Zhao, S.; Peng, Y.; Liu, J.; Wu, S. Tomato leaf disease diagnosis based on improved convolution neural network by attention module. Agriculture 2021, 11, 651. [Google Scholar] [CrossRef]
- Elhassouny, A.; Smarandache, F. Smart mobile application to recognize tomato leaf diseases using convolutional neural networks. In Collected Papers. Volume XI: On Physics, Artificial Intelligence, Health Issues, Decision Making, Economics, Statistics; Global Knowledge Publishing House: Chennai, India, 2019; p. 431. [Google Scholar]
- Ahmed, S.; Hasan, M.B.; Ahmed, T.; Sony MR, K.; Kabir, M.H. Less is more: Lighter and faster deep neural architecture for tomato leaf disease classification. IEEE Access 2022, 10, 68868–68884. [Google Scholar] [CrossRef]
- Yang, B.; Wang, Z.; Guo, J.; Guo, L.; Liang, Q.; Zeng, Q.; Zhao, R.; Wang, J.; Li, C. Identifying plant disease and severity from leaves: A deep multitask learning framework using triple-branch Swin Transformer and deep supervision. Comput. Electron. Agric. 2023, 209, 107809. [Google Scholar] [CrossRef]
- Li, X.; Li, S. Transformer help CNN see better: A lightweight hybrid apple disease identification model based on transformers. Agriculture 2022, 12, 884. [Google Scholar] [CrossRef]
- Hughes, D.; Salathé, M. An open access repository of images on plant health to enable the development of mobile disease diagnostics. arXiv 2015, arXiv:151108060. [Google Scholar]
- Zhu, J.-Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
- Alimanov, A.; Islam, M.B. Retinal Image Restoration using Transformer and Cycle-Consistent Generative Adversarial Network. In Proceedings of the 2022 International Symposium on Intelligent Signal Processing and Communication Systems (ISPACS), Penang, Malaysia, 22–25 November 2022. [Google Scholar]
- Johnson, J.; Alahi, A.; Fei-Fei, L. Perceptual losses for real-time style transfer and super-resolution. In Proceedings of the Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part II. Springer: Berlin/Heidelberg, Germany, 2016. [Google Scholar]
- Woo, S.; Debnath, S.; Hu, R.; Chen, X.; Liu, Z.; Kweon, I.S.; Xie, S. Convnext v2: Co-designing and scaling convnets with masked autoencoders. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023. [Google Scholar]
- Guo, J.; Han, K.; Wu, H.; Tang, Y.; Chen, X.; Wang, Y.; Xu, C. CMT: Convolutional neural networks meet vision transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
- Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021. [Google Scholar]
- Agarwal, M.; Gupta, S.K.; Biswas, K. Development of Efficient CNN model for Tomato crop disease identification. Sustain. Comput. Inform. Syst. 2020, 28, 100407. [Google Scholar] [CrossRef]
- Gonzalez-Huitron, V.; León-Borges, J.A.; Rodriguez-Mata, A.; Amabilis-Sosa, L.E.; Ramírez-Pereda, B.; Rodriguez, H. Disease detection in tomato leaves via CNN with lightweight architectures implemented in Raspberry Pi 4. Comput. Electron. Agric. 2021, 181, 105951. [Google Scholar] [CrossRef]
- Tm, P.; Pranathi, A.; SaiAshritha, K.; Chittaragi, N.B.; Koolagudi, S.G. Tomato leaf disease detection using convolutional neural networks. In Proceedings of the 2018 11th International Conference on Contemporary Computing (IC3), Noida, India, 2–4 August 2018. [Google Scholar]
- Maeda-Gutiérrez, V.; Galván-Tejada, C.E.; Zanella-Calzada, L.A.; Celaya-Padilla, J.M.; Galván-Tejada, J.I.; Gamboa-Rosales, H.; Luna-Garcia, H.; Magallanes-Quintanar, R.; Guerrero Mendez, C.A.; Olvera-Olvera, C.A. Comparison of convolutional neural network architectures for classification of tomato plant diseases. Appl. Sci. 2020, 10, 1245. [Google Scholar] [CrossRef]
- Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
Output Size | Layer Name | DVT-Ti | DVT-S | DVT-B |
---|---|---|---|---|
LFEM | , 14, stride = 2 | , 16, stride = 2 | , 16, stride = 2 | |
Patch Aggregation | , 42, stride = 2 | , 48, stride = 2 | , 64, stride = 2 | |
Stage 1 | ConvNext Bottleneck LMHSA FFN | |||
Patch Aggregation | , 84, stride = 2 | , 96, stride = 2 | , 128, stride = 2 | |
Stage 2 | ConvNext Bottleneck LMHSA FFN | |||
Patch Aggregation | , 168, stride = 2 | , 192, stride = 2 | , 256, stride = 2 | |
Stage 3 | ConvNext Bottleneck LMHSA FFN | |||
Patch Aggregation | , 336, stride = 2 | , 384, stride = 2 | , 512, stride = 2 | |
Stage 4 | ConvNext Bottleneck LMHSA FFN | |||
FC | , 1280 | |||
Classifier | (8, 11) | |||
Params(M) | 8.06 | 14.05 | 26.20 | |
FLOPs(G) | 1.1 | 1.4 | 3.78 |
Category | Our Private | PlantVillage | 2018 AI Challenger |
---|---|---|---|
Bacterial spot | 555 | 2127 | 13 |
Early blight | 440 | 1000 | 792 |
Late blight | 545 | 1909 | 1569 |
Leaf mold | 480 | 952 | 755 |
Mosaic virus | 325 | 373 | 298 |
Healthy | 310 | 1591 | 1381 |
Spectoria leaf spot | 790 | 1771 | 1403 |
Yellow leaf curl virus | 375 | 5357 | 4442 |
Two-spotted spider mite | — | 1676 | 975 |
Target spot | — | 1404 | 74 |
Powery mildew | — | — | 1497 |
Total | 3820 | 18,160 | 13,199 |
Parameter | CyTrGAN | DVT |
---|---|---|
Optimization algorithm | Adam | Adam |
Batch size | 1 | 64 |
Learning rate | 0.0001 | 0.0001 |
Epochs | 200 | 100 |
Dropout rate | 0.5 | 0.5 |
Momentum | 0.9 | 0.9 |
RMSprop | 0.999 | 0.999 |
Class | Cyclegan/Resnet | Cyclegan/Unet | CyTrGAN |
---|---|---|---|
Bacterial spot | 115.26 | 94.18 | 86.87 |
Early blight | 131.44 | 121.17 | 91.50 |
Healthy | 97.51 | 160.18 | 91.72 |
Late blight | 177.40 | 176.92 | 166 |
Leaf mold | 193.11 | 178.70 | 161.70 |
Mosaic virus | 106.05 | 97.50 | 104 |
Septoria leaf spot | 108.15 | 103 | 100.45 |
Two-spotted spider mite | 91.53 | 97.62 | 94.70 |
Target spot | 187.06 | 93.63 | 92.13 |
Yellow leaf curl virus | 139.70 | 135.60 | 111.39 |
Botanist | R-R | R-G | G-R | G-G | ||||||
---|---|---|---|---|---|---|---|---|---|---|
Accuracy1 | 51% | 49% | 53% | 47% | ||||||
Accuracy2 | C0 | C1 | C2 | C3 | C4 | C5 | C6 | C7 | C8 | C9 |
96% | 97% | 100% | 96% | 94% | 98% | 98% | 97% | 97% | 95% | |
vision researcher | R-R | R-G | G-R | G-G | ||||||
Accuracy1 | 53% | 48% | 52% | 49% | ||||||
Accuracy2 | C0 | C1 | C2 | C3 | C4 | C5 | C6 | C7 | C8 | C9 |
95% | 95% | 100% | 98% | 96% | 99% | 98% | 95% | 95% | 97% |
Model | Accuracy (%) | Parameters (M) | FLOPs (MFLOPs) | Model Size (MB) |
---|---|---|---|---|
VGG16 | 96.46% | 138.36 | 15,470.26 | 198 |
VGG19 | 96.04% | 143.67 | 19,632.06 | 548 |
Resnet50 | 83.72% | 25.56 | 4133.74 | 97.7 |
Resnet101 | 81.44% | 44.55 | 7866.44 | 170 |
Mobilenetv2 | 95.76% | 3.50 | 327.55 | 13.6 |
Vision Transformer | 98.82% | 85.81 | 16,862.87 | 327 |
Swin Transformer [30] | 99.10% | 27.53 | 4371.13 | 105 |
DVT-Ti | 99.45% | 8.06 | 1103.28 | 31 |
DVT-S | 99.67% | 14.02 | 1950.51 | 54 |
DVT-B | 99.76% | 26.20 | 3781.19 | 100 |
Class | F1-Score | Precision | Recall | |
---|---|---|---|---|
Bacteria spot | 0 | 1.00 | 1.00 | 1.0 |
Early blight | 1 | 0.95 | 0.95 | 0.95 |
Healthy | 2 | 0.99 | 0.99 | 0.99 |
Late blight | 3 | 0.97 | 0.97 | 0.97 |
Leaf mold | 4 | 0.97 | 0.98 | 0.97 |
Powdery mildew | 5 | 1.00 | 1.00 | 1.00 |
Septoria leaf spot | 6 | 0.98 | 0.99 | 0.98 |
Spider mite | 7 | 0.96 | 0.95 | 0.95 |
Target spot | 8 | 0.71 | 0.86 | 0.77 |
Mosaic virus | 9 | 0.98 | 0.97 | 0.97 |
Yellow leaf curl virus | 10 | 0.99 | 0.99 | 0.99 |
Class Label | F1-Score | Precision | Recall | |
---|---|---|---|---|
Bacterial spot | 0 | 0.95 | 0.97 | 0.93 |
Early blight | 1 | 0.95 | 0.95 | 0.95 |
Late blight | 2 | 0.95 | 0.92 | 0.99 |
Leaf mold | 3 | 0.93 | 0.93 | 0.93 |
Mosaic virus | 4 | 0.95 | 0.95 | 0.95 |
Healthy | 5 | 0.95 | 0.94 | 0.94 |
Septoria spot | 6 | 0.97 | 0.99 | 0.99 |
Yellow leaf curl virus | 7 | 0.96 | 0.96 | 0.96 |
Label | Label | F1-Score | Precision | Recall |
---|---|---|---|---|
Bacterial spot | 0 | 1.00 | 1.00 | 1.00 |
Early blight | 1 | 0.98 | 0.98 | 0.98 |
Healthy | 2 | 0.98 | 0.98 | 0.98 |
Late blight | 3 | 1.00 | 0.99 | 0.98 |
Leaf mold | 4 | 1.00 | 1.00 | 1.00 |
Mosaic virus | 5 | 0.98 | 0.99 | 0.99 |
Septoria leaf spot | 6 | 0.99 | 0.97 | 0.98 |
Spider mites | 7 | 1.00 | 1.00 | 1.00 |
Target Spot | 8 | 1.00 | 1.00 | 1.00 |
Yellow leaf curl virus | 9 | 1.00 | 1.00 | 1.00 |
Dataset | Accuracy | Precision | Recall | F1-Score |
---|---|---|---|---|
8 classes | ||||
Origin dataset | 95.4% | 0.95 | 0.94 | 0.94 |
Origin dataset + synthetic mages | 97.60% | 0.97 | 0.97 | 0.97 |
10 classes | ||||
Origin dataset | 99.45% | 0.98 | 0.99 | 0.99 |
Origin dataset + synthetic mages | 99.65% | 0.99 | 0.99 | 0.99 |
11 classes | ||||
Origin dataset | 98.30% | 0.97 | 0.98 | 0.97 |
Origin dataset + synthetic mages | 98.91% | 0.99 | 0.98 | 0.98 |
Reference | Dataset | Image Count | Class | Accuracy (%) | Model Size (MB) |
---|---|---|---|---|---|
Agarwal et al. [31] | PlantVillage | 18,160 | 10 | 98.70 | 0.208 |
Gonzalez-Huitron et al. [32] | PlantVillage | 18,160 | 10 | 95.24 | 138.4 |
Tm et al. [33] | PlantVillage | 18,160 | 10 | 94.85 | 156.78 |
Maeda-Gutierrez et al. [34] | PlantVillage | 18,160 | 10 | 99.39 | 23.06 |
Abbas et al. [15] | PlantVillage | 16,012 | 10 | 97.11 | 27.58 |
Ahmed et al. [21] | PlantVillage | 18,160 | 10 | 99.30 | 9.6 |
Our method | PlantVillage | 18,160 | 10 | 99.45 | 29 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Chen, Z.; Wang, G.; Lv, T.; Zhang, X. Using a Hybrid Convolutional Neural Network with a Transformer Model for Tomato Leaf Disease Detection. Agronomy 2024, 14, 673. https://doi.org/10.3390/agronomy14040673
Chen Z, Wang G, Lv T, Zhang X. Using a Hybrid Convolutional Neural Network with a Transformer Model for Tomato Leaf Disease Detection. Agronomy. 2024; 14(4):673. https://doi.org/10.3390/agronomy14040673
Chicago/Turabian StyleChen, Zhichao, Guoqiang Wang, Tao Lv, and Xu Zhang. 2024. "Using a Hybrid Convolutional Neural Network with a Transformer Model for Tomato Leaf Disease Detection" Agronomy 14, no. 4: 673. https://doi.org/10.3390/agronomy14040673
APA StyleChen, Z., Wang, G., Lv, T., & Zhang, X. (2024). Using a Hybrid Convolutional Neural Network with a Transformer Model for Tomato Leaf Disease Detection. Agronomy, 14(4), 673. https://doi.org/10.3390/agronomy14040673