Next Article in Journal
Structural Particularities of Gall Neoformations Induced by Monarthropalpus flavus in the Leaves of Buxus sempervirens
Next Article in Special Issue
Effects of Different Postharvest Treatments on Fruit Quality, Sucrose Metabolism, and Antioxidant Capacity of ‘Newhall’ Navel Oranges During Storage
Previous Article in Journal
Urban Flora Biodiversity of Some Continental Cities of the Po Plain (Emilia-Romagna, Northern Italy)
Previous Article in Special Issue
Spatial Distribution Characteristics of Micronutrients and Their Deficiency Effect on the Root Morphology and Architecture in Citrus Rootstock
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hyperspectral Imaging and Machine Learning for Huanglongbing Detection on Leaf-Symptoms

1
Faculty of Informatics, Kansai University, Osaka 569-1095, Japan
2
Electrical Engineering and Computer Science, Tottori University, Tottori 680-8552, Japan
3
Independent Researcher, Tsukuba 300-1252, Japan
4
Royal Project Foundation, 910 Moo 3, T. Maehia, Muang, Chiang Mai 50200, Thailand
*
Author to whom correspondence should be addressed.
Plants 2025, 14(3), 451; https://doi.org/10.3390/plants14030451
Submission received: 6 January 2025 / Revised: 27 January 2025 / Accepted: 1 February 2025 / Published: 3 February 2025
(This article belongs to the Special Issue Innovative Techniques for Citrus Cultivation)

Abstract

:
Huanglongbing is one of the most destructive diseases of citrus worldwide. Infected trees die due to the absence of practical cures. Thus, the removal of HLB-infected trees is one of the principal HLB managements for the regulation of disease spread. Here, we propose a non-destructive HLB detection method based on hyperspectral leaf reflectance. In total, 72 hyperspectral leaf images were collected in an HLB-invaded citrus orchard in Thailand and each image was visually distinguished into either any HLB symptom appearance (symptomatic) or no symptoms (asymptomatic) on the leaf. Principal component analysis was applied on the hyperspectral data and revealed 16 key wavelengths at red-edge to near-infrared regions (715, 718, 721, 724, 727, 730, 733, 736, 930, 933, 936, 939, 942, 945, 957, and 997 nm) that were characteristically differentiated in the symptomatic group. Seven models learnt on the spectral data at these 16 wavelengths were examined for the potential to separate these two image groups: random forest, decision tree, support vector machine, k-nearest neighbor, gradient boosting, logistic regression, linear discriminant. F1-score was employed to select the best-fit model to distinguish the two categories: random forest achieved the best score of 99.8%, followed by decision tree and k-nearest neighbor. The reliability of the visual grouping was evaluated by nearest neighbor matching and permutation test. These three models separated the two image categories as precisely as PCR results, indicating their potential as alternative tool instead of PCR.

Graphical Abstract

1. Introduction

Huanglongbing (HLB), or citrus greening disease, is caused by the bacterium Candidatus Liberibacter asiaticus, which is transmitted between plants by the vector insect Diaphorina citri in Asian countries [1]. HLB is recognized as the most destructive disease of citrus worldwide. As no effective measures have been established yet [1], the current HLB management consists of three major measures: use of pathogen-free materials and sanitation of fields to plant trees, vector control, and removal of infected trees. The conventional detection of HLB is based primarily on regular visual inspection of trees and DNA polymerase chain reaction (PCR) [2]. Visual inspection needs skills to determine HLB infection, while PCR requires specialized laboratory equipment with much cost [3]. Alternative diagnostic measures that are cost-effective, efficient, and highly accurate are urgent issues to develop. Optical image analyses could be a potential candidate for HLB detection [4].
Diagnosing plant diseases with RGB image analyses, particularly with convolutional neural networks (CNNs), has been enthusiastically explored in this decade [4]. These analyses rely on HLB symptoms appearing on trees: yellow blotchy mottling, thickening, and enlarged or corky veins of leaves [5], and wilting, stunted growth, or yellow shoots of the canopy [1]. Kalim et al. [6] achieved 87% precision in distinguishing HLB from other five diseases with combining machine learning and deep learning on 594 leaf images with these symptoms. Elaraby et al. [7] reached 94.3% accuracy in diagnosing six citrus diseases with an SGDM+AlexNet model that used 4745 leaf images, while Gómez-Flores et al. [8] used a VGG19 learning model in which they transferred 953 leaf image data distinguished into HLB-infected and not infected groups. This model correctly identified the HLB infection with 95% precision. Deep learning with numerous RGB images thus achieves high precision in detecting the HLB-infected trees. However, various background environmental noises are also taken in the images and the elimination of such noises negatively interferes with the image analyses [9]. Furthermore, the success in image analyzing depends on how appropriately the targeted samples are photographed, being determined by the experience on visual disease inspection [10].
Finer spectral data with finer division of wavelengths than RGB contribute to successfully excluding environmental noises and provide more effective image data for crop health, pest, and soil assessment in agriculture [4]. This technique utilizes the electromagnetic intensity of spectra that are determined by the chemical and physical properties in plants [11]. Two measures are adopted, depending on the differences in the number of wavelengths: multispectral analyses in which 4–16 wavelengths are used and hyperspectral analyses that the optical data are collected at 200–1000 wavelengths divided at every 1 to 7 nm (Figure 1). Owing to its finer partitioning of wavelengths to analyze, hyperspectral imaging usually gives more accuracy in separating different groups, in particular detecting asymptomatic infection [12]. For example, multispectral analyses with three spectral regions (520–600 nm, 630–690 nm, 760–900 nm) succeed in detecting HLB in an accuracy > 90% with four machine learning techniques on the image data that were obtained from aerial viewing of an orchard by a drone [13]. On the other hand, Deng et al. [14] analyzed hyperspectral data and extracted 13 specific wavelengths using six machine learning models, successfully distinguishing HLB infection with a precision of 96%, being more sensitive by 6% than the former.
Although hyperspectral cameras may specify the disease infection in an early stage, including HLB, they are costly and impractical for large-scale use [15]. Confining the data in wavelengths that contribute to the differentiation in the reflectance properties in HLB-infected trees, the size-down in image data could be possible, and more handy equipment for HLB detection may be able to develop. Various authors have published a number of articles devoted to the determination of HLB using hyperspectral remote sensing and have attempted to extract some characteristic wavelengths (Table 1, [14,16,17,18,19,20,21,22,23,24]). In this paper, we propose statistical methods to efficiently confine useful wavelengths from hyperspectral data and machine learning models to identify HLB infection.

2. Results

2.1. PCA for the Explanation of the Variance in the Data

The contributions of the first eight principal components (PCs) revealed by the PCA (principal component analysis) were 83.44%, 12.83%, 2.17%, 0.84%, 0.14%, 0.08%, 0.06%, and 0.04%, respectively (Figure 2a). Since the first four PCs explained over 99% of the variance in the data, we output their loading plot as shown in Figure 2b. The loadings represent the contribution of each wavelength to the distinction of image groupings, asymptomatic and symptomatic. Figure 2b shows relatively large loading variations in the 500–600 nm, near 700 nm, and 900–1000 nm regions. This suggests that the reflectance properties of the foliar hyperspectral data vary the most at these wavelengths, which may be critical for HLB detection classification. The loading curve for PC1 was smoother and usually represented the main differences in overall reflection or intensity. PC2–4 explained smaller proportions of the variance and showed more peaks and troughs, often representing minor but still important differences or possibly noise.
We then compared the distribution of asymptomatic and symptomatic data on PC1–4 two by two and output their scatter plots (Figure 3). Regardless of which two PCs are compared, there is a large overlap region, which indicates that PCA only is not yet able to separate the differences between the two categories well. Further wavelength optimization may be needed to exclude some redundant information. We analyzed the top eight wavelengths of PC1–4 and their loadings as shown in Table 2. Wavelengths closely regressed on PC1 in the range 715–736 nm, which corresponded to the red-edge region of citrus leaves. On PC2, the wavelengths at the 930–960 nm range were regressed, expanding over the near-infrared region. Wavelengths on PC3 included shorter wavelengths, 400, 710, and 933 nm, while on PC4 near infrared wavelengths 933–957 nm were related, mostly overlapping with PC2. Since PC1 and PC2 explained more than 96% of the variance and there was no overlap in the top eight wavelengths, a combination of these two PCs was chosen for further wavelength optimization. The PC1 score of the symptomatic group was negatively correlated against the PC2, while asymptotic groups were scattered near the origin in the coordinates with no apparent correlation.

2.2. Hyperspectral Data Wavelength Optimization

To further refine HLB characteristic wavelengths, we applied the wavelength optimization method based on incremental feature selection. By sequentially increasing the number of selected wavelengths according to their loading scores, we obtained a total of 64 wavelength combinations. These combinations were then input into the seven models for training and the F1 scores were compared for three conditions: full wavelengths, feature extraction, and wavelength optimization.
Nonlinear models (random forest (RF), decision tree, k-nearest neighbor (KNN), gradient boosting, and support vector machine (SVM)) consistently achieved F1 scores above 95%, while linear models (linear discriminant analysis (LDA) and logistic regression) performed poorly, with F1 scores below 40% after data reduction (Table 3). Among the nonlinear models, only RF, decision tree, KNN, and gradient boosting maintained high F1 scores (>95%) after wavelength optimization. The RF model integrating nine wavelengths achieved the highest F1 score of 99.8%, and the gradient boosting model integrating twelve wavelengths achieved the lowest F1 score of 96.2%. We took these four models and discussed them further in terms of their reliability.

2.3. Reliability of Expert System

A PCR-based study relied on precise PCR results to build a hyperspectral dataset of foliar leaves, which was learned and tested using six different machine learning models and achieved 96% accuracy on a 13-wavelength combination. As the four nonlinear models mentioned above performed well after data reduction, we conducted nearest neighbor matching for comparing wavelengths extracted in wavelength optimization by them with this 13-wavelength combination reported in the PCR-based study (Table 4). The discrepancies between PCR-based and model-selected wavelengths were 2.78 nm for RF, 3.14 nm for decision tree, 3.5 nm for KNN, and 5.0 nm for gradient boosting (Figure 4). Owing to their discrepancies smaller than the hyperspectral camera’s resolution of 7 (0 ± 3.5 nm), the RF, decision tree, and KNN models can be taken as effective alternatives to the PCR test for HLB detection. Therefore, these three models were selected for further analyses.
We analyzed the F1 scores, wavelengths used, and test time of these three models (Table 5). The decision tree with seven wavelengths completed the calculation in the shortest time, 7 ms, but the precision was comparable to that of the RF model. Although the KNN model required the longest time in the calculation, it utilized only four wavelengths with keeping the F1-score higher. These findings indicate that the RF, decision tree, and KNN models in which red to infrared reflectance data are integrated can recognize HLB infection with more than 97% precision. In particular, the RF and decision tree models almost perfectly separated image groups.
The availability of models may be considered in three aspects: precision in HLB determination, quickness in calculation time, and the reduction of used wavelengths (Figure 5). If focusing only on the precision, the RF model will be selected. On the other hand, if calculation time is the priority, the decision tree model should be the first choice. If simpler equipment is preferred, the KNN model should be considered because it uses only four wavelengths, which may be easily integrated into the multispectral analyzing system. Since all three models have high-level precision while the decision tree model has an outstanding processing speed, we would recommend prioritizing this model.
Permutation tests were conducted on these three models selected above to validate the statistical significance in their HLB determination. We randomly shuffled the true labels of the test set 10,000 times and evaluated them using the same trained models. All three models were significant (p < 0.01), meaning that RF, decision tree, and KNN models all performed HLB detection significantly better than random performance.

3. Discussion

The RF, decision tree, and KNN models indicated that red-edge to infrared wavelengths at 727 to 960 nm significantly contributed to determine leaves’ HLB symptoms from symptomatic leaves (Table 4). The 727 nm wavelength lies on the red-edge of visual light, which is typically closely related to the chlorophyll a content in plant leaves and their photosynthetic activities [25,26]. HLB infection disrupts chlorophyll metabolism, leading the optical characteristics at this wavelength to change and, accordingly, other waves around it also may be modified. Wavelengths at 930–960 nm fall within the near-infrared region and are associated with variations in leaf water content and cellular structure [27]. The results in this present study indicate that wavelengths differentiated in these bands may reflect the changes in either or both characters in HLB-infected leaves. The three models selected in our study thus could be integrated into the HLB-infection system as more cost-effective and efficient sensing technologies.
Meanwhile, there are some limitations in our study as well. We chose the study of Deng et al. [14] for comparison because their equipment, imaging conditions, and feature selection methods are similar to ours. This alignment can provide a more meaningful and fair comparison to highlight the reliability and applicability of our approach. However, accurate PCR tests should have been performed on the collected leaf samples to provide a more objective corroboration of the results of the expert inspection. Although our models successfully achieved almost perfect distinction between symptomatic and asymptomatic leaves with the integration of the two major components, in some cases, minor components often make important contributions to category-distinction. Ignorance of such minor components sometimes reduces the performance of models [28]. The existence of a partial overlap in the scatter plot of PC1 and PC2 (Figure 3a) suggests that leaf refraction in some HLB-infected leaves could not be differentiated from non-infected leaves based solely on the wavelengths explained by PC1. This insufficient separation in the HLB-infected leaves may be apparent in the early stages of infection, where few wavelengths lead noticeable changes in this stage. Such incomplete differentiation, if any, would result in the failing in the distinction of HLB infection by these three models. Therefore, the overlap might be attributed to how long the leaf has been infected by HLB: the shorter the time since the HLB infection, the less the wavelength traits, making early differentiation more challenging. Minor components other than PC1 and PC2 may contribute to distinguishing early infected leaves, which deserves further investigation. In addition, a mentioned nonlinear dimensionality reduction technique, t-SNE, has shown better results than PCA in feature extraction of some hyperspectral data [29,30]. We also plan to try to use this method to solve the overlapping problem mentioned above.
Our expert system is trained with images taken from multiple leaves and applied to the real-world recognition. Thus, multiple leaves can be determined into a presumed category at once by taking similar images. Compared to models that require taking pictures of one leaf at a time, the detection efficiency can be improved several times more precisely. Nonetheless, it is still inevitable to reach an incorrect determination or perform the determination less effectively. Recently, eye-tracking studies reveal experts assess HLB by scanning broader areas, suggesting the use of holistic cues not captured by current deep learning models [31]. Given the successful application of attention mechanisms in deep learning models in recent years [32], we propose that integrating expert diagnostic strategies can enhance model performance. By learning from the judgment experience of seasoned experts, deep learning models can establish key attention areas on the whole tree, improving learning efficiency and diagnostic accuracy. This approach has the potential to simplify the HLB diagnosis process and contribute to more effective disease management strategies, which is our future work.

4. Materials and Methods

Figure 6 illustrates our data processing flow, where foliar images were obtained with a hyperspectral camera (Specim IQ, SPECIM, Oulu, Finland) of 204 wavelengths partitioned at an interval of about 7 nm at 400–1000 nm. Based on expert visual diagnosis, hyperspectral pixel data were divided into two categories: any symptoms on leaves (symptomatic) or no symptom (asymptomatic). The pixel data were randomly split into training (80%) and test (20%) sets by label proportions. PCA was applied on every individual wavelength to extract specific wavelengths that were characteristic in foliar symptomatic appearances. PCA was applied to the training set only to avoid overfitting due to data leakage [33]. Then, an incremental feature selection method was conducted to explore the possible combinations of the PCA-determined key wavelengths, and the optimal wavelength combinations were determined on each of the seven machine learning models based on the F1-score metric. These models included five nonlinear models: RF, decision tree, SVM, KNN, and gradient boosting, and two linear models: logistic regression and LDA. Finally, nearest neighbor matching and permutation test methods were used to compare the HLB characteristic wavelengths extracted by the expert visual inspection and PCR-based detection to validate the reliability of our expert system.

4.1. Hyperspectral Image Data Collection

All images in this study were obtained in a citrus orchard, approximately 490 m2 located in Kuet Chang Sub-district, Mae Taeng District, Chiang Mai, Thailand (Figure 7). In this orchard, 7-year-old trees of the local tangerine cultivar, Sai Num Phung (Citrus reticulata), were cultivated, spaced 4 m from each other on a row and 3 m between rows.
Hyperspectral images were collected at 10:30 to 11:30 am on a sunny morning on 7 August 2024, with the hyperspectral camera fixed on a tripod. The distance between the front lens and imaged leaves was 40 cm and a neutral density filter (Ni060-40.5, MIDOPT, Palatine, IL, USA) was attached on the top of the objective lens to reduce light intensity. This filter reduced light intensity which would adversely affect image captures. A white reference panel (IQ White Reference, SPECIM, Oulu, Finland) was set beside plants for reflectance calibration every time the plants were imaged (Figure 8). Table 6 shows the devices and their specifications in this study. A total of 22 branch images were collected from 10 trees in this study.

4.2. Hyperspectral Data Preprocessing

A python-based preprocessing software developed by us was utilized to eliminate background noises in the images and extract area needed for the analyses using HSV (hue, saturation, value) thresholds and manual selection (Figure 9). This software extracted spectral data of 72 leaves in the 22 branch images. Expert visual assessment identified HLB symptoms in four out of the ten trees, resulting in 12 symptomatic and 60 asymptomatic leaves. All pixels from symptomatic leaves were labeled as “symptomatic” and those from asymptomatic leaves as “asymptomatic”. We eventually extracted 63,056 “symptomatic” pixels and 294,886 “asymptomatic” pixels. Considering the sample imbalance, we used stratified sampling in dividing the dataset, 80% from each of “symptomatic” and “asymptomatic” as the training set and the remaining 20% as the test set.

4.3. Hyperspectral Data Analysis and Modeling

4.3.1. Feature Extraction—PCA

PCA is a statistical technique used to simplify complex datasets while retaining as much important information as possible, widely used in fields like image processing, genetics, and finance to reduce data complexity and uncover meaningful patterns [34]. In this study, it was used to focus on wavelengths that could have differentiated optical characteristics of leaf reflectance due to the HLB symptoms. The PCs were created and calculated, and the proportion of variance explained by each component was used as a criterion for selecting key features. Often, just a few of these new components capture most of the information in the dataset. By focusing on these few components, the dataset can be simplified while preserving its essence. Before performing PCA, all spectral data were normalized to ensure uniform scaling across wavelengths.

4.3.2. Wavelength Optimization—Incremental Feature Selection

Since 16 spectral are the maximum wavelengths available for multispectral cameras, the extraction of wavelengths in the hyperspectral data was limited to or fewer than 16 wavelengths. However, 16 wavelengths rise to 216 − 1 = 65,535 combinations, except for the null combination, which impose heavy computational burden. In this study, this problem was avoided by adopting an incremental feature selection approach to generate targeted wavelength combinations (Table 7). This approach aligns with the data-driven insights provided by PCA, focusing on the most informative wavelengths from PCs rather than all available features [35]. Specifically, this method systematically explores combinations by gradually increasing the number of wavelengths from each component, ensuring a balanced representation of features while maintaining a logical structure.

4.3.3. Machine Learning Models

Seven machine learning models were assessed for how precisely they distinguished images in two groups, symptomatic or asymptomatic. Models examined in this study were five nonlinear models (RF [36], decision tree [37], KNN [38], gradient boosting [39], and SVM [40]) and two linear models (logistic regression [41] and LDA [42]). These methods are well-established in the current machine learning and were chosen due to their classification capabilities [14].

4.4. Evaluation of Models for Leaf Image Separation

4.4.1. F1 Score

F1 score was used to evaluate the precision of machine learning models in separating image data into the groups. This is a powerful tool for imbalanced datasets, as it adopts the harmonic mean of precision and recall, and effectively regulates both false positives and false negatives to provide a comprehensive performance measure [43]. The F1 score is calculated through the following three equations:
P r e c i s i o n   ( % ) = T P T P + F P × 100
R e c a l l   ( % ) = T P T P + F N × 100
F 1 s c o r e   ( % ) = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l × 100
where TPs (True Positives) means correctly identified positive cases; TNs (True Negatives) means correctly identified negative cases; FPs (False Positives) means negative cases incorrectly identified as positive; FNs (False Negatives) means positive cases incorrectly identified as negative [44].

4.4.2. Nearest Neighbor Matching

Nearest neighbor matching is a statistical technique commonly used in observational studies to pair each unit in a treatment group with the closest unit in a control group based on a predefined set of covariates or characteristics [45]. Given two wavelength sets, say A = {a1, a2, … an} and B = {b1, b2, …, bm}, the matching calculates the mean of the differences between each wavelength in A and its nearest wavelength in B as a measure of similarity:
D i s t a n c e = 1 n i = 1 n min a i b j   f o r   e a c h   a i A , b j B
To validate the effectiveness of our expert’s visual inspection, we used nearest neighbor matching to compare the HLB characteristic wavelengths extracted by our models with those reported in a PCR-based study. That study utilized hyperspectral imaging based on PCR testing to analyze the optimal wavelength combinations, and 13 HLB characteristic wavelengths (544, 718, 753, 760, 764, 930, 938, 943, 951, 969, 985, 998, and 999 nm) achieved the best performance of an accuracy of 96% [14]. Using nearest neighbor matching, we calculated the distance errors between the wavelengths extracted by each model and those from the PCR-based study.

4.4.3. Permutation Test

The permutation test is a robust, non-parametric statistical method which assesses whether the performance of a machine learning model is significantly better than random performance [46]. The procedure involves shuffling target labels to break the relationship between the input feature and the target variable, followed by re-evaluating the model on the permuted dataset multiple times (at least 1000 times is required for reliability) [46]. In this study, we randomly shuffled the true labels of the test set and evaluated the shuffled labels using the same trained models. This process was repeated 10,000 times. The frequencies of F1 scores greater than or equal to the scores achieved with the true labels were counted and the model’s statistical significance in the HLB determination was examined at p < 0.01 [46].

5. Conclusions

The following conclusions can be drawn from the study:
-
Our RF, decision tree, and KNN models are as reliable as PCR in identifying HLB.
-
Nonlinear models outperform linear models for HLB spectral data.
-
Using PCA for nonlinear models is effective for HLB feature extraction.
-
Decision tree model provides high accuracy with faster prediction, suitable for real-time applications.
-
KNN model shows promising potential for multispectral imaging applications.
-
The red-edge and near-infrared regions may be critical for HLB detection.
Future work will focus on applying these findings to multispectral cameras and analyzing expert visual patterns to develop deep learning models for whole-tree HLB diagnosis.

Author Contributions

Conceptualization, R.D., K.I. and T.H.; methodology, R.D. and K.I.; software, R.D.; validation, R.D., K.I. and A.P.; formal analysis, R.D., A.S. and K.I.; investigation, R.D., K.I., A.P., K.S., S.M. and T.H.; resources, R.D., K.S., S.M. and T.H.; data curation, R.D. and A.S.; writing—original draft preparation, R.D.; writing—review and editing, K.I. and T.H.; visualization, R.D. and A.S.; supervision, A.S. and K.I.; project administration, A.P. and T.H.; funding acquisition, A.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by JSPS KAKENHI in Japan, Grant Number JP22K12075.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. Software: PyCharm Community Edition 2024.2.1 for hyperspectral data analysis and machine learning modeling; Anaconda3 for managing library files.

Acknowledgments

We would like to express our gratitude to the Highland Research and Development Institute (Thailand) for the guidance and field applied for our data collection.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

CNN (convolutional neural network); HLB (Huanglongbing); KNN (k-nearest neighbors); LDA (linear discriminant analysis); PC (principal component); PCA (principal component analysis); PCR (polymerase chain reaction); RF (random forest); SVM (support vector machine).

References

  1. Tipu, M.M.H.; Masud, M.M.; Jahan, R.; Baroi, A.; Hoque, A.K.M.A. Identification of citrus greening based on visual symptoms: A grower’s diagnostic toolkit. Heliyon 2021, 7, e08387. [Google Scholar] [CrossRef] [PubMed]
  2. Deng, X.; Lan, Y.; Hong, T.; Chen, J. Citrus greening detection using visible spectrum imaging and C-SVC. Compt. Electron. Agric. 2016, 130, 177–183. [Google Scholar] [CrossRef]
  3. Cap, H.Q.; Suwa, K.; Fujita, E.; Kagiwada, S.; Uga, H.; Iyatomi, H. A deep learning approach for on-site plant leaf detection. In Proceedings of the 2018 IEEE 14th International Colloquium on Signal Processing & Its Applications, Penang, Malaysia, 9–10 March 2018; pp. 18–122. [Google Scholar]
  4. Yang, D.; Wang, F.; Hu, Y.; Lan, Y.; Deng, X. Citrus huanglongbing detection based on multi-modal feature fusion learning. Front. Plant Sci. 2021, 12, 809506. [Google Scholar] [CrossRef] [PubMed]
  5. Gottwald, T.R. Current epidemiological understanding of citrus Huanglongbing. Annu. Rev. Phytopathol. 2010, 48, 119–139. [Google Scholar] [CrossRef]
  6. Kalim, H.; Chug, A.; Singh, A.P. Citrus leaf disease detection using hybrid CNN-RF model. In Proceedings of the 4th International Conference on Artificial Intelligence and Speech Technology (AIST), Delhi, India, 9–10 December 2022; pp. 1–4. [Google Scholar]
  7. Elaraby, A.; Hamdy, W.; Alanazi, S. Classification of citrus diseases using optimization deep learning approach. Comput. Intell. Neurosci. 2022, 2022, 9153207. [Google Scholar] [CrossRef] [PubMed]
  8. Gómez-Flores, W.; Garza-Saldaña, J.J.; Varela-Fuentes, S.E. A huanglongbing detection method for orange trees based on deep neural networks and transfer learning. IEEE Access 2022, 10, 116686–116696. [Google Scholar] [CrossRef]
  9. Liu, X.; Min, W.; Mei, S.; Wang, L.; Jiang, S. Plant disease recognition: A large-scale benchmark dataset and a visual region and loss reweighting approach. IEEE Transact. Image Process. 2021, 30, 2003–2015. [Google Scholar] [CrossRef]
  10. Xia, F.; Xie, X.; Wang, Z.; Jin, S.; Yan, K.; Ji, Z. A novel computational framework for precision diagnosis and subtype discovery of plant with lesion. Front. Plant Sci. 2022, 12, 789630. [Google Scholar] [CrossRef] [PubMed]
  11. Li, L.; Zhang, Q.; Huang, D. A Review of Imaging Techniques for Plant Phenotyping. Sensors 2014, 14, 20078–20111. [Google Scholar] [CrossRef] [PubMed]
  12. Wang, K.; Dongmei, G.; Zhang, Y.; Deng, L.; Xie, R.; Lv, Q.; Yi, S.; Zheng, Y.; Ma, Y.; He, S. Detection of Huanglongbing (citrus greening) based on hyperspectral image analysis and PCR. Front. Agric. Sci. Eng. 2019, 6, 172–180. [Google Scholar] [CrossRef]
  13. Lan, Y.; Huang, Z.; Deng, X.; Zhu, Z.; Huang, H.; Zheng, Z.; Lian, B.; Zeng, G.; Tong, T. Comparison of machine learning methods for citrus greening detection on UAV multispectral images. Compt. Electron. Agric. 2020, 171, 105234. [Google Scholar] [CrossRef]
  14. Deng, X.; Huang, Z.; Zheng, Z.; Lan, Y.; Dai, F. Field detection and classification of citrus Huanglongbing based on hyperspectral reflectance. Compt. Electron. Agric. 2019, 167, 105006. [Google Scholar] [CrossRef]
  15. Podlesnykh, I.; Kovalev, M.; Platonov, P. Towards the future of ubiquitous hyperspectral imaging: Innovations in sensor configurations and cost reduction for widespread applicability. Technologies 2024, 12, 221. [Google Scholar] [CrossRef]
  16. Sankaran, S.; Mishra, A.; Maja, J.M.; Ehsani, R. Visible-near infrared spectroscopy for detection of Huanglongbing (HLB) Using a VIS-NIR Spectroscopy Technique. Comput. Electron. Agric. 2011, 77, 127–134. [Google Scholar] [CrossRef]
  17. Sankaran, S.; Ehsani, R. Visible-near infrared spectroscopy based citrus greening detection: Evaluation of spectral feature extraction techniques. Crop Prot. 2011, 30, 1508–1513. [Google Scholar] [CrossRef]
  18. Sankaran, S.; Maja, J.; Buchanon, S.; Ehsani, R. Huanglongbing (Citrus Greening) Detection Using Visible, Near Infrared and Thermal Imaging Techniques. Sensors 2013, 13, 2117–2130. [Google Scholar] [CrossRef] [PubMed]
  19. Li, X.; Lee, W.S.; Li, M.; Ehsani, R.; Mishra, A.R.; Yang, C.; Mangan, R.L. Spectral difference analysis and airborne imaging classification for citrus greening infected trees. Comput. Electron. Agric. 2012, 83, 32–46. [Google Scholar] [CrossRef]
  20. Kumar, A.; Lee, W.S.; Ehsani, R.J.; Albrigo, L.G.; Yang, C.; Mangane, R.L. Citrus greening disease detection using aerial hyperspectral and multispectral imaging techniques. J. Appl. Remote Sens. 2012, 6, 063542. [Google Scholar]
  21. Weng, H.; Lu, J.; Cen, H.; He, M.; Zeng, Y.; Hua, S.; Li, H.; Meng, Y.; Fang, H.; He, Y. Hyperspectral reflectance imaging combined with carbohydrate metabolism analysis for diagnosis of citrus Huanglongbing in different seasons and cultivars. Sens. Actuators B Chem. 2018, 275, 50–60. [Google Scholar] [CrossRef]
  22. Deng, X.; Zhu, Z.; Yang, J.; Zheng, Z.; Huang, Z.; Yin, X.; Wei, S.; Lan, Y. Detection of Citrus Huanglongbing Based on Multi-Input Neural Network Model of UAV HRS. Remote Sens. 2020, 12, 2678. [Google Scholar] [CrossRef]
  23. Menezes, J.; Dharmalingam, R.; Shivakumara, P. HLB Disease Detection in Omani Lime Trees Using Hyperspectral Imaging Based Techniques. In Recent Trends in Image Processing and Pattern Recognition; Springer Nature: Cham, Switzerland, 2024; pp. 67–81. [Google Scholar]
  24. Yan, K.; Song, X.; Yang, J.; Xiao, J.; Xu, X.; Guo, J.; Zhu, H.; Lan, Y.; Zhang, Y. Citrus huanglongbing detection: A hyperspectral data-driven model integrating feature band selection with machine learning algorithms. Crop Prot. 2025, 188, 107008. [Google Scholar] [CrossRef]
  25. Carter, G.A.; Knapp, A.K. Leaf optical properties in higher plants: Linking spectral characteristics to stress and chlorophyll concentration. Am. J. Bot. 2001, 88, 677–684. [Google Scholar] [CrossRef] [PubMed]
  26. Ustin, S.L.; Jacquemoud, S. How the optical properties of leaves modify the absorption and scattering of energy and enhance leaf functionality. In Remote Sensing of Plant Biodiversity; Cavender-Bares, J., Gamon, J.A., Townsend, P.A., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 349–384. [Google Scholar]
  27. Peñuelas, J.; Inoue, Y. Reflectance indices indicative of changes in water and pigment contents of peanut and wheat leaves. Photosynthetica 1999, 36, 355–360. [Google Scholar] [CrossRef]
  28. Zheng, J.; Rakovski, C. On the Application of Principal Component Analysis to Classification Problems. Data Sci. J. 2021, 20, 26. [Google Scholar] [CrossRef]
  29. Silva, R.; Melo-Pinto, P. t-SNE: A study on reducing the dimensionality of hyperspectral data for the regression problem of estimating oenological parameters. Artif. Intell. Agric. 2023, 7, 58–68. [Google Scholar] [CrossRef]
  30. Kang, Z.; Fan, R.; Zhan, C.; Wu, Y.; Lin, Y.; Li, K.; Qing, R.; Xu, L. The Rapid Non-Destructive Differentiation of Different Varieties of Rice by Fluorescence Hyperspectral Technology Combined with Machine Learning. Molecules 2024, 29, 682. [Google Scholar] [CrossRef]
  31. Kobayash, O.; Dong, R.; Shiraiwa, A. Eye-tracking data analysis for improving diagnosis performance in citrus greening disease. IEICE Tech. Rep. 2024, 124, 35–38. [Google Scholar]
  32. Dong, R.; Shiraiwa, A.; Pawasut, A.; Sreechun, K.; Hayashi, T. Diagnosis of Citrus Greening Using Artificial Intelligence: A Faster Region-Based Convolutional Neural Network Approach with Convolution Block Attention Module-Integrated VGGNet and ResNet Models. Plants 2024, 13, 1631. [Google Scholar] [CrossRef]
  33. Rosenblatt, M.; Tejavibulya, L.; Jiang, R.; Noble, S.; Scheinost, D. Data leakage inflates prediction performance in connectome-based machine learning models. Nat. Comm. 2024, 15, 1829. [Google Scholar] [CrossRef]
  34. Jolliffe, I. Principal component analysis. In International Encyclopedia of Statistical Science; Lovric, M., Ed.; Springer: Berlin/Heidelberg, Germany, 2011; pp. 1094–1096. [Google Scholar]
  35. Liu, H.; Setiono, R. Incremental feature selection. Appl. Intell. 1998, 9, 217–230. [Google Scholar] [CrossRef]
  36. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  37. Breiman, L.; Friedman, J.; Olshen, R.A.; Stone, C.J. Classification and Regression Trees, 1st ed.; Chapman and Hall/CRC: Boca Raton, FL, USA, 1984. [Google Scholar] [CrossRef]
  38. Cover, T.; Hart, P. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef]
  39. Jerome, H. Friedman. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  40. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  41. Kleinbaum, D.G. Logistic Regression: A Self-Learning Text; Springer: Berlin/Heidelberg, Germany, 1994. [Google Scholar]
  42. Liu, Z.P. Linear Discriminant Analysis. In Encyclopedia of Systems Biology; Dubitzky, W., Wolkenhauer, O., Cho, K.H., Yokota, H., Eds.; Springer: New York, NY, USA, 2013; pp. 1132–1133. [Google Scholar] [CrossRef]
  43. Powers, D. Evaluation: From precision, recall and F-measure to ROC, informedness, markedness & correlation. J. Mach. Learn. Technol. 2011, 2, 37–63. [Google Scholar]
  44. Fawcett, T. An introduction to ROC analysis. Pattern Recogn. Lett. 2006, 27, 861–874. [Google Scholar] [CrossRef]
  45. Clark, P.J.; Evans, F.C. Distance to Nearest Neighbor as a Measure of Spatial Relationships in Populations. Ecology 1954, 35, 445–453. [Google Scholar] [CrossRef]
  46. Ojala, M.; Garriga, G.C. Permutation tests for studying classifier performance. In Proceedings of the 2009 Ninth IEEE International Conference on Data Mining, Miami, FL, USA, 6–9 December 2009; pp. 908–913. [Google Scholar]
Figure 1. The schematic of RGB, multispectral, and hyperspectral images.
Figure 1. The schematic of RGB, multispectral, and hyperspectral images.
Plants 14 00451 g001
Figure 2. The contributions of the first eight principal components (a) and the loading plot for principal components 1–4 (b).
Figure 2. The contributions of the first eight principal components (a) and the loading plot for principal components 1–4 (b).
Plants 14 00451 g002
Figure 3. The scatter plots of different PCs between PC1 and PC4 (af).
Figure 3. The scatter plots of different PCs between PC1 and PC4 (af).
Plants 14 00451 g003aPlants 14 00451 g003b
Figure 4. Comparison of key wavelengths identified by the PCR-based study and our expert-based RF model.
Figure 4. Comparison of key wavelengths identified by the PCR-based study and our expert-based RF model.
Plants 14 00451 g004
Figure 5. Normalized performance comparison of classification models with F1 score, reversed test time, and reversed wavelengths used.
Figure 5. Normalized performance comparison of classification models with F1 score, reversed test time, and reversed wavelengths used.
Plants 14 00451 g005
Figure 6. The flowchart of hyperspectral data analysis and modeling in this study.
Figure 6. The flowchart of hyperspectral data analysis and modeling in this study.
Plants 14 00451 g006
Figure 7. The orchard in Thailand for this study (19°14′02″ N 98°53′14″ E, 611 m AMSL). This image is available from Google Earth. The Thai word in the figure is the local place name.
Figure 7. The orchard in Thailand for this study (19°14′02″ N 98°53′14″ E, 611 m AMSL). This image is available from Google Earth. The Thai word in the figure is the local place name.
Plants 14 00451 g007
Figure 8. RGB images of branches in symptomatic (a) and asymptomatic (b) groups created from the composite of hyperspectral images.
Figure 8. RGB images of branches in symptomatic (a) and asymptomatic (b) groups created from the composite of hyperspectral images.
Plants 14 00451 g008
Figure 9. Schematic of our preprocessing software. First, the region of interest is manually selected (blue region) (a). Background noise is then largely removed by HSV thresholding (b), with manual adjustment for finer noise removal (c). The blue boxes can be used to control the threshold in (b) and the size of the pixels when manually removing in (c). Finally, hyperspectral data of the selected region are extracted and saved.
Figure 9. Schematic of our preprocessing software. First, the region of interest is manually selected (blue region) (a). Background noise is then largely removed by HSV thresholding (b), with manual adjustment for finer noise removal (c). The blue boxes can be used to control the threshold in (b) and the size of the pixels when manually removing in (c). Finally, hyperspectral data of the selected region are extracted and saved.
Plants 14 00451 g009
Table 1. Previous studies using hyperspectral remote sensing for HLB detection.
Table 1. Previous studies using hyperspectral remote sensing for HLB detection.
Publication YearEquipment Wavelengths (nm)Feature Wavelengths (nm)Study Type
2011–2013 [16,17,18]350–2500537, 612, 638, 662, 688, 713, 763, 813, 998, 1066, 1120, 1148, 1296, 1445, 1472, 1546, 1597, 1622, 1746, 1898, 2121, 2172, 2348, 2471, 2493field
2012 [19]457–921650–850field and lab
2012 [20]457–921410–432, 440–509, 634–686,
734–927, 932, 951, 975, 980
field
2018 [21]379–1023493, 515, 665, 716, 739lab
2019 [14]400–1000544, 718, 753, 760, 764, 930, 938, 943, 951, 969, 985, 998, 999field
2020 [22]450–950,
325–1075
468, 504, 512, 516, 528, 536, 632, 680, 688, 852field
2024 [23]400–1000560, 678, 726, 750lab
2025 [24]325–1075375–425, 650–750, 890–925lab
Table 2. Top eight wavelengths of principal components 1–4 and their loadings.
Table 2. Top eight wavelengths of principal components 1–4 and their loadings.
PC1PC2PC3PC4
WavelengthLoadingWavelengthLoadingWavelengthLoadingWavelengthLoading
7270.07419330.11659330.13249450.1764
7240.07419360.11114000.13039420.1686
7300.07409300.1040709−0.12769330.1683
7210.07399570.1032712−0.12719480.1648
7330.07399420.10319360.12609510.1636
7180.07389390.10254030.12549540.1400
7360.07379450.1025706−0.12479570.1359
7150.07359970.1025715−0.12329360.1312
Table 3. F1 scores of seven classification models with symptomatic and asymptomatic datasets for full wavelengths, feature extraction, and wavelength optimization.
Table 3. F1 scores of seven classification models with symptomatic and asymptomatic datasets for full wavelengths, feature extraction, and wavelength optimization.
Classification ModelFull Wavelengths
(204 Wavelengths)
Feature Extraction
(16 Wavelengths)
Wavelength Optimization
(No. of Wavelengths)
NonlinearRF99.5%99.6%99.8% (9)
Decision tree99.1%99.2%99.3% (6)
KNN98.8%96.3%97.9% (4)
Gradient boosting97.7%95.6%96.2% (12)
SVM98.4%89.2%89.2% (16)
LinearLDA89.3%38.8%38.8% (15)
Logistic regression91.8%38.3%38.4% (14)
Table 4. Extracted wavelengths by RF, decision tree, KNN, and gradient boosting models and their wavelength discrepancies between the study by Deng et al. [14].
Table 4. Extracted wavelengths by RF, decision tree, KNN, and gradient boosting models and their wavelength discrepancies between the study by Deng et al. [14].
Classification ModelExtracted Wavelengths (nm)Discrepancies
Deng et al. [14]544, 718, 753, 760, 764, 930, 938, 943, 951, 969, 985, 998, 999
RF727, 930, 933, 936, 939, 942, 945, 957, 9972.78
Decision tree727, 930, 933, 936, 939, 942, 9573.14
KNN727, 930, 933, 9363.5
Gradient boosting721, 724, 727, 730, 733, 930, 933, 936, 939, 942, 945, 9575.0
Table 5. The F1-scores, wavelengths used and test time achieved by RF, decision tree and KNN models.
Table 5. The F1-scores, wavelengths used and test time achieved by RF, decision tree and KNN models.
Classification ModelF1-ScoreWavelengths UsedTest Time (ms)
RF99.8%9869
Decision tree99.3%77
KNN97.9%4899
Table 6. Specification of the hyperspectral camera used in this study.
Table 6. Specification of the hyperspectral camera used in this study.
DeviceSpecificationValue
Specim IQResolution512 × 512 pix
Wavelength range (204)397–1004 nm
Dimension207 × 91 × 74 mm
Pixel size17.58 μm × 17.58 μm
Calibration whiteboardReflectivity100%
Size10 × 10 cm
Neutral density filterAverage Transmission25%
Table 7. An example of an incremental feature selection approach focusing on the first two components, the primary and secondary principal components, PC1 and PC2, respectively.
Table 7. An example of an incremental feature selection approach focusing on the first two components, the primary and secondary principal components, PC1 and PC2, respectively.
PC1PC2Wavelength SelectionCounts
A1B1A1B1, A1B1B2, …, A1B1B2B3… Bnn
A2B2A1A2B1, A1A2B1B2, …, A1A2B1B2B3… Bnn
A3B3A1A2A3B1, A1A2A3B1B2, …, A1A2A3B1B2B3… Bnn
…………n
AnBnA1A2A3… AnB1, A1A2A3… AnB1B2, …, A1A2A3… AnB1B2B3… Bnn
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Dong, R.; Shiraiwa, A.; Ichinose, K.; Pawasut, A.; Sreechun, K.; Mensin, S.; Hayashi, T. Hyperspectral Imaging and Machine Learning for Huanglongbing Detection on Leaf-Symptoms. Plants 2025, 14, 451. https://doi.org/10.3390/plants14030451

AMA Style

Dong R, Shiraiwa A, Ichinose K, Pawasut A, Sreechun K, Mensin S, Hayashi T. Hyperspectral Imaging and Machine Learning for Huanglongbing Detection on Leaf-Symptoms. Plants. 2025; 14(3):451. https://doi.org/10.3390/plants14030451

Chicago/Turabian Style

Dong, Ruihao, Aya Shiraiwa, Katsuya Ichinose, Achara Pawasut, Kesaraporn Sreechun, Sumalee Mensin, and Takefumi Hayashi. 2025. "Hyperspectral Imaging and Machine Learning for Huanglongbing Detection on Leaf-Symptoms" Plants 14, no. 3: 451. https://doi.org/10.3390/plants14030451

APA Style

Dong, R., Shiraiwa, A., Ichinose, K., Pawasut, A., Sreechun, K., Mensin, S., & Hayashi, T. (2025). Hyperspectral Imaging and Machine Learning for Huanglongbing Detection on Leaf-Symptoms. Plants, 14(3), 451. https://doi.org/10.3390/plants14030451

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop