Next Article in Journal
Frequency Spectra Analysis of Drawbar Pulls Generated by Special Driving Wheels Improving Tractive Performance
Previous Article in Journal
Efficient Certificate-Less Aggregate Signature Scheme with Conditional Privacy-Preservation for Vehicular Ad Hoc Networks Enhanced Smart Grid System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Non-Destructive Detection Pilot Study of Vegetable Organic Residues Using VNIR Hyperspectral Imaging and Deep Learning Techniques

1
Department of Agricultural Engineering, National Institute of Agricultural Sciences, 310 Nongsaengmyeong-ro, Deokjin-gu, Jeonju 54875, Korea
2
Department of Biosystems Engineering, Kangwon National University, 1 Kangwondaehak-gil, Chuncheon 24341, Gangwon-do, Korea
3
Interdisciplinary Program in Smart Agriculture, Kangwon National University, Chuncheon 24341, Gangwon-do, Korea
4
Environmental Microbial and Food Safety Laboratory, Beltsville Agricultural Research Center, Agricultural Research Service, USDA, 10300 Baltimore Avenue, Beltsville, MD 20705, USA
*
Authors to whom correspondence should be addressed.
Sensors 2021, 21(9), 2899; https://doi.org/10.3390/s21092899
Submission received: 4 March 2021 / Revised: 19 April 2021 / Accepted: 19 April 2021 / Published: 21 April 2021
(This article belongs to the Section Chemical Sensors)

Abstract

:
Contamination is a critical issue that affects food consumption adversely. Therefore, efficient detection and classification of food contaminants are essential to ensure food safety. This study applied a visible and near-infrared (VNIR) hyperspectral imaging technique to detect and classify organic residues on the metallic surfaces of food processing machinery. The experimental analysis was performed by diluting both potato and spinach juices to six different concentration levels using distilled water. The 3D hypercube data were acquired in the range of 400–1000 nm using a line-scan VNIR hyperspectral imaging system. Each diluted residue in the spectral domain was detected and classified using six classification methods, including a 1D convolutional neural network (CNN-1D) and five pre-processing methods. Among them, CNN-1D exhibited the highest classification accuracy, with a 0.99 and 0.98 calibration result and a 0.94 validation result for both spinach and potato residues. Therefore, in comparison with the validation accuracy of the support vector machine classifier (0.9 and 0.92 for spinach and potato, respectively), the CNN-1D technique demonstrated improved performance. Hence, the VNIR hyperspectral imaging technique with deep learning can potentially afford rapid and non-destructive detection and classification of organic residues in food facilities.

1. Introduction

Contamination inspection of food facilities is indispensable to ensure food safety. While raw agricultural products are consumed without any processing, certain procedures, such as peeling, shredding, cutting, trimming, extruding, and sanitizing, require processed agricultural products [1]. After the completion of processing operations, certain amounts of organic materials may remain on the blades, cracks, or crevices in the facilities. This can produce infectious foodborne bacteria. Therefore, several studies investigated the risk of infection from food processing machines. Researchers collected bacterial samples from fresh-cut processing facilities after sanitization. The collected samples were incubated on a general growth media for 24 h, from which the mesophilic and psychrotrophic bacteria were isolated and identified. The studies reported that approximately 30% of more than 1000 isolated pathogen samples can potentially provide environmental protection to the foodborne pathogens [2]. Therefore, contamination of fresh-cut products can occur at any time during the processing of agricultural products [3,4].
Conventionally, the hygiene assessment in food processing facilities is performed by plating and incubating the samples on growth media for 24–48 h [5]. Typically, the cultivation of microorganisms is time-consuming and requires trained operators to perform intensive lab work. However, several rapid and non-destructive organic molecular component detection and classification techniques exist to evaluate food safety, particularly to ensure hygiene and sanitation in mass food processing facilities.
For decades, X-ray, near-infrared spectroscopy, and computer vision have contributed to the development of non-destructive safety inspection technologies [6,7]. For instance, Fourier transform near-infrared (FT-NIR) and FT-IR spectroscopic methods identified the unexpected contamination of onion powder by starch [8], and Raman spectroscopic technology differentiated between fake and real eggs [9]. Hyperspectral imaging (HSI) is an emerging technology that uses a data cube with spectral and spatial data to analyze organic residues. For instance, the melamine content in infant formulas was detected and isolated using HSI technology [10]. Furthermore, multiple HSI modalities, including visible and near-infrared (VNIR, 400–1000 nm) and short-wavelength IR waves (900–1700 nm), were used to detect and classify mislabeled fish fillets [11]. Additionally, multispectral laser-induced fluorescence imaging detected different dilutions of animal fecal matter on apples, and three dilutions of 1:2, 1:20, and 1:200 detected approximately 80% of fecal matter within 24 h after the application of the technique. However, the detection accuracy lowered when the apples were brushed and washed; the fecal matter detected were 100%, 30%, and 0% for the 1:2, 1:20, and 1:200 dilutions, respectively [12].
Conversely, the deep learning method performs reliably with multidimensional data, including hyperspectral data [13,14]. Convolutional neural networks (CNNs) exhibit accurate classification and feature extraction with 1D, 2D, and 3D data [15,16,17,18]. In particular, CNN-1D is considered ideal for processing signals and prediction models, such as medical electrocardiography signals [19], environmental sounds [20], and human activity recognition [21]. Additionally, CNN-1D classified the spectral data of foodborne bacteria using hyperspectral microscope imaging technology with higher accuracy (90%) than that of the machine learning methods [22].
This study aimed to develop a non-destructive technique to detect and classify the concentration of biological residues in spinach and potato juices on stainless-steel surfaces. The proposed technique acquired the 3D hypercube data using a VNIR HSI system. Furthermore, a CNN-1D and several chemometric methods demonstrated the detection and classification results of the potato and spinach droplets diluted to six different concentrations.

2. Materials and Methods

2.1. Sample Preparation

We purchased fresh potato (Solanum tuberosum) and spinach (Spinacia oleracea) from a local market to prepare the diluted residues. Initially, the products were cut and squeezed to extract the juice. The diluted juice samples were then placed on stainless-steel plates and dried for 24 h. The experimental analysis was performed on six samples, namely, a 100% undiluted original fresh juice and five dilutions of potato and spinach, at 20% (1:5), 10% (1:10), 5% (1:20), 2% (1:50), and 1% (1:100), prepared by adding distilled water to the juice. Approximately 10 μL of the diluted solutions were placed on the plate with 15 rows and 2 replicates using a pipette. Thus, the total number of diluted droplets was 90 (6 dilutions × 15 repeats) on the stainless-steel plate.

2.2. VNIR HSI System and Data Acquisition

The HSI system comprises a 14-bit electron multiplying charge-coupled device (EMCCD) camera (Luca R DL-604M, Andor Technology, South Windsor, CT, USA) with a shutter speed of 80 ms, coupled with a C-mount objective lens (F1.9 35 mm compact lens, Schneider Optics, Van Nuys, CA, USA). The VNIR spectra in the range of 400–1000 nm were acquired using a spectrophotometer (VNIR Hyperspec, Headwall Photonics, Inc.; Fitchburg, MA, USA) that was combined with the EMCCD camera. Additionally, two halogen lamps provided the lighting system. Each sample plate was placed on a linear motorized platform (Velmex, Inc.; Bloomfield, NY, USA) to convey the samples using a lab-built HSI system (ARS, USDA, Beltsville, MD, USA). The sample plates placed on the moving platform acquired the VNIR 3D hypercube data of the diluted droplets using a line-scan camera. Initially, the dark and white references were captured and applied to calibrate the raw images before acquiring the hyperspectral images. The VNIR HSI data contained images of 1000 × 1004 pixels in size and 128 bands in the range of 400–1000 nm.

2.3. Region of Interest (ROI) Selection

Figure 1 illustrates the flow chart of the image processing and the development of the classification models. To extract spectral data from the raw VNIR HSI images, an optimal region of interest (ROI) must be selected. The Otsu algorithm, principal component analysis (PCA), and U-net demonstrated the ROI selection results.
The Otsu method is a well-known image binarization algorithm that uses an image thresholding technique [23]. In this study, the threshold values for spinach and potato were determined as 110 and 98, respectively. On the other hand, PCA calculates the correlation with input data. However, the spectral data were extracted based on the ROI selected using the U-net method. Additionally, stainless-steel background (BG) spectral data were randomly selected from 12 regions apart from the sample droplets. To enhance the quality of the selected band image, a median filter and an image sharpening technique were applied. The dilutions of the potato and spinach droplets were considered as “Hundred”, “Twenty”, ”Ten”, ”Five”, ”Two”, and ”One” for the classification, corresponding to the dilutions of 100% (original fresh juice), 20% (1:5), 10% (1:10), 5% (1:20), 2% (1:50), and 1% (1:100), respectively.

2.4. U-Net for Feature Segmentation

Figure 2 illustrates the schematic of the U-net architecture applied in the image processing. U-net is a CNN developed for biomedical image segmentation [24], wherein the architecture comprises the encoding and decoding procedures. The left (box# 1–4) and right (box# 6–9) portions in the figure represent the encoding and decoding procedures, respectively. A convolution kernel exists throughout the procedure. The data extraction and compression occurs during encoding. Additionally, this model adopted the rectified linear unit (ReLU) as the activation function and a 2 × 2 max-pooling kernel with a dropout coefficient of 0.5 to achieve dimensionality reduction in the data. The input data were downsized by a quart during the encoding procedure and subsequently merged during up-sampling in the initial convolution of the decoding procedure. The concatenate function merges the dropout and up-sampling convolution.
In the decoding process, a 2 × 2 convolution layer was implemented to reconstruct a new feature map rather than the max-pooling layer. Moreover, the concatenate function (merging operation) was implemented with the corresponding feature maps (results of the dropout or convolution) from the encoding process to develop a feature map in each decoding layer. For instance, two 128-channel feature maps (from boxes #4 and #6) were merged using the concatenate function to generate a 256-channel feature map (box #7). At the final layer (box #9), a 1 × 1 convolution kernel transformed the final feature map to yield the output (mask image in binary mode).

2.5. Development of the Classification Model

The classification model was developed based on two strategies (Figure 1). As indicated in the figure, STEP #1 constitutes the chemometric methods that involve multivariate analysis methods and machine learning algorithms. Conversely, STEP #2 uses the CNN-1D algorithm. Table 1 presents the detailed model architecture and specifications.
Both linear and non-linear multivariate classification methods, such as linear discriminant analysis (LDA), partial least squares discriminant analysis (PLS-DA), support vector machine (SVM), decision tree (DT), least squares support vector machine (LSSVM), and random forest (RF), were used to analyze the results [25,26]. RF is an ensemble algorithm of DTs { T 1 ( X ) ,   ,   T n ( X ) }, wherein X = { x 1 ,   ,   x n } is an n-dimensional vector of properties associated with a dependent variable (spectrum of diluted droplets). The tree ensemble yields N outputs { Y 1 = T 1 ( X ) ,   ,   Y n = T n ( X ) }, wherein Y n ,   n = 1, …, N, represent the class predicted by the n-th tree for the input data [27,28]. The classification results were obtained using six preprocessing methods. No-P denotes no pre-processing; D1 and D2 are the 1st and 2nd derivatives, respectively, based on the Savitzky–Golay algorithm; MSC, MA, and NM represent the multiplicative scatter correction, moving average, and normalization, respectively. The accuracy and Cohen’s kappa coefficient were used to show the results of the used classificaiton methods based on the confusion matrix [29]. Cross-validation was performed to evaluate the accuracy of the classification models using the leave-one-out (LOO) method. All classification algorithms and pre-processing methods were coded using R (Ver. 3.6.2.), the statistical open-source environment and language. The model was developed using multiple classification packages, such as caret (Ver. 6.0–85), e1071 (Ver. 1.7–3), rpart (Ver. 4.1–15), kernlab (Ver. 2004), and randomForest (Ver. 4.6). The Otsu algorithm was performed using ImageJ (Ver. 1.53c), which is an open-source scientific image processing program.
To classify the spectral data obtained from the diluted residues, a CNN-1D model was developed based on the architecture and parameters presented in Table 1. The optimized CNN-1D algorithm comprises convolution, average pooling, max-pooling, dropout, and output. The activation function uses ReLU to produce an image from a linear model. Both average pooling and max-pooling are applied to reduce the dimensionality of the spectral data. The total number of parameters and repeated epochs were 123,967 and 5000, respectively. However, these values can vary depending on the state of convergence. Deep learning classification was performed using Python (Ver. 3.7.4), and the corresponding packages included Tensorflow (Ver. 2.2.0), Keras (Ver. 2.3.1), and Scikit-image (Ver. 0.17.1). The CNN-1D frameworks were developed and trained on a computer equipped with an i7-8750H (CPU), GeForce GTX1050 Ti (GPU), and 16 GB memory.

3. Results and Discussion

3.1. ROI Segmentation

We used a mask image to select the ROIs of the potato and spinach residues automatically from the corrected sample images. The mask image was developed to observe the segmentation results in the column regions (20 × 86 pixels) of six dilutions obtained from the 15 repeated raw images (391 × 86 pixels). Figure 3 depicts the mask images of the spinach and potato residues obtained from the Otsu algorithm, PCA, and U-net methods along with the sample raw images. In the case of spinach, the color of the residue Hundred (100%) in the raw image is different from that of the other diluted residues. While the Otsu and PCA masks segmented the entire sample in the five diluted residues, U-net produced all the samples with limited loss in image pixels. Although the droplets of potato exhibited different intensities between the residues of Hundred (100%) and One (1%) in the raw image, PCA and U-net produced appropriate results.

3.2. VNIR Spectral Characteristics

Figure 4 illustrates the mean spectra extracted from the HSI data obtained from the residues of the potato juice and the stainless-steel BG. The colored image depicts six dilutions of potato residues and the extracted region of the BG spectrum. The peaks in the mean spectrum were observed at 625, 720, 785, and 860 nm. Typically, most of the selected bands within the VNIR regions are associated with physiological substances, such as the CH, NH, and OH stretching, in the vibrational spectrum. For instance, absorption of anthocyanin and carotenoid occur at 650 and 680 nm, respectively [30,31]. Spectral bands at 690–710 nm and 760–800 nm represent the total chlorophyll bands, whereas the absorption bands at 705, 842, and 920 nm are associated with carbohydrates [31]. The band at 995 nm represents the 2nd vibration of the NH bonds in proteins or amino acids, whereas that at 880 nm constitutes the 3rd overtone absorption of CH. Additionally, the band relates to the 2nd overtone absorption of the OH and NH bonds at 750–900 nm and 962–1000 nm, respectively [32]. Figure 5 illustrates the score scattering attributes demonstrated in the principal components (PCs) during intuitive data analysis. We assigned seven colors to the residues based on the dilutions of potato and the BG surface. The first PC (PC1) denotes the variance of the potato residues at six dilutions and the BG spectral data as 99% and 1%, respectively. Conversely, the second PC (PC2) indicates the variance of the spinach residues at six dilutions and the BG spectral data as 98% and 2%, respectively. The original juice (Hundred, no dilution) was easily distinguishable in the PC score plots (black, circle). Moreover, a class of BG (yellow, square) was isolated from the diluted residues. Figure 5 shows that the low-dilution residues (<10%) of potato demonstrated overlapping clusters.

3.3. Classification Results

Six multivariate analysis methods and machine learning algorithms were used to classify the diluted residues on the stainless-steel surface. Table 2 and Table 3 present the classification results of the potato and spinach residues, respectively, considering the accuracy (A) and kappa coefficient (K) in the classification models.
In the case of potato residues, LSSVM and RF exhibited higher accuracies than 0.86 based on the pre-processing methods, such as No-P, D1, MA, and NM. Additionally, LDA demonstrated reasonable classification results at an accuracy of 0.83; however, the accuracies of PLS-DA and DT were less than 0.77. Conversely, the classification results obtained from SVM were of the highest accuracy at 0.90 (Table 2), and the detailed results for each of the diluted residues were 1.0, 0.89, 0.87, 0.93, 0.71, 0.94, and 0.95 for Hundred, Twenty, Ten, Five, Two, One, and BG, respectively. However, CNN-1D demonstrated improved accuracies compared to those of SVM in each of the residues at 1.0, 0.97, 0.96, 0.89, 0.90, 0.89, and 1.0 (Figure 6).
In the case of spinach residues, SVM exhibited the most accurate classification results with an accuracy of 0.92 (Table 3). Moreover, the accuracies of the classification results obtained from RF were higher than 0.9 in the case of D1 and MSC. While SVM classified the results of each of the residues at accuracies of 1.0, 0.96, 0.90, 0.88, 0.81, 0.85, and 0.98, CNN-1D demonstrated improved accuracies of 0.99, 0.98, 0.95, 0.91, 0.83, 0.93, and 1.0 (Figure 7).
To compare the results of the classification models, spectral data analysis was performed using the CNN-1D algorithm. Six diluted residues were classified, and the results were presented using a confusion matrix. In this study, the numbers of training epochs and parameters were 500 and 123,967, respectively. The mean absolute error and the loss were 0.0262 and 0.0093, respectively, after the model was trained. Figure 6 and Figure 7 depict the confusion matrices representing the prediction accuracies of the developed CNN-1D model applied to the validation dataset. While Figure 6 depicts the classification and validation results of the potato residues (Ac = 0.99, Av = 0.94), Figure 7 presents those of the spinach residues (Ac = 0.98, Av = 0.94). These results indicate that the CNN-1D improves the classification accuracy by 2–4% from 0.92 and 0.90 in the case of potato and spinach, respectively, using the chemometric method

4. Conclusions

To detect and classify the organic residues on a metal surface accurately, we developed a classification model using VNIR HSI technology and machine learning methods. We implemented deep learning methods, such as U-net and CNN-1D, to generate a mask image in the classification model. Owing to the enhanced ROI segmentation and fine-tuned parameters in the CNN layers, both the deep learning methods demonstrated improved classification accuracies in the case of diluted residues. The two mask image algorithms, such as Otsu and PCA, used an optimal thresholding based on a single intensity threshold, which is calculated by the difference between the inter-class variance and between-class variance or loading vector, respectively. These two methods use a single intensity threshold so as to have a fast, simple technique, whereas they tend to find it difficult to separate detailed image parts from background. In turn, U-net adopted encoding and decoding procedures for data reduction and optimal feature selection. Furthermore, data augmentation using the annotated images demonstrated a precise feature segmentation [24]. Typically, organic residues can potentially generate biofilms that cause extracellular polymeric substance production and maturation [33]. Therefore, this study can potentially afford the early detection of biofilms in food processing machines using VNIR HSI and machine learning at an accuracy of A = 0.94. However, further research is necessary to obtain diverse evidence to fine-tune the hyperparameters of the deep learning methods, particularly when multiple samples are considered. Additionally, the proposed model must be validated across areas and in more practical locations of the food industry.

Author Contributions

Y.S. contributed to the data analysis, model development, and writing of the manuscript; G.K. contributed revised the manuscript; J.L. and J.J. provided insights in model devlopment and revised manuscript; A.L., B.K. and C.M. collected parts of data and contributed to the revision of the manuscript; M.S.K. provided crucial insights in data analysis and contributed to the revision of the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This study was carried out with the support of “Research Program for Agricultural Science & Technology Development (Project No. PJ012216)”, National Institute of Agricultural Sciences, RDA, Korea.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Artes, F.; Gomez, P.A.; Artes-Hemandez, F. Physical, Physiological and Microbial Deterioration of Minimally Fresh Processed Fruits and Vegetables. Food Sci. Technol. Int. 2007, 13, 177–188. [Google Scholar] [CrossRef]
  2. Liu, N.T.; Lefcourt, A.; Nou, X.; Shelton, D.; Zhang, G.; Lo, Y.M. Native Microflora in Fresh-Cut Produce Processing Plants and Their Potentials for Biofilm Formation. J. Food Prot. 2013, 76, 827–832. [Google Scholar] [CrossRef]
  3. Lehto, M.; Kuisma, R.; Määttä, J.; Kymäläinen, H.-R.; Mäki, M. Hygienic level and surface contamination in fresh-cut vegetable production plants. Food Control 2011, 22, 469–475. [Google Scholar] [CrossRef]
  4. Jung, Y.; Jang, H.; Matthews, K.R. Effect of the food production chain from farm practices to vegetable processing on outbreak incidence. Microb. Biotechnol. 2014, 7, 517–527. [Google Scholar] [CrossRef]
  5. Jongenburger, I.; Reij, M.W.; Boer, E.P.J.; Gorris, L.G.M.; Zwietering, M.H. Factors influencing the accuracy of the plating method used to enumerate low numbers of viable micro-organisms in food. Int. J. Food Microbiol. 2010, 143, 32–40. [Google Scholar] [CrossRef] [Green Version]
  6. Chen, Q.; Zhang, C.; Zhao, J.; Ouyang, Q. Recent advances in emerging imaging techniques for non-destructive detection of food quality and safety. TrAC Trends Anal. Chem. 2013, 52, 261–274. [Google Scholar] [CrossRef]
  7. Zhang, B.; Huang, W.; Li, J.; Zhao, C.; Fan, S.; Wu, J.; Liu, C. Principles, developments and applications of computer vision for external quality inspection of fruits and vegetables: A review. Food Res. Int. 2014, 62, 326–343. [Google Scholar] [CrossRef]
  8. Lohumi, S.; Lee, S.; Lee, W.H.; Kim, M.S.; Mo, C.; Bae, H.; Cho, B.K. Detection of Starch Adulteration in Onion Powder by FT-NIR and FT-IR Spectroscopy. J. Agric. Food Chem. 2014, 62, 9246–9251. [Google Scholar] [CrossRef]
  9. Joshi, R.; Lohumi, S.; Joshi, R.; Kim, M.S.; Qin, J.; Baek, I.; Cho, B.-K. Raman spectral analysis for non-invasive detection of external and internal parameters of fake eggs. Sens. Actuators B Chem. 2020, 303, 127243. [Google Scholar] [CrossRef]
  10. Lim, J.; Kim, G.; Mo, C.; Kim, M.S.; Chao, K.; Qin, J.; Fu, X.; Baek, I.; Cho, B.K. Detection of melamine in milk powders using near-infrared hyperspectral imaging combined with regression coefficient of partial least square regression model. Talanta 2016, 151, 183–191. [Google Scholar] [CrossRef] [Green Version]
  11. Qin, J.; Vasefi, F.; Hellberg, R.S.; Akhbardeh, A.; Isaacs, R.B.; Yilmax, A.G.; Hwang, C.; Baek, I.; Schmidt, W.F.; Kim, M.S. Detection of fish fillet substitution and mislabeling using multimode hyperspectral imaging techniques. Food Control 2020, 114, 107234. [Google Scholar] [CrossRef]
  12. Lefcourt, A.M.; Kim, M.S.; Chen, Y.-R. Automated detection of fecal contamination of apples by multispectral laser-induced fluorescence imaging. Appl. Opt. 2003, 42, 3935–3943. [Google Scholar] [CrossRef] [Green Version]
  13. Qiu, Z.; Chen, J.; Zhao, Y.; Zhu, S.; He, Y.; Zhang, C. Variety Identification of Single Rice Seed Using Hyperspectral Imaging Combined with Convolutional Neural Network. Appl. Sci. 2018, 8, 212. [Google Scholar] [CrossRef] [Green Version]
  14. Pyo, J.; Duan, H.; Baek, S.; Kim, M.S.; Jeon, T.; Kwon, Y.S.; Lee, H.; Cho, K.H. A convolutional neural network regression for quantifying cyanobacteria using hyperspectral imagery. Remote Sens. Environ. 2019, 233, 111350. [Google Scholar] [CrossRef]
  15. Chen, X.; Kopsaftopoulos, F.; Wu, Q.; Ren, H.; Chang, F.-K. A Self-Adaptive 1D Convolutional Neural Network for Flight-State Identification. Sensors 2019, 19, 275. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Li, W.; Wu, G.; Zhang, F.; Du, Q. Hyperspectral Image Classification Using Deep Pixel-Pair Features. IEEE Trans. Geosci. Remote Sens. 2017, 55, 844–853. [Google Scholar] [CrossRef]
  17. Potena, C.; Narki, D.; Pretto, A. Fast and accurate crop and weed identification with summarized train sets for precision agriculture. Intell. Auton. Syst. 2017, 531, 105–121. [Google Scholar]
  18. Nagasubramanian, K.; Jones, S.; Singh, A.K.; Sarkar, S.; Singh, A.; Canapathysubramanian, B. Plant disease identification using explainable 3D deep learning on hyperspectral images. Plant Methods 2019, 15, 98. [Google Scholar] [CrossRef]
  19. Kiranyaz, S.; Ince, T.; Abdeljaber, O.; Avci, O.; Gabbouj, M. 1-D Convolutional Neural Networks for Signal Processing Applications. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 8360–8364. [Google Scholar]
  20. Abdoli, S.; Cardinal, P.; Lameiras Koerich, A. End-to-end environmental sound classification using a 1D convolutional neural network. Expert Syst. Appl. 2019, 136, 252–263. [Google Scholar] [CrossRef] [Green Version]
  21. Cho, H.; Yoon, S.M. Divide and Conquer-Based 1D CNN Human Activity Recognition Using Test Data Sharpening. Sensors 2018, 18, 1055. [Google Scholar]
  22. Kang, R.; Park, B.; Eady, M.; Ouyang, Q.; Chen, K. Classification of foodborne bacteria using hyperspectral microscope imaging technology coupled with convolutional neural networks. Appl. Microbiol. Biotechnol. 2020, 104, 3157–3166. [Google Scholar] [CrossRef]
  23. Otsu, N. A threshold selection method from gray-level histograms. IEEE Trans. Syst. Manand Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef] [Green Version]
  24. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  25. Sattlecker, M.; Stone, N.; Smith, J.; Bessant, C. Assessment of robustness and transferability of classification models built for cancer diagnostics using Raman spectroscopy. J. Raman Spectrosc. 2011, 42, 897–903. [Google Scholar] [CrossRef]
  26. Wakholi, C.; Kandpal, L.M.; Lee, H.; Bae, H.; Park, E.; Kim, M.S.; Mo, C.; Lee, W.H.; Cho, B.K. Rapid assessment of corn seed viability using short wave infrared line-scan hyperspectral imaging and chemometrics. Sens. Actuators B Chem. 2018, 255, 498–507. [Google Scholar] [CrossRef]
  27. Svetnik, V.; Liaw, A.; Tong, C.; Culberson, J.C.; Sheridan, R.P.; Feuston, B.P. Random Forest:  A Classification and Regression Tool for Compound Classification and QSAR Modeling. J. Chem. Inf. Comput. Sci. 2003, 43, 1947–1958. [Google Scholar] [CrossRef]
  28. Xia, J.; Ghamisi, P.; Yokoya, N.; Iwasaki, A. Random Forest Ensembles and Extended Multiextinction Profiles for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 202–216. [Google Scholar] [CrossRef] [Green Version]
  29. Vieira, S.M.; Kaymak, U.; Sousa, J.M.C. Cohen’s kappa coefficient as a performance measure for feature selection. In Proceedings of the International Conference on Fuzzy Systems, Barcelona, Spain, 18–23 July 2010; pp. 1–8. [Google Scholar]
  30. Gitelson, A.A.; Keydan, G.P.; Merzlyak, M.N. Three-band model for noninvasive estimation of chlorophyll, carotenoids, and anthocyanin contents in higher plant leaves. Geophys. Res. Lett. 2006, 33, L11402. [Google Scholar] [CrossRef] [Green Version]
  31. Siedliska, A.; Baranowski, P.; Zubik, M.; Mazurek, W.; Sosnowska, B. Detection of fungal infections in strawberry fruit by VNIR/SWIR hyperspectral imaging. Postharvest Biol. Technol. 2018, 139, 115–126. [Google Scholar] [CrossRef]
  32. Weng, S.; Tang, P.; Yuan, H.; Guo, B.; Yu, S.; Huang, L.; Xu, C. Hyperspectral imaging for accurate determination of rice variety using a deep learning network with multi-feature fusion. Spectrochim. Acta Part A Mol. Biomol. Spectrosc. 2020, 234, 118237. [Google Scholar] [CrossRef] [PubMed]
  33. Chmielewski, R.A.N.; Frank, J.F. Biofilm Formation and Control in Food Processing Facilities. Compr. Rev. Food Sci. Food Saf. 2003, 2, 22–32. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Flow chart of the data processing and schematic architecture of a 1D convolutional neural network. In order to develop the classification model, STEP#1 is using chemometric methods and STEP#2 is using the CNN-1D algorithm.
Figure 1. Flow chart of the data processing and schematic architecture of a 1D convolutional neural network. In order to develop the classification model, STEP#1 is using chemometric methods and STEP#2 is using the CNN-1D algorithm.
Sensors 21 02899 g001
Figure 2. Schematic flowchart of the U-net architecture for image masking.
Figure 2. Schematic flowchart of the U-net architecture for image masking.
Sensors 21 02899 g002
Figure 3. Target segmentation results of the spinach and potato residue droplets. The mask generated by U-net had less loss of the target image pixels.
Figure 3. Target segmentation results of the spinach and potato residue droplets. The mask generated by U-net had less loss of the target image pixels.
Sensors 21 02899 g003
Figure 4. Mean spectrum data of (A) potato and its locational (B) residues according to the dilutions and background (BG) pixels. (C) Mean spectra of spinach.
Figure 4. Mean spectrum data of (A) potato and its locational (B) residues according to the dilutions and background (BG) pixels. (C) Mean spectra of spinach.
Sensors 21 02899 g004aSensors 21 02899 g004b
Figure 5. PCA image on PC1 and PC2 space using the spectral data of potato (A) and spinach (B) from the residues and background ROI.
Figure 5. PCA image on PC1 and PC2 space using the spectral data of potato (A) and spinach (B) from the residues and background ROI.
Sensors 21 02899 g005
Figure 6. Confusion matrix of the classification result for potato according to the dilution of the residues using CNN-1D. The calibration result (Ac = 0.99) is (A) and the validation result (Av = 0.94) is (B).
Figure 6. Confusion matrix of the classification result for potato according to the dilution of the residues using CNN-1D. The calibration result (Ac = 0.99) is (A) and the validation result (Av = 0.94) is (B).
Sensors 21 02899 g006
Figure 7. Confusion matrix of the classification result for spinach dilutions of the residues using CNN-1D. The calibration result (Ac = 0.98) is (A) and the validation result (Av = 0.94) is (B).
Figure 7. Confusion matrix of the classification result for spinach dilutions of the residues using CNN-1D. The calibration result (Ac = 0.98) is (A) and the validation result (Av = 0.94) is (B).
Sensors 21 02899 g007
Table 1. Convolutional neural network and parameters.
Table 1. Convolutional neural network and parameters.
LayerTypeOutput Shape# of Parameter
Conv1d_1Conv1D124, 40240
Average_pooling1d_1 AveragePooling1D62, 400
Max_pooling1d_1MaxPooling1D31, 400
Conv1d_2Conv1D29, 202420
Average_pooling1d_2AveragePooling1D14, 200
Max_pooling1d_2MaxPooling1D7, 200
Flatten_1Flatten1400
Dense_1Dense50070,500
Dropout_1Dropout5000
Dense_2Dense10050,100
Dense_3Dense7707
Total parameters: 123,967, Trainable parameters: 123,967
Table 2. Result of the classification of the diluted residues of potato as the accuracy (A) and kappa coefficient (K) based on chemometric methods.
Table 2. Result of the classification of the diluted residues of potato as the accuracy (A) and kappa coefficient (K) based on chemometric methods.
MAPPHundredTwentyTenFiveTwoOneBGAK
LDANoP1.000.870.700.680.600.861.000.830.80
D11.000.870.700.680.600.861.000.830.80
D21.000.870.700.680.600.861.000.830.80
MSC0.980.850.780.390.420.580.820.710.66
MA1.000.870.700.680.600.861.000.830.80
NM1.000.860.710.660.620.861.000.830.80
PLSDANoP0.990.730.220.630.360.541.000.680.62
D10.990.640.010.710.190.071.000.570.49
D21.000.560.050.590.370.081.000.580.50
MSC0.990.800.670.700.250.650.960.600.52
MA0.990.730.220.630.360.541.000.680.62
NM1.000.750.240.670.360.561.000.690.63
SVMNoP1.000.890.870.930.710.940.950.900.88
D11.000.850.780.790.630.930.950.860.83
D21.000.860.710.820.600.920.950.850.82
MSC1.000.830.780.780.390.600.940.780.74
MA1.000.870.710.810.360.900.950.810.78
NM1.000.870.710.810.360.900.950.810.78
DTNoP1.000.730.560.680.400.870.950.760.71
D10.960.650.370.660.330.880.950.710.65
D20.990.700.320.610.310.900.950.700.65
MSC0.990.720.690.740.140.460.870.680.62
MA1.000.730.560.680.400.870.950.760.71
NM1.000.730.560.680.400.870.950.760.71
LSSVMNoP1.000.870.810.900.740.950.950.890.87
D11.000.860.660.810.670.980.950.850.83
D21.000.790.100.620.330.890.950.690.64
MSC0.990.800.670.700.250.650.960.740.69
MA1.000.880.810.880.730.940.950.890.87
NM1.000.870.780.870.690.940.950.880.86
RFNoP0.980.930.830.770.670.881.000.880.86
D10.980.920.730.760.740.941.000.880.86
D20.990.900.580.720.670.951.000.850.82
MSC0.980.910.720.590.430.490.990.760.72
MA0.980.930.830.770.670.881.000.880.86
NM0.980.880.810.760.670.881.000.870.84
MA: multivariate analysis; PP: preprocessing method; No-P: no-preprocessing; D1: 1st derivative; D2: 2nd derivative; MSC: multiplicative scatter correction; MA: moving average; NM: normalization; Hundred: 100% residue; Twenty: 20% diluted residue (1:5 dilution); Ten: 10% diluted residue (1:10 dilution); Five: 5% diluted residue (1:20 dilution); Two: 2% diluted residue (1:50 dilution); One: 1% diluted residue (1:100 dilution); A: Accuracy; K: kappa coefficient.
Table 3. Result of the classification of the diluted residues of spinach as the accuracy (A) and kappa coefficient (K) based on chemometric methods.
Table 3. Result of the classification of the diluted residues of spinach as the accuracy (A) and kappa coefficient (K) based on chemometric methods.
MAPPHundredTwentyTenFiveTwoOneBGAK
LDANoP1.000.920.830.770.770.741.000.880.86
D11.000.920.830.770.770.741.000.880.86
D21.000.920.830.770.770.741.000.880.86
MSC0.990.830.900.670.690.431.000.810.78
MA1.000.920.830.770.770.741.000.880.86
NM1.000.920.820.780.770.761.000.880.86
PLSDANoP1.000.960.870.370.660.741.000.830.80
D11.000.930.870.090.460.561.000.750.70
D21.000.960.880.180.600.701.000.800.76
MSC1.000.860.730.750.560.740.970.810.78
MA1.000.960.870.370.660.741.000.830.80
NM1.000.970.870.380.650.741.000.830.80
SVMNoP1.000.960.900.880.810.850.980.920.91
D11.000.900.760.820.760.780.980.870.85
D21.000.920.800.840.790.810.980.890.87
MSC1.000.840.780.790.770.720.980.860.83
MA1.000.760.640.710.760.710.980.810.78
NM1.000.760.640.710.760.710.980.810.78
DTNoP1.000.770.630.650.650.680.980.790.75
D11.000.850.710.740.600.780.980.830.80
D21.000.940.700.840.620.710.980.840.82
MSC1.000.890.720.800.720.660.980.840.81
MA1.000.770.630.650.650.680.980.790.75
NM1.000.770.630.650.650.680.980.790.75
LSSVMNoP1.000.890.820.790.700.750.980.860.84
D11.000.860.720.640.550.750.980.810.77
D21.000.900.730.710.470.700.980.810.77
MSC1.000.860.730.750.560.740.970.820.79
MA1.000.850.790.730.710.730.980.840.81
NM1.000.900.850.770.750.760.980.870.85
RFNoP1.000.860.840.780.820.841.000.890.87
D11.000.860.810.850.850.861.000.900.88
D21.000.890.810.880.790.761.000.890.87
MSC1.000.870.840.870.920.851.000.910.90
MA1.000.860.840.780.820.841.000.890.87
NM1.000.880.820.790.820.831.000.890.87
MA: multivariate analysis; PP: preprocessing method; No-P: no-preprocessing; D1: 1st derivative; D2: 2nd derivative; MSC: multiplicative scatter correction; MA: moving average; NM: normalization; Hundred: 100% residue; Twenty: 20% diluted residue (1:5 dilution); Ten: 10% diluted residue (1:10 dilution); Five: 5% diluted residue (1:20 dilution); Two: 2% diluted residue (1:50 dilution); One: 1% diluted residue (1:100 dilution); A: Accuracy; K: kappa coefficient.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Seo, Y.; Kim, G.; Lim, J.; Lee, A.; Kim, B.; Jang, J.; Mo, C.; Kim, M.S. Non-Destructive Detection Pilot Study of Vegetable Organic Residues Using VNIR Hyperspectral Imaging and Deep Learning Techniques. Sensors 2021, 21, 2899. https://doi.org/10.3390/s21092899

AMA Style

Seo Y, Kim G, Lim J, Lee A, Kim B, Jang J, Mo C, Kim MS. Non-Destructive Detection Pilot Study of Vegetable Organic Residues Using VNIR Hyperspectral Imaging and Deep Learning Techniques. Sensors. 2021; 21(9):2899. https://doi.org/10.3390/s21092899

Chicago/Turabian Style

Seo, Youngwook, Giyoung Kim, Jongguk Lim, Ahyeong Lee, Balgeum Kim, Jaekyung Jang, Changyeun Mo, and Moon S. Kim. 2021. "Non-Destructive Detection Pilot Study of Vegetable Organic Residues Using VNIR Hyperspectral Imaging and Deep Learning Techniques" Sensors 21, no. 9: 2899. https://doi.org/10.3390/s21092899

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop