Figure 1.
Hyperspectral imaging of clothing and scenery. (a) Hyperspectral capture. (b) Examples of clothing images (excerpt). (c) Examples of background images (excerpt).
Figure 1.
Hyperspectral imaging of clothing and scenery. (a) Hyperspectral capture. (b) Examples of clothing images (excerpt). (c) Examples of background images (excerpt).
Figure 2.
Example of a labeled image (a) and its corresponding pseudo-color (an artificially colored visualization) image (b).
Figure 2.
Example of a labeled image (a) and its corresponding pseudo-color (an artificially colored visualization) image (b).
Figure 3.
Hyperspectral reflectance curves, showing reflectance (vertical axis) over the visible–near-infrared range (horizontal axis) for 100 samples. (a) Clothing. (b) Inorganic background. (c) Plant background.
Figure 3.
Hyperspectral reflectance curves, showing reflectance (vertical axis) over the visible–near-infrared range (horizontal axis) for 100 samples. (a) Clothing. (b) Inorganic background. (c) Plant background.
Figure 4.
Distribution of the hyperspectral signals from clothing labeled into 39 categories. The data were analyzed using t-SNE, and circles are plotted at positions corresponding to the first and second components.
Figure 4.
Distribution of the hyperspectral signals from clothing labeled into 39 categories. The data were analyzed using t-SNE, and circles are plotted at positions corresponding to the first and second components.
Figure 5.
Multi-layer perceptron dataset creation workflow. Different colors and numbering indicate distinct label values in the dataset.
Figure 5.
Multi-layer perceptron dataset creation workflow. Different colors and numbering indicate distinct label values in the dataset.
Figure 6.
Sampling example, with moss-green dots indicating chosen pixels. (a) A pseudo-color image with 470, 580, 640 nm. (b) A labeled image. (c) A sampling example.
Figure 6.
Sampling example, with moss-green dots indicating chosen pixels. (a) A pseudo-color image with 470, 580, 640 nm. (b) A labeled image. (c) A sampling example.
Figure 7.
Multi-layer perceptron workflow and multi-label confusion matrix for 167 bands. The red and blue lines in the schematic represent, respectively, the relative magnitude and sign (positive in blue, negative in red) of the connection weights. These lines are illustrative only and do not reflect actual numerical weight values.
Figure 7.
Multi-layer perceptron workflow and multi-label confusion matrix for 167 bands. The red and blue lines in the schematic represent, respectively, the relative magnitude and sign (positive in blue, negative in red) of the connection weights. These lines are illustrative only and do not reflect actual numerical weight values.
Figure 8.
(a) Conversion to the clothing versus background matrix. (b) Example of the 2 × 2 confusion matrix for the 167-band model.
Figure 8.
(a) Conversion to the clothing versus background matrix. (b) Example of the 2 × 2 confusion matrix for the 167-band model.
Figure 9.
(a) Multi-label confusion matrix for 12-band multi-layer perceptron (MLP-12). (b) 2 × 2 confusion matrix for MLP-12.
Figure 9.
(a) Multi-label confusion matrix for 12-band multi-layer perceptron (MLP-12). (b) 2 × 2 confusion matrix for MLP-12.
Figure 10.
Relationship between band count and macro_avg.
Figure 10.
Relationship between band count and macro_avg.
Figure 11.
Flowchart of optimal wavelength set exploration (Steps 1–5 in the text). Different colored dots represent groups obtained via k-means clustering of candidate wavelength sets (for illustrative purposes only). See
Section 4.3.2 for details.
Figure 11.
Flowchart of optimal wavelength set exploration (Steps 1–5 in the text). Different colored dots represent groups obtained via k-means clustering of candidate wavelength sets (for illustrative purposes only). See
Section 4.3.2 for details.
Figure 12.
Example of convergence for the 4-band optimal wavelength set search (from left to right: initial state, iteration 1, iteration 2, iteration 5). (a) Principal component analysis (PCA) visualization of wave-set clusters. (b) Deviation (nm) of each band from the mean. The x-axis is the band index (1–4), and the y-axis is the offset in nm. Different colored lines indicate groups obtained via k-means clustering.
Figure 12.
Example of convergence for the 4-band optimal wavelength set search (from left to right: initial state, iteration 1, iteration 2, iteration 5). (a) Principal component analysis (PCA) visualization of wave-set clusters. (b) Deviation (nm) of each band from the mean. The x-axis is the band index (1–4), and the y-axis is the offset in nm. Different colored lines indicate groups obtained via k-means clustering.
Figure 13.
(a) Multi-label confusion matrix for optimal wavelength set with 4 bands (OWS4-1). (b) 2 × 2 confusion matrix for OWS4-1.
Figure 13.
(a) Multi-label confusion matrix for optimal wavelength set with 4 bands (OWS4-1). (b) 2 × 2 confusion matrix for OWS4-1.
Figure 14.
Passbands of the 4- and 5-band optimal wavelength sets (OWS). (a) OWS4-1. (b) OWS5-1. (c) OWS4-2. (d) OWS5-2.
Figure 14.
Passbands of the 4- and 5-band optimal wavelength sets (OWS). (a) OWS4-1. (b) OWS5-1. (c) OWS4-2. (d) OWS5-2.
Figure 15.
Multispectral reflectance at the four wavelengths of the optimal wavelength set (OWS4-1: 453, 556, 668, 708 nm) for (a) clothing, (b) inorganic background, and (c) plant background.
Figure 15.
Multispectral reflectance at the four wavelengths of the optimal wavelength set (OWS4-1: 453, 556, 668, 708 nm) for (a) clothing, (b) inorganic background, and (c) plant background.
Figure 16.
Performance variation when shifting the center wavelength of each of the 4 bands.
Figure 16.
Performance variation when shifting the center wavelength of each of the 4 bands.
Figure 17.
Performance variation for the 5-band set under center-wavelength shifts. Shifting the 3rd or 4th band alone had minimal impact (see text).
Figure 17.
Performance variation for the 5-band set under center-wavelength shifts. Shifting the 3rd or 4th band alone had minimal impact (see text).
Figure 18.
Performance variation when simultaneously widening the passband of all 4 bands. (a) performance of broadening passbands. (b) broadened passbands at 0.02 lower performance.
Figure 18.
Performance variation when simultaneously widening the passband of all 4 bands. (a) performance of broadening passbands. (b) broadened passbands at 0.02 lower performance.
Figure 19.
Example inference results at resolutions (a) multi-layer perceptron (MLP) at 512 × 512 pixels, (b) MLP at 64 × 64 pixels, and (c) YOLOv5m at 512 × 512 pixels, (d) YOLOv5m at 64 × 64 pixels. The MLP detects the boundary of clothing pixels and draws bounding boxes. It uses only spectral information from each pixel to decide whether it is clothing.
Figure 19.
Example inference results at resolutions (a) multi-layer perceptron (MLP) at 512 × 512 pixels, (b) MLP at 64 × 64 pixels, and (c) YOLOv5m at 512 × 512 pixels, (d) YOLOv5m at 64 × 64 pixels. The MLP detects the boundary of clothing pixels and draws bounding boxes. It uses only spectral information from each pixel to decide whether it is clothing.
Figure 20.
Inference speed, memory usage, and detection score. (a) Inference speed. The multi-layer perceptron (MLP) stays well below the 33 ms real-time threshold. (b) Memory usage. (c) Detection score.
Figure 20.
Inference speed, memory usage, and detection score. (a) Inference speed. The multi-layer perceptron (MLP) stays well below the 33 ms real-time threshold. (b) Memory usage. (c) Detection score.
Figure 21.
Examples of inference results from some wavelength set models with different band counts (12, 5, 4, and 3). Models with ≥4 bands perform well. (a) Pseudo-color (an artificially colored visualization) image. (b) 12-band multi-layer perceptron (MLP-12). (c) 5-band optimal wavelength set (OWS5-1) multi-layer perceptron. (d) 4-band OWS4-1 multi-layer perceptron. (e) 3-band OWS3-1 multi-layer perceptron.
Figure 21.
Examples of inference results from some wavelength set models with different band counts (12, 5, 4, and 3). Models with ≥4 bands perform well. (a) Pseudo-color (an artificially colored visualization) image. (b) 12-band multi-layer perceptron (MLP-12). (c) 5-band optimal wavelength set (OWS5-1) multi-layer perceptron. (d) 4-band OWS4-1 multi-layer perceptron. (e) 3-band OWS3-1 multi-layer perceptron.
Figure 22.
Example of a street scene containing a person with clothing not included in the training dataset. (a) Pseudo-color image. (b) 4-band optimal wavelength set (OWS4-1) multi-layer perceptron predictions. (c) Detailed labeling of the clothing region.
Figure 22.
Example of a street scene containing a person with clothing not included in the training dataset. (a) Pseudo-color image. (b) 4-band optimal wavelength set (OWS4-1) multi-layer perceptron predictions. (c) Detailed labeling of the clothing region.
Figure 23.
Samples of detected clothing not included in the training dataset. A single case misidentified as “plant” (red circle). (a) Pseudo-color image. (b) 4-band optimal wavelength set (OWS4-1) multi-layer perceptron predictions.
Figure 23.
Samples of detected clothing not included in the training dataset. A single case misidentified as “plant” (red circle). (a) Pseudo-color image. (b) 4-band optimal wavelength set (OWS4-1) multi-layer perceptron predictions.
Figure 24.
Example of a street scene containing background objects likely to be misclassified as clothing. (a) A scene including misclassified red and yellow objects. (b) Multi-label predictions by the 4-band optimal wavelength set (OWS4-1) multi-layer perceptron.
Figure 24.
Example of a street scene containing background objects likely to be misclassified as clothing. (a) A scene including misclassified red and yellow objects. (b) Multi-label predictions by the 4-band optimal wavelength set (OWS4-1) multi-layer perceptron.
Figure 25.
White (blue circle) or black (red circle) wool garments can be missed. (a) Pseudo-color image. (b) Multi-label predictions by the 4-band optimal wavelength set (OWS4-1) multi-layer perceptron. Gray cotton (green circle) garments can be missed. (c) Pseudo-color image. (d) Multi-label predictions by the 4-band optimal wavelength set (OWS4-1) multi-layer perceptron.
Figure 25.
White (blue circle) or black (red circle) wool garments can be missed. (a) Pseudo-color image. (b) Multi-label predictions by the 4-band optimal wavelength set (OWS4-1) multi-layer perceptron. Gray cotton (green circle) garments can be missed. (c) Pseudo-color image. (d) Multi-label predictions by the 4-band optimal wavelength set (OWS4-1) multi-layer perceptron.
Figure 26.
Examples of clothing materials deviating from this study’s “clothing hypothesis” (genuine leather and synthetic leather). (a) Pseudo-color image (genuine leather). (b) 4-band optimal wavelength set (OWS4-1) multi-layer perceptron inference (genuine leather). (c) Pseudo-color image (synthetic leather). (d) 4-band OWS4-1 multi-layer perceptron inference (synthetic leather).
Figure 26.
Examples of clothing materials deviating from this study’s “clothing hypothesis” (genuine leather and synthetic leather). (a) Pseudo-color image (genuine leather). (b) 4-band optimal wavelength set (OWS4-1) multi-layer perceptron inference (genuine leather). (c) Pseudo-color image (synthetic leather). (d) 4-band OWS4-1 multi-layer perceptron inference (synthetic leather).
Figure 27.
Overall inference examples from the 4-band optimal wavelength set (OWS4-1) multi-layer perceptron (top: pseudo-color (artificial color mapping) image; bottom: classification result with bright yellow for clothing and black for background). (a) Scenes not included in the training dataset. (b) Scenes used in training.
Figure 27.
Overall inference examples from the 4-band optimal wavelength set (OWS4-1) multi-layer perceptron (top: pseudo-color (artificial color mapping) image; bottom: classification result with bright yellow for clothing and black for background). (a) Scenes not included in the training dataset. (b) Scenes used in training.
Figure 28.
(a) 4-band reflectance and (b) hyperspectral reflectance for four types of green/blue garments.
Figure 28.
(a) 4-band reflectance and (b) hyperspectral reflectance for four types of green/blue garments.
Figure 29.
(a) 4-band and (b) hyperspectral reflectance for plants.
Figure 29.
(a) 4-band and (b) hyperspectral reflectance for plants.
Figure 30.
Reflectance at (a) 4 bands and (b) full hyperspectral data for white garments (polyester, cotton, and wool).
Figure 30.
Reflectance at (a) 4 bands and (b) full hyperspectral data for white garments (polyester, cotton, and wool).
Table 1.
List of 41 labels (label name, index, and R-channel intensity in the labeled image).
Table 1.
List of 41 labels (label name, index, and R-channel intensity in the labeled image).
Label Name 1 | Index | R-Channel Intensity |
---|
Inorganic background | 0 | 0 |
Plant background | 34 | 255 |
P-Fluorescent-Yellow | 1 | 53 |
P-Gray 2 | 2 | 245 |
P-White + C-White | 3 | 67 |
P-Black + C-Black | 4 | 204 |
P-Gray 1 + W-Gray 1 + C-Lt. Gray | 5 | 145 |
P-Fluorescent-Orange | 6 | 135 |
P-Blue + C-Blue | 7 | 219 |
P-Yellow + C-Yellow | 8 | 127 |
P-Navy 2 | 9 | 187 |
P-Navy 1 + C-Navy | 10 | 72 |
P-Red + C-Red + W-Red | 11 | 178 |
P-Green + C-Green | 12 | 248 |
C-Gray 2 | 13 | 244 |
C-Gray 1 | 14 | 189 |
C-Purple + W-Navy | 15 | 64 |
W-Black + W-Jet-Black | 16 | 197 |
W-Yellow | 17 | 128 |
W-Gray 2 | 18 | 65 |
W-Khaki + C-Lt. Olive | 19 | 171 |
W-Blue | 20 | 241 |
W-White | 21 | 153 |
W-Lt. Blue + P-Lt. Blue + C-Lt. Blue | 22 | 240 |
W-Green | 23 | 188 |
P-Sax Blue | 25 | 246 |
C-Lavender | 26 | 236 |
P-Lt. Green + C-Bright Green | 27 | 252 |
C-Light Beige + W-J. Sand Beige | 28 | 233 |
W-G. 1 Beige + W-T. 2 Beige | 29 | 227 |
W-T. 1 Ivory Beige | 30 | 214 |
P-Lt. Pink + C-Baby Pink + C-Lt. Pink | 35 | 251 |
W-G. 13 Cream Yellow | 36 | 206 |
P-MC. Pink | 38 | 253 |
P-Lt. Yellow + C-Lt. Yellow | 39 | 243 |
P-MC. Red | 40 | 239 |
A. Orange + Glim. Orange | 41 | 203 |
P-MC. Purple | 43 | 247 |
W-J. 6 Magenta | 45 | 221 |
C-Lt. Purple | 46 | 232 |
P-Lt. Purple | 47 | 249 |
Table 2.
Comparison of five machine learning approaches: radial basis function support vector machine (RBF-SVM), random forest, gradient boosting, adaptive boosting, and multi-layer perceptron (MLP).
Table 2.
Comparison of five machine learning approaches: radial basis function support vector machine (RBF-SVM), random forest, gradient boosting, adaptive boosting, and multi-layer perceptron (MLP).
Model | Accuracy Score | Precision | Recall | F1 Score |
---|
RBF-SVM | 0.922 | 0.88 | 0.525 | 0.658 |
Random Forest | 0.846 | 0.471 | 0.592 | 0.524 |
Gradient Boosting | 0.904 | 0.787 | 0.452 | 0.574 |
Adaptive Boosting | 0.901 | 0.776 | 0.437 | 0.559 |
MLP | 0.921 | 0.815 | 0.586 | 0.682 |
Table 3.
Results of preliminary experiments on the multi-layer perceptron (MLP) hidden units (6/10/16/32/64). Performance plateaued with 16 units × 2 layers.
Table 3.
Results of preliminary experiments on the multi-layer perceptron (MLP) hidden units (6/10/16/32/64). Performance plateaued with 16 units × 2 layers.
Model | Accuracy | Precision | Recall | F1 Score | Band1 (nm) | Band2 (nm) | Band3 (nm) | Band4 (nm) | Band5 (nm) | ... | Band20 (nm) |
---|
MLP: 20-6-6-49 | 0.884 | 0.94 | 0.837 | 0.885 | 413 | 438 | 463 | 488 | 513 | ... | 888 |
MLP: 20-10-10-49 | 0.917 | 0.95 | 0.893 | 0.92 | 413 | 438 | 463 | 488 | 513 | | 888 |
MLP: 20-16-16-49 | 0.928 | 0.965 | 0.899 | 0.931 | 413 | 438 | 463 | 488 | 513 | | 888 |
MLP: 20-32-32-49 | 0.924 | 0.967 | 0.889 | 0.927 | 413 | 438 | 463 | 488 | 513 | | 888 |
MLP: 20-64-64-49 | 0.935 | 0.97 | 0.908 | 0.938 | 413 | 438 | 463 | 488 | 513 | | 888 |
Table 4.
Accuracy/Precision/Recall/F1 Score for 167-band Multi-Layer Perceptron (MLP-167).
Table 4.
Accuracy/Precision/Recall/F1 Score for 167-band Multi-Layer Perceptron (MLP-167).
Model | Accuracy | Precision | Recall | F1 Score | Band1 (nm) | Band2 (nm) | Band3 (nm) | Band4 (nm) | Band5 (nm) | ... | Band167 (nm) |
---|
MLP-167 | 0.934 | 0.98 | 0.897 | 0.936 | 400 | 403 | 406 | 409 | 412 | ... | 900 |
Table 5.
Comparison of evaluation metrics for multi-layer perceptron (MLP) models with reduced dimensions by uniformly subsampling wavelengths or by principal component analysis (PCA).
Table 5.
Comparison of evaluation metrics for multi-layer perceptron (MLP) models with reduced dimensions by uniformly subsampling wavelengths or by principal component analysis (PCA).
Model | Accuracy | Precision | Recall | F1 Score |
---|
MLP-150 (400~900) | 0.924 | 0.976 | 0.881 | 0.926 |
MLP-70 (400~900) | 0.931 | 0.98 | 0.892 | 0.933 |
MLP-20 (400~900) | 0.934 | 0.978 | 0.898 | 0.936 |
MLP-12 (400~900) | 0.926 | 0.961 | 0.9 | 0.929 |
MLP-12 (430~770) | 0.947 | 0.970 | 0.931 | 0.950 |
MLP-9 (420~800) | 0.931 | 0.964 | 0.906 | 0.934 |
PCA-12 | 0.921 | 0.978 | 0.874 | 0.923 |
PCA-8 | 0.917 | 0.975 | 0.87 | 0.919 |
PCA-6 | 0.91 | 0.975 | 0.857 | 0.912 |
Table 6.
Final optimal wavelength set (OWS) combinations and evaluation metrics for 4-, 5-, and 3-band searches.
Table 6.
Final optimal wavelength set (OWS) combinations and evaluation metrics for 4-, 5-, and 3-band searches.
Model | Accuracy | Precision | Recall | F1 Score | Band1 (nm) | Band2 (nm) | Band3 (nm) | Band4 (nm) | Band5 (nm) |
---|
OWS4-1 | 0.947 | 0.97 | 0.932 | 0.95 | 453 | 556 | 668 | 708 | - |
OWS4-2 | 0.946 | 0.968 | 0.932 | 0.949 | 446 | 567 | 647 | 716 | - |
OWS5-1 | 0.947 | 0.97 | 0.932 | 0.95 | 444 | 556 | 623 | 652 | 709 |
OWS5-2 | 0.943 | 0.969 | 0.924 | 0.946 | 445 | 548 | 562 | 675 | 729 |
OWS3-1 | 0.93 | 0.945 | 0.926 | 0.935 | 445 | 581 | 713 | - | - |
Table 7.
Performance of Additional Wavelength Set Configurations: Agricultural-Use and Range-Limited (Visible-Only and Near-Infrared-Only) Examples.
Table 7.
Performance of Additional Wavelength Set Configurations: Agricultural-Use and Range-Limited (Visible-Only and Near-Infrared-Only) Examples.
Model | Accuracy | Precision | Recall | F1 Score | Band1 (nm) | Band2 (nm) | Band3 (nm) | Band4 (nm) | Band5 (nm) | ... | Band12 (nm) |
---|
Parrot Sequoia | 0.822 | 0.94 | 0.714 | 0.811 | 550 | 659 | 734 | 790 | - | - | - |
DJI P4 | 0.914 | 0.963 | 0.874 | 0.916 | 449 | 560 | 650 | 730 | 839 | - | - |
MicaSense RedEdge3 | 0.919 | 0.963 | 0.883 | 0.921 | 475 | 560 | 668 | 716 | 839 | - | - |
MLP-NIR12 | 0.859 | 0.951 | 0.778 | 0.855 | 651 | 673 | 694 | 716 | 738 | ... | 889 |
MLP-VIS12 | 0.885 | 0.944 | 0.835 | 0.886 | 412 | 435 | 458 | 482 | 505 | ... | 668 |
Table 8.
Comparison of speed, memory, and detection score for a sample test image.
Table 8.
Comparison of speed, memory, and detection score for a sample test image.
Model Name 1 | Input Image Size | Peak Memory (MB) | Inference Time (ms) | Detection Score |
---|
MLP | 64 × 64 | 1.02 | 0.5 | 1 |
MLP | 128 × 128 | 3.97 | 0.6 | 1 |
MLP | 512 × 512 | 63.04 | 1.3 | 1 |
Faster R-CNN | 64 × 64 | 462.91 | 44.7 | 0.98 |
Faster R-CNN | 128 × 128 | 463.05 | 43 | 1 |
Faster R-CNN | 512 × 512 | 459.56 | 42.4 | 1 |
EfficientDet | 64 × 64 | 119.65 | 184.7 | 0.93 |
EfficientDet | 128 × 128 | 121.15 | 146.8 | 0.96 |
EfficientDet | 512 × 512 | 134.77 | 145.4 | 0.97 |
YOLOv5s | 64 × 64 | 96.5 | 5.3 | 0.68 |
YOLOv5s | 128 × 128 | 96.9 | 6.6 | 0.85 |
YOLOv5s | 512 × 512 | 114.9 | 6.1 | 0.94 |
YOLOv5m | 64 × 64 | 214.57 | 7.2 | 0.78 |
YOLOv5m | 128 × 128 | 215.07 | 8.4 | 0.92 |
YOLOv5m | 512 × 512 | 225.74 | 7.2 | 0.96 |
YOLOv5x | 64 × 64 | 763.85 | 10.1 | 0.68 |
YOLOv5x | 128 × 128 | 764.58 | 10.6 | 0.93 |
YOLOv5x | 512 × 512 | 780.48 | 15.2 | 0.95 |