Next Article in Journal
Azimuth Full-Aperture Processing of Spaceborne Squint SAR Data with Block Varying PRF
Next Article in Special Issue
Automatic Recognition of Road Damage Based on Lightweight Attentional Convolutional Neural Network
Previous Article in Journal
Evaluation of Machine Learning Techniques for Traffic Flow-Based Intrusion Detection
Previous Article in Special Issue
Caps Captioning: A Modern Image Captioning Approach Based on Improved Capsule Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Industrial Anomaly Detection with Skip Autoencoder and Deep Feature Extractor

1
Department of Mechanical Engineering, National Taiwan University, Taipei 10617, Taiwan
2
3DFAMILY Technology Co., Ltd., New Taipei 23674, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(23), 9327; https://doi.org/10.3390/s22239327
Submission received: 6 November 2022 / Revised: 25 November 2022 / Accepted: 26 November 2022 / Published: 30 November 2022

Abstract

:
Over recent years, with the advances in image recognition technology for deep learning, researchers have devoted continued efforts toward importing anomaly detection technology into the production line of automatic optical detection. Although unsupervised learning helps overcome the high costs associated with labeling, the accuracy of anomaly detection still needs to be improved. Accordingly, this paper proposes a novel deep learning model for anomaly detection to overcome this bottleneck. Leveraging a powerful pre-trained feature extractor and the skip connection, the proposed method achieves better feature extraction and image reconstructing capabilities. Results reveal that the areas under the curve (AUC) for the proposed method are higher than those of previous anomaly detection models for 16 out of 17 categories. This indicates that the proposed method can realize the most appropriate adjustments to the needs of production lines in order to maximize economic benefits.

1. Introduction

Owing to the developments in convolutional neural networks (CNN) over recent years, classical CNN models such as the ResNeXt [1], Dual Path Networks [2], and EfficientNet [3] have been proposed, which has been proven to possess good image recognition abilities. Consequently, it is great for researchers to continue working toward importing CNN models into inspection applications and obtain good results, such as for industrial inspection [4,5], cryo-electron tomogram classification [6], lithium-ion battery electrode defect detection [7], solar cell surface defect inspection [8], bolt joints monitoring [9] and rolling bearing robust fault diagnosis [10].
However, the training of CNNs requires many defect samples and necessitates additional manpower for defect labeling, which makes it difficult to import deep learning technology into actual production lines. Although many data augmentation methods have been proposed thus far [11,12,13,14], the application of CNNs in industrial testing remains limited.
As a solution to overcome this limitation, anomaly detection technology, an unsupervised learning method, has garnered significant research attention. In recent studies, anomaly detection was mainly realized using an autoencoder (AE) [15] or a Generative Adversarial Network (GAN) [16]. The principle involves training neural networks through good products to realize better feature reconstruction abilities for good products. With regard to production lines, an abnormal score is defined as the difference between reconstructed and original images or features. Anomaly detection methods such as AnoGAN [17], GANomaly [18], Skip-GANomaly [19], and DFR [20] have been proven to possess certain anomaly detection abilities. Their model architecture is shown in Figure 1.
Although these models have certain abilities in industrial anomaly detection, their poor feature extraction and reconstruction abilities limit their practical applications. Recently, scholars have attempted to overcome these problems by introducing skip connections [19], deep feature reconstruction algorithms [20], and multiple autoencoders [21]. In this paper, an anomaly detection model with a skip autoencoder and a deep feature extractor is proposed. This model was proven to possess better inspection ability than previous anomaly detection models for different datasets and under different conditions.
This paper mainly discusses the following issues regarding the proposed method:
  • The MVTecAD, furniture wood, and mobile phone cover glass datasets for production lines were used to train and verify the proposed model, which was then compared with previous anomaly detection models.
  • Different feature extractors were used to train the proposed model, and optimal feature extractor selection under different requirements was discussed.
  • The proposed model was trained with different feature extract layers, and the corresponding effects were discussed.

2. Related Works

Over recent years, many excellent anomaly detection models have been proposed. In this study, four important anomaly detection models were selected for introduction. These models were trained and verified experimentally and then compared with the proposed method. The advantages and limitations of these models are shown in Table 1.

2.1. AnoGAN

AnoGAN [17] was the first deep learning algorithm that employed a GAN to detect anomalies. Its principle entails using a good product to train the model such that the network can generate good product images. During detection, the image that is most similar to the image to be measured is generated via an iterative method, and the defect score is defined based on the residual of the two images. However, this method requires a significant amount of computing resources because of the numerous iterations.

2.2. GANomaly

To address the problem of high computing resource consumption, Akcay et al. proposed an anomaly detection algorithm, GANomaly [18]. In addition to the concept of GAN-based anomaly detection, this approach employed an autoencoder as a generator. This autoencoder significantly reduces the computing resource consumption and time required for GANomaly. However, owing to the lack of the feature extraction and image reconstruction abilities, the model remains limited.

2.3. Skip-GANomaly

To overcome the problem of insufficient image reconstruction ability, Skip-GANomaly has been proposed [19]. The authors used the skip connection architecture to significantly enhance the image reconstruction ability. This model shows good detection ability for most categories in the MVTecAD dataset. However, in certain categories with more complex defects, the detection ability of Skip-GANomaly is limited, indicating that its ability to extract depth features warrants further improvements.

2.4. Deep Feature Reconstruction (DFR)

Deep feature reconstruction (DFR) [20] is an effective anomaly segmentation method that can inspect and segment anomalous images by applying a multi-scale regional feature generator. This generator can generate multiple representations from a pre-trained CNN model of an image, making it discriminative and beneficial for anomaly detection. By leveraging these descriptive regional features, the authors applied an efficient autoencoder and inspected the anomalous regions within images via fast feature reconstruction. This method is efficient and has considerable potential for practical applications. However, the feature reconstruction ability of the autoencoder in DFR is still unsuitable for detecting small defects. Therefore, the autoencoder of DFR should be improved.

3. Proposed Method

3.1. Model Architecture

The structure of the proposed method, as shown in Figure 2, comprises a pre-trained feature extractor and skip connection-based autoencoder. Inspired by DFR [20] and Skip-GANomaly [19], the feature extractor (FE) was designed as a pre-trained CNN (ResNeXt101 was used instead of VGG19 in the DFR) model. It can extract important features from an input image x. After feature extraction, these features are resized and concatenated as a three-dimensional tensor and imported to the autoencoder.
Furthermore, the proposed method uses the skip connection adopted in Skip-GANomaly. On importing this structure, the autoencoder (AE) in the proposed method is expected to realize better feature reconstruction than DFR. Furthermore, the introduction of this structure ensures that the model exhibits better stability and a higher anomaly detection ability.
During the training process, only good samples were input into the model. Therefore, the autoencoder in the model possessed better capability for reconstructing the characteristics of good products. During the detection process, provided the appropriate residual score is defined to express the characteristic tensor FE(x) and the tensor reconstructed via AE(FE(x)), the function of anomaly detection and segmentation can be realized.
Overall, the proposed method combines the advantages of DFR and Skip-GANomaly, which makes it suitable for feature extraction and reduction.

3.2. Training Process

In the training process, the training data x were inputed into the deep feature extractor, which was pre-trained using the ImageNet dataset [22]. Furthermore, the weighting of the feature extractor was locked during the training process. By being pre-trained with a large amount of image data, the feature extractor can effectively extract a significant part of the training data. This step allows the model to obtain more features in the training process and makes the model considerably better than the model trained using the image directly.
After feature extraction, the feature was resized and concatenated as a three-dimensional tensor and then imported to the skip-connect-based autoencoder. Furthermore, to ensure that the model has the best image reconstruction ability for good samples, the loss function, which is termed as contextual loss, was applied in model training. This loss function indicates the difference between the feature tensor FE(x) and tensor AE(FE(x)) reconstructed by the autoencoder.
Mathematically, distance L2 was used to clearly define the difference between these two tensors. Therefore, the contextual loss also employed this distance definition. It can be expressed as follows:
L = E x p x [ | | A E ( F E ( ( x ) ) F E ( x ) | | 2 ]
The aim of training is to minimize contextual loss, which ensures that the model achieves the best feature reduction performance for normal samples and further improves the anomaly detection ability of the model.

3.3. Detection Process

In the detection phase, the image x to be tested is first inputted into the FE feature extractor (.) and is reshaped and concatenated as the feature tensor FE(x). Next, the feature tensor was input into the autoencoder for reconstruction, and the tensor AE(FE(x)) was output. The residual map, R(FE(x), AE(FE(x))), was used to calculate distance L2 for the two vectors. This can be expressed as:
R ( F E ( x ) , A E ( F E ( x ) ) ) = | | A E ( F E ( ( x ) ) F E ( x ) | | 2
As the proposed method only uses good products for training, it shows better feature reconstruction ability for good products. Therefore, the residual score in the residual map R(FE(x), AE(FE(x))) of the good products was lower than that of the bad products.
During detection, by adjusting different threshold values θ , the part where the residual score is greater than the threshold value is defined as the abnormal area.

4. Experimental Setup

4.1. Datasets

In this study, in addition to MVTec AD [23], which is widely used in anomaly detection, furniture wood and mobile phone glass-cover datasets on an actual production line were also applied to evaluate the detection ability of the proposed model. The following is an introduction to the three datasets.

4.1.1. MVTec AD

MVTec AD [23] is currently one of the most widely used industrial datasets for anomaly detection [24,25,26,27,28,29]. The dataset contains 3629 training images and 1725 verification data points. It includes 15 categories: five of these correspond with texture detection items, whereas the remaining pertain to object detection items. These multiple types of components make MVTec AD more comprehensive in terms of evaluating the detection ability of a deep learning model. Sample data are shown in Figure 3.

4.1.2. Production Line Smartphone Glass-Cover Dataset

This is a collection of smartphone cover glass data captured via a line-scan camera. It contains 200 pieces of mobile phone glass-cover. The images were scanned and divided into 329 training and 54 test images. In the testing data, there are 27 abnormal images, which contain various common types of defects on the production line. To balance the data, an equal number of normal images were added to the testing data.
Notably, smartphone glass-covers in a production line easily absorb dust, which can be conveniently eliminated. Therefore, the samples containing dust in this dataset are regarded as normal samples, which makes the detection of these data considerably more difficult.

4.1.3. Production Line Furniture Wood Dataset

This is an anomaly detection dataset for furniture wood in an actual production line. It contains 3075 training and 740 test images, which were captured via a line scan camera. There are six tag categories in this dataset: normal products, holes, chalk, knots, and black. In the testing data, there are 370 abnormal images, including five types of this production line, and the number of each defect types is between 70 and 80. To balance the data, an equal number of normal images were added to the testing data. Compared to the wood dataset in MVTec AD, this dataset has more complex defect categories and characteristics, which are suitable for verifying the detection ability of deep learning models with difficult samples.

4.2. Training Process

During the development of the proposed method, in order to make the method better applied to real industrial inspection projects, we strive to find a set of hyperparameters that can have good detection ability in most cases. The learning rate is from 0.001 to 0.00001. In addition, the number of epochs was set from 50 to 500. Finally, it was observed that using Adam [30] as the optimizer, all test items can converge well in the case of 150 epochs and a learning rate of 0.0001. Therefore, in the experiment, this set of hyperparameters and optimizers was applied to minimize the loss function. In addition, different feature extraction backbones were applied, including MobileNet v3 small [31], MobileNet v3 large [31], VGG19 [20], resnext50 [1], and resnext101 [1]. Furthermore, the most suitable backbone under different conditions was discussed. In addition, there is a four-blocks hierarchy in ResNeXt 101. Higher-level blocks output semantic features, while lower-level blocks output image features. Therefore, the performance of different sets of output blocks is tested and discussed. Since the 4th block has a huge bias for the pre-trained dataset, only the first three blocks were used in the study. A system equipped with an Intel i7-10700k CPU and Nvidia RTX2080ti GPU was used to train and test the models in this study.

4.3. Evaluation Method

In this study, the performance of each model was evaluated using the area under the curve (AUC) of the receiver operating characteristics. The AUC can help in comprehensively evaluating the detection ability of a detector under various threshold values. It is also the most widely used evaluation method for anomaly detection and segmentation.

5. Experiment Results

5.1. MVTec AD Dataset

MVTec AD can be used to determine the detection ability of anomaly detection models in various projects. As listed in Table 2, the proposed method has the highest AUC for 14 of the 15 categories. Moreover, it is worth noting that the proposed method shows considerably better anomaly detection performance under four categories, including cables, metal nuts, tiles, and transformers, as compared with the previous detection methods. All these inspection categories have complex variations among the images, and the defects have irregular textures, which necessitates better feature extraction ability. Because AnoGAN, GANomaly, and Skip-GANomaly do not have a pre-trained feature extractor, they suffer from the problem of insufficient feature extraction ability when detecting these categories, thereby resulting in poor detection ability. However, although DFR has the architecture of a pre-trained feature extractor, the image reconstruction ability of its autoencoder is not qualified. Conversely, the advantages of the proposed method were remarkable because of its strong feature reconstruction ability and effective pre-trained feature extractor. The heat maps and segmented masks of the MVTec AD are shown in Figure 4. It can be observed from the figure that the proposed method can accurately segment the abnormal parts in the image.

5.2. Production Line Smartphone Glass-Cover and Furniture Wood Datasets

The smartphone glass-cover and furniture wood datasets can be used to test the anomaly detection ability of an anomaly detection model under actual detection environments. As summarized in Table 3, the proposed method features the best detection performance for both datasets. Dust is present on the good smartphone glass-cover products, whereas the furniture wood has more variable textures and defect types, both of which significantly increase the difficulty of data detection. Because the other four anomaly detection deep learning models do not have sufficient feature extraction and reconstruction capabilities, their disadvantages will be encountered when detecting such complex production line data. The heat maps and segmented masks of the proposed method are shown in Figure 4, indicating that the proposed method has significant anomaly detection and segmentation abilities.

5.3. Discussion of Inference Time

The inference times of AnoGAN, GANomaly, Skip-GANomaly, and the proposed method are shown in Table 4. It can be observed that the run time of AnoGAN is inappropriate for real-time anomaly detecting since the iterative algorithm. Although the inference times of GANomaly and Skip-GANomaly are faster than DFR and the proposed method, their AUC scores are inappropriate for industry cases. On the other hand, the proposed method and DFR can detect almost 100 pictures in one second and have excellent AUC scores. Therefore, they are most appropriate for production lines.

5.4. Discussion of Feature Extractor

The AUCs of different feature extractors, including MobileNet v3 Small, MobileNet v3 large, VGG19, ResNeXt50, and ResNeXt101, are listed in Table 5. Figure 5a shows the correlation between the average AUC and feature extractor. It is expected that the depth of the feature extractor is positively correlated with the AUC. However, it is notable that the average AUC of ResNeXt101 was only 0.07 higher than that of MobileNet v3 small. This feature enables the proposed method to be used effectively in edge-computing devices. Even if fewer computing resources are used, the proposed method can maintain good detection ability.
In addition, the use of feature extractor blocks is discussed. The correlation between the AUCs and block usage in ResNeXt101 is summarized in Table 6 and Figure 5b. It was observed that a greater number of blocks had a significant impact on the detection results. This is because the higher the number of blocks the neural network layer contains, the higher-dimensional the semantic information, which can be better used by the autoencoder. Therefore, if it is necessary to reduce computing resource consumption during actual detection cases, block 3 or a combination of blocks 2 and 3 can be selected to maintain a detection ability similar to that of the original model.

6. Conclusions

In this study, a new anomaly detection and segmentation model was proposed. This model is improved based on the DFR architecture. Its novelty and contributions are as follows. First, we use the skip-connection architecture to improve the feature reduction ability of the autoencoder in DFR. Second, we change the feature extractor from the VGG model to ResNeXt101 and explore the best combination of output blocks. Third, in addition to the opened dataset, two sets of production line data collected by our team are used to verify the performance of the model in actual industrial applications. With the skip connection and deep feature extractor, the proposed model exhibits good feature extraction and reconstruction abilities. Therefore, its performance for MVTec AD and two groups of production line anomaly detection datasets is significantly better than those of previous models. Furthermore, this study discussed the performance of the model in terms of computing resources. The results indicate that the proposed model can maintain good detection and segmentation abilities even if it is replaced with a lighter feature extractor. This implies that the proposed model shows good performance and is suitable for application in actual production line inspection tasks. However, the proposed method still has the following two limitations and challenges. First, since the model focuses on the detection of details, it is relatively poor in detecting macro-scale defects. Second, the method using deep learning will still take a long time to train. These are the directions for our team to study and improve in the future research.

Author Contributions

T.-W.T.: Conceptualization, Data curation, Formal analysis, Methodology, Software, Validation, Visualization, Writing—original draft. H.H.: Conceptualization, Data curation, Investigation, Resources, Supervision, Writing—review and editing. W.-R.H.: Conceptualization, Data curation, Methodology, Validation, Visualization, Writing—review and editing. K.-M.L.: Conceptualization, Investigation, Project administration, Supervision, Writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

National Science and Technology Council, Taiwan. Grant number: 111-2622-E-002-020-.

Acknowledgments

The authors would like to thank National Science and Technology Council (Taiwan) for financially supporting this research under contract nos. Grant number: 111-2622-E-002-020-.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1492–1500. [Google Scholar]
  2. Chen, Y.; Li, J.; Xiao, H.; Jin, X.; Yan, S.; Feng, J. Dual Path Networks. arXiv 2017, arXiv:1707.01629. [Google Scholar]
  3. Tan, M.; Le, Q. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; Volume 97, pp. 6105–6114. [Google Scholar]
  4. Avdelidis, N.P.; Tsourdos, A.; Lafiosca, P.; Plaster, R.; Plaster, A.; Droznika, M. Defects Recognition Algorithm Development from Visual UAV Inspections. Sensors 2022, 22, 4682. [Google Scholar] [CrossRef] [PubMed]
  5. Huang, Y.; Xiang, Z. RPDNet: Automatic Fabric Defect Detection Based on a Convolutional Neural Network and Repeated Pattern Analysis. Sensors 2022, 22, 6226. [Google Scholar] [CrossRef] [PubMed]
  6. Gubins, I.; Chaillet, M.L.; van der Schot, G.; Veltkamp, R.C.; Förster, F.; Hao, Y.; Bunyak, F. SHREC 2020: Classification in cryo-electron tomograms. Comput. Graph. 2020, 279–289. [Google Scholar] [CrossRef]
  7. Badmos, O.; Kopp, A.; Bernthaler, T.; Schneider, G. Image-based defect detection in lithium-ion battery electrode using convolutional neural networks. J. Intell. Manuf. 2020, 31, 885–897. [Google Scholar] [CrossRef]
  8. Chen, H.; Pang, Y.; Hu, Q.; Liu, K. Solar cell surface defect inspection based on multispectral convolutional neural network. J. Intell. Manuf. 2020, 31, 453–468. [Google Scholar] [CrossRef] [Green Version]
  9. Yuan, R.; Lv, Y.; Wang, T.; Li, S.; Li, H. Looseness monitoring of multiple M1 bolt joints using multivariate intrinsic multiscale entropy analysis and Lorentz signal-enhanced piezoelectric active sensing. Struct. Health Monit. 2022, 21, 2851–2873. [Google Scholar] [CrossRef]
  10. Yuan, R.; Lv, Y.; Lu, Z.; Li, S.; Li, H. Robust fault diagnosis of rolling bearing via phase space reconstruction of intrinsic mode functions and neural network under various operating conditions. Struct. Health Monit. 2022, 14759217221091131. [Google Scholar] [CrossRef]
  11. Dunphy, K.; Fekri, M.N.; Grolinger, K.; Sadhu, A. Data Augmentation for Deep-Learning-Based Multiclass Structural Damage Detection Using Limited Information. Sensors 2022, 22, 6193. [Google Scholar] [CrossRef] [PubMed]
  12. Wang, L.; Tang, D.; Liu, C.; Nie, Q.; Wang, Z.; Zhang, L. An Augmented Reality-Assisted Prognostics and Health Management System Based on Deep Learning for IoT-Enabled Manufacturing. Sensors 2022, 22, 6472. [Google Scholar] [CrossRef] [PubMed]
  13. Zhang, H.; Cao, J.; Zheng, D.; Yao, X.; Ling, B.W.K. Deep Learning-Based Synthesized View Quality Enhancement with DIBR Distortion Mask Prediction Using Synthetic Images. Sensors 2022, 22, 8127. [Google Scholar] [CrossRef] [PubMed]
  14. Stern, M.L.; Schellenberger, M. Fully Convolutional Networks for Chip-wise Defect Detection Employing Photoluminescence Images. J. Intell. Manuf. 2021, 32, 113–126. [Google Scholar] [CrossRef] [Green Version]
  15. An, J.; Cho, S. Variational Autoencoder based Anomaly Detection using Reconstruction Probability. SNU Data Min. Cent. 2015, 2, 1–18. [Google Scholar]
  16. Zenati, H.; Foo, C.S.; Lecouat, B.; Manek, G.; Chrasekhar, V.R. Efficient GAN-Based Anomaly Detection. arXiv 2019, arXiv:1611.05431. [Google Scholar]
  17. Schlegl, T.; Seeböck, P.; Waldstein, S.M.; Schmidt-Erfurth, U.; Langs, G. Unsupervised Anomaly Detection with Generative Adversarial Networks to Guide Marker Discovery. In Proceedings of the International Conference on Information Processing in Medical Imaging, Boone, NC, USA, 25–30 June 2017; pp. 147–157. [Google Scholar]
  18. Akcay, S.; Atapour-Abarghouei, A.; Breckon, T.P. GANomaly: Semi-supervised Anomaly Detection via Adversarial Training. In Proceedings of the Asian Conference on Computer Vision, Perth, WA, Australia, 2–6 December 2018; pp. 622–637. [Google Scholar]
  19. Akçay, S.; Atapour-Abarghouei, A.; Breckon, T.P. Skip-GANomaly: Skip Connected and Adversarially Trained Encoder-Decoder Anomaly Detection. In Proceedings of the International Joint Conference on Neural Networks, Budapest, Hungary, 14–19 July 2019. [Google Scholar]
  20. Yang, J.; Shi, Y.; Qi, Z. DFR: Deep Feature Reconstruction for Unsupervised Anomaly Segmentation. arXiv 2020, arXiv:2012.07122. [Google Scholar]
  21. Tang, T.W.; Kuo, W.H.; Lan, J.H.; Ding, C.F.; Hsu, H.; Young, H.T. Anomaly Detection Neural Network with Dual Auto-Encoders GAN and Its Industrial Inspection Applications. Sensors 2020, 20, 3336. [Google Scholar] [CrossRef] [PubMed]
  22. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Fei-Fei, L. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  23. Bergmann, P.; Fauser, M.; Sattlegger, D.; Steger, C. MVTec AD—A Comprehensive Real-World Dataset for Unsupervised Anomaly Detection. Int. J. Comput. Vis. 2021, 129, 1038–1059. [Google Scholar] [CrossRef]
  24. Liznerski, P.; Ruff, L.; Vandermeulen, R.A.; Franks, B.J.; Kloft, M.; Müller, K.R. Explainable Deep One-Class Classification. arXiv 2021, arXiv:2007.01760. [Google Scholar]
  25. Dehaene, D.; Frigo, O.; Combrexelle, S.; Eline, P. Iterative energy-based projection on a normal data manifold for anomaly localization. arXiv 2020, arXiv:2002.03734. [Google Scholar]
  26. Zavrtanik, V.; Kristan, M.; Skočaj, D. Reconstruction by inpainting for visual anomaly detection. Pattern Recognit. 2021, 112, 107706. [Google Scholar] [CrossRef]
  27. Dehaene, D.; Eline, P. Anomaly localization by modeling perceptual features. arXiv 2020, arXiv:2008.05369. [Google Scholar]
  28. Kim, J.H.; Kim, D.H.; Yi, S.; Lee, T. Semi-orthogonal Embedding for Efficient Unsupervised Anomaly Segmentation. arXiv 2021, arXiv:2105.21737. [Google Scholar]
  29. Chen, Y.; Tian, Y.; Pang, G.; Carneiro, G. Deep One-Class Classification via Interpolated Gaussian Descriptor. arXiv 2022, arXiv:2101.10043. [Google Scholar] [CrossRef]
  30. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  31. Howard, A.; Sandler, M.; Chu, G.; Chen, L.C.; Chen, B.; Tan, M.; Adam, H. Searching for MobileNetV3. arXiv 2019, arXiv:1905.02244. [Google Scholar]
Figure 1. Architectures of: (a) AnoGAN, (b) GANomaly, (c) Skip-GANomaly, and (d) DFR.
Figure 1. Architectures of: (a) AnoGAN, (b) GANomaly, (c) Skip-GANomaly, and (d) DFR.
Sensors 22 09327 g001
Figure 2. Structure of proposed method.
Figure 2. Structure of proposed method.
Sensors 22 09327 g002
Figure 3. MVTec AD, production line smartphone glass-cover, and furniture wood datasets for industrial inspection.
Figure 3. MVTec AD, production line smartphone glass-cover, and furniture wood datasets for industrial inspection.
Sensors 22 09327 g003
Figure 4. Segmented masks and heat maps of each category created using the proposed method.
Figure 4. Segmented masks and heat maps of each category created using the proposed method.
Sensors 22 09327 g004
Figure 5. Average AUCs of all categories with: (a) different pre-trained feature extractor; (b) different blocks of ResNeXt101 feature extractor.
Figure 5. Average AUCs of all categories with: (a) different pre-trained feature extractor; (b) different blocks of ResNeXt101 feature extractor.
Sensors 22 09327 g005
Table 1. Comparison of AnoGAN, GANomaly, Skip-GANomaly, and DFR.
Table 1. Comparison of AnoGAN, GANomaly, Skip-GANomaly, and DFR.
AnoGANGANomalySkip-GANomalyDFR
AdvantagesTraining the model without abnormal data.Much less inspection time than AnoGAN.Better ability of feature reconstructing.Great ability of feature extraction.
LimitationsRequires a significant amount of computing resources.Feature extracting and reconstructing abilities are limited.Cannot extract complex image feature.Limited ability of feature reconstruction.
Table 2. AUCs of MVTec AD dataset using AnoGAN, GANomaly, Skip-GANomaly, and the proposed method.
Table 2. AUCs of MVTec AD dataset using AnoGAN, GANomaly, Skip-GANomaly, and the proposed method.
CategoryAnoGANGANomalySkip-GANomalyDFRProposed Method
Bottle0.820.820.910.940.98
Cable0.770.640.650.870.95
Capsule0.850.750.720.970.99
Carpet0.550.830.520.980.98
Grid0.590.890.850.960.91
Hazelnut0.40.940.830.980.98
Leather0.640.810.820.990.99
Metal nut0.440.650.670.920.98
Pill0.760.670.80.950.98
Screw0.790.90.920.970.98
Tile0.520.650.680.890.97
Toothbrush0.880.850.780.970.99
Transistor0.780.70.810.780.87
Wood0.650.950.920.940.97
Zipper0.770.670.670.950.98
Average0.680.780.770.940.97
Table 3. AUCs of production line datasets using AnoGAN, GANomaly, Skip-GANomaly, and the proposed method.
Table 3. AUCs of production line datasets using AnoGAN, GANomaly, Skip-GANomaly, and the proposed method.
CategoryAnoGANGANomalySkip-GANomalyDFRProposed Method
Glass (PL)0.650.640.820.990.99
Wood (PL)0.710.840.80.920.94
Average0.680.740.810.960.97
Table 4. Inference Time of AnoGAN, GANomaly, Skip-GANomaly, and the proposed method.
Table 4. Inference Time of AnoGAN, GANomaly, Skip-GANomaly, and the proposed method.
AnoGANGANomalySkip-GANomalyDFRProposed Method
Time (ms)70252.682.8210.1011.20
Table 5. AUCs of proposed method with different pre-trained CNN feature extractors.
Table 5. AUCs of proposed method with different pre-trained CNN feature extractors.
CategoryMobileNet (S)MobileNet (L)VGG19ResNeXt50ResNeXt101
Bottle0.940.920.960.980.98
Cable0.870.880.920.940.95
Capsule0.940.940.980.990.99
Carpet0.880.90.980.950.98
Grid0.780.860.980.90.91
Hazelnut0.950.950.980.980.98
Leather0.980.980.980.980.99
Metal nut0.870.950.940.980.98
Pill0.920.920.970.960.98
Screw0.920.950.980.980.98
Tile0.980.950.910.970.97
Toothbrush0.950.940.980.980.99
Transistor0.740.660.790.820.87
Wood0.930.940.950.970.97
Zipper0.780.880.970.960.98
Glass (PL)0.990.990.990.990.99
Wood (PL)0.920.880.930.930.94
Average0.90.910.950.960.97
MobileNet (S)–MobileNet v3 Small, MobileNet (L)–MobileNet v3 Large.
Table 6. AUCs of proposed method with different feature extractor blocks.
Table 6. AUCs of proposed method with different feature extractor blocks.
CategoryBlock1Block2Block3Block1,2Block2,3Block1,2,3
Bottle0.840.960.970.940.970.98
Cable0.880.920.950.910.950.95
Capsule0.930.970.980.960.980.99
Carpet0.930.920.950.940.960.98
Grid0.880.910.870.920.90.91
Hazelnut0.970.960.970.970.980.98
Leather0.980.990.970.980.970.99
Metal nut0.960.960.950.960.970.98
Pill0.920.950.970.950.980.98
Screw0.970.970.960.970.970.98
Tile0.970.970.950.970.960.97
Toothbrush0.920.960.980.960.980.99
Transistor0.660.710.880.710.870.87
Wood0.960.960.940.960.960.97
Zipper0.920.940.960.940.970.98
Glass (PL)0.990.990.990.990.990.99
Wood (PL)0.880.890.920.890.930.94
Average0.920.940.950.940.960.97
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tang, T.-W.; Hsu, H.; Huang, W.-R.; Li, K.-M. Industrial Anomaly Detection with Skip Autoencoder and Deep Feature Extractor. Sensors 2022, 22, 9327. https://doi.org/10.3390/s22239327

AMA Style

Tang T-W, Hsu H, Huang W-R, Li K-M. Industrial Anomaly Detection with Skip Autoencoder and Deep Feature Extractor. Sensors. 2022; 22(23):9327. https://doi.org/10.3390/s22239327

Chicago/Turabian Style

Tang, Ta-Wei, Hakiem Hsu, Wei-Ren Huang, and Kuan-Ming Li. 2022. "Industrial Anomaly Detection with Skip Autoencoder and Deep Feature Extractor" Sensors 22, no. 23: 9327. https://doi.org/10.3390/s22239327

APA Style

Tang, T. -W., Hsu, H., Huang, W. -R., & Li, K. -M. (2022). Industrial Anomaly Detection with Skip Autoencoder and Deep Feature Extractor. Sensors, 22(23), 9327. https://doi.org/10.3390/s22239327

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop