Next Article in Journal
Loaded Plyometrics and Short Sprints with Change-of-Direction Training Enhance Jumping, Sprinting, Agility, and Balance Performance of Male Soccer Players
Next Article in Special Issue
A Roadmap to Integrate Digital Twins for Small and Medium-Sized Enterprises
Previous Article in Journal
Security of Blockchain-Based Supply Chain Management Systems: Challenges and Opportunities
Previous Article in Special Issue
Image Anomaly Detection Using Normal Data Only by Latent Space Resampling
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Data Augmentation to Achieve Noise Robustness and Overcome Data Deficiency for Deep Learning

Department of Electrical and Electronics Engineering, Pusan National University, Busan 46241, Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(12), 5586; https://doi.org/10.3390/app11125586
Submission received: 29 April 2021 / Revised: 4 June 2021 / Accepted: 15 June 2021 / Published: 17 June 2021

Abstract

:
Artificial intelligence technologies and robot vision systems are core technologies in smart factories. Currently, there is scholarly interest in automatic data feature extraction in smart factories using deep learning networks. However, sufficient training data are required to train these networks. In addition, barely perceptible noise can affect classification accuracy. Therefore, to increase the amount of training data and achieve robustness against noise attacks, a data augmentation method implemented using the adaptive inverse peak signal-to-noise ratio was developed in this study to consider the influence of the color characteristics of the training images. This method was used to automatically determine the optimal perturbation range of the color perturbation method for generating images using weights based on the characteristics of the training images. The experimental results showed that the proposed method could generate new training images from original images, classify noisy images with greater accuracy, and generally improve the classification accuracy. This demonstrates that the proposed method is effective and robust to noise, even when the training data are deficient.

1. Introduction

Recent advances in the Internet of Things, big data, cloud computing, and Industry 4.0 are rapidly revolutionizing the manufacturing industry. Industry 4.0, which is related to automation and digitization, is closely linked to smart factories. Smart factories describe the intelligent factory of the future and their core technologies are artificial intelligence (AI) and robot vision systems [1,2,3,4].
Owing to the development of AI-related technologies, such as deep learning algorithms, the computational power associated with the growth of graphics processing units, and the collection of big data using large-scale sensor networks, AI has been extensively researched. Furthermore, it has undergone rapid development. Robot vision is the prime perception channel used in manufacturing-related technologies [5]. Accordingly, vision systems have been used in a diverse range of applications in the manufacturing field such as inspection monitoring systems, manipulation, picking and placing, object recognition, and mobile robotics [6,7,8,9].
Specifically, combining deep learning, vision systems, and smart factories is a recent trend in the manufacturing sector. Deep learning has been utilized in various manufacturing applications, including autosorting systems, inspection systems, maintenance in mechanical manufacturing, fault classification and diagnosis, and classification systems. For example, to pick and place objects using a manipulator, the objects must be recognized and classified using deep learning and a robot vision system. Object recognition and classification can be achieved using deep learning [10,11,12,13,14].
However, there is a crucial problem in applying deep learning. For example, a network trained to classify objects into categories may be considered. In addition, some noise may be added to the image that falls under a specific category in the trained network. In this case, the noisy image may be misclassified, although the original image without noise is classified correctly. In other words, noise adversely affects the accuracy of image classification. Thus, deep learning models are vulnerable to noise attacks. An autonomous driving system applying deep learning can get into an accident if it is attacked with a noisy image. Such attacks are called adversarial attacks [15,16].
Another problem when applying deep learning algorithms is that it is necessary to prepare sufficient training images to train the deep learning network. Numerous diverse datasets that can be used as training images are available on the Internet. However, in some cases, it is difficult to collect the image data of unusual objects, which are not included in the training dataset of the deep learning model, as opposed to common items. Specifically, in the manufacturing field—for instance, in the electronics industry—many objects are uncommon. There is a limit to capturing and collecting the image data of unusual objects. Therefore, a data augmentation method that can generate images automatically is required to overcome data deficiency. Many researchers are studying data augmentation methods toward this end. Data augmentation improves the generalization capabilities of the deep learning network and the performance of the classification model [17,18].
In this paper, a data augmentation method is proposed to achieve noise robustness and overcome data deficiencies. Our study makes three main contributions to the literature as follows:
  • Automatic determination of the optimal perturbation range based on image similarity;
  • Weight calculation of color perturbation based on the characteristics of the color distribution of training images;
  • Data augmentation based on color perturbation and geometric transformations to compensate data deficiency and noisy images.
The proposed method suggests augmenting training data based on the concept of color jittering. Unlike conventional methods, the proposed method applies to perturb the color information by maintaining the similarity with an original image in the special range we suggest and considering color histogram information of the classified objects. Additionally, the reason why we consider the concept of color jittering is that the noise is related to pixel values. The proposed method automatically determines the optimal perturbation range by calculating the image similarity between the original and noisy images. In addition, we analyzed the color distribution based on the histograms of the training images. The weights were calculated using color perturbation based on the color distribution. We then generated new training images using the color perturbation method based on the image similarity to the calculated weights. Subsequently, the pretrained network was retrained using the images generated by the proposed method. To validate the proposed method, we compared its performance with that of the image classification results of the trained models using the test images.
The remainder of this paper is organized as follows: In Section 2, we review related studies on data augmentation, adversarial attacks, and similarity calculations. In Section 3, we describe the proposed data augmentation method that surmounts data deficiencies and guarantees noise robustness. In Section 4, we present the experimental results of the proposed method. Finally, we present our conclusions in Section 5.

2. Related Works

To implement the deep learning network to classify objects for a manipulator for picking and placing or assembling, sufficient training images must be prepared to train the deep learning network. This is because the deep learning network is trained using images for object classification. Accordingly, deficient training images may affect object classification accuracy. However, it is difficult to collect numerous training images, especially because uncommon objects in the manufacturing field are not included in training datasets. Therefore, it is essential to increase the number of training images through data augmentation [19].
Furthermore, noisy images affect classification accuracy. Even if a noisy image is very similar to the original image without noise, the deep learning network may misclassify the noisy images. Therefore, to avoid misclassification, the network must be trained such as to be noise robust. To compensate for these drawbacks, a data augmentation method that exploits image similarity and color perturbation based on the color characteristics of training images is proposed in the paper. This section deals with data augmentation, adversarial attacks, and image similarity. They are related to the proposed method for overcoming data deficiency and achieving noise robustness.

2.1. Data Augmentation

Sometimes data deficiency occurs because the images of objects that should be classified are insufficient or not included in the providing dataset. Data augmentation is being researched to overcome it. To increase the number of training images, data augmentation methods such as image rotation, flipping, scaling, cropping, color jittering, and shearing can be applied [20,21,22].
Color space transformation methods, such as color perturbation, edge enhancement, and principal component analysis (PCA), are also used. The color perturbation method can be applied by extracting a single-color channel or adding a random value to the color channel. It can also be performed using a color histogram or PCA of the color channel [19]. Therefore, we augmented the training data to counter data deficiency based on color perturbation because this method is simple and has a relatively short computation time.
In our experiment, the objects for image classification were placed on a conveyor. Then, we considered the rotation of the geometric transformation using (1).
R o t a t i o n = cos Θ sin Θ sin Θ cos Θ

2.2. Adversarial Attack

Recent studies have revealed that deep learning algorithms are vulnerable to adversarial attacks. In 2013, it was found that a certain barely perceptible perturbation could maximize the error in the classification result. Therefore, adversarial attacks indicate that deep learning models may have inherent weaknesses [15,16].
Examples of adversarial attacks are shown in Figure 1. We only added random noise to images with Gaussian-distributed random numbers, and the original images and noisy images were classified using pretrained networks [23]. The experimental results indicate that, after adding noise to images of hot-dogs, the obtained predictions were incorrect: tarantula, cockroach, and mousetrap. The experimental results prove that noisy images may be labeled incorrectly, although the original image may be labeled correctly.
Thus, noise may affect the prediction results adversely and degrade the performance of the classification model. Especially, in cases of object classification in the manufacturing sector, adversarial attacks would reduce the classification accuracy by causing a malfunction that is related to defect and longer working time. To achieve robustness against adversarial attacks to avoid misclassification, we consider Gaussian-distributed random numbers for a weighted color perturbation method.

2.3. Image Similarity Calculation

To implement a data augmentation method that overcomes data deficiency and is robust against noise, we considered tuning the values of pixels in an image using Gaussian-distributed random numbers. We applied Gaussian noise because it is similar to actual noise. The criterion for the perturbation range was automatically determined by the proposed method. To determine a reasonable range for color perturbation, the peak signal-to-noise ratio (PSNR) was used to calculate the weights of the original and noisy images.
The similarity between two images was calculated using the PSNR (2) as follows [24,25]:
P S N R = 10 log 10 M A X I 2 M S E = 20 log 10 M A X I M S E = 20 log 10 M A X I 10 log 10 M S E
In (2), M A X I indicates the maximum fluctuation based on the type of image. As the pixel values of our training images ranged from 0 to 255, the value of M A X I was 255. Equation (3) represents the M S E calculation.
M S E = 1 m n i = 1 m j = 1 n I i , j K i , j 2 .
This equation demonstrates that the M S E is associated with the difference between the two images, I and K. Furthermore, this implies a correlation between the P S N R and the difference between both images. Based on this concept, the PSNR was applied to determine the perturbation range and generate new images using a color perturbation matrix.

3. Proposed Method

A data augmentation method using color perturbation based on the color characteristics of the image and Gaussian noise was proposed to make a deep learning network robust against adversarial attacks. The proposed data augmentation method can overcome data deficiencies and improve the classification accuracy. Use of a suitable data augmentation method is key to image classification.
Therefore, the proposed method focuses on color perturbations. To apply color perturbation, we tuned the pixels of the image using the inverse PSNR, Gaussian-distributed random numbers, and weights calculated by histogram. An overview of the proposed method is presented in Figure 2.
As shown in this figure, the proposed method can be divided into six parts as follows: capturing image; decision of perturbation range (Figure 3); weight calculation (Figure 6); image preprocessing; color perturbation; geometric transformation. To establish the training image set, the objects to be classified were captured using the vision sensor. Based on the captured images, the perturbation value was calculated to determine the color perturbation range randomly. Further, the weights were calculated by using the histogram of the captured images. These weights were used when calculating the perturbation matrix. Background elimination was performed in image preprocessing. Next, color perturbation and geometric transformation were carried out to augment the training images.
First, we collected training images captured using a vision sensor. We loaded the training images to calculate the perturbation range automatically. The proposed method suggests a technique for determining an ideal perturbation range. This process is illustrated in detail in Figure 3.
The perturbation range was determined using images extracted arbitrarily from the training dataset. Random noise was generated and incorporated into an image. The PSNR was calculated using the original image and the noisy image. As mentioned previously, the PSNR indicates the similarity between the two images. We used the characteristics to determine the perturbation range. The PSNR was calculated for all randomly extracted images. Following the PSNR calculation, the perturbation range was determined for inverse PSNR data augmentation.
Next, we generated new training images by the adaptive inverse PSNR data augmentation. This process, which is related to the inverse PSNR, was based on the following calculation (4).
10 log 10 1 m n i = 1 m j = 1 n I i , j K i , j 2 = 20 log 10 M A X I 2 P S N R , i = 1 m j = 1 n I i , j K i , j 2 = m n 10 2 log 10 M A X I 2 0.1 P S N R
This equation, which was obtained from (2), yielded the optimal range for color perturbation. As we considered a data augmentation method based on the value of image similarity, we implemented the PSNR equation inversely.
After loading the training images into categories and determining the perturbation value, the alpha channel was extracted from an image. The alpha channel is a criterion for isolating an object from its background. The region of an object is the region of interest (ROI). ROI is extracted to tune only the pixels of an object in an image.
In addition, we analyzed the color characteristics of the training images in each category based on their histograms to improve the classification accuracy. First, we focused on color, a notable characteristic of an image. Figure 4 shows the RGB channel histograms, which are calculated by accumulating the pixel values of the images in each category. These images were included in the training images for object classification using deep learning. However, it was difficult to find distinct characteristics because black was the dominant color of some objects. Therefore, we rewrote the histogram, excluding the black objects. The rewritten histogram is shown in Figure 5.
Figure 5 shows some objects and their histogram distributions, with respect to the RGB color space. As shown in the figure, the training images were generally red and green, rather than blue. Therefore, we determined that the color perturbation method gave more weight to blue than to red and green. In the proposed method, the abovementioned process was implemented automatically, as shown in Figure 6.
The ideal weights for color perturbation were calculated, as depicted in Figure 6. The weight values indicated that the characteristics of the training images by histogram results were considered during the color perturbation. The summation of the weight values of each channel was three, as shown in Equation (5).
W r e d _ c h a n n e l + W g r e e n _ c h a n n e l + W b l u e _ c h a n n e l = 3
The weights of the W r e d _ c h a n n e l , W g r e e n _ c h a n n e l , and W b l u e _ c h a n n e l were initially one. In the case of our training images, the blue channel had a higher weight value than the red and green channels. After histogram analysis, as shown in Figure 6, the weights were changed by the histogram distribution. In our experiment, the W b l u e _ c h a n n e l had a higher value than the W r e d _ c h a n n e l and W g r e e n _ c h a n n e l because the training images often contain red or green colors.
Next, the perturbation matrix is the perturbation value based on the perturbation range, and the weights were multiplied. A three-dimensional perturbation matrix was generated using Gaussian-distributed random numbers. It was multiplied by the perturbation value and the weights, and combined with a multiplied matrix and an image. The combined image was rotated from 0° to 360°. Rotation was used because the objects were randomly laid on the conveyor. Subsequently, the generated image was saved. This process was repeated until the final range decision value and final image of the final category. When this process was completed, the training dataset was generated through adaptive inverse PSNR data augmentation, and used to retrain the pretrained network for object classification. Transfer learning was deployed to reduce the training time and train the network efficiently [23].
To verify this concept by applying the weights calculated based on the histogram of the objects to be classified, we deployed the color perturbation method by changing the pixels in each channel of the color space. The process is depicted in Figure 7.
The training images were loaded; we isolated the alpha channel from the image to extract the ROI in which an object was contained. The alpha channel was calculated during the preprocessing step through a background elimination process. To eliminate the image background, leaving only the object, we combined the RGB color space and CIE L*a*b* color space. Although the RGB color space is a well-known color space, it is difficult to isolate the object from the background using only the RGB color space. Therefore, we also deployed the CIE L*a*b* color space, which consists of two chromaticity layers and one luminosity layer. One chromaticity layer was related to the red and green axes, and the other was related to the blue and yellow axes. Therefore, we found and eliminated background regions based on background color information using the RGB and CIE L*a*b* color spaces [26].
Next, a one-dimensional perturbation matrix was generated using Gaussian-distributed random numbers. This matrix was multiplied by the range of decision values. The perturbation matrix was combined with the target component, which was extracted by separating the color channel from the image. After combining the perturbation matrix and the target component, the generated image was saved. This process was repeated when the range decision value was the final value. In our experiment, to determine whether the higher weight value of the blue color was effective, all three components—that is, RGB—were individually tested. The experimentally verified results are explained in detail in Section 4.

4. Experimental Results

The goal of our experiment was to classify the objects into ten categories for grasping the manipulator. We applied a deep learning network for object classification. However, because the objects were not common items, it was difficult to collect sufficient training images to train the deep learning network. Therefore, using the proposed method, training images were generated to improve classification accuracy. There were 32 training images for each category. There were ten categories. The proposed data augmentation method increased the number of training images. In addition, transfer learning was applied for effective learning and to reduce the computational time because it uses a pretrained model that has already learned the features of numerous images. We selected VGGNet with 19 weights layers for transfer learning. VGGNet was already trained by the images of 1000 categories. We cut off and retrained the network using generated images by the methods [23,27].
To verify the proposed method, we compared the classification accuracy of the network trained using the proposed method and the conventional methods. In addition, we trained and compared the deep learning network using the original data and data generated using the proposed method. The test results are presented in the tables. Test images were captured supplementally by placing the objects arbitrarily. They were not included in the training dataset. The classification accuracy is calculated by inputting test images into the trained model.
Table 1 presents the classification accuracy of the network trained using original images. As abovementioned, there were 320 training images, 32 for each category. The experimental results show that the classification accuracy was low when noiseless test images were used because the original images were few. In general, the classification accuracy of noisy images was less than that of noiseless images. Thus, the classification accuracy of noisy images resulted in lower values.
Table 2 presents the classification accuracy of the network trained by the conventional method using original images. Data augmentation was achieved by changing the range of color jittering. In general, color jittering is conducted by adding random values to a color channel. Based on this concept, color jittering #1 #3 add the random values in terms of RGB color space. Color jittering #1 adds positive values to the pixel values of an original image. Color jittering #2 and #3 add positive or negative values with different ranges [19]. When the conventional method was deployed, the number of training images increased by nine times, from 320 to 2880. The experimental results showed that the classification accuracy of the noisy images improved slightly. However, the classification accuracy tended to be low.
To improve the classification accuracy, we eliminated the background and cropped an image by considering the center of the object. Table 3 presents the classification accuracy of the network trained using preprocessed images. The classification accuracy improved slightly over that of the original images and noisy images in Table 1.
Next, we implemented a data augmentation method based on the inverse PSNR, with respect to the RGB color space and rotation. The number of training images increased by 216 times. Table 4 presents the classification accuracy of the network trained using the inverse PSNR, with respect to the RGB color spac, and rotation of the images. As shown in the table, the performance of the trained network improved significantly and the classification accuracy was higher than that of the network trained using the original images and the preprocessed images. Specifically, the classification performance on the noisy images increased significantly. Furthermore, the classification performance on the noiseless images increased.
As mentioned in Section 3, because the objects for image classification generally had more red and green colors than blue, we determined that in the color perturbation method, a higher weight would most likely result in blue. To provide reasonable evidence, we trained the network using images obtained by applying the inverse PSNR to three channels individually. The classification results are listed in Table 5.
Comparing the above experimental results, the classification results obtained by applying the inverse PSNR to the B channel were better than those obtained from the images with the inverse PSNR applied to the R and G channels. Therefore, we tuned the weights to increase the weight of the B channel and decrease the weights of the R and G channels. We tested two cases in which the weights were tuned. The results are presented in Table 6.
As shown in the table, the classification accuracy was higher than that of the network trained by other methods. Specifically, the classification performance on the noisy images was higher than 70%. A comparison of tables revealed that the proposed method had the best performance. It is evident that applying weights based on the color tendency of the training images was effective. To validate the proposed method, the network was trained with an equal number of training images of conventional methods. For this, we randomly selected 2880 training images from the images generated by the proposed method. The model is trained three times using selected training images. In case of noiseless images, the average of the classification accuracy of the model trained by the 2880 training images from the images generated by the proposed method was 87.3%. The classification performance on the noisy images was 67.0%. This additional experiment shows that the proposed method could augment the training images effectively.
Figure 8 represents the comparison of the classification accuracy, with respect to average value and best performance. This figure shows that the more the proposed algorithms are combined, the higher the classification accuracy. Furthermore, the proposed method outperformed the conventional method. Therefore, the proposed method could improve the classification accuracy of both noisy and noiseless test images.
We compared the confusion matrices of the experimental results for the training images in Figure 9 and Figure 10. The left and right columns represent the classification results for the original and noisy test images, respectively. The experimental results for the original test images showed that the classification accuracy improved significantly after application of the adaptive inverse PSNR. As depicted in the left column of Figure 9 and Figure 10, the classification accuracy was higher than 90%. The experimental results of the noisy test images showed that the classification accuracy also improved after application of the adaptive inverse PSNR. Specifically, the classification accuracy of the model trained using the original images was 25%. However, the proposed method showed a significant improvement, from 25% to 76%. Therefore, it was demonstrated that the proposed method could effectively compensate for data deficiencies and was robust to noisy images.
Table 7 presents the precision, recall, specificity, and F1-score of the conventional method and the proposed method. Precision, recall, specificity, and F1-score were calculated using (6).
Precision = True Positive True Positive + False Positive Recall = True Positive True Positive + False Negative Specificity = True Negative True Negative + False Positive F 1 - score = 2 Recall × Precision Recall + Precision
As shown in the table, in case of the proposed method, precision and recall values have a value closer to one than the conventional method. Therefore, the proposed method could result higher classification performance than the conventional method.
The additional experiment was conducted about the illumination change of the images when captured. One of common applications of the smart factory is AOI inspection. It is important to maintain a stable lighting source for AOI inspection. As per the experimental results, in case of the trained model using original images, the classification accuracy is 45.89%. The classification accuracy of the trained model by the conventional method is 43.67%, and that of the trained model by the proposed method is 82.89%. The experimental results show that the classification performance of the proposed method is higher than other trained models; namely, it represents that the proposed method could be robust against the illumination change.

5. Conclusions

We developed a data augmentation method to overcome training data deficiency, achieve robustness against noise that causes classification errors, and improve the image classification accuracy. To examine the effect of the characteristics of the training images, we employed the adaptive inverse PSNR, which automatically determined the ideal perturbation range and weights for the color perturbation method, which were used to generate new training images. Weights are useful for classifying noisy images because the weights are greater than the color distribution.
Experiments were performed to compare the classification models trained by the original data, the data generated by the conventional method, and the data generated by the proposed method using the test images. The experimental results showed that the images generated based on the adaptive inverse PSNR using the rotation method were more effective than the training data generated by the proposed method. In addition, the experimental results showed that the proposed method could generate new images as training data effectively, was robust against noisy images, and improved the image classification accuracy. In future, we plan to develop a data augmentation method that considers the characteristics of each image to overcome data deficiencies in the manufacturing domain. Further, we will apply the proposed method in other AI models such as object detection or semantic segmentation and so on.

Author Contributions

Conceptualization, E.K. and H.L.; methodology, E.K.; software, E.K.; validation, E.K., H.L. and J.K.; formal analysis, E.K.; investigation, E.K.; resources, E.K. and J.K.; data curation, E.K. and J.K.; writing—original draft preparation, E.K.; writing—review and editing, S.K.; visualization, E.K. and H.L.; supervision, S.K.; project administration, S.K.; funding acquisition, S.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Technology Innovation Program (10073147, Development of Robot Manipulation Technology by Using Artificial Intelligence) funded By the Ministry of Trade, Industry & Energy (MOTIE, Korea).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to undergoing project.

Acknowledgments

This research was supported by BK21PLUS, Creative Human Resource Education and Research Programs for ICT Convergence in the 4th Industrial Revolution, and was supported by the Technology Innovation Program (10073147, Development of Robot Manipulation Technology by Using Artificial Intelligence) funded By the Ministry of Trade, Industry & Energy (MOTIE, Korea) and was supported by Korea Institute for Advancement of Technology (KIAT) grant funded by the Korea Government (MOTIE) (P0008473, HRD Program for Industrial Innovation).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Tao, F.; Qi, Q. New IT driven service-oriented smart manufacturing: Framework and characteristics. IEEE Trans. Syst. Man Cybern. Syst. 2017, 49, 81–91. [Google Scholar] [CrossRef]
  2. Ghobakhloo, M. The future of manufacturing industry: A strategic roadmap toward Industry 4.0. J. Manuf. Technol. Manag. 2018, 29, 910–936. [Google Scholar] [CrossRef] [Green Version]
  3. Hozdić, E. Smart factory for industry 4.0: A review. Int. J. Mod. Manuf. Technol. 2015, 7, 28–35. [Google Scholar]
  4. Shi, Z.; Xie, Y.; Xue, W.; Chen, Y.; Fu, L.; Xu, X. Smart factory in Industry 4.0. Syst. Res. Behav. Sci. 2020, 37, 607–617. [Google Scholar] [CrossRef]
  5. Frese, U.; Hirschmüller, H. Special issue on robot vision: What is robot vision? J. Real Time Image Process. 2015, 10, 597–598. [Google Scholar] [CrossRef] [Green Version]
  6. Edinbarough, I.; Balderas, R.; Bose, S. A vision and robot based on-line inspection monitoring system for electronic manufacturing. Comput. Ind. 2005, 56, 986–996. [Google Scholar] [CrossRef]
  7. Nair, A.; Chen, D.; Agrawal, P.; Isola, P.; Abbeel, P.; Malik, J.; Levine, S. Combining self-supervised learning and imitation for vision-based rope manipulation. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; pp. 2146–2153. [Google Scholar]
  8. Pedersen, M.R.; Nalpantidis, L.; Andersen, R.S.; Schou, C.; Bøgh, S.; Krüger, V.; Madsen, O. Robot skills for manufacturing: From concept to industrial deployment. Robot. Comput. Integr. Manuf. 2016, 37, 282–291. [Google Scholar] [CrossRef]
  9. Zakhama, A.; Charrabi, L.; Jelassi, K. Intelligent Selective Compliance Articulated Robot Arm robot with object recognition in a multi-agent manufacturing system. Int. J. Adv. Robot. Syst. 2019, 16, 1–15. [Google Scholar] [CrossRef] [Green Version]
  10. Wang, T.; Yao, Y.; Chen, Y.; Zhang, M.; Tao, F.; Snoussi, H. Auto-sorting system toward smart factory based on deep learning for image segmentation. IEEE Sens. J. 2018, 18, 8493–8501. [Google Scholar]
  11. Li, L.; Ota, K.; Dong, M. Deep learning for smart industry: Efficient manufacture inspection system with fog computing. IEEE Sens. J. 2018, 14, 4665–4673. [Google Scholar] [CrossRef] [Green Version]
  12. Pech, M.; Vrchota, J.; Bednář, J. Predictive Maintenance and Intelligent Sensors in Smart Factory. Sensors 2021, 21, 1470. [Google Scholar] [CrossRef]
  13. Lee, K.B.; Cheon, S.; Kim, C.O. A convolutional neural network for fault classification and diagnosis in semiconductor manufacturing processes. IEEE Trans. Semicond. Manuf. 2017, 30, 135–142. [Google Scholar] [CrossRef]
  14. Kwon, O.; Kim, H.G.; Ham, M.J.; Kim, W.; Kim, G.; Cho, J.; Kim, N.I.; Kim, K. A deep neural network for classification of melt-pool images in metal additive manufacturing. J. Intell. Manuf. 2020, 31, 375–386. [Google Scholar] [CrossRef]
  15. Madry, A.; Makelov, A.; Schmidt, L.; Tsipras, D.; Vladu, A. Towards deep learning models resistant to adversarial attacks. arXiv 2017, arXiv:1706.06083. [Google Scholar]
  16. Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; Fergus, R. Intriguing properties of neural networks. arXiv 2013, arXiv:1312.6199. [Google Scholar]
  17. Acción, Á.; Argüello, F.; Heras, D.B. Dual-Window Superpixel Data Augmentation for Hyperspectral Image Classification. Appl. Sci. 2020, 10, 8833. [Google Scholar] [CrossRef]
  18. Baldominos, A.; Saez, Y.; Isasi, P. A survey of handwritten character recognition with mnist and emnist. Appl. Sci. 2019, 9, 3169. [Google Scholar] [CrossRef] [Green Version]
  19. Shorten, C.; Khoshgoftaar, T.M. A survey on image data augmentation for deep learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  20. Qi, H.; Liang, Y.; Ding, Q.; Zou, J. Automatic Identification of Peanut-Leaf Diseases Based on Stack Ensemble. Appl. Sci. 2021, 11, 1950. [Google Scholar] [CrossRef]
  21. Urbonas, A.; Raudonis, V.; Maskeliūnas, R.; Damaševičius, R. Automated identification of wood veneer surface defects using faster region-based convolutional neural network with data augmentation and transfer learning. Appl. Sci. 2019, 9, 4898. [Google Scholar] [CrossRef] [Green Version]
  22. Hussain, Z.; Gimenez, F.; Yi, D.; Rubin, D. Differential data augmentation techniques for medical imaging classification tasks. In Proceedings of the AMIA Annual Symposium Proceedings, Washington, DC, USA, 6–8 November 2017; pp. 979–984. [Google Scholar]
  23. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  24. Hore, A.; Ziou, D. Image quality metrics: PSNR vs. SSIM. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 2366–2369. [Google Scholar]
  25. Huynh-Thu, Q.; Ghanbari, M. Scope of validity of PSNR in image/video quality assessment. Electron. Lett. 2008, 44, 800–801. [Google Scholar] [CrossRef]
  26. Kim, E.K.; Lee, H.; Kim, J.Y.; Kim, S. Data Augmentation Method by Applying Color Perturbation of Inverse PSNR and Geometric Transformations for Object Recognition Based on Deep Learning. Appl. Sci. 2020, 10, 3755. [Google Scholar] [CrossRef]
  27. Hussain, M.; Bird, J.J.; Faria, D.R. A study on cnn transfer learning for image classification. In Proceedings of the UK Workshop on Computational Intelligence, Nottingham, UK, 5–7 September 2018; pp. 191–202. [Google Scholar]
Figure 1. Comparison of classification results for original image and noisy image.
Figure 1. Comparison of classification results for original image and noisy image.
Applsci 11 05586 g001
Figure 2. Overview of the proposed method.
Figure 2. Overview of the proposed method.
Applsci 11 05586 g002
Figure 3. Process of perturbation range calculation.
Figure 3. Process of perturbation range calculation.
Applsci 11 05586 g003
Figure 4. Red, green, and blue (RGB) channel histograms related to images in each category.
Figure 4. Red, green, and blue (RGB) channel histograms related to images in each category.
Applsci 11 05586 g004
Figure 5. Partial training images and histogram of images in its category.
Figure 5. Partial training images and histogram of images in its category.
Applsci 11 05586 g005
Figure 6. Partial training images and histogram of images in its category.
Figure 6. Partial training images and histogram of images in its category.
Applsci 11 05586 g006
Figure 7. Process of color perturbation for components of color space.
Figure 7. Process of color perturbation for components of color space.
Applsci 11 05586 g007
Figure 8. Comparison of classification accuracy.
Figure 8. Comparison of classification accuracy.
Applsci 11 05586 g008
Figure 9. Comparison of confusion matrix #1.
Figure 9. Comparison of confusion matrix #1.
Applsci 11 05586 g009
Figure 10. Comparison of confusion matrix #2.
Figure 10. Comparison of confusion matrix #2.
Applsci 11 05586 g010
Table 1. Classification accuracy of the network trained using original images.
Table 1. Classification accuracy of the network trained using original images.
Number of
Experiments
MethodNumber of
Training Images
Classification Accuracy
NoiselessNoisyTotal
1Training by
original images
32048.0%25.0%36.5%
243.0%29.0%36.0%
344.0%25.0%34.5%
Table 2. Classification accuracy of the network trained using the conventional method.
Table 2. Classification accuracy of the network trained using the conventional method.
Number of
Experiments
MethodNumber of
Training Images
Classification Accuracy
NoiselessNoisyTotal
1Training using images
obtained through
color jittering #1
288041.0%35.0%38.0%
247.0%37.0%42.0%
341.0%32.0%36.5%
1Training using images
obtained through
color jittering #2
288041.0%37.0%39.0%
243.0%39.0%41.0%
342.0%39.0%40.5%
1Training using images
obtained through
color jittering #3
288029.0%36.0%32.5%
231.0%32.0%31.5%
329.0%32.0%30.5%
Table 3. Classification accuracy of the network trained using preprocessed images.
Table 3. Classification accuracy of the network trained using preprocessed images.
Number of
Experiments
MethodNumber of
Training Images
Classification Accuracy
NoiselessNoisyTotal
1Training using
preprocessed
Images
32063.0%31.0%47.0%
252.0%26.0%39.0%
360.0%44.0%52.0%
Table 4. Classification accuracy of the network trained using images obtained through inverse PSNR with RGB color space and rotation.
Table 4. Classification accuracy of the network trained using images obtained through inverse PSNR with RGB color space and rotation.
Number of
Experiments
MethodNumber of
Training Images
Classification Accuracy
NoiselessNoisyTotal
1Training using
inverse PSNR + ROT
images
69,12090.0%60.0%75.0%
293.0%62.0%77.5%
390.0%66.0%78.0%
Table 5. Classification accuracy of the network trained using images obtained by applying the inverse PSNR to three channels individually.
Table 5. Classification accuracy of the network trained using images obtained by applying the inverse PSNR to three channels individually.
Number of
Experiments
MethodNumber of
Training Images
Classification Accuracy
NoiselessNoisyTotalAverage
1Training using
R PSNR images
192068.0%30.0%49.0%47.5%
257.0%34.0%45.5%
356.0%40.0%48.0%
4Training using
G PSNR images
192055.0%26.0%40.5%46.8%
560.0%42.0%51.0%
661.0%37.0%49.0%
7Training using
B PSNR images
192066.0%46.0%56.0%51.2%
858.0%38.0%48.0%
967.0%32.0%49.5%
Table 6. Classification accuracy of the network trained using images obtained by the adaptive inverse PSNR and RGB color space and rotation.
Table 6. Classification accuracy of the network trained using images obtained by the adaptive inverse PSNR and RGB color space and rotation.
Number of
Experiments
MethodNumber of
Training Images
Classification Accuracy
NoiselessNoisyTotal
1Training using
adaptive inverse
PSNR + ROT images
69,12092.0%76.0%84.0%
291.0%73.0%82.0%
393.0%72.0%82.5%
Table 7. Precision, recall, specificity, and F1-score of the conventional method and the proposed method.
Table 7. Precision, recall, specificity, and F1-score of the conventional method and the proposed method.
NoiselessNoisy
PrecisionRecallSpecificityF1-ScorePrecisionRecallSpecificityF1-Score
Conventional Method0.750.300.990.431.000.201.000.33
0.780.700.980.740.241.000.660.39
1.000.101.000.181.000.101.000.18
0.470.800.900.590.500.200.980.29
0.301.000.740.470.470.800.900.59
1.000.501.000.670.310.800.800.44
-0.001.00--0.001.00-
0.350.900.810.500.800.400.990.53
1.000.101.000.180.330.100.980.15
0.750.300.990.431.000.101.000.18
Proposed Method1.001.001.001.001.001.001.001.00
1.001.001.001.001.000.701.000.82
1.001.001.001.001.000.901.000.95
0.671.000.940.800.361.000.800.53
1.000.701.000.821.000.101.000.18
0.880.700.990.781.000.501.000.67
0.900.900.990.900.670.800.960.73
1.000.901.000.951.000.701.000.82
1.001.001.001.001.001.001.001.00
0.911.000.990.950.820.900.980.86
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kim, E.; Kim, J.; Lee, H.; Kim, S. Adaptive Data Augmentation to Achieve Noise Robustness and Overcome Data Deficiency for Deep Learning. Appl. Sci. 2021, 11, 5586. https://doi.org/10.3390/app11125586

AMA Style

Kim E, Kim J, Lee H, Kim S. Adaptive Data Augmentation to Achieve Noise Robustness and Overcome Data Deficiency for Deep Learning. Applied Sciences. 2021; 11(12):5586. https://doi.org/10.3390/app11125586

Chicago/Turabian Style

Kim, Eunkyeong, Jinyong Kim, Hansoo Lee, and Sungshin Kim. 2021. "Adaptive Data Augmentation to Achieve Noise Robustness and Overcome Data Deficiency for Deep Learning" Applied Sciences 11, no. 12: 5586. https://doi.org/10.3390/app11125586

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop