Next Article in Journal
Detection of Oil Spill in SAR Image Using an Improved DeepLabV3+
Previous Article in Journal
Green Wearable Sensors and Antennas for Bio-Medicine, Green Internet of Things, Energy Harvesting, and Communication Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Study on Data Preprocessing for Machine Learning Based on Semiconductor Manufacturing Processes

by
Ha-Je Park
1,
Yun-Su Koo
2,
Hee-Yeong Yang
1,
Young-Shin Han
3,* and
Choon-Sung Nam
1,*
1
Department of Software Convergence Engineering, Inha University, 100 Inha-ro, Michuhol-gu, Incheon 22212, Republic of Korea
2
Department of Mechatronics Engineering, Inha University, 100 Inha-ro, Michuhol-gu, Incheon 22212, Republic of Korea
3
Frontier College, Inha University, 100 Inha-ro, Michuhol-gu, Incheon 22212, Republic of Korea
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(17), 5461; https://doi.org/10.3390/s24175461 (registering DOI)
Submission received: 24 June 2024 / Revised: 14 August 2024 / Accepted: 21 August 2024 / Published: 23 August 2024
(This article belongs to the Section Industrial Sensors)

Abstract

:
Various data types generated in the semiconductor manufacturing process can be used to increase product yield and reduce manufacturing costs. On the other hand, the data generated during the process are collected from various sensors, resulting in diverse units and an imbalanced dataset with a bias towards the majority class. This study evaluated analysis and preprocessing methods for predicting good and defective products using machine learning to increase yield and reduce costs in semiconductor manufacturing processes. The SECOM dataset is used to achieve this, and preprocessing steps are performed, such as missing value handling, dimensionality reduction, resampling to address class imbalances, and scaling. Finally, six machine learning models were evaluated and compared using the geometric mean (GM) and other metrics to assess the combinations of preprocessing methods on imbalanced data. Unlike previous studies, this research proposes methods to reduce the number of features used in machine learning to shorten the training and prediction times. Furthermore, this study prevents data leakage during preprocessing by separating the training and test datasets before analysis and preprocessing. The results showed that applying oversampling methods, excluding KM SMOTE, achieves a more balanced class classification. The combination of SVM, ADASYN, and MaxAbs scaling showed the best performance with an accuracy and GM of 85.14% and 72.95%, respectively, outperforming all other combinations.

1. Introduction

The semiconductor industry is a high-value-added sector in modern society. Semiconductors are essential components for products ranging from small electronic devices, such as PCs, smartphones, and tablets to large electronic devices like cars [1,2].
The semiconductor manufacturing process involves producing silicon wafers, fabricating integrated circuits, packaging these integrated circuits, assembling the final products, and conducting tests [3]. In such a process, predicting good and defective products during manufacturing is essential to enhance productivity and reduce costs [3,4].
SECOM [5], provided by UCI (University of California, Irvine, CA, USA), consists of real data generated from semiconductor manufacturing processes and is used for research in classifying good and defective products. The dataset contains 41,951 missing values, accounting for approximately 4.54% of the total data. Excluding time and labels, it comprises 590 features in total. The dataset is relatively small, with only 1568 instances, including 1464 instances of the good product class and 104 instances of the defective product class. This results in a highly imbalanced dataset, with a class ratio of approximately 1:0.07 between good and defective products.
Classifying good and defective products using semiconductor manufacturing process data, such as SECOM, require removing and estimating missing values and reducing dimensionality [6]. In addition, the class ratio of training data should be balanced by resampling before machine learning [7]. Without such preprocessing steps, machine learning results in low data classification performance [8,9]. Finally, when evaluating the model performance, it is essential to use evaluation metrics such as the geometric mean (GM) of sensitivity and specificity in addition to the traditional accuracy metric. Accuracy means the results are biased towards the majority class [10,11].
The process described above, if carried out without split training and testing data before preprocessing and data analysis such as handling missing values and dimension reduction, would result in the leakage of information from the testing data into the training data, leading to compromised data integrity and overfitting of the test data [12].
In recent years, previous research has utilized SECOM for classifying good and defective products in semiconductor manufacturing processes. Three dimensionality reduction methods, including Boruta, Multivariate Adaptive Regression Splines, and principal component analysis (PCA), were compared and analyzed in combination with the undersampling and oversampling techniques in [4]. The performance of Correlation-Based Feature Selection, PCA, and cases without dimensionality reduction was compared in [13]. After reducing dimensionality using Recursive Feature Elimination, various models were compared and analyzed in [14]. These studies primarily focus on dimensionality reduction and resampling techniques using SECOM. However, the methods and timing of separating training and test data differ across studies, and the lack of confusion matrices makes it difficult to verify the possibility of biased results toward certain classes. They did not use or could not be calculated using the geometric mean. Therefore, it is difficult to directly compare and evaluate previous studies. Specifically, in [13], oversampling was applied to the entire dataset, leading to information from the training data leaking into the test data and resulting in the inclusion of non-existent data in the test set, which poses a significant issue. All three studies [4,13,14] used a Normalizer for scaling and did not investigate whether the application of different scaling methods could potentially improve model performance. We separate the training and test data, apply resampling only to the training data, and use the geometric mean to evaluate the classification performance of each class in a balanced manner. Finally, we investigated various scaling methods to identify the most suitable scaling technique for each model.
Thus, this study classified good and defective products during the semiconductor manufacturing process utilizing semiconductor manufacturing process data and machine learning. SECOM is utilized for this purpose. Furthermore, SECOM was split into training and testing data, and research on dataset integrity was conducted. This research examined methods for handling missing values and dimension reduction techniques to decrease the features used in machine learning for semiconductor manufacturing process data, such as SECOM. In addition, this study examined resampling methods and performance evaluation techniques for imbalanced data.

2. Related Research

2.1. Missing Value

Missing values occur in the real world for various reasons. For example, survey respondents may fail to answer specific questions, and participants in clinical trials may drop out due to drug allergies or death. Moreover, manual data entry can result in input errors [7]. Semiconductor manufacturing process data can also contain missing values for various reasons, such as machine errors, sensor malfunctions, data transmission errors, and mistakes by facility managers.
Data analysis results can be skewed if missing values are not handled, and certain patterns of missing data can introduce bias during machine learning, reducing the generalization performance of a model [8]. This issue can be addressed by removing samples or features with missing values or replacing missing values with specific values [7,8]. On the other hand, removing missing values can lead to the loss of important information, potentially affecting the learning process. Moreover, deleting samples might be more beneficial for the learning process than replacing them with specific values when the proportion of missing values is high. Nevertheless, imputing specific values in place of missing values, such as the mean, is used when the proportion of missing values for each feature is low. This approach allows the model to utilize the characteristics of the remaining data for learning.

2.2. Dimensionality Reduction

SECOM is a high-dimensional dataset with 590 features. In high-dimensional data, the number of features used in machine learning increases, leading to the curse of dimensionality. The curse of dimensionality refers to the phenomenon where, as the number of dimensions increases, the distance between data points increases, reducing density and increasing noise, decreasing the generalization performance of the model and causing overfitting. In addition, the computational burden on the model increases, leading to longer training times. Therefore, reducing the dimensionality of high-dimensional data to a lower-dimensional space is necessary to address the curse of dimensionality by reducing the number of features [15].
Methods for dimensionality reduction include feature selection and feature extraction [15]. Feature selection involves choosing features from existing ones based on an analysis of the correlation between independent and dependent variables. This method has lower computational complexity and better generalization characteristics but ignores the dependencies between features [16]. Principal component analysis (PCA) is a representative feature extraction method that generates new features by extracting the principal components to maintain maximum variance in the data. PCA preserves as much information as possible from the original features before dimensionality reduction, but it assumes that the data are embedded in a globally or approximately linear low-dimensional space [15].

2.3. Handling Data Imbalance

Imbalanced data refers to datasets where the class distribution is not uniform. SECOM is also an imbalanced dataset, with a higher proportion of good product classes than defective ones. Imbalanced datasets cause a lack of information in the minority class, leading to a bias towards the majority class, and poor classification performance in machine learning [6,9]. Therefore, resampling the data to balance the class proportions is necessary to mitigate this.
Resampling methods include undersampling, which reduces the majority class samples to match the minority class samples, and oversampling, which increases the minority class samples to match the majority class samples [10,17]. Both methods can yield better results than the original dataset. On the other hand, undersampling may result in a loss of essential sample data, while oversampling increases the dataset size, leading to higher computational costs [10].

2.4. Scaling

SECOM does not disclose which equipment was used for data collection. Similarly, semiconductor manufacturing process data are collected using various types of equipment, resulting in different units and ranges for the collected data. In machine learning, the results can be biased towards data with larger units or wider ranges if the maximum, minimum, or distribution of the data varies, degrading the performance [18].
Preventing such bias requires scaling the data to match the ranges. Scaling methods include normalization and standardization [19]. Standardization transforms the data such that the mean is 0 and there is a standard deviation of 1. Normalization adjusts the data range to be within 0 and 1, or −1 and 1, or other specified ranges.

2.5. Model Classification Performance Metrics

Model classification performance evaluation metrics include various methods such as accuracy and precision [10]. These metrics are calculated based on the confusion matrix and indicate the classification performance of the model. In binary classification problems, the confusion matrix consists of true negative (TN), false positive (FP), false negative (FN), and true positive (TP) as listed in Table 1. Each element provides information about the predictions and outcomes of the model. A TN indicates that the model predicted a negative result, and the actual result was also negative; hence, the result is true. An FP indicates that the model predicted positive, but the actual result was negative; hence, the result is false. An FN indicates that the model predicted negative, but the actual result was positive; hence, the result is false. A TP indicates that the model predicted positive, and the actual result was positive; hence, the result is true.
Accuracy represents the proportion of true outcomes in the overall predictions of the model, reflecting the classification performance across all classes, as expressed in Equation (1) [10,11]. In a dataset with imbalanced classes, if all predictions for the negative class are true and for the positive class there are five true and five false predictions, the accuracy still appears as 99%, indicating a bias towards the majority class, as shown in the confusion matrix in Table 2. In this scenario, the accuracy for the positive class was only 50%, but the overall accuracy was 99%. Therefore, for datasets with uneven class distributions, metrics, such as sensitivity and specificity, should be used alongside accuracy to evaluate the classification performance of the model [10].
A c c u r a c y = ( T N + T P ) ( T N + F P + F N + T P )
Sensitivity (also known as recall) is the proportion of true outcomes for the positive class predictions (Equation (2)), and specificity is the proportion of true outcomes for the negative class predictions, as shown in Equation (3) [10]. The sensitivity and specificity reflect the classification performance of the model for a single class. On the other hand, using the results for only one class does not provide a complete picture of the performance of a model across all classes. Therefore, the GM of the sensitivity and specificity should be used to address this issue [10,11].
S e n s i t i v i t y = T P ( T P + F N )
S p e c i f i c i t y = T N ( T N + F P )
Equation (4) expresses the GM, which is determined by using sensitivity and specificity to include the prediction results for two classes and then taking the square root to calculate the geometric mean of the ratio between sensitivity and specificity. For example, when sensitivity and specificity are 50% and 100%, respectively, as shown in Table 2, the GM is 70.7%, reflecting the balanced performance of the model with imbalanced data. Therefore, when evaluating the classification performance of models using imbalanced datasets, such as semiconductor manufacturing process data, it is important to use metrics, such as GM, in addition to accuracy to assess classification performance of the model.
G e o m e t r i c   M e a n = S e n s i t i v i t y × S p e c i f i c i t y

3. Research Method

This paper examined the optimal combination of preprocessing methods and machine learning models using SECOM to predict good and defective products from semiconductor manufacturing process data. In addition, it conducted a comparative evaluation against the project by Samsung SDS [20].
Figure 1 shows the research processes of Samsung and this paper. As shown in Figure 1a, Samsung performed preprocessing, such as replacing missing values and dimensionality reduction, without separating the training and test data. However, unlike Samsung and previous studies, this paper assumed that if the separation of training and test data is not conducted as the first step, information from the training data could leak into the test data during data analysis and preprocessing, leading to overfitting of the model. Thus, this paper separated the data into training and test data at a ratio of 7 to 3 to ensure data integrity, as shown in Figure 1b and detailed in Table 3. Missing values were removed or replaced to clean the data, and the number of features was reduced for dimensionality reduction. The model was evaluated after training by resampling the training data to uniformly adjust the class ratio before applying machine learning. This study used the following six models: Logistic Regression (LR), Support Vector Machine (SVM), K Nearest Neighbor (KNN), Decision Tree (DT), Random Forest (RF), and Extreme Gradient Boosting (XGB). These models demonstrated improved performance through preprocessing of imbalanced datasets, such as those from semiconductor manufacturing processes [21,22,23].

3.1. Data Cleaning and Dimensionality Reduction

Before processing the raw data, SECOM was divided into training and test data. Missing values were deleted or replaced based on the training data. Figure 2 presents the distribution of the number of missing values according to the feature in the training data. After deleting 31 features with more than 300 missing values per feature, the remaining missing values were replaced with feature-specific averages. The test data also had the same features deleted as in the training data, and missing values were replaced with the feature-specific averages from the training data. This study assumed that the independent and dependent variables have a linear relationship because SECOM does not provide the type, form, and unit of collected data. Therefore, missing values were replaced with feature-specific averages to reduce their effect on machine learning.
The dataset remained high-dimensional with 558 features, even after data cleaning. Feature selection methods using Pearson correlation coefficients were applied to reduce the number of features. A correlation coefficient above 80% indicates a very strong correlation, above 60% indicates a strong correlation, above 40% indicates a moderate correlation, and above 20% indicates a weak correlation [24]. In the case of SECOM, which contains 1568 total data instances, the number of instances per class is insufficient, resulting in very low correlation coefficients with the dependent variable. Since there were no features with a correlation coefficient above 20%, only those with a correlation coefficient above 10% were selected. The Pearson correlation coefficients between the independent and dependent variables were calculated based on the learning data. Absolute values were used to select sensors 21, 59, 64, 100, 122, 330, 407, 408, 411, 412, 413, and 486 with a correlation coefficient of 10% or more, reducing the dimension to 12 features and selecting the same features for the test data. Figure 3 shows the correlation coefficients by feature.
The machine learning model was trained after data cleaning and dimensionality reduction to compare the training data before and after resampling. The average accuracy of the five models, excluding DT, was 93.2%, with sensitivity biased towards 0% for good products, as listed in Table 4. DT has a lower accuracy of 88.11% compared to other models, but it shows a GM of 30.1%, which is higher than other models. However, the sensitivity remains low at 9.68%, indicating poor performance in classifying defective products.

3.2. Resampling

The class imbalance problem in SECOM was addressed by applying effective oversampling methods, as demonstrated by the approach in [10]. The methods utilized included the Synthetic Minority Oversampling Technique (SMOTE), Adaptive Synthetic Sampling (ADASYN), Borderline SMOTE (BL SMOTE), K-Means SMOTE (KM SMOTE), and Support Vector Machine SMOTE (SVM SMOTE), all of which demonstrated improved performance after resampling [11,17,21]. Table 5 lists the ratio of good and defective products in the training data after applying each oversampling method.
Table 6 presents the test results after machine learning after oversampling the training data with an oversampling method and model. The combination of SVM, SMOTE, ADASYN, and BL SMOTE achieved the highest sensitivity at 70.97% based on sensitivity, as shown in Table 6. On the other hand, the average accuracy was 32.76%, indicating that the classification performance was biased towards defective products. The combination of LR and SMOTE classified good and bad products in a balanced manner, with a GM of 69.5%, the highest among all combinations. However, when using KM SMOTE with the RF and XGB models, the sensitivity does not increase even after oversampling. In the LR, KNN, and DT models, the sensitivity is significantly lower with KM SMOTE than with other oversampling methods: 19.35% for LR, 3.23% for KNN, and 12.9% for DT. However, in the case of SVM with SVM SMOTE, it has the lowest sensitivity at 6.45%. Therefore, since KM SMOTE results in biased outcomes toward good products compared to other oversampling methods, it is excluded from subsequent research. SVM SMOTE, on the other hand, is not excluded because it increased sensitivity in models other than SVM.

3.3. Scaling

Figure 4 shows the value distribution by feature after oversampling. The range of the data is larger than that of other features in the case of sensor 21. Scaling was performed to match the range of the data using the Standard, MinMax, MaxAbs, Robust, and Normalizer [18].
Figure 5 shows the distribution of accuracy and the GM before and after scaling for each model after oversampling. Figure 5a,c show the distribution of accuracy after scaling for LR and SVM, and Figure 5b,d show the distribution of the GM after scaling of LR and SVM. The accuracy distribution increased compared to other scaling methods when LR and SVM used a Normalizer. On the other hand, in the case of the GM, the distribution decreased compared to that before the application of the Normalizer. For LR, the accuracy distribution after scaling decreased except for when using the Normalizer, but the GM distribution increased. MaxAbs achieved the highest GM and had a high median value. After scaling, SVM showed an increased distribution in the accuracy and GM, with MaxAbs being stable and achieving the highest GM because of the small distribution of accuracy and the GM. In the case of KNN, the accuracy distribution increased for scaling methods other than the Normalizer, as shown in Figure 5e. Figure 5f shows the distribution of the GM after scaling. The distribution of the Standard method is lower than before scaling, while the Normalizer has the highest median value and is stable because of its narrow distribution, although the highest value is achieved with Robust scaling. Tree-based models, such as DT, RF, and XGB, show a slight decrease in accuracy with the Normalizer, as shown in Figure 5g, but other scaling methods result in a slight change compared to before scaling. The Normalizer increased the distribution and maximum value of the GM as shown in Figure 5h. Thus, the scaling results vary from model to model; a single scaling method cannot be applied universally to all models. Therefore, this study selected a combination of the model, oversampling method, and scaling method by applying all scaling methods.

3.4. Comparative Evaluation

The results of this research were compared with the project process and results of Samsung. The biggest difference is that in the case of Samsung, missing values processing and dimensionality reduction were performed before the training data and test data were separated. In addition, Samsung selected features with a correlation coefficient of 6.9% or more and reduced them to 40 features. In this study, missing value processing and dimensionality reduction were carried out after separating the training data and the test data. In addition, the dimension was reduced to 12 features with a correlation coefficient of 10% or more.
Table 7 and Figure 6 list the results of Samsung and the optimal combination results of this paper by model. In the case of Samsung, the combination of XGB, SMOTE, and Normalize showed 85.99% accuracy and 76.9% GM. The best combination in this study was SVM, ADASYN, and MaxAbs, with an accuracy of 85.14%, which was 0.85% lower than that of Samsung, and a GM of 3.95%. On the other hand, the training time and prediction time of the model were reduced via the machine learning process using 12 features, 28 fewer features than Samsung. In addition, the process of this study suggests that the model was less likely to overfit the test data, and the generalization performance for the new test data was higher than Samsung’s process because the training data and the test data were separated and then missing value processing and dimensionality reduction were performed.

4. Conclusions

The semiconductor industry is one of the most highly valued industries for the future in modern society, making it crucial to accurately classify good and defective products in the semiconductor manufacturing process. The semiconductor manufacturing process is highly complex, continuously generating hundreds of types of measurement data to improve yield. Using this vast amount of measured semiconductor manufacturing process data is essential to automated quality classification because it can significantly enhance the efficiency of the manufacturing process.
The purpose of this study is to improve the efficiency of defect detection by predicting defective products in advance using data generated from the semiconductor manufacturing process. To achieve this, this paper investigates methods for predicting defective products during the semiconductor manufacturing process using machine learning. This paper analyzed semiconductor manufacturing process data based on SECOM, focusing on handling missing values, dimensionality reduction, and resampling preprocessing techniques, with the aim of improving the efficiency of the semiconductor manufacturing process.
We compared and analyzed the results of 150 different combinations by utilizing six models, five oversampling methods, and five scaling methods for this study. As a result, the model that utilized SVM with ADASYN for oversampling and MaxAbs for scaling demonstrated the highest performance, with an accuracy of 85.14% and GM of 72.95%. This result highlights the importance of the combination of resampling methods, scaling techniques, and models in dealing with imbalanced data.
In the case of Samsung, the XGB model, when scaled using Normalizer and oversampled with SMOTE, achieved an accuracy of 85.99% and GM of 76.9%, which is slightly higher than the results of this study. However, since the process of Samsung involved handling missing values and performing dimensionality reduction before separating the training and test data, this may have influenced the learning process and the resulting outcomes. In contrast, we separated the training and test data before handling missing values and performing dimensionality reduction, ensuring that the test data remained unaffected. The proposed method allows for a more objective evaluation and demonstrates that the classification performance of the model, as measured by the GM, was not biased toward either the good or defective class. Our proposed method achieved similar accuracy using only 12 features, which is less than Samsung’s 40 features. In other words, it was confirmed that this method is capable of efficiently predicting defective products with minimal features. From the results of analyzing various combinations of scaling methods and models, we confirmed that one scaling method does not consistently improve the performance of all models. We analyzed the optimal scaling methods that contribute to performance improvement for each model, and proposed scaling techniques that can maximize performance specific to each model.
The machine learning models trained on our preprocessed data did not achieve an accuracy of over 90%. Thus, it is expected that future research could attain higher accuracy using KNN for missing data imputation, machine learning-based dimensionality reduction techniques, deep learning methods, and hyperparameter optimization. Finally, it is necessary to verify the generalizability of this research method using other semiconductor manufacturing process data.

Author Contributions

Conceptualization, C.-S.N.; methodology, C.-S.N. and H.-J.P.; software, H.-J.P.; validation, C.-S.N., H.-Y.Y., Y.-S.K. and H.-J.P.; formal analysis, C.-S.N., H.-Y.Y., Y.-S.K. and H.-J.P.; investigation, C.-S.N., H.-Y.Y., Y.-S.K. and H.-J.P.; resources, C.-S.N., H.-Y.Y., Y.-S.K. and H.-J.P.; data curation, C.-S.N. and H.-J.P.; writing—original draft preparation C.-S.N., Y.-S.K. and H.-J.P.; writing—review and editing, C.-S.N.; visualization, C.-S.N. and H.-J.P.; supervision, C.-S.N. and Y.-S.H.; project administration, C.-S.N. and Y.-S.H.; funding acquisition, C.-S.N. and Y.-S.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Innovative Human Resource Development for Local Intellectualization program through the Institute of Information & Communications Technology Planning & Evaluation (IITP) grant funded by the Korea government (MSIT) (IITP-2024-RS-2023-00259678).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kamal, K.Y. The Silicon Age: Trends in Semiconductor Devices Industry. J. Eng. Sci. Technol. Rev. 2022, 15, 1791–2377. [Google Scholar] [CrossRef]
  2. Ahmad, A. Automotive Semiconductor Industry-Trends, Safety and Security Challenges. In Proceedings of the 2020 8th International Conference on Reliability, Infocom Technologies and Optimization (Trends and Future Directions) (ICRITO), Noida, India, 4–5 June 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1373–1377. [Google Scholar] [CrossRef]
  3. Munirathinam, S.; Ramadoss, B. Predictive Models for Equipment Fault Detection in the Semiconductor Manufacturing Process. IACSIT Int. J. Eng. Technol. 2016, 8, 273–285. [Google Scholar] [CrossRef]
  4. Moldovan, D.; Cioara, T.; Anghel, I.; Salomie, I. Machine Learning for Sensor-Based Manufacturing Processes. In Proceedings of the 2017 13th IEEE International Conference on Intelligent Computer Communication and Processing (ICCP), Cluj-Napoca, Romania, 7–9 September 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 147–154. [Google Scholar] [CrossRef]
  5. McCann, M.; Johnston, A. SECOM. UCI Machine Learning Repository. 2008. Available online: https://archive.ics.uci.edu/dataset/179/secom (accessed on 17 June 2024). [CrossRef]
  6. Kumar, P.; Bhatnagar, R.; Gaur, K.; Bhatnagar, A. Classification of Imbalanced Data: Review of Methods and Applications. In Proceedings of the IOP Conference Series: Materials Science and Engineering, Online, 12–14 November 2021; IOP Publishing: Bristol, UK, 2021; Volume 1099, p. 012077. [Google Scholar] [CrossRef]
  7. Emmanuel, T.; Maupong, T.; Mpoeleng, D.; Semong, T.; Mphago, B.; Tabona, O. A Survey on Missing Data in Machine Learning. J. Big Data 2021, 8, 140. [Google Scholar] [CrossRef] [PubMed]
  8. Ren, L.; Wang, T.; Seklouli, A.S.; Zhang, H.; Bouras, A. A Review on Missing Values for Main Challenges and Methods. Inf. Syst. 2023, 119, 102268. [Google Scholar] [CrossRef]
  9. Hasib, K.M.; Iqbal, M.S.; Shah, F.M.; Mahmud, J.A.; Popel, M.H.; Showrov, M.I.H.; Ahmed, S.; Rahman, O. A Survey of Methods for Managing the Classification and Solution of Data Imbalance Problem. arXiv 2020, arXiv:2012.11870. [Google Scholar] [CrossRef]
  10. García, V.; Sánchez, J.S.; Mollineda, R.A. On the Effectiveness of Preprocessing Methods When Dealing with Different Levels of Class Imbalance. Knowl.-Based Syst. 2012, 25, 13–21. [Google Scholar] [CrossRef]
  11. Pratama, I.; Pristyanto, Y.; Prasetyaningrum, P.T. Imbalanced Class Handling and Classification on Educational Dataset. In Proceedings of the 2021 4th International Conference on Information and Communications Technology (ICOIACT), Yogyakarta, Indonesia, 30–31 August 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 180–185. [Google Scholar] [CrossRef]
  12. Bouke, M.A.; Abdullah, A. An Empirical Study of Pattern Leakage Impact During Data Preprocessing on Machine Learning-Based Intrusion Detection Models Reliability. Expert Syst. Appl. 2023, 230, 120715. [Google Scholar] [CrossRef]
  13. El Mourabit, Y.; El Habouz, Y.; Zougagh, H.; Wadiai, Y. Predictive system of semiconductor failures based on machine learning approach. Int. J. Adv. Comput. Sci. Appl. 2020, 11, 199–203. [Google Scholar] [CrossRef]
  14. Pradeep, D.; Vardhan, B.V.; Raiak, S.; Muniraj, I.; Elumalai, K.; Chinnadurai, S. Optimal Predictive Maintenance Technique for Manufacturing Semiconductors using Machine Learning. In Proceedings of the 2023 3rd International Conference on Intelligent Communication and Computational Techniques (ICCT), Jaipur, India, 19–20 January 2023; pp. 1–5. [Google Scholar] [CrossRef]
  15. Jia, W.; Sun, M.; Lian, J.; Hou, S. Feature Dimensionality Reduction: A Review. Complex Intell. Syst. 2022, 8, 2663–2693. [Google Scholar] [CrossRef]
  16. Zebari, R.; Abdulazeez, A.; Zeebaree, D.; Zebari, D.; Saeed, J. A Comprehensive Review of Dimensionality Reduction Techniques for Feature Selection and Feature Extraction. J. Appl. Sci. Technol. Trends 2020, 1, 56–70. [Google Scholar] [CrossRef]
  17. Rendon, E.; Alejo, R.; Castorena, C.; Isidro-Ortega, F.J.; Granda-Gutierrez, E.E. Data Sampling Methods to Deal with the Big Data Multi-Class Imbalance Problem. Appl. Sci. 2020, 10, 1276. [Google Scholar] [CrossRef]
  18. Ahsan, M.M.; Mahmud, M.P.; Saha, P.K.; Gupta, K.D.; Siddique, Z. Effect of Data Scaling Methods on Machine Learning Algorithms and Model Performance. Technologies 2021, 9, 52. [Google Scholar] [CrossRef]
  19. Sharma, V. A Study on Data Scaling Methods for Machine Learning. Int. J. Glob. Acad. Sci. Res. 2022, 1, 31–42. [Google Scholar] [CrossRef]
  20. Brightics, AI. Available online: https://www.brightics.ai/community/knowledge-sharing/detail/7059 (accessed on 17 June 2024).
  21. Al Helal, M.; Haydar, M.S.; Mostafa, S.A.M. Algorithms Efficiency Measurement on Imbalanced Data Using Geometric Mean and Cross Validation. In Proceedings of the 2016 International Workshop on Computational Intelligence (IWCI), Dhaka, Bangladesh, 12–13 December 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 110–114. [Google Scholar] [CrossRef]
  22. Khushi, M.; Shaukat, K.; Alam, T.M.; Hameed, I.A.; Uddin, S.; Luo, S.; Yang, X.; Reyes, M.C. A Comparative Performance Analysis of Data Resampling Methods on Imbalance Medical Data. IEEE Access 2021, 9, 109960–109975. [Google Scholar] [CrossRef]
  23. Tanha, J.; Abdi, Y.; Samadi, N.; Razzaghi, N.; Asadpour, M. Boosting Methods for Multi-Class Imbalanced Data Classification: An Experimental Review. J. Big Data 2020, 7, 70. [Google Scholar] [CrossRef]
  24. Liu, Y.; Mu, Y.; Chen, K.; Li, Y.; Guo, J. Daily activity feature selection in smart homes based on Pearson correlation coefficient. Neural Process. Lett. 2020, 51, 1771–1787. [Google Scholar] [CrossRef]
Figure 1. Research processes: (a) Samsung; (b) this paper.
Figure 1. Research processes: (a) Samsung; (b) this paper.
Sensors 24 05461 g001
Figure 2. Distribution of missing values by feature.
Figure 2. Distribution of missing values by feature.
Sensors 24 05461 g002
Figure 3. Feature correlations with dependent variable.
Figure 3. Feature correlations with dependent variable.
Sensors 24 05461 g003
Figure 4. Distribution of values by feature after oversampling.
Figure 4. Distribution of values by feature after oversampling.
Sensors 24 05461 g004
Figure 5. Accuracy and GM distribution before and after scaling by model: (a) accuracy distribution in LR; (b) GM distribution in LR; (c) accuracy distribution in SVM; (d) GM distribution in SVM; (e) accuracy distribution in KNN; (f) GM distribution in KNN; (g) accuracy distribution in Tree-based models; (h) GM distribution in Tree-based models.
Figure 5. Accuracy and GM distribution before and after scaling by model: (a) accuracy distribution in LR; (b) GM distribution in LR; (c) accuracy distribution in SVM; (d) GM distribution in SVM; (e) accuracy distribution in KNN; (f) GM distribution in KNN; (g) accuracy distribution in Tree-based models; (h) GM distribution in Tree-based models.
Sensors 24 05461 g005
Figure 6. Comparison of the results of the optimal combination of Samsung’s results and this paper.
Figure 6. Comparison of the results of the optimal combination of Samsung’s results and this paper.
Sensors 24 05461 g006
Table 1. Confusion matrix for binary classification.
Table 1. Confusion matrix for binary classification.
Predict
NegativePositive
ActualNegativeTNFP
PositiveFNTP
Table 2. Confusion matrix for imbalanced data.
Table 2. Confusion matrix for imbalanced data.
Predict
NegativePositive
ActualNegative10000
Positive55
Table 3. Class ratio of training data to test data.
Table 3. Class ratio of training data to test data.
Training DataTest Data
Good1023440
Defective7331
Ratio1:0.07
Table 4. Model-specific results after data cleaning and dimensionality reduction.
Table 4. Model-specific results after data cleaning and dimensionality reduction.
ModelAccuracyGMSensitivitySpecificity
LR0.9299000.9955
SVC0.9342001
KNN0.9342001
DT0.88110.30100.09680.9364
RF0.9342001
XGB0.9278000.9932
Table 5. Number and ratio of good and defective products after oversampling.
Table 5. Number and ratio of good and defective products after oversampling.
SMOTEADASYNBL SMOTEKM SMOTESVM SMOTE
Good10231023102310231023
Defective102399910231024595
Ratio1:11:0.981:11:11:0.58
Table 6. Application of the oversampling methods by model.
Table 6. Application of the oversampling methods by model.
ModelSamplingAccuracyGMSensitivitySpecificity
LRSMOTE0.81530.69500.58060.8318
ADASYN0.80890.69210.58060.8250
BL SMOTE0.82170.67910.54840.8409
KM SMOTE0.84290.41470.1935 *0.8886
SVM SMOTE0.85560.63190.45160.8841
SVMSMOTE0.31630.45260.70970.2886
ADASYN0.31420.45080.70970.2864
BL SMOTE0.35240.48190.70970.3273
KM SMOTE0.83860.37810.16130.8864
SVM SMOTE0.92570.25230.0645 *0.9864
KNNSMOTE0.72400.59880.48390.7409
ADASYN0.72400.55910.41940.7455
BL SMOTE0.77920.55890.38710.8068
KM SMOTE0.92990.17900.0323 *0.9932
SVM SMOTE0.81950.49880.29030.8568
DTSMOTE0.81740.52430.32260.8523
ADASYN0.83010.44390.22580.8727
BL SMOTE0.85560.53680.32260.8932
KM SMOTE0.88320.34760.1290 *0.9364
SVM SMOTE0.83860.44620.22580.8818
RFSMOTE0.89600.49280.25810.9409
ADASYN0.90230.42930.19350.9523
BL SMOTE0.90450.39280.16130.9568
KM SMOTE0.93240.00000.0000 *1.0000
SVM SMOTE0.91720.39560.16130.9705
XGBSMOTE0.88320.38810.16130.9341
ADASYN0.88320.45810.22580.9295
BL SMOTE0.88540.38860.16130.9364
KM SMOTE0.91720.00000.0000 *0.9818
SVM SMOTE0.89810.39140.16130.9500
* Worst sensitivity by sampling method.
Table 7. Results of the optimal combination of Samsung’s results and this paper by model.
Table 7. Results of the optimal combination of Samsung’s results and this paper by model.
ModelSamplingScalerAccuracyGM
SamsungXGBSMOTENormalize0.85990.7690
paper combination 1 LRSMOTEMaxAbs0.78340.7383
paper combination 2 * SVMADASYNMaxAbs0.85140.7295
paper combination 3 KNNSMOTERobust0.74310.6259
paper combination 4 DTBL SMOTENormalize0.84930.6294
paper combination 5RFSMOTEMaxAbs0.89810.5227
paper combination 6 XGBBL SMOTENormalize0.88960.5201
* Best combination in the paper.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Park, H.-J.; Koo, Y.-S.; Yang, H.-Y.; Han, Y.-S.; Nam, C.-S. Study on Data Preprocessing for Machine Learning Based on Semiconductor Manufacturing Processes. Sensors 2024, 24, 5461. https://doi.org/10.3390/s24175461

AMA Style

Park H-J, Koo Y-S, Yang H-Y, Han Y-S, Nam C-S. Study on Data Preprocessing for Machine Learning Based on Semiconductor Manufacturing Processes. Sensors. 2024; 24(17):5461. https://doi.org/10.3390/s24175461

Chicago/Turabian Style

Park, Ha-Je, Yun-Su Koo, Hee-Yeong Yang, Young-Shin Han, and Choon-Sung Nam. 2024. "Study on Data Preprocessing for Machine Learning Based on Semiconductor Manufacturing Processes" Sensors 24, no. 17: 5461. https://doi.org/10.3390/s24175461

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop