Next Article in Journal
Morphological Evolution and Damping Properties of Dynamically Vulcanized Butyl Rubber/Polypropylene Thermoplastic Elastomers
Next Article in Special Issue
A Multiscale Study of CFRP Based on Asymptotic Homogenization with Application to Mechanical Analysis of Composite Pressure Vessels
Previous Article in Journal
N-(2-Arylethyl)-2-methylprop-2-enamides as Versatile Reagents for Synthesis of Molecularly Imprinted Polymers
Previous Article in Special Issue
Preparation of the Carbonized Zif−8@PAN Nanofiber Membrane for Cadmium Ion Adsorption
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Quality Prediction and Abnormal Processing Parameter Identification in Polypropylene Fiber Melt Spinning Using Artificial Intelligence Machine Learning and Deep Learning Algorithms

Department of Materials Science and Engineering, National Taiwan University of Science and Technology, Taipei 10607, Taiwan
*
Author to whom correspondence should be addressed.
Polymers 2022, 14(13), 2739; https://doi.org/10.3390/polym14132739
Submission received: 8 June 2022 / Revised: 27 June 2022 / Accepted: 30 June 2022 / Published: 4 July 2022

Abstract

:
Melt spinning machines must be set up according to the process parameters that result in the best end product quality. In this study, artificial intelligence algorithms were employed to create a system that detects abnormal processing parameters and suggests strategies to improve quality. Polypropylene (PP) was selected as the experimental material, and the quality achieved by adjusting the melt spinning machine’s processing parameter settings was used as the basis for judgement. The processing parameters included screw temperature, gear pump temperature, die head temperature, screw speed, gear pump speed, and take-up speed as the six control factors. The four quality characteristics included fineness, breaking strength, elongation at break, and elastic energy modulus. In the first part of our study, we applied fast deep-learning characteristic grid calculations on a 440-item historical data set to train a deep learning neural network and determine methods for multi-quality optimization. In the second part, with the best processing parameters as a benchmark, and given abnormal quality data derived from processing parameter settings deviating from these optimal values, several machine learning and deep learning methods were compared in their ability to find the settings responsible for the abnormal data, which was randomly split into a 210-item training data set and a 210-item verification data set. The random forest method proved to be the best at identifying responsible parameter settings, with accuracy rates of single and double identification classifications together of 100%, for single factor classification of 98.3%, and for double factor classification of 96.0%, thereby confirming that the diagnostic method proposed in this study can effectively predict product abnormality and find the parameter settings responsible for product abnormality.

Graphical Abstract

1. Introduction

Modern synthetic fibers are mainly manufactured by three methods: dry spinning, wet spinning, or melt spinning. Among them, melt spinning is the most commonly used in the industry due to its low cost and the stability of the process [1,2]. However, due to the wear and tear of equipment and the resulting maintenance or replacement of different parts of the melt spinning process, product quality can be expected to deviate from the original. Inability to achieve the preset best quality and abnormality of the process may also be due to a number of other factors, such as the experience of the operator of the machine and parameter settings, among other issues. Because of the above factors, the causes of abnormality are also more difficult to analyze. Currently, the industry almost completely relies on the expertise of technologists to solve the problem.
If a specialized analysis technique can be developed to explore the link between abnormal quality and processing parameters, it will significantly enhance the maintenance of product quality, reduce abnormal analysis downtime, and stabilize process control. In order to accomplish these goals, we present a set of techniques in this work that engineers may use as criteria to determine whether product quality is declining and to quickly identify the causes of it, enhancing product quality and lowering manufacturing costs. Therefore, the goal of this study is to investigate the best melt spinning machine process parameter settings and to pinpoint those that lead to improper processing. Since PP fiber is currently one of the three major synthetic fibers that are mostly used in the production of knitwear and plush items and are produced by melt spinning, this study selects PP as the experimental material and samples from melt spinning machines as the research object. A deep learning neural network [3] is used to find the optimal processing parameters determinant of multiple quality characteristics and, in the case of quality changes, to indicate which processing parameter changes are responsible for the abnormal quality and thus solve the problem. We hope to provide the industry with a good and efficient diagnostic method for the analysis of quality abnormalities.
The selection of machining parameters in the industry is based on the experience and intuition of the engineer responsible for the machine. However, the trial-and-error process adopted is error-prone and time-consuming, so it is not optimal for complex manufacturing processes [3,4]. Chen et al. [5] used Taguchi parametric design methods combined with neural networks to optimize processing parameter settings for plastic injection molding. The experimental results showed that their neural network model not only effectively reduces the time taken for optimal processing parameter setting, but also results in more reliable product quality. The same approach is widely used in a number of different fields, including computing, science, engineering, medicine, the environment, agriculture, mining, weather prediction, business, and even art [6]. Majumdar and Majumdar [7] compared three modeling approaches for forecasting the breaking elongation of ring spun cotton yarns (mathematical, statistical, and artificial neural network models). The main reason why neural networks can successfully predict product quality is that they have the ability to learn any nonlinear mapping between input and output, and compared with traditional regression and statistical models, the results are more definitive [8]. The larger the amount of historical data, the more accurate the neural network can become [8,9].
Artificial intelligence has recently been integrated to the analytical tools of many researchers in order to shorten experimentation times and lower costs. Although abnormality diagnosis systems have been used in various industries, they are rarely met with in polymer melt spinning processing. This is surprising, since in melt spinning, as in other production processes, artificial intelligence classifiers can count different features and classify them according to different characteristics, using decision tree, support vector machine, random forest [10], or neural network rules, among others, to detect faults.
Sugumaran and Ramachandran [11] extracted and analyzed eigenvalues through vibration signals and used a decision tree to select 3 excellent features that could identify normal and abnormal eigenvalues from 11 eigenvalues. They then input them into fuzzy classifiers, showing that they can be applied to fault diagnosis. Decision trees are used not only for fault diagnosis, but also for other diagnostic purposes, like in medicine and prediction of mechanical strength [12,13]. Zimmerman et al. [14] used a CART decision tree for analysis and found that their CART-developed test, although less accurate than PCR, had good sensitivity and good predictive performance for influenza.
The random forest method has further improved the accuracy of machine learning decision trees and has been widely used in the field of artificial intelligence classification in recent years, having great advantages over other artificial intelligence algorithms. Nafees et al. [15] integrated the linear model of random forest to create a study which aims to create modelling tools for estimating the compressive and tensile strengths of plastic concrete. It is also widely used in industrial diagnosis. Cerrada et al. [16] applied random forests and genetic algorithms to do fault diagnosis and classification of gears. With a large number of features as input, the accuracy rate was as high as 97%. Sanchez et al. [17] used a deep random forest to diagnose failures of gearboxes and compared it with other machine learning algorithms such as support vector machines and the k-nearest neighbor algorithm, confirming that the random forest was the best for their data. Beyond industry, random forest classification has also achieved very good results in many other fields, such as medical analysis [18] and biological field [19]. In the field of artificial intelligence, effective classifiers can be developed not only with random forests, but also with the aforementioned neural networks. Artificial neural networks use nonlinear activation functions and have weighted outputs from multiple neurons into deeper neuron layers, with multiple layers connected in sequence to increase learning accuracy [20]. Ali et al. [21] studying the non-stationary nonlinear characteristics of rolling bearing vibration signals, used a neural network to classify bearing defects. Their experimental results showed that this method can reliably classify defects.
In this study, we proposed a set of methodologies that engineers may use as criteria to determine whether product quality is deteriorating and to immediately assist them in identifying causes, ultimately increasing product quality and lowering manufacturing costs. As a result, the goal of this research is to find the best process parameter settings for melt spinning machines and to identify those that cause anomalous processing.

2. Methods and Materials

The experimental method and artificial intelligence steps are shown in Figure 1, including melt spinning experiment, data pre-processing, artificial intelligence classifier, single and double anomaly identification, and classification results.

2.1. Random Forest

The random forest is composed of decision trees, and each decision tree in the random forest is not related [22]. After obtaining the random forest, when a new input sample enters, each decision tree in the random forest makes a judgment to predict which class the sample should belong to, and finally, each decision tree votes to predict which class the sample belongs to. Although each decision tree in the random forest obtained by this algorithm is very weak, the combination of each decision tree works very well. This method is also called ensemble learning [23]. Its calculation steps:
(1)
Define a random sample of size n, and randomly select n data from the data set.
(2)
From the selected n data, a decision tree is trained, d features are randomly extracted for each node in the decision tree, and then the features are used to divide the node.
(3)
Repeat steps 1~2 k times with improvements. The more commonly used improvement is Adaboost.
(4)
Summarize the predictions of all decision trees and decide the result of this classification by voting majority or weighted voting.
AdaBoost is an algorithm that improves boosting [24] and is the most commonly used one today. The idea is to increase the weight of the samples misclassified by the previous decision trees so that each time a new decision tree is trained it can focus on training data that is easily misclassified. Each decision tree uses weighted voting instead of the average voting mechanism. Weak classifiers with higher accuracy have larger weights, and weak classifiers with lower accuracy have lower weights.
First, a set of training data is given:
( { ( x 1 , y 1 ) , ( x 2 , y 2 ) , , ( x n , y n ) } )
Assuming that the weight of the kth time is w k i , the weight of each sample of the first decision tree classifier is:
w 1 i = 1 n
First train the first decision tree classifier f k ( x ) with weights w k i .
Assuming that L decision tree classifiers are trained, when training the kth one:
a k = 0.5 * In ( 1 ε k ε k )
w k + 1 i = { w k i * e a k     i f     f k ( x i ) y i w k i * e a k       i f     f k ( x i ) = y i
The ε k is the error of the kth decision tree classifier.
Finally, f L ( x ) decision tree classifiers are obtained, and the results of all decision tree classifiers are weighted to vote:
H ( x ) = s i g n ( K = 1 L a k f k ( x ) )

2.2. Neural Network

A neural network mimics the human brain, in a structure consisting of thousands of interconnected neurons. A neuron can be connected to multiple neurons in its rear layer for output, and multiple neurons in a front layer for reception. Neurons and perceptron’s are mathematical functions that multiplies input data from the input layer (×1, ×2…) by a weight (w1, w2…) and adds a bias (b) to the weighted inputs (hidden layer). The result is then put via an activation function (f) to introduce nonlinearity to the network. All incoming data points receive a weight, are multiplied and added together, and passed to a non-linear activation function. An example of a single-layer neural network-like architecture is shown in Figure 2.
The output of each neuron:
a = f ( w p + b )
After calculating the loss between the output layer of the neural network and the correct value, the neural network will modify the neural network parameters through the back-propagation algorithm. Since neural networks are inherently nonlinear, consisting of multiple inputs and multiple outputs, they are suitable for modeling complex nonlinear systems.
In order to find the smallest loss function, the training of neural network often uses the gradient descent algorithm to achieve optimization.
w i = w i 1 γ L W
where W is the weight parameter; γ is the learning rate; L is the loss function. L W is the gradient of the loss function to the weight parameter.

2.3. Activation Functions

Activation functions are functions that are used in neural networks to compute the weighted total of input and biases, which is then used to determine whether or not a neuron can activate [25]. They are used to control the outputs of our neural networks in a variety of areas, including object recognition and classification, as well as other domains, to name a few, with early research findings demonstrating unequivocally that good activation function selection improves neural network computing results. ReLU, Mish, and Sigmoid Functions are the activation functions employed in this work.

2.4. Optimization Techniques

Optimization is one of the important aspects of deep learning, as it helps a model train better when the weights are modified, so that it can reduce the loss error and also handle the dimensionality problem during back propagation. Ruder [26] investigated the convergence time, number of fluctuations, and parameter update rate of multiple stochastic gradient descent-based optimization algorithms, including SGD and SGD with momentum, Adam, and RMSProp, using varying numbers of iterations and particular values of the test function.

2.5. Materials

PP is an outstanding textile material which is purchased from polyacrylic polymer Globalene 6331 (LCY Chemical Corp., Taipei, Taiwan). The chemical structure is shown in Figure 3. It features abrasion resistance, flexibility, high strength resistance, light weight, strong antistatic character, good chemical resistance, and a low cost. Acid and alkali resistance, water repellency, quick drying, bacteria repellency, below thermal conductivity, warmth retention, low glass transition point, low temperature resistance, low energy consumption, low CO2 emission, decomposability, no dyeing wastewater pollution, and oil absorption are among its advantages over PET, Nylon 6, and Nylon 66 fibers. PP fiber is now widely utilized and has gained economic relevance in the production of home furnishings and industrial uses. The PP fiber is one of three primary synthetic fibers and is generally made via melt spinning. In the vertical or horizontal screw extruder, the acrylic resin is heated and molten, then extruded by nozzle through the metering pump and cooled to fiber in the air [27,28].

3. Experiment Plan

In this study, a melt spinning machine was used as the research machine, and PP was used as the material. The melt spinning machine uses heating to make the material appear in a molten state. The material was then conveyed through the screw and the gear pump, so that it is continuously extruded through the spinning nozzle, and wound into a drum through the roller. Since the melt spinning machine includes feeder, screw heating zone, gear pump, spinning nozzle, and take up system, as shown in Figure 4, the screw temperature, gear pump temperature, die head temperature, screw speed, gear pump speed, and take-up speed are designed as process parameters to discuss the quality of the fiber process. At the same time, a group of neural networks was trained using historical experimental data to predict the multiple characteristics of quality on the basis of various processing parameter values.
Half the quality characteristics obtained were used as input feature values for the training of a quality abnormality classifier. After training, the other half were input as test samples to confirm whether the classifier could identify processing parameters responsible for the quality abnormality. Assuming the identification was successful, the diagnosis system for the melt spinning machine was complete.

3.1. Materials Analysis

PP was selected as the experimental material in the study because of its characteristics, namely, easy processing, mechanical strength, strong elasticity, resistance to staining, lightness, and low price. Before the experiment, it was necessary to find its melting point and thermal cracking point to plan settings for the machine. To learn the temperature of the thermal cracking point, a thermogravimetric analyzer was used, with a thermal differential analyzer used to measure its melting point. A thermogravimetric analysis diagram and differential scanning calorimetry (DSC) diagram are shown in Figure 5 and Figure 6. It can be seen that the thermal cracking point of PP is about 400 °C, so this temperature was not exceeded during the experiment, as it risked contaminating the machine. The melting point is about 166 °C. Therefore, the temperature was kept above this level during the experiment. Temperatures lower than this temperature also risked damaging the machine.

3.2. Multi-Quality Characteristic Prediction

3.2.1. Experimental Data

A total of 440 historical melt spinning measurement records were used as data in the study. There were six processing parameters, namely, screw temperature, gear pump temperature, die head temperature, screw speed, gear pump speed, and take-up speed. The corresponding quality characteristics were fineness, breaking strength, elongation at break, and modulus of resilience. These 440 samples were randomly split into a 330-item training data set and a 110-item validation data set, using the k-fold cross-validation method to produce the best model for subsequent analysis when the model was finally optimized.

3.2.2. Data Processing

Due to the different units of measurement of all the input independent variables, the output dependent variables, and the different value size characteristics, comparability was impaired. To solve this problem, the data was first normalized for this experiment. The range of processing parameters of the original independent variables is shown in Table 1.

3.2.3. Neural Network Training

A neural network was used for the prediction of the multiple quality characteristics. The input variables were the processing parameters of the melt spinning machine: screw temperature, gear pump temperature, die head temperature, screw speed, gear pump speed, and take-up speed. The output variables were the corresponding quality characteristics: fineness, breaking strength, elongation at break, and modulus of resilience.
The architecture of the neural network is shown in Figure 7. The number of hidden layers and the number of neurons in each layer of the neural network were variables used to find the best results using the grid search method.
In order to avoid the problem of over-fitting in neural networks, this study added the dropout method between hidden layers for regularization. However, the experimental results showed that the use of dropout regularization in the neural network only slightly improved the prediction of values. In order to normalize output values in the range 0 to 1, the output layer adopted the sigmoid function as shown in Equation (8). The activation function of the remaining layers was different from the commonly-used ReLU activation function [29]. The novel Mish function was used instead [30], as shown in Equation (9). The experimental process is shown in Figure 8.
s i g m o i d ( x ) = 1 1 + e x
M i s h ( x ) = x tanh ( ln ( 1 + e x ) )
As can be seen from Figure 8, the Mish function behaves differently than the commonly-used ReLU and sigmoid functions [31]. It can effectively and quickly converge the loss, and it is less prone to the problem of network weakening when the loss increases in a long iteration.

3.2.4. Evaluation Criteria and Training Results

In order to evaluate the performance of the neural network, the study adopted the commonly-used statistical standards, mean absolute error (MAE) and root mean squared error (RMSE), as evaluation methods. The formulas are shown in Equations (10) and (11). The lower the value, the better the performance of the neural network.
M A E = 1 n i = 1 N | y ^ i y i y i |
R M S E = i = 1 N ( y ^ i y i ) 2 / N
where N represents the quality characteristic quantity, and y ^ i and y i represent the predicted value and the actual value, respectively.
The mean absolute error and root mean square error of the training results for the validation data sets with neural network grid search are shown in Table 2 and Table 3, respectively.
It is observed from Table 2 and Table 3 that the mean absolute error and the root mean square error can be reduced by increasing the number of neurons and the hidden layers. However, if the number of these two is increased too much, it will lead to overfitting, so that the error in the validation dataset increases. After using the grid search method [32] to obtain the optimal number of hidden layers and neurons in the neural network, we added dropout regularization between the hidden layers in the neural network to reduce the problem of overfitting, as shown in Figure 9.
Despite trying to reduce overfitting in the neural network through the use of dropout regularization, the experimental results showed it only slightly reduced errors in predicted numerical values in the validation data set. Then, we used three algorithms to improve gradient descent for the final optimization of the neural network, namely SGDM [33], RMSProp [34], and Adam [35]. In order to solve the common problem of the basic gradient descent falling into a local optimal solution and not being able to escape, we made use of the SGDM gradient descent method with momentum.
V i = β V i 1 γ L W
W i = W i 1 + V i
The above equation is the SGDM gradient descent formula with momentum. Compared with the basic SGDM gradient descent method, directional velocity V and momentum β are added.
Starting by testing with a smaller number of iterations, as shown in Figure 10, we found that the RMSProp algorithm reduced error rapidly at the beginning, but its performance declined after a larger number of iterations. The Adam algorithm and the SGDM algorithm, on the other hand, behaved similarly and converged effectively on a smaller error rate. Therefore, the number of iterations was extended in the experiment, and only the performance of the Adam algorithm and the SGDM algorithm was compared. The results are shown in Figure 11.
After 1000 iterations, the Adam algorithm and the SGDM algorithm had quite similar training curves, with mean absolute errors of 0.0713 and 0.0709, and root mean square errors of 0.0917 and 0.0905, respectively. Although there was not much difference between the two, the SGDM algorithm reduced the loss slightly more effectively than the Adam algorithm. The detailed optimal training process is shown in Table 4.
Finally, the quality characteristic values predicted by the neural network are shown in Table 5 as a figure between 0 and 1 calculated by the sigmoid function. The results show that the neural network effectively predicted the multiple quality characteristic values that particular processing parameter settings would produce. Not only could expected results be predicted before an experimental run, but the search for optimization parameters could also be further carried out.
As can be seen from Figure 12, Figure 13, Figure 14 and Figure 15, the neural network model could successfully and effectively predict the effect of various combinations of processing parameters on the corresponding fineness, breaking strength, elongation at break, and modulus of resilience quality characteristics.
Compared with the traditional Taguchi analysis method, which requires the setting up of orthogonal arrays, carrying out main effect analysis, analysis of variance, confirmatory tests, and undertaking other time-consuming steps [5], the neural network model conducts self-training and learning with past historical data, meaning it can analyze the data more efficiently.
It was the aim of this study to obtain optimal processing parameters for minimum fineness and maximum breaking strength, elongation at break, and modulus of resilience. Therefore, because of the high speed at which deep learning neural networks could be calculated (one data prediction does not require one millisecond), the grid search method was used to exhaustively find the best combination of processing parameters with a mean square error minimum for fineness and a maximum one for breaking strength, elongation at break and modulus of resilience. The result of the search was that when the screw temperature is 180 °C, the gear pump temperature is 220 °C, the die head temperature is 240 °C, the screw speed is 7.5 rpm, the gear pump speed is 15 rpm, and the take-up speed is 700 rpm, the output quality characteristic is that the fineness is 243 Denier (D), the breaking strength is 3.4 N/mm2, the breaking elongation is 643%, and the elastic energy modulus is 9.13 N/mm2, as shown in Table 6.

3.3. Creating Historical Data and Abnormal Samples

Having obtained the best parameters with the neural network grid search method, this set of processing parameters was used as experimental parameters to produce one data set made up of 20 normal samples, as shown in Table 7. Then, to produce two more abnormal experimental data sets, for the first with one abnormal parameter setting and the second with two abnormal settings, one or two processing parameters were changed in sequence, as shown in Table 8. Therefore, for the first single-factor abnormal samples for the A setting there, the value was Abnormal 1 and for the second samples it was Abnormal 2, with the remaining processing parameters staying the same. For two-factor abnormal samples, two processing parameters were changed at a time, with the others staying the same. Therefore, for ones where the A and B settings were changed, at the same time, they were Abnormal 1 and then they were Abnormal 2. The remaining processing parameters were changed according to this same rule.
In this study, for both normal and abnormal data sets, the generation of 20 samples was taken as a standard, so a total of 20 samples were produced for the best parameters, There were twenty samples for each single-factor abnormal processing parameter (both Abnormal 1 and Abnormal 2), and 20 samples for each two-factor abnormal sample pairing (both Abnormal 1 and Abnormal 2). A total of 20 normal samples, 120 (=6 × 20) single-factor abnormal samples, and 300 (=15 × 20) two-factor abnormal samples were obtained.

3.4. Abnormal Processing Parameter Classifier Model Training

In order to determine which processing parameters cause quality characteristic abnormalities, and to improve the process yield of the melt spinning machine, the 420 samples of the abnormal data sets and 20 of the normal data set were applied to train an artificial intelligence classifier, using the neural network to generate quality characteristic predictions. Input feature x i , as shown in Equation (14), was the differences between the four quality characteristics y i (fineness, breaking strength, elongation at break, elastic energy modulus) actually obtained in the abnormal sample data and the corresponding y ^ i predictions of the neural network quantity, as shown in Table 9. The output was one-hot encoded classification results for the processing parameters of screw temperature, gear pump temperature, die head temperature, screw speed, gear pump speed, and take-up speed, as shown in Table 10.
x i = ( y ^ i y i ) x
When training the classifier, if all the data are directly classified as the result of one or two factors, errors are likely because some values are too close to allow abnormal processing parameter settings to be correctly judged. In order to improve the accuracy of the classifier, it was necessary to divide the abnormal samples into one-factor or two-factor groups first, and then treat them separately. Therefore, this study needed to use a total of three classifiers to make predictions. When the classifier was classified as single or double, the output was a single or two-factor abnormality; when a single-factor classification was used, the output was the corresponding abnormal processing parameter type; when a two-factor classification was used, the output was the corresponding two abnormal processing parameter types.
Because relevant studies in the literature have achieved outstanding performance with a number of different classifiers, this study compared these commonly-used machine learning methods with the results found here. Classification methods such as decision trees, random forests, support vector machines, and neural network methods such as deep learning are all possible bases for comparison. Since each of these classifier methods have many hyper parameters to be adjusted, the grid search method was used to find the best parameter combination of each method. A random forest architecture diagram is shown in Figure 16.

3.4.1. Single and Double Identification

For determining the difference for each sample between the actual four quality characteristics of the abnormal sample data, and the four quality characteristics predicted by the neural network quantity, an array of 4 values was used as the input feature of the classifier. There were 420 abnormal samples and 20 normal samples in the data set. The data set was randomly divided into a 220-item training data set and a 220-item verification data set. First, the training data set was used to construct a model, and then the verification data set was used for testing. The purpose of the verification data set was to determine whether the melt spinning abnormal diagnosis system could adequately detect processing parameters responsible for abnormalities or not. Training also used grid search and cross-validation to find the best settings for each classifier. Evaluation of the classifier used the detection success rate, which is the ratio of the number of samples correctly identified to the total number of samples in all the validation data sets, as shown in Equation (15). The larger the value, the better. The detection success rates of various classification methods are shown in Table 11. The success of the random forest classification method can be seen in the confusion matrix in Figure 17, where 0 means no abnormality, 1 means a single abnormality, and 2 means a double abnormality. In the validation data set, the method was the best at identifying single and double abnormalities, and there were no lack of abnormalities, single abnormalities, and double abnormalities in the 220 samples which it missed. All predictions were correct. As for grid search hyperparameters, a random forest with eight decision trees and a maximum depth of four worked best.
Detection   success   rate = Correct   number   of   samples   to   be   tested All   sample   numbers

3.4.2. One-Factor Classification

Single-factor abnormal discrimination can be carried out after single-double identification is performed and has indicated there is one abnormal parameter setting. In development of the study’s classifier, there were a total of 120 samples, 60 were randomly selected for training, and the other 60 formed the verification data set. The results for single-factor classification, shown in Table 12, were similar to the single-double identification ones, and the use of random forest classification was again the best. From the confusion matrix in Figure 18, where 0 to 5 represent screw temperature, gear pump temperature, die head temperature, screw speed, gear pump speed, and take-up speed, respectively, it can be seen that only one gear pump speed abnormality in the 60-item verification data set was misjudged as an abnormal screw temperature, so the detection success rate was as high as 98.3%. Identification by the random forest classifier of which processing parameter caused the abnormality worked very well. As for grid search hyperparameters, the random forest with twelve decision trees and a maximum depth of four performed best.

3.4.3. Two-Factor Classification

Finally, a two-factor abnormality classifier was developed with a total of 300 samples, 150 randomly selected for the training data set and 150 for the verification data set. The experimental results are shown in Table 13. It can be seen that the random forest classifier performed much better than the decision tree and neural network ones. Its confusion matrix is shown in Figure 19, with 0 to 4 representing the combination of screw temperature and the remaining 5 abnormal processing parameters, 5 to 8 representing the combination of gear pump temperature and the remaining 4 abnormal processing parameters, 9 to 11 representing the combination of die temperature and the other three abnormal processing parameters, 12 and 13 representing the combination of the screw speed and the other two abnormal processing parameters, and 14 representing the abnormal processing parameter combination of the gear pump speed and the take-up speed. It can be observed that only 6 misjudgments were made of verification data set samples, and the detection success rate was as high as 96.0%. These 6 misjudged samples were made up of a combination of different abnormal processing parameters, with no parameter appearing in more than one sample, indicating that model overfitting is not a problem. As for the grid search hyperparameters, the random forest with twenty-four decision trees and a maximum depth of six worked best.
It can be concluded that the random forest achieved the best accuracy rate, with results similar to those in the literature. With comparatively few data samples, it also exhibited greater anti-overfitting resistance. This study’s random forest classifier was compared with the use of the decision tree classifier and the four values obtained by the RAM method as input features in their ability to identify abnormal processing parameters. As can be seen from the final overall accuracy rate in Table 14, the results for both are outstanding. However, using the deep learning neural network and random forest classifier of this study to do abnormal processing parameter detection avoids the need to compute Hotelling’s T2 for abnormal product detection first, by directly classifying the data on the basis of the results obtained from the deep learning neural network. To use the decision tree with the RAM method too much process calculation is required [36,37]. For example, supposing there is some bias in the calculation of Hotelling’s T2 at the beginning. This will lead to indirect errors in the RAM method and the feature input of the decision tree, resulting in misjudgment of the final detection result. In addition, the calculation time required for the final abnormal processing parameter detection in this study is only 0.08 s, meaning it is more efficient in comparison.

4. Conclusions

This study applied a deep learning neural network and random forest from machine learning in artificial intelligence to the optimal quality prediction of multiple quality parameters and quality abnormality diagnosis of melt spinning machines. It included six processing parameters and four qualities. The conclusions are as follows.
(1)
The deep learning neural network is used for experiments, 440 pieces of historical data are trained, and multiple quality optimization parameters are searched by using the characteristic grid of deep learning rapid calculation. The deep learning neural network was used to generate quality predictions, trained on a 440-item historical data set, and multiple quality optimization parameters were searched for using rapid deep-learning characteristic grid calculations. Compared with the traditional Taguchi analysis method, the neural network model conducts self-training and learning using past historical data, which means the research can proceed faster, analysis is more efficient, and conclusions are more robust, because a calculation error in one step will not affect the overall detection system.
(2)
This research compared several artificial intelligence machines learning and deep learning classifiers that have obtained outstanding results in the related literature, and finally selected the random forest as being the best, because its classifier belongs to ensemble learning, and the classifier is resistant to overfitting. Its ability to detect the cause of quality problems was better than that of other classifiers. As an indication the success rate of single and double identification was 100%, the success rate of single factor classification was 98.3%, and the success rate of double factor classification was 96.0%. It can be seen that the proposed method offers an effective way to identify the problematic machine settings, causing problems in quality control after the engineer has measurements of the abnormality so that the settings can be quickly modified to improve production yield.
(3)
This study applied the methods of artificial intelligence to the development of an abnormal processing PP fiber melt spinning parameter identification system which can quickly find abnormal settings and reduce unnecessary cost and waste. In the future, different online detection systems matching the capabilities of this system for various other kinds of material will be added to the resources available to production engineers seeking to apply the developed identification system for its functions of selection and evaluation.

Author Contributions

Conceptualization, C.-F.J.K. and Y.-S.L.; methodology, C.-F.J.K. and Y.-S.L.; validation, A.K.G., Y.-S.L. and C.-F.J.K.; formal analysis, A.K.G.; investigation, Y.-S.L., A.K.G.; resources, C.-F.J.K.; data curation, A.K.G.; writing—original draft preparation, Y.-S.L.; writing—review and editing, A.K.G. and C.-F.J.K.; visualization, A.K.G.; supervision, C.-F.J.K.; project administration, C.-F.J.K.; funding acquisition, C.-F.J.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry of Science and Technology of Taiwan under Grant No. 110-2622-E-011-012 and the APC was funded by Author voucher code (319c911c64fecf02).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kuo, C.F.J.; Syu, S.S.; Lin, C.H.; Peng, K.C. The application of principal component analysis and gray relational method in the optimization of the melt spinning process using the cooling air system. Text Res. J. 2013, 83, 371–380. [Google Scholar] [CrossRef]
  2. Asmatulu, R.; Khan, W.S. Synthesis and Applications of Electrospun Nanofibers, Micro and Nano Technologies; Elsevier Publishing: Amsterdam, The Netherlands, 2019; pp. 1–15. [Google Scholar]
  3. Iba, H.; Nasimul, N. Deep Neural Evolution; Springer: Berlin/Heidelberg, Germany, 2020. [Google Scholar]
  4. Lam, Y.C.; Zhai, L.Y.; Tai, K.; Fok, S.C. An evolutionary approach for cooling system optimization in plastic injection moulding. Int. J. Prod. Res. 2004, 42, 2047–2061. [Google Scholar] [CrossRef]
  5. Chen, W.C.; Wang, M.W.; Fu, G.L.; Chen, C.T. Optimization of plastic injection molding process via Taguchi’s parameter design method, BPNN, and DFP. In Proceedings of the 2008 International Conference on Machine Learning and Cybernetics, IEEE, Kumming, China, 12–15 July 2008; Volume 6, pp. 3315–3321. [Google Scholar]
  6. Abiodun, O.I.; Jantan, A.; Omolara, A.E.; Dada, K.V.; Mohamed, N.A.; Arshad, H. State-of-the-art in artificial neural network applications: A survey. Heliyon 2018, 4, e00938. [Google Scholar] [CrossRef] [Green Version]
  7. Majumdar, P.K.; Majumdar, A. Predicting the breaking elongation of ring spun cotton yarns using mathematical, statistical, and artificial neural network models. Text. Res. J. 2004, 74, 652–655. [Google Scholar] [CrossRef]
  8. Dave, V.S.; Dutta, K. Neural network-based models for software effort estimation: A review. Artif. Intell. Rev. 2014, 42, 295–307. [Google Scholar] [CrossRef]
  9. Bai, Y.; Li, C.; Sun, Z.; Chen, H. Deep neural network for manufacturing quality prediction. In Proceedings of the 8th IEEE Prognostics and System Health Management Conference, Harbin, China, 9–12 July 2017. [Google Scholar]
  10. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  11. Sugumaran, V.; Ramachandran, K.I. Automatic rule learning using decision tree for fuzzy classifier in fault diagnosis of roller bearing. Mech. Syst. Signal Processing 2007, 21, 2237–2247. [Google Scholar] [CrossRef]
  12. Wang, Q.; Ahmad, W.; Ahmad, A.; Aslam, F.; Mohamed, A.; Vatin, N.I. Application of Soft Computing Techniques to Predict the Strength of Geopolymer Composites. Polymers 2022, 14, 1074. [Google Scholar] [CrossRef]
  13. Nafees, A.; Amin, M.N.; Khan, K.; Nazir, K.; Ali, M.; Javed, M.F.; Aslam, F.; Musarat, M.A.; Vatin, N.I. Modeling of mechanical properties of silica fume-based green concrete using machine learning techniques. Polymers 2021, 14, 30. [Google Scholar] [CrossRef]
  14. Zimmerman, R.K.; Balasubramani, G.K.; Nowalk, M.P.; Eng, H.; Urbanski, L.; Jackson, M.L.; Jackson, L.A.; McLean, H.Q.; Belongia, E.A.; Monto, A.S.; et al. Classification and regression tree (CART) analysis to predict influenza in primary care patients. BMC Infect. Dis. 2016, 16, 503. [Google Scholar] [CrossRef] [Green Version]
  15. Nafees, A.; Javed, M.F.; Khan, S.; Nazir, K.; Farooq, F.; Aslam, F.; Musarat, M.A.; Vatin, N.I. Predictive Modeling of Mechanical Properties of Silica Fume-Based Green Concrete Using Artificial Intelligence Approaches: MLPNN, ANFIS, and GEP. Materials 2021, 14, 7531. [Google Scholar] [CrossRef] [PubMed]
  16. Cerrada, M.; Zurita, G.; Cabrera, D.; Sánchez, R.V.; Artés, M.; Li, C. Fault diagnosis in spur gears based on genetic algorithm and random forest. Mech. Syst. Signal Process. 2016, 70, 87–103. [Google Scholar] [CrossRef]
  17. Li, C.; Sanchez, R.; Zurita, G.; Cerrada, M.; Cabrera, D.; Vásquez, R.E. Gearbox fault diagnosis based on deep random forest fusion of acoustic and vibratory signals. Mech. Syst. Signal Process. 2016, 76, 283–293. [Google Scholar] [CrossRef]
  18. Esmaily, H.; Tayefi, M.; Doosti, H.; Ghayour-Mobarhan, M.; Nezami, H.; Amirabadizadeh, A. A comparison between decision tree and random forest in determining the risk factors associated with type 2 diabetes. J. Res. Health Sci. 2018, 18, 412. [Google Scholar]
  19. Yanjun, Q. Random forest for bioinformatics. In Ensemble Machine Learning; Springer: Boston, MA, USA, 2012; pp. 307–323. [Google Scholar]
  20. Ansoategui, F.J.; Campa, C.; Díez, M. Influence of the machine tool compliance on the dynamic performance of the servo drives. Int. J. Adv. Manuf. Technol. 2017, 90, 2849–2861. [Google Scholar] [CrossRef]
  21. Ali, J.B.; Fnaiech, N.L.; Chebel-Morello, S.B.; Fnaiech, F. Application of empirical mode decomposition and artificial neural network for automatic bearing fault diagnosis based on vibration signals. Appl. Acoust. 2015, 89, 16–27. [Google Scholar]
  22. Hastie, T.R.; Friedman, T.J. Random forests. In The Elements of Statistical Learning; Springer: New York, NY, USA, 2019; pp. 587–604. [Google Scholar]
  23. Dietterich, T.G. Ensemble learning. Handb. Brain Theory Neural Netw. 2002, 2, 110–125. [Google Scholar]
  24. Ying, C.; Qi-Guang, M.; Jia-Chen, L.; Lin, G. Advance and prospects of AdaBoost algorithm. Acta Autom. Sin. 2013, 39, 745–758. [Google Scholar]
  25. Nwankpa, C.; Ijomah, W.; Gachagan, A.; Marshall, S. Activation functions: Comparison of trends in practice and research for deep learning. arXiv 2018, arXiv:1811.03378. [Google Scholar]
  26. Ruder, S. An overview of gradient descent optimization algorithms. arXiv 2016, arXiv:1609.04747. [Google Scholar]
  27. Elmaaty, T.A.; Taweel, F.E.I.; Elsisi, H.; Okubayashi, S. Water free dyeing of polypropylene fabric under supercritical carbon dioxide and comparison with its aqueous analogue. J. Supercrit. Fluids 2018, 139, 114–121. [Google Scholar] [CrossRef]
  28. Elmaaty, T.A.; Sofan, M.; Elsisi, H.; Kosbar, T.; Negm, E.; Hirogaki, K.; Tabata, I.; Hori, T. Optimization of an eco-friendly dyeing process in both laboratory scale and pilot scale supercritical carbon dioxide unit for polypropylene fabrics with special new disperse dyes. J. CO2 Util. 2019, 33, 365–371. [Google Scholar] [CrossRef]
  29. Gonzalez, T.F. Handbook of Approximation Algorithms and Metaheuristics; Chapman and Hall/CRC: Boca Raton, FL, USA, 2007. [Google Scholar]
  30. Misra, D. Mish: A self-regularized non-monotonic activation function. arXiv 2019, arXiv:1908.08681. [Google Scholar]
  31. Barry, P. Sigmoid functions and exponential Riordan arrays. arXiv 2017, arXiv:1702.04778. [Google Scholar]
  32. Bergstra, J.; Bengio, Y. Random search for hyper-parameter optimization. J. Mach. Learn. Res. 2012, 13, 281–305. [Google Scholar]
  33. Schaul, T.; Antonoglou, I.; Silver, D. Unit tests for stochastic optimization. arXiv 2013, arXiv:1312.6055. [Google Scholar]
  34. Wang, Y.; Liu, J.; Mišić, J.; Mišić, V.B.; Lv, S.; Chang, X. Assessing optimizer impact on dnn model sensitivity to adversarial examples. IEEE Access 2019, 7, 152766–152776. [Google Scholar] [CrossRef]
  35. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  36. Saraswat, S.; Yadava, G.S. An overview on reliability, availability, maintainability and supportability (RAMS) engineering. Int. J. Qual. Reliab. Manag. 2008, 25, 330–344. [Google Scholar] [CrossRef]
  37. Kuo, C.F.J.; Huang, C.C.; Yang, C.H. Integration of multivariate control charts and decision tree classifier to determine the faults of the quality characteristic(s) of a melt spinning machine used in polypropylene as-spun fiber manufacturing Part I: The application of the Taguchi method and principal component analysis in the processing parameter optimization of the melt spinning process. Text. Res. J. 2021, 91, 1815–1829. [Google Scholar]
Figure 1. Flowchart for anomaly detection.
Figure 1. Flowchart for anomaly detection.
Polymers 14 02739 g001
Figure 2. Single-layer neural network architecture.
Figure 2. Single-layer neural network architecture.
Polymers 14 02739 g002
Figure 3. Chemical structure of PP.
Figure 3. Chemical structure of PP.
Polymers 14 02739 g003
Figure 4. Melt spinning machine process.
Figure 4. Melt spinning machine process.
Polymers 14 02739 g004
Figure 5. Thermogravimetric analysis of PP.
Figure 5. Thermogravimetric analysis of PP.
Polymers 14 02739 g005
Figure 6. DSC analysis of PP.
Figure 6. DSC analysis of PP.
Polymers 14 02739 g006
Figure 7. The architecture of the neural network.
Figure 7. The architecture of the neural network.
Polymers 14 02739 g007
Figure 8. The effect of common activation functions used in the hidden layer on the performance.
Figure 8. The effect of common activation functions used in the hidden layer on the performance.
Polymers 14 02739 g008
Figure 9. The effect of adding dropout between hidden layers on performance.
Figure 9. The effect of adding dropout between hidden layers on performance.
Polymers 14 02739 g009
Figure 10. Neural network using different gradient descent methods for comparison.
Figure 10. Neural network using different gradient descent methods for comparison.
Polymers 14 02739 g010
Figure 11. The comparison of Adam and SGDM algorithm.
Figure 11. The comparison of Adam and SGDM algorithm.
Polymers 14 02739 g011
Figure 12. Comparison of predicted and actual values in the fineness verification data set after normalization.
Figure 12. Comparison of predicted and actual values in the fineness verification data set after normalization.
Polymers 14 02739 g012
Figure 13. Comparison of predicted and actual values in the fracture strength verification data set after normalization.
Figure 13. Comparison of predicted and actual values in the fracture strength verification data set after normalization.
Polymers 14 02739 g013
Figure 14. Comparison of predicted and actual values in the elongation at break verification data set after normalization.
Figure 14. Comparison of predicted and actual values in the elongation at break verification data set after normalization.
Polymers 14 02739 g014
Figure 15. Comparison of predicted and actual values in the modulus of resilience verification data set after normalization.
Figure 15. Comparison of predicted and actual values in the modulus of resilience verification data set after normalization.
Polymers 14 02739 g015
Figure 16. Random forest architecture.
Figure 16. Random forest architecture.
Polymers 14 02739 g016
Figure 17. The confusion matrix of the classification results of random forest single and double identification.
Figure 17. The confusion matrix of the classification results of random forest single and double identification.
Polymers 14 02739 g017
Figure 18. Confusion matrix of random forest single-factor classification results.
Figure 18. Confusion matrix of random forest single-factor classification results.
Polymers 14 02739 g018
Figure 19. Confusion matrix of random forest two-factor classification results.
Figure 19. Confusion matrix of random forest two-factor classification results.
Polymers 14 02739 g019
Table 1. The range of melt spinning machine processing parameters.
Table 1. The range of melt spinning machine processing parameters.
RangeScrew TemperatureGear Pump TemperatureDie Head TemperatureScrew Speed Gear Pump SpeedTake-Up Speed
Lowest160 °C200 °C210 °C5 rpm15 rpm300 rpm
Highest200 °C240 °C250 °C10 rpm25 rpm700 rpm
Table 2. Mean absolute error of validation data sets for neural network grid search.
Table 2. Mean absolute error of validation data sets for neural network grid search.
Neurons in Each Layer20304050
No. of Hidden Layers
20.0840.0790.0880.085
30.0780.0780.0790.080
40.0760.0740.0730.074
50.0750.0750.0750.077
Table 3. Root mean square error of validation data sets for neural network grid search.
Table 3. Root mean square error of validation data sets for neural network grid search.
Neurons in Each Layer20304050
No. of Hidden Layers
20.1040.1020.1050.104
30.1010.1010.1020.102
40.0980.0950.0930.096
50.0970.0970.0980.101
Table 4. Best training results of neural network classes.
Table 4. Best training results of neural network classes.
Basic Neural NetworkReLUMishDropoutAdamRMSPropSGDMMAERMSE
TT T 0.0750.097
T T T 0.0730.091
T TTT 0.0710.092
T TT T 0.0720.093
T TT T0.0710.091
(Notes: T stands for use).
Table 5. The outputs of the neural network.
Table 5. The outputs of the neural network.
QualityFinenessBreaking StrengthElongation at BreakModulus of Resilience
No.
10.37560.47500.77920.3485
20.49110.69830.14020.9094
30.60380.19090.66640.8443
40.79080.120010.7692
50.16880.388700.5886
60.712810.22880.1552
70.64790.53970.04660.7645
80.422700.71210.5876
90.50200.68140.53380.6714
100.11600.40720.68501
110.42360.59230.70310
120.90140.51120.42810.9307
130.10140.88920.89020.9345
140.79150.74020.67980.4314
1500.26480.14860.4834
160.49650.39880.78970.6255
1710.61180.57820.6722
180.15560.55480.50800.5609
190.18750.94080.76260.8439
200.08910.89040.83770.9105
Table 6. Predicted quality data for optimal processing parameters.
Table 6. Predicted quality data for optimal processing parameters.
Fineness (dB)Breaking Strength (dB)Elongation at Break (dB)Modulus of Resilience (dB)
Predication value0.1830.8720.9470.935
Denormalized value2433.46439.13
Table 7. Twenty data samples with the best parameters.
Table 7. Twenty data samples with the best parameters.
Best Parameter Data
QualityFineness
(Diner)
Breaking Strength
(N/mm2)
Elongation at Break
(%)
Modulus of Resilience
(N/mm2)
Samples
12363.1641.9729.03
22372.8648.3059.40
32373.4648.3579.39
42312.8644.2249.28
52413648.2659.45
62493.6635.9239.52
72272.9642.8458.74
82313.6641.2188.82
92323.6646.8019.30
102383.5645.2169.36
112413.5643.5069.03
122313.5640.7259.48
132363641.3789.03
142473.4646.7769.49
152512.8643.3939.79
162453.5642.9428.99
172402.7641.1559.54
182403.6642.8119.20
192343.6640.9119.04
202573.6646.9268.87
Table 8. Abnormal sample processing parameter settings.
Table 8. Abnormal sample processing parameter settings.
ABCDEF
Screw Temperature (°C)Gear Pump Temperature (°C)Die Head Temperature (°C)Screw Speed (rpm)Gear Pump Speed (rpm)Take-Up Speed (rpm)
Normal1802202407.515700
Abnormal
1
190200220520300
Abnormal
2
2002102301025500
Table 9. Quality data for 10 abnormal samples.
Table 9. Quality data for 10 abnormal samples.
QualityFineness
(dB)
Breaking Strength
(dB)
Elongation at Break
(dB)
Modulus of Resilience
(dB)
Sets
12242643.7918.97
22492.8655.6677.66
35621.8591.1979.14
45982.9642.0686.99
53162.2520.8318.92
62833.3531.7916.09
75512.8645.0448.73
83473.1647.5418.96
92543.2606.2699.60
102962.2638.9888.93
Table 10. Classification results for the corresponding abnormal samples of Table 9.
Table 10. Classification results for the corresponding abnormal samples of Table 9.
Processing ParameterScrew TemperatureGear Pump TemperatureDie Head TemperatureScrew SpeedGear Pump SpeedTake-Up Speed
Sets
1010000
2100100
3000110
4000001
5010010
6100001
7001000
8000100
9100000
10011000
Table 11. Comparison of single and double identification and classification methods.
Table 11. Comparison of single and double identification and classification methods.
MethodSingle and Double Identification Detection Success Rate
Decision tree98.5%
Radom forest100%
Support vector machine98.1%
Neural network98.1%
Table 12. Comparison of single factor classification methods.
Table 12. Comparison of single factor classification methods.
MethodSingle Factor Classification Detection Success Rate
Decision tree95.0 %
Radom forest98.3 %
Neural network96.8 %
Table 13. Comparison of two-factor classification results.
Table 13. Comparison of two-factor classification results.
MethodTwo-Factor Classification Detection Success Rate
Decision tree91.8%
Radom forest96.0%
Neural network89.3%
Table 14. Overall classification accuracy comparison.
Table 14. Overall classification accuracy comparison.
Single and Double IdentificationOne-Factor ClassificationTwo-Factor Classification
This research100%98.3%96.0%
Decision tree + RAM method98.60%98.3%95.3%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gope, A.K.; Liao, Y.-S.; Kuo, C.-F.J. Quality Prediction and Abnormal Processing Parameter Identification in Polypropylene Fiber Melt Spinning Using Artificial Intelligence Machine Learning and Deep Learning Algorithms. Polymers 2022, 14, 2739. https://doi.org/10.3390/polym14132739

AMA Style

Gope AK, Liao Y-S, Kuo C-FJ. Quality Prediction and Abnormal Processing Parameter Identification in Polypropylene Fiber Melt Spinning Using Artificial Intelligence Machine Learning and Deep Learning Algorithms. Polymers. 2022; 14(13):2739. https://doi.org/10.3390/polym14132739

Chicago/Turabian Style

Gope, Amit Kumar, Yu-Shu Liao, and Chung-Feng Jeffrey Kuo. 2022. "Quality Prediction and Abnormal Processing Parameter Identification in Polypropylene Fiber Melt Spinning Using Artificial Intelligence Machine Learning and Deep Learning Algorithms" Polymers 14, no. 13: 2739. https://doi.org/10.3390/polym14132739

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop