Next Article in Journal
Variable Switching Frequency for ZVS over Wide Voltage Range in Dual Active Bridge
Previous Article in Journal
Exploration of Eye Fatigue Detection Features and Algorithm Based on Eye-Tracking Signal
Previous Article in Special Issue
Forester: Approximate Processing of an Imperative Procedure for Query-Time Exploratory Data Analysis in a Relational Database
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Data-Driven Prediction Model for Analysis of Sensor Data

by
Ognyan Yotov
and
Adelina Aleksieva-Petrova
*
Faculty of Computer Systems and Technologies, Technical University of Sofia, 1000 Sofia, Bulgaria
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(10), 1799; https://doi.org/10.3390/electronics13101799
Submission received: 27 March 2024 / Revised: 28 April 2024 / Accepted: 30 April 2024 / Published: 7 May 2024

Abstract

:
In view of Industry 4.0, data generation and analysis are challenges. For example, machine health monitoring and remaining useful life prediction use sensor signals, which are difficult to analyze using traditional methods and mathematical techniques. Machine and deep learning algorithms have been used extensively in Industry 4.0 to process sensor signals and improve the accuracy of predictions. Therefore, this paper proposes and validates the data-driven prediction model to analyze sensor data, including in the data transformation phase Principal Component Analysis tested by Fourier Transformation and Wavelet Transformation, and the modeling phase based on machine and deep learning algorithms. The machine learning algorithms used for tests in this research are Random Forest Regression (RFR), Multiple Linear Regression (MLR), and Decision Tree Regression (DTR). For the deep learning comparison, the algorithms are Deep Learning Regression and Convolutional network with LeNet-5 Architecture. The experimental results indicate that the models show promising results in predicting wear values and open the problem to further research, reaching peak values of 92.3% accuracy for the first dataset and 62.4% accuracy for the second dataset.

1. Introduction

Industry 4.0 refers to the fourth industrial revolution, which is characterized by the integration of advanced technologies into various industries. Machine learning (ML), a subset of artificial intelligence, plays a crucial role in this revolution. It enables efficient analysis of large volumes of data, identifying correlations, patterns, and insights that can drive innovation and improve processes across different fields, such as machine health monitoring, quality control, anomaly detection, and others.
Industry 4.0 relies on two main concepts—automation and data [1]—which are interdependent; the development of one correlates to the development of the other and vice versa. Data-driven approaches need huge amounts of data, which are conducted on servers located in data centers or the cloud. Data-driven techniques such as knowledge-based models, life expectancy models, and neural networks can efficiently create behavioral models. However, these methods primarily rely on historical data spanning from a healthy state to a failure state of the system being monitored [2]. This approach increases system complexity and helps with system failure [3]. The development in the field of sensorics and wireless technologies enabled the generation of data for efficient analysis in different industry applications.
Analysis of sensor signals with machine learning is becoming a fundamental part of the optimization of production processes because of the complexity of these signals and the advantages and flexibility of machine and deep learning algorithms. Complex neural network architectures, mathematical transformations, and regularizations enable these algorithms to find correlations and patterns in high-dimensional data.
For many machine learning based on predictive maintenance applications in various industrial processes feature selection represents an important part. The feature selection methods, such as Fisher score, mRMR, and ReliefF, show different results according to the size of the datasets. The stability of all the methods suffered in some cases under the small datasets, but with large datasets, Fisher’s score has the stability and predictive performance in almost all the individual cases [4].
Deep Learning (DL) has become a rapidly developed and researched part of ML. For analysis and processing with large datasets, deep learning algorithms are showing promising results [5,6,7]. Also, it shows excellent performance in different areas, such as object recognition, speech recognition, state prediction, and image segmentation. The benefits of deep learning applications in manufacturing are optimization of operation costs, productivity improvement, fault diagnosis, downtime optimization, and improvement of processes.
To exploit the advantages of these approaches, we propose a data-driven prediction model to analyze sensor data using machine and deep learning algorithms. The data utilized in this study come from a shearing-cutting operation and is collected through sensors located on the press. To validate the proposed model, the test case will be a prediction of different wear conditions of a punch during the stamping process. During operation, the punch of a stamping press is critical to the quality of the produced goods. The wear of the punch plays a crucial role in achieving the desired quality. The conditions are dynamically changing during the process—wear proactively hinders the quality of the product.
Our main contributions can be summarized as follows:
  • First, we propose a data-driven model to analyze and extract data and features from the datasets. Using mathematical transformations, such as PCA-Analysis, for dimensionality reduction and optimization. After this optimization, the dataset undergoes Fourier and Wavelet transformations to shift the approach to the dataset without losing data;
  • The transformed and optimized datasets are used in ML and DL algorithms for prediction and classification—benchmarking them against each other for comparison;
  • The datasets are generated from test benches designed according to real-world standards.
The rest of the paper is organized as follows. Section 2 presents the related works in the application of machine and deep learning in manufacturing as production quality and its optimization. Section 3 describes the proposed data-driven prediction model to analyze sensor data using machine and deep learning algorithms. Section 4 explains the data transformation phase, which includes Principal component analysis, which is tested by Fourier transformation and Wavelet transformation. Section 5 is the modeling phase based on machine and deep learning algorithms, and Section 6 explains the validation phase. In Section 7, the results of method validation are given and discussed. Section 8 is conclusions.

2. Related Works

In this paper, we use sensor signal analysis in order to be able to predict the state of machine health, which can lead to an increase in production quality and optimization. This is linked to two interrelated processes: machine monitoring and anomaly detection. Both processes use different ML and DL models, requiring the combination of many different techniques. For example, the integration of techniques like Pearson correlation coefficient analysis, sequential bidirectional long short-term memory networks, and enhanced remora optimization algorithm introduces a framework used for accurate temperature prediction of the permanent magnet synchronous machine drives [8].
The tool condition monitoring is part of the process of machine monitoring. In the [9] authors discuss how monitoring systems for tool conditions contribute to optimizing machining processes by detecting signs of tool failures and estimating remaining tool life. The paper also highlights the current limitations and blank spots in implementing machine learning, deep learning, and IoT technologies in tool-embedded tool condition monitoring, such as big data handling and generalization of models, as well as latency in cloud computing. Also, the fast and low-cost solution for tool condition monitoring is provided in [10]. The authors used an Arduino node to acquire sound and current consumption signals through low-cost and non-invasive sensors, but again, they built their model through a machine learning node based on a BeagleBoneBlack unit to build the tool wear model using a logistic regression classifier.
In addition, a condition monitoring tool is used for detecting and preventing sudden tool failures [11]. The approach is built on a Discrete Wavelet Transform (DWT) lifting scheme to extract a time-frequency representation of the AErms signals and a long short-term memory autoencoder to compress and reconstruct the DWT features. The suggested approach maximizes the utilization of tool life while ensuring the protection of machined parts, leading to reduced machining downtime and costs.
Machine health monitoring is also part of the machine monitoring process and is affected by the improvements in the fields of sensors and the Internet of Things (IoT). The ability to generate cost-effective data and analyze them finds application in the manufacturing industry [12]. By constantly monitoring systems, a vast amount of data can be gathered and analyzed, opening up a new area of exploration [13]. The study offers a comprehensive overview of the use of ML algorithms in the field of electronics and highlights the state-of-the-art advancements in various electronic components.
To predict future health indicators for prognostic and health management purposes, a combination of the long short-term memory model and an attention mechanism is employed in an iterative manner [14]. Moreover, the similarity method is applied to estimate the predicted remaining useful life distribution of the examined sensor-equipped machine based on historical data from multiple sensor-equipped machines’ health indicators. Also, two distinct ML algorithms are used for structural health monitoring—one involving polynomial regression and the other a shallow neural network [15]. Both are supervised methods involving training with certain data and testing with a supplementary subset of data. The two techniques were first optimized for performance in terms of training and testing and then compared in accuracy. The study confirmed the effectiveness of Machine Learning in identifying impacts.
On the other hand, anomaly detection techniques can be applied in machine health monitoring, which are used to identify unusual patterns or outliers that may indicate potential faults, malfunctions, or abnormal behavior in the monitored machines. Anomaly detection is a process that involves identifying data values or sequences that significantly deviate from the majority of other observations, which are considered normal data. A self-supervised learning (SSL) framework for time-series anomaly detection in an Industrial Internet of Things system is proposed, which consists of two augmentation techniques in time-series data that capture two different patterns of original samples before feeding them to the classifier [16]. The classifier uses a one-dimension convolutional neural network to learn the characteristics of normal data.
The combination of DWT with high and low-frequency separation processing and Long Short-Term Memory (LSTM) for anomaly detection is introduced to extract multi-current signal features and detect anomalies based on weight updating the LSTM network [17]. Experiments on real motor bearing faults and permanent magnet synchronous motor stator faults datasets demonstrate the method’s effectiveness in fusing current features and detecting anomalies.
The primary difficulty in identifying temporal anomalies lies in creating models or systems capable of retaining previous encounters or interpreting patterns from extended and intricate time series data. Distinguishing temporal anomalies should not be mistaken for predicting sensor noise problems, as noise fluctuations or sudden changes in signal amplitude or frequency can be effectively addressed using established digital signal processing techniques like Fast Fourier Transform (FFT) and Radon Fourier Transform (RFT) [18].

3. Data-Driven Prediction Model to Analyze Sensor Data

The proposed data-driven prediction model to analyze sensor data using machine and deep learning algorithms is shown in Figure 1. It is based on the Knowledge Discovery in Time Series for Engineering Applications (KDT-EA), which is a process of generating knowledge from all steps of data management while analyzing and processing datasets consisting of time series (e.g., sensor signals). KDT-EA was developed by the Institute for Production Engineering and Forming Machines at the Technical University of Darmstadt, and it is based on the Knowledge Discovery in Databases (KDD) process [19].
The diagram shows the main phases, which include Data Acquisition, Data Preparation, Data Transformation, Modeling, and Validation. Data Acquisition is composed of two processes: data collection and data cleaning to prepare the dataset. The data preparation phase includes feature selection and data normalization. However, the focus of our research is the Data Transformation, Modeling, and Validation phase, where we proposed different solutions, which are described in the following sections.
The data used in this research are from a shear-cutting process and is generated via sensors on the press. The sensors gather force data during the process. Figure 2a,b shows the difference between no wear and critical wear and the different force curves for different wear conditions.
The two datasets are labeled with different wear conditions. Table 1 shows the difference between both datasets.

4. Data Transformation Phase

For Data Transformation Principal Component Analysis (PCA) is used. This is a statistical process that allows the user to analyze large quantities of tabular datasets by smaller fractions of “summary indices”. PCA is the fundamental block of multivariate data analysis based on projection methods. The most important application of PCA is the ability to represent multivariate tabular data as a smaller set, thus easing the observation and detection of trends, jumps, or clusters. This analysis can detect relationships and correlations between observations and variables. The PCA approach allows the analysis of data with missing values, imprecisions or categorical data, making it flexible and with application in various fields. PCA’s goal is to extract the important information from the dataset and to represent it via the smaller indices called principal components.
In statistical terms, Principal Component Analysis finds lines, planes, and hyper-planes in the K-dimensional space, which can approximate the data in the best way. The approximation via a line or a plane for given data points, on the basis of least squares approximation, makes the variance as large as possible.
After the above explained steps, the first principal component (PC1) is ready to be calculated. PC1 is the line in the K-space with the best approximation according to the least squares rule. Each yellow dot in Figure 3 is an observation that can be projected onto this line, thus getting a value along the PC-line.
The second principal component (PC2) is represented again via a line in the same K-space, and the line is orthogonal to the first PC-line. Passing through the origin (mean value), PC2 helps the approximation.
Data transformation with PCA is performed only after feature extraction using TSFRESH/TSFEL libraries.

4.1. Fourier Transformation

For these tests, a Fourier Transform is performed to achieve better results. The Fourier Transform (FT) is a mathematical transformation that transfers the function from the time domain to the frequency domain. The magnitude of this transform is equal to the amount of presence of a given frequency in the original function. In mathematical terms, the Fourier Transform is explained by the following Equation:
f ^ ( ω ) = f t e j ω t d t
  • f ^ is the continuous Fourier Transform of the function f;
  • ω is the angular frequency;
  • j is the imaginary unit.
Figure 4 shows the difference between them for the force signal used in the tests. The frequency domain graph shows only the first section of the signal (0–200) samples because of scaling issues.

4.2. Wavelet Transformation

Fourier transformation has a high resolution in the frequency domain, but in the time domain, it has a resolution of zero. Thus, Fourier transformation gives exact information about the frequencies present but no information about the time they occurred. Figure 5 shows the difference between two signals in time and frequency domain.
As seen, the two frequency spectra contain the exact number of peaks—four—meaning the information after Fourier Transformation is insufficient.
The better approach for analyzing signals with a dynamic frequency spectrum is the Wavelet Transform. This transformation has high resolution in both domains’ frequency and time, giving information not only about the frequencies present but also about the time they have occurred. This is achieved by using and working with different scales. First, the signal is analyzed with large scale and “large” features are extracted, afterwards with smaller scales in order to extract smaller features. Figure 6 shows the comparison between different transformations with the signal in the time domain as a reference. An indication of the size of the features is the scale and orientation of the blocks; as seen, the original time series has high resolution in the time domain but low in the frequency domain. This enables the detection of small features in the time domain and none in the frequency domain. The Fourier Transformation enables the opposite—high resolution in frequency and low in time series. An in-between step is the Short Time Fourier Transformation, which has a medium-sized resolution in both domains.
On the contrary, Wavelet Transform offers:
  • High resolution for small frequency values in the frequency domain and low resolution in the time domain;
  • Low resolution for large frequencies in the frequency domain and high resolution in the time domain.
The Wavelet Transform uses a function called Wavelets. The difference is that Fourier Transformation uses a sine wave that stretches in the interval between − and + infinity, Figure 7. A Wavelet is localized in time.
Because of this localization in time, it is possible to multiply the original signal with the Wavelet at different times. Starting from the beginning of the signal and moving towards the end. This process is called convolution. After the Wavelet transform of the one-dimensional signal, the output is in a two-dimensional form (time-scale) called a scaleogram. The dimension is called scale because the term frequency is reserved for the Fourier Transform. This is the reason behind the two axes of the scaleogram—time and scale. In comparison to Fourier Transform, where only one type of wave is used, there are many types of wavelets that are suited for different signals. Depending on the original signal, one can tailor a wavelet to the needs of the specific situation for best results. Only two mathematical conditions need to be fulfilled, meaning it is relatively easy to generate a completely new wavelet for a particular case. The two mathematical constraints are as follows:
  • Finite energy—Localization in time and frequency, integrative, and the inner product between Wavelet and signal always exists;
  • Zero mean in time-domain—ensuring it is integrative and the inverse of the Wavelet transform can be calculated.

5. Modeling Phase

In recent years, the collection of data for important variables over time has been a critical aspect in the progress of smart manufacturing systems, which are made possible through the invaluable contribution of sensors [20]. The experimental framework adopts DL algorithms to categorize Multivariate Time-Series data as either failure/unusual events or regular events. To address the issue of imbalanced data, data balancing techniques such as ensemble learning with undersampling and synthetic minority oversampling techniques were employed. Furthermore, along with DL algorithms like Convolutional Neural Network (CNN) and LSTM, ML algorithms like Support Vector Machine (SVM) and K-nearest neighbor (KNN) were also employed. The findings indicate that CNN is potentially the most effective algorithm for accurately classifying this dataset into two categories, and it outperforms both traditional approaches and other deep learning algorithms [20].

5.1. Traditional Machine Learning Algorithms

The first test, as a benchmark for future reference, is the traditional machine learning algorithms for regression and classification. The results will be used as baselines for comparison between shallow and deep learning. The tests will be separated into two categories—the first dataset and the second dataset. The data used for both cases are the output data from the TSFRESH library. The TSFRESH library extracts features from a given time series, such as the absolute energy of the time series and the highest value of the time series—maximum, mean change, median, and variance.
After the extraction, a correlation analysis is performed. The correlation analysis is performed in four steps, as seen below:
  • Feature extraction;
  • Calculation of Pearson’s correlation coefficient for each feature with the wear radius;
  • Sorting of correlation coefficients;
  • Using n strongest correlating features for tests;
The algorithms used for regression tests are Random Forest Regression (RFR), Multiple Linear Regression (MLR), and Decision Tree Regression (DTR).

5.2. Deep Learning Regression Model

The structure of the Deep Learning Regression Model is constant for all test cases, meaning it performs well for the whole range of features 1 − 500. The architecture is shown in Table 2.

5.3. Convolutional Neural Network with Wavelet Transform

The advantage of a Convolutional Neural Network (CNN) lies in the ability and efficiency to learn characteristic patterns of labels in images. This type of network can also analyze the two-dimensional CWT coefficients as pixels of an image.
In addition, all coefficient matrices are resized to a squared shape (64 × 64). This step is used because of the different lengths of the signal depending on the hub speed of the press, and CNN works with constant input shapes in Figure 8a,b. The chosen architecture is the LeNet-5 architecture in Figure 9.

6. Validation Phase

The coefficient of determination R2 is used in statistical models where prediction is the main purpose. It measures the predicted values based on the proportion of total variation of predictions. R2 score shows to what degree the variance of one variable explains the variance of the second, in comparison to the correlation coefficient, which indicates the strength of the relationship between independent and dependent variables. For example, an R2 score of 0.50 shows that half of the observed variation can be explained by the model’s inputs. The interval of this score is [0:1]. The general mathematical definition of the R2 score is as follows:
R 2 = 1 S S r e s S S t o t
  • S S t o t = i y i y ¯ 2 —Total sum of squares (proportional to the variance of the data), where y i is each individual score and y ¯ is the mean of all scores;
  • S S r e s = i y i f i 2 —Residual sum of squares, where y i is each individual score and f i is the predicted value.
Another criterion on which we measure the performance of a regression model is the Mean Squared Error (MSE). It measures the average of the squares of the error between the predicted and actual values. For predictions, MSE’s mathematical expression is as follows:
M S E = 1 n i = 1 n Y i Y ^ 2
  • Y i —Vector of actual values;
  • Y ^ —Vector of predicted values.

7. Results and Discussion

7.1. Traditional Machine Learning Algorithms—Regression

Shallow learning algorithms show promising results on the first dataset, with the R2 score being in the interval of [0,91:0,99] in Table 3. All algorithms tend to achieve better results with a higher number of features extracted, with the drawback of hardware limitations after 50 features. The same trend goes on for the mean squared error (MSE) where a constant decline is observed with the increase of features selected for the test.
The second dataset is proven to be difficult to learn for the classification test case in Table 4, but the regression algorithms are performing comparably to the ones on the first dataset. The R2 score and MSE are in marginally the same intervals as for the first dataset, meaning shallow learning algorithms perform well on both datasets.

7.2. Deep Learning Model—Regression

Deep learning models have the advantage of being able to analyze larger datasets in comparison to traditional machine learning algorithms. For this case, the regression model is able to handle data from all sensors without losing the ability to learn efficiently. As seen from Table 5 and Table 6, the R2 score of the regression model is between 0, 95 and 0, 99.

7.3. Convolutional Neural Network with Wavelet Transform

Convolutional neural networks (CNN) excel in the ability to extract features from images. This advantage over other neural networks is utilized in these tests. Table 7 shows the architecture used for both datasets.
This approach shows promising results in the field of classification. Figure 10 and Figure 11 show the training process. The data are split 70/30 between training and testing. The hyperparameters were chosen empirically.
This architecture, with a combination of hyperparameters, showed the best results despite requiring higher computation power. The ability to analyze big datasets and extract features efficiently enabled the CNN to achieve promising results. The accuracy of the trained model is 92.3% for the first dataset with six different wear conditions.
The second dataset is more complex than the first because of the number of different wear conditions, in this case, 15. The model shows promising results, achieving 62.4% accuracy across all classes.

8. Conclusions

Having the information on the wear state of the punch will be crucial for the manufacturer, with the approach not only being viable in this isolated case but for all manufacturing processes where mechanical wear occurs. Deep learning models will be benchmark tested against traditional machine learning algorithms in order for a conclusion to be drawn about the prospects of the application of deep learning models in the context of manufacturing and sensor signal processing. The drawback of deep learning and neural networks is the computational power needed for the training process, which, in some cases, compared to traditional machine learning algorithms, is exponentially larger. This balance between accuracy and complexity will be subject to testing during this research.
After all tests are conducted, the prospects of deep learning algorithms in both scenarios—regression and classification are promising. Sensor signals are complex to analyze, and based on the volume of data that need to be analyzed, a different machine-learning approach needs to be chosen. Due to hardware limitations, the number of features and length of data is limited for traditional shallow learning models. However, their advantage lies in their efficiency in terms of computational resource requirements. In cases with large and multidimensional datasets, deep learning algorithms, because of their neural network topologies and mathematical optimizations, excel in detecting patterns and correlations.
The tests show promising results in both regression and classification problems. Sensor signals are difficult to approach and analyze, and with the increase in volume, traditional machine learning algorithms are not achieving satisfactory results. Therefore, deep learning algorithms need to be used. As seen from the results, until a certain volume of data or, in this case, features, are reached, traditional machine learning algorithms excel at predicting and classifying, but once a certain threshold is exceeded, deep learning algorithms are to be preferred. Deep learning models need large datasets in order to “learn” and make better predictions/classifications. The last tests show promising results in terms of classification—92.3% accuracy for the first dataset and 62.4% for the second, despite the complexity of both. The confusion matrix of the second dataset shows that all inaccurate predictions are close to the value, meaning that in a real-world application, the “usable” accuracy of the model is far greater.
This application of deep learning in engineering processes, especially in manufacturing, can be further researched to optimize deep learning architectures. Pattern recognition via convolutional neural networks in the context of sensor signal classification is a field that can yield industry-changing results for the transition to Industry 4.0.

Author Contributions

Conceptualization, A.A.-P. and O.Y.; methodology, A.A.-P. and O.Y.; software, O.Y.; validation, O.Y.; formal analysis, O.Y.; investigation, A.A.-P. and O.Y.; resources, A.A.-P. and O.Y.; writing—original draft preparation, O.Y.; writing—review and editing, A.A.-P.; visualization, O.Y.; supervision, A.A.-P.; project administration, A.A.-P.; funding acquisition, A.A.-P. All authors have read and agreed to the published version of the manuscript.

Funding

The research work presented in the paper is funded by European Union-NextGenerationEU via the National Recovery and Resilience Plan of the Republic of Bulgaria under project BG-RRP-2.004-0005 “Improving the research capacity and quality to achieve international recognition and reSilience of TU-Sofia (IDEAS)”.

Data Availability Statement

The data are not publicly available due to restrictions.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wan, J.; Tang, S.; Li, D.; Wang, S.; Liu, C.; Abbas, H. A Manufacturing Big Data Solution for Active Preventive Maintenance. IEEE Trans. Ind. Inform. 2017, 13, 2039–2047. [Google Scholar] [CrossRef]
  2. Soualhi, A.; Lamraoui, M.; Elyousfi, B.; Razik, H. PHM SURVEY: Implementation of prognostic methods for monitoring industrial systems. Energies 2022, 15, 6909. [Google Scholar] [CrossRef]
  3. Antonini, M.; Pincheira, M.; Vecchio, M.; Antonelli, F. An Adaptable and Unsupervised TinyML Anomaly Detection System for Extreme Industrial Environments. Sensors 2023, 23, 2344. [Google Scholar] [CrossRef] [PubMed]
  4. Assafo, M.; Städter, J.P.; Meisel, T.; Langendörfer, P. On the Stability and Homogeneous Ensemble of Feature Selection for Predictive Maintenance: A Classification Application for Tool Condition Monitoring in Milling. Sensors 2023, 23, 4461. [Google Scholar] [CrossRef] [PubMed]
  5. Zhang, Q.; Yang, L.T.; Yan, Z.; Chen, Z.; Li, P. An Efficient Deep Learning Model to Predict Cloud Workload for Industry Informatics. IEEE Trans. Ind. Inform. 2018, 14, 3170–3178. [Google Scholar] [CrossRef]
  6. Zhang, Q.; Yang, L.T.; Chen, Z.; Li, P. A Tensor-Train Deep Computation Model for Industry Informatics Big Data Feature Learning. IEEE Trans. Ind. Inform. 2018, 14, 3197–3204. [Google Scholar] [CrossRef]
  7. Ota, K.; Dao, M.S.; Mezaris, V.; de Natale, F.G.B. Deep Learning for Mobile Multimedia. ACM Trans. Multimed. Comput. Commun. Appl. 2017, 13, 1–22. [Google Scholar] [CrossRef]
  8. Latif, A.; Mehedi, I.M.; Vellingiri, M.T.; Meem, R.J.; Palaniswamy, T. Enhanced Remora Optimization with Deep Learning Model for Intelligent PMSM Drives Temperature Prediction in Electric Vehicles. Axioms 2023, 12, 852. [Google Scholar] [CrossRef]
  9. Tran, M.-Q.; Doan, H.P.; Vu, V.Q.; Vu, L.T. Machine learning and IoT-based approach for tool condition monitoring: A review and future prospects. Measurment 2023, 207, 112351. [Google Scholar] [CrossRef]
  10. Failing, J.M.; Abellán-Nebot, J.V.; Benavent Nácher, S.; Rosado Castellano, P.; Romero Subirón, F. A Tool Condition Monitoring System Based on Low-Cost Sensors and an IoT Platform for Rapid Deployment. Processes 2023, 11, 668. [Google Scholar] [CrossRef]
  11. Hassan, M.; Ahmad, S.; Helmi, A. A Real-Time Deep Machine Learning Approach for Sudden Tool Failure Prediction and Prevention in Machining Processes. Sensors 2023, 23, 3894. [Google Scholar] [CrossRef]
  12. Wang, J.; Ma, Y.; Zhang, L.; Gao, R.X.; Wu, D. Deep learning for smart manufacturing: Methods and applications. J. Manuf. Syst. 2018, 48, 144–156. [Google Scholar] [CrossRef]
  13. Bhat, D.; Muench, S.; Roellig, M. Application of Machine Learning Algorithms in Prognostics and Health Monitoring of Electronic Systems: A Review. e-Prime—Adv. Electr. Eng. Electron. Energy 2023, 4, 100166. [Google Scholar] [CrossRef]
  14. Yan, J.; He, Z.; He, S. A deep learning framework for sensor-equipped machine health indicator construction and remaining useful life prediction. Comput. Ind. Eng. 2022, 172, 108559. [Google Scholar] [CrossRef]
  15. Dipietrangelo, F.; Nicassio, F.; Scarselli, G. Structural Health Monitoring for impact localization via machine learning. Mech. Syst. Signal Process. 2023, 183, 109621. [Google Scholar] [CrossRef]
  16. Tran, D.H.; Nguyen, V.L.; Nguyen, H.; Jang, Y.M. Self-supervised learning for time-series anomaly detection in Industrial Internet of Things. Electronics 2022, 11, 2146. [Google Scholar] [CrossRef]
  17. Tang, M.; Liang, L.; Zheng, Z.; Chen, J.; Chen, D. Anomaly Detection of Permanent Magnet Synchronous Motor Based on Improved DWT-CNN Multi-Current Fusion. Sensors 2024, 24, 2553. [Google Scholar] [CrossRef]
  18. Chirayil Nandakumar, S.; Mitchell, D.; Erden, M. Anomaly Detection Methods in Autonomous Robotic Missions. Sensors 2024, 24, 1330. [Google Scholar] [CrossRef] [PubMed]
  19. Kubik, C.; Molitor, D.A.; Becker, M.; Groche, P. Knowledge Discovery in Engineering Applications Using Machine Learning Techniques. J. Manuf. Sci. Eng. 2022, 144, 091003. [Google Scholar] [CrossRef]
  20. Rahman, M.M.; Farahani, M.A.; Wuest, T. Multivariate Time-Series Classification of Critical Events from Industrial Drying Hopper Operations: A Deep Learning Approach. J. Manuf. Mater. Process. 2023, 7, 164. [Google Scholar] [CrossRef]
Figure 1. Data-driven prediction model for analysis of sensor data with main phases: Data Acquisition, Data Preparation, Data Transformation, Modeling, and Validation.
Figure 1. Data-driven prediction model for analysis of sensor data with main phases: Data Acquisition, Data Preparation, Data Transformation, Modeling, and Validation.
Electronics 13 01799 g001
Figure 2. (a) Difference between no wear and critical wear. (b) Difference between force curves for all wear conditions from ri = 0 to ri = 0.6 mm., with 0.1 mm steps in-between.
Figure 2. (a) Difference between no wear and critical wear. (b) Difference between force curves for all wear conditions from ri = 0 to ri = 0.6 mm., with 0.1 mm steps in-between.
Electronics 13 01799 g002aElectronics 13 01799 g002b
Figure 3. PCA Visualization and PC1 line in the K-space with the best approximation according to the least squares rule.
Figure 3. PCA Visualization and PC1 line in the K-space with the best approximation according to the least squares rule.
Electronics 13 01799 g003
Figure 4. Force curve in (a) time-domain and (b) frequency-domain for the force signal used in the tests.
Figure 4. Force curve in (a) time-domain and (b) frequency-domain for the force signal used in the tests.
Electronics 13 01799 g004
Figure 5. Signals and frequency spectrum of a signal contain four frequencies at all times: four different frequencies at four different times.
Figure 5. Signals and frequency spectrum of a signal contain four frequencies at all times: four different frequencies at four different times.
Electronics 13 01799 g005
Figure 6. Overview of different transformations in comparison to the original signal in the time domain.
Figure 6. Overview of different transformations in comparison to the original signal in the time domain.
Electronics 13 01799 g006
Figure 7. Comparison between a sine-wave, cosine wave and Wavelet.
Figure 7. Comparison between a sine-wave, cosine wave and Wavelet.
Electronics 13 01799 g007
Figure 8. (a) Force curve in time domain; (b) scaleogram of force curve.
Figure 8. (a) Force curve in time domain; (b) scaleogram of force curve.
Electronics 13 01799 g008
Figure 9. CNN Architecture (LeNet-5).
Figure 9. CNN Architecture (LeNet-5).
Electronics 13 01799 g009
Figure 10. Confusion Matrix of CNN-Wavelet Model (first dataset).
Figure 10. Confusion Matrix of CNN-Wavelet Model (first dataset).
Electronics 13 01799 g010
Figure 11. Confusion Matrix of CNN-Wavelet Model (second dataset).
Figure 11. Confusion Matrix of CNN-Wavelet Model (second dataset).
Electronics 13 01799 g011
Table 1. Dataset information.
Table 1. Dataset information.
First DatasetSecond Dataset
Number of strokes2020
Stroke speed100–500 strokes per minute100–500 strokes per minute
Wear conditions0–0.6 mm (0.1 mm steps)0–0.75 mm (0.05 mm steps)
Table 2. Deep Learning regression model architecture.
Table 2. Deep Learning regression model architecture.
Name of LayerNumber of Neurons
Input layer16
Hidden Layer 132
Hidden Layer 216
Hidden Layer 332
Hidden Layer 416
Output layer1
Table 3. R2—Score of Regression Models (left) and MSE of Regression Models (right) (first dataset).
Table 3. R2—Score of Regression Models (left) and MSE of Regression Models (right) (first dataset).
Number of FeaturesRFRMLRDTRNumber of FeaturesRFRMLRDTR
10.946780.917690.9236210.039480.03940.0394
30.978390.9224530.970330.002090.003230.003
50.984210.946650.9724250.000850.003050.00117
100.995180.979160.99576100.000590.001310.00135
150.998390.993650.99576150.000070.000270.00008
200.998390.99410.99788200.000070.000240.0008
250.99860.994650.99788250.000070.000230.00017
500.998110.99750.99788500.000080.00010.00008
Table 4. R2—Score of Regression Models (left) and MSE of Regression Models (right) (second dataset).
Table 4. R2—Score of Regression Models (left) and MSE of Regression Models (right) (second dataset).
Number of FeaturesRFRMLRDTRNumber of FeaturesRFRMLRDTR
10.948630.91530.9336810.00260.004260.00333
30.948360.919210.9336830.00260.004060.00333
50.993380.936940.9917150.000330.003170.0042
100.996440.973410.99585100.00180.001340.00021
150.998340.978840.99834150.000080.001060.00008
200.998480.981610.99751200.000080.000920.00013
250.997450.981350.99309250.000130.000940.00035
500.998580.995370.99641500.000070.000230.00018
Table 5. MSE and R2—Score of Deep Learning Regression Model (first dataset).
Table 5. MSE and R2—Score of Deep Learning Regression Model (first dataset).
Number of FeaturesMSER2-Score
20.009640.74599
30.001790.95286
50.002230.74129
100.000620.98343
150.000440.98830
200.000690.98186
250.000430.99370
1000.000410.98917
2000.000410.98932
3000.000510.98663
5000.000510.98663
Table 6. MSE and R2—Score of Deep Learning Regression Model (second dataset).
Table 6. MSE and R2—Score of Deep Learning Regression Model (second dataset).
Number of FeaturesMSER2-Score
20.029180.40860
30.025650.48017
50.001620.96714
100.004510.90859
150.000890.98809
200.000740.98505
250.012500.74673
1000.011850.75983
2000.000200.99585
3000.000720.98536
5000.006620.86577
Table 7. Architecture of 2-D CNN Classification Model with Wavelet Transform.
Table 7. Architecture of 2-D CNN Classification Model with Wavelet Transform.
Layer TypeOutput ShapeNumber of Parameters
Convolutional layer(64, 64, 32)160
Max Pooling 2D layer(32, 32, 32)0
Convolutional layer(32, 32, 64)2112
Max Pooling 2D layer(16, 16, 64)0
Flatten16,3840
Fully-connected layer1282,097,290
Fully-connected layer546966
Fully-connected layer6330
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yotov, O.; Aleksieva-Petrova, A. Data-Driven Prediction Model for Analysis of Sensor Data. Electronics 2024, 13, 1799. https://doi.org/10.3390/electronics13101799

AMA Style

Yotov O, Aleksieva-Petrova A. Data-Driven Prediction Model for Analysis of Sensor Data. Electronics. 2024; 13(10):1799. https://doi.org/10.3390/electronics13101799

Chicago/Turabian Style

Yotov, Ognyan, and Adelina Aleksieva-Petrova. 2024. "Data-Driven Prediction Model for Analysis of Sensor Data" Electronics 13, no. 10: 1799. https://doi.org/10.3390/electronics13101799

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop