Next Article in Journal
An Autonomous Vehicle Behavior Decision Method Based on Deep Reinforcement Learning with Hybrid State Space and Driving Risk
Previous Article in Journal
Novel Surveillance View: A Novel Benchmark and View-Optimized Framework for Pedestrian Detection from UAV Perspectives
Previous Article in Special Issue
Geometry-Based Synchrosqueezing S-Transform with Shifted Instantaneous Frequency Estimator Applied to Gearbox Fault Diagnosis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Vibration-Based Anomaly Detection for Induction Motors Using Machine Learning

1
Department of Electrical Engineering, COMSATS University Islamabad Abbottabad Campus, Abbottabad 22060, Pakistan
2
Department of Defense Systems Engineering, Sejong University, Gwangjin-gu 05006, Republic of Korea
3
Marine Domain & Security Research Department, Korea Institute of Ocean Science & Technology, Busan 49111, Republic of Korea
4
Department of Electrical Engineering, Mirpur University of Science and Technology, Mirpur AJK 10250, Pakistan
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(3), 773; https://doi.org/10.3390/s25030773
Submission received: 21 November 2024 / Revised: 17 January 2025 / Accepted: 24 January 2025 / Published: 27 January 2025
(This article belongs to the Special Issue Fault Diagnosis and Vibration Signal Processing in Rotor Systems)

Abstract

:
Predictive maintenance of induction motors continues to be a significant challenge in ensuring industrial reliability and minimizing downtime. In this study, machine learning techniques are utilized to enhance fault diagnosis through the use of the Machinery Fault Database (MAFAULDA). A detailed extraction of statistical features was performed on multivariate time-series data to capture essential patterns that could indicate potential faults. Three machine learning algorithms—deep neural networks (DNNs), support vector machines (SVMs), and K-nearest neighbors (KNNs)—were applied to the dataset. Optimization strategies were carefully implemented along with oversampling techniques to improve model performance and handle imbalanced data. The results achieved through these models are highly promising. The SVM model demonstrated an accuracy of 95.4%, while KNN achieved an accuracy of 92.8%. Notably, the combination of deep neural networks with fast Fourier transform (FFT)-based autocorrelation features produced the highest performance, reaching an impressive accuracy of 99.7%. These results provide a novel approach to machine learning techniques in enhancing operational health and predictive maintenance of induction motor systems.

1. Introduction

Industrial systems across manufacturing, energy, and transportation sectors are powered by induction motors [1]. Operational efficiency is directly impacted by their reliability, where significant economic losses, production delays, and potential safety hazards can be caused by unexpected failures. While predictive maintenance has been established as a crucial strategy in mitigating these challenges, limitations have been found in traditional diagnostic methods that rely on manual inspection and periodic maintenance. Through recent advances in machine learning and signal processing, this landscape has been transformed and more accurate, proactive approaches to equipment monitoring have been enabled [2].
Catastrophic failures can occur when faults in induction motors, arising from complex interactions between mechanical and electrical components, are left undetected [3,4]. Vibration and acoustic emission signals have been identified as essential indicators for fault detection, through which statistical techniques and probabilistic models have been widely adopted over recent decades [5,6]. Intelligent solutions for predictive maintenance in rotating machinery have been implemented by manufacturers, particularly where remote monitoring applications are concerned [7]. Vibration-based condition monitoring has been established as a cornerstone, through which potential faults can be identified by tracking vibration pattern variations [8].
Fault detection strategies have been fundamentally reshaped by data-driven methodologies, where artificial intelligence and advanced signal processing are incorporated [9]. Among the various failure modes by which motor performance can be compromised, imbalance has been identified as a uniquely complex challenge for which sophisticated diagnostic approaches are required. Unlike simpler mechanical failures, the imbalance can be manifested through multiple configurations—mass imbalance, geometric imbalance, and couple imbalance—by which distinct detection challenges are presented and motor reliability and operational efficiency are significantly impacted.
Fault detection reliability is challenged by several obstacles: high-frequency signal noise, dynamic operating conditions, and complex signal propagation mechanisms. In variable industrial environments, where significant signal interference and operational variability are introduced, conventional approaches have been found particularly challenging [10]. By computational constraints, imbalance fault diagnosis is further complicated, especially in real-time processing scenarios where immediate results must be delivered without system performance being compromised. Advanced signal-processing techniques have been developed through recent research, by which precise fault indicators can be extracted under diverse operational conditions [11]. Traditional limitations have been overcome by combining fast Fourier transform (FFT), wavelet transforms, and machine learning algorithms.
Feature extraction has been identified as a critical challenge in imbalance fault analysis. Through hybrid methodologies developed by researchers, time-domain and frequency-domain transformations have been integrated to capture subtle mechanical irregularities [12]. Generalizable features have been emphasized, by which consistency can be maintained across different motor configurations and operating conditions. Detection precision has been enhanced through supervised learning algorithms, ensemble classification methods, and adaptive neural network architectures [13].
The imbalance fault detection field has been continuously advanced through machine learning applications. Particular promise in handling complex motor imbalance characteristics has been demonstrated by deep learning and ensemble methods [14]. By these approaches, fundamental limitations in existing diagnostic methods have been addressed and more robust, adaptable detection models have been developed through which generalization across varied motor configurations and operating conditions can be achieved.
The potential of artificial intelligence in motor fault diagnosis has been highlighted by recent studies [15,16,17]. In our research, deep neural networks (DNNs), support vector machines (SVMs), and K-nearest neighbors (KNNs) have been explored to address the binary classification between “normal” and “imbalance” conditions. In previous work, SVM was combined with long short-term memory (LSTM) networks [18], although a deep understanding of fault-specific patterns was required. Statistical features from vibration and current signals have been employed with SVM in other studies [19], while KNN has demonstrated effectiveness in diagnosing faults in rotating machinery [20,21].
In this work, a novel approach has been developed where traditional and advanced machine learning models are combined for imbalance fault detection. SVM with time-domain features has been employed for binary classification, by which its effectiveness with simpler feature sets can be leveraged, while DNN with FFT-based autocorrelation features has been applied to capture complex fault signatures in the frequency domain. The frequency domain analysis is performed using autocorrelation computed through FFT-based convolution, where the signal is convolved with its time-reversed version. This approach leverages FFT’s computational efficiency for calculating autocorrelation, enabling effective capture of periodic patterns and temporal dependencies in the vibration signals. Using the MAFAULDA dataset [22] and random oversampling techniques [23], the following significant improvements were achieved: SVM accuracy increased from 85.9% to 95.4%; KNN achieved 92.8% accuracy; and our DNN implementation with FFT-based features reached an accuracy of 99.7%. These results align with research where frequency-domain advantages in fault detection have been highlighted [24,25,26].
In this paper, existing research has been extended through the practical application of AI in predictive maintenance. The focus is on vibration-based datasets, preprocessing data/features, and specific fault conditions. Through the MAFAULDA dataset’s alignment with studied fault conditions, both relevance and accuracy in our findings are ensured. Figure 1 illustrates the flowchart of the proposed methods.
The remainder of this paper is organized as follows: In Section 2, the experimental setup, dataset characteristics, and methods are presented, including details of the machinery fault simulator and data acquisition system. Section 3 describes the feature extraction process and the statistical parameters used for fault detection. The feature signal analysis approach and its implementation are detailed in Section 4. Section 5 presents the classification methodology, including the implementation of SVM, KNN, and DNN algorithms along with a discussion on practical considerations and performance evaluations of these algorithms. The experimental results and detailed analysis of each classifier’s performance are presented in Section 6. Section 7 concludes the paper with key findings and suggestions for future research.

2. Materials and Methods

2.1. Experimental Dataset

The performances of the proposed prediction models were evaluated using the MAFAULDA dataset, which simulates undesirable failure scenarios in rotating machinery, including misalignment, unbalance, and bearing failures. This dataset was collected by the Signals, Multimedia, and Telecommunication Laboratory using the Machine Fault Simulator (MFS), also known as the SpectraQuest Alignment/Balance Vibration Trainer [27], as shown in Figure 2. Although an imbalanced dataset, the MAFAULDA database is specifically designed to assess the performance of diagnostic methods under various operating conditions and fault scenarios. Table 1 provides the default specifications of the MFS trainer [28], which was used to generate the data for this study.

2.2. Receiver Operating Characteristic (ROC) Analysis

The receiver operating characteristic (ROC) analysis is a graphical representation of classifier performances across different classification thresholds, enabling a comprehensive model assessment. In addition, the area under the curve (AUC) analysis is used to summarize overall classifier performance. It ranges from 0 to 1, with the following interpretations: 0.5 represents no discriminative capability, 0.5–0.7 represents poor discrimination, 0.7–0.8 represents good discrimination, 0.8–0.9 represents excellent discrimination and values greater than 0.9 indicate outstanding performance. The ROC illustrates the true positive rate (TPR) and false positive rate (FPR), which evaluate the model’s ability to correctly identify normal and faulty conditions.

2.3. Methodological Approach to Fault Classification

The comprehensive fault classification methodology employed three advanced machine learning algorithms: SVM, KNN, and DNN. The random classifier, which serves as a baseline predictor by assigning data points to classes with equal probability, is assumed to be a reference model for SVM and KNN classifiers. Each algorithm is systematically evaluated using a robust set of performance metrics such as accuracy, precision, recall, and F1 score to ensure a comprehensive analysis of fault detection capabilities.

2.3.1. Support Vector Machines (SVMs)

Support vector machines represent a powerful machine learning approach that is particularly effective for complex, high-dimensional datasets [29]. The SVM algorithm can handle both linear and non-linear classifications by optimizing the hyperplane separation between different classes. For normalizing the input features during preprocessing, the data are transformed to zero mean x ¯ and unit variance σ using Equation (1):
X scaled = X x ¯ σ
The SVM optimization process balances decision boundary smoothness with training point classification precision. The hyperparameters C and γ , detailed in Table 2, are carefully tuned to optimize model performance using Equation (2). The SVM implementation explored three key configurations: unoptimized linear SVM (Figure 3a), which represents the initial baseline performance; optimized SVM with RBF kernel (Figure 3b), which shows significant performance improvement; and oversampled optimized SVM (Figure 3c), which demonstrates the highest discriminative capability.
SVM ( X scaled , C , γ ) = arg min w , b , ξ C i = 1 n ξ i + 1 2 w 2 + i = 1 n β i ξ i + i = 1 n α i 1 y i ( w · x i + b ) + ξ i ,
where ξ i denotes a slack variable that measures the degree of misclassification, C denotes the regularization parameter that controls the trade-off between achieving a low error on the training data and minimizing the model complexity for better generalization, w is a weight vector of the hyperplane, b denotes a bias of the hyperplane, w represents the squared norm of w that controls the margin size (the distance between the separating a hyperplane and the support vectors), α i and β i are dual coefficients associated with the constraints that depend on the Lagrange multipliers in the optimization formulation, y i denotes the class label of the i th training sample, x i denotes an input feature of the i th training sample in the feature dataset X and n represents the total number of training samples.
Figure 3a shows that the area under the ROC is below the random classifier, which indicates that the classifier is not performing well. Figure 3b shows the ROC curve of optimized SVM, which improves the accuracy of both training and testing. Thus, the area under the ROC becomes AUC = 0.88 as compared to that of random classifier AUC = 0.50 , which illustrates excellent discrimination. Figure 3c shows the ROC curve of an oversampled optimized SVM, which achieves an overall accuracy of 95.4% on testing (Table 5, see Section 5). This also demonstrates the model’s high predictive capability. The ROC curve of the classifier illustrates that the model achieved an AUC of 0.98, which indicates outstanding discriminative ability between the normal and imbalanced classes.

2.3.2. K-Nearest Neighbors (KNNs)

The K-nearest neighbors algorithm represents an instance-based learning technique for classifying unknown data points [30,31]. The KNN algorithm determines class membership based on the majority vote of the k-nearest neighbors. We explored the following two primary distance metrics [32]: Manhattan distance (L1 norm), given by p 1 ( x 1 , x 2 ) = i = 1 n | x i y i | , and Euclidean distance (L2 norm), given by p 2 ( x 1 , x 2 ) = i = 1 n ( x i y i ) 2 . The hyperparameters were systematically explored using grid search, as outlined in Table 3. The KNN approach was evaluated using multiple configurations, namely, unoptimized KNN (Figure 4a), which served as the initial performance assessment; optimized KNN (Figure 4b), which showed improved classification accuracy; and oversampled optimized KNN (Figure 4c), which provided enhanced discriminative power.
The ROC curve of unoptimized KNN is shown in Figure 4a, indicating a slightly low AUC of 0.78 , which shows that the model is not good enough to predict good results. The ROC of optimized KNN shows AUC = 0.86 , which is a good discrimination result, and better than the unoptimized KNN, as depicted in Figure 4b. The optimized KNN achieved an accuracy of 89.8% on the test data and approximately 95% on the training data in the confusion matrix, as depicted in Figure 15b and discussed in Section 6.
The over-sampled optimized KNN has AUC = 0.96 as shown in Figure 4c, which indicates the outstanding performance of the model, and its overall accuracy is 92.8% on the testing data. This ROC-AUC score level implies that the model has a high level of accuracy in differentiating the normal and imbalanced (faulty) classes.

2.3.3. Deep Neural Networks (DNNs)

We used the dense neural network algorithm [33] to develop the DNN architecture, as shown in Figure 5, for a novel fault classification. The network was fed with autocorrelation features computed through FFT-based convolution. The FFT-based autocorrelation choice for feature extraction offers several key advantages in fault detection. Through autocorrelation analysis, temporal patterns in vibration signals can be effectively captured even when these patterns are obscured by noise in the time domain. This approach is particularly powerful for machinery fault detection, as recurring patterns in the signal are emphasized, making fault signatures more distinguishable. By correlating the signal with itself at different time lags, this method helps isolate fault-specific patterns from the complex vibration signatures typical of induction motors, where multiple mechanical components operate simultaneously. The FFT-based implementation of autocorrelation provides computational efficiency while maintaining sensitivity to periodic patterns. Moreover, this approach demonstrates robustness to phase shifts and timing variations in the signal, a crucial advantage in real-world operating conditions where exact synchronization may not be guaranteed. These characteristics complement DNN’s deep feature learning capabilities particularly well, enabling the network to identify and learn from the most relevant temporal patterns for accurate fault classification.
The input layer of the DNN has twenty-one nodes, matching the number of dimensions in the input data pertaining to faults. The hidden layers have five subsequent layers consisting of 32, 64, 128, 64, and 32 neurons, respectively. Each layer uses the rectified linear unit (ReLU) activation function as depicted in Figure 5. The output layer consists of two neurons corresponding to the normal and fault classes. In the output layer, we used the softmax activation function, which offers probability distributions for precise fault classification. Adam is an optimization technique that may be used to update dense neural network weights iteratively based on training data, replacing the traditional stochastic gradient descent approach [33]. We used cross-entropy as the loss function to evaluate the accuracy measure. We stopped training if the loss metric did not improve after two consecutive epochs, preventing an overfitting curve. Thus, the proposed DNN is specifically designed to improve fault classification skills and leverages the model principles to automatically identify complex patterns from the input data.

3. Feature Extraction

This section presents our approach to statistical feature extraction for fault diagnosis. We employed eleven statistical features that effectively characterized the distribution and patterns in vibration data. The extracted features included mean, standard deviation, quartile medians (Q1, Q2, and Q3), minimum and maximum (peak-to-peak), kurtosis, skewness, root mean square (RMS), and energy. For parameter optimization, we utilized kurtosis as an advanced input feature, measuring signal sharpness to produce optimized features. The detailed optimization process using kurtosis is discussed in [34]. These optimization techniques enable comprehensive assessment of the data structure, particularly in identifying peaked, flat, or directionally biased distributions. Skewness quantifies distribution symmetry, with zero indicating normal or symmetric distributions, while kurtosis measures the relative heaviness of distribution tails compared to normal distributions. Both metrics provide crucial statistical characteristics for fault detection. The mean and standard deviation describe the central tendency and data spread, respectively, with variance and standard deviation being particularly useful for measuring data distribution during feature extraction [35]. Our methodology applies these statistical features to differentiate between expected (normal) and unusual (imbalance) patterns in time-domain data. The dataset comprises 2550 feature windows, with 70% allocated for training and 30% for testing. Each window spans 132 samples and is advanced with 25% overlap, generating multiple statistical features that characterize the vibration patterns under diverse operating conditions. The mathematical formulations for all statistical features are presented in Table 4.

4. Feature Signal Analysis for Fault Detection

The MAFAULDA dataset contains vibration data collected through industrial IMI sensors (601A01 and 604B31 accelerometers), which were positioned on the MFS to capture vibrations in radial, axial, and tangential directions. The data acquisition system included a Monarch Instrument MT-190 tachometer, Shure SM81 microphone, and National Instruments NI-9234 data acquisition modules operating at a 51.2 kHz sampling rate. The analysis extracted eleven fundamental features from the raw vibration data: mean, standard deviation, minimum, maximum, kurtosis, skewness, root mean square (RMS), energy, and median values (25%, 50%, 75%). The SVM and KNN methods applied these features to both raw and processed data, allowing the algorithms to learn and adapt to underlying patterns for effective classification. The DNN implementation uniquely incorporated autocorrelation analysis of accelerometer data, computed through FFT-based convolution, which captured temporal patterns and recurring signal characteristics crucial for fault detection.
The dataset comprises vibration sequences at fixed rotation speeds from 254 to 3686 rpm, with approximately 60 rpm increments. These sequences were sampled at 50 kHz and analyzed using a sliding window approach. Windows of 132 samples were advanced with 25% overlap between consecutive windows, corresponding to a step size of 99 samples, resulting in 2550 feature windows spanning approximately five seconds of data. The 60 rpm increments were carefully chosen to provide comprehensive coverage of operational speeds while maintaining manageable data volume. The imbalance fault simulations were conducted using loads ranging from 6 g to 35 g. Under normal operation, the rotational frequency was consistently maintained for each load value below 30 g. These operating conditions maintained a constant speed within each sequence, allowing for reliable fault signature identification. Figure 6 shows the analysis of extracted features under normal rotational speeds. Loads equal to or exceeding 30 g limited the system’s ability to achieve rotational frequencies above 3300 rpm, revealing important characteristics for fault detection. The progressive load conditions (6 g to 35 g) reveal important system dynamics that are crucial for practical fault detection implementations.
Different imbalance conditions show distinct behavioral patterns. At the 6 g imbalance (Figure 7), the mean, standard deviation, and kurtosis properties remained stable despite slight perturbations. The 10 g imbalance analysis (Figure 8) shows sustained rotational properties. Skewness and RMS features demonstrated characteristic changes in the system’s vibrational behavior under an increased imbalance load.
At the 15 g imbalance (Figure 9), variations in maximum, energy, and median parameters show how the system responded to the increased load. The 20 g load imbalance (Figure 10) created significant dynamical variations across all features, especially in skewness and kurtosis. The 25 g imbalance analysis (Figure 11) reveals the system’s response to substantial imbalances through standard deviation and minimum trends.
The analysis is expanded to a 15 g load imbalance as shown in Figure 9. Changes in important parameters, such as the maximum, energy, and median, provide a detailed analysis of how the system reacts to more extreme imbalance situations. The maximum value directly indicates peak vibration amplitudes, which typically increase under severe imbalance conditions. This parameter is crucial as it reveals potential threshold violations that could indicate severe mechanical stress on the system. The energy value, calculated as the sum of squared signal amplitudes provides a comprehensive measure of the overall vibrational intensity. Under extreme imbalance conditions, energy values show significant increases reflecting the higher vibrational content across the entire signal duration. This makes energy a sensitive indicator of severe imbalance conditions.
Figure 10 shows the effects of a 20 g load imbalance, resulting in dynamical variations in all features. Skewness and kurtosis are two characteristics that might show significant fluctuations, suggesting that the system is sensitive to increasing imbalance levels. In Figure 11, variations in all extracted features are represented under a 25 g imbalance. Similarly, trends in standard deviation and minimum provide insight into the system’s capability to manage important imbalance issues.
Figure 12 shows limitations to the feature performance by analyzing the effects of higher load imbalance. The 30 g load analysis reveals constrained rotational frequencies, limiting frequencies above 3300 rpm, and variations in maximum and energy features, indicating system stability issues when subjected to high loads. The dynamics of all extracted features under a severe load of a 35 g imbalance are examined in the last exploration as shown in Figure 13. At a 35 g load imbalance, significant changes appear across all features, particularly in mean, skewness, and RMS values, showing system behavior at operational limits. Each sequence spans five seconds at a 50 kHz sampling rate, providing detailed data for fault detection. In Figure 6, Figure 7, Figure 8, Figure 9, Figure 10, Figure 11, Figure 12 and Figure 13, the time is taken in microseconds.

5. Practical Consideration of SVM, KNN, and DNN Algorithms

5.1. Implementation Framework

Our implementation leverages Python’s robust machine learning libraries. For classification tasks, we utilized scikit-learn’s support vector classifier (SVC) for the SVM implementation and KNeighborsClassifier for KNN. The framework incorporates GridSearchCV for hyperparameter optimization through a systematic grid search, enhancing both SVM and KNN algorithms. To address dataset imbalance, we employed RandomOverSampler from the imbalanced-learn library. The implementation also relies on essential Python libraries: matplotlib.pyplot for visualization, pandas for data manipulation, NumPy for numerical operations, and SciPy for statistical measures like kurtosis and skewness. Feature standardization was achieved using scikit-learn’s StandardScaler, ensuring zero mean and unit variance for input features.

5.2. Performance Metrics

Our evaluation framework employs standard metrics to assess model performance. These metrics provide detailed insights into how well the models handle true positives (TPs), true negatives (TNs), false positives (FPs), and false negatives (FNs). The ROC curves, shown in Figure 3 and Figure 4, illustrate the trade-off between sensitivity and specificity across classification thresholds. The key metrics include the following: Accuracy, which involves measuring overall classification performance as expressed in Equation (3),
Accuracy = T P + T N T P + F N + T N + F P ,
Precision is calculated by Equation (4), which is crucial when false positives are costly,
Precision = T P T P + F P ,
Recall is obtained by Equation (5), which is essential when false negatives must be minimized,
Recall = T P T P + F N ,
and F1 score, which provides a balanced measure of precision and recall is computed by using the following Equation:
F 1 score = 2 × ( Precision × Recall ) Precision + Recall

5.3. Comparative Analysis

Our experimental results demonstrate the effectiveness of optimization and oversampling strategies across all implemented algorithms. As shown in Table 5, the SVM implementation showed consistent improvement through optimization, with the baseline model achieving 87% in training and 85.9% in testing accuracy. The optimized version improved to 93% in training and 90.4% in testing accuracy, while the addition of oversampling further enhanced performance to 97.5% in training and 95.4% in testing accuracy. The KNN implementation followed a similar trajectory of improvement. The baseline model achieved 94% in training and 87.4% in testing accuracy; with optimization, it maintained 95% training accuracy and improved testing accuracy to 89.8%. The oversampled version demonstrated the best performance, reaching 95.7% in training and 92.8% in testing accuracy.
Most notably, our DNN implementation with autocorrelation features, computed using FFT-based convolution, showed exceptional performance. While the direct time-domain analysis achieved respectable results (97% in training, 95% in testing accuracy), the FFT-based autocorrelation features approach significantly improved performance to 99.8% in training and 99.7% in testing accuracy. These results underscore the importance of both optimization strategies and appropriate data preprocessing in fault diagnostics. The consistent improvement across all algorithms, particularly with oversampling and FFT-based autocorrelation feature analysis, suggests robust potential for real-world applications in predictive maintenance systems.

6. Illustration of Experimental Results

6.1. SVM

Our experimental validation employed the MAFAULDA dataset [22] to thoroughly evaluate model performance. The classification results are best understood through confusion matrices, presented in Figure 14, which provide a comprehensive view of how well our models distinguish between normal and fault conditions. The detailed performance metrics derived from Figure 14a–c are summarized in Table 6. Among our implementations, the oversampled optimized SVM demonstrated remarkable performance, correctly identifying 344 faulty cases and 385 normal cases, while minimizing misclassifications to just 33 false positives and 2 false negatives. This translated to an impressive 95.4% accuracy, with the model achieving 91.2% precision and 99.4% recall. While the optimized SVM also showed strong performance with 97.2% precision, the unoptimized linear SVM, despite perfect precision (1.00), fell short in overall effectiveness with an accuracy of 85.8%, correctly identifying 656 positive cases but also recording 108 false negatives.

6.2. KNN

Our KNN implementation underwent similar rigorous testing, with results visualized in Figure 15. The oversampled optimized variant proved most effective, successfully identifying 325 faulty and 384 normal cases, although it produced 52 false positives and 3 false negatives. Table 7 presents the complete performance metrics, calculated using Equations (3)–(6). The oversampled approach achieved 92.8% accuracy with an exceptional 99.1% recall, while the base-optimized KNN showed strong discrimination capabilities with 637 correct positive identifications. The unoptimized version, while achieving a respectable 87.4% accuracy, demonstrated room for improvement through our optimization strategies.

6.3. DNN

In developing DNN architecture using the dense NN method, we leveraged FFT-based autocorrelation features to enhance the model’s ability to recognize fault patterns. This approach processes raw vibration data through FFT-based autocorrelation analysis, where fault signatures become more distinguishable through the enhancement of temporal patterns and periodic relationships in the signal. As demonstrated in Figure 16, this transformation significantly improved classification performance. By effectively filtering noise and focusing on relevant frequency components, our DNN achieved an outstanding 99.7% accuracy, compared to 95% in time-domain analysis as shown in Table 5.
Based on the FFT-based autocorrelation fed data, DNN successfully identified 655 faulty and 107 normal cases, producing 01 false positive and 01 false negative. The significance of the performance becomes clear when considering that the dataset was not balanced through oversampling. This approach achieved a remarkable accuracy of 99.7% with an exceptional 99.9% precision and 99.8% recall. The confusion matrix is depicted in Figure 16.

7. Conclusions

This paper introduces effective methods for predicting induction motor faults, aiming to minimize losses and prevent disasters through advanced fault detection techniques. By leveraging SVM, KNN, and DNN, we precisely classify motor states as either “normal” or “imbalance”. Our findings demonstrate significant improvements over previous studies, notably with accuracy for optimized oversampled SVM reaching 95.4% and optimized oversampled KNN achieving an accuracy of 92.8%. These methods, enhanced by optimization and oversampling strategies, contribute significantly to fault detection in induction motors. The best-performing algorithm was DNN with FFT-based implementation of autocorrelation, achieving an impressive 99.7% accuracy. This aligns with trends in technology-driven fault identification and offers substantial benefits for industrial processes such as increased efficiency, reduced downtime, and enhanced safety
In future work, the model’s robustness will be enhanced by testing it on the testbed under varying load conditions and real-world noise. The training dataset will also be expanded with diverse fault scenarios, and adaptive learning techniques will be incorporated to improve the generalization and reliability of the model.

Author Contributions

N.K.: programming and data curation; I.U.: methodology and writing—original draft; J.S. and S.M.: resources, investigation, and validation of the research; S.A.M. and W.-G.K.: visualization, validation, data curation, writing—review and editing draft, project administration, and funding. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea (NRF) funded by the Korean Government Ministry of Science and ICT (MSIT) under Grant RS-2022-00166977.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Available on the public repository website: See reference [24].

Acknowledgments

This work was conducted in the Department of Electrical and Computer Engineering, COMSATS University Islamabad, Abbottabad Campus, and the Department of Defense Systems Engineering at Sejong University, Seoul, Republic Korea.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Penrose, A. Industrial Electromechanical Systems: A Comprehensive Overview. Int. J. Ind. Eng. 2022, 45, 112–125. [Google Scholar]
  2. Sharma, M.; Kumar, R. Machine Learning in Predictive Maintenance: A Systematic Review. J. Predict. Maint. 2021, 18, 45–67. [Google Scholar]
  3. Mian, T.; Anurag, C.; Fatima, S.; Panigrahi, B.K. Artificial intelligence of things based approach for anomaly detection in rotating machines. Comput. Electr. Eng. 2023, 109, 108760. [Google Scholar] [CrossRef]
  4. Ameer, H.S.; Ungku, A.B.-U.A. A review on fault detection and diagnosis of industrial robots and multi-axis machines. Results Eng. 2024, 23, 102397. [Google Scholar]
  5. Lin, S.L. Intelligent Fault Diagnosis and Forecast of Time-Varying Bearing Based on Deep Learning VMD-DenseNet. Sensors 2021, 21, 7467. [Google Scholar] [CrossRef]
  6. He, J.; Wu, P.; Tong, Y.; Zhang, X.; Lei, M.; Gao, J. Bearing Fault Diagnosis via Improved One-Dimensional Multi-Scale Dilated CNN. Sensors 2021, 21, 7319. [Google Scholar] [CrossRef]
  7. Chi, Y.; Dong, Y.; Wang, Z.J.; Yu, F.R.; Leung, V.C.M. Knowledge-Based Fault Diagnosis in Industrial Internet of Things: A Survey. IEEE Internet Things J. 2022, 9, 12886–12900. [Google Scholar] [CrossRef]
  8. Almutairi, K.M.; Sinha, J.K. Experimental Vibration Data in Fault Diagnosis: A Machine Learning Approach to Robust Classification of Rotor and Bearing Defects in Rotating Machines. Machines 2023, 11, 943. [Google Scholar] [CrossRef]
  9. Sharma, D.K.; Brahmachari, S.; Singhal, K.; Gupta, D. Data driven predictive maintenance applications for industrial systems with temporal convolutional networks. Comput. Ind. Eng. 2022, 169, 108213. [Google Scholar] [CrossRef]
  10. Gao, Z.; Ding, S.X.; Cecati, C. Real-time fault diagnosis and fault-tolerant control. IEEE Trans. Industrial Electron. 2015, 62, 3752–3756. [Google Scholar] [CrossRef]
  11. Jose, J.P.; Ananthan, T.; Prakash, N.K. Ensemble Learning Methods for Machine Fault Diagnosis. In Proceedings of the 2022 Third International Conference on Intelligent Computing Instrumentation and Control Technologies, Kannur, India, 11–12 August 2022; pp. 1127–1134. [Google Scholar]
  12. Hou, J.; Lu, X.; Zhong, Y.; He, W.; Zhao, D.; Zhou, F. A comprehensive review of mechanical fault diagnosis methods based on convolutional neural network. J. Vibroeng. 2023, 26, 44–65. [Google Scholar] [CrossRef]
  13. Wen, Y.; Rahman, M.F.; Xu, H.; Tseng, T.-L.B. Recent Advances and Trends of Predictive Maintenance from Data-Driven Machine Prognostics Perspective. Measurement 2022, 187, 110276. [Google Scholar] [CrossRef]
  14. Chen, L.; Li, X. IoT and Edge Computing in Predictive Maintenance. IEEE Internet Things J. 2022, 9, 13456–13470. [Google Scholar]
  15. Zamudio-Ramírez, I.; Osornio-Ríos, R.A.; Antonino-Daviu, J.A.; Quijano-Lopez, A. Smart-sensor for the automatic detection of electromechanical faults in induction motors based on the transient stray flux analysis. Sensors 2020, 20, 1477. [Google Scholar] [CrossRef]
  16. Misra, S.; Kumar, S.; Sayyad, S.; Bongale, A.; Jadhav, P.; Kotecha, K.; Abraham, A.; Gabralla, L.A. Fault detection in induction motor using time domain and spectral imaging-based transfer learning approach on vibration data. Sensors 2022, 22, 8210. [Google Scholar] [CrossRef]
  17. Lu, H.; Li, Y.; Mu, S.; Wang, D.; Kim, H.; Serikawa, S. Motor Anomaly Detection for Unmanned Aerial Vehicles Using Reinforcement Learning. IEEE Internet Things J. 2018, 5, 2315–2322. [Google Scholar] [CrossRef]
  18. Vos, K.; Peng, Z.; Jenkins, C.; Shahriar, M.R.; Borghesani, P.; Wang, W. Vibration-based anomaly detection using LSTM/SVM approaches. Mech. Syst. Signal Process. 2022, 169, 108752. [Google Scholar] [CrossRef]
  19. Purushottam, G.; Rajiv, T. A support vector machine based fault diagnostics of Induction motors for practical situation of multi-sensor limited data case. Measurement 2019, 135, 694–711. [Google Scholar]
  20. Aguo, L.; Ming, J.Z. Gear crack level identification based on weighted K nearest neighbor classification algorithm. Mech. Syst. Signal Process. 2009, 23, 1535–1547. [Google Scholar]
  21. Lu, J.; Qian, W.; Li, S.; Cui, R. Enhanced K-Nearest Neighbor for Intelligent Fault Diagnosis of Rotating Machinery. Appl. Sci. 2021, 11, 919. [Google Scholar] [CrossRef]
  22. Cu, V.X. “MAFAULDA Full”, Kaggle. 25 May 2022. Available online: https://www.kaggle.com/datasets/vuxuancu/mafaulda-full/data (accessed on 25 October 2023).
  23. Alzghoul, A.; Jarndal, A.; Alsyouf, I.; Bingamil, A.A.; Ali, M.A.; AlBaiti, S. On the Usefulness of Pre-processing Methods in Rotating Machines Faults Classification using Artificial Neural Network. J. Appl. Comput. Mech. 2021, 7, 254–261. [Google Scholar]
  24. Wang, Z.; Li, J.; Xu, Z.; Yang, S.; He, D.; Chan, S. Application of Deep Neural Network with Frequency Domain Filtering in the Field of Intrusion Detection. Int. J. Intell. Syst. 2023, 2023, 8825587. [Google Scholar] [CrossRef]
  25. Mayaki, M.Z.A.; Riveill, M. Machinery Anomaly Detection Using Artificial Neural Networks and Signature Feature Extraction. In Proceedings of the 2023 International Joint Conference on Neural Networks (IJCNN 2023), Gold Coast, Australia, 18–23 June 2023; pp. 1–10. [Google Scholar]
  26. Perales Gomez, A.L.; Fernandez Maimo, L.; Huertas Celdran, A.; Garcia Clemente, F.J. SUSAN: A Deep Learning based anomaly detection framework for sustainable industry. Sustain. Comput. Informatics Syst. 2023, 37, 100842. [Google Scholar] [CrossRef]
  27. SpectraQuest, Inc. Available online: http://www.spectraquest.com (accessed on 5 October 2023).
  28. Nayak, C.; Pathak, V.K.; Kumar, S.; Athnekar, P. Design and development of machine fault simulator (MFS) for fault diagnosis. Int. J. Recent Adv. Mech. Eng. (IJMECH) 2015, 4, 77–84. [Google Scholar] [CrossRef]
  29. Kumar, D.; Sanaullah, M.U.; Kapal, D.; Sunder, A.K.; Naveed, A.B.; Tanweer, H. Towards soft real-time fault diagnosis for edge devices in industrial IoT using deep domain adaptation training strategy. J. Parallel Distrib. Comput. 2022, 160, 90–99. [Google Scholar] [CrossRef]
  30. Anava, O.; Levy, K.Y. k*-Nearest Neighbors: From Global to Local. In Proceedings of the 30th Conference on Neural Information Processing Systems (NIPS) 2016, Barcelona, Spain, 5–10 December 2016. [Google Scholar]
  31. Achraf, E.B.; Noura, J.; Omar, M.; Abdelkader, H. Adaptive K values and training subsets selection for optimal K-NN performance on FPGA. J. King Saud Univ.-Comput. Inf. Sci. 2024, 36, 102081. [Google Scholar]
  32. García-Pedrajas, N.; Castillo, J.D.; Cerruela-García, G. A proposal for local k values for k-nearest neighbor rule. IEEE Trans. Neural Netw. Learn. Syst. 2015, 28, 470–475. [Google Scholar] [CrossRef]
  33. Sandeep, K.R.; Rajesh, K.P. Accelerated Singular Value Decomposition (ASVD) using momentum based Gradient Descent Optimization. J. King Saud Univ. Comput. Inf. Sci. 2021, 33, 447–452. [Google Scholar]
  34. Pestana-Viana, D.; Zambrano-López, R.; De Lima, A.A.; De Prego, M.T.; Netto, S.L.; Da Silva, E.A.B. The Influence of Feature Vector on the Classification of Mechanical Faults Using Neural Networks. In Proceedings of the 7th IEEE Latin American Symposium on Circuits and Systems (LASCAS) 2016, Florianopolis, Brazil, 28 February–2 March 2016. [Google Scholar]
  35. Altaf, M.; Akram, T.; Khan, M.A.; Iqbal, M.; Ch, M.M.I.; Hsu, C.-H. A new statistical features based approach for bearing fault diagnosis using vibration signals. Sensors 2022, 22, 2012. [Google Scholar] [CrossRef]
Figure 1. The flowchart of the proposed work: red-dashed box implies the implementation of optimization and oversampling using SVM and KNN; blue-dashed box implies the FFT-based feature extraction applied on DNN.
Figure 1. The flowchart of the proposed work: red-dashed box implies the implementation of optimization and oversampling using SVM and KNN; blue-dashed box implies the FFT-based feature extraction applied on DNN.
Sensors 25 00773 g001
Figure 2. The SpectraQuest machine fault simulator.
Figure 2. The SpectraQuest machine fault simulator.
Sensors 25 00773 g002
Figure 3. Receiver operating characteristic (ROC) curves of SVM. (a) Linear SVM ROC. (b) Optimized SVM ROC. (c) Optimized SVM using oversampling ROC.
Figure 3. Receiver operating characteristic (ROC) curves of SVM. (a) Linear SVM ROC. (b) Optimized SVM ROC. (c) Optimized SVM using oversampling ROC.
Sensors 25 00773 g003aSensors 25 00773 g003b
Figure 4. Receiver operating characteristic (ROC) curves of KNN. (a) Unoptimized KNN ROC, (b) Optimized KNN ROC, (c) Oversampled Optimized KNN ROC.
Figure 4. Receiver operating characteristic (ROC) curves of KNN. (a) Unoptimized KNN ROC, (b) Optimized KNN ROC, (c) Oversampled Optimized KNN ROC.
Sensors 25 00773 g004
Figure 5. Deep neural network architecture.
Figure 5. Deep neural network architecture.
Sensors 25 00773 g005
Figure 6. Analysis of 11 features in normal rotational sequences (737 rpm to 3686 rpm).
Figure 6. Analysis of 11 features in normal rotational sequences (737 rpm to 3686 rpm).
Sensors 25 00773 g006
Figure 7. Feature analysis under 6 g imbalance: consistently maintained rotations.
Figure 7. Feature analysis under 6 g imbalance: consistently maintained rotations.
Sensors 25 00773 g007
Figure 8. Feature analysis under the 10 g imbalance: stable rotational characteristics.
Figure 8. Feature analysis under the 10 g imbalance: stable rotational characteristics.
Sensors 25 00773 g008
Figure 9. Feature analysis under the 15 g imbalance: assessing rotational stability.
Figure 9. Feature analysis under the 15 g imbalance: assessing rotational stability.
Sensors 25 00773 g009
Figure 10. Feature analysis under the 20 g imbalance: dynamical variations.
Figure 10. Feature analysis under the 20 g imbalance: dynamical variations.
Sensors 25 00773 g010
Figure 11. Feature analysis under the 25 g imbalance: investigating system resilience.
Figure 11. Feature analysis under the 25 g imbalance: investigating system resilience.
Sensors 25 00773 g011
Figure 12. Feature analysis under a 30 g imbalance: limitations on rotational frequencies.
Figure 12. Feature analysis under a 30 g imbalance: limitations on rotational frequencies.
Sensors 25 00773 g012
Figure 13. Feature analysis under a 35 g imbalance.
Figure 13. Feature analysis under a 35 g imbalance.
Sensors 25 00773 g013
Figure 14. Faulty and normal classification using SVM.
Figure 14. Faulty and normal classification using SVM.
Sensors 25 00773 g014aSensors 25 00773 g014b
Figure 15. Faulty and normal classification using KNN.
Figure 15. Faulty and normal classification using KNN.
Sensors 25 00773 g015aSensors 25 00773 g015b
Figure 16. Faulty and normal classification using DNN.
Figure 16. Faulty and normal classification using DNN.
Sensors 25 00773 g016
Table 1. Machinery fault simulator (MFS) specifications.
Table 1. Machinery fault simulator (MFS) specifications.
ParameterValue
Motor & 1/4 CV DC
Frequency Range700–3600 rpm
System Weight22 kg
Axis Diameter16 mm
Rotor15.24 cm
Bearings Distance390 mm
Data Acquisition System
AccelerometersThree IMI 601A01 model
One IMI 604B31 model
TachometerMonarch MT-190
Microphone sensorShure SM81
Data Acquisition ModulesTwo NI 9234
Sample Rate51.2 kHz
Sampling ParameterValue
Sequence Duration5 s
Sampling Rate50 kHz
Samples per Sequence250,000
Sequence Types
Normal SequencesRotation speed: 737 to 3686 rpm
Imbalance FaultsLoad values: 6 g to 35 g
Table 2. Hyperparameter grid for optimized SVM.
Table 2. Hyperparameter grid for optimized SVM.
HyperparameterValues
C0.1, 1, 10, 100
γ 0.01, 0.1, 1, 10
Table 3. Hyperparameters used in optimized KNN.
Table 3. Hyperparameters used in optimized KNN.
HyperparameterValues
n _ n e i g h b o r s 3, 5, 7, 9
p1 (Manhattan), 2 (Euclidean)
Table 4. Statistical formulas of features.
Table 4. Statistical formulas of features.
FeatureFormula
  Mean x ¯ = 1 N i = 1 N ( x i )
  Standard Deviation σ = 1 N ( i = 1 N ( x i x ¯ ) 2 )
  Minimum min ( x i )
  First Quartile (Q1) ( Q 1 = 1 4 ( N + 1 ) th )
  Median (Q2) ( Q 2 = 1 2 ( N + 1 ) th )
  Third Quartile (Q3) ( Q 3 = 3 4 ( N + 1 ) th )
  Maximum max ( x i )
  Kurtosis i = 1 N ( x i x ¯ ) 4 ( N 1 ) σ 4
  Skewness i = 1 N ( x i x ¯ ) 3 ( N 1 ) σ 3
  Root Mean Square (RMS) 1 N ( i = 1 N ( x i ) 2 )
  Energy i = 1 N | x i | 2
Table 5. Comparison of algorithms.
Table 5. Comparison of algorithms.
AlgorithmTrainingTesting
Accuracy Accuracy
Unoptimized SVM87%85.9%
Optimized SVM93%90.4%
Oversampled optimized SVM97.5%95.4%
Unoptimized KNN94%87.4%
Optimized KNN95%89.8%
Oversampled optimized KNN95.7%92.8%
Time-domain based DNN97%95%
FFT based DNN99.8%99.7%
Table 6. Calculation of accuracy, precision, recall, and F1 score of the model using SVM.
Table 6. Calculation of accuracy, precision, recall, and F1 score of the model using SVM.
AlgorithmTPFPFNTNAccuracyPrecisionRecallF1 Score
Linear SVM656010800.861.000.860.92
Optimized SVM6342251570.900.970.930.95
Oversampled optimized SVM3443323850.950.910.990.95
Table 7. Calculation of accuracy, precision, recall, and F1 score of the model using KNN.
Table 7. Calculation of accuracy, precision, recall, and F1 score of the model using KNN.
AlgorithmTPFPFNTNAccuracyPrecisionRecallF1 Score
Unoptimized KNN6173957510.870.940.920.93
Optimized KNN6372553490.900.960.920.94
Oversampled optimized KNN3255233840.930.860.990.92
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ullah, I.; Khan, N.; Memon, S.A.; Kim, W.-G.; Saleem, J.; Manzoor, S. Vibration-Based Anomaly Detection for Induction Motors Using Machine Learning. Sensors 2025, 25, 773. https://doi.org/10.3390/s25030773

AMA Style

Ullah I, Khan N, Memon SA, Kim W-G, Saleem J, Manzoor S. Vibration-Based Anomaly Detection for Induction Motors Using Machine Learning. Sensors. 2025; 25(3):773. https://doi.org/10.3390/s25030773

Chicago/Turabian Style

Ullah, Ihsan, Nabeel Khan, Sufyan Ali Memon, Wan-Gu Kim, Jawad Saleem, and Sajjad Manzoor. 2025. "Vibration-Based Anomaly Detection for Induction Motors Using Machine Learning" Sensors 25, no. 3: 773. https://doi.org/10.3390/s25030773

APA Style

Ullah, I., Khan, N., Memon, S. A., Kim, W.-G., Saleem, J., & Manzoor, S. (2025). Vibration-Based Anomaly Detection for Induction Motors Using Machine Learning. Sensors, 25(3), 773. https://doi.org/10.3390/s25030773

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop