Next Article in Journal
Novel Algorithm for Radon Real-Time Measurements with a Pixelated Detector
Previous Article in Journal
Low-Cost Force Sensors Embedded in Physical Human–Machine Interfaces: Concept, Exemplary Realization on Upper-Body Exoskeleton, and Validation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dynamic Noise Reduction with Deep Residual Shrinkage Networks for Online Fault Classification

1
School of Computing, Engineering and Built Environment, Glasgow Caledonian University, 70 Cowcaddens Road, Glasgow G4 0BA, UK
2
Innovation Centre for Online Systems, 7 Townsend Business Park, Bere Regis BH20 7LA, UK
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(2), 515; https://doi.org/10.3390/s22020515
Submission received: 22 November 2021 / Revised: 1 January 2022 / Accepted: 6 January 2022 / Published: 10 January 2022
(This article belongs to the Topic Artificial Intelligence in Sensors)

Abstract

:
Fault signals in high-voltage (HV) power plant assets are captured using the electromagnetic interference (EMI) technique. The extracted EMI signals are taken under different conditions, introducing varying noise levels to the signals. The aim of this work is to address the varying noise levels found in captured EMI fault signals, using a deep-residual-shrinkage-network (DRSN) that implements shrinkage methods with learned thresholds to carry out de-noising for classification, along with a time-frequency signal decomposition method for feature engineering of raw time-series signals. The approach will be to train and validate several alternative DRSN architectures with previously expertly labeled EMI fault signals, with architectures then being tested on previously unseen data, the signals used will firstly be de-noised and a controlled amount of noise will be added to the signals at various levels. DRSN architectures are assessed based on their testing accuracy in the varying controlled noise levels. Results show DRSN architectures using the newly proposed residual-shrinkage-building-unit-2 (RSBU-2) to outperform the residual-shrinkage-building-unit-1 (RSBU-1) architectures in low signal-to-noise ratios. The findings show that implementing thresholding methods in noise environments provides attractive results and their methods prove to work well with real-world EMI fault signals, proving them to be sufficient for real-world EMI fault classification and condition monitoring.

1. Introduction

Power generation equipment and assets used in high-voltage (HV) power production plants are prone to developing faults; if these faults are undetected they can lead to breakdowns, in turn causing health and safety hazards, legal issues and incurring major losses such as fines and large-scale power outages [1]. Condition monitoring is carried out on HV assets for early fault detection and breakdown prevention, preventing aforementioned losses, and other unwanted outcomes. Condition monitoring is carried out manually by experts [2] observing electromagnetic interference (EMI) data in differing forms based upon individual preferences then following the observations present faults are classified, this method is often used to detect partial-discharge (PD) in assets [3]. The current expert-led manual approach to condition monitoring is problematic operationally due to high cost, sole reliance on experts to detect and classify faults, and lack of continuous monitoring. This leads to faults going unnoticed when experts are not available to carry out condition monitoring practices. An automated approach to condition monitoring will not only reduce dependence on experts but will also allow condition monitoring to be practiced continuously; this will prevent faults from going unnoticed and becoming malfunctions. Reference [4] outlines that the spectrum analysis of specific conditions can lead to more automated methods of data analysis. A continuous automated approach to condition monitoring is implemented using pattern recognition techniques by [5]. With machine learning (ML) also being implemented to focus on condition monitoring using EMI data, [6] implemented a deep residual neural network to classify EMI signals, [7] implements capsule networks to classify EMI signals. Although the produced results from these systems prove to be acceptable, there is room to increase classification results specifically in the presence of interference noise sources. Many of the downfalls of these systems arise from the varying settings found in the data collection process, as these signals are collected to diagnose a variety of assets hence all of the collected signals are subject to different levels of noise.
The work in this study proposes an automated classification system to classify several EMI fault signals in various noise levels. The produced ML system will classify EMI fault signals by using data in the form of Stockwell (S) transform time-frequency (TF) decomposition matrices, introduced by [8], as inputs. The ML system will implement a residual neural network (ResNet), an architecture built for image recognition proposed by [9], alongside various thresholding functions. Thresholding functions, also known as shrinkage functions, are used for signal de-noising; their functionality focuses on thresholding parameters often chosen by signal processing experts, to outline and alter values they believe to be in noise ranges; they then use the thresholding functions and their chosen parameters to create filters. The systems produced in this study implement architectures that learn thresholding parameters through model training, creating an automated de-noising filter. This automation allows de-noising without expert insight into the data being observed. Previous work using the learned thresholding method has been carried out by [10] focusing on the soft thresholding method and applying it to perform condition monitoring on mechanical power generation assets, the work implements learned thresholding by using a ResNet architecture with residual shrinkage blocks. The work in this study will build upon the residual-shrinkage-building-unit from [10] (RSBU) by implementing, to the author’s knowledge, previously unexplored methods of thresholding in the condition monitoring field, building a new residual-building-unit-2 (RSBU-2) architecture to perform thresholding functions that require two thresholding parameters and applying the built methods on EMI data. The EMI fault signal dataset used in this study is collected from real-world operational HV assets, containing 8 fault classes. This data was used to train and test 7 different learned thresholding systems, 4 of which contained a single learned thresholding parameter and the remaining 3 consisted of two learned thresholding parameters.
The following sections of this paper will outline the methods used in this study and will be structured as follows; Section 2 will introduce the S transform, Section 3 describes the thresholding functions implemented throughout the study outlining their learned parameters, Section 4 introduces and explains the two models used in the study, Section 5 outlines the experimental set-up of the research describing the dataset and tools used, Section 6 presents and discusses the research findings outlining how models were evaluated; finally, Section 7 will conclude the findings of the results with insights into future work.

2. Stockwell Transform

The feature engineering method used in this study is the S transform, introduced by [8]; this TF distribution is created to build upon the continuous wavelet transform (CWT) while avoiding some of the disadvantages of the method and the S transform is also considered to be a generalisation of the short-time Fourier Transform (STFT).
This study uses the S transform to create TF decompositions of raw time-series EMI signals. The produced TF decomposition retains frequency dependant resolution from the original time-series signals. The retention of this frequency resolution proves to be highly desirable in regards to EMI fault signals from real-world assets due to their non-stationary characteristics. The S transform is selected due to its ability to detect low frequency components of a signal while also providing the ability to detect short lived high frequency components of the signal under analysis, a quality that is not prevalent in the STFT.

Discrete Stockwell Transform

As the signals used throughout this study are of discrete time nature, the derived TF transform must cohere to this. The discrete time S transform for these signals can be found by performing operations on the Fourier transform, the discrete Fourier transform X [ k ] of signal x [ n ] is found using Equation (1).
X [ k ] = n = 0 N 1 x [ n ] e j 2 π n k N .
The S transform can then be found by carrying out operations on the Fourier transform from Equation (1). This is outlined in Equation (2), where n 0 and ρ = m = n = 0 , 1 , , N 1 .
S [ ρ , n ] = m = 0 N 1 X [ m + n ] e 2 π 2 m 2 n 2 e j 2 π m ρ N .
However, Equation (2) does not include the calculation for n = 0 . To accommodate for this when n = 0 is encountered Equation (3) is used.
S [ ρ , 0 ] = 1 N m = 0 N 1 x [ m ] .
The produced S transform then undergoes further processing to convert the imaginary product of Equation (2) to a modulus product; this is done by taking the absolute value of the S transform | S [ ρ , n ] | . The feature extraction process through the S transform method can be seen in Figure 1.

3. Thresholding Methods

Thresholding or shrinkage functions are primarily used for de-noising as observed in [11,12]. The thresholding process consists of creating value ranges which correspond to noise then, dependent upon the thresholding method values both within and outwith these ranges, they are altered. An example of which being the commonly used soft thresholding method from [13]; this thresholding method deems values between the thresholding parameter and zero unimportant, hence all values within this range are set to zero and values outwith the thresholding parameter range are given non-zero values. The thresholding parameter ranges are often chosen by signal processing experts, with expertise in the data at hand. Deep learning reduces dependence on signal processing experts for selecting optimal thresholding parameters for de-noising, and carries out thresholding by learning thresholding parameters through gradient descent, producing thresholding-based filters relevant to the data being observed. Such data-oriented filters should intuitively lead to classification with greater accuracy and allow classifications to occur in high noise environments. The thresholding methods used throughout this study will be used to alter features in the feature map that is undergoing thresholding deemed to correspond to noise; this is outlined in Figure 2.
The work carried out in this study will observe several thresholding methods; soft thresholding and hard thresholding [13], firm thresholding [14], Garrote thresholding [15], quadratic curve thresholding (QCT) [16] and Hyper trim (H-trim) thresholding [17]. Thresholding parameters for all of the thresholding methods will be learned using deep learning, graphical representations of the thresholding methods used are found in Figure 3.

3.1. Soft Thresholding

The soft thresholding method sets values within the given threshold ranges to zero and values outwith these ranges retain their original values minus the threshold parameter value. This relationship is shown in Equation (4), where γ represents the learned threshold parameter.
δ ( w ) = 0 , | w | < γ ( | w | γ ) · s g n ( w ) , | w | γ .

3.2. Hard Thresholding

Hard thresholding sets values within threshold ranges to zero and while values outwith the threshold ranges retain their original values; this relationship is shown in Equation (5).
δ ( w ) = 0 , | w | < γ w , | w | γ .

3.3. Garrote Thresholding

Originally introduced to overcome downfalls found in the soft and hard thresholding methods, it has a non-linear approach to values outwith its single thresholding parameter range. Values within the thresholding range are set to zero while values outwith this range are altered non-linearly, as shown in Equation (6).
δ ( w ) = 0 , | w | < γ w γ 2 w , | w | γ .

3.4. Firm Thresholding

Firm thresholding consists of two thresholding parameters; this leads to the alteration of values in three ranges. Values within zero and the lower thresholding parameter are set to zero, values within the lower and upper threshold parameters are altered accordingly, and values above the upper threshold limit retain their original values. These relationships are outlined in Equation (7), where γ represents the lower threshold limit and λ represents the upper threshold limit.
δ ( w ) = 0 , | w | γ λ ( | w | γ ) λ γ · s g n ( w ) , γ | w | < λ w , | w | λ .

3.5. Hyper-Trim Thresholding

H-trim thresholding is a method introduced to outperform soft thresholding, prior to the introduction of H-trim, firm thresholding was seen to outperform soft thresholding in electrocardiogram (ECG) signal denoising although firm thresholding had two thresholding parameters compared to H-trims single thresholding parameter. Equation (8) outlines the H-trim thresholding method.
δ ( w ) = 0 , | w | < γ t a n h ( w ) , | w | γ .

3.6. Quadratic Curve Thresholding

QCT is a method introduced to overcome the weaknesses found in soft and hard thresholding. Overcoming the Gibbs phenomenon, outlined in [18], caused by an overshoot at jump discontinuities when thresholds are achieved in hard thresholding and distortion found when | w | γ in soft thresholding [16]. QCT contains two thresholding parameters and three coefficients, Equation (9) shows the mathematical relationships for QCT. Where γ and λ represent the lower and upper thresholding parameters respectively and the coefficients used are represented by a, b, and c.
δ ( w ) = 0 , | w | < γ a w 2 + b w + c γ , γ | w | < λ w , | w | λ .
Values within the lower threshold range are set to zero, values within the lower and upper threshold range are assigned a quadratic curve value and finally values greater than the upper threshold range retain their original values. Given that ( γ , 0 ) and ( λ , λ ) are points on the same quadratic curve, the set of equations in Equation (10) can be found.
0 = a γ 2 + b γ + c λ = a λ 2 + b λ + c .
Equation (10) can then be further simplified to describe b and c with respect to the thresholding ranges, this is shown in Equations (11) and (12) respectively. Showing three parameters to be assigned values, γ and λ , which are the learned thresholding parameters and a, which is a coefficient to be assigned a value, a = 1 was used for the purposes of this study.
b = λ λ γ a ( λ + γ )
c = γ λ [ 1 γ λ + a ] .

4. Model Architecture

The studies throughout this work implemented two neural-network architectures based upon the DRSN proposed in [10]. Architectures are developed using a deep ResNet architecture along with RSBUs, ResNet architectures developed in [9] that are a form of convolutional neural network (CNN) that are widely used in the image classification field due to their state-of-the-art performance on image classification tasks. The main alteration in ResNets in comparison to conventional CNNs is the presence of identity skip connections which reduce training error and loss, due to their ability to eliminate the vanishing or exploding gradient problem.
This study builds upon the work in [10] by implementing further thresholding methods, with both single and two thresholding parameters, to tackle EMI signal classification tasks, building upon the DRSN architecture to accommodate thresholding methods with two learned thresholds and observing the benefits of using these alternative forms of learned thresholding methods.
The overall DRSN architecture used in this study can be found in Figure 4, the architecture is shown to stack RSBU blocks to gradually decrease noise-related features. Alterations in the RSBU are proposed for learning two thresholding parameters. This new architecture is referred to as RSBU-2 with the original architecture for learning a single thresholding parameter being referred to as RSBU-1.

4.1. RSBU-1

Designed to learn a single thresholding parameter, the RSBU-1 architecture can be found in Figure 5. The channel-wise threshold parameter γ ( c ) is found by following several steps. Firstly, through finding the absolute value of the global average pooling (GAP), the feature map is reduced to a one-dimensional vector. GAP provides many benefits to the overall architecture and the operation calculates the mean value from each channel of the input feature map, in turn reducing the number of weights to be used in the following FC layer reducing the probability of encountering over-fitting. The one-dimensional vector is then propagated to a two-layer fully connected (FC) network and the output of this network can be found in Figure 5 as α ( c ) , which is the scaling parameter of the cth neuron and it is scaled to the range ( 0 , 1 ) using Equation (13), where z ( c ) represents the feature of the cth neuron; this relationship is shown as the Sigmoid layer in Figure 5.
α ( c ) = 1 1 + e z ( c ) .
The discovered scaling parameter α ( c ) is then used to find the thresholding parameter of the cth channel of the feature map γ ( c ) through element-wise multiplication of α ( c ) and the output of the absolute GAP shown as E ( h , w ) [ w ( h , w , c ) ] , this relationship is found in Equation (14), where h, w and c show the indexes of the height, width and channels of the feature map w.
γ ( c ) = α ( c ) · E ( h , w ) [ w ( h , w , c ) ] .
The RSBU-1 architecture was used to find the single thresholding parameters for soft, hard, Garrote, and H-trim thresholding.

4.2. RSBU-2

The RSBU-2 architecture, found in Figure 6, was created to accommodate a second thresholding parameter referred to in this work as λ . The first channel-wise thresholding parameter γ ( c ) of RSBU-2 is found identically to the methods explored in RSBU-1 and outlined in Section 4.1. The discovery of the second channel-wise thresholding parameter λ ( c ) is outlined in this sub-section.
First, the feature map w undergoes absolute GAP, the resultant one-dimensional vector is propagated to a two-layer FC network and the produced output of the FC network q ( c ) is scaled using Equation (15) producing the scaling parameter of the cth channel β ( c ) .
β ( c ) = 1 1 + e q ( c ) .
λ ( c ) is then found by carrying out an element-wise multiplication between the calculated β ( c ) and the absolute GAP of w shown as E ( h , w ) [ w ( h , w , c ) ] ; this relationship is outlined in Equation (16).
λ ( c ) = β ( c ) · E ( h , w ) [ w ( h , w , c ) ] .
The RSBU-2 architecture was used to find two learned thresholding parameters γ and λ , and was used to implement firm and QCT thresholding.

5. Experimental Set-Up

The work in this study used real-world fault signals to produce an automated noise-robust classification model using ML methods, this section will outline the necessary steps taken to process the signal data in preparation for model ingestion outlining the steps taken to produce datasets at various known noise levels, feature engineering, in the form of TF decomposition methods, to create data that can be ingested into the classification models, and how the implemented classification models underwent training, validation, and testing. Figure 7 outlines the feature extraction and classification system used to produce fault predictions from EMI signals at varying noise levels.

5.1. Data Set

The EMI signals used in this study were collected in adherence with the Comité International Spécial des Pertrubations Radioélectriques (CISPR) 16 standard from [19], and implemented the EMI signal collection technique from [20]. The collected data were in the form of time-resolved signals from real-world operational assets sampled at a rate of 24000 samples/second and signals were labelled with their corresponding faults through manual expert analysis. The raw time series signals can be seen in Figure 8. Eight fault classes were prevalent in the dataset signals; PD, Exciter, Arcing, Data-Modulation, Processing Noise, Random Noise, and Micro-Sparking. A balanced dataset was produced from the collected fault signals and each class was represented with 261 instances with a signal length of 4000 samples per signal.
The signals in the original dataset then underwent further pre-processing to produce 11 datasets with known noise levels, this was carried out by denoising the original signals using a symlet4 wavelet with a posterior median threshold rule from [21] implemented using the wavelet denoising (wdenoise) method from Matlab and noise variance was estimated based on the highest-resolution wavelet coefficients. Following the denoising of the original signals, additive white Gaussian noise was added at desired levels chosen to be 5 , 4 , 3 , 2 , 1 , 0, 1, 2, 3, 4 and 5 dB signal-to-noise-ratio (SNR), the process carried out to create the varying noise level datasets is shown in Figure 9. All 11 known noise datasets were also split into further sub-sections to produce training, validation, and testing splits which contained 70 % , 15 % , and 15 % of signals from their original datasets, respectively.
Prior to model ingestion, the signals in the dataset underwent feature engineering to become two-dimensional real-valued TF mappings, this decomposition was performed using the S transform outlined in Section 2 and then taking the modulus of the produced imaginary TF mapping. Following the feature engineering step, the datasets consist of 261 two-dimensional TF mappings for each fault class, along with the relevant expertly assigned fault label.

5.2. Model Training & Testing

The studies required the training of six different learned thresholding architectures and two conventional ML methods, a 50-layer ResNet model (ResNet50) and a 16-layer Visual Geometry Group (VGG16) model proposed in [22], along with eleven alternate known noise datasets. Each architecture and dataset pairing was run ten times, this led to 880 training instances. Model training was implemented using Tensorflow [23], models were trained by reducing categorical cross-entropy loss along with a momentum optimiser with the momentum co-efficient set to 0.9 , as recommended by [9], over 650 epochs, the learning rate alternated from 0.1 in the first 260 epochs, then 0.01 in the following 260 epochs, and 0.001 in the final 130 epochs. L2 regularization is used in order to increase test accuracy and reduce the effects of over-fitting; as explained in [24], the L2 regularization penalty term co-efficients were set to 0.0001 staying consistent with classical ResNets [9]. The model with the greatest validation accuracy over all of the epochs in the training instance was selected as the model used for testing. Models with the best validation accuracy from all 880 training runs were tested, and average testing accuracies, and the receiver operating characteristic area under the curve (ROC-AUC) scores were procured through grouping architectures with their model and data pairs, each having ten instances. Testing accuracies were calculated using binary accuracy, which is the division of the total number of correct predictions by the number of test samples. Mathematically described in Equation (17), where TP, TN, FP, and FN, represent true positives and negatives and false positives and negatives respectively. The ROC-AUC score is the area under the probability curve, also known as the ROC; the ROC-AUC score is limited between 0 1 and a value closer to 1 shows a model with greater class separability, meaning the model is better at distinguishing between classes.
Acc = c o r r e c t p r e d i c t i o n s a l l p r e d i c t i o n s = T P + T N T P + T N + F P + F N

6. Results & Discussion

The mean test accuracies over the ten runs are provided by testing the thresholding architectures against corresponding known noise level data and the results can be found in Table 1. Figure 10 outlines the corresponding ROC-AUC scores of the various thresholding methods, implemented through the proposed DRSN networks, and the conventional ML architectures, with the alternating known noise level datasets. The results show thresholding models with two thresholding parameters to outperform their single thresholding parameter counterparts in high-noise environments when considering mean test accuracies. The results also outline the benefits of exploring further thresholding methods in the HV condition monitoring field, as to the author’s knowledge soft thresholding is the only learned thresholding method used in the condition monitoring field prior to this study and it is shown to be outperformed in a majority of noise environments in EMI signal data. The DRSN networks proposed also show vast performance improvements when compared to state-of-the-art conventional ML classification architectures showing double figure improvements in mean test accuracies in all noise environments, with further support when comparing mean ROC-AUC scores.

7. Conclusions

The work in this study produced several thresholding methods to tackle the noisy environments in which HV condition monitoring, through EMI signal analysis, incurs. Providing methods that automate the denoising and classification process of EMI signals, the results of the work found the benefits of exploring novel thresholding methods in the field, how these thresholding methods can be used with ML to produce a non-expert dependant system to carry out denoising prior to classification, and the work produced an architecture to allow for the learning of several thresholding parameters when required for thresholding methods which require more than one learned parameter. The architecture proposed in this work showed significant improvements in current conventional ML classification methods.
The systems produced in this work can be used to develop an adaptable condition monitoring software to reduce dependence on both HV condition monitoring experts and signal processing experts. Creating an adaptable and robust fault classification system will benefit users and producers of electricity by ensuring the prevention of devastating losses. The work from the study can be built upon by observing further thresholding methods and exposing these methods to assets in real-world settings.

Author Contributions

All authors of this paper contributed to the work. A.S., I.M., P.B., A.N. and G.M. conceived and designed the experiments and helped draw conclusions; A.S. performed the experiments; A.S., I.M. and G.M. analyzed the data; P.B. contributed data/reagents/materials/analysis tools; A.S. wrote the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Doble Engineering Company grant number 18-135.

Institutional Review Board Statement

Not Applicable.

Informed Consent Statement

Not Applicable.

Data Availability Statement

The data used in this research is private due to the client confidentiality policy of Doble Engineering Company.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
HVHigh-voltage
EMIElectro-magnetic interference
DRSNDeep residual shrinkage networks
RSBUResidual shrinkage building unit
PDPartial discharge
MLMachine Learning
SStockwell
ResNetResidual Neural Network
CWTContinuous wavelet transform
TFTime-frequency
QCTQuadratic curve thresholding
ECGElectrocardiogram
CNNConvolutional neural network
GAPGlobal average pooling
FCFully connected
CISPRComité International Spécial des Pertrubations Radioélectriques
SNRSignal-to-noise-ratio
VGGVisual Geometry Group

References

  1. Barrios, S.; Buldian, D.; Comech, M.P.; Gilbert, I.; Orue, I. Partial Discharge Classification Using Deep Learning Methods—Survey of Recent Progress. Energies 2019, 12, 2485. [Google Scholar] [CrossRef] [Green Version]
  2. Timperley, J.E.; Vallejo, J.M. Condition assessment of electrical apparatus with emi diagnostics. In Proceedings of the IEEE Petroleum and Chemical Industry Committee Conference (PCIC), Houston, TX, USA, 5–7 October 2015; pp. 1–8. [Google Scholar]
  3. Timperley, J.E.; Buchanan, D.; Vallejo, J. Electric generation condition assessment with electromagnetic interference analysis. IEEE Trans. Ind. Appl. 2017, 54, 1921–1929. [Google Scholar] [CrossRef]
  4. Timperley, J.E. Audio spectrum analysis of EMI patterns. In Proceedings of the 2007 Electrical Insulation Conference and Electrical Manufacturing Expo, Nashville, TN, USA, 22–24 October 2007; pp. 39–41. [Google Scholar]
  5. Mitiche, I.; Morison, G.; Hughes-Narborough, M.; Nesbitt, A.; Boreham, P.; Stewart, B.G. Classification of Partial Discharge Signals by Combining Adaptive Local Iterative Filtering and Entropy Features. Sensors 2018, 18, 406. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Mitiche, I.; Jenkins, M.D.; Boreham, P.; Nesbitt, A.; Morison, G. Deep complex neural network learning for high-voltage insulation fault classification from complex bispectrum representation. In Proceedings of the 27th European Signal Processing Conference (EUSIPCO), A Coruna, Spain, 2–6 September 2019. [Google Scholar]
  7. Salimy, A.; Mitiche, I.; Boreham, P.; Nesbitt, A.; Morison, G. Low Complexity Classification of Power Asset Faults for Real Time IoT-based Diagnostics. In Proceedings of the IEEE Global Conference on Artificial Intelligence and Internet of Things (GCAIoT), Dubai, United Arab Emirates, 12–16 December 2020. [Google Scholar]
  8. Stockwell, R.G.; Mansinha, L.; Lowe, R.P. Localization of the complex spectrum: The S transform. IEEE Trans. Signal Process. 1996, 44, 998–1001. [Google Scholar] [CrossRef]
  9. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 23–26 June 2016; pp. 770–778. [Google Scholar]
  10. Zhao, M.; Zhong, S.; Fu, X.; Tang, B.; Pecht, M. Deep Residual Shrinkage Networks for Fault Diagnosis. IEEE Trans. Ind. Informat. 2020, 16, 4681–4690. [Google Scholar] [CrossRef]
  11. Donoho, D.L. De-noising by soft-thresholding. IEEE Trans. Inf. Theory 1995, 41, 613–627. [Google Scholar] [CrossRef] [Green Version]
  12. Isogawa, K.; Ida, T.; Shiodera, T.; Takeguchi, T. Deep shrinkage convolutional neural network for adaptive noise reduction. IEEE Signal Process. Lett. 2018, 25, 224–228. [Google Scholar] [CrossRef]
  13. Donoho, D.L.; Johnstone, I.M. Ideal Spatial Adaptation via Wavelet Shrinkage. Biometrica 1994, 81, 425–455. [Google Scholar] [CrossRef]
  14. Gao, H.; Bruce, A.G. Waveshrink with Firm Shrinkage. Stat. Sin. 1997, 7, 855–874. [Google Scholar]
  15. Gao, H. Wavelet Shrinkage Denoising Using the Non-Negative Garrote. J. Comput. Graph. Stat. 1998, 7, 469–488. [Google Scholar]
  16. He, H.; Wang, Z.; Tan, Y. Noise reduction of ECG signals through genetic optimized wavelet threshold filtering. In Proceedings of the 2015 IEEE International Conference on Computational Intelligence and Virtual Environments for Measurement Systems and Applications (CIVEMSA), Shenzhen, China, 12–14 June 2015; pp. 1–6. [Google Scholar]
  17. Poornachandra, S.; Kumaravel, N. Hyper-trim shrinkage for denoising of ECG signal. Digit. Signal Process 2005, 15, 317–327. [Google Scholar] [CrossRef]
  18. Hewitt, E.; Hewitt, R.E. The Gibbs-Wilbraham phenomenon: An episode in fourier analysis. Arch. Hist. Exact Sci. 1979, 21, 129–160. [Google Scholar] [CrossRef]
  19. Specification for Radio Disturbance and Immunity Measurement Apparatus and Methods-Part 1: Radio Disturbance and Immunity Measuring Apparatus, IEC: 2015, CISPR 16-1-1, Part 1-1. Available online: https://webstore.iec.ch/publication/23387 (accessed on 7 January 2022).
  20. Timperley, J.E.; Vallejo, J. Condition assessment of electrical apparatus with emi diagnostics. IEEE Trans. Ind. Appl. 2017, 53, 693–699. [Google Scholar] [CrossRef]
  21. Johnstone, I.M.; Silverman, B.W. Needles and Straw in Haystacks: Empirical Bayes Estimates of Possibly Sparse Sequences. Ann. Stat. 2004, 32, 1594–1649. [Google Scholar] [CrossRef] [Green Version]
  22. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  23. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. TensorFlow: A System for Large-scale Machine Learning. In Proceedings of the 12th USENIX Symposium on Operating Systems Design and Implementation (OSDI ’16), Savannah, GA, USA, 2–4 November 2016; pp. 265–283. [Google Scholar]
  24. LeCun, Y.; Bengio, Y.; Hinton, G. Deep Learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Feature extraction process through the Stockwell transform time-frequency decomposition.
Figure 1. Feature extraction process through the Stockwell transform time-frequency decomposition.
Sensors 22 00515 g001
Figure 2. Illustration outlining the feature thresholding method.
Figure 2. Illustration outlining the feature thresholding method.
Sensors 22 00515 g002
Figure 3. Graphical representation of thresholding methods (a)−soft thresholding, (b)−hard thresholding, (c)−firm thresholding, (d)−Garrote thresholding, (e)−quadratic curve thresholding, (f)−hyper-trim thresholding.
Figure 3. Graphical representation of thresholding methods (a)−soft thresholding, (b)−hard thresholding, (c)−firm thresholding, (d)−Garrote thresholding, (e)−quadratic curve thresholding, (f)−hyper-trim thresholding.
Sensors 22 00515 g003
Figure 4. Overall deep residual shrinkage network.
Figure 4. Overall deep residual shrinkage network.
Sensors 22 00515 g004
Figure 5. Residual Shrinkage Building Unit-1 Architecture.
Figure 5. Residual Shrinkage Building Unit-1 Architecture.
Sensors 22 00515 g005
Figure 6. Residual Shrinkage Building Unit-2 Architecture.
Figure 6. Residual Shrinkage Building Unit-2 Architecture.
Sensors 22 00515 g006
Figure 7. Flow diagram showing the proposed system, outlining feature extraction and classification of EMI signals at varying noise level inputs.
Figure 7. Flow diagram showing the proposed system, outlining feature extraction and classification of EMI signals at varying noise level inputs.
Sensors 22 00515 g007
Figure 8. Raw time-series EMI fault signals, representing each prevalent class.
Figure 8. Raw time-series EMI fault signals, representing each prevalent class.
Sensors 22 00515 g008
Figure 9. Flow diagram picturing the steps taken to create datasets with varying known noise levels.
Figure 9. Flow diagram picturing the steps taken to create datasets with varying known noise levels.
Sensors 22 00515 g009
Figure 10. Bar graph of average receiving operating characteristics area under the curve score of alternative thresholding, ResNet50 and VGG16 methods. DRSN with RSBU-1 is used to implement Soft, Hard and Garrote thresholding, DRSN with RSBU-2 is used to implement H-trim, Firm and QCT thresholding.
Figure 10. Bar graph of average receiving operating characteristics area under the curve score of alternative thresholding, ResNet50 and VGG16 methods. DRSN with RSBU-1 is used to implement Soft, Hard and Garrote thresholding, DRSN with RSBU-2 is used to implement H-trim, Firm and QCT thresholding.
Sensors 22 00515 g010
Table 1. Average Testing accuracies of thresholding, ResNet50 and VGG16 models with alternate noise level data. Best performance is presented in bold font. DRSN with RSBU-1 is used to implement Soft, Hard and Garrote thresholding, DRSN with RSBU-2 is used to implement H-trim, Firm and QCT thresholding.
Table 1. Average Testing accuracies of thresholding, ResNet50 and VGG16 models with alternate noise level data. Best performance is presented in bold font. DRSN with RSBU-1 is used to implement Soft, Hard and Garrote thresholding, DRSN with RSBU-2 is used to implement H-trim, Firm and QCT thresholding.
Data Set (dB SNR)SoftHardGarroteH-trimFirmQCTResNet50VGG16
−553.54%52.63%53.72%53.94%54.71%56.09%43.21%42.12%
−458.10%55.51%57.59%57.48%60.07%60.84%43.25%42.34%
−358.32%59.74%60.62%58.61%60.15%61.68%45.84%43.98%
−262.85%60.18%63.10%61.39%60.88%64.42%45.69%44.27%
−162.48%60.29%61.13%60.36%62.85%62.70%47.78%46.75%
067.12%67.15%67.77%65.77%68.50%67.59%51.02%52.52%
175.07%74.56%76.06%73.91%74.64%76.13%49.42%51.75%
279.12%77.15%77.70%76.02%77.52%74.38%54.34%59.34%
379.31%76.53%80.55%77.88%78.72%79.20%58.94%62.26%
485.84%83.21%86.42%83.58%85.00%85.00%65.36%70.99%
584.38%80.47%83.58%81.57%81.28%83.83%68.14%72.48%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Salimy, A.; Mitiche, I.; Boreham, P.; Nesbitt, A.; Morison, G. Dynamic Noise Reduction with Deep Residual Shrinkage Networks for Online Fault Classification. Sensors 2022, 22, 515. https://doi.org/10.3390/s22020515

AMA Style

Salimy A, Mitiche I, Boreham P, Nesbitt A, Morison G. Dynamic Noise Reduction with Deep Residual Shrinkage Networks for Online Fault Classification. Sensors. 2022; 22(2):515. https://doi.org/10.3390/s22020515

Chicago/Turabian Style

Salimy, Alireza, Imene Mitiche, Philip Boreham, Alan Nesbitt, and Gordon Morison. 2022. "Dynamic Noise Reduction with Deep Residual Shrinkage Networks for Online Fault Classification" Sensors 22, no. 2: 515. https://doi.org/10.3390/s22020515

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop