Next Article in Journal
Visual Object Tracking Using Structured Sparse PCA-Based Appearance Representation and Online Learning
Next Article in Special Issue
Underwater Target Detection and 3D Reconstruction System Based on Binocular Vision
Previous Article in Journal
A Novel Scheme for MIMO-SAR Systems Using Rotational Orbital Angular Momentum
Previous Article in Special Issue
A Bimodal Model to Estimate Dynamic Metropolitan Population by Mobile Phone Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Partial Discharge Recognition with a Multi-Resolution Convolutional Neural Network

1
State Key Laboratory of Electrical Insulation and Power Equipment, School of Electrical Engineering, Xi’an Jiaotong University, Xi’an 710049, China
2
The Global Energy Interconnection Development and Cooperation Organization (GEIDCO), Beijing 100031, China
*
Authors to whom correspondence should be addressed.
Sensors 2018, 18(10), 3512; https://doi.org/10.3390/s18103512
Submission received: 24 September 2018 / Revised: 11 October 2018 / Accepted: 16 October 2018 / Published: 18 October 2018
(This article belongs to the Special Issue Multi-Sensor Fusion and Data Analysis)

Abstract

:
Partial discharge (PD) is not only an important symptom for monitoring the imperfections in the insulation system of a gas-insulated switchgear (GIS), but also the factor that accelerates the degradation. At present, monitoring ultra-high-frequency (UHF) signals induced by PDs is regarded as one of the most effective approaches for assessing the insulation severity and classifying the PDs. Therefore, in this paper, a deep learning-based PD classification algorithm is proposed and realized with a multi-column convolutional neural network (CNN) that incorporates UHF spectra of multiple resolutions. First, three subnetworks, as characterized by their specified designed temporal filters, frequency filters, and texture filters, are organized and then intergraded by a fully-connected neural network. Then, a long short-term memory (LSTM) network is utilized for fusing the embedded multi-sensor information. Furthermore, to alleviate the risk of overfitting, a transfer learning approach inspired by manifold learning is also present for model training. To demonstrate, 13 modes of defects considering both the defect types and their relative positions were well designed for a simulated GIS tank. A detailed analysis of the performance reveals the clear superiority of the proposed method, compared to18 typical baselines. Several advanced visualization techniques are also implemented to explore the possible qualitative interpretations of the learned features. Finally, a unified framework based on matrix projection is discussed to provide a possible explanation for the effectiveness of the architecture.

1. Introduction

Gas-insulated switchgears (GISs) are widely used as the major control and protection equipment in medium to ultra-high voltage substations, due to their superior compactness, high reliability, strong dielectric strength, and maintenance-free properties. However, unavoidable imperfections and deterioration in the insulation system pose serious threats to the reliability and safety of the GIS and whole power grids [1,2,3]. The occurrence of partial discharges (PDs) is not only one of the main characteristics that can effectively reflect the inner dielectric flaws, but is also the cause of the accelerated degradation [4,5]. Therefore, it is imperative to assess the potential correlation between the PD patterns and the defect types, so that corresponding maintenance activities can be taken before a complete breakdown.
The conventional monitoring methods of PD could be summarized into three main categories: chemical, electrical, and acoustic [6]. Among these, ultra-high-frequency (UHF) signals are becoming increasingly important due to their high sensitivity, better immunity against interference, and the ability for real-time monitoring [7,8]. Considering the stochastic nature of PDs, machine learning has long been the mainstream method for the diagnosis of UHF signals. The existing literature covers a wide range of techniques, including Support Vector Machines (SVM) [9], Neural Networks (NN) [10,11,12], cluster analysis [13], fuzzy logic [14] and so on. A more detailed taxonomy of NNs could give rise to methodologies such as the self-organizing map (SOM) [10], Ensemble neural networks (ENN) [11], and Probabilistic Neural Networks (PNN) [12]. Compared with classifiers, feature extraction plays a more dominant role in the success of PD recognition due to the high dimensionality of the PD data. A gamut of techniques have been applied to both the time-resolved partial discharge (TRPD) and phase-resolved partial discharge (PRPD) patterns [6], of which the statistical parameters [15], the time-power ratio map [16], fractal features [17], Fourier transform [18], wavelet analysis [12], Hilbert-Huang transform (HHT) [19], and S-transform [20] are a few. In addition, some advanced signal processing techniques such as compressed sensing [21] and principal component analysis (PCA) [22] are also utilized. A summary of these implementations is shown in Table 1.
Although the cooperation of features and algorithms indicate a considerable degree of success, it is pertinent to note that most of them suffer from a fundamental limitation: the carefully selective feature extraction process usually calls for much human expertise and hence, it is subject to human error. The significant randomness of the electromagnetic waves caused by propagation and reflection makes the process even more intricate [23]. Besides, even with all these efforts, the features designed for a particular task are not guaranteed to be the optimal choices for the other application scenarios and datasets. Therefore, learning features directly from the dataset itself may be a better choice.
Among all of the analysis methods, time–frequency (TF) analysis is especially applicable, due to its invertibility, and its ability to provide both time and frequency resolutions [24]. From the perspective of manifold learning, the intrinsic advantage of the TF method is to obtain continuity and compactness in a high-dimensional space at the cost of dimensionality. However, the high dimensionality poses a substantial impediment for the traditional machine learning models. The matrix compression algorithms for dimensionality reduction such as the two-dimensional PCA (2DPCA) [25] and Non-negative matrix factorization (NMF) [26] inevitably share similar obstacles with the other feature engineering-based methodologies, such as adaptation and generalization.
Fortunately, the recent advances in deep learning make it possible to extract high-level features automatically from high-dimensional sensory data, and they demonstrate dramatic success in various areas such as natural language processing, image classification, and auto-driving, of which the Convolutional Neural Network (CNN) has been designed for vision-related tasks [27,28]. Therefore, the combination of the CNN and TF maps becomes a potentially promising option for the PD signal recognition tasks. Nevertheless, although it sounds transparent, directly applying CNN for TF maps is not a good idea, due to the inherently non-local and serial characteristics of TF maps. In our preliminary work [29], we proposed a CNN of frequency matching to obtain the semantic meaning. In this paper, we further improve the framework into multiple columns to realize multi-resolution diagnosis, thus promoting the diagnostic accuracy to a new level. The major contributions are summarized as follows.
  • We propose a deep multi-column architecture that incorporates multi-resolution information for the recognition of UHF signals with specified temporal filters, frequency filters, and texture filters. Multi-sensor fusion based on LSTM is also utilized to further improve the diagnostic accuracy.
  • To alleviate the risk of overfitting induced by the expanded structure, a spectral embedding technique based on manifold learning is utilized for the pretraining. We demonstrate that the transfer learning-based training process is able to achieve remarkable performance with limited data.
  • A unified matrix projection-based framework is proposed to enhance the interpretability of the proposed model.
The rest of this paper is organized as follows. Some related theories are introduced first in Section 2. Section 3 provides the details of both the model architectures and the transfer learning process. The platform presented in Section 4 provides the necessary dataset for verifying the proposed method, and Section 5 describes the recognition results. Based on the experimental results, Section 6 discusses the effectiveness of the proposed model. Finally, Section 7 concludes the paper.

2. Related Work

In deep learning, the most straightforward way to improve the model performance has shifted from feature engineering to network architecture designing. ResNet [30] proposed by Microsoft is a good example, which introduces an extra column with residual mapping, and has state-of-the-art performance. Although the early implementation of the multi-column neural network only aimed at accelerating the training process with multiple Graphics Processing Units (GPUs) [28], it has been realized that the several columns can become specified experts for different features. For example, the Inception model [31] is a successful multi-column deep neural network (DNN) to approximate the optimal sparse structure of the input with different convolutional kernels. The Xception model [32] goes even further by decoupling both the cross-channel and spatial correlations to obtain a multi-branch structure. In addition, Saining exposed a new dimension called cardinality to standardize the multi-branch structure [33]. Ling attempted to train a multi-column neural network with spectral clustering to embed the hierarchical manifolds into a compact vector [34].
Although sharing the similar structure of multiple columns, the multi-resolution theories focus more on the structures of different scales. Sander attempted to learn music features by extracting the mel-spectrograms with different windows [35]. Michael utilized CNN to predict future images from different resolutions and concluded that a lower resolution was able to capture the entire context, while a higher resolution recovered more details [36]. The similar idea was also shared by MMDenseNet [37], which enables the neural network to model both the fine-grained structures and the long contexts efficiently. In the Deep Belief Network, the multi-resolution structures also showed their superiority in image-related tasks [38]. Furthermore, the idea of multiple resolutions has long been integrated into the underlying design of neural networks. The pooling layers and the variants, such as subsequent spatial pyramid pooling all attempt to obtain the larger space by reducing the resolution of the intermediary layers [39].

3. Proposed Method

3.1. Gabor Representations and Multi-Resolution Analysis

Fourier transform is a powerful tool to provide the frequency distributions of complex signals. However, we must simply remark that the one-dimensional solutions are delocalized in time and not comprehensive enough. The bidimensional functions that provide both the frequency and time localizations are preferred. To introduce the time resolution, a time-localized window is utilized to suppress the signal outside the neighborhood, which is the well-known short-time Fourier transform (STFT) [24]:
F x ( t , v ; h ) = + x ( u ) h * ( u t ) e x p [ j 2 π v u ] d u
where x ( u ) is the original signal, and h is the analysis window. t and v represent the time and frequency, respectively. In practical applications, the discrete STFT is preferred, which is given by:
F x [ n , m ; h ] = F x ( n t 0 , m v 0 ; h ) = + x ( u ) h * ( u n t 0 ) e x p [ j 2 π m v 0 u ] d u
where m , n Z . t 0 and v 0 are the time and frequency steps. Specifically, the STFT with Gaussian windows is able to theoretically maximize the concentration in the TF plane by approaching the lower bound of the Heisenberg–Gabor inequality [24]:
T × B 1
Here, T is the time duration, and B is the normalized bandwidth. The STFT of this particular case is also called the Gabor transform. The most important inference of the Heisenberg-Gabor inequality is the trade-off between time and frequency resolutions in accordance with the width of the analysis window. An illustrative example is shown in Figure 1.
As shown in Figure 1, the long window gives rise to harmonic information, while the pulses emerge from the short time window. In this study, the spectrograms of different resolutions are all employed for the diagnosis of the UHF signals. It should be noted that the concept of multiple resolutions is different from the wavelet or S-transform, which still generate only a single resolution for each frequency band.
Furthermore, a semiquantitative method is proposed as an assistance criteria for choosing the proper analysis windows, considering the morphology characteristics of the STFT spectrums. As indicated by Figure 2, centering on the max value, the time and frequency ranges before reaching a certain threshold are calculated, which can be regarded as a rough description of the peak shape. Intuitively, the greater the time range is, the less the time resolution is retained, while the frequency isolation is more obvious. The frequency range varies in an opposite way. The proper window can be selected based on the trade-off between the time and frequency resolutions, depending on their priorities.

3.2. Convolutional Neural Network and the Design of Convolutional Kernels

In recent years, CNN and its derivatives have been regarded as the state-of-the-art choices for computer vision tasks. The key ideas in CNN include the concepts of local receptive fields, shared weights, and spatial subsampling. Specifically, in each convolutional layer, the feature maps are first convolved with several kernels. After the biases, activation function and sometimes, the pooling operation, the outputs are then fed into the next convolutional layer, which can be summarized as:
x j l = f ( i M j x i l 1 k i j l + b j l )
where donates the convolutional operator. x j l 1 and x i l are the input and output of the lth layer, respectively, and b is the bias. M j is a selection of the feature maps, and k i j l is the element of the kernel matrix.
The activation function f is chosen as the Relu function for the convenience of the gradient calculation. A simple parameter analysis can be given as follows. Assuming that the volume of input is W 1 × H 1 × D 1 , after applying convolutions of k filters (without zero padding), the new volume of W 2 × H 2 × D 2 is:
W 2 = ( W 1 F 1 ) S 1 + 1
H 2 = ( H 1 F 2 ) S 2 + 1
D 2 = k
where F 1 and F 2 are the width and height of the filter cores, and S 1 and S 2 are the strides. One special characteristic of CNN is the shift invariance. However, although dramatically useful in image recognition, the invariance features pose an impediment for the recognition of Gabor representations that own a clear physical meaning. Any shift of a generally rectangular filter may distort the original TF distributions and confuse the CNN. To alleviate the restriction, three kinds of convolutional kernels are designed in accordance with the spectrograms of different resolutions, namely frequency filters, temporal filters, and texture filters.
As shown by the red rectangles in Figure 3, the temporal filters are designed for maps of higher time resolutions, and they share the same widths with the maps, while the frequency filters are suitable for higher frequency resolutions, and they are as high as the spectrograms. The texture filters are of medium size as supplements.
The designing principles of the special convolutional filters may be explained qualitatively with orthogonality, since the convolutional operation is basically a template matching process. The first step is to make the filter cover either a complete frequency axis or a whole time axis, thereby allowing for perfect frequency or temporal matching. Second, by assuming that the filter is also a kind of special signal, it is assumed that orthogonality exists between the unfolding direction of the filters and the spectrograms, to maximize the diversity in a filter as much as possible. Following this idea, the frequency filters are applied to the Gabor maps of higher frequency resolutions, while the temporal filters are applied to the higher time resolutions. Another advantage of the frequency and temporal filters refers to Equations (5) and (6). The outputs of the frequency and temporal filters are only one-dimensional, which dramatically decreases the model’s size.

3.3. Architecture of the Proposed Model

As has been mentioned before, TF maps of different resolutions can reveal different rhythms in the 2-dimensional space. It is hypothesized that it would be of great benefit to explicitly provide TF maps of different resolutions as inputs to a deep model. This combines the multiple views with deep learning’s ability to extract intrinsic properties from raw inputs, so that the discriminative features could be learned while maintaining explanatory factors with physical meaning.
Accordingly, the proposed multi-resolution network is summarized in Figure 4. It mainly contains three parts, namely, single-resolution embedding, multi-resolution fusion, and multi-sensor fusion. Here, the first part is realized by three subnetworks; the second part is a fully-connected neural network attached at the end of the subnetworks, and Long Short-Term Memory (LSTM) [40] is implemented for multi-sensor fusion.
1. Single-resolution embeddings with the subnetworks
Each branch of the subnetworks has its own micro-structure, which is basically a CNN with the convolutional operation. The main differences lie in the specified designed frequency filters, temporal filters, and texture filters. After the convolutional filter, an activation function, a pooling layer, and a dropout layer are arranged hierarchically as a basic building block, which repeats itself twice before connecting to a flattening layer. Finally, a fully-connected layer is attached at the end of each subnetwork to generate an embedding vector as the intermediate feature space.
2. Multi-resolution information fusion using the fully-connected neural network
The fully connected neural network is a normal multi-layer perceptron (MLP) containing two layers of 100 units and 50 units, respectively, that are attached after the concatenation of the three subnetworks. The model automatically chooses the best filters by assigning them higher weights, and the information of different resolutions is merged to offer comprehensive representations of compact forms.
3. Multi-sensor information fusion using the LSTM
The propagation and reflections of UHF signals in the GIS can significantly influence the characteristics of the signals, especially at the L corner for the inconsistency of the wave impedance [41]. Therefore, the characteristics of sensors of different positions can reveal different patterns. By fusing the information from different sensors, it is possible to obtain a more comprehensive understanding of the PD patterns.
Recurrent neural network (RNNs) [42] is the most popular choice for sequence mining tasks such as speech recognition, natural language translation, and video recognition. As shown in Figure 5, in each time step t , the hidden state h t is updated by a function of the last hidden state h t 1 and the new input x t :
  h t = t a n h ( W x t + U h t 1 + b )
where the hidden state h t is a d-dimensional vector, and b is the bias. W and U are the weights. Tanh is the activation function. LSTM is a branch of RNN, which introduces a memory cell to preserve the information for long sequences, and relieves the exploding or vanishing gradients problem in the original RNN [40]. The basic unit of LSTM can be explained as a collection of a block input x t , a memory cell c t , and a hidden state h t , controlled by an input gate i t , a forget gate f t , and an output gate o t as shown in Figure 5, with the following relationships:
i t = σ ( W ( i ) x t + U ( i ) h t 1 + b i )
  f t = σ ( W ( f ) x t + U ( f ) h t 1 + b f )  
  o t = σ ( W ( o ) x t + U ( o ) h t 1 + b o )  
  u t = t a n h ( W ( u ) x t + U ( u ) h t 1 + b u )  
  c t = i t u t + f t c t 1  
  h t = o t t a n h ( c t )  
where σ is the sigmoid function and donates the point multiplication. W ( i ) , W ( f ) , W ( o ) , W ( u ) , U ( i ) , U ( f ) , U ( o ) , and U ( u ) are the weights. b i , b f , b o , and b u are the biases. Intuitively, the input gate and the output gate control how much the old memories and new information are recombined as the new memory cell. The hidden state h t is a gated, partial view of the internal memory of the cell. By varying the gate values in different time steps, the LSTM can keep itself up-to-date while maintaining the most discriminable information acquired from different time steps. Finally, the features from different sensors are integrated for information fusion to give the final diagnosis results.
As the net structure is very complex, more sophisticated techniques are required to relieve the risk of overfitting. Details of the training process will be presented in the next part.

3.4. Model Training by Transfer Learning

Although the multi-resolution views and the multi-steam structures gain great potential for more comprehensive diagnosis, a larger number of parameters accompanies its bigger size, making the model more prone to overfitting. Thus, a transfer learning framework was proposed that classifies the training process into the three stages of single-resolution embedding, multi-resolution fusion, and multi-sensor fusion, in accordance with the network’s structure.
Transfer learning is a technique focusing on transferring knowledge gained from one problem to a different but related task [43]. The parameters of the target network are partly or all initialized from another network that is pretrained from the source task, while the nontransferred weights are randomly initialized. In most cases, transfer learning is done by weight transfer.
The usage of transfer learning was more flexible in this study, where the source and target problem use the same dataset. The key idea is relative independence among columns of different resolutions. The same dataset is much less prone to overfitting when it is applied to a smaller problem. Therefore, it is beneficial if the single-column is trained first independently. The knowledge is then transferred to the higher levels for more sophisticated targets. The details of the three learning stages are shown in Figure 6.
  • The single-column features extracted by each subnetwork are learned using manifold embedding.
  • The weights of the MLP attached at the end of the three subnetworks are randomly initialized, and the whole structure is fine-tuned.
  • The weights gained from step (2) are frozen as feature extractors. Only the weights of the LSTM are updated for multi-sensor fusion.

3.4.1. Single Column Embedding with Manifold Learning

It is worth noting that the subnetwork that embeds the high-dimensional data into a fixed vector is similar to a general dimensionality reduction problem that maps the high dimensional points onto a low dimensional manifold to make the similar points adjacent. Although some general dimensionality reduction frameworks exist [44,45], the same key point is to preserve the distances of the original vertex pairs. Given the data x 1 , , x U , the loss function parameterized by α and embedding function f is:
L o s s = i , j U L ( f ( x i , α ) f ( x j , α ) W i j )  
where W i j is the element of a pairwise distance matrix in a class-dependent or class-independent way. For example, Multidimensional scaling (MDS) [46] can be used in this pattern by defining W i j as the Euclidean distance. For the specified problem, the distances are defined in a class-dependent way as follows. W i j = 0 if i and j belong to the same class, and 1 otherwise.
Furthermore, to make the loss function learnable and more suitable for the subnetworks, a Siamese network structure [47] is implemented to realize the pairwise affinity calculation. A Siamese network is a special neural network with two symmetric branches, as shown in Figure 7. Each time, only one subnetwork is fabricated into the Siamese structure for pretraining with the following loss function:
  L o s s = { f ( x i , α ) f ( x j , α ) 2 i f   W i j = 0 m a x ( 0 , m f ( x i , α ) f ( x j , α ) 2 ) i f   W i j = 1
where m is the margin. The loss function is different from MDS in two ways. First, the distance is class-dependent. Second, the losses are chosen as the hinge losses for better generalization. After accomplishing the model construction, Adadelta optimizer is utilized for the weights update.

3.4.2. Transfer Learning for Cascaded Training

For the other two steps, the weights of the MLP that were attached at the end of the three subnetworks were randomly initialized first, and then another layer with the same node number of the classes was connected to make a complete single-sensor diagnosis model. After fine-tuning with Adadelta, the weights gained from step (2) were frozen, and they only acted as feature extractors for the information fusion of different sensors, while the weights of the LSTM were made to be trainable. The convergence speeds of the two steps were likely to be much faster compared with the first step, due to the existing class-dependent knowledge. In addition, some advanced training techniques such as early stopping and heuristic learning rate adaption were also utilized.

4. PD Laboratory Setup

For the purpose of the experimental study, a simulative L-shaped GIS tank with a coaxial cylindrical conductor was implemented to simulate the realistic equipment, as shown in Figure 8 and Figure 9a, with the central conductor diameter being 90 mm, and the enclosure tank diameter being 320 mm. The tank was filled with SF6 gas of 0.1 MPa to simulate the propagation medium.
Four planar equiangular spiral antennas (PESAs) with impedance transformers were installed inside the GIS chamber in the hand holes parallel to the axis of the GIS bus-bar, as shown in Figure 8 and Figure 9b. The PESA utilized in this study is an ultra-wideband antenna that can be installed internally in the GIS. Its outside and inside radii are 109 mm and 2 mm, respectively, with the substrate thickness being 1 mm. The other key parameters that can affect the performance of the PESA sensors include the relative dielectric constant, the spiral growth rate, and the rotation angle, which are chosen as 2.65, 0.364, and 3.5 π , respectively. In practice, the most commonly quoted parameter to evaluate the performance of the antenna is the reflection coefficient S11, which represents how much power is reflected from the antenna. The range of the bandwidth that satisfies S11 < −10 dB was 0.8 GHz to 3.5 GHz from the lower limit to the higher limit for the PESAs. Besides, the S11 parameter ranged between −8 dB to −6 dB from 0.2 GHz and 0.7 GHz, and close to −10 dB in the 0.7 GHz to 0.8 GHz range, which satisfied the measurement requirements of the UHF signals. The detailed S11 curve can be found in our previous study [48]. The geometrical dimensions of the GIS tank and the relative distances between the PESAs are also illustrated in Figure 8.
High voltage was introduced by an insulating bushing connected with a non-PD testing transformer (Xinyuan Electric, Yangzhou, China, YDTW-30/150) with an amplitude of 0–150 kV, and a capacity of 30 kVA. The four PESAs received the UHF waves successively, based on their locations. Furthermore, a four-channel digital oscilloscope (Tektronix, Beaverton, OR, USA, DPO7354C, 3.5 GHz, 40 GS/s) was employed to acquire and record the TPRD UHF patterns.
Five types of PDs were fabricated to simulate the various insulation defects, including the floating electrode, a metal protrusion on the conductor and the tank, surface contamination, and free metal particles. Besides, three relative angles between the defect position and the PESA antennas, which were 0°, 90°, and 180°, were integrated with the five defect types in the defect simulations.
As shown in Figure 10, the floating electrode defect was simulated by fixing two adjacent copper nuts to an insulated bolt that was attached on the conductor. The nuts were 3 mm away from the conductor, and the distance between the copper nuts was 1 mm. Two kinds of metal protrusions were replicated by a 30 mm needle adhered on the high-voltage conductor and tank, respectively. Surface contamination was replicated by a piece of aluminum foil (2 × 20 mm2) adhered to the surface of the spacer. Finally, the free particles defect was produced by connecting an insulation tube that contained some aluminum particles inside the conductor, with the length of the insulation tube being 20 mm. The first four types of defects were arranged at three relative angles between the defect positions and the sensor position for 0°, 90°, and 180°, as shown in Figure 8, generating 13 PD conditions in total (the free metal particles were only simulated at 90°). Figure 11 shows some typical acquisitions of the UHF signals within 100 ns, and their corresponding frequency spectra.

5. Experiment Evaluations

A detailed analysis was carried out to ascertain the model’s discriminatory capability. Both the relative angles and defect types were taken into consideration, including the floating electrode (0°, 90°, and 180°), the metal protrusion on the conductor (0°, 90°, and 180°), the metal protrusion on the tank (0°, 90°, and 180°), the surface contamination (0°, 90°, and 180°), and the free particles (90° only), thus creating 13 defect modes. There were 1386 samples in total, of which 20% were used for testing. Two cases were designed for recognition with the same dataset, namely, (1) the combined diagnosis of positions and defect types together, which had 13 classes in total, and (2) diagnosis of the defect types only, which has five classes only.

5.1. Implementation Details

First, the lengths of the analysis windows were chosen based on the frequency decline range and the time decline range described in Figure 2, and 50% of the max value was chosen as the threshold. Instead of using the ranges directly, their proportions in the whole analysis time axis and frequency axis were calculated for ease of comparison. The time range limit and frequency range limit were chosen as 100 ns and 3 GHz, respectively. The sampling rate was 10 GHz, indicating that 0.1 ns is added when the window length increases by 1. Figure 12 shows the variations of the percentages of the average time and frequency ranges. It is obvious that as the window length increased, the frequency resolution increases as well, while the time resolution showed an opposite trend. In order to retain enough resolution information for both the axis, 30% was selected as the threshold. It is observed that the window lengths were near 6 and 100 at this point. Therefore, the temporal window and frequency window were chosen as 6 and 100, respectively. Besides, the crossing point was obtained near 30, which was selected as the window for the texture channel.
Besides, in order to further reduce the input dimension, a bigger stride was used along the time axis for the frequency channel, while a bigger frequency stride was utilized for the temporal channel in STFT, which could be seen as a subsampling. A cut-off of 2 GHz was also utilized in the temporal channel and the texture channel, considering that the details in the high-frequency band were trivial in their diagnosis. Finally, the windows and the input shapes of the three subnetworks are summarized in Table 2. Considering the limited amount of data, the neural network architecture was wide, but it had only a few convolutional maps per column. The detailed net structures are also shown in Table 2. The three columns are the time column, frequency column, and texture column, respectively.
The learning algorithm setup was as follows: the Adadelta optimizer with a learning rate of 1.0 for the pretraining and MLP embedding stages, and a Rmsprop optimizer for LSTM with a learning rate of 0.001 and epsilon of 1 × 10−6. The pretraining was carried out for 1000, 800, and 400 iterations, with the patience being 100, 50, and 50. All networks were realized by Keras and NVIDIA 750TI GPUs (NVIDIA, Santa Clara, CA, USA). The tests averaged over five sets of initialization parameters were reported as the final results.

5.2. PD Pattern Recognition Results

5.2.1. Diagnosis Accuracies

The recognition accuracies of the two cases of (1) diagnosis of both the defect positions and types, and (2) the defect types only, are shown in Table 3 and Table 4. Overall, we achieved 97.51% testing accuracy for the recognition of both the angles and defect types, and 98.20% for the defect types only, based on the experimental dataset.
Furthermore, the diagnostic accuracies of single resolutions and single sensors with partial configurations are also presented as quantitative measures of the informativeness. The following points may be observed, based on the recognition accuracies.
  • Multi-resolution diagnosis leads to a smaller misclassification rate than by using only the single resolution information.
  • Sensor1 has the best accuracy among all the sensors with respect to the shortest distance to the defect location. The information loss can be quite significant after the L-shaped corner.
  • A higher frequency resolution is more valuable for single resolution diagnosis.
  • The multi-sensor combination is not certainly better than a single sensor without appropriate methods for the increasing the dimensionality and the confusing information carried by different sensors.

5.2.2. Loss Curves and Time Consumption

To diagnose the effectiveness of transfer learning, the loss curves and time consumptions of the single column training, multi-column training, and the LSTM training are shown in Figure 13 and Figure 14. It is obvious that during the pretraining stage, the convergence speed of the texture channel was the fastest, while the frequency channel ranked second, and the temporal channel was the slowest. However, the time consumption was the other way around. The texture channel called for the longest training time. This was attributed to the number of dot calculations for each iteration being much more than the other two kinds of kernels. Besides, both the iteration number and the time consumption of the second and third training stages were much less, compared with their previous steps. For the final LSTM information fusion stage, only less than 10 s was needed.
The results in Table 3 and Table 4, and Figure 13 and Figure 14 show the exhaustive role played by transfer learning. It is significant that during the three training stages, the diagnostic accuracy increased, while both the training iterations and time decreased dramatically, even with a larger network. This dynamic indicates that the knowledge learned from earlier stages was successfully transferred to a higher level with less proneness to overfitting.

5.2.3. Visualization of the Convolutional Filters

Neural networks are usually considered as black boxes. However, with the help of visualization methods, it is possible to obtain good qualitative interpretations of the black boxes of deep networks. Besides, visualization is also an effective way to diagnose the misclassifications and evaluate the degree of overfitting. In this study, the Activation Maximization [49] method was utilized to visualize the learned filters. The idea of Activation Maximization is quite simple: generating an image to maximize the activation of a hidden unit as:
  x * = argmax x   s . t .   x = ρ h i j ( θ ,   x )  
where x * is the visualization result, and θ donates the fixed neural network parameters. Thus, h i j ( θ ,   x ) is the activation values of unit j in layer i with fixed parameters θ and bounded norm ρ . For simplicity and the convenience of comparison, only the first layer of filters were visualized, as shown in Figure 15.
It is clear that the learned convolutional filters were quite smooth in space, thus indicating that the training was sufficient. The temporal filters were sensitive to pulses along the time axis, while the frequency filters showed clear patterns of different frequency combinations. Besides, the diversity of patterns was richer in the visualization results of the texture filters, where both the temporal and frequency patterns revealed their importance. It is also encouraging that both the temporal filters and frequency filters gained significant resolution power over multiple scales, as shown by the different densities of the lines.
Besides, a direct visualization of the outputs of the filters was helpful in illustrating the automatic feature extraction and the information compression of the filters. An illustrative example is shown in Figure 16. The outputs of the frequency filter and the temporal filter were only one-dimensional, while the texture filter transferred the two-dimensional input to a smaller feature map. That is also why the texture filter calls for the longest training time.

5.2.4. Visualization of the Learned Embedding Features

Furthermore, by using a t-Distributed Stochastic Neighbor Embedding (t-SNE) projection algorithm, the 50-dimensional features learned by the subnetworks were reduced to three dimensions, as shown in Figure 17 and Figure 18 for the cases of 13 and five classes, respectively. It was observed that the features learned automatically by the CNN of different defects fell into distinct clusters in the three-dimensional space, where each cluster showed clear manifold characteristics of a continuous distribution in the space.

5.3. Comparison with the Baselines

To further verify the effectiveness, some baselines were implemented. Instead of focusing on the specified methods, we first extracted the common framework utilized in the practice of UHF signal recognition, which were two-dimensionalization, matrix compression, decomposition, features extraction, feature selection (FS), and the final classifier. Second, by using typical models in certain steps, representative methods could be obtained. A description of some implemented techniques is as follows.
  • Two-dimensionalization: Wavelet spectrum and Hibert–Huang spectrum;
  • Matrix compression: NMF and 2DPCA;
  • Signal Decomposition: Wavelet transform and Hibert–Huang transform (HHT);
  • Features extraction: The extracted time and frequency (T&F) features include the max value, root mean square deviation, standard deviation, skewness, kurtosis, and the peak-to-peak value in the time domain. The frequency features include the mean frequency, frequency center, root mean square frequency, and the standard deviation frequency. Besides, the entropy is also calculated.
  • Feature selection: ExtraTrees Classifier and LinearSVC feature selection;
  • Classifier: Finally, both the SVM and DNN of dense layers are chosen as the final classifiers, due to their high representativeness in engineering practice.
Through the combination of the different techniques, 18 diagnosis models are summarized in Table 5 as the baselines, including both simple methods and complicated flows. However, the recognition accuracies are not certainly proportional to the complexity of the methods.
All of the hyperparameters are selected based on cross-validation to the best of our knowledge. The final recognition accuracies of (1) combined the diagnosis of positions and defect types together, 13 classes, and (2) diagnosis of the defect types only; five classes are shown in Table 6 and Table 7, respectively.
The accuracy comparisons may provide some insights into the fundamental differences between the traditional feature-based methods and the deep learning approaches, as well as some essential flaws of the feature engineering methods.
  • The misclassification rate of the proposed multi-resolution CNN clearly indicates its superior capability compared with the baseline methods. STFT’s performance is slightly better than the Wavelet spectrum in certain resolutions, and much better than the HHT map.
  • The best performance in the baseline methods is gained by the combination of STFT, NMF, and SVM, which is 92.09% for the 13 classes, and 93.88% for the five classes. Firstly, it is observed that NMF slightly outperforms 2DPCA, which may be explained as follows. In the experiments, the NMF gains the best performance with four dimensions, compared with the 12 dimensions of the 2DPCA, indicating that the redundancy is huge in the STFT maps. Thus, searching for a global projection in the redundancy data may be not a good idea. Second, the features extraction usually does not increase the model performance, which reflects the eternal contradiction among the learning ability of the model, the input dimensions, the distribution of the data, and the discernibility of the features.
  • The performances of SVM and DNN are similar when utilizing the same input features. Therefore, for traditional feature-based recognition methods, the recognition accuracy relies much on the discriminative features.
  • Similar to the conclusion drawn from Table 3 and Table 4, the performance of multiple sensors is not certainly better than the single sensor, especially when the nearest single sensor can gain good diagnostic accuracy.

6. Discussion

Although the deep networks have gained tremendous performance in many areas, low interpretability is always an obstacle to applying them in some highly regulated environments, due to the uncontrollability that comes with the unclear mechanism. Sometimes, the training of the neural networks becomes a trial-and-error problem. In recent years, a great number of studies have focused on the interpretability of deep learning, and great progress has been made in both the theories and experiments. In the discussion, instead of pursuing a general solution, we only concentrate attention on the special CNN structures that are implemented here, and try to explain its effectiveness from the perspective of manifold learning.
A comprehensive comparison of the similarities between the filters learned by the CNN and the matrix projection-based methods, such as 2DPCA, is presented to provide some insights into the advantages of deep learning methods, compared with the feature extraction solutions. It is quite interesting to notice the similarity between the projection operation in 2DPCA and the convolutions in CNN. In 2DPCA, an image A of m × n matrix is projected into a lower dimension by:
Y k = A X k
where X k is a vector of n dimensions, and k is the index. A visualization example of the 2DPCA’s compressed result of the STFT spectrum is shown in Figure 19. It is easy to notice its similarity with the outputs in Figure 16, while the information concentration in Figure 16 is better.
Moreover, it is also possible to reconstruct the original image from the compressed information by:
A ^ = k = 1 d Y k X k T
where d is the compression dimension. Therefore, in 2DPCA, it is assumed that any input can be approximately reconstructed as a weighted sum of smaller collections. In summary, the key points in 2DPCA include:
(1)
The projection operation for compressing,
(2)
The assumption that complex images are composed of eigenimages.
It is quite interesting that similar characteristics are shared by the proposed temporal and frequency filters.
(1)
The dot production between the frequency filters and the spectrum, and the similar operation between the temporal filters and the spectrum can be seen as a projection operation,
(2)
In the CNN, it is also assumed that images are formed from elements of the lower level, like pixels to more composite representations [28].
Therefore, a common framework that distinguishes different categories by matrix projection may be extracted to unify the deep structure and 2DPCA. The CNN and 2DPCA both try to classify different categories by using matrix projections, and the main differences lies in their supervision methods, where the CNN is supervised by labels, while 2DPCA is supervised by the inner product in a high-dimensional space. Furthermore, a special case of NMF that aims at reducing the reconstruction errors by projecting could also be defined to fit this framework. The main specialty lies in that the loss function is defined at the single sample, and it uses the mean square error between the original images and the reconstructed ones.
Moreover, based on this framework, the 2DPCA can also be reformulated and approximated as a special case of the CNN with the following setups: (1) Use only a single layer of convolutions, and the width of the convolutional kernels should be restricted at one. (2) The loss function is defined as the sum of the pairwise Euclidean distance losses to force the model to maintain its relative relations in higher space. Some new algorithms may be developed based on this idea.
Therefore, the superiority of the proposed may be explained as follows. Based on the same framework, only the CNN is goal-oriented at the final classifying task, while the 2DPCA and NMF aim at the other specified goals. It is usually hard to determine the effectiveness of these specified goals until obtaining the final results. Is it better to maintain the global information as 2DPCA, or to keep the local completeness as in NMF? In addition, the multiple projections also equip the CNN with a better learning ability to handle more complex problems.

7. Conclusions

Inspired by the phenomenon that the PD spectrograms of different resolutions can capture the information of different patterns, both multi-resolution and multi-sensor fusion algorithms are proposed. They are realized by a multi-column deep CNN characterized by specified filters and a sequential LSTM network. A multi-stage transfer learning framework is also presented to train the model with limited data. Several conclusions may be drawn based on the detailed analysis.
  • The proposed multi-resolution model gains accuracies of 97.51% and 98.20% on the tasks of diagnosing (1) the positions and defect types together, and (2) the defect types only; thus indicating its clear superiority compared with the baselines.
  • The loss curves, time consumptions, and diagnostic accuracies show the effectiveness of transfer learning, which successfully transfers knowledge from a lower level to a higher level with less proneness to overfitting.
  • The comparisons with the baseline methods reveal several fundamental flaws of the feature based methods, such as the difficulties in choosing the most suitable approach, and the curse of dimensionality when encountering too many features from multiple sensors.
  • A matrix projection framework is proposed to enhance the interpretability of the proposed deep network structure.

Author Contributions

G.L. and X.W. conceived and designed the experiments; X.L. provided the experimental data; G.L. wrote the paper; A.Y. revised the contents and reviewed the manuscript; M.R. provided theoretical guidance and supported the study.

Funding

This work was supported by the National Natural Science Foundation of China (No. 51521065).

Acknowledgments

The author gratefully acknowledges the support of the K. C. Wong Education Foundation.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Metwally, I.A. Technology progress in high-voltage gas-insulated substations. IEEE Potentials 2010, 29, 25–32. [Google Scholar] [CrossRef]
  2. Riechert, U.; Holaus, W. Ultra high-voltage gas-insulated switchgear—A technology milestone. Eur. Trans. Electr. Power 2012, 22, 60–82. [Google Scholar] [CrossRef]
  3. Schichler, U.; Koltunowicz, W.; Endo, F.; Feser, K.; Giboulet, A.; Girodet, A.; Hama, H.; Hampton, B.; Kranz, H.-G.; Lopez-Roldan, J. Risk assessment on defects in GIS based on PD diagnostics. IEEE Trans. Dielectr. Electr. Insul. 2013, 20, 2165–2172. [Google Scholar] [CrossRef]
  4. Stone, G.C. Partial discharge diagnostics and electrical equipment insulation condition assessment. IEEE Trans. Dielectr. Electr. Insul. 2005, 12, 891–904. [Google Scholar] [CrossRef]
  5. Bartnikas, R. Partial discharges. Their mechanism, detection and measurement. IEEE Trans. Dielectr. Electr. Insul. 2002, 9, 763–808. [Google Scholar] [CrossRef]
  6. Wu, M.; Cao, H.; Cao, J.; Nguyen, H.L. An overview of state-of-the-art partial discharge analysis techniques for condition monitoring. IEEE Electr. Insul. Mag. 2015, 31, 22–35. [Google Scholar] [CrossRef]
  7. Kurrer, R.; Feser, K. The application of ultra-high-frequency partial discharge measurements to gas-insulated substations. IEEE Trans. Power Deliv. 1998, 13, 777–782. [Google Scholar] [CrossRef]
  8. High Voltage Test Techniques—Measurement of Partial Discharges by Electromagnetic and Acoustic Methods; IEC TS 62478:2016; BSI: London, UK, 2016.
  9. Umamaheswari, R.; Sarathi, R. Identification of partial discharges in gas-insulated switchgear by ultra-high-frequency technique and classification by adopting multi-class support vector machines. Electr. Power Compon. Syst. 2011, 39, 1577–1595. [Google Scholar] [CrossRef]
  10. Darabad, V.; Vakilian, M.; Blackburn, T.; Phung, B. An efficient PD data mining method for power transformer defect models using SOM technique. Int. J. Electr. Power Energy Syst. 2015, 71, 373–382. [Google Scholar] [CrossRef]
  11. Mas’Ud, A.A.; Stewart, B.; McMeekin, S. Application of an ensemble neural network for classifying partial discharge patterns. Electr. Power Syst. Res. 2014, 110, 154–162. [Google Scholar] [CrossRef]
  12. Evagorou, D.; Kyprianou, A.; Lewin, P.; Stavrou, A.; Efthymiou, V.; Metaxas, A.; Georghiou, G. Feature extraction of partial discharge signals using the wavelet packet transform and classification with a probabilistic neural network. IET Sci. Meas. Technol. 2010, 4, 177–192. [Google Scholar] [CrossRef]
  13. Wang, K.; Li, J.; Zhang, S.; Liao, R.; Wu, F.; Yang, L.; Li, J.; Grzybowski, S.; Yan, J. A hybrid algorithm based on s transform and affinity propagation clustering for separation of two simultaneously artificial partial discharge sources. IEEE Trans. Dielectr. Electr. Insul. 2015, 22, 1042–1060. [Google Scholar] [CrossRef]
  14. Zhu, M.-X.; Xue, J.-Y.; Zhang, J.-N.; Li, Y.; Deng, J.-B.; Mu, H.-B.; Zhang, G.-J.; Shao, X.-J.; Liu, X.-W. Classification and separation of partial discharge ultra-high-frequency signals in a 252 kV gas insulated substation by using cumulative energy technique. IET Sci. Meas. Technol. 2016, 10, 316–326. [Google Scholar] [CrossRef]
  15. Li, L.; Tang, J.; Liu, Y. Partial discharge recognition in gas insulated switchgear based on multi-information fusion. IEEE Trans. Dielectr. Electr. Insul. 2015, 22, 1080–1087. [Google Scholar] [CrossRef]
  16. Albarracín, R.; Robles, G.; Martinez-Tarifa, J.M.; Ardila-Rey, J. Separation of sources in radiofrequency measurements of partial discharges using time–power ratio maps. ISA Trans. 2015, 58, 389–397. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Li, J.; Jiang, T.; Harrison, R.F.; Grzybowski, S. Recognition of ultra high frequency partial discharge signals using multi-scale features. IEEE Trans. Dielectr. Electr. Insul. 2012, 19, 1412–1420. [Google Scholar] [CrossRef]
  18. Wang, X.; Li, X.; Rong, M.; Xie, D.; Ding, D.; Wang, Z. UHF Signal Processing and Pattern Recognition of Partial Discharge in Gas-Insulated Switchgear Using Chromatic Methodology. Sensors 2017, 17, 177. [Google Scholar] [CrossRef] [PubMed]
  19. Gu, F.-C.; Chang, H.-C.; Chen, F.-H.; Kuo, C.-C.; Hsu, C.-H. Application of the Hilbert-Huang transform with fractal feature enhancement on partial discharge recognition of power cable joints. IET Sci. Meas. Technol. 2012, 6, 440–448. [Google Scholar] [CrossRef]
  20. Dai, D.; Wang, X.; Long, J.; Tian, M.; Zhu, G.; Zhang, J. Feature extraction of GIS partial discharge signal based on S-transform and singular value decomposition. IET Sci. Meas. Technol. 2016, 11, 186–193. [Google Scholar] [CrossRef]
  21. Majidi, M.; Fadali, M.S.; Etezadi-Amoli, M.; Oskuoee, M. Partial discharge pattern recognition via sparse representation and ANN. IEEE Trans. Dielectr. Electr. Insul. 2015, 22, 1061–1070. [Google Scholar] [CrossRef]
  22. Khan, Y. Partial discharge pattern analysis using PCA and back-propagation artificial neural network for the estimation of size and position of metallic particle adhering to spacer in GIS. Electr. Eng. 2016, 98, 29–42. [Google Scholar] [CrossRef]
  23. Li, X.; Wang, X.; Yang, A.; Xie, D.; Ding, D.; Rong, M. Propogation characteristics of PD-induced UHF signal in 126 kV GIS with three-phase construction based on time–frequency analysis. IET Sci. Meas. Technol. 2016, 10, 805–812. [Google Scholar] [CrossRef]
  24. Auger, F.; Flandrin, P.; Gonçalvès, P.; Lemoine, O. Time-Frequency Toolbox; CNRS France: Paris, France; Rice University: Houston, TX, USA, 1996. [Google Scholar]
  25. Zhang, S.; Li, C.; Wang, K.; Li, J.; Liao, R.; Zhou, T.; Zhang, Y. Improving recognition accuracy of partial discharge patterns by image-oriented feature extraction and selection technique. IEEE Trans. Dielectr. Electr. Insul. 2016, 23, 1076–1087. [Google Scholar] [CrossRef]
  26. Wang, K.; Liao, R.; Yang, L.; Li, J.; Grzybowski, S.; Hao, J. Optimal features selected by NSGA-II for partial discharge pulses separation based on time-frequency representation and matrix decomposition. IEEE Trans. Dielectr. Electr. Insul. 2013, 20, 825–838. [Google Scholar] [CrossRef]
  27. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436. [Google Scholar] [CrossRef] [PubMed]
  28. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. In Proceedings of the Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–8 December 2012; pp. 1097–1105. [Google Scholar]
  29. Li, G.; Rong, M.; Wang, X.; Li, X.; Li, Y. Partial discharge patterns recognition with deep Convolutional Neural Networks. In Proceedings of the Condition Monitoring and Diagnosis, Xi’an, China, 25–28 September 2016; pp. 324–327. [Google Scholar]
  30. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  31. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  32. Chollet, F. Xception: Deep Learning With Depthwise Separable Convolutions. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  33. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated residual transformations for deep neural networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5987–5995. [Google Scholar]
  34. Shao, L.; Wu, D.; Li, X. Learning Deep and Wide: A Spectral Method for Learning Deep Networks. IEEE Trans. Neural Netw. Learn. Syst. 2017, 25, 2303–2308. [Google Scholar] [CrossRef] [PubMed]
  35. Dieleman, S.; Schrauwen, B. Multiscale approaches to music audio feature learning. In Proceedings of the 14th International Society for Music Information Retrieval Conference (ISMIR-2013), Pontifícia Universidade Católica do Paraná, Curitiba, Brazil, 4–8 November 2013; pp. 116–121. [Google Scholar]
  36. Mathieu, M.; Couprie, C.; LeCun, Y. Deep multi-scale video prediction beyond mean square error. arXiv Preprint, 2015; arXiv:1511.05440. [Google Scholar]
  37. Takahashi, N.; Mitsufuji, Y. Multi-Scale multi-band densenets for audio source separation. In Proceedings of the 2017 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA), New Paltz, NY, USA, 15–18 October 2017; pp. 21–25. [Google Scholar]
  38. Tang, Y.; Mohamed, A.-R. Multiresolution deep belief networks. In Proceedings of the Artificial Intelligence and Statistics, La Palma, Canary Islands, Spain, 21–23 April 2012; pp. 1203–1211. [Google Scholar]
  39. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [PubMed]
  40. Greff, K.; Srivastava, R.K.; Koutník, J.; Steunebrink, B.R.; Schmidhuber, J. LSTM: A Search Space Odyssey. IEEE Trans. Neural Netwo. Learn. Syst. 2015, 28, 2222–2232. [Google Scholar] [CrossRef] [PubMed]
  41. Li, X.; Wang, X.; Xie, D.; Wang, X.; Yang, A.; Rong, M. Time–frequency analysis of PD-induced UHF signal in GIS and feature extraction using invariant moments. IET Sci. Meas. Technol. 2018, 12, 169–175. [Google Scholar] [CrossRef]
  42. Mikolov, T.; Karafiát, M.; Burget, L.; Cernocký, J.; Khudanpur, S. Recurrent neural network based language model. In Proceedings of the INTERSPEECH 2010—11th Annual of Conference of the International Speech Communication Association, Makuhari, Chiba, Japan, 26–30 September 2010; pp. 1045–1048. [Google Scholar]
  43. Pan, S.J.; Yang, Q. A Survey on Transfer Learning. IEEE Trans. Knowl. Data Eng. 2010, 22, 1345–1359. [Google Scholar] [CrossRef] [Green Version]
  44. Weston, J.; Ratle, F.; Mobahi, H.; Collobert, R. Deep learning via semi-supervised embedding. In Neural Networks: Tricks of the Trade; Springer: Berlin/Heidelberg, Germany, 2012; pp. 639–655. [Google Scholar]
  45. Bengio, Y.; Paiement, J.F.; Vincent, P.; Delalleau, O.; Roux, N.L.; Ouimet, M. Out-of-sample extensions for LLE, IsoMap, MDS, Eigenmaps, and Spectral Clustering. Adv. Neural Inf. Process. Syst. 2004, 16, 177–184. [Google Scholar]
  46. Williams, C.K.I. On a Connection between Kernel PCA and Metric Multidimensional Scaling. Mach. Learn. 2002, 46, 11–19. [Google Scholar] [CrossRef] [Green Version]
  47. Chopra, S.; Hadsell, R.; Lecun, Y. Learning a Similarity Metric Discriminatively, with Application to Face Verification. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005; pp. 539–546. [Google Scholar]
  48. Li, T.; Rong, M.; Zheng, C.; Wang, X. Development simulation and experiment study on UHF partial discharge sensor in GIS. IEEE Trans. Dielectr. Electr. Insul. 2012, 19, 1421–1430. [Google Scholar] [CrossRef]
  49. Erhan, D.; Bengio, Y.; Courville, A.; Vincent, P. Visualizing Higher-Layer Features of a Deep Network; Technical Report 1341; University of Montreal: Montreal, QC, Canada, 2009; pp. 1–13. [Google Scholar]
Figure 1. Gabor representations of different time and frequency resolutions: (a) windows of different lengths; (b) spectrogram of higher time resolution; (c) spectrogram of higher frequency resolution; (d) spectrum of medium resolution.
Figure 1. Gabor representations of different time and frequency resolutions: (a) windows of different lengths; (b) spectrogram of higher time resolution; (c) spectrogram of higher frequency resolution; (d) spectrum of medium resolution.
Sensors 18 03512 g001
Figure 2. Schematic diagram of the time range and frequency range before reaching the threshold.
Figure 2. Schematic diagram of the time range and frequency range before reaching the threshold.
Sensors 18 03512 g002
Figure 3. Schematic diagram of the different convolutional filters and their corresponding outputs: (a) temporal filter; (b) frequency filter; (c) texture filter.
Figure 3. Schematic diagram of the different convolutional filters and their corresponding outputs: (a) temporal filter; (b) frequency filter; (c) texture filter.
Sensors 18 03512 g003
Figure 4. Structure of the multi-resolution CNN and the LSTM fusion.
Figure 4. Structure of the multi-resolution CNN and the LSTM fusion.
Sensors 18 03512 g004
Figure 5. Architectures of LSTM and a general RNN.
Figure 5. Architectures of LSTM and a general RNN.
Sensors 18 03512 g005
Figure 6. Three stages of the knowledge transfer process.
Figure 6. Three stages of the knowledge transfer process.
Sensors 18 03512 g006
Figure 7. Siamese network for embedding the feature maps into a constant vector.
Figure 7. Siamese network for embedding the feature maps into a constant vector.
Sensors 18 03512 g007
Figure 8. Schematic diagram of the GIS tank, PESA sensors, and the experimental circuit.
Figure 8. Schematic diagram of the GIS tank, PESA sensors, and the experimental circuit.
Sensors 18 03512 g008
Figure 9. GIS experimental platform: (a) GIS tank and the positions of the UHF sensors; (b) the installation of the UHF sensors inside the tank.
Figure 9. GIS experimental platform: (a) GIS tank and the positions of the UHF sensors; (b) the installation of the UHF sensors inside the tank.
Sensors 18 03512 g009
Figure 10. Typical PD defect patterns: (a) floating electrode; (b) metal protrusion on the conductor; (c) metal protrusion on the tank; (d) surface contamination; (e) free metal particles.
Figure 10. Typical PD defect patterns: (a) floating electrode; (b) metal protrusion on the conductor; (c) metal protrusion on the tank; (d) surface contamination; (e) free metal particles.
Sensors 18 03512 g010
Figure 11. Typical UHF waveforms: (a) UHF signals of five different defect types; (b) UHF signals acquired from the four channels; (c) UHF signals of the three relative angles.
Figure 11. Typical UHF waveforms: (a) UHF signals of five different defect types; (b) UHF signals acquired from the four channels; (c) UHF signals of the three relative angles.
Sensors 18 03512 g011
Figure 12. Percentages of the time and frequency decline ranges along the window lengths.
Figure 12. Percentages of the time and frequency decline ranges along the window lengths.
Sensors 18 03512 g012
Figure 13. Loss curves of the three stages of the transfer learning: (a) The defect types and positions; (b) the defect types only.
Figure 13. Loss curves of the three stages of the transfer learning: (a) The defect types and positions; (b) the defect types only.
Sensors 18 03512 g013
Figure 14. Time consumptions of the three stages of the transfer learning: (a) The defect types and positions; (b) the defect types only.
Figure 14. Time consumptions of the three stages of the transfer learning: (a) The defect types and positions; (b) the defect types only.
Sensors 18 03512 g014
Figure 15. Visualization of the convolutional filters: (a) temporal filters; (b) frequency filters; (c) texture filters.
Figure 15. Visualization of the convolutional filters: (a) temporal filters; (b) frequency filters; (c) texture filters.
Sensors 18 03512 g015
Figure 16. The outputs of the first layer of convolutional filters. (a) The output of the temporal filter; (b) the output of the frequency filter; (c) the output of the texture filter.
Figure 16. The outputs of the first layer of convolutional filters. (a) The output of the temporal filter; (b) the output of the frequency filter; (c) the output of the texture filter.
Sensors 18 03512 g016
Figure 17. Visualization of the learned features for both the defects and positions: (a) Temporal channel; (b) frequency channel; (c) texture channel.
Figure 17. Visualization of the learned features for both the defects and positions: (a) Temporal channel; (b) frequency channel; (c) texture channel.
Sensors 18 03512 g017
Figure 18. Visualization of the learned features for the defect types only: (a) Temporal channel; (b) frequency channel; (c) texture channel.
Figure 18. Visualization of the learned features for the defect types only: (a) Temporal channel; (b) frequency channel; (c) texture channel.
Sensors 18 03512 g018
Figure 19. Visualization of the compressed results of the 2DPCA.
Figure 19. Visualization of the compressed results of the 2DPCA.
Sensors 18 03512 g019
Table 1. Summary of the relevant features and classifiers that have been used for partial discharge (PD) recognition.
Table 1. Summary of the relevant features and classifiers that have been used for partial discharge (PD) recognition.
ReferencePD TypesFeature ExtractionClassifierRecognition Accuracy
Umamaheswari and Sarathi [9]Four types of artificial insulation defectsPartial power analysisSVMAverage of 95.25%
Darabad et al. [10]Ten types of PD sourcesTexture features + PCASOMGrouped data visualization
Mas’ud et al. [11]Six PD fault geometries in oil, air, and poly-ethylene-terephthalateStatistical parameters from the PRPD patternsENNAverage 95%
Evagorou et al. [12]Four types of artificial PDs in oil and air.Wavelet packet transform + statistical parametersPNN97.49%, 91.9%, 100%, and 99.8%
Wang et al. [13]Four types of artificial defect models in the oil and air.S-transformAffinity propagation clustering99.67%
Li et al. [14]Four kinds of typical defects in a 252 kV GISCumulative energy in time and frequency domain + mathematical morphological gradientFuzzy maximum-likelihood algorithm98%
Li et al. [15]Four kinds of defects in GISStatistical parameters of both the TRPD and PRPD patternsDempster–Shafer evidence theory + Neural network97.25%
Albarracín et al. [16]Separation of PDs with electromagnetic noisesPower ratios and the signal timesGrouped data visualizationThe separation criteria was given
Li et al. [17]Six types of artificial insulation defect models in the oil, air, and paper-fiber interface.Wavelet packages + fractal features + Linear discriminant analysisFinding the closest centroid99.4, 94.5, 99.4, 91.9, 87.5, and 97.7%, respectively.
Wang et al. [18]Three typical PD defect patterns in a 252 kV GISFourier transform + Chromatic methodologySVM86.67%
Gu et al. [19]Three common defect types of 25 kV cross-linked polyethylene (XLPE) power cableHHT + Fractal parametersNN100%, when 5% random white noise
Dai et al. [20]Four kinds of artificial defects in GISS-transform + Singular value decomposition (SVD)SVM98.33%
Majidi et al. [21]Air voids with dimensions of 1, 1.5, and 2 mm1-norm, 2-norm, and infinity-norm of the statistical featuresSparse representation vector99.7%, 92.9%, 94.0%, and 81.6% for four different scenarios
Khan et al. [22]Parallel-plane stainless steel electrodes in SF6 with 10 different particle lengths and positionsStatistical features + PCANN88%
Table 2. Detailed network structure.
Table 2. Detailed network structure.
IndexWindowInput ShapeCNN StructuresMLPLSTM
Column 1630 × 50(Conv2 × 50−MaxPooling2 × 1−Dropout)−(Conv2 × 1−MaxPooling2 × 1−Dropout)−(Dense100−Dense50)Flatten–Concatenation–Dense100–Dense50LSTM with 16 inner nodes–Flatten–Dense50–Output
Column 2100150 × 36(Conv150 × 4−MaxPooling1 × 2−Dropout)–(Conv1 × 3–MaxPooling1 × 2−Dropout)–(Dense100−Dense50)
Column 330100 × 64(Conv50 × 4−MaxPooling2 × 2−Dropout)−(Conv2 × 2−MaxPooling2 × 2−Dropout)−(Dense100−Dense50)
Table 3. Diagnostic accuracies and the intermediate outputs for classifying both the positions and defect types.
Table 3. Diagnostic accuracies and the intermediate outputs for classifying both the positions and defect types.
IndexStructureSingle Sensor Diagnosis (%)Multi-Sensor (%)
Sensor1Sensor2Sensor3Sensor4
1Temporal column85.7286.9466.6271.9487.91
2Frequency column94.5793.5685.4383.7895.35
3Texture column94.4293.0584.7886.3093.70
4Multi-resolution96.0895.0791.0190.1897.51
Table 4. Diagnostic accuracies and the intermediate outputs for classifying the defect types only.
Table 4. Diagnostic accuracies and the intermediate outputs for classifying the defect types only.
IndexStructureSingle sensor Diagnosis (%)Multi-Sensor (%)
Sensor1Sensor2Sensor3Sensor4
1Temporal column89.5390.8681.2281.0494.53
2Frequency column95.5896.2990.1487.8196.51
3Texture column96.1595.0789.5387.4895.79
4Multi-resolution97.3797.0991.2990.7998.20
Table 5. Typical diagnosis models for comparisons.
Table 5. Typical diagnosis models for comparisons.
IndexSummaryTwo-Dimensionalization AnalysisFeature ExtractionFeature SelectionClassifier
Two-DimensionalizationMatrix CompressionDecompositionFeatures
1Map + CNNHHT Map----CNN
2Wavelet Map----CNN
3Raw input----YesSVM
4----YesDNN
5T&F features---T&FYesSVM
6---T&FYesDNN
7Wavelet + T&F--waveletT&FYesSVM
8--waveletT&FYesDNN
9HHT + T&F--HHTT&FYesSVM
10--HHTT&FYesDNN
11STFT + NMFSTFTNMF -YesSVM
12STFTNMF -YesDNN
13SFTF + NMF + T&FSTFTNMF T&FYesSVM
14STFTNMF T&FYesDNN
15SFTF + 2DPCASTFT2DPCA YesSVM
16STFT2DPCA YesDNN
17SFTF + 2DPCA + T&FSTFT2DPCA T&FYesSVM
18STFT2DPCA T&FYesDNN
Table 6. Diagnostic accuracies of the baseline methods for classifying both defect types and positions.
Table 6. Diagnostic accuracies of the baseline methods for classifying both defect types and positions.
IndexSummarySingle Sensor (%)Multi-Sensor (%)
Sensor1Sensor2Sensor3Sensor4
1HHT spectrum + CNN65.5861.7345.9448.4960.79
2Wavelet spectrum + CNN90.7291.8384.8284.4690.61
3FS + SVM89.2087.4167.2766.1866.18
4FS + DNN89.1786.5867.6662.9978.77
5T&F + FS + SVM55.0455.0441.7349.6472.30
6T&F + FS + DNN55.5856.7343.2046.3774.31
7Wavelet + T&F + FS + SVM61.8755.7647.4849.6476.61
8Wavelet + T&F + FS + DNN65.0758.0246.6250.5876.87
9HHT + T&F + FS + SVM37.3332.7322.6632.0145.68
10HHT + T&F + FS + DNN38.4136.7630.1433.4542.58
11STFT + NMF + FS + SVM90.2988.8572.6668.7092.09
12STFT + NMF + FS + DNN90.9390.5876.1271.0491.55
13STFT + NMF + T&F + FS + SVM77.7071.5856.8357.9185.97
14STFT + NMF + T&F + FS + DNN80.1472.9159.5758.0984.86
15STFT + 2DPCA + FS + SVM87.4184.5357.1955.3987.05
16STFT + 2DPCA + FS + DNN87.6284.1060.2855.3685.89
17STFT + 2DPCA + T&F + FS + SVM84.5376.2549.2849.6482.37
18STFT + 2DPCA + T&F + FS + DNN84.0676.4754.6049.7881.72
Table 7. Diagnostic accuracies of the baseline methods for classifying the defect types only.
Table 7. Diagnostic accuracies of the baseline methods for classifying the defect types only.
IndexSummarySingle Sensor (%)Multi-Sensor (%)
Sensor1Sensor2Sensor3Sensor4
1HHT Spectrum + CNN73.0274.160.6860.8676.47
2Wavelet Spectrum + CNN92.4193.8190.3289.8291.94
3FS + SVM92.4590.2973.3771.9481.29
4FS + DNN90.7289.2876.8076.1677.48
5T&F + FS + SVM64.3960.0752.5262.9581.29
6T&F + FS + DNN67.6368.7156.8766.2282.01
7Wavelet + T&F + FS + SVM68.3566.9157.5561.5182.73
8Wavelet + T&F + FS + DNN72.6067.6360.0765.0081.85
9HHT + T&F + FS + SVM48.9250.0042.8148.5656.83
10HHT + T&F + FS + DNN54.9354.1444.8253.5360.07
11STFT + NMF + FS + SVM91.0190.6582.3777.7093.88
12STFT + NMF + FS + DNN92.5990.7283.6776.7393.48
13STFT + NMF + T&F + FS + SVM79.5074.8268.3569.7888.49
14STFT + NMF + T&F + FS + DNN84.0378.1370.3668.7189.28
15STFT + 2DPCA + FS + SVM87.7687.7666.1967.6387.41
16STFT + 2DPCA + FS + DNN89.7186.0170.1471.9486.47
17STFT + 2DPCA + T&F + FS + SVM86.6980.5864.0363.3187.05
18STFT + 2DPCA + T&F + FS + DNN85.7982.4161.8365.3284.38

Share and Cite

MDPI and ACS Style

Li, G.; Wang, X.; Li, X.; Yang, A.; Rong, M. Partial Discharge Recognition with a Multi-Resolution Convolutional Neural Network. Sensors 2018, 18, 3512. https://doi.org/10.3390/s18103512

AMA Style

Li G, Wang X, Li X, Yang A, Rong M. Partial Discharge Recognition with a Multi-Resolution Convolutional Neural Network. Sensors. 2018; 18(10):3512. https://doi.org/10.3390/s18103512

Chicago/Turabian Style

Li, Gaoyang, Xiaohua Wang, Xi Li, Aijun Yang, and Mingzhe Rong. 2018. "Partial Discharge Recognition with a Multi-Resolution Convolutional Neural Network" Sensors 18, no. 10: 3512. https://doi.org/10.3390/s18103512

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop