Next Article in Journal
GHG Emissions Mitigation in the European Union Based on Labor Market Changes
Next Article in Special Issue
Editorial to the Special Issue “AI Applications to Power Systems”
Previous Article in Journal
Mitigating the Piston Effect in High-Speed Hyperloop Transportation: A Study on the Use of Aerofoils
Previous Article in Special Issue
Online Steady-State Security Awareness Using Cellular Computation Networks and Fuzzy Techniques
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimization Techniques for Mining Power Quality Data and Processing Unbalanced Datasets in Machine Learning Applications

by
Alvaro Furlani Bastos
* and
Surya Santoso
Department of Electrical and Computer Engineering, The University of Texas at Austin, Austin, TX 78712, USA
*
Author to whom correspondence should be addressed.
Energies 2021, 14(2), 463; https://doi.org/10.3390/en14020463
Submission received: 24 December 2020 / Revised: 12 January 2021 / Accepted: 14 January 2021 / Published: 16 January 2021
(This article belongs to the Special Issue AI Applications to Power Systems)

Abstract

:
In recent years, machine learning applications have received increasing interest from power system researchers. The successful performance of these applications is dependent on the availability of extensive and diverse datasets for the training and validation of machine learning frameworks. However, power systems operate at quasi-steady-state conditions for most of the time, and the measurements corresponding to these states provide limited novel knowledge for the development of machine learning applications. In this paper, a data mining approach based on optimization techniques is proposed for filtering root-mean-square (RMS) voltage profiles and identifying unusual measurements within triggerless power quality datasets. Then, datasets with equal representation between event and non-event observations are created so that machine learning algorithms can extract useful insights from the rare but important event observations. The proposed framework is demonstrated and validated with both synthetic signals and field data measurements.

1. Introduction

The application of machine learning algorithms has expanded noticeably in many fields in the last few decades, especially due to the increased power and reduced expense of computation, the growth of field data collection and the advent of novel techniques to process and analyze large datasets. This trend has also been observed in power systems, where most machine learning applications are related to distributed energy resources (such as solar, wind, and storage) and smart grid control. Such examples include the following: load and demand forecasts [1,2], electricity production forecasts [2], solar radiation forecasts [1], wind speed/power forecasts [3], automated control of smart grids [2], management of electric vehicle fleets [1], predictive maintenance [4], fault detection and location [3,5,6] and power quality disturbance classification [3].
Data for machine learning applications in power systems can be acquired from multiple sources. A common source of field data measurements is power quality monitors (PQM), which record instantaneous voltage and current waveforms with a high time resolution (hundreds of samples per cycle). The latest version of these devices allows the addition of precise and synchronized time stamps to the measured data, expanding the suitability of the recorded data to more advanced applications [7]. Traditionally, PQMs employ a limited set of triggering features to detect disturbances within the dataset and, once they have been detected, store a few waveform cycles as individual events [8]. More recently, however, a triggerless data acquisition approach has emerged, where all waveform samples are stored for further analysis.
The main advantage of this approach is that even inconspicuous disturbances are successfully captured [9]; on the other hand, triggerless PQMs generate voluminous datasets and require large data storage capabilities [10,11]. Further, most of the data correspond to the steady-state operation of the power system, whereas only a small part of the recorded data shows disturbances. In other words, the dataset is highly unbalanced, with the steady-state observations heavily outnumbering the disturbance observations, which, as will be discussed later, can cause performance deterioration in most machine learning algorithms. Thus, one of the focuses of this paper is dataset rebalancing, such that the disturbance and non-disturbance classes are equally represented in the input dataset prior to its use by a machine learning algorithm.

1.1. Disturbance Detection in Power System Datasets

Multiple techniques have been proposed over the years for detecting disturbances in PQM data, and they are broadly classified into two categories: in the first category, the trigger mechanism is based on the magnitude of a time series (e.g., overvoltage, overcurrent, signal rate of rise and root-mean-square (RMS) voltage variations) [12] or employs time–frequency and time–scale transformations to decompose the signal into several subbands (e.g., short-time Fourier transform and wavelet transform) [13,14]; the second category is composed of methods based on prominent signal residuals, which are obtained through time-varying mathematical models (e.g., autoregressive (AR) models and Kalman filters) or direct data comparison (e.g., point-by-point or cycle-by-cycle comparison) [15].
It has been shown that these techniques are effective for detecting conspicuous disturbances (i.e., cases where the underlying system event causes transients in the voltage and/or current waveforms) [16]. On the other hand, they are unable to detect most inconspicuous disturbances [17,18], hindering their suitability for the processing of triggerless PQM datasets. Further, they might be sensitive to harmonics, sampling frequency and other user-selected parameters (such as a mother wavelet for the detector based on a wavelet transform). These drawbacks often result in disturbances being missed by the detector, especially those that are very short and/or subtle.
Although waveforms collected by PQMs are valuable assets for power system analysis, these raw measurements might not directly provide useful information for disturbance identification and classification [12]. In fact, various power system events might not cause conspicuous disturbances in the PQM waveforms; instead, they are characterized by an abrupt step change in the RMS voltage profile. Common examples of power system events that belong to this category include capacitor switching de-energizing [19], transformer tap-changing, voltage regulator operation and switching of large loads [13].
Thus, RMS voltage step changes have been proposed as an alternative triggering feature to detect events (both conspicuous and inconspicuous) within PQM datasets [8,12]. This task, however, is complicated by the fact that the magnitude of these RMS voltage step changes is often quite small (even less than 0.5% of the nominal voltage). Moreover, the presence of rapidly varying fluctuations in an RMS voltage profile hinders the detection of RMS voltage step changes. Therefore, prior to being used in the disturbance identification process, the RMS voltage profile must be processed to remove those rapid voltage fluctuations. The desired output of this process is an RMS voltage profile with a high signal-to-noise ratio and sharp edges during the step changes [8], which is another focus of this paper.

1.2. Contributions

This paper proposes a framework for the detection of RMS voltage step changes and rebalancing of highly unbalanced PQM datasets. Its main contributions are as follows: (a) the proposal of a strategy for filtering RMS voltage profiles such that rapidly varying noise is removed or significantly attenuated, whilst preserving the steep edges of the RMS voltage profile caused by switching events; (b) the automatic detection of RMS voltage step changes in the filtered RMS voltage profile, so that both conspicuous and inconspicuous events within a PQM dataset are identified; and (c) the proposal of a framework for rebalancing highly unbalanced PQM datasets.
All optimization problems presented in this paper are implemented and solved in Pyomo [20].

1.3. Article Organization

The remainder of this paper is organized as follows. Section 2 discusses the effects of highly unbalanced datasets in machine learning training and proposes a strategy for rebalancing highly unbalanced PQM datasets. Section 3 presents a literature review on the filtering of RMS voltage profiles and detection of RMS voltage step changes. Section 4 describes the proposed approaches for RMS voltage profile filtering and dataset rebalancing, as well as presenting the PQM datasets analyzed throughout this paper. Section 5 demonstrates the performance of the proposed framework through field data measurements and Section 6 addresses some final considerations.

2. The Problem of Unbalanced Datasets in Machine Learning Training

The main goal of any machine learning algorithm is to learn patterns directly from the data through some computational methods, without relying on a predetermined physical model or some other strong assumptions about the data features. In general, the performance of these algorithms improves as the amount and variability of available samples increase [21,22]. The growth in popularity of machine learning applications is a direct consequence of the rise in big data, as most rule-based models are inadequate to extract insight from such large, complex and ever-changing datasets. Some real-world machine learning applications already in use include such diverse fields as the following [23]:
  • Computational finance, for credit scoring and algorithmic trading;
  • Image processing and computer vision, for face recognition, motion detection and object detection;
  • Computational biology, for tumor detection, drug discovery and DNA sequencing;
  • Energy production, for price and load forecasting;
  • Automotive, aerospace and manufacturing, for predictive maintenance;
  • Natural language processing.
Machine learning methods are broadly classified into two categories: supervised learning, where the algorithm tries to establish a mapping between input features and output targets so that it can be used to predict the output target for future input features; and unsupervised learning, where there is no output target and the goal is to group and interpret the data based only on its input features. As will become clear in the following discussion, the focus of this paper is on supervised learning—either in terms of classification (i.e., the output variable is categorical/discrete) or regression (i.e., the output variable is continuous).
A machine learning application is often divided into three stages—training, validation and testing—with the input dataset split into three corresponding subsets as well [21,22]. Figure 1 represents the general workflow of a typical machine learning application. First, the training set (which usually is the largest of the three subsets) is used to train a machine learning model. The performance of the resulting model is then evaluated using the validation set. If its performance is satisfactory with respect to some metric, the current model is considered as the final version of the machine learning model; otherwise, an iterative loop of successive training and validation stages is executed to incrementally improve the model’s predictive power until the desired performance is achieved. This training/validation loop consists of hyper-parameter tuning (if the selected algorithm has any hyper-parameters) or even the selection of an entirely different algorithm. Due to the large number of machine learning algorithms, this step involves some trial-and-error, as there is no one-size-fits-all approach in machine learning (i.e., there is no algorithm that outperforms all other counterpart algorithms for all types of application, datasets size and types of data or desired insights).
Once the final machine learning model has been selected, the test set is used to produce its performance metrics. It is important to emphasize that the test set should not overlap with the training and validation sets, as the goal in this evaluation step is to estimate the predictive power of the final model on samples that have not been used to fine-tune the model’s parameters.
In this paper, we focus on a pre-processing step to be employed prior to any of those three stages in an attempt to improve the performance of the machine learning algorithm. More specifically, our focus is on the handling and processing of highly unbalanced input datasets; i.e., cases in which the observations in the training dataset belonging to one class heavily outnumber the observations in the other class. A general overview of this pre-processing step is shown in Figure 2; the components of the dataset rebalancing block are detailed in Figure 5.
Highly unbalanced datasets in machine learning training might influence the model performance and often result in a phenomena called the accuracy paradox. This occurs when the accuracy measure simply reflects the underlying class distribution, rather than learning the actual patterns present in the dataset. Most standard machine learning algorithms are developed under the assumption that the class distributions are roughly balanced. When presented with unbalanced datasets, these algorithms fail to capture the effects of severe class distribution skewness [24], as well as experience difficulties learning the concepts related to the minority class [25].
For example, consider a binary classification problem where the training dataset is composed of 95% of observations for Class 1 and only 5% of observations for Class 2. Most of the machine learning algorithms tend to be biased toward the majority class. If an algorithm classifies a new observation based only on the majority class in the training set (Class 1 in this case), its accuracy would be 95%, which is an excellent value for most practical applications. This approach, however, does not take into account the features of each observation; i.e., there is no actual learning during the training stage, and the final machine learning model is likely to have low predictive accuracy on new observations.
The drawbacks caused by unbalanced datasets might be even worse than is apparent [26]. For example, consider the study presented in [27], where the goal is to predict voltages throughout a distribution network. Not surprisingly, most of the target values in the training dataset are around 1.0 pu, with only a few observations for which the target value is less than 0.95 pu or greater than 1.05 pu. However, prediction accuracy is more important for these scenarios with extreme voltage values (the minority class) than scenarios with voltages around 1.0 pu (the majority class). This difference in prediction accuracy importance is due to the fact that the scenarios with very low or very high voltages are those in which a voltage control device has to operate.
There are multiple practical examples in which class unbalance is quite common and even expected to occur. The minority class often represents rare but important events [25]. A well-known example is represented by the datasets of credit card transactions, where nearly all transactions were authorized by the card holder (not-fraud class), while only a few transactions belong to the fraud class. A similar situation is observed in power system measurements: most of the measurements correspond to steady-state conditions (non-event), while only a few of them are events.
Multiple strategies have been proposed for handling unbalanced datasets, including the following [24,28,29]:
  • Collect more data;
  • Explore alternative performance metrics, such as the confusion matrix, precision, recall, F-score, Cohen’s kappa and receiver operating characteristic (ROC) curves [30];
  • Resample the dataset (either through under-sampling or over-sampling, depending on the dataset’s initial size);
  • Generate synthetic observations;
  • Investigate penalized models, where additional costs are imposed on the misclassification of the minority class during training and a higher cost of prediction is associated with rarity [31];
  • Reconstruct the training dataset, where the minority observations are identified through anomaly or change detection.
This paper employs the resampling and change detection approaches to construct balanced training datasets. Given an RMS voltage profile, a training dataset is constructed as follows:
(1)
Partition the input profile into multiple equal-length segments and determine which contain significant changes in the RMS voltage levels; a significant change is defined as an RMS voltage step change greater than a pre-specified threshold, which will be introduced in later sections. Each one of these selected segments corresponds to one observation of the minority class (event) in the training dataset—let n E denote the number of such observations;
(2)
Among the segments without a significant change in the RMS voltage level (non-event), randomly select n E segments to form the majority class (non-event) in the training dataset.
Note that the minority and majority classes are used in the steps above only for consistency with the previous discussion. In fact, the newly created training dataset is evenly balanced between the two classes.

3. The State-of-the-Art

As mentioned in Section 1, this paper focuses on the detection of substantial changes in RMS voltage profiles, so that datasets with a more balanced ratio between events and non-events can be obtained for use in the training and validation stages of a machine learning application pipeline. The most straightforward method to detect such changes in an RMS voltage profile is based on RMS voltage gradients. There are also other alternative detectors proposed in the literature, and these are discussed below.

3.1. The RMS Voltage Gradient Profile Detection Approach

Let the vector V R n represent an RMS voltage profile with a one-sample time resolution; then, the corresponding RMS voltage gradient profile is defined as
Δ V k = V k V k p N , for k = p N + 1 , , n
where N is the number of waveform samples per cycle. The quantity p N controls which RMS voltage values are compared to each other; it is recommended to adopt p 2 [13] so that there is at least a one-cycle gap between the waveform samples used to compute V k and V k p N . Otherwise, the magnitude of the RMS voltage gradient might be smaller than the true magnitude of the step change when those sets of waveform samples contain a mix of both event and non-event data, possibly causing the event to be undetected [12]. On the other hand, adopting p 2 guarantees that any waveform transients lasting less than one cycle will have dissipated and that at least one value in the Δ V profile captures the true magnitude of the step change. The computation of the RMS voltage gradient profile is illustrated in Figure 3 for p = 2 .
In the RMS voltage gradient approach, an event is detected whenever the following condition is satisfied:
Δ V k > δ step , for k = p N + 1 , , n
where δ step is a pre-specified threshold for the triggering criteria. The chosen value for this threshold has great impacts on the detector’s performance, as unsuitable values might cause multiple false positives ( δ step is too small) or false negatives ( δ step is too large).
In this paper, δ step is selected based on well-known characteristics of power systems; more specifically, we consider switching events that cause the most subtle change in RMS voltage profiles, as described below:
  • Voltage regulators are devices that adjust the voltage level by changing the tap positions in an autotransformer. In general, they provide a −10% to +10% regulation range with 32 steps, where each step represents ±0.625% of the nominal voltage [32].
  • Switched capacitor banks cause voltage variations, the magnitudes of which depend on the capacitor bank size and the short-circuit capacity at the bank location. For practical scenarios, the voltage variation falls between 0.36% and 4% of the nominal voltage [19,32,33,34].
Based on this discussion, we adopted δ step = 0.0018 pu, which follows the rule of thumb of setting the threshold as half of the minimum-expected step change [35]. This detection technique has been shown to achieve high accuracy, especially in cases where the signal-to-noise ratio of the RMS voltage profile is high (i.e., low noise levels) [12,36]. On the other hand, this detector fails if the RMS voltage profile has high noise levels or it has not been properly filtered.

3.2. Alternative Standard Detector

In the 2015 update, the International Electrotechnical Commission (IEC) added the concept of a rapid voltage change (RVC) to one of its standards [14]. An RVC is defined as an abrupt transition between two RMS voltage values, and its detection is performed as follows:
1
Compute the arithmetic mean of the immediately preceding RMS voltage values:
V ¯ k = 1 2 f p = k 2 f + 1 k V k
where f is the system frequency (either 50 or 60 Hz).
2
Flag a new RMS voltage value as part of an RVC if it deviates from V ¯ k by more than a given threshold δ RVC :
V k V ¯ k > δ RVC Flag as RVC
The RVC threshold δ RVC is set by the user according to the desired application; the standard recommends considering values in the range of 0.01 pu to 0.06 pu. Due to the computation of arithmetic means, this detection approach behaves similarly to linear filtering, which, as discussed in the next section, has the drawback of blurring out the steep edges of the signal.

3.3. RMS Voltage Profile Filtering

The event detectors described previously can exhibit great performance degradation if the RMS voltage profiles are contaminated with noise. In the context of this paper, the following are factors that contribute to noise corruption:
  • Noise introduced by the measurement device;
  • Varying system frequency, which results in incorrect RMS voltage computations, as N waveform samples do not correspond to an integer number of cycles [36,37];
  • Small load variations, which create intermittent variations in the RMS voltage profile and have the potential to hinder the detection of the events of interest.
Thus, a low-pass filtering technique must be applied to the RMS voltage profiles as a pre-processing step [35]. Linear filters, such as a moving average filter, have been shown to be effective in removing or attenuating rapidly varying noise while preserving the slowly varying signal. However, they blur out any steep edges of the signal [8,38,39], such as RMS voltage step changes, making this type of filter unfit for applications based on the detection of switching events [8].
On the other hand, median filters are well-known as suitable options for signals that contain sharp edges [39,40]. The performance of median filters can be further improved through an iterated and multiscale filtering approach, where multiple median filters are applied sequentially from a fine scale (narrow window) to a coarse scale (wide window). The goal of this process is to increase the signal-to-noise ratio at each stage such that the advantages of median filtering can be leveraged at increasingly low noise levels [12,39]. Previous work has compared the performance of single-stage and three-stage median filters applied to RMS voltage profiles around capacitor switching instants. It has been shown that both filters successfully attenuate the signal noise while preserving the RMS step changes in most cases; however, the three-stage median filter provided a faster transition between the steady-state levels prior and posterior to the switching instant [12]. This study also presented scenarios in which median filtering (both single- and three-stage) fails; for example, if the signal varies linearly (i.e., not a constant value) immediately before the step change, median filtering is not able to accurately track the signal.

4. Methodology

As mentioned in Section 1, techniques for properly filtering RMS voltage profiles are one of the main contributions of this paper. This section describes the proposed filtering approach, which is demonstrated through test signals.

4.1. Problem Setup

This subsection presents the data analyzed in the paper (both field measurements and synthetic signals), as well as definitions and formulations that are used in later sections.

4.1.1. Data

Field Measurements
The field measurements analyzed in this study consisted of 28-min continuous power quality data (voltage and current waveforms) collected at the feeder head of a 25 kV, 60 Hz radial distribution system with multiple parallel feeders. The power quality monitor was installed immediately downstream of the substation transformer, and its sampling frequency was 7.68 kHz (i.e., 128 waveform samples per cycle). The entire monitoring period contained eight major switching events: four capacitor energizing operations and four capacitor de-energizing operations. Further, some relatively large load switching events were also observed, although they had a smaller impact on the RMS voltage profile compared to capacitor switching events.
Synthetic Signals
Synthetic signals were also used in this study because they contained information about the true RMS voltage value without noise contamination at each time instant. The following signals are analyzed in later sections:
  • Signal 1: The voltage level in a distribution system was in a quasi-stationary condition at 0.996 pu for 1 second. At that time instant, a capacitor bank was energized, instantaneously increasing the RMS voltage to 1.0 pu. After another 1 second had elapsed, the capacitor bank was de-energized and the RMS voltage level returned to 0.996 pu.
  • Signal 2: The voltage level in a distribution system was in a quasi-stationary condition at 1.0 pu for 1 second. At that time instant, the load size connected to the system increased gradually over 1 second, causing the RMS voltage to drop linearly to 0.996 pu. This voltage drop triggered the energizing of a capacitor bank, instantaneously increasing the voltage level back to 1.0 pu. Note: this is the scenario in which median filtering was unable to track the original signal, as mentioned in Section 3.3.
These synthetic signals represented RMS voltage profiles with a half-cycle time resolution, so that each second contained 120 RMS voltage values (for a 60 Hz system). Further, each signal also contained additive noise originating from a normal distribution with zero-mean and standard deviation equal to 0.00025 pu. Figure 4 depicts both synthetic signals.

4.1.2. RMS Profile Computation

Let a sampled waveform signal be represented by a vector z; then, its RMS value at instant k, Z k , is defined as
Z k = 1 N s = k N + 1 k z s 2 1 2
where N is the number of samples per cycle in the waveform signal. Industrial standards recommend updating an RMS voltage profile every half-cycle ( N / 2 samples) [14,41]; profiles with this time resolution will be indicated as Z ( 1 / 2 ) in the rest of this paper.
On the other hand, computing a new RMS value once every waveform sample becomes available might result in hundreds or thousands of updates per second. The high computational burden involved in this approach is often pointed out as a drawback of having RMS profiles with such a high time resolution [13,14]. However, it has been shown that a recursive approach eliminates such issues [36]. In the recursive approach, the RMS value at instant k is computed as
Z k = Z k 1 2 + 1 N z k 2 z k N 2 1 2
This recursive approach will be employed throughout the paper wherever RMS profile computation with a high time resolution is necessary.

4.1.3. Vector Norms

For a given vector z R n , its l 1 -norm (Manhattan norm) and l 2 -norm (Euclidean norm) are defined according to Equations (7) and (8), respectively:
z 1 = i = 1 n | z i |
z 2 = i = 1 n z i 2 1 2
Note that in the following sections, the squared Euclidean norm z 2 2 is preferred over z 2 in order to avoid the square root operator.

4.2. Proposed Approach

Figure 5 depicts an overview of the PQM dataset rebalancing framework proposed in this paper. First, the input voltage waveforms were converted into the corresponding RMS voltage profiles (Section 4.1.2), which were filtered to remove/attenuate additive noise (Section 4.3). The filtered RMS voltage profiles were segmented into fixed-length, non-overlapping windows (in this study, we set each window length to 1 s). Each one of these segments was classified as an event or non-event, using the RMS voltage gradient profile approach that was introduced in Section 3.1. After all segments were classified into one of the two categories, dataset rebalancing was performed as described in Section 2. Finally, the resulting dataset could be used for the training/validation of machine learning algorithms.

4.3. Data Filtering

This section describes the filtering of time series through optimization techniques. Consider a signal represented by the vector x R n , where each coefficient x i represents the signal value sampled at the i-th time instant and the sampling interval is fixed. Without loss of generality, it is often assumed that the signal does not vary too rapidly for most of the time, as was the case for the signals analyzed in this study, so that x i x i + 1 .
As commonly observed in field measurements, the signal x is corrupted by an additive noise ν , i.e., x cor = x + ν . Note that x cor is observable by measurement devices, whereas the true underlying signal x is unknown. The additive noise ν can be modeled based on known characteristics of the process under study; however, for generality, it will be assumed that it follows an unknown distribution, has a small amplitude and varies much more rapidly than the signal x [42].
The objective of time series filtering is to produce an estimate x * of the original signal x, given the corrupted signal x cor ; this process is also called signal reconstruction or de-noising. The reconstructed signal x * should be similar to the corrupted signal and smooth; i.e., the rapidly varying noise is removed or significantly attenuated. The closeness between the corrupted and reconstructed signals is often measured with respect to the l 2 -norm, and a penalty function ϕ is used to assess the non-smoothness of the reconstructed signal. Thus, this signal filtering problem can be formulated as a convex vector optimization problem [42], as follows:
x * = argmin x ^ R n F ( x ^ , x cor )
where the objective function
F ( x ^ , x cor ) = x ^ x cor 2 ϕ ( x ^ )
is a vector. Its first component, F 1 = x ^ x cor 2 represents a measure of fit or consistency between the corrupted and estimated signals, whereas the second component, F 2 = ϕ ( x ^ ) , measures the roughness or lack of smoothness of the estimate x ^ . The function ϕ : R n R is convex and often given as some norm. Note, however, that F 1 and F 2 do not need to be measured with respect to the same norm, and this fact will be exploited later to produce better estimates for x * . In problems involving l 2 -norms, it is common practice to consider the corresponding squared norms [42], so that the nonlinearities caused by square roots are removed from the problem formulation; thus, we will adopt F 1 = x ^ x cor 2 2 .
The formulation presented in Equation (9) corresponds to a multi-objective optimization problem, where each component can be interpreted as different scalar objectives. The goal is to minimize each one of these components; however, they represent competing objectives, and a decrease in F 1 is accompanied by an increase in F 2 and vice-versa.
A standard approach for solving such optimization problems is called scalarization or regularization, where the objective function in Equation (9) is reformulated as λ T F ( x ^ , x cor ) = λ 1 x ^ x cor 2 2 + λ 2 ϕ ( x ^ ) for any weight vector λ > 0 [42]. Note that λ T F ( x ^ , x cor ) is scalar-valued and convex, since it is a weighted sum of convex functions [43]. Therefore, the reformulated problem is an ordinary scalar convex optimization problem, which can be solved easily.
The weight vector λ has a great influence on the filtering process as it controls the smoothness level of the output signal, and choosing a suitable value is critical to achieving the desired level of noise removal [44]. In general, each choice of λ results in a different estimate x * [42]. Let λ = [ 1 , δ ] T , for some δ > 0 ; as δ varies over [ 0 , ) , the solution of the equivalent scalar optimization problem traces out the optimal trade-off curve (or Pareto curve) between minimizing each component F 1 and F 2 separately. Figure 6 depicts a typical Pareto curve for a bi-criterion vector optimization problem, where values of components F 1 and F 2 are plotted on the horizontal and vertical axes, respectively.
For any δ , the slope of the Pareto curve represents the local optimal trade-off between the two objectives: if the slope is steep, small changes in F 1 are accompanied by large changes in F 2 , and vice-versa [42]. In other words, a Pareto curve allows us to determine how large one of the objectives must be in order to have the other one be small. Thus, the filtering of signals with a low signal-to-noise ratio (high noise levels) requires a larger δ [44].
In the extremes of a Pareto curve, we have the following interpretation:
  • δ = 0 : there is no penalty associated with the roughness of the output signal; thus, no smoothing is performed and x * = x cor . This scenario corresponds to the endpoint at the left in the Pareto curve, and it represents the smallest possible value of F 1 without any consideration of F 2 .
  • δ : a stronger emphasis is placed on the smoothness of the output signal, at the expense of disregarding the similarity between the corrupted and estimated signals; for a sufficiently large δ , x * becomes a constant signal. This scenario corresponds to the endpoint at the right in the Pareto curve, and it represents the smallest possible value of F 2 without any consideration of F 1 .
Choosing a suitable δ is a compromise between F 1 and F 2 . In practice, its value is chosen empirically by analyzing the Pareto curve and selecting a value such that a small decrease in one objective is accompanied by a small increase in the other objective [42]. The δ values that satisfy this requirement form the knee of the Pareto curve.
In the next sections, we present different strategies for quantifying the smoothness of the filtered signal; i.e., we present formulations for the component F 2 = ϕ ( x ^ ) of the objective function.

4.3.1. Quadratic Smoothing

The most straightforward roughness measure of a signal is given in terms of the sum of squares of differences. The quadratic smoothing function is defined as
F 2 = ϕ quad ( x ^ ) = i = 1 n 1 ( x ^ i + 1 x ^ i ) 2 = D x ^ 2 2
where D R ( n 1 ) × n is the bidiagonal matrix
D = [ r ] 1 1 0 0 0 0 0 1 1 0 0 0 0 0 0 1 1 0 0 0 0 0 1 1 ( n 1 ) × n
and represents an approximation to the first-order differentiation operator. As ϕ quad ( x ^ ) is defined in terms of a l 2 -norm, its squared value is used in the optimization problem, as discussed previously.
The estimate x * is the solution to the following unconstrained scalar optimization problem:
minimize x ^ R n x ^ x cor 2 2 + δ D x ^ 2 2
where δ > 0 parametrizes the optimal trade-off curve between x ^ x cor 2 2 and D x ^ 2 2 . This formulation corresponds to a quadratic problem, which can be solved very efficiently [42].
Figure 7a shows the Pareto curve for δ [ 0 , 500 ] for the synthetic signal 1 defined in Section 4.1.1, where it can be observed that δ 2 is the optimal weight. Figure 7b depicts three smoothed signals on the optimal trade-off curve:
  • δ = 0.2 (under-filtering): the weight associated with the output signal roughness is too small; although the steep edges in the signal are preserved, there is almost no reduction in the signal noise.
  • δ = 2 (optimal): this scenario represents the optimal trade-off between corrupted and estimated signals similarity and noise reduction; however, the noise level in the filtered signal is still quite high.
  • δ = 100 (over-filtering): an excessive weight is placed on the signal smoothness, resulting in over-filtering; the similarity between the corrupted and estimated signals is rather low.
Figure 8 shows the filtering results for the synthetic signal 2, and the discussion presented above is also valid for this test case.
This analysis shows that quadratic smoothing either removes the rapidly varying noise or preserves the steep signal edges, but not both; in fact, quadratic smoothing behaves as a low-pass filter. Thus, this technique is not suitable for the category of signals analyzed in this paper.

4.3.2. Total Variation Smoothing

Given the limitations of quadratic smoothing discussed previously, this section describes a smoothing function that is effective at removing/attenuating the signal noise, while still preserving the steep edges of the original signal [45,46]. In this case, the signal smoothness is measured according to the following function:
F 2 = ϕ tv ( x ^ ) = i = 1 n 1 x ^ i + 1 x ^ i = D x ^ 1
which is called the total variation of x ^ R n . Note that, compared to ϕ quad in Equation (11), ϕ tv is not squared, as it is given in terms of a l 1 -norm and there are no square root terms to be removed.
The estimate x * is the solution of the following unconstrained scalar optimization problem:
minimize x ^ R n x ^ x cor 2 2 + δ D x ^ 1
The optimization problem in Equation (15) cannot be easily solved because the l 1 -norm is non-differentiable [47]. The following problem reformulation is based on [43]. First, for simplicity, we introduce a new variable y i = x ^ i + 1 x ^ i , i = 1 , , n 1 , so that ϕ tv ( x ^ ) = i = 1 n 1 y i .
Let y i = y i + y i , i = 1 , , n 1 , where y i + and y i are variables constrained to be nonnegative. It can be shown that these two variables cannot be simultaneously nonzero; i.e., at least one of the variables y i + and y i is zero for each index i. Therefore,
y i = y i + , if y i 0 ( y i + 0 , y i = 0 ) y i , if y i < 0 ( y i + = 0 , y i > 0 ) y i = y i + + y i
By replacing y i in Equation (15), the following alternative formulation is obtained:
minimize x ^ , y , y + , y R n x ^ x cor 2 2 + δ i = 1 n 1 y i + + y i subject to y i = x ^ i + 1 x ^ i , i = 1 , , n 1 y i = y i + y i , i = 1 , , n 1 y i + 0 , i = 1 , , n 1 y i 0 , i = 1 , , n 1
which is a constrained, convex and differentiable optimization problem.
Figure 9 demonstrates the filtering of synthetic signal 1 through total variation smoothing. Figure 9a shows the Pareto curve for δ [ 0 , 5 ] , where it can be observed that δ 0.004 is the optimal weight. Figure 9b depicts three smoothed signals on the optimal trade-off curve:
  • δ = 0.0002 (under-filtering): the weight associated with the output signal roughness is too small, meaning that there is almost no reduction in the signal noise.
  • δ = 0.004 (optimal): this scenario represents the optimal trade-off between corrupted and estimated signal similarity and noise reduction.
  • δ = 0.2 (over-filtering): an excessive weight is placed on the signal smoothness, resulting in over-filtering; due to the large penalty associated with variations in the signal, the magnitude of the step change in the filtered signal is much smaller than the magnitude of the true step change.
Figure 10 shows the filtering results for the synthetic signal 2, and the discussion presented above is also valid for this test case. Further, unlike median filtering, total variation smoothing was able to track this piecewise linear signal.
This analysis shows that total variation smoothing exhibits great performance in noise reduction without blurring the sharp transitions of the original signal, as long as the weight δ has been properly selected.

4.3.3. Quadratic vs. Total Variation Smoothing

As discussed in the previous sections, total variation smoothing shows better performance in the filtering of RMS voltage profiles when compared to quadratic smoothing. In this section, we explore and compare the characteristics of these two smoothing operators in order to justify the superiority achieved by total variation smoothing.
Both ϕ quad and ϕ tv , which were defined in Equations (11) and (14), respectively, assign large penalty costs to rapidly varying x ^ . However, the quadratic smoothness function assigns a relatively small penalty to small values of x ^ i + 1 x ^ i [48]. For example, if x ^ i + 1 x ^ i = 0.001 , then the penalties assigned by the quadratic and total variation smoothness functions are 10 6 and 10 3 , respectively. In other words, the quadratic smoothing operator tolerates some variation in the filtered signal, whereas the total variation smoothing operator is subject to a much larger penalty if such signal variations exist, meaning that it enforces x ^ i + 1 x ^ i 0 for almost all i’s.
In general, the following characteristics are observed in the solutions of optimization problems with penalty functions [42]:
  • l 2 -norm penalty: D x ^ 2 has many non-zero small entries and relatively few larger ones;
  • l 1 -norm penalty: D x ^ 1 has many zero or very small entries and more larger ones.
The optimization problem scalarized with an l 1 -norm is a heuristic for finding a solution in which D x ^ 1 is sparse. As D represents an approximation to the first-order differentiation operator, total variation smoothing is biased toward solutions in which the filtered signal is linear or piecewise linear.
This behavior can be observed in Figure 11, which depicts the histogram of x i + 1 x i for the filtered signals computed in Section 4.3.1 (Figure 7b, quadratic smoothing with δ = 2 ) and Section 4.3.2 (Figure 9b, total variation smoothing with δ = 0.004 ), respectively. As expected, quadratic smoothing allows some x ^ i + 1 x ^ i to be greater than zero, which correspond to the smooth transition around the steep edges of the original signal. On the other hand, almost all x ^ i + 1 x ^ i in Figure 11b are approximately zero, except for two values that correspond to the two step changes present in the original signal.

5. Results

This section demonstrates the application of the proposed framework (i.e., total variation smoothing) using field data collected at the feeder head of a 25 kV radial distribution system, which is described in Section 4.1.1. Based on the results in Section 4.3.2, we adopted δ = 0.0035 . Figure 12a shows the unfiltered and filtered RMS voltage profiles for the entire 28-min measurement interval, whereas Figure 12b shows the corresponding RMS voltage gradient profiles. Using the triggering threshold δ step = 0.0018 pu (Section 3.1), all RMS voltage step changes were successfully detected without any false positives. The root causes of the detected RMS voltage step changes were capacitor de-energizing (events 1, 2, 5 and 6) and capacitor energizing (events 3, 4, 7 and 8). Further, the unfiltered RMS voltage gradient profile did not create any false positives either; however, the gradient values were much larger compared to the filtered cases (as large as 0.0015 pu), indicating that the unfiltered profile might create false positives for some datasets.
Detailed views of the unfiltered and filtered RMS voltage profiles are shown in Figure 13 for four scenarios: capacitor de-energizing, capacitor energizing, load energizing and steady-state. Note that the filtered profile did not contain rapidly varying noise and its step changes were not affected, as initially desired. The unfiltered RMS voltage profile in Figure 13b contained a spike immediately after the RMS voltage step change, which was due to high-frequency transients in the voltage waveform caused by a capacitor energizing operation. On the other hand, total variation smoothing successfully removed this spike. This is an important advantage of using the filtered profile, as the magnitude of the step change might be one of the features employed by the machine learning algorithm (the magnitude given by the unfiltered profile is about 50% larger than the correct value).
Both unfiltered and filtered RMS voltage profiles were segmented into non-overlapping 1 s windows and classified as an event or non-event, as described in Section 4.2. Table 1 shows the distribution of classes before and after the rebalancing of the PQM dataset.
Before dataset rebalancing, less than 0.5% of the observations in the input dataset corresponded to power system disturbances; in this case, machine learning algorithms are very unlikely to be able to extract any useful information about the minority class. On the hand, the dataset was perfectly balanced using the framework proposed in this paper. It should be noted, however, that the rebalanced dataset contained only 16 observations, which is often considered too small for successfully training machine learning algorithms. One solution would be to select more observations for the majority class, as long as the resulting dataset does not become highly unbalanced. Another solution consists of collecting more PQM data; the field data considered in this paper represent only 28 minutes of measurement, whereas utilities have access to much longer measurement intervals (days, weeks or even months).

6. Conclusions

The RMS voltage profile filtering proposed in this paper was shown to be robust for removing/attenuating rapidly varying signal noise without blurring out the RMS voltage step changes due to switching events. By combining filtering and step change detection techniques, both conspicuous and inconspicuous events present in a PQM dataset can be successfully identified. Detecting such events is the basis for rebalancing highly unbalanced PQM datasets, consequently improving the performance of machine learning algorithms that use these datasets in their training and validations stages.
As observed in Figure 7b, Figure 8b, Figure 9b and Figure 10b, the parameter δ has a great effect on the RMS voltage profile filtering process. Further, the optimal value for δ depends on the noise level present in the signal; i.e., scenarios with a higher signal-to-noise ratio (low noise level) require a lower δ value. Therefore, the optimal δ value adopted in this paper might not be the most suitable choice for field measurements collected at other locations, as the signal-to-noise ratio might be different.
Future research directions include the development of techniques for automatically determining the optimal δ for each dataset. For example, for a given RMS voltage profile, such techniques would first analyze only a short segment of the profile for multiple δ values in order to construct the Pareto curve. Then, the optimal δ would be the value corresponding to the knee of the Pareto curve, as shown in Figure 6. Once this optimal value has been determined, the whole RMS voltage profile would be filtered through total variation smoothing. It is important to emphasize that optimization applications based on the Pareto curve in all fields—and not only power systems—empirically determine the optimal δ by visually inspecting the Pareto curve. Thus, a technique for automatically determining this value would represent a meaningful contribution.

Author Contributions

The conceptualization, methodology, software development, data analysis and writing of the manuscript was done by A.F.B.; S.S. was the advisor mentoring the project. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ARAutoregressive
MLMachine learning
PQMPower quality monitor
puPer unit
RMSRoot-mean-square
ROCReceiver operating characteristic
RVCRapid voltage change

References

  1. Mosavi, A.; Salimi, M.; Ardabili, S.F.; Rabczuk, T.; Shamshirband, S.; Varkonyi-Koczy, A.R. State of the art of machine learning models in energy systems: A systematic review. Energies 2019, 12, 1301. [Google Scholar] [CrossRef] [Green Version]
  2. Kumar, N.M.; Chand, A.A.; Malvoni, M.; Prasad, K.A.; Mamun, K.A.; Islam, F.R.; Chopra, S.S. Distributed energy resources and the application of AI, IoT, and blockchain in smart grids. Energies 2020, 13, 5739. [Google Scholar] [CrossRef]
  3. Perez-Ortiz, M.; Jimenez-Fernandez, S.; Gutierrez, P.A.; Alexandre, E.; Hervas-Martinez, C.; Salcedo-Sanz, S. A review of classification problems and algorithms in renewable energy applications. Energies 2016, 9, 607. [Google Scholar] [CrossRef]
  4. Bastos, A.F.; Santoso, S. Condition monitoring of circuit-switchers for shunt capacitor banks through power quality data. IEEE Trans. Power Deliv. 2019, 34, 1499–1507. [Google Scholar] [CrossRef]
  5. Ananthan, S.N.; Bastos, A.F.; Santoso, S.; Chirapongsananurak, P. Model-based approach integrated with fault circuit indicators for fault location in distribution systems. In Proceedings of the IEEE Power and Energy Society General Meeting, Atlanta, GA, USA, 4–8 August 2019; pp. 1–5. [Google Scholar]
  6. Ananthan, S.N.; Bastos, A.F.; Santoso, S. Novel system model-based fault location approach using dynamic search technique. IET Gener. Transm. Distrib. 2021. [Google Scholar] [CrossRef]
  7. Bastos, A.F.; Santoso, S.; Freitas, W.; Xu, W. SynchroWaveform measurement units and applications. In Proceedings of the IEEE Power and Energy Society General Meeting, Atlanta, GA, USA, 4–8 August 2019; pp. 1–5. [Google Scholar]
  8. Bastos, A.F.; Lao, K.W.; Todeschini, G.; Santoso, S. Novel moving average filter for detecting rms voltage step changes in triggerless PQ data. IEEE Trans. Power Deliv. 2018, 33, 2920–2929. [Google Scholar] [CrossRef] [Green Version]
  9. Xu, W. Experiences on Using Gapless Waveform Data and Synchronized Harmonic Phasors. In Panel Session in IEEE Power and Energy Society General Meeting; Technical Report; IEEE Power & Energy Society: Piscataway, NJ, USA, 2015. [Google Scholar]
  10. Silva, L.; Kapisch, E.; Martins, C.; Filho, L.; Cerqueira, A.; Duque, C.; Ribeiro, P. Gapless power-quality disturbance recorder. IEEE Trans. Power Deliv. 2017, 32, 862–871. [Google Scholar] [CrossRef]
  11. Li, B.; Jing, Y.; Xu, W. A generic waveform abnormality detection method for utility equipment condition monitoring. IEEE Trans. Power Deliv. 2017, 32, 162–171. [Google Scholar] [CrossRef]
  12. Bastos, A.F.; Freitas, W.; Todeschini, G.; Santoso, S. Detection of inconspicuous power quality disturbances through step changes in rms voltage profile. IET Gener. Transm. Distrib. 2019, 13, 2226–2235. [Google Scholar] [CrossRef]
  13. Bollen, M.; Gu, I. Signal Processing of Power Quality Disturbances; Wiley: Hoboken, NJ, USA, 2006. [Google Scholar]
  14. IEC. IEC Electromagnetic Compatibility: Testing and Measurements Techniques—Power Quality Measurement Methods; IEC: London, UK, 2015; Standard 61000-4-30. [Google Scholar]
  15. Bastos, A.F.; Santoso, S. Universal waveshape-based disturbance detection in power quality data using similarity metrics. IEEE Trans. Power Deliv. 2020, 35, 1779–1787. [Google Scholar] [CrossRef]
  16. Santoso, S.; Powers, E.J.; Grady, W.M.; Parsons, A.C. Power quality disturbance waveform recognition using wavelet-based neural classifier—Part 1: Theoretical foundation. IEEE Trans. Power Deliv. 2000, 15, 222–228. [Google Scholar] [CrossRef]
  17. Bastos, A.F.; Lao, K.W.; Todeschini, G.; Santoso, S. Accurate identification of point-on-wave inception and recovery instants of voltage sags and swells. IEEE Trans. Power Deliv. 2019, 34, 551–560. [Google Scholar] [CrossRef] [Green Version]
  18. Bastos, A.F.; Santoso, S.; Todeschini, G. Comparison of methods for determining inception and recovery points of voltage variation events. In Proceedings of the IEEE Power and Energy Society General Meeting, Portland, OR, USA, 5–10 August 2018; pp. 1–5. [Google Scholar]
  19. Bastos, A.F.; Santoso, S. Identifying switched capacitor relative locations and energizing operations. In Proceedings of the IEEE Power and Energy Society General Meeting, Boston, MA, USA, 17–21 July 2016; pp. 1–5. [Google Scholar]
  20. Hart, W.E.; Laird, C.D.; Watson, J.P.; Woodruff, D.L.; Hackebeil, G.A.; Nicholson, B.L.; Siirola, J.D. Pyomo—Optimization Modeling in Python, 2nd ed.; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2017; Volume 67. [Google Scholar]
  21. Bishop, C.M. Pattern Recognition and Machine Learning, 1st ed.; Springer: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  22. Duda, R.O.; Hart, P.E.; Stork, D.G. Pattern Classification, 2nd ed.; Wiley-Interscience: Hoboken, NJ, USA, 2000. [Google Scholar]
  23. Gron, A. Hands-On Machine Learning with Scikit-Learn and TensorFlow: Concepts, Tools, and Techniques to Build Intelligent Systems, 1st ed.; O’Reilly Media, Inc.: Newton, MA, USA, 2017. [Google Scholar]
  24. He, H.; Garcia, E.A. Learning from imbalanced data. IEEE Trans. Knowl. Data Eng. 2009, 21, 1263–1284. [Google Scholar]
  25. Batista, G.E.A.P.A.; Prati, R.C.; Monard, M.C. A study of the behavior of several methods for balancing machine learning training data. ACM SIGKDD Explor. Newsl. 2004, 6, 20–29. [Google Scholar] [CrossRef]
  26. Torgo, L.; Ribeiro, R.P. Precision and Recall for Regression; Discovery Science; Springer: Berlin/Heidelberg, Germany, 2009; pp. 332–346. [Google Scholar]
  27. Bastos, A.F.; Santoso, S.; Krishnan, V.; Zhang, Y. Machine learning-based prediction of distribution network voltage and sensor allocation. In Proceedings of the IEEE Power and Energy Society General Meeting, Montreal, QC, Canada, 2–6 August 2020; pp. 1–5. [Google Scholar]
  28. Brownlee, J. 8 Tactics to Combat Imbalanced Classes in Your Machine Learning Dataset. 2015. Available online: https://machinelearningmastery.com/tactics-to-combat-imbalanced-classes-in-your-machine-learning-dataset/ (accessed on 20 December 2020).
  29. Chawla, N.V. Data Mining for Imbalanced Datasets: An Overview. In Data Mining and Knowledge Discovery Handbook; Maimon, O., Rokach, L., Eds.; Springer: Boston, MA, USA, 2005; pp. 853–867. [Google Scholar]
  30. Tamayo, S.C.; Luna, J.M.F.; Huete, J.F. On the use of weighted mean absolute error in recommender systems. In Proceedings of the Workshop on Recommendation Utility Evaluation, Dublin, Ireland, 9 September 2012; pp. 24–26. [Google Scholar]
  31. Torgo, L.; Ribeiro, R.P.; Pfahringer, B.; Branco, P. SMOTE for Regression. In Lecture Notes in Computer Science, Proceedings of the Progress in Artificial Intelligence, Azores, Portugal, 9–12 September 2013; Springer: Berlin/Heidelberg, Germany, 2013; pp. 378–389. [Google Scholar]
  32. Short, T.A. Electric Power Distribution Handbook; CRC Press: Boca Raton, RL, USA, 2003. [Google Scholar]
  33. IEEE. IEEE Guide for Application of Shunt Power Capacitors; Standard 1036; IEEE: Piscataway, NJ, USA, 2010. [Google Scholar]
  34. Bastos, A.F.; Biyikli, L.; Santoso, S. Analysis of power factor over correction in a distribution feeder. In Proceedings of the IEEE Power and Energy Society Transmission and Distribution Conference and Exposition, Dallas, TX, USA, 3–5 May 2016; pp. 1–5. [Google Scholar]
  35. Gustafsson, F. Adaptive Filtering and Change Detection; John Wiley & Sons: Hoboken, NJ, USA, 2000. [Google Scholar]
  36. Bastos, A.F.; Santoso, S. Root-mean-square profiles under varying power frequency: Computation and applications. In Proceedings of the IEEE Power and Energy Society General Meeting, Atlanta, GA, USA, 4–8 August 2019; pp. 1–5. [Google Scholar]
  37. Bastos, A.F.; Kim, T.; Santoso, S.; Grady, W.M.; Gravois, P.; Miller, M.; Kadel, N.; Schmall, J.; Huang, S.H.; Blevins, B. Frequency retrieval from PMU data corrupted with pseudo-oscillations during off-nominal operation. In Proceedings of the North American Power Symposium, Tempe, AZ, USA, 11–14 April 2021; pp. 1–6. [Google Scholar]
  38. Smith, S. The Scientist and Engineer Guide to Digital Signal Processing; California Tech. Pub.: San Diego, CA, USA, 1997. [Google Scholar]
  39. Castro, E.A.; Donoho, D.L. Does median filtering truly preserve edges better than linear filtering? Ann. Stat. 2009, 37, 1172–1206. [Google Scholar] [CrossRef] [Green Version]
  40. Huber, P.J. Robust estimation of a location parameter. Ann. Math. Stat. 1964, 35, 73–101. [Google Scholar] [CrossRef]
  41. IEEE. IEEE Guide for Voltage Sag Indices; Standard 1564; IEEE: Piscataway, NJ, USA, 2014. [Google Scholar]
  42. Boyd, S.; Vandenberghe, L. Convex Optimization, 7th ed.; Cambridge University Press: Cambridge, UK, 2009. [Google Scholar]
  43. Bertsimas, D.; Tsitsiklis, J.N. Introduction to Linear Optimization, 1st ed.; Athena Scientific: Belmont, MA, USA, 1997. [Google Scholar]
  44. Selesnick, I.W.; Bayram, I. Total Variation Filtering. 2010. Available online: https://eeweb.engineering.nyu.edu/iselesni/lecture_notes/TV_filtering.pdf (accessed on 20 December 2020).
  45. Rudin, L.I.; Osher, S.; Fatemi, E. Nonlinear total variation based noise removal algorithms. Phys. D Nonlinear Phenom. 1992, 60, 259–268. [Google Scholar] [CrossRef]
  46. Little, M.A.; Jones, N.S. Sparse Bayesian step-filtering for high-throughput analysis of molecular machine dynamics. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Dallas, TX, USA, 14–19 March 2010; pp. 4162–4165. [Google Scholar]
  47. Chambolle, A. An algorithm for total variation minimization and applications. J. Math. Imaging Vis. 2004, 20, 89–97. [Google Scholar]
  48. Strong, D.; Chan, T. Edge-preserving and scale-dependent properties of total variation regularization. Inverse Probl. 2003, 19, S165–S187. [Google Scholar] [CrossRef]
Figure 1. General workflow of a typical machine learning application.
Figure 1. General workflow of a typical machine learning application.
Energies 14 00463 g001
Figure 2. General overview of the work presented in this paper.
Figure 2. General overview of the work presented in this paper.
Energies 14 00463 g002
Figure 3. Illustration of the root-mean-square (RMS) voltage gradient profile computation for p = 2 .
Figure 3. Illustration of the root-mean-square (RMS) voltage gradient profile computation for p = 2 .
Energies 14 00463 g003
Figure 4. Synthetic signals analyzed throughout this study (both before and after noise addition).
Figure 4. Synthetic signals analyzed throughout this study (both before and after noise addition).
Energies 14 00463 g004
Figure 5. Overview of the power quality monitor (PQM) dataset rebalancing framework proposed in this paper.
Figure 5. Overview of the power quality monitor (PQM) dataset rebalancing framework proposed in this paper.
Energies 14 00463 g005
Figure 6. Typical Pareto curve for a bi-criterion vector optimization problem.
Figure 6. Typical Pareto curve for a bi-criterion vector optimization problem.
Energies 14 00463 g006
Figure 7. Results of filtering the corrupted synthetic signal 1 through quadratic smoothing. (a) Pareto curve. (b) Estimated signals representing under-filtering ( δ = 0.2 ), optimal filtering ( δ = 2 ) and over-filtering ( δ = 100 ).
Figure 7. Results of filtering the corrupted synthetic signal 1 through quadratic smoothing. (a) Pareto curve. (b) Estimated signals representing under-filtering ( δ = 0.2 ), optimal filtering ( δ = 2 ) and over-filtering ( δ = 100 ).
Energies 14 00463 g007
Figure 8. Results of filtering the corrupted synthetic signal 2 through quadratic smoothing. (a) Pareto curve. (b) Estimated signals representing under-filtering ( δ = 0.2 ), optimal filtering ( δ = 2 ) and over-filtering ( δ = 100 ).
Figure 8. Results of filtering the corrupted synthetic signal 2 through quadratic smoothing. (a) Pareto curve. (b) Estimated signals representing under-filtering ( δ = 0.2 ), optimal filtering ( δ = 2 ) and over-filtering ( δ = 100 ).
Energies 14 00463 g008
Figure 9. Results of filtering the corrupted synthetic signal 1 through total variation smoothing. (a) Pareto curve. (b) Estimated signals representing under-filtering ( δ = 0.0002 ), optimal filtering ( δ = 0.004 ) and over-filtering ( δ = 0.2 ).
Figure 9. Results of filtering the corrupted synthetic signal 1 through total variation smoothing. (a) Pareto curve. (b) Estimated signals representing under-filtering ( δ = 0.0002 ), optimal filtering ( δ = 0.004 ) and over-filtering ( δ = 0.2 ).
Energies 14 00463 g009
Figure 10. Results of filtering the corrupted synthetic signal 2 through total variation smoothing. (a) Pareto curve. (b) Estimated signals representing under-filtering ( δ = 0.0002 ), optimal filtering ( δ = 0.003 ) and over-filtering ( δ = 0.1 ).
Figure 10. Results of filtering the corrupted synthetic signal 2 through total variation smoothing. (a) Pareto curve. (b) Estimated signals representing under-filtering ( δ = 0.0002 ), optimal filtering ( δ = 0.003 ) and over-filtering ( δ = 0.1 ).
Energies 14 00463 g010
Figure 11. Histogram of the first derivative amplitudes for the filtered synthetic signal 1 using the optimal δ value for each scenario. (a) Quadratic smoothing with δ = 2 . (b) Total variation smoothing with δ = 0.004 .
Figure 11. Histogram of the first derivative amplitudes for the filtered synthetic signal 1 using the optimal δ value for each scenario. (a) Quadratic smoothing with δ = 2 . (b) Total variation smoothing with δ = 0.004 .
Energies 14 00463 g011
Figure 12. Results from the field data. (a) Unfiltered and filtered RMS voltage profiles; the filtered profile was obtained through total variation smoothing with δ = 0.0035 . (b) Unfiltered and filtered RMS voltage gradient profiles, where the numbers in circles represent event IDs.
Figure 12. Results from the field data. (a) Unfiltered and filtered RMS voltage profiles; the filtered profile was obtained through total variation smoothing with δ = 0.0035 . (b) Unfiltered and filtered RMS voltage gradient profiles, where the numbers in circles represent event IDs.
Energies 14 00463 g012
Figure 13. Detailed view of the unfiltered and filtered RMS voltage profiles for the field data. (a) Capacitor de-energizing (event 1). (b) Successive capacitor energizing (events 7 and 8). (c) Load energizing (between events 3 and 4). (d) Steady-state (between events 4 and 5).
Figure 13. Detailed view of the unfiltered and filtered RMS voltage profiles for the field data. (a) Capacitor de-energizing (event 1). (b) Successive capacitor energizing (events 7 and 8). (c) Load energizing (between events 3 and 4). (d) Steady-state (between events 4 and 5).
Energies 14 00463 g013
Table 1. Class distribution for the PQM dataset before and after rebalancing.
Table 1. Class distribution for the PQM dataset before and after rebalancing.
Before RebalancingAfter Rebalancing
Majority class (Event)1672 (99.52%)8 (50%)
Minority class (Non-Event)8 (0.48%)8 (50%)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Furlani Bastos, A.; Santoso, S. Optimization Techniques for Mining Power Quality Data and Processing Unbalanced Datasets in Machine Learning Applications. Energies 2021, 14, 463. https://doi.org/10.3390/en14020463

AMA Style

Furlani Bastos A, Santoso S. Optimization Techniques for Mining Power Quality Data and Processing Unbalanced Datasets in Machine Learning Applications. Energies. 2021; 14(2):463. https://doi.org/10.3390/en14020463

Chicago/Turabian Style

Furlani Bastos, Alvaro, and Surya Santoso. 2021. "Optimization Techniques for Mining Power Quality Data and Processing Unbalanced Datasets in Machine Learning Applications" Energies 14, no. 2: 463. https://doi.org/10.3390/en14020463

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop