1. Introduction
China has launched a magnificent plan to promote intelligent processes for the distribution systems in response to extreme weather and renewable energy consumption. In contrast to the attention-attracting high-voltage distribution network construction, the majority stock of 380 V-below low-voltage power station areas (LVPSAs) remain passive and have lower levels of automation due to prohibitive investment in upgrading the infrastructure. Thus, it is of great significance to explore a capital-efficient scheme to improve the self-healing ability of LVPSAs to power interruption events, which tend to be increasingly intensive with frequent extreme weather.
At this stage, extensive research has been conducted on the fault self-healing of medium-voltage distribution networks [
1,
2]. However, due to different communication environments, the results achieved cannot fully meet the needs of low-voltage distribution network fault self-healing. Network reconstruction is a common method for distribution networks to achieve fault self-healing. Combined with the structural characteristics of low-voltage distribution networks, the issue of low-voltage network topology reconstruction has also attracted attention. For example, the literature [
3] proposes a network reconstruction method for medium- and low-voltage distribution network coordination based on neural networks for the problem of phase-to-phase load balancing. The literature [
4] proposes a multi-objective tie switch optimization configuration method for low-voltage photovoltaic accommodation problems, but none of them consider the change in load demand during low-voltage fault repair and the difficulty in obtaining low-voltage line parameters.
Low-voltage fault repair time generally lasts for several hours. Short-term load demand changes have a significant impact on fault recovery strategies [
5]. Current short-term load forecasting technologies mainly include traditional data-driven, artificial intelligence, and hybrid forecasting methods. Traditional data-driven methods are represented by methods such as the autoregressive integrated moving average (ARIMA) and modal decomposition. Artificial intelligence methods [
6,
7] include the long short-term memory network (LSTM) [
8,
9], convolutional neural network [
10], and an adversarial neural network method [
11]. Note that model training for these methods often relies on large numbers of samples and computing resources. In contrast, incremental learning methods developed in recent years [
12,
13] can obtain high-precision short-term load forecasting with less computing resources by supplementing real-time samples. From the perspective of communication and computing power costs, incremental learning prediction has good potential for application in low-voltage load prediction problems with a large number of nodes and relatively simple nature components.
From the above literature summary, it is evident that research on the self-healing and recovery of LVPSAs is still very limited. Issues such as communication technology and the configuration of self-healing equipment for LVPSAs remain in their early stages. Existing self-healing techniques developed for medium-voltage distribution systems often rely on precise line parameters. Meanwhile, short-term load forecasting methods, despite leveraging advanced intelligent techniques, typically require extensive historical load records and long training times. In this regard, it is necessary to consider communication cost and quality to propose efficient load prediction and self-healing methods which are viable to deploy on the lightweight computing hardware.
Being aware of the typical structure of LVPSAs and investment limitations of complex communication infrastructure, this paper proposes a fault self-healing strategy for LVPSAs based on fog computing architecture. In comparison with the existing literature, the main contributions of this paper are highlighted as follows:
A fog computing load forecasting method, used to offer load demand inputs of the self-healing strategy, is proposed based on a dynamic aggregation of incremental learning models. This forecasting method embeds two weighted ultra-short-term load forecasting techniques with complementary characteristics and mines real-time load to learn incrementally.
A line-parameter-free self-healing strategy for a faulted LVPSA is formulated as a three-stage mixed integer program, which of the slave subproblem is forged as a type-1 Wasserstein distributionally robust optimization (p-1WDRO) accounting for inevitable load forecast errors in a data-driven manner.
An algorithm to expedite the model-resolving process is proposed by means of the column-and-constraint generation algorithm. To this end, integer variables involved in the slave problem are eliminated a priori to help derive the equivalent Karush–Kuhn–Tucker conditions with the embedded p-1WDRO problem.
The rest of this paper is organized as follows.
Section 2 describes the fog communication framework and operating hypotheses designed for applying the proposed LVPSA self-healing strategy.
Section 3 elaborates on the ultra-short-term load forecast method. And, the proposed LVPSA self-healing strategy, along with the resolving algorithm, is described in
Section 4, which is followed by case studies and conclusions making up
Section 5 and
Section 6, respectively.
2. Framework and Hypotheses
2.1. Concept and Hypotheses
To avoid ambiguity with respect to the current research status of renewable energy penetrated distribution networks, it is first clarified that the LVPSA studied in this article refers to the traditional distribution (or service)-transformer-supplied radial 380 V power station area in the absence of any distributed energy resources. Such a topology prevails in most practical urban low-voltage distribution networks in China. Starting from the transformer, the downstream structure of the concerned LVPSAs is outlined in
Figure 1 and explained as follows: outlet wires of the transformer are tapped through the
outlet cabinet to the
distribution box and then led out from the distribution box to the
light-power cabinet generally residing in the building. Finally, the outlet wires of the light-power cabinet are designated phase by phase to the home through
branch boxes. Note, each location of each box/cabinet forms a branch node. According to the operation and management requirements of the power grid department of China, distribution boxes are generally only equipped with knife switches with fuses. On the other hand, the other cabinets can be equipped with remote-controlled circuit breakers and monitoring communication units to perform the functions of “telemetry, remote signaling, and remote control” on demand. Furthermore, through integration with the
intelligent fusion terminal and lightweight edge computing, advanced functions such as fault clearing, fault isolation, switching operations, etc., are enabled. Considering the above-mentioned software and hardware environment and data acquisition conditions of the LVPSAs, the LVPSA self-healing strategy devised in this paper is based on the following assumptions:
- ①
It is assumed that the node voltages of the LVPSA are dominated by the upstream power grid and always maintain rated values (1 p.u.) being insensitive to low-voltage power fluctuations.
- ②
Since the node voltage is assumed to be a rated constant
, the line loss of the low-voltage feeder can be assumed to be approximately proportional to the branch power with rated value
. The proportional coefficient is termed the branch line loss coefficient, denoted by
, and expressed as
where
r is the alternative current resistance in per-unit length.
- ③
The power factor of the load is constant.
- ④
Except for distribution boxes, breakers can be configured between other distribution cabinets of the same type.
- ⑤
The light-power cabinet can be equipped with a fog computing device to collect power information reported by household intelligent meters and support lightweight advanced applications, such as ultra-short-term load forecasting.
2.2. Fog Communication Architecture
The self-healing communication technology deployed in LVPSAs needs to meet the requirements of distributed communication, low latency, and high efficiency. Fog computing is a narrow-band edge computing architecture that can provide data transmission, calculation, and storage [
14,
15]. It was proposed by Cisco in 2014. The self-healing communication architecture supporting LVPSA self-healing is shown in
Figure 1. In sum, the overall communication architecture is divided into three layers, namely the terminal layer, fog computing layer, and cloud computing layer.
The terminal layer, deployed at the user meter, is responsible for periodically collecting electrical parameters such as power, voltage, and current. It utilizes the OpenADR protocol and adheres to standards like IEEE 802.11 (WLAN) [
16], and IEEE 802.15.4 (LR-WPAN) [
17]. Data are transmitted to the fog computing layer through a secure, encrypted, wireless network.
The fog computing layer, deployed within the light-power cabinet, integrates computing resources such as central processors, graphics processors, and storage devices as needed. Acting as a transfer hub, it collects and decrypts user meter data while facilitating the exchange of graphic and text-based instructions. This layer supports lightweight advanced applications, including but not limited to load data cleaning and encryption, feature extraction, and data backup. As a cornerstone of distributed computing, the fog computing layer enables customized data uploads, effectively reducing the reliance on cloud communication bandwidth and minimizing transmission delays. For information transmission, in addition to carrier communication, long-range broadband wireless technologies such as WiMAX and 5G are viable options.
The cloud computing layer is deployed in the intelligent fusion terminal, enabling communication with the fog computing unit in the light-power cabinet. It provides stable and reliable information storage, distributed computing task decomposition, and supports computationally intensive applications such as optimizing the LVPSA self-healing strategy and issuing fault isolation commands.
To minimize the cost of establishing and maintaining a private communication network for large-scale low-voltage user information, communication across the three information layers can leverage third-party network broadband. By utilizing slicing and network partitioning techniques, private network emulation can be achieved efficiently.
2.3. Framework of Self-Healing Strategy
Based on the described fog computing communication architecture, the self-healing strategy of LVPSAs is primarily activated in response to faults occurring between the low-voltage side outlet of the distribution transformer and the power cabinet. The execution logic of the architecture is as follows: when the protection mechanism trips the upstream switch of the fault, the monitoring unit in the downstream power cabinet detects the fault and triggers a "loss of voltage and current" event signal. This signal is then broadcast to other fog computing units within the same power station area. Upon receiving the event signal, the fog computing units preprocess locally stored load samples, predict node power demand, and report the results to the intelligent fusion terminal. The intelligent fusion terminal subsequently optimizes the LVPSA self-healing strategy and issues commands to open or close the remote-controlled low-voltage switches accordingly.
The proposed fog computing architecture is characterized by its event-driven nature, where fog computing tasks and data communication are triggered only by specific events, such as fault occurrences or model update instructions. During normal operation, when no faults occur, the system remains in a low-power standby state, performing tasks like incremental model learning. Given that fault self-healing involves adjustments to network status, it is essential to predict load demand at both upstream and downstream fault locations during power restoration. The fog computing framework discussed in this article primarily focuses on ultra-short-term load power prediction during fault repair intervals. The following section introduces the proposed ultra-short-term load prediction method, specifically designed for fog computing environments.
3. Fog Computing Load Forecast
Under the framework of fog computing load forecasting, it is necessary to consider the limited capabilities of floating-point calculations and the data storage of a fog computing unit. This paper proposes a fog computing load forecasting method based on the dynamic aggregation of incremental learning models. This method embeds two independent ultra-short-term load forecasting models. In terms of comprehensive factors such as the calculation burden, data-recording storage, and prediction accuracy, this paper selects the method proposed in the literature [
13] as a candidate of the two embedded methods, represented by
fk. Inspired by this method, another load forecasting method based on neighborhood sample replacement is derived to be the other embedded candidate, represented by
fk-near. The two embedded models independently carry out incremental learning online based on real-time load to avoid hoarding a large number of samples and communication interactions. Their respective load prediction values are then dynamically weighted and aggregated into the final output load prediction results [
18]. The dynamic weights used in this article are progressively updated as per (
2):
where
T denotes the number of historical streaming power demand samples.
represents the true load demand samples at time instance
j,
represents the forecasting value via the two embedded methods, and
denotes the Euclidean norm.
Given
, the proposed fog computing load forecasting method produces the sequence
forecast values as (
3)
The following part is devoted to a detailed explanation of
fk-near based on a schematic data structure definition in
Figure 2:
Given a set of historical load power sequences of a certain node, , for instance, starting from any sample, say , continuously intercepting the subsequent samples to form a label training sample, denoted as , where the first q intercepted samples including are called feature samples, denoted as , and the rest sample is called the true load counterpart of the feature sample, denoted as . According to the above definition of label training samples, in theory, no more than label training samples can be extracted from to form a training sample pool, which is . The degree of similarity of label training samples is calibrated by the Euler distance of their feature samples. The closer the value is to 0, the higher the similarity shall be. The k neighbors of a label training sample (or interchangeably, a feature sample) are k numbers of other label training samples (or feature samples) with decreasing similarity.
Referring to the above definition, let
represent the number of neighborhood samples associated with the feature sample
with the closest Euler distance to the input feature sample
,
represent the
kth closest historical feature sample in the neighborhood of
; note that
and
represent the true load counterpart of
at the
j period, and then the load prediction value defined in this article is expressed as [
13]:
where
The connotation of (
4) is that given a training sample pool, the load forecast is defined to be the weighted sum of true value counterparts of
k neighbors of the label training sample with the highest similarity to the feature sample. The similarities are quantified by (
5) and (
6), respectively. Note that the weights differ from those of [
13] in the additional constant of “1”, which is introduced to have the denominator avoiding zero.
When substituting (
4)∼(
6) into (
2) for consecutive load forecasting, forecast errors of individual methods can be monitored based on real-time load. When the error exceeds the prescribed threshold, the newly acquired real-time data can be used to update the label training sample pool. In view of statistical errors of time-series data, this paper uses the Mean Absolute Scaled Error (MASE) as shown in (
7):
Based on eMASE, the error criterion for the sample update can be designed, such as eMASE > 15, so that the incremental learning of the feature sample neighborhood is triggered by the error criterion while keeping the volume of the training sample pool unchanged. The specific steps are given below.
First, define a matrix to collect quantitative similarities between all feature samples in the current sample pool, such as
where
represents the similarity between feature samples
i and
j. Then, define a new column vector collecting the similarity between
q newly arrived real-time power demands and every feature sample in the current sample pool, denoted by
, bring it into (
9), and replace
in the current sample pool with
in the case of
:
The following method can be used to minimally update
to save the computational overhead based on (
9). First, define a vector
, and then replace the column
and row
of
with
and
respectively.
In conclusion,
fk-near, developed from [
13] in this paper, is intended for complement its father method. The complementation is mainly reflected in the fact that in the case of load sequences with relatively gentle trend changes, both methods utilize a small number of samples in the neighborhood to quickly output prediction (see (
4)) through information fusion, maintaining high accuracy and prediction efficiency. On the other hand, when the load sequence mutates, the real-time cumulative prediction error
eMASE monitored by
fk-near,j, once beyond a threshold, will trigger the replacement and update of neighborhood samples, and compels
fk-near,j to use the latest true load values instead of some historical ones long time ago to sustain the feature diversity of the limited training samples. Both embedded methods use real-time data for incremental self-learning. The predicted values of the two methods coincide in the early stage of the training pool, albeit they are diverse over time due to different mechanisms adopted by the two methods to update their training pools, respectively. Based on the short-term lookback performance of the prediction errors of the two methods, the weights of the two methods are consecutively learned via (
2) toward the best, as shown later, and the stability of the overall prediction performance can be maintained much better.
4. Self-Healing Strategy of the LVPSA
The self-healing strategy of the LVPSA leverages the load prediction curve from each light-power cabinet’s fog computing unit during the fault repair period, the current status of low-voltage switches, and the estimated fault duration as inputs. These inputs are used to determine the optimal operation strategy for the low-voltage remote-controlled breakers. The specific model is detailed as follows.
The optimization goal is set to minimize the maximum value of the total power loss plus three-phase power imbalance subject to the worst load forecast errors; that is,
where
denotes the joint probability distribution of node load demand forecast errors, which falls into the ambiguity set defined by the type-
p Wasserstein ball with a radius of ϵ. The nominal distribution involved in defining
will be discussed later.
represents the node number of the concerned LVPSA, and
M represents the penalty coefficient of load shedding. Considering that it is not common for low-voltage power demand to be dispatched smoothly, binary
is introduced to indicate the breaker operation of load
i at moment
t.
means that the load is fully supplied, while
means disconnecting the load with the grid.
is the fog computing node power prediction; that is,
, and
is the forecast error following
. Finally,
and
and
are auxiliary variables introduced to achieve three-phase load balancing at all times.
4.1. Phase Power Balance Constraints
Since a low-voltage user is generally supplied through a single phase, an adjustment of the network switch status may alter load distribution from phase to phase. Three-phase load balance should always be maintained as much as possible. To do this, set constraints
where
denotes the phase sending power of the transformer.
4.2. Kirchhoff’s Law
Considering that line parameters in low-voltage station areas are generally difficult to obtain accurately [
19], this paper proposes line-parameter-free power flow constraints as shown in (
12):
where
represents the branch power injected into node
j,
represents the power-source injected power into node
j, and
or
represent the set of nodes connected by the branches sharing node
j as the power receiving or sending end.
is the forecast error, which follows the joint probability distribution
. The branch power loss needs to be accounted for according to the branch power direction. The specific principle is explained as follows by referring to
Figure 3. Suppose that the reference direction of branch power
is flowing into node
j: if the actual flow direction is consistent with the positive direction (
Figure 3a), then according to the assumption ②, the branch loss is calculated as
. On the other hand, when the actual flow direction is opposed to the reference (
Figure 3b), the power loss is included in
, and thus it should be
. Therefore, the constraints on
can be specifically implemented through the big-M method as follows. First, an auxiliary binary variable
is introduced to represent the branch power direction. When it is consistent with the specified reference direction, it is taken as
; otherwise,
. The specific expression is given in (
13):
where
denotes the branch rated capacity. Based on the definition of
, the above constraints on
can be realized as
4.3. Switch Status Constraints
Without a loss of generality,
represents the operating status of the branch switch between node
m and node
n at time
t.
represents the initial state of the switch after fault isolation. The set of nodes either without automatic switches or isolated due to a fault is represented by
. Therefore, the switch status constraints should firstly be expressed as
During the fault recovery period, to avoid the risk of secondary power outage to restored users, which could be ascribed to frequent low-voltage switch operations, it is stipulated that “all switches will operate at most once during the fault self-healing period”. Specially,
where
denotes the set of complete nodes.
At the same time, in order to always keep the network operating in a radial manner, we can identify all the possible rings. Note that there should be at least three nodes comprising a ring due to the topology feature of the LVPSA under study. Denote the node set belonging to the
ith ring by
, then
where Card(·) denotes the cardinality operator.
Compared to the virtual line topology constraint method [
18], the radiation topology limitation method, which utilizes (
18) in conjunction with (
16), offers several advantages. It eliminates the need to introduce additional auxiliary variables into the model, reduces the feasible domain of integer variables, and effectively prevents the formation of any loops, regardless of the status of branches and switches. All loops can be pre-identified offline after the installation of facilities such as switches and lines, thereby avoiding the consumption of online computing resources. Furthermore, constraints on the state linkage operation of three-phase switches can be incorporated as needed.
4.4. Boundary Conditions
Equation (
19) represents the branch active power and loss constraint restricted by the switch state, and
is the constant branch power factor due to assumption ③.
In summary, the switch state reconstruction decision model based on fog computing load prediction proposed in this article consists of (
10)∼(
19), and the decision variables include
,
,
,
,
, and
. The intact model falls into the “wait-and-see” problem, where the switch states should be decided “here and now”, accounting for the speculated worst effect of the load forecasts. We propose an efficacious method to resolve it in the following section.
4.5. Solution Algorithm
First, we lump all the decision variables in the form of vectors denoted by bold symbols, respectively, and compact the original model (
10)∼(
19) as a three-stage “Master-Slave” optimization problem below:
And let
, then
where
signifies the resolved value of
from (
20) after the
zth iteration.
Next, let us pick up the thread of defining a necessary nominal distribution for
. In order to impose the prior hypotheses as little as possible, the empirical distribution of forecast errors fitting the latest training sample pool of
fk-near is employed to play the role of the nominal distribution. By doing this, the nominal distribution varies with the incrementally learning training sample pool. Formally,
where δ(·) is the Dirac function and
To solve the above “Master-Slave” problem, substitute
p = 1 into (
21) to obtain a type-1 WDRO problem. As per [
20] (Theorem 9 therein), the type-1 WDRO on the condition of (
22) is equivalent to
Note that (
23) is also a two-stage problem with
involving integers of
and
, and we can relax these integers and add constraints [
21] like
. By doing this, KKT conditions can viably be derived for
to convert (
23) into a single-layer maximization problem.
At last, once upon resolving
from (
23), the column-and-constraint generation algorithm can be employed to add each group of auxiliary variables and constraints with fixed
as well as
into (
20) to proceed with the iteration until convergence criteria, e.g., the gap of the objective value, are satisfied. A relevant flowchart, as shown in
Figure 4, outlines the above resolving process.