Next Article in Journal
AI-Driven Blockchain and Federated Learning for Secure Electronic Health Records Sharing
Previous Article in Journal
An Improved TOPSIS Method Using Fermatean Fuzzy Sets for Techno-Economic Evaluation of Multi-Type Power Sources
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multi-Channel Δ-BiLSTM Framework for Short-Term Bus Load Forecasting Based on VMD and LOWESS

1
School of Electrical and Information Engineering, Changsha University of Science & Technology, Changsha 410114, China
2
State Key Laboratory of Disaster Prevention & Reduction for Power Grid, Changsha University of Science & Technology, Changsha 410114, China
3
Hubei Engineering and Technology Research Center for AC/DC Intelligent Distribution Network, School of Electrical Engineering and Automation, Wuhan University, Wuhan 430072, China
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(23), 4772; https://doi.org/10.3390/electronics14234772
Submission received: 14 November 2025 / Revised: 1 December 2025 / Accepted: 2 December 2025 / Published: 4 December 2025

Abstract

Short-term bus load forecasting in distribution networks faces severe challenges of non-stationarity, high-frequency disturbances, and multi-scale coupling arising from renewable integration and emerging loads such as centralized EV charging. Conventional statistical and deep learning approaches often exhibit instability under abrupt fluctuations, whereas decomposition-based frameworks risk redundancy and information leakage. This study develops a hybrid forecasting framework that integrates variational mode decomposition (VMD), locally weighted scatterplot smoothing (LOWESS), and a multi-channel differential bidirectional long short-term memory network (Δ-BiLSTM). VMD decomposes the bus load sequence into intrinsic mode functions (IMFs), residuals are adaptively smoothed using LOWESS, and effective channels are selected through correlation-based redundancy control. The Δ-target learning strategy enhances the modeling of ramping dynamics and abrupt transitions, while Bayesian optimization and time-sequenced validation ensure reproducibility and stable training. Case studies on coastal-grid bus load data demonstrate substantial improvements in accuracy. In single-step forecasting, RMSE is reduced by 65.5% relative to ARIMA, and R2 remains above 0.98 for horizons h = 1–3, with slower error growth than LSTM, RNN, and SVM. Segment-wise analysis further shows that, for h = 1 , the RMSE on the fluctuation, stable, and peak segments is reduced by 69.4%, 62.5%, and 62.4%, respectively, compared with ARIMA. The proposed Δ-BiLSTM exhibits compact error distributions and narrow interquartile ranges, confirming its robustness under peak-load and highly volatile conditions. Furthermore, the framework’s design ensures both transparency and reliable training, contributing to its robustness and practical applicability. Overall, the VMD–LOWESS–Δ-BiLSTM framework achieves superior accuracy, calibration, and robustness in complex, noisy, and non-stationary environments. Its interpretable structure and reproducible training protocol make it a reliable and practical solution for short-term bus load forecasting in modern distribution networks.

1. Introduction

In modern power systems, the increasing penetration of distributed energy sources and the ongoing electrification of end-use sectors, coupled with the rapid growth of new loads such as centralized charging and demand response, have led to the emergence of non-stationary, highly disturbed, and multi-scale coupled characteristics in bus loads within distribution networks. The frequency of peak mutations and “spike” noise has risen significantly, thereby increasing uncertainty and engineering constraints in short-term forecasting [1,2,3]. Short-term bus load forecasting not only provides boundary conditions for power flow and reactive voltage but also serves as a critical input for reserve scheduling, peak shaving, and electricity price incentives, directly impacting operational safety and economic efficiency [4,5].
To improve stability in noisy and non-stationary scenarios, existing studies generally follow three approaches. The first is statistical or traditional machine learning, which relies on structural priors and handcrafted features. These methods are simple and interpretable, but they struggle to adapt to holidays, weather variations, and ramp-up segments, especially under nonlinear dynamics or when features are incomplete [6,7]. Deep learning, with end-to-end representation learning, captures temporal dependencies with high accuracy. Several studies have explored various deep learning models for load forecasting, such as Recurrent Neural Networks (RNN), Long Short-Term Memory (LSTM) networks, and their bidirectional variants (BiLSTM), which have demonstrated strong performance in handling temporal dependencies and predicting future loads. Additionally, methods like Adaptive Fusion Domain-Cycling Variational Generative Adversarial Networks (VGAN) and Adaptive Variational Autoencoding Generative Adversarial Networks (VAE-GAN) have been investigated, showcasing their potential for improving model robustness and generalization in forecasting tasks. Furthermore, multi-modal techniques, such as those applied in arc detection for railway systems, have shown promise in enhancing forecasting accuracy by incorporating multiple data sources. These approaches highlight the growing interest and success of deep learning models in forecasting applications, particularly in scenarios with complex and dynamic load behaviors [8,9,10,11]. The “decomposition–learning” paradigm, such as empirical mode decomposition (EMD), complete ensemble empirical mode decomposition with adaptive noise (CEEMDAN), variational mode decomposition (VMD), wavelet transform (WT), and seasonal-trend decomposition based on loess (STL), first separates multi-scale components or performs denoising, and then processes each component individually before combining the results. This approach helps reduce component interference and balances high and low frequencies [12,13,14]. However, wavelets suffer from subjectivity in selecting frequency bands and layers, while EMD-based methods are prone to mode mixing and endpoint effects. If all components are modeled and combined without discrimination, noise and redundancy may be introduced, reducing robustness [15,16]. Engineering practice also shows that errors often accumulate in areas with compounded seasonal patterns and short-term pulses, as well as in ramp-up segments during electricity usage peaks, extreme weather, and concentrated charging access [3]. Furthermore, improper prior settings in over-parameterized models can lead to sensitivity to initialization and search range, affecting reproducibility and generalization [17].
The short-term bus load forecasting primarily faces three challenges:
(1)
Multi-scale/Multi-phase coupling. Single-channel end-to-end models face challenges in simultaneously capturing trends, seasonality, and pulses, often resulting in a trade-off between these factors. Full-modeling after decomposition tends to dilute effective signals and increases the risk of overfitting [18,19].
(2)
A discrepancy exists between forecasted targets and actual scheduling requirements. Solely minimizing point value prediction errors fails to adequately constrain change rate or ramp behavior, resulting in error concentration during peak and abrupt-change periods [20].
(3)
Lack of robustness and reproducibility in the process. Decomposing followed by splitting or cross-temporal normalization often leads to information leakage [21]. Mixed frameworks have a high degree of freedom, and results are sensitive to random seeds and training budgets [22,23].
To address these issues, this paper proposes a multi-channel Δ-BiLSTM (Differential Bidirectional Long Short-term Memory Network) framework, beginning with mode extraction using VMD. Residuals are then smoothed and denoised using LOWESS (Locally Weighted Scatterplot Smoothing). Channel selection is performed based on correlations beyond the starting point. Δ-targets are employed to model change rates and phase mismatches via a BiLSTM. Robust training and reproducible evaluation are ensured through lightweight Bayesian optimization and time-sequenced validation. Compared with existing VMD/EMD-based hybrid models and single-channel deep networks, VMD–LOWESS–Δ-BiLSTM uniquely applies VMD–LOWESS only in a leakage-free training pipeline and couples correlation-screened multi-channel inputs with Δ-target BiLSTM under a unified training budget, jointly tackling noise, redundancy, and ramp dynamics.
An overview of the VMD-LOWESS-Δ-BiLSTM forecasting workflow is shown in Figure 1. The innovations of the proposed framework are as follows:
Multi-scale robust input. The method applies “VMD + LOWESS” processing to the training segment to suppress noise and anomalous fluctuations, thereby mitigating interference and preventing data leakage. This approach enables multi-scale input that captures both trends and transient disturbances.
Scheduling-consistent Δ-target learning. By integrating Δ-targets with a BiLSTM network, this approach enhances the model’s ability to capture rate-of-change patterns and ramp transitions, leading to significantly improved prediction accuracy during peak and abrupt-change periods.
Robust training and reproducible evaluation. The framework enhances training stability and ensures result reproducibility through foresight-based channel selection, lightweight Bayesian optimization, and time-sequenced validation.
The remaining structure of this paper is organized as follows: Section 2 outlines the methodological foundations, including VMD, LOWESS residual smoothing, Δ-target construction, LSTM/BiLSTM networks, multi-channel construction, correlation-based channel selection, and Bayesian optimization integrated with robust training strategies. Section 3 presents the integrated framework, which combines the aforementioned modules into a multi-channel Δ-BiLSTM architecture. It further elaborates on the end-to-end process and describes the leakage-prevention training/validation protocols. Section 4 covers experiments, including data preprocessing, parameter configurations, and evaluation metrics. It compares the proposed approach against baseline models, assessing performance during peak and abrupt-change periods as well as computational complexity. Section 5 gives the conclusions.

2. Methodological Foundations

2.1. Variational Mode Decomposition

Variational mode decomposition (VMD) is a widely used signal processing method for noise reduction in power systems. It works by iteratively solving for mode functions at different time scales and their corresponding central frequencies, resulting in several intrinsic mode functions (IMFs). Through selective reconstruction of these IMFs, the noise reduction effect can be achieved. The denoising process using VMD for bus load can be divided into two parts: construction and decomposition [24].
For the noise and spikes present in the original bus load sequence, VMD is used to stabilize the noisy load sequence, decomposing it into multiple stationary subsequences and an unstable residual. To prevent optimistic evaluation results caused by information leakage, a leakage-prevention process is applied [25]. Finally, reconstruction with the IMFs results in a multi-channel input that preserves both scale information and robustness.
All numerical experiments and model training in this study were implemented using MATLAB R2023b and Python 3.11.8.

2.1.1. Sub-Sequence Analytic Signal and Frequency Mixing

The rules for VMD are as follows:
(1) Let the noisy bus load sequence be X t , where t     [ 1 ,   T ] , and T is the length of the noisy bus load sequence.
(2) The noisy bus load sequence X ( t ) is decomposed by VMD into subsequences u k ( t ) , where k     [ 1 ,   K ] , and K represents the number of subsequences after decomposition.
(3) The sub-sequence u k ( t ) is converted into an analytic signal via the Hilbert transform:
[ δ t + j π t ] u k t
where δ ( t ) is the Dirac delta function, * denotes the convolution operator, and u k ( t ) is the sub-sequence of the bus load sequence after applying VMD;
(4) The analytic signal is demodulated to shift the frequency component of each mode to the baseband:
δ t + j π t u k t e j ω k t
where ω k is the central frequency.

2.1.2. Sub-Sequence Baseband Minimum Bandwidth Model

After converting the analytic signals of the bus load subsequences to the baseband, the objective is to obtain the minimum bandwidth frequency. A decomposition constraint condition is established, and the Lagrangian operator is introduced for iterative calculation:
u k , ω k , λ = k = 1 K t δ t + j π t u k t e j ω k t 2 2 + X t k = 1 K u k t 2 2 + λ t , X t k = 1 K u k t
where λ ( t ) is the Lagrangian multiplier.
Initialization is performed for u ^ k 1 , ω k 1 , and λ ^ 1 . When ω 0 , u ^ k , ω k , and λ ^ n are subsequently updated:
u k n + 1 ω = f ^ ω i k u ^ i n + 1 ω + λ ^ n ω 2 1 + 2 ω ω k n 2
ω k = 0 ω u ^ k n + 1 ω 2 d ω 0 u ^ k n + 1 ω 2 d ω
λ ^ n + 1 ω = λ ^ n ω + f ^ ω k u ^ k n + 1 ω
Until the condition k u ^ k n + 1 u ^ k n 2 2 / u ^ k n 2 2 < ε is satisfied, where ε is the preset error tolerance and n is the number of iterations.

2.1.3. Sub-Sequence Smoothing Optimization

The bus load sequence X(t) after VMD is expressed as follows:
X t = i = 1 K 1 I M F i t + r t
where I M F i t represents the decomposed intrinsic mode functions, and r t is the residual.
The residual of the bus load sequence is smoothed using the LOWESS method, which operates by assigning weights to proximal data points within a local section to ensure a smooth fit. The local polynomial regression equation for the smoothed residual sequence r t of the bus load sequence after VMD is as follows:
r t = a x i + b , i 1 , T
where xt is the variable of the sequence, a and b are the regression weights of the sequence.
The sequence weights of this smoothed sequence regression are solved by defining the cost function as follows:
J a , b = 1 N i = 1 N ω i y i a x i b 2
where N is the number of neighboring points set in the smoothing operation, and ω i is the cubic weight of the point x i relative to its neighboring points within the operation range.
The cubic weighting function is expressed as follows:
ω i x = 1 x x i d i 3 3
where x is the current center point, x i is a neighboring point within the bandwidth, and d i is the distance from x i to the current point along the horizontal axis.
The cubic weights obtained above are used for locally weighted linear regression. The parameters a and b are obtained by setting the partial derivatives of the cost function J a , b with respect to a and b to zero:
a J a , b = 1 N 2 i = 1 N ω i x i y i a x i b = 0 b J a , b = 1 N 2 i = 1 N ω i y i a x i b = 0
Through Equation (11), the regression equation gives the predicted value y ^ i corresponding to x i . After moving to the next point x i + 1 , the iterative step is repeated to obtain the prediction value y ^ i + 1 .
The smoothed residual is then reconstructed together with the decomposed subsequences. r t and I M F i t are combined to reconstruct the denoised bus load sequence L o a d t :
L o a d t = i = 1 n I M F i t + r t

2.1.4. Multi-Channel Input and Leakage Prevention

Let the original sequence { X ( t ) } t = 1 T be divided into training, validation, and testing segments in chronological order:
D tr = { 1 ,   , T tr } , D va = { T tr + 1 ,   , T va } , D te = { T va + 1 ,   , T }
where Ttr, Tva, and T denote the endpoints of the training, validation, and total sequence, respectively. Dtr, Dva, and Dte are the index sets of the three time segments.
At each stage l { tr , va , te } , only the visible prefix { X ( t ) } t T l is used to independently perform VMD and residual LOWESS, yielding { X ( t ) } k = 1 K and r ( l ) ( t ) , respectively. This then yields the following:
X V + L ( l ) ( t ) = k = 1 K I M F k ( l ) ( t ) + r ~ ( l ) ( t ) , t T l , l { tr , va , te }
where I M F k ( t ) ( t ) denotes the k-th mode obtained by applying VMD on the historical prefix [ 1 ,   T l ] of the sequence x ( t ) . K is the number of modes. r ( l ) ( t ) is the residual from the same decomposition. r ~ ( l ) ( t )   =   LOWESS ( r ( l ) ( t ) ) is the smoothed residual by LOWESS. X V + L ( l ) ( t ) represents the reconstructed denoised sequence by “VMD + LOWESS”.
The K modes and one smoothed residual are combined into C   =   K   +   1 channels. The channel vector is defined as follows:
z ( l ) ( t ) = [ I M F 1 ( l ) ( t ) ,   , I M F K ( l ) ( t ) , r ( l ) ( t ) ] T R C , C = K + 1
With window length L , the input tensor is constructed as follows:
X t ( l ) = [ z ( l ) ( t L + 1 ) ,   , z ( l ) ( t ) ] R C × L , t = L ,   , T l H
The corresponding point-wise prediction target is defined as follows:
y t ( l ) = [ X ( t + 1 ) ,   , X ( t + H ) ] T
where z ( l ) ( t ) is the channel vector, X t ( l ) is the input tensor of length L ending at window t , and y t ( l ) is the point-wise observation target of length H .
The mean and standard deviation for each channel are calculated only from the training set samples { X t ( tr ) } , and are then applied consistently across all stages for normalization:
X ~ t ( l ) ( c , · ) = X t ( l ) ( c , · ) μ c σ c + ε n u m , c = 1 ,   , C , l { tr , va , te } .
where μc and σc are the mean and standard deviation of channel c , respectively. These values are calculated exclusively from the training data { X t ( tr ) } , and then fixed for validation and testing. The term ε n u m   >   0 is a numerical stabilizer, and X ~ t ( l ) denotes the normalized input.

2.2. Recurrent Neural Networks

A bidirectional long short-term memory network (BiLSTM) is a variant of LSTM that can learn contextual information from both forward and backward directions. It trains sequential data more effectively through full feature extraction compared with traditional recurrent neural networks. Under the integration of multi-channel inputs and in combination with Δ-target learning, the BiLSTM model is well suited to handle the highly non-stationary and noisy characteristics of bus load data, achieving improved forecasting accuracy and stability.

2.2.1. LSTM Unit

Suppose there are k synchronous input sequences. The LSTM unit employs three gates: a forget gate, an input gate, and an output gate. The input vector xt carries the network’s input values; the hidden state ht delivers the output to the next layer; the cell state ct stores the internal state of the LSTM unit for long-term memory.
The LSTM steps are as follows:
(1) Forget gate update. It determines how much of the past memory should be retained.
f t = σ ( W f x t + U f h t 1 + b f )
(2) Input gate update. It regulates how much new information should be written into the cell state.
i t = σ W i h t 1 + U i x t + b i
c ˜ t = tanh ( W c x t + U c h t 1 + b c )
(3) Cell state update. It combines the retained memory and the newly added candidate memory.
c t = f t c t 1 + i t c ˜ t
(4) Output gate update. It determines how much of the memory will be exposed at the current time step.
o t = σ ( W o x t + U o h t 1 + b o )
(5) Hidden state update. It calculates the proportion of memory exposed to the output at the current time step.
h t = o t tanh ( c t )
(6) Final prediction update. It generates the current predicted output.
y ^ t = σ V h t + c
where c t is the cell state, h t is the hidden state, x t is the input and W, U, and b are the recurrent weight, input weight, and bias, respectively. σ x is the activation function of each gate, which maps the input into a probability value.
By dynamically balancing the operations of “retain-write-read” through the gating mechanism, the LSTM model can stably propagate information across relatively long time scales.

2.2.2. BiLSTM Architecture

Based on the above, within a fixed-length window, “backward” contextual information is introduced to enhance discrimination of local patterns without crossing the window boundary.
h t = LSTM   ( x 1 : t ) , h t = LSTM   ( x L : t )
where LSTM and LSTM represent the forward and backward LSTM units applied within the window [ t L + 1 , t ] , ensuring that no information outside the window is accessed.
The forward and backward hidden representations are then concatenated to form a more complete temporal representation of the window:
h t ( bi ) = [ h t ; h t ] R 2 H

2.2.3. Δ-Target and Multi-Channel Sample Organization

To use the bidirectional representations produced by BiLSTM within a single window for regression prediction, the hidden states over the window are first aggregated into a fixed-length representation, yielding the vector at the input of the regression head:
g t = Pool ( { h τ ( bi ) } τ = t L + 1 t ) R 2 H
where { h τ ( bi ) } are the bidirectional hidden states and Pool ( ) by default assigned the last hidden state h τ ( bi ) . Alternatively, average pooling or other strategies can be applied within the window, provided that the boundary is not exceeded.
The regression head then maps g t linearly to predict the “future h-step increment”:
Δ y ^ t + h = w g t + b .
where w R 2 H , and b R are trainable parameters.
To enhance the aggregation representation, an attention mechanism can be introduced within the window to weight different time steps adaptively:
e τ = v tanh ( W a h τ ( bi ) ) , α τ = exp ( e τ ) κ exp ( e κ ) ; g t = τ α τ h τ ( bi )
where W a , v are the attention parameters, and α τ is normalized only within the window to ensure leakage-free evaluation.
To strengthen the modeling of change rates and ramp points, this paper reformulates the point-wise prediction into a relative difference target:
Δ y t + h = y t + h y t + h 1
The mean squared error (MSE) for optimization.
L MSE = 1 N 0 n = 1 N ( Δ y ^ t n + h Δ y t n + h ) 2
where N0 is the number of samples. During the inference stage, the ground truth value at the end of the window is used as the baseline, and the accumulated increments are added back to restore the point prediction:
y ^ t + h = y t + j = 1 h Δ y ^ t + j
where yt is the true load at time step t. Multi-step prediction can be performed either recursively or in parallel, depending on requirements.
Finally, to ensure that the input sufficiently incorporates multiple scales as well as exogenous factors, we construct multi-channel window tensors and apply leakage-free normalization. The concatenated form of the window channels is as follows:
X t = x t L + 1 : t | r ~ ( t L + 1 : t ) | u k 1 ( t L + 1 : t ) , , u k M ( t L + 1 : t ) | z t L + 1 : t
where r ˜ denotes the residual smoothed by LOWESS within the window, x is the original sequence, u k m are the M IMF channels selected from VMD, z denotes exogenous variables (e.g., calendar factors), and indicates concatenation across channels. To prevent statistical leakage, the mean and standard deviation of each channel are estimated only on the training block, then fixed and applied to validation/testing. This yields the following:
X t ( c ) = X t ( c ) μ c train σ c train , c = 1 , , C sel
where X t ( c ) denotes the sub-sequence segment of length L for channel c. μ c train and σ c train are the mean and standard deviation of channel c, respectively, computed only from the training block.
The model finally takes X t ( c ) as input, outputs Δ y ^ t + h and restores the point prediction y ^ t + h through Equation (33).

2.3. Channel Value Metrics and Selection

2.3.1. Forward Correlation

For the c-th candidate channel sequence in the training stage, denoted as x c ( tr ) ( t ) , the target is the original point value sequence X ( t ) . To evaluate its contribution after the “forecast origin,” forward correlations are defined for each horizon:
ρ c ( h ) = corr x c ( tr ) ( t ) , X ( t + h ) , t = L , , T tr h
where H is the maximum forecasting horizon.
Here, the correlation coefficients are computed on the training samples. The inputs and targets are standardized using the means and standard deviations obtained from Equation (18), i.e., with z-score normalization. To account for multiple-step forecasting, the absolute correlations at different horizons are aggregated with weights { w h } h = 1 H , yielding a single score for channel c .
s c = h = 1 H w h ρ c ( h ) , h = 1 H w h = 1
To capture potential nonlinear dependencies, mutual information and gray relational degree can also be computed and fused with weights:
I c = h = 1 H w h MI x c ( tr ) ( t ) , X ( t + h ) , γ c = h = 1 H w h γ c ( h ) γ c ( h ) = 1 N h t Δ min + ζ Δ max Δ c ( t , h ) + ζ Δ max , Δ c ( t , h ) = X ( t + h ) x c ( tr ) ( t ) , ζ ( 0 , 1 ]
where γ c ( h ) is computed based on the difference sequence Δ c ( t , h ) using a three-parameter formulation. Nh is the number of comparable samples.
All three metrics are normalized to the range [0, 1], with the maximum value across channels set to 1. The comprehensive score is then given by the following:
s ˜ c = α s ^ c + β I ^ c + ( 1 α β ) γ ^ c , α , β [ 0 , 1 ]

2.3.2. Redundancy-Aware Channel Subset Selection

The number of candidate channels is C = K + 1 . Given the channel budget M budget , the upper bound is set as follows:
M max = min C , M budget
First, channels are ranked by their scores s ˜ c , and the top M 0 M max are retained for secondary selection. To reduce redundancy caused by overlapping frequency bands, a forward stepwise selection with a penalty is applied. Specifically, given the current selected set S , each step selects:
c = arg max c S s ˜ c λ 1 | S | j S corr x c ( tr ) , x j ( tr ) , λ [ 0 , 1 ]
Until the target number of channels is reached | S | = M M max . In the equation λ [ 0 , 1 ] controls the strength of redundancy penalization, and the correlations are calculated using the training samples. Denote the final selected set by S = { c 1 ,   , c M } .

2.4. Lightweight Bayesian Optimization and Stable Training

To avoid large-scale brute-force search and to ensure fair comparison under a unified training budget, this study adopts lightweight Bayesian optimization on the validation set targets. Stable training protocols such as early stopping, gradient clipping, regularization, and fixed random seeds are used to guarantee reproducibility and comparability within the leakage-free boundary.

2.4.1. Objective Functions and Evaluation Protocol

After determining the channel set and constructing the multi-channel input windows, the performance of the Δ-BiLSTM model mainly depends on the hyperparameter set θ. The hyperparameters most relevant to training and sensitive to performance are included in the search space:
θ = { hidden , lr , L 2 , dropout , clip , batch } Ω
For a given hyperparameter set θ, after training the model on the training set, evaluation on the validation set is performed under the “forward” protocol. For each horizon h , the index set of comparable validation samples at the horizon h in the validation stage is as follows:
I va ( h ) = { t L t T vae h }
where Tvae denotes the endpoint of the validation set, and L is the input window length.
The corresponding root mean squared error (RMSE) at horizon h during the validation stage is calculated as follows:
RMSE h ( va ) ( θ ) = 1 | I va ( h ) | t I va ( h ) y ^ t + h ( θ ) y t + h 2 1 / 2
The point-wise predictions are restored by accumulating increments:
y ^ t + h ( θ ) = y t + j = 1 h Δ y ^ t + j ( θ )
where y ^ t + h denotes the point forecast of y t + h , and the baseline value yt is taken from the end of the sliding window to ensure causality.
The step-weighted validation objective J ( θ ) is aggregated:
J ( θ ) = h = 1 H w h RMSE h ( va ) ( θ )
where the step weights w h are non-negative and normalized to unity.
When the difference in J between two hyperparameter sets θ 1 and θ 2 is smaller than a tolerance ε, symmetric mean absolute percentage error (sMAPE), and mean absolute error (MAE) are further used as tie-breakers:
sMAPE h ( va ) ( θ ) = 100 % | I va ( h ) | t I va ( h ) | y ^ t + h ( θ ) y t + h | 1 2 | y ^ t + h ( θ ) | + | y t + h |
MAE h ( va ) ( θ ) = 1 | I va ( h ) | t I va ( h ) | y ^ t + h ( θ ) y t + h |
To further improve robustness, the validation set can be partitioned into Q consecutive sub-windows, where metrics are computed for each sub-window and then aggregated. The stability-penalized evaluation across sub-windows J ¯ ( θ ) is given by the following:
J ¯ ( θ ) = 1 Q q = 1 Q J q ( θ ) + η 1 Q q = 1 Q J q ( θ ) J ¯ ( θ ) 2 , η 0
where J q ( θ ) is the objective value on the q-th sub-window, and η 0 is the penalty weight for fluctuation.

2.4.2. Search Space and Lightweight Search Strategy

Bayesian optimization models J ( θ ) with surrogates, recording the mean and standard deviation of the objective function after the t-th round as μ t ( θ ) and σ t ( θ ) , respectively.
The acquisition function selects the next evaluation point, typically using expected improvement (EI) or lower confidence bound (LCB). For the minimization problem, EI is defined as follows:
EI t ( θ ) = ( J min , t μ t ( θ ) ) Φ J min , t μ t ( θ ) σ t ( θ ) + σ t ( θ ) ϕ J min , t μ t ( θ ) σ t ( θ )
where J min , t = m i n i t J ( θ i ) , Φ ( ) and ϕ ( ) denote the cumulative distribution function and probability density function of the standard normal distribution, respectively.
LCB is expressed as follows:
LCB t ( θ ) = μ t ( θ ) κ t σ t ( θ ) , κ t > 0
To embody the lightweight search strategy, each training run adopts the same budget set:
B = { E max , P , N iter }
where E max is the maximum number of epochs, P is the early stopping patience, and N iter is the number of iterations per epoch.
This ensures comparability across different θ . The final optimal solution is recorded as follows:
θ = arg min θ Ω J ¯ ( θ ) ( o r   J ( θ ) )

2.4.3. Stable Training and Reproducibility Protocols

All evaluations strictly follow the leakage-free boundary defined in Section 2.1.4. Channel normalization statistics are computed only during the training stage as defined in Equation (18), and remain fixed for the validation and test stages. The training loss adopts the weighted mean squared error on the Δ-target:
L train ( θ ) = h = 1 H λ h 1 I tr ( h ) t I tr ( h ) Δ y ^ t + h ( θ ) Δ y t + h 2 , h = 1 H λ h = 1
When combined with L2 regularization and dropout:
L ( θ ) = L train ( θ ) + L 2 W 2 2
During backpropagation, L2-norm gradient clipping is applied to prevent exploding gradients:
g g min 1 , clip g 2
Early stopping is also applied. If the validation objective shows no improvement within P consecutive rounds (with tolerance ε), training is terminated, and the model with the best validation weight is retained. All evaluations are conducted with fixed random seeds and device settings to ensure reproducibility. Testing is performed only once, strictly using the θ obtained from the training–validation procedure, along with the corresponding normalization parameters and the selected channel set S .

3. Combined Forecasting Framework

3.1. Overall Procedure and End-to-End Mapping

At each time step t, within the causal sliding window { t L + 1 , , t } , the “VMD + residual LOWESS” procedure is executed to obtain K band-limited components and one smoothed residual [26]. Together they form a candidate feature vector of dimension C = K + 1:
IMF 1 ( t ) , , IMF K ( t ) , r ˜ ( t ) = D K , α , τ , tol , Niter , span X t L + 1 : t R C
where D denotes the operator for “VMD + residual LOWESS”, and r ˜ ( t ) is the residual smoothed by LOWESS within the window. The parameters K , α , τ , tol , Niter and the LOWESS span are predetermined or chosen from small candidate sets, focusing the search on core hyperparameters of the learner.
Candidate channels are standardized independently during training using z-score statistics, and then selected according to the forward correlation measures and redundancy control in Section 2.3, yielding M effective channels. The c-th channel after normalization x ˜ c ( t ) is given by the following:
x ˜ c ( t ) = x c ( t ) μ c ( tr ) σ c ( tr ) , S = Select { s ˜ c } c = 1 C , M , λ
where μ c ( tr ) and σ c ( tr ) are the mean and standard deviation estimated and fixed during the training stage, respectively. Select ( ) enotes the “TOP-M + redundancy penalty” channel selection.
According to S , the multi-channel input tensor is constructed and trained with the Δ-target using a BiLSTM architecture:
X t = [ x ˜ t L + 1 : t ( c ) ] c S R L × M , Δ y ^ t + 1 : t + H = f θ ( X t )
where X t is the input tensor of length L and dimension M, and f θ ( ) denotes the Δ-BiLSTM model trained under configuration θ , outputting the H -step incremental prediction Δ y ^ t + 1 : t + H .
By compressing the above steps into a single operator, the end-to-end mapping is expressed as follows:
y ^ t + 1 : t + H = R ° f θ ° W S ° N μ ( tr ) , σ ( tr ) ° D K , α , τ , tol , Niter , span X t L + 1 : t
where N μ ( tr ) , σ ( tr ) denotes per-channel normalization driven by training stage statistics, W S represents channel selection with window concatenation, R indicates the “Δ-to-point” restoration, and D K , α , τ , tol , Niter , span is the VMD + residual LOWESS decomposition operator. The optimal hyperparameters are determined by the unified objective function in Equation (53).

3.2. Training and Validation Protocol

Building on the mapping in Section 3.1, this section specifies the protocol for non-leakage training–validation testing, and the model selection criteria. The RMSE at horizon h as defined in Equation (44) is computed, and the aggregated validation objective is given by Equation (46). When robustness is required, Equation (49) is used. Based on these computations, θ is obtained, and the validation-optimal weight is fixed as the sole checkpoint for testing [27].
The test procedure is executed only once and is excluded from any selection or early stopping. The data generation and normalization procedures of the validation stage are replicated. Δ y ^ t + 1 : t + H is subsequently obtained using f θ ( X t ) , and a “Δ-to-point” restoration is performed according to Equation (45). This approach ensures no information leakage while guaranteeing fairness, comparability, and reproducibility.

3.3. Complexity Analysis

The end-to-end framework mainly consists of “causal decomposition + channel scoring + Δ-BiLSTM forward/backward training”. At any reporting time t, within the causal window { t L + 1 ,   , t } , the computational complexity of VMD and residual LOWESS can be approximated as follows:
cost VMD + LOWESS O ( K L log L Niter ) + O ( L )
where Niter is the number of iterations in the VMD process.
Channel scoring and selection are only performed during training. Its overall scale is determined by the candidate channel number C = K + 1, sequence length H, and training length. The complexity can be approximated as follows:
cost score O ( C H T tr ) + O ( M ( C M ) T tr )
The former corresponds to forward-looking correlation or mutual information, whereas the latter refers to forward-looking selection with redundancy penalties. In comparison, the computational cost of the network training can usually be ignored.
For the bidirectional LSTM, the single-layer forward and backward unfolding with sliding window length L, channel number M, and hidden units hidden , has a complexity approximately given by the following:
cost BiLSTM   per   batch O 2 L hidden 2 + hidden M
where the constant 2 arises from the bidirectional structure.
Under the unified training budget of maximum epochs and batch size, the training time scale satisfies the following:
T train E N iter L hidden 2 + hidden M
The main computational cost is determined by the gate states and parameter matrices. It scales approximately quadratically with the hidden dimension, and linearly with both L and M. The inference phase only involves forward propagation, with the same order of complexity as above but with a smaller constant factor.

4. Case Study Analysis

The load power measurement data of the main transformer bus in Area #1 was selected as the dataset, containing a total of 864 time steps, with one time step equal to 5 min. This dataset represents a coastal region along China’s southeast coast with abundant wind and photovoltaic resources, characterized by high renewable energy penetration. Compared to inland regions with lower renewable integration, it captures the complex load fluctuations and system challenges typical of high renewable penetration areas, providing a more accurate reflection of their impact on power system behavior [28].
The dataset includes three typical load fluctuation days. However, due to the similar fluctuation patterns observed across the three days, and to enhance computational efficiency, 400 time steps, totaling 34 h of data, were selected, which include a typical day. This selection was made to ensure that the data remained representative of typical operating conditions while reducing the computational burden.

4.1. Time Series Decomposition and Residual Smoothing

4.1.1. VMD Results and Multi-Scale Features

The case study data was chronologically divided into training, validation, and test sets, comprising 70%, 15%, and 15% of the total data, respectively. Only the training set statistics were used for differencing and normalization, so as to avoid information leakage.
During the training stage, VMD was applied, yielding six intrinsic mode functions (IMFs) and one residual component. Figure 2 illustrates the VMD of the training series into six IMFs and a residual. The higher-order IMFs capture short-term fluctuations and noise, whereas the residual and lower-frequency components preserve the long-term trend and smooth variations in the load.
As shown in Figure 2:
(1) IMF 1 IMF 3 capture intraday low-frequency fluctuations, while IMF 4 IMF 6 gradually shift to higher central frequencies, identifying peaks and rapid oscillations.
(2) The residual mainly reflects slow-varying trends and contains ultra-low-frequency or slowly changing information not captured by the IMFs .
A comparison is conducted between the decomposed and the original sequences to verify correct decomposition and reconstruction, numerical stability, and consistency.
Figure 3 presents the reconstruction performance of VMD across the train, validation, and test segments. The reconstructed series closely overlaps with the original load curves in all phases, indicating that the selected IMF components and residual effectively preserve both short-term variations and long-term trends without introducing noticeable distortion.
As shown in Figure 3, the VMD reconstruction signal almost entirely overlaps with the original load profile across the training, validation, and test sets, preserving trends and turning points. In the training segment, there is a dip between samples 50–70 and peaks around 120–140 and 190–220. The validation segment declines smoothly, while the test segment rises. This confirms the fidelity of the decomposition and the numerical stability of the reconstruction.

4.1.2. Channel Value Evaluation and Selection

As confirmed in Figure 3, the full VMD reconstruction (∑IMFs + Residual) shows nearly complete overlap with the original series; the focus thus shifts from decomposition validation to channel utilization. On the training segment, each decomposed component (i.e., IMF1-IMF6 and the residual) is evaluated using two complementary criteria, and a compact subset is retained. The results are summarized in Table 1.
(i)
Foresight correlation
With a window length of L = 24, multi-step incremental targets are constructed as follows:
Δ y h ( t ) = y ( t + h ) y ( t ) , h { 1 , 2 , 3 }
For each channel F i ( t ) , after alignment (discarding the first L samples), the absolute Pearson correlations | r i ( h ) | = corr ( F i , Δ y h ) are computed and then averaged across all horizons. This yields the following:
Score i corr = 1 3 h = 1 3 | r i ( h ) |
The absolute value measures the strength of the linear relationship regardless of sign, and the use of future increments prevents information leakage.
(ii)
Energy ratio
For each channel, energy is defined as E i = t F i ( t ) 2 , and the energy ratio is given by E i / E y , where E y = t y ( t ) 2 denotes the energy of the raw series. Since VMD modes are not orthogonal, this ratio serves as an indicator of prominence relative to the raw energy rather than a normalized share. Thus, the sum of ratios across modes need not equal one and may exceed unity, as observed for IMF5 and IMF6 in Table 1. This metric is used jointly with correlation to avoid retaining components that are highly correlated but energetically negligible.
(iii)
Selection rule (Top-M)
Channels are ranked by Score i corr and ties are broken by the energy ratio. The Top-M channels, where M = 6, are retained (the residual is treated as a candidate channel but is specifically smoothed in Section 4.1.3). The “Selected” column in Table 1 marks the retained channels with a 1.
According to Table 1, IMF2 attains the highest foresight correlation on the training segment (≈0.5236), followed by IMF1 (≈0.3427) and IMF3 (≈0.3203). The Residual shows moderate correlation (≈0.1982) and is retained both by the selection rule and for subsequent smoothing. In contrast, IMF4 exhibits very weak correlation (≈0.0185) and a negligible energy ratio (≈1.29 × 10−4), making it the only channel not selected. Although IMF5 and IMF6 display relatively lower correlations (≈0.2063 and ≈0.0216, respectively), their energy ratios are large (≈3.015 and ≈1.506), indicating significant oscillatory content on the raw energy scale. Retaining them helps preserve multi-scale structure for ablation and reconstruction. Thus, the retained set is {IMF1, IMF2, IMF3, IMF5, IMF6, Residual}.

4.1.3. Residual LOWESS and Spike Suppression

Residuals are smoothed only by applying LOWESS to suppress slowly varying noise, while preserving the multi-scale structural features already captured by the IMFs . Using the training segment as the benchmark, the reconstruction error RMSE and roughness are jointly considered across spans to evaluate the reconstruction. The optimal span is selected according to the following objective function J = λ RMSE norm + 1 λ Roughness norm λ = 0.4 .
Figure 4 shows the span selection process for LOWESS, where RMSE to the raw training sequence (blue curve) and roughness measured by second-difference energy (orange curve) are jointly evaluated. As the span increases, RMSE grows monotonically while roughness decreases sharply and then stabilizes, indicating a trade-off between fidelity to the raw signal and smoothness of the residual.
From Figure 4, it can be observed that the comprehensive index reaches its minimum when span ≈ 9. This span is then fixed and subsequently applied consistently during re-decomposition and reconstruction at the validation and testing stages. Consequently, the residual component and its corresponding LOWESS-smoothed counterpart can be obtained for comparison.
The residual and LOWESS comparison is shown in Figure 5. As shown in Figure 5, across the training, validation, and test panels, the residual is overlaid with its residual–LOWESS counterpart. Peaks, troughs, and turning points remain aligned. The residual and the LOWESS curves almost overlap, indicating that no significant structural bias has been introduced. Meanwhile, compared with the residual, the curve after LOWESS processing is smoother, showing better stability and less fluctuation.
To intuitively illustrate the advantages of the proposed processing method, four curves are compared during the training stage: the original series, VMD complete reconstruction, VMD-LOWESS, and the weighted moving average (WMA, as a baseline).
In Figure 6, the original raw signal on the training segment is contrasted with three denoising methods: plain VMD (without LOWESS), a weighted moving average (WMA) baseline, and the proposed VMD-LOWESS. The upper-right inset replots the curves within the magnified region, while the lower-right inset reports the MAE computed over the same region.
From Figure 6, it can be observed that after the “VMD + LOWESS” processing, spikes and local fluctuations in the input sequence are effectively suppressed, while the integrity of the overall trend and extreme values is maintained. This aligns with the proposed multi-scale denoising input method, which provides cleaner, more stable, and leakage-preventive input features for subsequent model construction.
During the training stage, a fine-grained grid search over K × span was further conducted. Figure 7 illustrates the composite objective J evaluated on the training segment across a grid of LOWESS span and VMD modes. The color and height encode the normalized magnitude of J, where lower values indicate better performance.
As shown in Figure 7, although roughness decreases with the increase of span , the RMSE rises, leading to a deterioration of the composite metric J . This indicates that excessive smoothing will cause the loss of useful information. When K increases from 3 to 6, the objective J is significantly improved, but further increases lead to diminishing returns and higher decomposition costs.
In summary, setting K = 6 and span 9 provides a favorable trade-off between accuracy and computational complexity.

4.2. Experimental Setup and Evaluation

Building on the decomposition and channel selection in Section 4.1, this section further conducts forecasting experiments and model selection. The entire process follows a chronological partition. All normalization parameters and channel statistics are fixed within the training phase and then consistently applied to the validation/testing phase, avoiding any form of information leakage.

4.2.1. Δ Target Definition and Experimental Protocol

The parameter ranges are as follows: hidden { 32 , 64 ,   , 256 } , lr [ 10 4 , 5 × 10 3 ] (on a logarithmic scale), L 2 [ 10 6 , 10 2 ] , dropout [ 0 , 0.5 ] , clip [ 0.5 , 5 ] , batch { 32 , 64 , 128 } . The forecasting target is expressed in Δ form, with inputs retaining the multi-channel structure of “ IMFs + LOWESS residuals”. The network adopts a single-layer BiLSTM, with training computations combined with early stopping. Key training items and computational records are listed in Table 2.
The test set evaluation includes the following metrics: RMSE, MAE, sMAPE, and R2. Among these, both RMSE and MAE are expressed in kilowatts (kW), consistent with the unit of the load. sMAPE is expressed as a percentage. R2 is used to assess the goodness of fit and facilitates comparisons across different forecast horizons.

4.2.2. Single-Step Prediction Result Analysis

Under this protocol, the single-step prediction performance is first examined. Figure 8 shows the comparison between predicted and actual values at the test stage with step length h = 1.
From Figure 8, the following can be seen:
(1) The overall trend and turning points are accurately captured, with the predicted curve almost overlapping the actual curve.
(2) After a slight lag at certain positions before the peaks, the model aligns well with the peaks and subsequent decline phases.
(3) The enlarged window indicates that the model maintains smoothness during rapid climbing stages without obvious overshooting, which meets the expected design goals.

4.2.3. Multi-Step Forecasting Error Propagation

Subsequently, the forecast horizon is extended to h = 2 and h = 3, as shown in Figure 9. It presents the model’s multi-step performance across horizons, displaying bar plots of RMSE, sMAPE, and R 2 computed on the test set.
The bar chart in Figure 9 illustrates that the error accumulates and increases step by step. RMSE and sMAPE grow progressively from h = 1 to h = 3, while R2 remains at a relatively high level. This indicates that the use of the Δ-target training strategy enables the model to capture the dynamic variations in rapid climbing and sudden changes more sharply. Even during the peak surge and high-load transition stages, Δ-BiLSTM still maintains smooth and stable predictions, validating the effectiveness of the consistency adjustment in the proposed Δ-target learning innovation.
The corresponding detailed numerical values are listed in Table 3.

4.2.4. Time Window and Step Length Sensitivity Analysis

To ensure a more stable time correlation setting, under the same leakage-free pipeline, for window lengths L { 28 , 56 , 112 } and maximum step length h max 1 , 2 , 3 , a grid search is performed. Each L is first segmented and processed, followed by decomposition, LOWESS, and Top-M channel selection in the training phase. Then, for each h h max , a Δ-model is trained separately, and the validation RMSE weighted by step length is used as the selection metric.
Figure 10 visualizes the validation weighted RMSE on the L , h max grid. For each cell in the grid, the leakage-free pipeline is executed, Δ-models are trained for all h h max , and their RMSE values are aggregated with step length weights. As shown in Figure 10, a distinct performance valley appears around L ≈ 56 with h max 2 . Extending L beyond 56 or increasing hmax to three raises the weighted RMSE. Therefore, L = 56 and h max = 2 are adopted as a balanced choice that maintains accuracy while limiting model complexity.
Figure 11 summarizes the validation weighted RMSE for three input window lengths L { 28 , 56 , 112 } . For each L , the bars represent the maximum forecast horizon h max { 1 , 2 , 3 } , corresponding to the step-weighted aggregation of RMSE for horizons up to h max . As shown in Figure 11, when L = 56 , the model consistently maintains consistently low error for all three window lengths. In contrast, for the same L , the weighted RMSE rises significantly as hmax increases, reflecting the cumulative difficulty of multi-step joint modeling. The complete grid results are presented in Table 4. It should be noted that if a certain combination is only trained up to hmax = 1 or 2, the uncovered horizons are marked as “-”, which does not affect the conclusions. Based on the consistency results of Table 4, subsequent comparisons adopt the configuration L = 56 , with hmax also determined by the same weighted criterion.
In summary, under the unified framework of “decomposition-selection-Δ representation” and leakage-free evaluation, the model demonstrates strong trend depiction and peak-tracking ability in both single-step and multi-step scenarios, with L = 56 serving as the more robust time window.

4.3. Results and Discussion

4.3.1. Model Comparison and Advantages

Before comparing the baseline models (such as SVM, RawLSTM, RNN, ARIMA, etc.) with the proposed method (Δ-BiLSTM), it is essential to clarify that all models were trained using the same configuration to ensure fairness and comparability of the results. Specifically, all models employed the same early stopping strategy, training epochs (60 epochs), learning rate (0.001), dropout (0.2), L2 regularization (1.00 × 10−5), gradient clipping (1), batch size (64), and random seed (123), ensuring identical convergence behavior and training stability. However, the proposed method (Δ-BiLSTM) applied VMD and LOWESS during the data preprocessing phase to remove noise and capture multi-scale features, whereas the baseline models were trained directly on the raw data without these additional processing steps. This allowed the proposed method to better handle noise and abrupt fluctuations, resulting in improved prediction accuracy and robustness.
With this unified configuration and training budget, Δ-BiLSTM is compared with baseline models such as RawLSTM and RNN (Recurrent Neural Network) [29]. The results are presented in Table 5.
As shown in Table 5, the proposed Δ-BiLSTM model achieves superior performance on the three core metrics of R2, RMSE, and sMAPE in most cases, with particularly significant advantages at h = 1 and h = 2. Specifically, when h = 1 and h = 2, RMSE decreases by 65.5% and 38.26%, respectively, compared to the best baseline model, ARIMA (AutoRegressive Integrated Moving Average). When the step length is extended to h = 3, although the overall error increases, the rate of error growth is lower than that of the baseline methods, while high predictive stability is maintained. This indicates that the proposed method not only demonstrates outstanding accuracy in single-step forecasting but also maintains robust advantages in multi-step forecasting.

4.3.2. Robustness Analysis

From the perspective of robustness analysis, Figure 12 presents the boxplot distribution of absolute errors across the horizon. The three boxplots show the test set absolute error (in kW) distributions of five models at forecast horizons h = 1, 2, and 3. The boxes denote the interquartile range (IQR), with medians shown in red and means in black; the whiskers extend to 1.5 × IQR, and circles mark outliers.
As shown in Figure 12, under the three scenarios with h = 1 to 3, Dispersion increases with horizon for all methods, while Δ-BiLSTM achieves the lowest median and tightest IQR at h = 1 and remains competitive as h grows (margins narrow at h = 3). Δ-BiLSTM maintains the lowest median and the narrowest interquartile range (IQR) in the distribution of absolute errors, while the 95% quantile is significantly lower than that of the benchmark models, indicating the strongest robustness.
By contrast, the RNN exhibits higher tail risks in long-horizon forecasts, while the SVM and RawLSTM show slightly better medians than ARIMA but still suffer from wider distribution spreads. Overall, Δ-BiLSTM demonstrates superior accuracy and robustness compared to traditional machine learning models and less sophisticated deep learning baseline models.

4.3.3. Error Distribution Characteristics and Prediction Comparison of Multiple Algorithms

Figure 13 and Figure 14 further reveal the spatial distribution of forecasting errors across time and prediction horizons. Here, Algo1 denotes Δ-BiLSTM, Algo2 denotes SVM, Algo3 denotes RawLSTM, Algo4 denotes RNN, and Algo5 denotes ARIMA. As shown in Figure 13, the paired heatmaps display test errors along the sequence. Rows correspond to Algo1 to Algo5, which map to Δ-BiLSTM, SVM, RawLSTM, RNN, and ARIMA, respectively. The left panel shows absolute error (kW), and the right panel shows relative error (%). Warmer colors denote larger error magnitudes. The Δ-BiLSTM row is predominantly composed of cooler shades, whereas RNN and RawLSTM show recurring warm patterns and localized hotspots. Overall, Δ-BiLSTM exhibits the lowest spatial error profile on both absolute and relative scales.
As shown in Figure 14, the left shows the residual series at h = 1 for five algorithms. The right shows the corresponding predictions at the same horizon. It is observed that Algorithm 1 (Δ-BiLSTM) produces errors that are most tightly concentrated around zero, with the narrowest amplitude and no visible drift. Its prediction trace closely matches the measured values in terms of peak height and ramp timing. In contrast, competing methods (e.g., Algorithm 4) show larger excursions or offsets. These patterns substantiate the advantage of the proposed VMD-LOWESS + Δ-BiLSTM pipeline in delivering unbiased and stable single-step forecasts.
In summary, under the same experimental protocol and unified evaluation criteria, the proposed Δ-BiLSTM model outperforms baseline models across four aspects: overall accuracy, calibration, robustness, and spatial consistency. These results align with the experimental findings presented in Section 4.1 and Section 4.2, further demonstrating that Δ-BiLSTM, when trained under unified training protocols and evaluated without leakage, exhibits stronger consistency and stability. The results not only highlight its superiority over baseline models but also show that in repeated experiments and under varying prediction horizons, it maintains robust stability.

4.3.4. Segment-Wise Improvement Under Different Operating Conditions

To further quantify the improvement amplitude of the proposed Δ-BiLSTM under different operating conditions, the 5 min load dataset is divided into three segments along the sample-index axis, as shown in Figure 15.
The segmentation is based on both the load level and the local variability. Specifically, the active power series P ( k ) is first processed with a sliding window of 12 samples (1 h) to compute the local standard deviation σ ( k ) . Continuous intervals with σ ( k ) 1.0 kW and a duration longer than 20 samples are regarded as stable segments, which correspond to three relatively flat periods in the dataset (sample indices 20–64, 138–192, and 313–348). Among the remaining samples, intervals with a load level P ( k ) 67 kW are identified as peak segments, forming two sustained peak-load periods (indices 107–137 and 193–245). The rest of the samples, which mainly cover the ramp-up and ramp-down transitions between valley and peak, are classified as the fluctuation segment.
As shown in Figure 15, the RMSE of the h = 1 forecasts are evaluated separately on the stable, fluctuation, and peak segments. Since ARIMA still achieves the best overall accuracy among the conventional benchmark models in this study, the comparison in this subsection focuses on Δ-BiLSTM versus ARIMA. For h = 1 , Δ-BiLSTM reduces the RMSE on the fluctuation segment by 69.4%, and achieves 62.5% and 62.4% RMSE reduction on the stable and peak segments, respectively, compared with ARIMA.
These segment-wise results indicate that Δ-BiLSTM not only improves the prediction accuracy in normal stable periods, but also significantly suppresses error amplification during rapid load changes and peak-load conditions. In particular, the model shows a clear advantage in tracking high-frequency variations and sustained peaks while maintaining low error levels, which is crucial for short-term dispatch, reserve allocation, and secure operation of distribution networks.

4.3.5. Corresponding Effects of Contribution Points

In this section, the performance of Δ-BiLSTM is compared with four variant models: Abl w/o Delta, Abl w/o ChSel, Abl w/o LOWESS, and Abl w/o BiLSTM. Specifically, we quantify the contribution of each component to the overall performance, focusing on the percentage increase in RMSE. The analysis is based on h = 1 for consistency across the different models.
The results are presented in Table 6.
As shown in Table 6.
(i) Abl w/o Delta removes the Δ-target component, which is designed to model the rate of change and abrupt transitions in bus load. After removing this component, the RMSE increases by 8.00%, highlighting the importance of capturing dynamic transitions and fluctuations in the load sequence.
(ii) Abl w/o ChSel removes the Channel Selection component, which helps filter out redundant input features based on their correlation. The removal of ChSel leads to an RMSE increase of 4.43%, emphasizing the value of selecting effective channels for improving model performance.
(iii) Abl w/o LOWESS removes the LOWESS component, which is used to smooth residuals and reduce noise in the data. Removing LOWESS results in a significant RMSE increase of 17.91%, demonstrating the importance of noise reduction in ensuring stability and reliability in the forecasts.
(iv) Abl w/o BiLSTM replaces BiLSTM with LSTM, a unidirectional model that only processes past data, whereas BiLSTM is bidirectional and can take both past and future information into account. This change leads to the largest RMSE increase of 29.47%, emphasizing the critical role BiLSTM plays in enhancing the model’s ability to capture long-term and short-term dependencies in the data.
Based on the above calculation results, it shows that each part of Δ-BiLSTM contributes significantly to its performance. In particular, the Δ-target and BiLSTM components are crucial for capturing the dynamic nature of the load and improving prediction accuracy. Removing these components results in significant performance degradation, especially in scenarios involving abrupt changes and complex time series data. The Channel Selection and LOWESS steps also play important roles in improving model robustness and accuracy by reducing redundancy and noise.

5. Conclusions

In this paper, we address the challenges of non-stationarity, irregular fluctuations, and multi-scale coupling in short-term bus load forecasting for distribution networks, and propose a multi-channel Δ-BiLSTM forecasting framework based on VMD and residual LOWESS. By performing VMD and residual smoothing only during the training phase, combined with forward-looking correlation screening and redundancy suppression, we construct a multi-channel input embedding that captures both deterministic and stochastic features. A Δ-BiLSTM network is then employed to realize multi-step forecasting and abrupt-change tracking. With Bayesian optimization and early stopping, together with robustness-oriented training protocols, the model ensures reproducibility and fairness across different horizons.
Quantitative results on the coastal-grid bus load dataset show that the prediction accuracy is substantially improved under a unified training budget. In single-step forecasting, Δ-BiLSTM reduces the test set RMSE by 65.5% and 38.3% at horizons h = 1 and h = 2 , respectively, compared with the best conventional benchmark ARIMA, while maintaining higher R 2 and lower sMAPE across all horizons. The boxplot analysis further shows that Δ-BiLSTM produces the tightest absolute error distributions with the lowest medians and 95% quantiles, confirming that the error band is more concentrated, smoother, and lower-variance than those of SVM, RawLSTM, RNN, and ARIMA.
To refine the analysis of improvement amplitude and link it to operating conditions, the 5 min load dataset is divided into stable, fluctuation, and peak segments according to the load level and local variability. For h = 1 , Δ-BiLSTM reduces the RMSE by 69.4% on the fluctuation segment and by 62.5% and 62.4% on the stable and peak segments, respectively, relative to ARIMA. In addition, ablation experiments show that removing the Δ-target, channel selection, residual LOWESS, and BiLSTM components leads to RMSE increases of 8.00%, 4.43%, 17.91% and 29.47%, respectively, on the test set. These results clarify the corresponding effects of the contribution points: Δ-targets and the bidirectional architecture are crucial for tracking ramping dynamics and abrupt transitions, while channel selection and LOWESS significantly enhance robustness by suppressing redundant information and high-frequency noise.
From an application perspective, the proposed VMD–LOWESS–Δ-BiLSTM framework is especially suitable for short-term bus load forecasting in distribution networks with strong high-frequency disturbances, significant peak–valley transitions, and limited historical data. In such scenarios, conventional statistical models or single-channel deep networks tend to underestimate peak levels and amplify errors during rapid ramps, whereas the proposed method maintains low error levels and stable calibration by combining multi-scale inputs, Δ-target learning, and leakage-free training protocols. This makes the framework a practical tool for short-term dispatch, reserve allocation, and security assessment in feeders with high renewable penetration and emerging loads such as centralized EV charging.
Nevertheless, this study still has several limitations. The case study is conducted on a single coastal-grid feeder with a moderate data length and a limited set of exogenous variables, and the Δ-BiLSTM learner is restricted to a relatively compact architecture for the sake of reproducibility. Future work will extend the framework to larger-scale datasets, multi-node shared forecasting models, and explore alternative sequence learners such as Transformer-based architectures and hybrid expert (mixture-of-experts) models. In addition, integrating richer exogenous information (e.g., high-resolution meteorological data and market signals) and conducting formal forecast-comparison tests under different operating regimes will further improve the generality and interpretability of the proposed approach.

Author Contributions

Conceptualization, methodology, software, writing—original draft, validation, formal analysis, investigation, Y.G.; resources, J.Z.; data curation, L.W. and J.Z.; writing—review and editing, L.W. and J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This paper is supported by the National College Students Innovation and Entrepreneurship Training Program (Grant No. S202410536024) and the National Key Research and Development Program of China (Grant No. 2024YFE0115600).

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to research security.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Eren, Y.; Küçükdemiral, I. A comprehensive review on deep learning approaches for short-term load forecasting. Renew. Sustain. Energy Rev. 2024, 189, 114031. [Google Scholar] [CrossRef]
  2. Morais, L.B.S.; Aquila, G.; de Faria, V.A.D.; Lima, L.M.M.; Lima, J.W.M.; de Queiroz, A.R. Short-term load forecasting using neural networks and global climate models. Appl. Energy 2023, 348, 121439. [Google Scholar] [CrossRef]
  3. Wang, C.; Zhao, H.; Liu, Y.; Fan, G. Minute-level ultra-short-term power load forecasting based on time-series data features. Appl. Energy 2024, 372, 123801. [Google Scholar] [CrossRef]
  4. Chan, J.W.; Yeo, C.K. A Transformer-based approach to electricity load forecasting. Electr. J. 2024, 37, 107370. [Google Scholar] [CrossRef]
  5. Hasan, M.; Mifta, Z.; Papiya, S.J.; Roy, P.; Dey, P.; Salsabil, N.A.; Chowdhury, N.-U.; Farrok, O. A State-of-the-Art Comparative Review of Load Forecasting Methods: Characteristics, Perspectives, and Applications. Energy Convers. Manag. X 2025, 26, 100922. [Google Scholar] [CrossRef]
  6. Iqbal, M.S.; Adnan, M.; Mohamed, S.E.G.; Tariq, M. A hybrid deep learning framework for short-term load forecasting. Results Eng. 2024, 22, 103560. [Google Scholar] [CrossRef]
  7. Fan, G.-F.; Han, Y.-Y.; Li, J.-W.; Peng, L.-L.; Yeh, Y.-H.; Hong, W.-C. A hybrid model for deep learning short-term power load forecasting. Expert Syst. Appl. 2024, 234, 122012. [Google Scholar] [CrossRef]
  8. Zhong, B.; Yang, L.; Li, B.; Ji, M. Short-term power grid load forecasting based on VMD-SE-BiLSTM-Attention hybrid model. Int. J. Low-Carbon Technol. 2024, 19, 1951–1958. [Google Scholar] [CrossRef]
  9. Wang, X.; Jiang, H.; Wu, Z.; Yang, Q. Adaptive variational autoencoding generative adversarial networks for rolling bearing fault diagnosis. Adv. Eng. Inform. 2023, 56, 102027. [Google Scholar] [CrossRef]
  10. Wang, X.; Jiang, H.; Zeng, T.; Dong, Y. An adaptive fused domain cycling variational generative adversarial network for machine fault diagnosis under data scarcity. Inf. Fusion 2025, 56, 103616. [Google Scholar] [CrossRef]
  11. Yan, J.; Cheng, Y.; Zhang, F.; Zhou, N.; Wang, H.; Jin, B.; Wang, M.; Zhang, W. Multimodal imitation learning for arc detection in complex railway environments. IEEE Trans. Instrum. Meas. 2025, 74, 1–13. [Google Scholar] [CrossRef]
  12. Dong, Z.; Tian, Z.; Lv, S. Decomposition and weighted correction for short-term load forecasting. Appl. Soft Comput. 2025, 162, 111863. [Google Scholar] [CrossRef]
  13. Han, M.; Lee, J.-S. IVMD-BiLSTM with differential mechanism for load forecasting. Appl. Soft Comput. 2024, 152, 110657. [Google Scholar] [CrossRef]
  14. Zare, M.; Nouri, N.M. End-effects mitigation in empirical mode decomposition. Mech. Syst. Signal Process. 2023, 200, 110205. [Google Scholar] [CrossRef]
  15. Yu, M.; Yuan, H.; Li, K.; Deng, L. Noise cancellation via TVF-EMD with adaptive thresholding. Algorithms 2023, 16, 296. [Google Scholar] [CrossRef]
  16. Shalby, E.M.; Abdelaziz, A.Y.; Ahmed, E.S.; Rashad, B.A.-E. A guide to selecting wavelet decomposition level and functions. Sci. Rep. 2025, 15, 82025. [Google Scholar] [CrossRef]
  17. Coroneo, L.; Iacone, F. Testing for equal predictive accuracy with strong dependence. Int. J. Forecast. 2025, 41, 1073–1092. [Google Scholar] [CrossRef]
  18. Guenoukpati, A.; Agbessi, A.P.; Salami, A.A.; Bakpo, Y.A. Hybrid Long Short-Term Memory Wavelet Transform Models for Short-Term Electricity Load Forecasting. Energies 2024, 17, 4914. [Google Scholar] [CrossRef]
  19. Wang, Y.; Li, H.; Jahanger, A.; Li, Q.; Wang, B.; Balsalobre-Lorente, D. A novel ensemble electricity load forecasting system based on a decomposition-selection-optimization strategy. Energy 2024, 312, 133524. [Google Scholar] [CrossRef]
  20. Li, H.; Li, S.; Wu, Y.; Xiao, Y.; Pan, Z.; Liu, M. Short-term power load forecasting for integrated energy system based on a residual and attentive LSTM-TCN hybrid network. Front. Energy Res. 2024, 12, 1384142. [Google Scholar] [CrossRef]
  21. Harvey, D.I.; Leybourne, S.J.; Zu, Y. Tests for equal forecast accuracy under heteroskedasticity. J. Appl. Econ. 2024, 39, 850–869. [Google Scholar] [CrossRef]
  22. Harvey, D.I.; Leybourne, S.J.; Zu, Y. Testing for equal average forecast accuracy in possibly unstable environments. J. Bus. Econ. Stat. 2024, 43, 643–656. [Google Scholar] [CrossRef]
  23. Liu, F.; Chen, L.; Zheng, Y.; Feng, Y. A Prediction Method with Data Leakage Suppression for Time Series. Electronics 2022, 11, 3701. [Google Scholar] [CrossRef]
  24. Wu, Y.; Meng, X.; Zhang, J.; He, Y.; Romo, J.A.; Dong, Y.; Lu, D. Effective LSTMs with seasonal-trend decomposition (STL, LOESS) and adaptive learning. Expert Syst. Appl. 2024, 236, 121202. [Google Scholar] [CrossRef]
  25. Zhuang, Q.; Gao, L.; Zhang, F.; Ren, X.; Qin, L.; Wang, Y. MIVNDN: Ultra-Short-Term Wind Power Prediction Method with MSDBO-ICEEMDAN-VMD-Nons-DCTransformer Net. Electronics 2024, 13, 4829. [Google Scholar] [CrossRef]
  26. Li, K.; Huang, W.; Hu, G.; Li, J. Ultra-short-term power load forecasting based on CEEMDAN-SE + LSTM. Energy Build. 2023, 279, 112666. [Google Scholar] [CrossRef]
  27. Sun, H.; Yu, Z.; Zhang, B. Research on short-term power load forecasting based on VMD + GRU/CNN-Attention. PLoS ONE 2024, 19, e0306566. [Google Scholar] [CrossRef]
  28. Ouyang, T.; He, Y.; Li, H.; Sun, Z.; Baek, S. Modeling and forecasting short-term power load with copula model and deep belief network. IEEE Trans. Emerg. Top. Comput. Intell. 2019, 3, 127–136. [Google Scholar] [CrossRef]
  29. Zelios, V.; Mastorocostas, P.; Kandilogiannakis, G.; Kesidis, A.; Tselenti, P.; Voulodimos, A. Short-Term Electric Load Forecasting Using Deep Learning: A Case Study in Greece with RNN, LSTM, and GRU Networks. Electronics 2025, 14, 2820. [Google Scholar] [CrossRef]
Figure 1. An overview of the VMD-LOWESS-Δ-BiLSTM forecasting workflow.
Figure 1. An overview of the VMD-LOWESS-Δ-BiLSTM forecasting workflow.
Electronics 14 04772 g001
Figure 2. VMD on of the training segment (K = 6)—Top: raw series; Middle: IMF1-IMF6 with central frequency ω k ; Bottom: Residual.
Figure 2. VMD on of the training segment (K = 6)—Top: raw series; Middle: IMF1-IMF6 with central frequency ω k ; Bottom: Residual.
Electronics 14 04772 g002
Figure 3. VMD reconstruction accuracy for the bus load series: (a) training, (b) validation, and (c) test segments.
Figure 3. VMD reconstruction accuracy for the bus load series: (a) training, (b) validation, and (c) test segments.
Electronics 14 04772 g003aElectronics 14 04772 g003b
Figure 4. Trade-off between reconstruction RMSE and roughness for different LOWESS spans on the training set. The blue circles denote RMSE to the raw training series (left axis), and the orange squares denote roughness (right axis).
Figure 4. Trade-off between reconstruction RMSE and roughness for different LOWESS spans on the training set. The blue circles denote RMSE to the raw training series (left axis), and the orange squares denote roughness (right axis).
Electronics 14 04772 g004
Figure 5. Comparison between the VMD residual and its LOWESS-smoothed counterpart for (a) training, (b) validation, and (c) test segments.
Figure 5. Comparison between the VMD residual and its LOWESS-smoothed counterpart for (a) training, (b) validation, and (c) test segments.
Electronics 14 04772 g005
Figure 6. Denoising comparison on the training segment (Raw/VMD-plain/WMA/VMD–LOWESS) with zoom-in MAE.
Figure 6. Denoising comparison on the training segment (Raw/VMD-plain/WMA/VMD–LOWESS) with zoom-in MAE.
Electronics 14 04772 g006
Figure 7. Grid search of the composite objective J over the number of VMD modes K and the LOWESS span on the training set.
Figure 7. Grid search of the composite objective J over the number of VMD modes K and the LOWESS span on the training set.
Electronics 14 04772 g007
Figure 8. Comparison between measured and predicted bus load on the test set for horizon h = 1 , with a zoomed-in view of a peak-load period.
Figure 8. Comparison between measured and predicted bus load on the test set for horizon h = 1 , with a zoomed-in view of a peak-load period.
Electronics 14 04772 g008
Figure 9. Horizon-wise test performance of Δ-BiLSTM on the bus load dataset (RMSE, sMAPE, and R2).
Figure 9. Horizon-wise test performance of Δ-BiLSTM on the bus load dataset (RMSE, sMAPE, and R2).
Electronics 14 04772 g009
Figure 10. Validation weighted RMSE on the ( L , h m a x ) grid under the leakage-free pipeline.
Figure 10. Validation weighted RMSE on the ( L , h m a x ) grid under the leakage-free pipeline.
Electronics 14 04772 g010
Figure 11. Weighted RMSE distribution on (L, hmax) plane.
Figure 11. Weighted RMSE distribution on (L, hmax) plane.
Electronics 14 04772 g011
Figure 12. Boxplots of test set absolute errors | e | (in kW) for five models (Δ-BiLSTM, SVM, RawLSTM, RNN, and ARIMA) at horizons h = 1, 2 and 3. The red horizontal lines denote the medians, the black dots denote the sample means, and the circles mark outliers.
Figure 12. Boxplots of test set absolute errors | e | (in kW) for five models (Δ-BiLSTM, SVM, RawLSTM, RNN, and ARIMA) at horizons h = 1, 2 and 3. The red horizontal lines denote the medians, the black dots denote the sample means, and the circles mark outliers.
Electronics 14 04772 g012
Figure 13. Heatmaps of absolute error (left) and percentage relative error (right) for five algorithms across sample points.
Figure 13. Heatmaps of absolute error (left) and percentage relative error (right) for five algorithms across sample points.
Electronics 14 04772 g013
Figure 14. Comparison of residuals (left) and predicted versus measured bus load (right) at h = 1 for five algorithms (Δ-BiLSTM, SVM, RawLSTM, RNN, and ARIMA).
Figure 14. Comparison of residuals (left) and predicted versus measured bus load (right) at h = 1 for five algorithms (Δ-BiLSTM, SVM, RawLSTM, RNN, and ARIMA).
Electronics 14 04772 g014
Figure 15. Segmented load profile with stable, fluctuation, and peak segments.
Figure 15. Segmented load profile with stable, fluctuation, and peak segments.
Electronics 14 04772 g015
Table 1. Correlation, energy ratio, and selection results.
Table 1. Correlation, energy ratio, and selection results.
ComponentCorrelationEnergy RatioSelected
IMF10.34265090.781733791
IMF20.52361280.008356741
IMF30.32032910.043661881
IMF40.01847560.000128830
IMF50.20626513.01516681
IMF60.02164551.50612631
Residual0.1982360.1660741
Table 2. Training configuration for Δ-BiLSTM by horizon (h = 1 to 3).
Table 2. Training configuration for Δ-BiLSTM by horizon (h = 1 to 3).
hHiddenlrDropoutL2ClipEpochsPatienceBatchSeed
1640.0010.21.00 × 10−5160864123
2640.0010.21.00 × 10−5160864123
3640.0010.21.00 × 10−5160864123
Table 3. Test set performance across the horizon (h = 1, 2, and 3).
Table 3. Test set performance across the horizon (h = 1, 2, and 3).
hRMSEMAER2sMAPE
10.5092820.4088920.9981070.922703
21.6617381.4318420.9798482.940043
31.5257871.2953970.983012.896958
Table 4. Weighted RMSE across ( L , h max ) combinations.
Table 4. Weighted RMSE across ( L , h max ) combinations.
LhmaxRMSEweightRMSE_h1RMSE_h2RMSE_h3RMSEstd
2810.8537880.853788--0.0
2821.5374690.8418341.885286-0.737832
2832.0856470.8257131.7051262.7593060.968112
5610.3734260.373426--0.0
5620.5766760.3965640.666732-0.191037
5630.83690.4081130.6439331.1084690.35634
11210.3943420.394342--0.0
11220.8123560.4587810.909494-0.375977
11231.3543930.5643880.9972451.8558270.657311
Table 5. Model performance comparison with baseline models.
Table 5. Model performance comparison with baseline models.
hModelRMSEMAEsMAPER2
1Δ-BiLSTM0.73890.60551.16480.9972
1SVM2.36862.17354.50110.9716
1RawLSTM2.38472.03144.09680.9712
1RNN2.97612.19924.29400.9552
1ARIMA2.15291.94744.16970.9766
2Δ-BiLSTM1.30951.08222.08450.9913
2SVM2.32842.14134.37420.9725
2RawLSTM2.15142.05584.20510.9765
2RNN3.03202.29674.51440.9534
2ARIMA2.12111.96344.06360.9772
3Δ-BiLSTM1.87371.54952.98980.9822
3SVM1.82271.60843.37300.9831
3RawLSTM2.02521.74853.32130.9792
3RNN3.25212.44474.85040.9463
3ARIMA2.18301.89684.20160.9758
Table 6. Model performance comparison with four variant models.
Table 6. Model performance comparison with four variant models.
hModelRMSEMAEsMAPER2
1Δ-BiLSTM0.73890.60551.16480.9972
1Abl_w_o_Delta0.79770.69811.39820.9816
1Abl_w_o_ChSel0.77160.67171.37170.9828
1Abl_w_o_LOWESS0.87140.77161.47160.9713
1Abl_w_o_BiLSTM0.95540.75541.55540.9668
2Δ-BiLSTM1.30951.08222.08450.9913
2Abl_w_o_Delta1.49151.09172.19190.9727
2Abl_w_o_ChSel1.47281.08722.16290.9866
2Abl_w_o_LOWESS1.57661.12672.27670.9734
2Abl_w_o_BiLSTM1.65341.15362.35370.9675
3Δ-BiLSTM1.87371.54952.98980.9822
3Abl_w_o_Delta1.94271.61013.21020.9733
3Abl_w_o_ChSel1.93351.60473.20750.9793
3Abl_w_o_LOWESS1.97961.82363.67270.9705
3Abl_w_o_BiLSTM1.98661.84673.69470.9659
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Guo, Y.; Wang, L.; Zhao, J. A Multi-Channel Δ-BiLSTM Framework for Short-Term Bus Load Forecasting Based on VMD and LOWESS. Electronics 2025, 14, 4772. https://doi.org/10.3390/electronics14234772

AMA Style

Guo Y, Wang L, Zhao J. A Multi-Channel Δ-BiLSTM Framework for Short-Term Bus Load Forecasting Based on VMD and LOWESS. Electronics. 2025; 14(23):4772. https://doi.org/10.3390/electronics14234772

Chicago/Turabian Style

Guo, Yeran, Li Wang, and Jie Zhao. 2025. "A Multi-Channel Δ-BiLSTM Framework for Short-Term Bus Load Forecasting Based on VMD and LOWESS" Electronics 14, no. 23: 4772. https://doi.org/10.3390/electronics14234772

APA Style

Guo, Y., Wang, L., & Zhao, J. (2025). A Multi-Channel Δ-BiLSTM Framework for Short-Term Bus Load Forecasting Based on VMD and LOWESS. Electronics, 14(23), 4772. https://doi.org/10.3390/electronics14234772

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop