Next Article in Journal
Region-Based Approach for Machining Time Improvement in Robot Surface Finishing
Next Article in Special Issue
Milling Machine Fault Diagnosis Using Acoustic Emission and Hybrid Deep Learning with Feature Optimization
Previous Article in Journal
VEIDEA: A Comprehensive Framework for Implementing Building Information Modeling-Based Value Engineering Within a Common Data Environment in Construction Projects
Previous Article in Special Issue
Condition Monitoring of Electrical Transformers Using the Internet of Things: A Systematic Literature Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Time Series Forecasting of Thermal Systems Dispatch in Legal Amazon Using Machine Learning

by
William Gouvêa Buratto
1,*,
Rafael Ninno Muniz
2,3,
Rodolfo Cardoso
3,
Ademir Nied
1,
Carlos Tavares da Costa, Jr.
2 and
Gabriel Villarrubia Gonzalez
4
1
Electrical Engineering Graduate Program, Department of Electrical Engineering, Santa Catarina State University (UDESC), Joinville 89219-710, Brazil
2
Electrical Engineering Graduate Program, Department of Electrical Engineering, Federal University of Pará (UFPA), Belém 66075-110, Brazil
3
Production Engineering Master Program, Institute of Science and Technology, Federal Fluminense University, Rio das Ostras 28895-532, Brazil
4
Expert Systems and Applications Lab, Faculty of Science, University of Salamanca, 37008 Salamanca, Spain
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(21), 9806; https://doi.org/10.3390/app14219806
Submission received: 30 August 2024 / Revised: 20 October 2024 / Accepted: 25 October 2024 / Published: 27 October 2024

Abstract

:
This paper analyzes time series forecasting methods applied to thermal systems in Brazil, specifically focusing on diesel consumption as a key determinant. Recognizing the critical role of thermal systems in ensuring energy stability, especially during low rain seasons, this study employs bagged, boosted, and stacked ensemble learning methods for time series forecasting focusing on exploring consumption patterns and trends. By leveraging historical data, the research aims to predict future diesel consumption within Brazil’s thermal energy sector. Based on the bagged ensemble learning approach a mean absolute percentage error of 0.089% and a coefficient of determination of 0.9752 were achieved (average considering 50 experiments), showing it to be a promising model for the short-time forecasting of thermal dispatch for the electric power generation system. The bagged model results were better than for boosted and stacked ensemble learning methods, long short-term memory networks, and adaptive neuro-fuzzy inference systems. Since the thermal dispatch in Brazil is closely related to energy prices, the predictions presented here are an interesting way of planning and decision-making for energy power systems.

1. Introduction

Electricity generation and its costs in an isolated system off-grid are mostly related to fuel engines with diesel or natural gas consumption [1], mainly because these energy types are easily transported and they are non-intermittent energy sources in contrast to solar or wind power plants which can be integrated [2]. It is challenging to achieve accurate generation prediction of these cited renewable sources [3], while fossil fuel has significant disadvantages as it generates pollutant emissions and there is a need to refine the oil that is imported in part from foreign countries in the thermal plants of the Brazilian Amazon Region [4].
The interdependence between the nexus of greenhouse emissions, economic growth, financial development, and load demand-supply in the past with the future is important in managing policy recommendations and/or establishing strategies for the corporate sector to adopt business operations with better support technologies to improve profit with ecological efficiency [5]. Determining scenarios that could improve the reduction in carbon dioxide or coordinate carbon capture systems to reach targets considering uncertainties and promoting the best possible adaptability involves following the convergence results of these four parameters [6].
The extraction, conversion, and distribution of crude oil as diesel form part of a logistic value energy chain that is complex, and the implementation necessary for electricity generation in the power grid for a domestic load to supply isolated regions from this resource brings higher costs for the electrical system in a short time period [7]. In trying to avoid these disadvantages, the integration of renewable energy sources by applying an optimal dispatch algorithm can be evaluated and assist responsible companies in controlling greenhouse emissions from fossil fuels as well as reducing import dependence. To avoid these disadvantages, modeling the load capacity forecast is essential to plan the energy requirements of the analyzed region and the grid modifications needed [8].
Economic dispatch is the process of determining the optimal combination of power output from different generation units to meet the electrical load demand at the lowest cost while satisfying various system and operational constraints [9]. The key difference between the low-carbon form of economic dispatch and the common (or traditional) form lies in prioritizing environmental concerns, particularly carbon emissions, in the decision-making process [10]. While the traditional economic dispatch aims at cost minimization based on fuel and operational costs, the low-carbon form integrates environmental considerations, specifically targeting a reduction in carbon emissions, thus aligning the dispatch process with broader environmental and sustainability goals [11].
Low-carbon economic dispatch depends on energy conversion appliances which affect the exchange power and thermal load, storage, and carbon trading and require to be integrated with the demand response and scenarios of the predicted future load [12]. These parameters can help to determine losses and increase investment mechanisms to reduce costs and increase the profit-aggregating influence factors for carbon intensity, benefiting from the scale of technologies that could contribute to reaching lower emission targets in future scenarios [13].
A major concern about sustainability in the Amazon region is related to the use of diesel oil as the main source of energy. Figure 1 shows the used diesel oil for electricity generation in the legal Amazon by region states in 2021. According to this information, the further away from the industrialized center, the more fossil fuel consumption is observed. This is concerning since these regions are where there are the largest areas of preserved vegetation and the presence of large thermoelectric plants increases the level of pollution in the area, thus affecting the local ecosystem [14].
Traditional diesel engine power plants, compared to those based on new concepts, such as gasification with the Rankine cycle from biomass forest, are still cheaper to implement [15]. The new Rankine-gasification power plants have 1.2 times higher cost for implementation than diesel thermoelectric plants currently used despite heat wasted through exhaust gases in the general diesel operation model [16].
However, the annual operation and maintenance costs using diesel have doubled in price compared with biomass use based on research performed in an isolated village aiming for decarbonization. When natural gas pipelines are implemented, this could reduce costs by forty percent and carbon dioxide emissions by twenty-five percent in Amazonia isolated systems [17].
Fossil fuel load forecasting is important for the adaptability and stability of regions that have high dependence as evaluated by Li et al. [18], who applied diversified ensemble learning methods to reduce delayed dispatch and provide strategically scheduled storage for natural gas.
These methods can enhance hybrid generation using fossil fuels and renewable resources by integrating photovoltaic and diesel engines to reduce expenditure. The algorithms of Kanaan et al. [19] utilize improved strategies for storage, decreasing the daily carbon emission costs of a power plant by 29%, and have the potential for application in isolated regions of the legal Amazon by avoiding battery degradation.
Evaluating the consumption management and generation and adoption of technologies that relocate electrical load is a solution to increase energy sustainability [20], Within this context, forecasting time series using ensemble learning approaches is an alternative to improve the reliability of the electrical system to achieve a better environmental-economic coefficient [21]. Ensemble learning is the fusion of several predictive models to enhance collective performance [22]. It starts by assembling a varied array of foundational forecasting models, like decision trees, support vector machines, support vector regressors, and others. Each of these base models offers distinct perspectives on the data, which the ensemble method consolidates [23]. Specifically, for generation forecasting, ensemble learning leverages diverse algorithmic strengths to attain heightened accuracy and resilience in predictions [24].
There is a trend towards using deep learning models, given their ability to comprehend non-linear patterns [25]. However, these models can require more computational effort and are not always the best alternative. Considering the promising results of using ensemble models, this paper aims to present a comparison between ensemble learning methods and models based on deep learning, such as the long short-term memory (LSTM) model.
Considering that the use of thermal energy results in higher energy costs for consumers [26], especially when there is a shortage of raw materials, forecasting its use can help with energy planning. Based on this premise, the main theme of this paper is the forecasting of thermal systems dispatch. The analysis focuses on the legal Amazon since new energy sources are being explored and implementation of several insulated systems in this region and their more efficient use can improve energy supply capacity.
Given the major importance of thermal systems generation in northern Brazil, this paper proposes a thorough evaluation of ensemble learning methods for dispatch forecast for thermal systems in this region. The main contributions of this paper are:
  • The time series forecasting of thermal dispatch is an indicator that can be used for evaluating energy prices, which is an important concern in insulated thermal systems since the prices are related.
  • The bagging ensemble learning method proved to be an appropriate structure, having a lower computational demand compared to the LSTM network based on deep learning. The bagging approach proved to be a better approach than the boosting and stacking ensemble approaches.
  • The ensemble learning approaches were shown to be stable having few differences when several experiments were performed considering random initialization.
The remainder of this paper is organized as follows: In Section 2, related works regarding time series forecasting applications are discussed. Section 3 explains the proposed method. In Section 4, the results of the application of the proposed method are analyzed, and finally, in Section 5, a conclusion is drawn, and directions for future work are suggested.

2. Related Work

The use of artificial intelligence to analyze electrical power systems is increasing due to the extent to which these models have to deal with non-linearities [27]. Nowadays, deep learning has been widely applied in several fields [28,29,30,31]. Regarding power systems [32], its application can focus on the identification of faults in the power grid based on classification [33] or the prediction of its condition based on time series analysis [34]. Given the advantages of having earlier knowledge about a system’s condition, time series forecasting of energy is a way to improve decision-making in electrical power management [35].
In the work by Pazikadin et al. [36], a review is presented about solar irradiance measurement instrumentation and power generation forecasting over the past five years, specifically focusing on the integration of artificial neural networks (ANNs). They discuss the pivotal role of ANNs in enhancing the accuracy of solar power forecasts, addressing their capabilities, limitations, and areas for further improvement. The synthesis of five years of research provides insights into the progress, emerging trends, and future directions in harnessing solar energy through instrumentation and predictive modeling, contributing to a broader understanding and development of renewable energy systems.
Ahmed and Khalid’s research [37] focuses on forecasting models applied to various renewable energy sources, including solar, wind, hydroelectric, and other emerging technologies. It involves a targeted analysis of ANN-based models and their practical implications in optimizing the integration and management of renewable energy systems. Their work proposes new advances in power system forecasting for greater efficiency and stability in energy systems.
Focused on improving the supply of electricity, Corso et al. [38] presented a study on the application of convolutional neural network (CNN)-based models for classifying faults in electrical insulators, an approach that was also applied in [39] through a hybrid approach that combines several models. In [40], CNNs are combined with LSTM for fault detection in electrical machines.
The CNN-LSTM can also be used for time series forecasting [41]. In [42], an interpretable method is proposed to improve fault classification. In [43], an object detection method based on deep learning is combined with the Quasi-ProtoPNet to have an improved architecture to classify faults in the power grid, enhancing the reliability of the electrical power system. Deep learning strategies have been successfully applied to time series forecasting [44].
In [45], ANNs are applied for forecasting electricity generation, and in the work of Akhter et al. [46], the landscape of forecasting photovoltaic power generation through machine learning and metaheuristic techniques is analyzed. They highlight the effectiveness, adaptability, and areas for improvement in machine learning and metaheuristic techniques, thus providing practical applications in optimizing the efficiency of photovoltaic energy systems. Nazir et al. [47] presented a review of five-year wind generation forecasting methodologies, focusing on ANN applications. In their paper, they explore the evolution and limitations of ANNs for energy forecasting, providing ideas on possible implementations of these models, and highlighting their potential for renewable energy forecasting.
Regarding hybrid methods, many authors are applying filters combined with forecasting models to enhance their capacities and generalization. For instance, in [48], the Hodrick–Prescott filter, and in [49,50,51], the wavelet transform, are applied for denoising. All these studies proved that the use of filters in pre-processing the time series is a promising strategy when trend forecasting is the major goal of the analysis.
Yamasaki et al. [52] explore the application of optimized hybrid ensemble learning approaches in the context of very short-term load forecasting. According to them, by leveraging the strengths of hybridization, the optimized ensemble learning models demonstrate superior performance in accurately predicting short-term electricity demand. Also using a hybrid model, Stefenon et al. [53] used a group method of data handling combined with the Christiano–Fitzgerald random walk filter. The results highlighted that a denoising strategy is a promising way to enhance the architecture of these models.
The methods used here could be applied to other fields, such as in [54], where the kernel representations are evaluated, making it possible to capture changing regimes. In [55], a regime-switching method with non-linear representation is proposed; the authors proved that the model can produce promising performance results in stock market forecasting.
According to Klaar et al. [56], the use of a wavelet to decompose the signal combined with an attention mechanism is an outstanding strategy to enhance the prediction of the LSTM network. This combination is also applied in [57], where the predictions concern the level of dams for hydroelectric power plants. In [58], the wavelet with the LSTM is evaluated without the attention mechanism; using this strategy, the results were better than for other models, showing that the combination of these two techniques is a good strategy.
The research by Stefenon et al. [59] proposes a hybrid forecasting approach for modeling and predicting the complex dynamics of Italian electricity spot prices. They combine the strengths of the Prophet forecasting model with seasonal trend decomposition using LOESS (locally estimated scatterplot smoothing), known for its ability to handle seasonality, holidays, and long-term trends. The integration of these two methodologies enhances the accuracy and robustness of electricity spot price forecasts.
The work of Klaar et al. [60] focuses on enhancing the accuracy of energy price forecasting in Latin America, with a specific case study conducted in Mexico. They propose a novel approach by integrating ensemble learning methods and seasonal decomposition techniques to optimize the structure of forecasting models. The motivation behind this research is the dynamic and often unpredictable nature of energy markets, particularly in the context of Latin American countries where energy pricing is influenced by various economic, environmental, and regulatory factors.
In the work of Júnior et al. [61], the authors explore the landscape of forest bioelectricity generation in Brazil, investigating its distribution patterns and temporal-spatial dependencies. The study combines geographical information systems and time-series data, to assess the distribution of forest bioelectricity plants across different regions of the country. Their research offers a foundation for policymakers, energy planners, and stakeholders to make informed decisions that promote the efficient and sustainable integration of this renewable energy source into Brazil’s broader energy landscape.
In [62], a hybrid approach for energy consumption forecasting within the context of smart grids is presented. The proposed hybrid model combines the strengths of machine learning and optimization techniques to enhance forecasting accuracy and efficiency. When evaluating the available forecasting models, it is a hard task to define the most appropriate structure to perform the predictions. According to Stefenon et al. [63], the ensemble learning methods may have better performance than deep learning since they are based on weak learners that usually require less computational effort to be computed.

Dataset

The data considered relate to the thermal generation due to dispatch in the northern region of Brazil and are provided by the national system operator (in Portuguese Operador Nacional do Sistema—ONS). The original dataset is available at: https://dados.ons.org.br/dataset/geracao-termica-despacho-2 (accessed on 25 November 2023). The total generation relates to 16 thermal power plants (Aparecida Parte I, Cristiano Rocha, Geramar I, Geramar II, Jaraqui, Manauara, Maranhão III, Maranhão IV, Maranhão V, Mauá 3, MC2 Nova Venécia 2, Parnaíba IV, Parnaíba V, Ponta Negra, Porto do Itaqui, and Tambaqui).
Scheduled and verified generation data for thermal power plants dispatched by the ONS on an hourly basis are provided. This time series is presented in Figure 2. Given an hourly basis, 2160 records are considered from 18 August (0 h) to 15 November (23 h) of 2023, resulting in a period of evaluation of 90 days. The time series forecasting presented here is one step ahead, meaning that the predictions would be presented one hour earlier.
The models presented in this paper can be applied to other datasets since the analysis is carried out in relation to the variation in the time series (see Figure 2). Especially concerning the analysis of variation in diesel consumption, it is necessary to understand whether non-linearities (noise in the time series) affect the applicability of the model. In these cases, it is necessary to use input filters to attenuate the noise.

3. Proposed Method

A common ensemble method for time series forecasting involves combining predictions from multiple base models to improve forecasting accuracy [64]. Let { y t } represent a univariate time series where y t denotes the value at time t. Suppose we have M base forecasting models denoted as f 1 , f 2 , , f M , which can be any suitable time series forecasting models [65].
For a given time series dataset up to time T, the ensemble forecast y ^ T at time T is computed by aggregating the predictions of the base models:
y ^ T = 1 M i = 1 M f i ( T )
where f i ( T ) represents the forecast generated by the i-th base model at time T. Various aggregation methods can be used for combining predictions, such as simple averaging, or more sophisticated techniques like stacking or boosting applied to time series forecasting [66]. Ensemble methods aim to leverage the diversity of individual models to create a more accurate and robust forecast by reducing potential biases or errors inherent in any single forecasting model [67]. In this paper, the stacking, bagging, and boosting ensemble learning models are considered.

3.1. Stacking Ensemble Learning Model

Let X be the training dataset with input features represented as X = { x 1 , x 2 , , x n } and corresponding labels Y = { y 1 , y 2 , , y n } . Let M be the number of base models, and each base model h i is represented by:
h i : X Y i , i = 1 , 2 , , M
where Y i denotes the predictions made by the i-th base model [68].
The predictions of the base models are combined to form a new dataset X , where each instance x j is given by:
x j = { h 1 ( x j ) , h 2 ( x j ) , , h M ( x j ) } , j = 1 , 2 , , n .
A meta-learner g is then trained on the new dataset X using the true labels Y:
g : X Y .
The final prediction y ^ for a new input x new is obtained by applying the meta-learner g on the predictions of the base models for x new :
y ^ new = g { h 1 ( x new ) , h 2 ( x new ) , , h M ( x new ) } .
The architecture of the stacking ensemble learning approach is presented in Figure 3.

3.2. Boosting Ensemble Learning Model

The boosting model is an ensemble method where multiple weak learners are combined to create a strong learner [69]. Let T denote the number of iterations or base models. Initialize the weights of the training instances uniformly: D 1 ( i ) = 1 n for i = 1 , 2 , , n .
For each iteration t = 1 , 2 , , T : Train a weak learner h t on the weighted training data X with weights D t ( i ) . Calculate the weighted error ϵ t of h t :
ϵ t = i = 1 n D t ( i ) · ( h t ( x i ) y i )
where is the indicator function. Compute the importance of h t :
α t = 1 2 ln 1 ϵ t ϵ t .
Update the weights D t + 1 ( i ) for the next iteration:
D t + 1 ( i ) = D t ( i ) · exp α t · y i · h t ( x i ) Z t
where Z t is a normalization factor to ensure that the weights sum up to 1 [70]. The architecture of the boosting ensemble model is shown in Figure 4.

3.3. Bagging Ensemble Learning Model

Bagging (Bootstrap Aggregating) is an ensemble method where B base models are trained on different subsets of the training data. Let B denote the number of base models [71]. Each base model h i is trained on a bootstrap sample X i (a randomly sampled subset with replacement from X) and is represented by:
h i : X i Y i , i = 1 , 2 , , B
where Y i denotes the predictions made by the i-th base model on its respective bootstrap sample X i [70].
The final prediction y ^ new for a new input x new is obtained by aggregating the predictions of all base models:
y ^ new = Aggregation ( h 1 ( x new ) , h 2 ( x new ) , , h B ( x new ) ) .
Common aggregation methods include averaging the predictions for regression problems or taking a majority vote (or averaging probabilities) for classification problems. In this paper, aggregation through averaging was applied to the forecasting of time series. The architecture of this model is shown in Figure 5.

3.4. Parameters

For the hyperparameters evaluation, the linear (LIN) radial basis function (RBF) [72], and polynomial (POLY) kernel functions are considered. The sequential minimal optimization (SMO), iterative single data algorithm (ISDA) [73], and L1 soft-margin minimization by quadratic programming (L1QP) solver are evaluated [74]. Besides the error measures, the total time (training and testing) for each experiment is presented.
The L1QP involves solving the optimization task in its entirety using traditional optimization methods, such as the interior-point method [75]. The Wolfe dual problem that is evaluated in the optimization process is calculated by the following equation:
max α i = 1 m α i 1 2 i = 1 m j = 1 m α i α j y i y j x i , x j
s . t . : 0 α i C , i = 1 , , m
i = 1 m α i y i = 0
where α i is the dual variable that lies within the box.
SMO improves the computational efficiency by solving a simpler sequence of optimization problems. The SMO algorithm works on the dual form of the problem, solving the dual problem for two points at a time, thereby updating two dual variables at each step [48]. Instead of solving the complete optimization problem, ISDA processes one data point at a time. This makes its functionality similar to SMO, but with the key difference of updating only one multiplier at a time. To determine which multiplier to update, ISDA selects the point that most significantly violates the Karush–Kuhn–Tucker (KKT) conditions [76].
The kernel functions (K) applied in the support vector regression, which are the weak learners used in the ensemble learning methods evaluated here are as follows: the linear (LIN) (2), radial basis function (RBF) (3), and polynomial (POLY) (4). It should be noted that the selection of the transformation to be applied is determined by the type of data and is often performed by experimentation [74]. In this paper, these functions are evaluated to obtain the best application of the compared methods.
K ( x j , x k ) = x j x k .
K ( x j , x k ) = exp ( x j x k 2 ) .
K ( x j , x k ) = 1 + x j x k q .

3.5. Benchmarking

For the proposed method, the LSTM is considered. LSTM networks are a type of recurrent neural network architecture designed to model sequential data while addressing the vanishing or exploding gradient problem in traditional ANNs [77]. Given a sequence of input data, an LSTM unit maintains a hidden state vector h t and a cell state vector c t at time step t [25].
The LSTM unit consists of several gates that control the flow of information [78]. The input gate ( i t ) controls the flow of new information into the cell state, the forget gate ( f t ) controls the flow of old information to be forgotten from the cell state, and the output gate ( o t ) controls the flow of information from the cell state to the hidden state for output [79]. The computations in an LSTM unit are described as follows:
i t = σ ( W i · [ h t 1 , x t ] + b i ) f t = σ ( W f · [ h t 1 , x t ] + b f ) o t = σ ( W o · [ h t 1 , x t ] + b o ) c ˜ t = tanh ( W c · [ h t 1 , x t ] + b c ) c t = f t c t 1 + i t c ˜ t h t = o t tanh ( c t )
where W and b are weight matrices and bias vectors specific to each gate, c ˜ t is the candidate cell state, and σ and tanh are the sigmoid and hyperbolic tangent activation functions [80]. The LSTM network was employed for benchmarking using three optimizers as follows: root mean square propagation (RMSProp) [81], adaptive moment estimation (ADAM) [82], and stochastic gradient descent with momentum (SGDM) [83].
An adaptive neuro-fuzzy inference system (ANFIS) is also evaluated in comparison with the best method found in this work, considering the Sugeno FIS using fuzzy c-means (FCM), subtractive clustering, and grid partition. The ANFIS is a hybrid model that integrates the strengths of both ANNs and fuzzy logic principles, making it particularly effective for time series forecasting [84].
ANFIS harnesses the learning capabilities of neural networks to optimize the fuzzy inference system’s parameters, thereby enhancing the model’s ability to handle non-linear and complex patterns inherent in time series data [85]. ANFIS employs a Takagi–Sugeno-type fuzzy inference mechanism, which facilitates the generation of a fuzzy rule base from the input data, enabling a nuanced mapping of inputs to outputs. ANFIS stands out as a robust and versatile tool for time series forecasting, offering significant improvements in predictive performance and reliability [86].
The evaluation flowchart followed in this paper is presented in Figure 6. The proposed application is based on ensemble learning models that combine weak learners to create a meta-learner. For this, the bagging, boosting, and stacking approaches are used. A statistical analysis was carried out considering 50 experiments with each model to evaluate the variability in the results considering a random initialization of the network. The maximum, minimum, mean, standard deviation, and variance values are presented. The LSTM and ANFIS methods were used for comparative analysis (benchmarking).

3.6. Evaluation Setup

For comparison, the data are split using 70% for training and 30% for testing. All experiments were computed using an i5-7300HQ, with 20 GB of RAM, using the Matlab (2019a) software. For future comparisons, the ensemble learning models applied in this paper can be found at: https://github.com/vhrique/ELT (accessed on 24 October 2024). The stacked, boosted, and bagged ensemble learning models are analyzed. For model assessment, the mean squared error (MSE), mean absolute error (MAE), mean absolute percentage error (MAPE), and coefficient of determination are used, given by:
MSE = 1 n i = 1 n y i y ^ i 2
MAE = 1 n i = 1 n y i y ^ i
error = 1 n i = 1 n y i y ^ i y i × 100
R 2 = 1 i = 1 n y i y ^ i 2 i = 1 n y i y ¯ i 2
where y is the observed value, y ^ is the predicted value, and y ¯ is the average of the observed values [48].

4. Results and Discussion

In this section, the results of the predictions using the ensemble learning methods are discussed and a final statistical evaluation is presented considering the best parameters of each structure.

4.1. Ensemble Learning Model Evaluation

The results of the stacked ensemble learning model are presented in Table 1. Using the linear kernel function, all compared solvers had better results of error (MSE, MAE, and MAPE). The results using the polynomial kernel function were much higher, and this kernel function should be avoided for the stacked ensemble model. The best solver for this model was the SMO, which was slightly better than the other solvers considering the linear kernel function.
Given the poor results of the stacked ensemble learning model, other approaches are evaluated. In Table 2, the results of the prediction performances of the boosted ensemble learning model are presented considering the same variation of structures used previously.
Regarding the hyperparameters, the results of the boosted model were similar to the stacked ensemble learning model. Here, the linear kernel function was the best selection considering the error measures. However, for this model, the L1QP is the best solver and this structure has been used further for the final assessments.
In Table 3, the complete evaluation of the bagged model is presented. These results were more promising given that a MAPE of 0.073% and a coefficient of determination of 0.9752 were achieved. These results were achieved considering the SMO solver with the linear kernel function; additionally, the MSE and the MAE were better than the compared models and configurations.
In Figure 7, a comparison between the observed and the predicted values is presented for all the evaluated ensemble learning methods. In Figure 7A, major differences occurred when there were higher values of schedule generation, resulting in higher errors, making the stacked ensemble learning model unsuitable for this evaluation.
The boosted model (Figure 7B) was better than the previously evaluated staked model; however, there were still major differences between the predicted and the observed values. The bagged model, presented in Figure 7C, had promising results with predictions close to the desired output. The prediction results presented here consider one step ahead (1 h), which is adequate for a schedule generation problem. In the following, a statistical assessment of these ensemble learning models considering their best setup parameters is presented.

4.2. Statistical Evaluation

For the bagged and stacked ensemble learning approaches, the best solver was the SMO, and for the boosted, the best was the L1QP; all these results were considering the linear kernel function. Considering the best configuration for each model, a statistical analysis is presented in Table 4. This evaluation considers 50 runs with random initialization of the weights.
The bagging ensemble learning method has proved that in addition to having a lower error compared to other models, it is stable since for the MSE, the variation in the maximum error and minimum error was 2.38%, considering 50 experiments with random initialization. For this model, the coefficient of determination remains at 0.97 in all experiments.
Given the error results of the bagging model, the standard deviation was also low, proving that this structure has an acceptable performance regarding its robustness. The more stable model was the stacking ensemble learning method; however, its performance regarding the error was inferior to the bagging, which was the more suitable model for the task covered in this paper.

4.3. Comparative Analysis

In this subsection, a comparative analysis with variations of the LSTM and ANFIS models is presented. Specifically, for the LSTM, the RMSprop, Adam, and SGDM optimizers are evaluated, and the evaluated parameter is the number of hidden units, with 10, 50, 100, and 500 hidden units. For the ANFIS, the FCM, grid partition, and subtractive clustering are considered; for the FCM, the parameter evaluated is the number of clusters. Table 5 shows the results of this comparison in relation to the bagged ensemble learning model.
The ANFIS model using the grid partition only converged using two membership functions (MFs) for the input variables; when using more MFs, the model did not converge. When subtractive clustering is evaluated, no combination of parameters results in the convergence of the model. For this reason, the grid partition and subtractive clustering are not presented in Table 5. This issue also occurred when using CNN-LSTM; models based on deep learning did not result in acceptable prediction values.
For the LSTM, the best results were achieved using 500 hidden units in both RMSprop and SGDM. An attempt was made to use more hidden units but the model did not converge when a greater number of hidden units was used. For the Adam optimizer, the best result was using 100 hidden units. Among the optimizers used for the LSTM, the best result was using Adam; however, this result was inferior in all the metrics evaluated compared to the bagged ensemble model, proving, in this comparative analysis, that the approach used in this article is the most suitable to be applied to this challenge.

4.4. Comparison to Other Research

Comparatively, Ribeiro et al. [23] had an MSE of 388.46 (RMSE was originally presented), an R2 of 0.9544, and a symmetric MAPE of 0.0418 for three months ahead using a self-adaptive decomposition and heterogeneous ensemble learning method. The authors highlight that using more steps ahead results in greater difficulty in predicting future values. In this paper, a step ahead is considered, and this analysis is suitable for the proposed problem, resulting in better performance according to the results discussed here.
A very short-term load forecasting was also discussed by Yamasaki et al. [52]; with the promising results pointed out in their work, the use of ensemble learning methods, such as those applied in this paper. is proving to be a potentially valuable solution. The ensemble stacking model was used in [24]; this model was not the best solution here, emphasizing that a complete evaluation with several ensemble learning models is necessary for the best prediction strategy.
Noise reduction methods were applied by Stefenon et al. [53], Ribeiro et al. [87], and Moreno et al. [88], which may be interesting for improving the capabilities of the prediction model. However, if denoising is not used properly, it can result in unnecessary attenuation of the signal. Predicting time series relative to the original signal is a better strategy because no information is lost.

5. Conclusions

By employing time series forecasting techniques, this paper provides insights into the patterns, trends, and dependencies associated with diesel-powered thermal systems. Understanding these dynamics is pivotal for energy planners, policymakers, and industry stakeholders to make informed decisions regarding energy resource allocation, infrastructure development, and environmental sustainability.
Considering a MAPE of 0.089% and a coefficient of determination of 97.52% achieved by the bagged ensemble learning model (average considering 50 experiments), besides its robustness proved by statistical evaluation, this model is shown to be promising for short-term time series forecasting. By having more precise predictions about the use of thermal systems, it is possible to improve the management of the feedstock to be used for burning and generating electricity. This can prevent a shortage of the necessary inputs, resulting in a price spike in this type of generation.
This research lays the groundwork for informed decision-making, emphasizing the importance of balancing energy needs, environmental considerations, and long-term sustainability in Brazil’s energy landscape. Stakeholders can steer the nation towards a more resilient, diversified, and environmentally conscious energy future. Further research can be undertaken, including an evaluation of the relations between different regions since when there is a lack of rain and the hydropower dams have lower levels, generation using thermal systems is used in all regions of the country. For time series with higher noise, filters with higher intensity may be a solution that can be applied to related tasks.

Author Contributions

Conceptualization, software, formal analysis, writing—original draft preparation, W.G.B.; methodology, data curation, R.N.M.; writing—review and editing, supervision, R.C., A.N., C.T.d.C.J. and G.V.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work is part of the project ‘Self-adaptive platform based on intelligent agents for the optimization and management of operational processes in logistic ware-houses’ (PLAUTON) PID2023-151701OB-C21, funded by MCIN/AEI/10.13039/501100011033/FEDER, EU. The authors would like to thank the Coordination for the Improvement of Higher Education Personnel (CAPES) for the scholarship of the first author and the Council for Scientific and Technological Development (CNPq) for the grant of the fourth author. This study was financed (i) in part by CAPES under the doctoral scholarship number 88887.808258/2023-00, and (ii) by CNPq under grant number 310447/2021-6.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy.

Acknowledgments

Thanks are due to Stefenon, S. F. for his important contribution to this research.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Thirugnanam, K.; Kerk, S.K.; Yuen, C.; Liu, N.; Zhang, M. Energy management for renewable microgrid in reducing diesel generators usage with multiple types of battery. IEEE Trans. Ind. Electron. 2018, 65, 6772–6786. [Google Scholar] [CrossRef]
  2. Lei, L.; Tan, Y.; Dahlenburg, G.; Xiang, W.; Zheng, K. Dynamic energy dispatch based on deep reinforcement learning in IoT-driven smart isolated microgrids. IEEE Internet Things J. 2020, 8, 7938–7953. [Google Scholar] [CrossRef]
  3. Das, I.; Cañizares, C.A. Renewable energy integration in diesel-based microgrids at the Canadian arctic. Proc. IEEE 2019, 107, 1838–1856. [Google Scholar] [CrossRef]
  4. de Oliveira, A.K.V.; de Azevedo, K.L.R.; Dos Santos, D.O.; Aghaei, M.; Rüther, R.; Orabona, R.; Naspolini, H. Assessing the Potential of Green Hydrogen in Decarbonizing Off-Grid Amazonian Communities. In Proceedings of the 2023 International Conference on Future Energy Solutions (FES), Vaasa, Finland, 12–14 June 2023; IEEE: New York, NY, USA, 2023; pp. 1–6. [Google Scholar] [CrossRef]
  5. Adebayo, T.S.; Uhunamure, S.E.; Shale, K. A time-varying approach to the nexus between environmental related technologies, renewable energy consumption and environmental sustainability in South Africa. Sci. Rep. 2023, 13, 4860. [Google Scholar] [CrossRef]
  6. Zhang, B.; Wu, X.; Ghias, A.M.; Chen, Z. Coordinated carbon capture systems and power-to-gas dynamic economic energy dispatch strategy for electricity–gas coupled systems considering system uncertainty: An improved soft actor–critic approach. Energy 2023, 271, 126965. [Google Scholar] [CrossRef]
  7. Ahmed, I.; Rehan, M.; Basit, A.; Hong, K.S. Greenhouse gases emission reduction for electric power generation sector by efficient dispatching of thermal plants integrated with renewable systems. Sci. Rep. 2022, 12, 12380. [Google Scholar] [CrossRef]
  8. Akinyemi, A.S.; Musasa, K.; Davidson, I.E. Analysis of voltage rise phenomena in electrical power network with high concentration of renewable distributed generations. Sci. Rep. 2022, 12, 7815. [Google Scholar] [CrossRef]
  9. dos Santos, K.V.; Colonetti, B.; Finardi, E.C.; Zavala, V.M. Accelerated dual dynamic integer programming applied to short-term power generation scheduling. Int. J. Electr. Power Energy Syst. 2023, 145, 108689. [Google Scholar] [CrossRef]
  10. Zhang, G.; Wang, W.; Chen, Z.; Li, R.; Niu, Y. Modeling and optimal dispatch of a carbon-cycle integrated energy system for low-carbon and economic operation. Energy 2022, 240, 122795. [Google Scholar] [CrossRef]
  11. Xiang, Y.; Wu, G.; Shen, X.; Ma, Y.; Gou, J.; Xu, W.; Liu, J. Low-carbon economic dispatch of electricity-gas systems. Energy 2021, 226, 120267. [Google Scholar] [CrossRef]
  12. Long, Y.; Li, Y.; Wang, Y.; Cao, Y.; Jiang, L.; Zhou, Y.; Deng, Y.; Nakanishi, Y. Low-carbon economic dispatch considering integrated demand response and multistep carbon trading for multi-energy microgrid. Sci. Rep. 2022, 12, 6218. [Google Scholar] [CrossRef] [PubMed]
  13. Li, J.; Luo, Y.; Wei, S. Long-term electricity consumption forecasting method based on system dynamics under the carbon-neutral target. Energy 2022, 244, 122572. [Google Scholar] [CrossRef]
  14. Muniz, R.N.; Stefenon, S.F.; Buratto, W.G.; Nied, A.; Meyer, L.H.; Finardi, E.C.; Kuhl, R.M.; S’a, J.A.S.d.; Rocha, B.R.P.d. Tools for measuring energy sustainability: A comparative review. Energies 2020, 13, 2366. [Google Scholar] [CrossRef]
  15. Brandão, P.C.; de Souza, A.L.; Rousset, P.; Simas, F.N.B.; de Mendonça, B.A.F. Forest biomass as a viable pathway for sustainable energy supply in isolated villages of Amazonia. Environ. Dev. 2021, 37, 100609. [Google Scholar] [CrossRef]
  16. Morawski, A.P.; de Araújo, L.R.; Schiaffino, M.S.; de Oliveira, R.C.; Chun, A.; Ribeiro, L.C.; Santos, J.J.C.S.; Donatelli, J.L.M.; Cunha, C.C.M. On the suitable superstructure thermoeconomic optimization of a waste heat recovery system for a Brazilian diesel engine power plant. Energy Convers. Manag. 2021, 234, 113947. [Google Scholar] [CrossRef]
  17. Barbosa, M.O.; Peyerl, D.; Mendes, A.B. The economic and environmental benefits of adopting natural gas in isolated systems of Amazonas state, Brazil. Environ. Dev. 2023, 47, 100889. [Google Scholar] [CrossRef]
  18. Li, F.; Zheng, H.; Li, X.; Yang, F. Day-ahead city natural gas load forecasting based on decomposition-fusion technique and diversified ensemble learning model. Appl. Energy 2021, 303, 117623. [Google Scholar] [CrossRef]
  19. Kanaan, L.; Ismail, L.S.; Gowid, S.; Meskin, N.; Massoud, A.M. Optimal energy dispatch engine for PV-DG-ESS hybrid power plants considering battery degradation and carbon emissions. IEEE Access 2023, 11, 58506–58515. [Google Scholar] [CrossRef]
  20. Taheri, S.I.; Salles, M.B.; Costa, E.C. Optimal cost management of distributed generation units and microgrids for virtual power plant scheduling. IEEE Access 2020, 8, 208449–208461. [Google Scholar] [CrossRef]
  21. Yu, H.; Yang, Y.; Li, B.; Liu, B.; Guo, Y.; Wang, Y.; Guo, Z.; Meng, R. Research on the community electric carbon emission prediction considering the dynamic emission coefficient of power system. Sci. Rep. 2023, 13, 5568. [Google Scholar] [CrossRef]
  22. Galicia, A.; Talavera-Llames, R.; Troncoso, A.; Koprinska, I.; Martínez-Álvarez, F. Multi-step forecasting for big data time series based on ensemble learning. Knowl.-Based Syst. 2019, 163, 830–841. [Google Scholar] [CrossRef]
  23. Ribeiro, M.H.D.M.; Stefenon, S.F.; de Lima, J.D.; Nied, A.; Mariani, V.C.; Coelho, L.S. Electricity price forecasting based on self-adaptive decomposition and heterogeneous ensemble learning. Energies 2020, 13, 5190. [Google Scholar] [CrossRef]
  24. Ribeiro, M.H.D.M.; da Silva, R.G.; Moreno, S.R.; Mariani, V.C.; dos Santos Coelho, L. Efficient bootstrap stacking ensemble learning model applied to wind power generation forecasting. Int. J. Electr. Power Energy Syst. 2022, 136, 107712. [Google Scholar] [CrossRef]
  25. Sopelsa Neto, N.F.; Stefenon, S.F.; Meyer, L.H.; Ovejero, R.G.; Leithardt, V.R.Q. Fault prediction based on leakage current in contaminated insulators using enhanced time series forecasting models. Sensors 2022, 22, 6121. [Google Scholar] [CrossRef] [PubMed]
  26. Alba, E.L.; Oliveira, G.A.; Ribeiro, M.H.D.M.; Rodrigues, E.O. Electricity Consumption Forecasting: An Approach Using Cooperative Ensemble Learning with SHapley Additive exPlanations. Forecasting 2024, 6, 839–863. [Google Scholar] [CrossRef]
  27. Stefenon, S.F.; Seman, L.O.; da Silva, L.S.A.; Mariani, V.C.; dos Santos Coelho, L. Hypertuned temporal fusion transformer for multi-horizon time series forecasting of dam level in hydroelectric power plants. Int. J. Electr. Power Energy Syst. 2024, 157, 109876. [Google Scholar] [CrossRef]
  28. Starke, L.; Hoppe, A.F.; Sartori, A.; Stefenon, S.F.; Santana, J.F.D.P.; Leithardt, V.R.Q. Interference recommendation for the pump sizing process in progressive cavity pumps using graph neural networks. Sci. Rep. 2023, 13, 16884. [Google Scholar] [CrossRef]
  29. Surek, G.A.S.; Seman, L.O.; Stefenon, S.F.; Mariani, V.C.; Coelho, L.S. Video-based human activity recognition using deep learning approaches. Sensors 2023, 23, 6384. [Google Scholar] [CrossRef]
  30. Stefenon, S.F.; Seman, L.O.; Schutel Furtado Neto, C.; Nied, A.; Seganfredo, D.M.; Garcia da Luz, F.; Sabino, P.H.; Torreblanca González, J.; Quietinho Leithardt, V.R. Electric field evaluation using the finite element method and proxy models for the design of stator slots in a permanent magnet synchronous motor. Electronics 2020, 9, 1975. [Google Scholar] [CrossRef]
  31. dos Santos, G.H.; Seman, L.O.; Bezerra, E.A.; Leithardt, V.R.Q.; Mendes, A.S.; Stefenon, S.F. Static attitude determination using convolutional neural networks. Sensors 2021, 21, 6419. [Google Scholar] [CrossRef]
  32. Larcher, J.H.K.; Stefenon, S.F.; dos Santos Coelho, L.; Mariani, V.C. Enhanced multi-step streamflow series forecasting using hybrid signal decomposition and optimized reservoir computing models. Expert Syst. Appl. 2024, 255, 124856. [Google Scholar] [CrossRef]
  33. Singh, G.; Stefenon, S.F.; Yow, K.C. Interpretable visual transmission lines inspections using pseudo-prototypical part network. Mach. Vis. Appl. 2023, 34, 41. [Google Scholar] [CrossRef]
  34. Medeiros, A.; Sartori, A.; Stefenon, S.F.; Meyer, L.H.; Nied, A. Comparison of artificial intelligence techniques to failure prediction in contaminated insulators based on leakage current. J. Intell. Fuzzy Syst. 2022, 42, 3285–3298. [Google Scholar] [CrossRef]
  35. Stefenon, S.F.; Kasburg, C.; Freire, R.Z.; Silva Ferreira, F.C.; Bertol, D.W.; Nied, A. Photovoltaic power forecasting using wavelet Neuro-Fuzzy for active solar trackers. J. Intell. Fuzzy Syst. 2021, 40, 1083–1096. [Google Scholar] [CrossRef]
  36. Pazikadin, A.R.; Rifai, D.; Ali, K.; Malik, M.Z.; Abdalla, A.N.; Faraj, M.A. Solar irradiance measurement instrumentation and power solar generation forecasting based on Artificial Neural Networks (ANN): A review of five years research trend. Sci. Total Environ. 2020, 715, 136848. [Google Scholar] [CrossRef] [PubMed]
  37. Ahmed, A.; Khalid, M. A review on the selected applications of forecasting models in renewable power systems. Renew. Sustain. Energy Rev. 2019, 100, 9–21. [Google Scholar] [CrossRef]
  38. Corso, M.P.; Stefenon, S.F.; Singh, G.; Matsuo, M.V.; Perez, F.L.; Leithardt, V.R.Q. Evaluation of visible contamination on power grid insulators using convolutional neural networks. Electr. Eng. 2023, 105, 3881–3894. [Google Scholar] [CrossRef]
  39. Souza, B.J.; Stefenon, S.F.; Singh, G.; Freire, R.Z. Hybrid-YOLO for classification of insulators defects in transmission lines based on UAV. Int. J. Electr. Power Energy Syst. 2023, 148, 108982. [Google Scholar] [CrossRef]
  40. Borré, A.; Seman, L.O.; Camponogara, E.; Stefenon, S.F.; Mariani, V.C.; Coelho, L.S. Machine fault detection using a hybrid CNN-LSTM attention-based model. Sensors 2023, 23, 4512. [Google Scholar] [CrossRef]
  41. Buratto, W.G.; Muniz, R.N.; Nied, A.; Barros, C.F.d.O.; Cardoso, R.; Gonzalez, G.V. Wavelet CNN-LSTM time series forecasting of electricity power generation considering biomass thermal systems. IET Gener. Transm. Distrib. 2024, 1–15. [Google Scholar] [CrossRef]
  42. Stefenon, S.F.; Seman, L.O.; Klaar, A.C.R.; Ovejero, R.G.; Leithardt, V.R.Q. Hypertuned-YOLO for interpretable distribution power grid fault location based on EigenCAM. Ain Shams Eng. J. 2024, 15, 102722. [Google Scholar] [CrossRef]
  43. Stefenon, S.F.; Singh, G.; Souza, B.J.; Freire, R.Z.; Yow, K.C. Optimized hybrid YOLOu-Quasi-ProtoPNet for insulators classification. IET Gener. Transm. Distrib. 2023, 17, 3501–3511. [Google Scholar] [CrossRef]
  44. da Silva, E.C.; Finardi, E.C.; Stefenon, S.F. Enhancing hydroelectric inflow prediction in the Brazilian power system: A comparative analysis of machine learning models and hyperparameter optimization for decision support. Electr. Power Syst. Res. 2024, 230, 110275. [Google Scholar] [CrossRef]
  45. Vargas, S.A.; Esteves, G.R.T.; Maçaira, P.M.; Bastos, B.Q.; Cyrino Oliveira, F.L.; Souza, R.C. Wind power generation: A review and a research agenda. J. Clean. Prod. 2019, 218, 850–870. [Google Scholar] [CrossRef]
  46. Akhter, M.N.; Mekhilef, S.; Mokhlis, H.; Mohamed Shah, N. Review on forecasting of photovoltaic power generation based on machine learning and metaheuristic techniques. IET Renew. Power Gener. 2019, 13, 1009–1023. [Google Scholar] [CrossRef]
  47. Nazir, M.S.; Alturise, F.; Alshmrany, S.; Nazir, H.M.J.; Bilal, M.; Abdalla, A.N.; Sanjeevikumar, P.; Ali, Z.M. Wind generation forecasting methods and proliferation of artificial neural network: A review of five years research trend. Sustainability 2020, 12, 3778. [Google Scholar] [CrossRef]
  48. Seman, L.O.; Stefenon, S.F.; Mariani, V.C.; Coelho, L.S. Ensemble learning methods using the Hodrick–Prescott filter for fault forecasting in insulators of the electrical power grids. Int. J. Electr. Power Energy Syst. 2023, 152, 109269. [Google Scholar] [CrossRef]
  49. Stefenon, S.F.; Freire, R.Z.; Meyer, L.H.; Corso, M.P.; Sartori, A.; Nied, A.; Klaar, A.C.R.; Yow, K.C. Fault detection in insulators based on ultrasonic signal processing using a hybrid deep learning technique. IET Sci. Meas. Technol. 2020, 14, 953–961. [Google Scholar] [CrossRef]
  50. Yang, S.; Liu, J. Time-Series Forecasting Based on High-Order Fuzzy Cognitive Maps and Wavelet Transform. IEEE Trans. Fuzzy Syst. 2018, 26, 3391–3402. [Google Scholar] [CrossRef]
  51. Ustundag, B.B.; Kulaglic, A. High-Performance Time Series Prediction With Predictive Error Compensated Wavelet Neural Networks. IEEE Access 2020, 8, 210532–210541. [Google Scholar] [CrossRef]
  52. Yamasaki, M.; Freire, R.Z.; Seman, L.O.; Stefenon, S.F.; Mariani, V.C.; dos Santos Coelho, L. Optimized hybrid ensemble learning approaches applied to very short-term load forecasting. Int. J. Electr. Power Energy Syst. 2024, 155, 109579. [Google Scholar] [CrossRef]
  53. Stefenon, S.F.; Seman, L.O.; Sopelsa Neto, N.F.; Meyer, L.H.; Mariani, V.C.; Coelho, L.d.S. Group method of data handling using Christiano-Fitzgerald random walk filter for insulator fault prediction. Sensors 2023, 23, 6118. [Google Scholar] [CrossRef]
  54. Xu, K.; Chen, L.; Patenaude, J.M.; Wang, S. Kernel representation learning with dynamic regime discovery for time series forecasting. In Pacific-Asia Conference on Knowledge Discovery and Data Mining; Springer: Singapore, 2024; pp. 251–263. [Google Scholar] [CrossRef]
  55. Xu, K.; Chen, L.; Patenaude, J.M.; Wang, S. Rhine: A regime-switching model with nonlinear representation for discovering and forecasting regimes in financial markets. In 2024 SIAM International Conference on Data Mining (SDM); SIAM: Bangkok, Thailand, 2024; pp. 526–534. [Google Scholar] [CrossRef]
  56. Klaar, A.C.R.; Stefenon, S.F.; Seman, L.O.; Mariani, V.C.; Coelho, L.S. Optimized EWT-Seq2Seq-LSTM with attention mechanism to insulators fault prediction. Sensors 2023, 23, 3202. [Google Scholar] [CrossRef]
  57. Stefenon, S.F.; Seman, L.O.; Aquino, L.S.; dos Santos Coelho, L. Wavelet-Seq2Seq-LSTM with attention for time series forecasting of level of dams in hydroelectric power plants. Energy 2023, 274, 127350. [Google Scholar] [CrossRef]
  58. Branco, N.W.; Cavalca, M.S.M.; Stefenon, S.F.; Leithardt, V.R.Q. Wavelet LSTM for fault forecasting in electrical power grids. Sensors 2022, 22, 8323. [Google Scholar] [CrossRef] [PubMed]
  59. Stefenon, S.F.; Seman, L.O.; Mariani, V.C.; Coelho, L.S. Aggregating prophet and seasonal trend decomposition for time series forecasting of Italian electricity spot prices. Energies 2023, 16, 1371. [Google Scholar] [CrossRef]
  60. Klaar, A.C.R.; Stefenon, S.F.; Seman, L.O.; Mariani, V.C.; Coelho, L.S. Structure optimization of ensemble learning methods and seasonal decomposition approaches to energy price forecasting in Latin America: A case study about Mexico. Energies 2023, 16, 3184. [Google Scholar] [CrossRef]
  61. Júnior, E.P.S.; Martins, K.D.L.D.C.; Silva, M.V.B.D.; Maurício, C.F.B.; Menezes, R.S.C.; Coelho Junior, L.M. Forest Bioelectricity in Brazil: Distribution and Spatial-Time Dependence. IEEE Access 2022, 10, 132822–132835. [Google Scholar] [CrossRef]
  62. Hafeez, G.; Alimgeer, K.S.; Qazi, A.B.; Khan, I.; Usman, M.; Khan, F.A.; Wadud, Z. A Hybrid Approach for Energy Consumption Forecasting With a New Feature Engineering and Optimization Framework in Smart Grid. IEEE Access 2020, 8, 96210–96226. [Google Scholar] [CrossRef]
  63. Stefenon, S.F.; Bruns, R.; Sartori, A.; Meyer, L.H.; Ovejero, R.G.; Leithardt, V.R.Q. Analysis of the ultrasonic signal in polymeric contaminated insulators through ensemble learning methods. IEEE Access 2022, 10, 33980–33991. [Google Scholar] [CrossRef]
  64. Moon, J.; Jung, S.; Rew, J.; Rho, S.; Hwang, E. Combination of short-term load forecasting models based on a stacking ensemble approach. Energy Build. 2020, 216, 109921. [Google Scholar] [CrossRef]
  65. Júnior, D.S.d.O.S.; de Mattos Neto, P.S.; de Oliveira, J.F.; Cavalcanti, G.D. A hybrid system based on ensemble learning to model residuals for time series forecasting. Inf. Sci. 2023, 649, 119614. [Google Scholar] [CrossRef]
  66. Zhang, S.; Chen, Y.; Zhang, W.; Feng, R. A novel ensemble deep learning model with dynamic error correction and multi-objective ensemble pruning for time series forecasting. Inf. Sci. 2021, 544, 427–445. [Google Scholar] [CrossRef]
  67. Liu, H.; Yu, C.; Wu, H.; Duan, Z.; Yan, G. A new hybrid ensemble deep reinforcement learning model for wind speed short term forecasting. Energy 2020, 202, 117794. [Google Scholar] [CrossRef]
  68. Stefenon, S.F.; Ribeiro, M.H.D.M.; Nied, A.; Mariani, V.C.; Coelho, L.S.; Leithardt, V.R.Q.; Silva, L.A.; Seman, L.O. Hybrid wavelet stacking ensemble model for insulators contamination forecasting. IEEE Access 2021, 9, 66387–66397. [Google Scholar] [CrossRef]
  69. Tan, M.; Yuan, S.; Li, S.; Su, Y.; Li, H.; He, F. Ultra-Short-Term Industrial Power Demand Forecasting Using LSTM Based Hybrid Ensemble Learning. IEEE Trans. Power Syst. 2020, 35, 2937–2948. [Google Scholar] [CrossRef]
  70. Ribeiro, M.H.D.M.; dos Santos Coelho, L. Ensemble approach based on bagging, boosting and stacking for short-term prediction in agribusiness time series. Appl. Soft Comput. 2020, 86, 105837. [Google Scholar] [CrossRef]
  71. Andiojaya, A.; Demirhan, H. A bagging algorithm for the imputation of missing values in time series. Expert Syst. Appl. 2019, 129, 10–26. [Google Scholar] [CrossRef]
  72. Li, T.; Liu, X.; Lin, Z.; Morrison, R. Ensemble offshore Wind Turbine Power Curve modelling—An integration of Isolation Forest, fast Radial Basis Function Neural Network, and metaheuristic algorithm. Energy 2022, 239, 122340. [Google Scholar] [CrossRef]
  73. Kecman, V.; Zigic, L. Algorithms for direct L2 support vector machines. In Proceedings of the IEEE International Symposium on Innovations in Intelligent Systems and Applications (INISTA) Proceedings, Alberobello, Italy, 23–25 June 2014; IEEE: New York, NY, USA, 2014; pp. 419–424. [Google Scholar] [CrossRef]
  74. Stefenon, S.F.; Ribeiro, M.H.D.M.; Nied, A.; Yow, K.C.; Mariani, V.C.; Coelho, L.S.; Seman, L.O. Time series forecasting using ensemble learning methods for emergency prevention in hydroelectric power plants with dam. Electr. Power Syst. Res. 2022, 202, 107584. [Google Scholar] [CrossRef]
  75. Wambacq, J.; Ulloa, J.; Lombaert, G.; François, S. Interior-point methods for the phase-field approach to brittle and ductile fracture. Comput. Methods Appl. Mech. Eng. 2021, 375, 113612. [Google Scholar] [CrossRef]
  76. Haeser, G.; Ramos, A. Constraint qualifications for Karush–Kuhn–Tucker conditions in multiobjective optimization. J. Optim. Theory Appl. 2020, 187, 469–487. [Google Scholar] [CrossRef]
  77. Kasburg, C.; Stefenon, S.F. Deep learning for photovoltaic generation forecast in active solar trackers. IEEE Lat. Am. Trans. 2019, 17, 2013–2019. [Google Scholar] [CrossRef]
  78. Chimmula, V.K.R.; Zhang, L. Time series forecasting of COVID-19 transmission in Canada using LSTM networks. Chaos Solitons Fractals 2020, 135, 109864. [Google Scholar] [CrossRef] [PubMed]
  79. Stefenon, S.F.; Kasburg, C.; Nied, A.; Klaar, A.C.R.; Ferreira, F.C.S.; Branco, N.W. Hybrid deep learning for power generation forecasting in active solar trackers. IET Gener. Transm. Distrib. 2020, 14, 5667–5674. [Google Scholar] [CrossRef]
  80. Wang, W.; Shao, J.; Jumahong, H. Fuzzy inference-based LSTM for long-term time series prediction. Sci. Rep. 2023, 13, 20359. [Google Scholar] [CrossRef]
  81. Elshamy, R.; Abu-Elnasr, O.; Elhoseny, M.; Elmougy, S. Improving the efficiency of RMSProp optimizer by utilizing Nestrove in deep learning. Sci. Rep. 2023, 13, 8814. [Google Scholar] [CrossRef]
  82. Ahmed, F.R.; Alsenany, S.A.; Abdelaliem, S.M.F.; Deif, M.A. Development of a hybrid LSTM with chimp optimization algorithm for the pressure ventilator prediction. Sci. Rep. 2023, 13, 20927. [Google Scholar] [CrossRef]
  83. Fernandes, F.; Stefenon, S.F.; Seman, L.O.; Nied, A.; Ferreira, F.C.S.; Subtil, M.C.M.; Klaar, A.C.R.; Leithardt, V.R.Q. Long short-term memory stacking model to predict the number of cases and deaths caused by COVID-19. J. Intell. Fuzzy Syst. 2022, 6, 6221–6234. [Google Scholar] [CrossRef]
  84. Salehi, S. Employing a Time Series Forecasting Model for Tourism Demand Using ANFIS. J. Inf. Organ. Sci. 2022, 46, 157–172. [Google Scholar] [CrossRef]
  85. Fatemi, S.E.; Parvini, H. The impact assessments of the ACF shape on time series forecasting by the ANFIS model. Neural Comput. Appl. 2022, 34, 12723–12736. [Google Scholar] [CrossRef]
  86. Stefenon, S.F.; Freire, R.Z.; Coelho, L.S.; Meyer, L.H.; Grebogi, R.B.; Buratto, W.G.; Nied, A. Electrical insulator fault forecasting based on a wavelet neuro-fuzzy system. Energies 2020, 13, 484. [Google Scholar] [CrossRef]
  87. Ribeiro, M.H.D.M.; da Silva, R.G.; Moreno, S.R.; Canton, C.; Larcher, J.H.K.; Stefenon, S.F.; Mariani, V.C.; dos Santos Coelho, L. Variational mode decomposition and bagging extreme learning machine with multi-objective optimization for wind power forecasting. Appl. Intell. 2024, 1, 1. [Google Scholar] [CrossRef]
  88. Moreno, S.R.; Seman, L.O.; Stefenon, S.F.; dos Santos Coelho, L.; Mariani, V.C. Enhancing wind speed forecasting through synergy of machine learning, singular spectral analysis, and variational mode decomposition. Energy 2024, 292, 130493. [Google Scholar] [CrossRef]
Figure 1. Diesel oil for electricity generation by Amazon legal region states (GWh). Source: Brazilian Energy Balance, Ministry of Mines and Energy of Brazil.
Figure 1. Diesel oil for electricity generation by Amazon legal region states (GWh). Source: Brazilian Energy Balance, Ministry of Mines and Energy of Brazil.
Applsci 14 09806 g001
Figure 2. Time series of the thermal generation scheduled for the northern region of Brazil.
Figure 2. Time series of the thermal generation scheduled for the northern region of Brazil.
Applsci 14 09806 g002
Figure 3. Stacking ensemble model architecture.
Figure 3. Stacking ensemble model architecture.
Applsci 14 09806 g003
Figure 4. Boosting ensemble model architecture.
Figure 4. Boosting ensemble model architecture.
Applsci 14 09806 g004
Figure 5. Bagging ensemble model architecture.
Figure 5. Bagging ensemble model architecture.
Applsci 14 09806 g005
Figure 6. Evaluation procedure applied in this paper.
Figure 6. Evaluation procedure applied in this paper.
Applsci 14 09806 g006
Figure 7. Predicted versus observed values: (A) Stacking; (B) Boosting; (C) Bagging ensemble models.
Figure 7. Predicted versus observed values: (A) Stacking; (B) Boosting; (C) Bagging ensemble models.
Applsci 14 09806 g007
Table 1. Results of stacked ensemble learning model.
Table 1. Results of stacked ensemble learning model.
SolverKernelMSEMAEMAPE (%)R2Time (s)
ISDALIN1.49  × 10 5 1.76  × 10 2 7.030.41981.28
RBF3.78  × 10 5 3.13  × 10 2 1.32  × 10 1 0.47350.65
POLY1.03  × 10 12 1.58  × 10 5 5.39  × 10 3 -40.89
L1QPLIN1.49  × 10 5 1.76  × 10 2 7.020.420210.53
RBF4.06  × 10 5 3.28  × 10 2 1.39  × 10 1 0.581212.62
POLY1.25  × 10 9 3.67  × 10 3 1.21  × 10 2 -9.33
SMOLIN1.48  × 10 5 1.76  × 10 2 7.020.42182.70
RBF4.06  × 10 5 3.27  × 10 2 1.39  × 10 1 0.580910.57
POLY8.92  × 10 9 1.64  × 10 4 5.51  × 10 2 -95.33
Best results are in bold.
Table 2. Results of boosted ensemble learning model.
Table 2. Results of boosted ensemble learning model.
SolverKernelMSEMAEMAPE (%)R2Time (s)
ISDALIN1.25  × 10 5 1.20  × 10 2 5.500.511131.82
RBF4.48  × 10 5 4.06  × 10 2 1.94  × 10 1 0.74616.74
POLY4.43  × 10 7 2.06  × 10 3 7.56  × 10 1 -1136.64
L1QPLIN5.19  × 10 4 1.03  × 10 2 4.960.7977179.57
RBF4.66  × 10 5 4.15  × 10 2 1.98  × 10 1 0.8181167.07
POLY5.03  × 10 7 1.86  × 10 3 6.86  × 10 1 -214.36
SMOLIN8.51  × 10 4 1.54  × 10 2 7.880.668249.58
RBF4.66  × 10 5 4.17  × 10 2 2.00  × 10 1 0.81564.74
POLY6.83  × 10 7 2.30  × 10 3 8.70  × 10 1 -1847.39
Best results are in bold.
Table 3. Results of bagged ensemble learning model.
Table 3. Results of bagged ensemble learning model.
SolverKernelMSEMAEMAPE (%)R2Time (s)
ISDALIN6.39  × 10 3 4.201.02  × 10 1 0.97519.71
RBF4.21  × 10 5 3.42  × 10 2 1.48  × 10 1 0.64031.91
POLY1.00  × 10 6 2.65  × 10 2 1.00  × 10 1 -430.01
L1QPLIN6.39  × 10 3 4.239.71  × 10 2 0.975136.75
RBF4.48  × 10 5 3.55  × 10 2 1.54  × 10 1 0.748241.035
POLY4.03  × 10 5 6.01  × 10 1 1.910.572137.813
SMOLIN6.36  × 10 3 3.777.34  × 10 2 0.975216.97
RBF4.49  × 10 5 3.56  × 10 2 1.54  × 10 1 0.74941.52
POLY4.43  × 10 5 6.99  × 10 1 2.24-865.87
Best results are in bold.
Table 4. Statistical analysis of ensemble learning approaches.
Table 4. Statistical analysis of ensemble learning approaches.
StackingBoostingBagging
MSEMax1.48  × 10 5 1.40  × 10 5 6.44  × 10 3
Min1.48  × 10 5 3.80  × 10 4 6.29  × 10 3
Mean1.48  × 10 5 8.57  × 10 4 6.38  × 10 3
Std Deviation8.82  × 10 11 2.63  × 10 4 3.35  × 10 1
Variance7.78  × 10 21 6.90  × 10 8 1.12  × 10 3
MAEMax1.76  × 10 2 2.14  × 10 2 4.57
Min1.76  × 10 2 7.90  × 10 1 3.7
Mean1.76  × 10 2 1.62  × 10 2 4.04
Std Deviation1.15  × 10 13 3.25  × 10 1 1.83  × 10 1
Variance1.32  × 10 26 1.06  × 10 3 3.34  × 10 2
MAPEMax7.021.19  × 10 1 1.22  × 10 1
Min7.023.657.25  × 10 2
Mean7.028.448.90  × 10 2
Std Deviation3.59  × 10 15 2.051.18  × 10 2
Variance1.29  × 10 29 4.191.39  × 10 4
R2Max4.22  × 10 1 8.52  × 10 1 9.75  × 10 1
Min4.22  × 10 1 4.56  × 10 1 9.75  × 10 1
Mean4.22  × 10 1 6.66  × 10 1 9.75  × 10 1
Std Deviation-1.02  × 10 1 1.31  × 10 4
Variance-1.05  × 10 2 1.71  × 10 8
Table 5. Comparison of ensemble bagged method to LSTM model.
Table 5. Comparison of ensemble bagged method to LSTM model.
ModelMethodEvaluated
Parameter
MSEMAEMAPE (%)R2Time (s)
LSTMRMSprop101.65  × 10 5 1.81  × 10 2 7.310.357642.21
506.06  × 10 4 8.31  × 10 1 2.890.763427.81
1005.11  × 10 4 8.21  × 10 1 3.110.800629.40
5002.88  × 10 4 6.33  × 10 1 2.870.887655.55
Adam101.22  × 10 5 1.55  × 10 2 6.330.524627.65
503.31  × 10 4 5.29  × 10 1 2.070.871029.76
1001.06  × 10 4 2.00  × 10 1 0.710.958628.72
5001.75  × 10 4 5.38  × 10 1 2.190.931756.04
SGDM101.25  × 10 5 1.58  × 10 2 6.500.513325.97
505.88  × 10 4 8.03  × 10 1 3.060.770526.16
1004.79  × 10 4 6.65  × 10 1 2.390.813126.72
5001.26  × 10 4 1.14  × 10 1 0.290.951053.84
ANFISFCM101.71  × 10 4 6.00  × 10 1 2.660.93346.11
202.59  × 10 4 7.87  × 10 1 3.490.899111.16
301.20  × 10 7 1.53  × 10 3 6.48  × 10 1 -17.15
402.85  × 10 9 6.35  × 10 2 2.60  × 10 1 -25.99
502.16  × 10 10 5.82  × 10 3 2.51  × 10 2 -37.936
Bagged
Model
--6.36  × 10 3 4.048.95 × 10 2 0.975215.59
Best results are in bold.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Buratto, W.G.; Muniz, R.N.; Cardoso, R.; Nied, A.; da Costa, C.T., Jr.; Gonzalez, G.V. Time Series Forecasting of Thermal Systems Dispatch in Legal Amazon Using Machine Learning. Appl. Sci. 2024, 14, 9806. https://doi.org/10.3390/app14219806

AMA Style

Buratto WG, Muniz RN, Cardoso R, Nied A, da Costa CT Jr., Gonzalez GV. Time Series Forecasting of Thermal Systems Dispatch in Legal Amazon Using Machine Learning. Applied Sciences. 2024; 14(21):9806. https://doi.org/10.3390/app14219806

Chicago/Turabian Style

Buratto, William Gouvêa, Rafael Ninno Muniz, Rodolfo Cardoso, Ademir Nied, Carlos Tavares da Costa, Jr., and Gabriel Villarrubia Gonzalez. 2024. "Time Series Forecasting of Thermal Systems Dispatch in Legal Amazon Using Machine Learning" Applied Sciences 14, no. 21: 9806. https://doi.org/10.3390/app14219806

APA Style

Buratto, W. G., Muniz, R. N., Cardoso, R., Nied, A., da Costa, C. T., Jr., & Gonzalez, G. V. (2024). Time Series Forecasting of Thermal Systems Dispatch in Legal Amazon Using Machine Learning. Applied Sciences, 14(21), 9806. https://doi.org/10.3390/app14219806

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop