Next Article in Journal
A Survey on Industrial Internet of Things Security: Requirements, Attacks, AI-Based Solutions, and Edge Computing Opportunities
Previous Article in Journal
VIDAR-Based Road-Surface-Pothole-Detection Method
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dual-Path Information Fusion and Twin Attention-Driven Global Modeling for Solar Irradiance Prediction

1
Faculty of Information Engineering and Automation, Kunming University of Science and Technology, Kunming 650500, China
2
Key Laboratory of Artificial Intelligence in Yunnan Province, Kunming 650500, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(17), 7469; https://doi.org/10.3390/s23177469
Submission received: 21 July 2023 / Revised: 23 August 2023 / Accepted: 23 August 2023 / Published: 28 August 2023
(This article belongs to the Section Remote Sensors)

Abstract

:
Accurate prediction of solar irradiance holds significant value for renewable energy usage and power grid management. However, traditional forecasting methods often overlook the time dependence of solar irradiance sequences and the varying importance of different influencing factors. To address this issue, this study proposes a dual-path information fusion and twin attention-driven solar irradiance forecasting model. The proposed framework comprises three components: a residual attention temporal convolution block (RACB), a dual-path information fusion module (DIFM), and a twin self-attention module (TSAM). These components collectively enhance the performance of multi-step solar irradiance forecasting. First, the RACB is designed to enable the network to adaptively learn important features while suppressing irrelevant ones. Second, the DIFM is implemented to reinforce the model’s robustness against input data variations and integrate multi-scale features. Lastly, the TSAM is introduced to extract long-term temporal dependencies from the sequence and facilitate multi-step prediction. In the solar irradiance forecasting experiments, the proposed model is compared with six benchmark models across four datasets. In the one-step predictions, the average performance metrics RMSE, MAE, and MAPE of the four datasets decreased within the ranges of 0.463–2.390 W / m 2 , 0.439–2.005 W / m 2 , and 1.3–9.2%, respectively. Additionally, the average R 2 value across the four datasets increased by 0.008 to 0.059. The experimental results indicate that the model proposed in this study exhibits enhanced accuracy and robustness in predictive performance, making it a reliable alternative for solar irradiance forecasting.

1. Introduction

As environmental pollution and the energy crisis worsen, focus on renewable energy is growing. Governments worldwide are prioritizing and supporting dependable and financially sustainable electric power systems. The International Energy Agency’s (IEA) 2023 Electricity Market Report reveals a nearly 11% growth in global renewable energy capacity in 2022, with solar capacity surging almost 18%. However, despite the growth in solar energy’s contribution to renewable energy, its penetration into the mainstream energy market remains modest [1]. This is primarily attributed to the inherent intermittency and instability caused by the fluctuating nature of solar radiation. Such volatility in solar photovoltaic (PV) systems significantly impacts the stability and reliability of the power supply, underscoring the need for enhanced predictability and security, while accurate forecasting of solar irradiance faces challenges in data acquisition and the influence of multiple variables plays a vital role in addressing these issues. Employing real-time electricity dispatch based on precise predictions of solar irradiance can enhance energy efficiency, stabilize the power supply, and reduce the costs of electricity [2].
In recent years, numerous models for predicting solar radiation have emerged, including physical models [3], statistical models [4], machine learning models, deep learning models [5], and hybrid models [6,7]. Physical models use solar radiation transfer equations and atmospheric principles to forecast solar radiation with meteorological data. For example, Angstrom [8] estimated total solar radiation at specific locations by analyzing the relationship between global solar radiation and sunshine duration. Meanwhile, Whillier [9] assumed constant atmospheric transmittance throughout the day and estimated hourly radiation values from daily radiation values through data analysis and theoretical derivation. However, these physical models encounter computational difficulties and have limited accuracy under diverse climatic conditions. In contrast, statistical models like ARIMA [10], SARIMA [11], VAR, and VARMA [12] use historical solar radiation data for forecasting. They handle time series data effectively, capturing trends and periodicity, as evidenced in previous work by Belmahdi et al. [13] and Shadab et al. [14]. The accuracy of these statistical models heavily depends on dataset quality and parameter selection, often overlooking the influence of external factors on solar radiation.
With the growing volume of data and advancements in machine learning, researchers have increasingly used neural network models to predict solar radiation. These methods mainly include both traditional machine learning and deep learning approaches. Among conventional machine learning techniques, commonly used methods include the support vector machine (SVM) [15,16], decision tree (DT) [17,18], artificial neural network (ANN) [19,20], and random forest (RF) [21,22]. However, these individual machine learning models often struggle to capture the complex patterns and nonlinear relationships in solar radiation time series data. This leads to suboptimal information utilization, incorrect feature selection, and significant prediction biases. To overcome these issues, researchers have combined optimization algorithms with machine learning models [23,24] for sequence forecasting. For example, Natgunanathan et al. [25] utilized RF algorithm and digital twin technology to predict and optimize power generation within the renewable energy microgrid project at the solar farm. Even though using optimization algorithms improves the feature selection capacity of individual machine learning models, they still face challenges with high-dimensional, large-scale, and complex-structured data. Furthermore, deep learning models have gained attention for their ability to automatically extract features, strong generalization capabilities, and expertise in handling large-scale data. For example, Kazem et al. [26] predicted the current and power of the grid-connected photovoltaic system (GCPV) using a full recurrent neural network (FRNN) combined with principal component analysis (PCA). In a similar vein, models like recurrent neural networks (RNN) [27,28], long short-term memory (LSTM) networks [29,30], transformer [31], and time convolutional networks (TCN) [32] are utilized in solar radiation time series research. Shekar et al. [33] conducted multi-step forecasting of solar radiation using the LSTM model. In a similar vein, Gao et al. [34] introduced the use of gated recurrent unit (GRU) for solar irradiance prediction. LSTM and GRU models, with their gating mechanisms and memory units, improve conventional RNNs’ ability to manage long sequences and dependencies. However, this complexity adds to computational demands, slowing down processing with intricate data such as solar radiation and possibly hindering the capture of continuous long-term trends.
Hybrid Models typically combine two or more different types of models [35,36]. The hybrid models primarily combine CNN, RNN, LSTM, and the attention mechanism, leveraging their unique advantages to enhance prediction accuracy. For example, Kumari et al. [37] demonstrated the effectiveness of an LSTM–CNN hybrid model for short-term global horizontal irradiance (GHI) prediction. Similarly, Zang et al. [38] introduced a cascaded CNN–LSTM structure for spatiotemporal correlation, using CNN for spatial feature extraction and LSTM for temporal dependencies. The cascaded structure improved understanding of spatial relationships and solar radiation time series data. To address the intermittency and instability of solar radiation, Gao et al. [39] utilized a combined CEEMDAN–CNN–LSTM model, achieving significant improvements in prediction performance, while the application of hybrid models is pivotal for solar radiation forecasting and shows promise, they also introduce complexity. Such complexity can result in slower processing speeds and challenges in capturing continuous long-term trends, especially with intricate data like solar irradiance.
Solar radiation’s dynamic interplay with meteorological variables results in unique traits. One notable characteristic is the overlapping of cycles caused by seasonal oscillations. This complexity imposes higher requirements on prediction models. Specifically, these models need to deal with intricate interrelations among multiple temporal intervals or spatial positions. Moreover, these models need to distribute attention across different features at various levels, thereby enhancing their capability to understand and identify complex relationships and patterns. The Attention Mechanism, which allows for the adaptive focus on different parts of a sequence, has been successfully integrated with deep learning models such as RNN and LSTM. This combination has shown the value of attention-based models in addressing these complexities, providing improved accuracy and efficiency. For example, Qin et al. [40] developed an approach that blends the time attention mechanism with RNN to capture the long-term temporal dependencies in time series data. Similarly, Pan et al. [41] integrated the attention mechanism with LSTM for the prediction of photovoltaic power generation. The integration of attention mechanisms augmented the prowess of the LSTM model for long-term sequence prediction. Aslam et al. [42] proposed a method that combines attention with LSTM, bolstering the model’s performance by assigning disparate weights to the importance of features. Zhang et al. [43] employed attention mechanisms with GRU for feature extraction, enabling a comparison of data before and after faults and capturing changes in different fault locations. The integration of attention mechanisms has helped to overcome some limitations of RNN and LSTM in feature modeling. However, the design of LSTM and GRU with gate structures introduces a new challenge. During sequence data processing, this model structure can lead to information loss. Consequently, critical information may be omitted, hindering the overall performance.
To address the aforementioned issues, this study propose a deep-learning-based solar radiation forecasting model, integrating three key components: a residual attention time convolution block (RACB), a dual-path information fusion module (DIFM), and a twin self-attention module (TSAM). Specifically, time series data typically contain multiple features, each with a potential impact on target value prediction. Although a correlation analysis is performed on the input data and some low-correlation features are removed, the existing feature selection method still falls short in accurately determining the significance of each feature. To tackle this issue, a feature attention is incorporated into RACB, enabling the network to adaptively discern the importance of each channel during the feature extraction process. By allocating different weights to different channels, the model effectively concentrates on the features most pertinent to the prediction target. Second, the DIFM comprises two parallel time series convolutional networks: the local feature extraction network (LFEN) and the dilated feature extraction network (DFEN). The LFEN is employed to extract local features from the time series, while the DFEN captures a wider context information by expanding the receptive field. By integrating features from both paths, the model is designed to boost its feature representation ability and strengthen its robustness against input data variations. Finally, seasonal fluctuations often lead to overlapping signals from different periods, complicating the model’s ability to extract effective feature representations. To effectively manage this situation, the model needs to learn long-term dependencies in the data. To address this issue, the TSAM is proposed, which encompasses channel self-attention and sequence self-attention. This module aims to globally model the features produced by the DIFM from both a sequence and a channel perspective, thereby enhancing the predictive capabilities of the model.
In summary, the contributions of this study encompass the following three aspects:
  • A novel solar radiation forecasting model is proposed that innovatively incorporates a residual attention time convolution block (RACB). This design enables the model to selectively enhance significant features for irradiance prediction while diminishing the importance of irrelevant ones;
  • A dual-path information fusion module (DIFM) is proposed, adeptly integrating both local features and broader contextual information. Through the consolidation of features from distinct scales, the module enhances the model’s representational capacity, thereby bolstering its robustness against variations in input data;
  • A twin self-attention module (TSAM) is designed. By modeling long-distance dependencies in channels and sequences dimensions, the predictive capability of the model is improved. Experimental results on several public datasets demonstrate the effectiveness of the methods proposed in this study.
The remainder of this study is structured as follows: Section 2 describes the detailed structure of the model. In Section 3, the performance of the proposed model is evaluated through experiments on several publicly available datasets. Finally, in Section 4, a summary and discussion of the entire study is provided.

2. Methodology

2.1. Overview of the Proposed Method

To tackle the technical challenges in solar irradiance forecasting, this study proposes a method based on dual-path information fusion and twin self-attention for predicting solar irradiance. As depicted in Figure 1, the proposed method incorporates three key components: (1) the residual attention temporal convolution block (RACB), characterized by incorporating feature attention into the temporal convolution block, aimed at adaptively learning important features; (2) the dual-path information fusion module (DIFM) is designed to augment the model’s feature representation capacity by synergistically merging local features from distinct branches with wider contextual information; and (3) the twin self-attention module (TSAM), specifically formulated for global modeling, with an aim to forge long-distance dependencies across both channel and sequence dimensions, thereby contributing to enhanced prediction accuracy of the model.

2.2. Residual Attention Temporal Convolution Block

When carrying out time series prediction tasks, the importance of different features varies. These features impact the prediction results to different degrees, even though preliminary feature screening is conducted before inputting them into the network. However, the manually selected features still struggle to effectively balance the importance of each feature. To enhance the ability of convolutional networks to handle time series modeling and forecasting with a large number of features, this study designs a RACB. This module introduces a feature attention mechanism that can adaptively learn and understand the importance of each feature.
Figure 2 presents the RACB, with the kernel size and dilation factor represented by k and d, respectively. Consider an input time series X R C × N , where C represents the number of features and N represents the sequence length. This model applies the RACB to extract features from X, which is defined as follows:
F = σ F a t t n ( F ( X ) ) + F 1 × 1 ( X )
where X = x 1 ( i ) , x 2 ( i ) , , x N ( i ) , i { 1 , 2 , , C } , F ( · ) signifies two 1D convolution layers, interspersed with WeightNorm, ReLU activation, and Dropout operations. σ denotes the ReLU activation function, F 1 × 1 ( · ) refers to the 1 × 1 convolution operation, while F a t t n ( · ) denotes the feature attention. Formally, let us assume that F c o n v R C × N represents the features output by the convolution layer F ( · ) . The global max pooling function F g m p ( · ) and the global temporal pooling function F g t p ( · ) are employed to compress F c o n v into two low-rank tensors: F g t p F c o n v R C × 1 and F g m p F c o n v R C × 1 , respectively. Subsequently, a fully connected layer F F C α , which shares weights, is used to reduce dimensionality and refine useful information. By performing an addition operation, a tensor ϕ R C / 2 × 1 is obtained, representing the global information of the original input F c o n v . This can be formulated as follows:
ϕ = F F C a F g m p F c o n v + F F C a F g t p F c o n v
Next, a fully connected layer, denoted as F F C β , is employed to calculate the low-rank tensor η R C × 1 , which is utilized to adjust the weights of the input features. This is defined as follows:
η = F s f m F F C β ( ϕ )
where F s f m ( · ) represents the Softmax operation. The features, denoted as F c o n v , are readjusted by η , as follows:
F ˜ c o n v = η F c o n v
where ⊙ denotes the multiplication and F ˜ c o n v R C × N . In this way, the network is designed to adaptively focus on the features most informative for predicting solar irradiance, while suppressing those features that contribute less to the prediction.

2.3. Dual-Path Information Fusion Module

To capture both short-term and long-term patterns in time series, this study introduces a DIFM composed of a local feature extraction network (LFEN) and a dilated feature extraction network (DFEN). The LFEN is specifically designed to extract local features in the time series, while the DFEN is structured to capture a wider range of contextual information by increasing its receptive field. The features extracted by the LFEN are embedded into those extracted by the DFEN, achieving deep fusion of features at different scales. The structure of the LFEN, as depicted in Figure 1, comprises three cascaded RACBs. For a given input sequence X R C × N , the convolution kernel size k for each layer is fixed at 3, and the dilation factor d is fixed at 1. The entire process can be defined as follows:
F L F = F L F E N ( X )
where F L F E N ( · ) denotes the LFEN. The structure of the DFEN is shown in Figure 1. It consists of three RACBs. In the RACB, dilated convolution is introduced to expand the receptive field of the convolution kernel, to better capture a wider range of contextual information in the sequence. The convolution kernel size k for each block is fixed at 2, and the dilation factors d are sequentially 1, 2, and 4. The entire process can be defined as follows:
F D F = F D F E N ( X )
where F D F E N ( · ) denotes the DFEN. Subsequently, the features from both paths are fused as defined below:
F f u s e = F 1 × 1 F L F , F D F
where [ · , · ] represents the concatenation operation, and F 1 × 1 ( · ) denotes a 1 × 1 convolution operation.

2.4. Twin Self-Attention Module

Through the DIFM, the multi-scale fusion feature F f u s e is obtained. To further enhance the global context information, the TSAM is employed to model the feature F f u s e , which consists of two independent self-attention mechanisms: channel self-attention and sequence self-attention. Specifically, the channel self-attention is first introduced, as depicted in Figure 3a. For the input feature, to capture the global dependency on the channel dimension, 1 × 1 point-wise convolution and 3 × 3 depth-wise convolution are utilized to generate Q 1 = F D W C Q 1 F P W C Q 1 F f u s e , K 1 = F D W C K 1 F P W C K 1 F f u s e , and V 1 = F D W C V 1 F P W C V 1 F f u s e , which encapsulate rich local context information. F P W C ( · ) and F D W C ( · ) , respectively, represent point-wise convolution and depth-wise convolution. Then, a multiplication is performed on Q 1 and K 1 to generate the channel attention matrix A c R C × C . The entire calculation process for channel self-attention is defined as follows:
F ˜ c = F 1 × 1 Softmax Q 1 · K 1 T / α 1 · V 1
where F f u s e and F ˜ c are the input and output feature maps, respectively. Q 1 R C × N , K 1 T R N × C , V 1 R C × N . α 1 is a learnable parameter used to control the magnitude of the output between Q 1 and K 1 . To capture the global dependencies in the sequence dimension, sequence self-attention is introduced, as illustrated in Figure 3b. Another set of 1 × 1 point-wise convolution and 3 × 3 depth-wise convolution is used to generate Q 2 = F D W C Q 2 F P W C Q 2 F f u s e , K 2 = F D W C K 2 F P W C K 2 F f u s e , and V 2 = F D W C V 2 F P W C V 2 F f u s e . Then, a multiplication is performed on Q 2 and K 2 to generate the sequence attention matrix A s R N × N . The entire calculation process for sequence self-attention is defined as follows:
F ˜ s = F 1 × 1 V 2 · Softmax K 2 T · Q 2 / α 2
where F f u s e and F ˜ s are the input and output feature maps, respectively. Q 2 R C × N , K 2 T R N × C , V 2 R C × N . α 2 is a learnable parameter used to control the magnitude of the output between Q 2 and K 2 . Subsequently, the features F ˜ c and F ˜ s are fused and combined with the initial input F f u s e . This operation can be formalized as follows:
F ˜ = F 1 × 1 F ˜ c , F ˜ s + F f u s e
where [ · , · ] denotes the concatenation operation, and F 1 × 1 ( · ) signifies a 1 × 1 convolution operation. By implementing the aforementioned processes, long-term dependencies across various positions within the sequence are effectively captured.

2.5. Loss Function

In this work, the mean squared error (MSE) is chosen as the loss function to optimize the model. The MSE loss function measures the average of the squared differences between the predicted values of the model and the true target values. It serves as a common metric to quantify prediction errors. The specific calculation formula is as follows:
M S E = 1 n i = 1 n ( Y i Y ^ i ) 2
where Y i represents the i-th element in the target sequence, Y ^ i denotes the corresponding element in the prediction sequence generated by the neural network, and n denotes the length of the sequence.

3. Experiment

3.1. Datasets and Data Processing

Datasets: The study utilizes a dataset from the National Solar Radiation Database (NSRDB), available at https://nsrdb.nrel.gov/ (accessed on 5 February 2023). To ensure comprehensive experimentation, data from four geographically distinct locations were selected, specifically from Busan, Karnataka, Nevada, and Yunnan. Each of these datasets includes solar irradiance data recorded at 30-min intervals. Table 1 provides detailed information about these datasets, including their geographical longitude and latitude, climatic conditions, dataset size, and data resolution.
Data preprocessing: A normalization process was applied to the solar irradiance time series to transform diverse value ranges into a uniform 0–1 range. The formula for this normalization is:
f x i norm = f x i f x i min f x i max f x i min
where f x i min represents the minimum value in the data, f x i max denotes the maximum value, and f x i corresponds to the original value. Additionally, since the Global Horizontal Irradiance (GHI) is zero during the nighttime and does not require prediction, only data with GHI values greater than zero were retained. Moreover, to extract the features most closely related to solar irradiance from the dataset, the Pearson Correlation Coefficient (PCC) between solar irradiance and meteorological variables was calculated [44,45]. The value of the PCC lies between +1 and −1, where +1 signifies a perfect positive linear relationship between two variables, while −1 indicates a perfect negative linear relationship. Table 2 presents our results from the correlation coefficient calculation, reflecting the degree of correlation between each variable and GHI. The formula for calculating the PCC is as follows:
PCC X , G = E X μ X G μ G σ X σ G
where PCC X , G represents the Pearson correlation coefficient between variables X and G. The term E stands for expectation, whereas μ X and μ G , respectively, denote the mean values of variables X and G. Additionally, σ X and σ G represent the standard deviations of variables X and G, respectively. The PCC was averaged across all four datasets [46], and features with an average value greater than 0.1 were considered as input features, as depicted in Figure 4. Ultimately, 13 features were selected as input: GHI, Temperature, Dew Point, DHI, DNI, Wind Speed, Clearsky DHI, Clearsky DNI, Clearsky GHI, Cloud Type, Solar Zenith Angle, Precipitable Water, and Relative Humidity.

3.2. Evaluation Metrics and Experimental Setup

Evaluation metrics: To provide a quantitative comparison of the proposed model, this study utilizes information from the first ’t’ moments to predict the GHI value at the ’t + n’ moment. Several metrics are employed, including the Root Mean Square Error (RMSE), Mean Absolute Error (MAE), Mean Absolute Percentage Error (MAPE), and R-squared ( R 2 ), to facilitate an accurate analysis of the predictive outcomes.
The MAE represents the mean of the absolute differences between the predicted and actual values. The formula is:
MAE = 1 m i = 1 m G i G ^ i
The RMSE is the square root of the mean of the squared differences between the predicted and actual values. A smaller RMSE value indicates a closer approximation of the predicted value to the actual one. The formula is:
RMSE = 1 m i = 1 m G i G ^ i 2
The MAPE is the mean of the absolute percentage errors between the predicted and actual values. A lower MAPE value indicates a closer approximation of the predicted value to the actual one. The formula is:
MAPE = 100 % m i = 1 m G i G ^ i G i
The R 2 quantifies the proportion of the variance in the dependent variable that can be predicted from the independent variable(s). A value closer to 1 signifies a better fit of the model to the observed data. The formula is:
R 2 = 1 i = 1 m G i G ^ i 2 i = 1 m G i G i ¯ 2
where G i represents the actual value of solar radiation irradiance, G ^ i denotes the predicted value of solar radiation irradiance, G i ¯ is the mean of G i , and m represents the total number of samples in the dataset.
Experimental setup: The algorithm in this study was developed using the PyTorch 1.12.1 framework and Python 3.8 on a server configured with a 12th Gen Intel (R) Core (TM) i9-12900K, an NVIDIA GeForce RTX 3090 GPU, and running the Ubuntu 20.04 LTS operating system. During the training process, the model parameters were optimized using the AdamW optimizer with a batch size of 256 and a learning rate of 0.001. The data was divided into training, validation, and test sets in a 6:2:2 ratio, and the model is trained for a total of 100 epochs. To identify the optimal length of the input sequence, various input sequence lengths were tested on the Nevada dataset to predict one-step GHI values. The RMSE values of each model served as the criterion for comparison. As shown in Figure 5, the figure depicts the RMSE values of one-step predictions of each model at different input lengths. Upon analysis of the results, it was determined that the prediction outcome was most optimal when the input length was 57. Consequently, the length of the input sequence was finalized as 57 for this study.

3.3. Result and Analysis

To evaluate the performance of the proposed model, systematic experiments were executed across distinct prediction intervals: 1, 2, 4, 6, 8, and 10 steps. Comprehensive insights can be derived from the experimental findings. The outcomes for the GHI time series prediction on the Busan dataset are delineated in Table 3. It is evident that the proposed approach consistently surpasses other models in predictive accuracy across varied intervals and evaluation criteria. For a one-step-ahead prediction, the model registered scores of 2.192 W / m 2 , 1.517 W / m 2 , 0.022, and 0.979 for RMSE, MAE, MAPE, and R 2 , respectively. Noteworthy reductions in RMSE were observed when juxtaposed with models such as TCN [47], LSTM [48], LSTM-Attention [49], CNN-LSTM [50], and Transformer [51]. A discernible trend is that as the prediction horizon extends from 1 to 10 steps, there is a commensurate decline in the performance metrics for all models, underscoring the inherent challenge of forecasting over longer intervals. The Karnataka dataset’s results, tabulated in Table 4, further underscore the consistent superiority of the proposed model. In the context of a one-step prediction, the model’s performance metrics were impressive, and when compared against benchmarks like TCN, LSTM, and others, it exhibited noteworthy reductions in RMSE. For example, in the one-step-ahead prediction, the method yielded scores of 2.195 W / m 2 , 1.503 W / m 2 , 0.029, and 0.992 on the RMSE, MAE, MAPE, and R 2 metrics, respectively. Compared with TCN, LSTM, LSTM-Attention, CNN-LSTM, and Transformer, the model reduced the RMSE by 0.585 W / m 2 , 0.518 W / m 2 , 0.960 W / m 2 , 3.283 W / m 2 , and 5.880 W / m 2 , respectively. The overarching observation from the table is the amplification of prediction errors as the forecasting range elongates, highlighting the intricacies associated with GHI forecasting. The results for the Nevada and Yunnan datasets are cataloged in Table 5 and Table 6. These findings further corroborate the proposed model’s exemplary predictive prowess over diverse horizons and metrics. Collectively, the empirical evidence affirms the model’s unparalleled capability in forecasting solar irradiance across varied temporal granularities, consistently outshining competitors across diverse datasets.
Figure 6 and Figure 7 show the one-step prediction results of various models across four datasets, represented as line graphs and scatter plots. Specifically, Figure 6 presents the prediction value curves for the four datasets. In contrast, Figure 7 displays the corresponding scatter plots of these prediction values. A detailed analysis of these charts yields several key observations: In the line graphs, the proposed model closely aligns with the actual irradiance curves, especially at inflection points where significant irradiance fluctuations occur. This highlights the strong fitting capability of the model. In the scatter plots, the predicted values from the model predominantly cluster around the diagonal line, signifying superior tracking performance. From the perspective of the datasets, the method effectively adapts to the unique characteristics inherent in each. The model maintains consistent predictive performance across different datasets, as evidenced in both line graphs and scatter plots. The prediction value curves showcase the efficiency of this model in forecasting solar irradiance fluctuations, and the scatter plots affirm its precision in capturing the linear trends within the data.
Figure 8 illustrates the predictive performance of various models across four different datasets. The RMSE line graph reveals that the proposed model consistently achieves the lowest RMSE values across all four datasets with minimal fluctuations. This indicates its superior error reduction performance. The MAPE bar chart shows that the proposed model records the smallest MAPE value for the Busan dataset and the largest for the Nevada dataset. Nonetheless, it consistently outperforms other models across all datasets. The R 2 graph demonstrates that the proposed model consistently records R 2 values near 1 across all datasets, surpassing other models in each instance. These findings collectively affirm the superior performance of the proposed model. Figure 9 illustrates the relationship between the model’s evaluation metrics and the change in prediction time step length. Despite the increase in MAE value as the prediction step length increases, the radar chart for MAE shows that the magnitude of change is relatively small, and the proposed model’s overall performance surpasses that of other models. Even in the Nevada dataset, where performance tends to be lower, the MAE value of the proposed model remains superior to those of other models. The MAPE bar chart clearly demonstrates that the proposed model maintains lower MAPE values across all four datasets, and achieves lower MSE and RMSE values in most instances. The lollipop chart for R 2 indicates that the proposed model consistently yields R 2 values closest to 1 across all datasets, suggesting a superior fitting capability compared to other models. In general, as the prediction time step length increases, the R 2 values of all six models decrease, and MSE, RMSE, and MAE values increase, indicating a decline in predictive performance. Nonetheless, the proposed model excels in most evaluation metrics, demonstrating consistently lower prediction errors, higher goodness-of-fit, and smaller percentage errors across various prediction time step lengths.

3.4. Further Exploration

To thoroughly evaluate the performance of the prediction model in a multi-step sequence prediction, analysis was conducted on a randomly selected fourth step. In the Busan dataset, the model delivered an RMSE of 3.144 W / m 2 , an MAE of 2.144 W / m 2 , a MAPE of 0.031, and an R 2 of 0.956. When pitted against algorithms like TCN, LSTM, LSTM-Attention, CNN-LSTM, and Transformer, the proposed model demonstrated a reduction in RMSE by 0.512 W / m 2 , 0.545 W / m 2 , 0.699 W / m 2 , 1.832 W / m 2 , and 2.651 W / m 2 , respectively. For the Karnataka dataset, the proposed model achieved RMSE reductions of 0.922 W / m 2 , 1.317 W / m 2 , 1.131 W / m 2 , 5.591 W / m 2 , and 7.171 W / m 2 . Similarly, in the Nevada and Yunnan datasets, the model exhibited RMSE reductions of 1.020 W / m 2 , 0.764 W / m 2 , 0.840 W / m 2 , 1.962 W / m 2 , 6.782 W / m 2 , and 0.743 W / m 2 , 0.623 W / m 2 , 0.375 W / m 2 , 3.033 W / m 2 , and 6.698 W / m 2 , respectively. Considering other performance indicators, the proposed model stood out, registering lower MAE and MAPE scores and attaining higher R 2 values. The prediction curves and scatter plots of this model across the four datasets are demonstrated in Figure 10 and Figure 11. Figure 10 displays the irradiance curves for these datasets, emphasizing particular time intervals. It is evident that the prediction model aligns remarkably with the actual curve, highlighting its superior forecasting capabilities for solar irradiance fluctuations. When benchmarked against other models, the proposed model shines in its tracking performance. It is particularly noteworthy during significant shifts, illustrating robust regression abilities and a negligible variance from actual values. This emphasizes its adeptness in utilizing time features. The scatter plots, as shown in Figure 11, for all datasets depict a close-knit distribution of the model’s predicted values around the diagonal. In comparison to other models, the proposed model demonstrates a denser distribution, closely trailed by the TCN and LSTM models. This further confirms the model’s exceptional proficiency in multi-step sequence predictions.

3.5. Ablation Study

In this subsection, an ablation study on the Busan dataset was undertaken to validate the efficacy of the introduced techniques. The findings, tabulated in Table 7, feature optimal results emphasized in bold. The table employs specific nomenclature for clarity. “w/o LFEN” indicates the removal of the LFEN, “w/o DFEN” represents the absence of the DFEN, “w/o RACB” signifies the omission of the RACB, “w/o TSAM” means the TSAM was excluded, and the term “Ours” is employed to denote the full-fledged model. All experiments were designed around a one-step prediction scenario. Upon evaluating the results, the RMSE values for the different models were discerned. The full model exhibited an RMSE of 2.19 W / m 2 . In contrast, the “w/o LFEN”, “w/o DFEN”, “w/o RACB”, and “w/o TSAM” variants registered RMSEs of 2.540, 2.435, 2.313, and 2.467 W / m 2 , respectively. Corresponding MAE values were documented as 1.517, 1.793, 1.680, 1.635, and 1.761 W / m 2 . Concurrently, MAPE metrics stood at 0.022, 0.027, 0.025, 0.023, and 0.024, with R 2 values being 0.979, 0.972, 0.974, 0.976, and 0.973. A key observation from the one-step prediction outcomes is the RMSE’s sensitivity to the removal of any module. The variance ranged between 0.122 W / m 2 and 0.349 W / m 2 . Notably, LFEN’s omission resulted in the most pronounced increase, highlighting its pivotal role in discerning local patterns within time-series datasets. Furthermore, in multi-step forecasts, the exclusion of TSAM consistently inflated the RMSE metric, underlining the significance of this module in enhancing the model’s predictive prowess.
Figure 12 depicts the fluctuation patterns of each model’s metric indicators as a function of prediction step length. The abscissa signifies the prediction step length, whereas the ordinate conveys the metric values. It is evident from the trends that, across different time scales, excluding any individual module leads to a surge in the RMSE, MAE, and MAPE values, while concurrently causing a decline in the R 2 value. Specifically, for one-step predictions, the omission of the LFEN escalates the MSE, MAE, and MAPE to their peak values, simultaneously dragging the R 2 to its lowest. Conversely, in the context of multi-step predictions, the absence of the TSAM results in maximal RMSE, MAE, and MAPE values, with the R 2 reaching its nadir.

4. Conclusions and Discussion

This study aims to develop a multi-step sequence prediction model for solar irradiance forecasting. To achieve this, a deep learning-based dual-path information fusion and twin attention-driven solar irradiance forecasting model is proposed. This model comprises three components: the RACB, the DIFM, and the TSAM. The RACB is designed to enable the network to adaptively learn important features while suppressing irrelevant ones. Following that, the DIFM is introduced to reinforce the model’s robustness against input data variations and integrate multi-scale features. Finally, the TSAM is introduced to extract long-term temporal dependencies within the sequence, thereby enhancing multi-step prediction capabilities.
The experimental results indicate that the model proposed in this study can accurately predict irradiance data. In both one-step and multi-step predictions, it significantly outperforms other models. Furthermore, when compared to other models such as TCN, LSTM, LSTM-Attention, CNN-LSTM, and Transformer, the proposed model exhibits superior and more robust performance across datasets from four different regions. For instance, for one-step predictions, the model reports RMSE values of 2.192 W / m 2 , 2.195 W / m 2 , 2.508 W / m 2 , and 2.238 W / m 2 across the four datasets. When compared to algorithms such as TCN, LSTM, LSTM-Attention, CNN-LSTM, and Transformer, the proposed model achieves average RMSE reductions of 0.381 W / m 2 , 0.212 W / m 2 , 0.256 W / m 2 , 1.052 W / m 2 , and 1.855 W / m 2 , underscoring its consistent performance across various datasets. For other multi-step predictions, the proposed model also surpasses other models, with the performance gap becoming more noticeable as prediction steps increase. Additionally, scatter plots and curve diagrams provide a more visual representation of the precision of the proposed method. Lastly, ablation studies were conducted to validate the effectiveness of the DFEN, the LFEN, the RACB, and TSAM. The results highlight that, when removing these components, there is a substantial decline in performance compared to the full model. For example, when LFEN is removed, RMSE increases from 2.19 W / m 2 to 2.540 W / m 2 .
The prediction methods employed in this study have been summarized, and several conclusions are drawn. Firstly, solar irradiance is influenced by a multitude of factors, and the RACB effectively extracts essential features from this complex interplay. Secondly, these influencing factors exhibit distinct characteristics across different time scales. The DIFM captures these features across multiple scales, enhancing the model’s robustness and enabling precise predictions. Lastly, for extended-range forecasting, the TSAM addresses long-term dependencies in the data sequence, facilitating more accurate forecasting of future irradiance data. In contrast, other models like TCN, LSTM, and RNN do not consider the impacts of diverse factors and the importance of features across various scales. The model proposed in this study provides valuable support for photovoltaic power generation systems, a pivotal step toward the development of intelligent grid systems.
While the model excels in forecasting solar irradiance over multiple steps, there is room for refining its precision. Its primary reliance on past irradiance data might pose challenges in anticipating abrupt irradiance shifts. For even greater accuracy, future studies might consider broadening data sources and optimizing the neural network’s design. Addressing the intricate dynamics of irradiance patterns could demand meticulous feature engineering and model tweaks. Given the notable weather variations across different regions and seasons, it seems prudent to design models tailored to specific locales and times of the year. Such bespoke modeling, factoring in the unique climatic nuances of each area and season, stands to boost predictive accuracy.

Author Contributions

Conceptualization, Y.Y. and Z.T.; methodology, Y.Y. and Z.L.; software, Y.Y. and Z.L.; validation, Z.T. and J.H.; formal analysis, Z.L.; investigation, Y.Y. and Z.T.; data curation, Y.Y.; writing—original draft preparation, Y.Y., Z.L. and Z.T.; writing—review and editing, Y.Y., Z.L. and Z.T.; visualization, Y.Y.; supervision, Z.T. and Y.Z.; project administration, Y.Y., X.S. and Z.T.; funding acquisition, J.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research is funded by the National Natural Science Foundation of China, grant number 82160347, and the National Energy Science and Technology Project, grant number CSIEKJ230700104.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The source code is available at https://github.com/89Mansions/DFTD, accessed on 21 August 2023.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bett, P.E.; Thornton, H.E. The climatological relationships between wind and solar energy supply in Britain. Renew. Energy 2016, 87, 96–110. [Google Scholar] [CrossRef]
  2. Kaur, A.; Nonnenmacher, L.; Pedro, H.T.; Coimbra, C.F. Benefits of solar forecasting for energy imbalance markets. Renew. Energy 2016, 86, 819–830. [Google Scholar] [CrossRef]
  3. Dolara, A.; Leva, S.; Manzolini, G. Comparison of different physical models for PV power output prediction. Sol. Energy 2015, 119, 83–99. [Google Scholar] [CrossRef]
  4. Prema, V.; Rao, K.U. Development of statistical time series models for solar power prediction. Renew. Energy 2015, 83, 100–109. [Google Scholar] [CrossRef]
  5. Kumari, P.; Toshniwal, D. Deep learning models for solar irradiance forecasting: A comprehensive review. J. Clean. Prod. 2021, 318, 128566. [Google Scholar] [CrossRef]
  6. Diagne, M.; David, M.; Lauret, P.; Boland, J.; Schmutz, N. Review of solar irradiance forecasting methods and a proposition for small-scale insular grids. Renew. Sustain. Energy Rev. 2013, 27, 65–76. [Google Scholar] [CrossRef]
  7. Liu, J.; Huang, X.; Li, Q.; Chen, Z.; Liu, G.; Tai, Y. Hourly stepwise forecasting for solar irradiance using integrated hybrid models CNN-LSTM-MLP combined with error correction and VMD. Energy Convers. Manag. 2023, 280, 116804. [Google Scholar] [CrossRef]
  8. Angstrom, A. Solar and terrestrial radiation. Q. J. R. Meteorol. Soc. 1924, 50, 121–126. [Google Scholar] [CrossRef]
  9. Whillier, A. The determination of hourly values of total solar radiation from daily summations. Arch. Für Meteorol. Geophys. Und Bioklimatol. Ser. B 1956, 7, 197–204. [Google Scholar] [CrossRef]
  10. Atique, S.; Noureen, S.; Roy, V.; Subburaj, V.; Bayne, S.; Macfie, J. Forecasting of total daily solar energy generation using ARIMA: A case study. In Proceedings of the 2019 IEEE 9th Annual Computing and Communication Workshop and Conference (CCWC), Las Vegas, NV, USA, 7–9 January 2019; Volume 7, pp. 114–119. [Google Scholar] [CrossRef]
  11. Yatiyana, E.; Rajakaruna, S.; Ghosh, A. Wind speed and direction forecasting for wind power generation using ARIMA model. In Proceedings of the 2017 Australasian Universities Power Engineering Conference (AUPEC), Melbourne, Australia, 19–22 November 2017; Volume 219, pp. 1–6. [Google Scholar] [CrossRef]
  12. Gülerce, M.; Ünal, G. Electricity price forecasting using multiple wavelet coherence method: Comparison of ARMA versus VARMA. Int. J. Financ. Eng. 2018, 5, 1850004. [Google Scholar] [CrossRef]
  13. Belmahdi, B.; Louzazni, M.; Bouardi, A.E. One month-ahead forecasting of mean daily global solar radiation using time series models. Optik 2020, 219, 165207. [Google Scholar] [CrossRef]
  14. Shadab, A.; Ahmad, S.; Said, S. Spatial forecasting of solar radiation using ARIMA model. Remote Sens. Appl. Soc. Environ. 2020, 20, 100427. [Google Scholar] [CrossRef]
  15. Zeng, J.; Qiao, W. Short-term solar power prediction using a support vector machine. Renew. Energy 2013, 52, 118–127. [Google Scholar] [CrossRef]
  16. Belaid, S.; Mellit, A. Prediction of daily and mean monthly global solar radiation using support vector machine in an arid climate. Energy Convers. Manag. 2016, 118, 105–118. [Google Scholar] [CrossRef]
  17. Ma, R.; Boubrahimi, S.F.; Hamdi, S.M.; Angryk, R.A. Solar flare prediction using multivariate time series decision trees. In Proceedings of the 2017 IEEE International Conference on Big Data, 11–14 December 2017; pp. 2569–2578. [Google Scholar] [CrossRef]
  18. Rahul.; Gupta, A.; Bansal, A.; Roy, K. Solar Energy Prediction using Decision Tree Regressor. In Proceedings of the 2021 5th International Conference on Intelligent Computing and Control Systems (ICICCS), Madurai, India, 6–8 May 2021; pp. 489–495. [Google Scholar] [CrossRef]
  19. Mittal, M.; Bora, B.; Saxena, S.; Gaur, A.M. Performance prediction of PV module using electrical equivalent model and artificial neural network. Sol. Energy 2015, 176, 104–117. [Google Scholar] [CrossRef]
  20. Kumar, S.; Kaur, T. Development of ANN Based Model for Solar Potential Assessment Using Various Meteorological Parameters. Energy Procedia 2016, 90, 587–592. [Google Scholar] [CrossRef]
  21. Liu, J.; Cao, M.; Bai, D.; Zhang, R. Solar radiation prediction based on random forest of feature-extraction. IOP Conf. Ser. Mater. Sci. Eng. 2019, 658, 012006. [Google Scholar] [CrossRef]
  22. Babar, B.; Luppino, L.T.; Boström, T.; Anfinsen, S.N. Random forest regression for improved mapping of solar irradiance at high latitudes. Sol. Energy 2020, 198, 81–92. [Google Scholar] [CrossRef]
  23. Bae, K.Y.; Jang, H.S.; Sung, D.K. Hourly Solar Irradiance Prediction Based on Support Vector Machine and Its Error Analysis. IEEE Trans. Power Syst. 2017, 32, 935–945. [Google Scholar] [CrossRef]
  24. Prasad, R.; Ali, M.; Kwan, P.; Khan, H. Designing a multi-stage multivariate empirical mode decomposition coupled with ant colony optimization and random forest model to forecast monthly solar radiation. Appl. Energy 2019, 236, 778–792. [Google Scholar] [CrossRef]
  25. Natgunanathan, I.; Mak-Hau, V.; Rajasegarar, S.; Anwar, A. Deakin microgrid digital twin and analysis of AI models for power generation prediction. Energy Convers. Manag. X 2023, 18, 100370. [Google Scholar] [CrossRef]
  26. Kazem, H.A.; Yousif, J.H.; Chaichan, M.T.; Al-Waeli, A.H.; Sopian, K. Long-term power forecasting using FRNN and PCA models for calculating output parameters in solar photovoltaic generation. Heliyon 2022, 8, e08803. [Google Scholar] [CrossRef]
  27. Alzahrani, A.; Shamsi, P.; Dagli, C.; Ferdowsi, M. Solar Irradiance Forecasting Using Deep Neural Networks. Procedia Comput. Sci. 2017, 114, 304–313. [Google Scholar] [CrossRef]
  28. Pang, Z.; Niu, F.; O’Neill, Z. Solar radiation prediction using recurrent neural network and artificial neural network: A case study with comparisons. Renew. Energy 2020, 156, 279–289. [Google Scholar] [CrossRef]
  29. Obiora, C.N.; Ali, A.; Hasan, A.N. Forecasting Hourly Solar Irradiance Using Long Short-Term Memory (LSTM) Network. In Proceedings of the 2020 11th International Renewable Energy Congress (IREC), Hammamet, Tunisia, 29–31 October 2020; pp. 1–6. [Google Scholar] [CrossRef]
  30. Hossain, M.S.; Mahmood, H. Short-Term Photovoltaic Power Forecasting Using an LSTM Neural Network and Synthetic Weather Forecast. IEEE Access 2020, 8, 172524–172533. [Google Scholar] [CrossRef]
  31. Zhang, L.; Wilson, R.; Sumner, M.; Wu, Y. Advanced multimodal fusion method for very short-term solar irradiance forecasting using sky images and meteorological data: A gate and transformer mechanism approach. Renew. Energy 2023, 0960-1481, 118952. [Google Scholar] [CrossRef]
  32. Song, Z.; Brown, L.E. Multi-dimensional Evaluation of Temporal Neural Networks on Solar Irradiance Forecasting. In Proceedings of the 2019 IEEE Innovative Smart Grid Technologies—Asia (ISGT Asia), Chengdu, China, 21–24 May 2019; pp. 4192–4197. [Google Scholar] [CrossRef]
  33. Shekar, D.D.; Hiremath, A.C.; Keshava, A.; Vinatha, U. Solar Irradiance forecasting using Recurrent Neural Networks. In Proceedings of the 2022 IEEE Region 10 Symposium (TENSYMP), Mumbai, India, 1–3 July 2022; pp. 1–6. [Google Scholar] [CrossRef]
  34. Gao, B.; Huang, X.; Shi, J.; Tai, Y.; Xiao, R. Predicting day-ahead solar irradiance through gated recurrent unit using weather forecasting data. J. Renew. Sustain. Energy 2019, 11, 4. [Google Scholar] [CrossRef]
  35. Huang, X.; Li, Q.; Tai, Y.; Chen, Z.; Zhang, J.; Shi, J.; Gao, B.; Liu, W. Hybrid deep neural model for hourly solar irradiance forecasting. Renew. Energy 2021, 171, 1041–1060. [Google Scholar] [CrossRef]
  36. Guermoui, M.; Melgani, F.; Gairaa, K.; Mekhalfi, M.L. A comprehensive review of hybrid models for solar radiation forecasting. J. Clean. Prod. 2020, 258, 120357. [Google Scholar] [CrossRef]
  37. Kumari, P.; Toshniwal, D. Long short term memory–convolutional neural network based deep hybrid approach for solar irradiance forecasting. Appl. Energy 2021, 295, 117061. [Google Scholar] [CrossRef]
  38. Zang, H.; Liu, L.; Sun, L.; Cheng, L.; Wei, Z.; Sun, G. Short-term global horizontal irradiance forecasting based on a hybrid CNN-LSTM model with spatiotemporal correlations. Renew. Energy 2020, 160, 26–41. [Google Scholar] [CrossRef]
  39. Gao, B.; Huang, X.; Shi, J.; Tai, Y.; Zhang, J. Hourly forecasting of solar irradiance based on CEEMDAN and multi-strategy CNN-LSTM neural networks. Renew. Energy 2020, 162, 1665–1683. [Google Scholar] [CrossRef]
  40. Qin, Y.; Song, D.; Chen, H.; Cheng, W.; Jiang, G.; Cottrell, G. A Dual-Stage Attention-Based Recurrent Neural Network for Time Series Prediction. arXiv 2017, arXiv:1704.02971. [Google Scholar] [CrossRef]
  41. Pan, C.; Tan, J.; Feng, D.; Li, Y. Very Short-Term Solar Generation Forecasting Based on LSTM with Temporal Attention Mechanism. In Proceedings of the 2019 IEEE 5th International Conference on Computer and Communications (ICCC), Chengdu, China, 6–9 December 2019; pp. 267–271. [Google Scholar] [CrossRef]
  42. Aslam, M.; Lee, S.J.; Khang, S.H.; Hong, S. Two-Stage Attention Over LSTM With Bayesian Optimization for Day-Ahead Solar Power Forecasting. IEEE Access 2021, 9, 107387–107398. [Google Scholar] [CrossRef]
  43. Zhang, F.; Liu, Q.; Liu, Y.; Tong, N.; Chen, S.; Zhang, C. Novel Fault Location Method for Power Systems Based on Attention Mechanism and Double Structure GRU Neural Network. IEEE Access 2020, 8, 75237–75248. [Google Scholar] [CrossRef]
  44. Castangia, M.; Aliberti, A.; Bottaccioli, L.; Macii, E.; Patti, E. A compound of feature selection techniques to improve solar radiation forecasting. Expert Syst. Appl. 2021, 178, 114979. [Google Scholar] [CrossRef]
  45. Lyu, C.; Eftekharnejad, S.; Basumallik, S.; Xu, C. Dynamic Feature Selection for Solar Irradiance Forecasting Based on Deep Reinforcement Learning. IEEE Trans. Ind. Appl. 2023, 59, 533–543. [Google Scholar] [CrossRef]
  46. Jebli, I.; Belouadha, F.Z.; Kabbaj, M.I.; Tilioua, A. Prediction of solar energy guided by pearson correlation using machine learning. Energy 2021, 224, 120109. [Google Scholar] [CrossRef]
  47. Bai, S.; Kolter, J.Z.; Koltun, V. An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv 2018, arXiv:1803.01271. [Google Scholar] [CrossRef]
  48. Graves, A. Long short-term memory. In Supervised Sequence Labelling with Recurrent Neural Networks; Springer: Berlin/Heidelberg, Germany, 2012; pp. 37–45. [Google Scholar] [CrossRef]
  49. Liu, J.; Gao, H.; Yan, Q.; Bai, B.; Wang, X. Short-term Load Forecasting Method Based on LSTM-Attention Mechanism. In Proceedings of the 2021 IEEE Sustainable Power and Energy Conference (iSPEC), Nanjing, China, 23–25 December 2021; pp. 3837–3842. [Google Scholar] [CrossRef]
  50. Kim, T.Y.; Cho, S.B. Predicting residential energy consumption using CNN-LSTM neural networks. Energy 2019, 182, 72–81. [Google Scholar] [CrossRef]
  51. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar] [CrossRef]
Figure 1. The overall structure of the proposed method.
Figure 1. The overall structure of the proposed method.
Sensors 23 07469 g001
Figure 2. The proposed residual attention temporal convolution block (RACB), where k denotes the kernel size and d denotes the dilation factor.
Figure 2. The proposed residual attention temporal convolution block (RACB), where k denotes the kernel size and d denotes the dilation factor.
Sensors 23 07469 g002
Figure 3. Architecture of the twin self-attention module (TSAM).
Figure 3. Architecture of the twin self-attention module (TSAM).
Sensors 23 07469 g003
Figure 4. Pearson correlation value range plot between solar radiation and meteorological information.
Figure 4. Pearson correlation value range plot between solar radiation and meteorological information.
Sensors 23 07469 g004
Figure 5. RMSE of various models for different input lengths.
Figure 5. RMSE of various models for different input lengths.
Sensors 23 07469 g005
Figure 6. Prediction curve graphs of different prediction models one step ahead for four different datasets.
Figure 6. Prediction curve graphs of different prediction models one step ahead for four different datasets.
Sensors 23 07469 g006
Figure 7. Scatter plots of predicted values and actual values one step ahead for different prediction models on four different datasets.
Figure 7. Scatter plots of predicted values and actual values one step ahead for different prediction models on four different datasets.
Sensors 23 07469 g007
Figure 8. Performance evaluation metrics of various prediction models one step ahead for four different datasets.
Figure 8. Performance evaluation metrics of various prediction models one step ahead for four different datasets.
Sensors 23 07469 g008
Figure 9. Performance evaluation metrics of different prediction models for multiple steps on four different datasets.
Figure 9. Performance evaluation metrics of different prediction models for multiple steps on four different datasets.
Sensors 23 07469 g009
Figure 10. Prediction curve graphs of different prediction models four steps ahead for four different datasets.
Figure 10. Prediction curve graphs of different prediction models four steps ahead for four different datasets.
Sensors 23 07469 g010
Figure 11. Scatter plots of predicted values and actual values four steps ahead for different prediction models on four different datasets.
Figure 11. Scatter plots of predicted values and actual values four steps ahead for different prediction models on four different datasets.
Sensors 23 07469 g011
Figure 12. Performance evaluation metrics for ablation experiments.
Figure 12. Performance evaluation metrics for ablation experiments.
Sensors 23 07469 g012
Table 1. General information of the four datasets.
Table 1. General information of the four datasets.
LocationLatitude ( )Longitude ( )Size (Years)Climate Type (Koppen Classification)Data Resolution
Busan, South Korea35.2101129.02075Temperate MonsoonHalf-Hourly
Karnataka, India13.074577.56535Tropical MonsoonHalf-Hourly
Nevada, United States36.1811−115.13463Cold DesertHalf-Hourly
Yunnan, China25.0918102.90365Temperate MonsoonHalf-Hourly
Table 2. Pearson correlation analysis of solar radiation and meteorological information.
Table 2. Pearson correlation analysis of solar radiation and meteorological information.
FeatureBusanKarnatakaNevadaYunnanAverage
GHI1.0001.0001.0001.0001.000
Temperature0.1940.6490.4940.4700.452
Dew Point−0.026−0.341−0.028−0.209−0.151
DHI0.5820.5250.4030.4870.499
DNI0.7600.7250.7390.6830.727
Pressure−0.0190.049−0.044−0.141−0.039
Wind Direction0.064−0.140−0.1140.2560.016
Wind Speed−0.0830.0680.2020.4530.160
Clearsky DHI0.4530.6090.7330.3310.531
Clearsky DNI0.7260.7670.7600.6930.736
Clearsky GHI0.7860.8860.9360.7420.837
Cloud Type−0.338−0.263−0.265−0.305−0.293
Ozone0.182−0.0270.0350.0820.068
Solar Zenith Angle−0.737−0.847−0.922−0.688−0.799
Precipitable Water−0.152−0.2410.038−0.192−0.137
Relative Humidity−0.553−0.601−0.491−0.732−0.594
Hour−0.0070.001−0.0170.0620.010
Table 3. Forecasting performance of Busan Dataset.
Table 3. Forecasting performance of Busan Dataset.
Time StepEvaluation Metrics Model
TCNLSTMLSTM-AttentionCNN-LSTMTransformerOurs
1-step aheadRMSE2.3142.4412.5694.1473.5832.192
MAE1.6111.6341.8393.1992.7041.517
MAPE0.0220.0230.0260.0460.0380.022
R 2 0.9760.9740.9710.9240.9430.979
2-step aheadRMSE2.8362.7303.3145.1525.2192.540
MAE2.0501.8182.3883.9894.0951.705
MAPE0.0280.0250.0330.0570.0540.024
R 2 0.9650.9670.9520.8830.8800.972
4-step aheadRMSE3.6563.6893.8434.9765.7953.144
MAE2.6092.7312.8193.7974.4572.144
MAPE0.0380.0390.0410.0530.0630.031
R 2 0.9410.9400.9350.8910.8520.956
6-step aheadRMSE4.3554.2584.5555.8357.4253.817
MAE3.2843.0813.3674.3405.8692.649
MAPE0.0460.0440.0490.0650.0820.038
R 2 0.9160.9200.9090.8500.7570.936
8-step aheadRMSE4.9665.2264.9938.3178.3074.598
MAE3.6353.8033.6816.7686.4943.266
MAPE0.0530.0530.0530.0970.0920.047
R 2 0.8910.8800.8900.6950.6960.907
10-step aheadRMSE5.4205.5876.3847.73311.1384.981
MAE3.9873.9594.9466.0569.0903.554
MAPE0.0580.0580.0710.0860.1310.051
R 2 0.8710.8630.8200.7370.4530.891
Table 4. Forecasting performance of Karnataka Dataset.
Table 4. Forecasting performance of Karnataka Dataset.
Time StepEvaluation Metrics Model
TCNLSTMLSTM-AttentionCNN-LSTMTransformerOurs
1-step aheadRMSE2.7802.7133.1555.4788.0752.195
MAE1.9011.7292.2874.1586.0881.503
MAPE0.0350.0320.0500.0770.1150.029
R 2 0.9870.9880.9830.9490.8900.992
2-step aheadRMSE3.2893.3083.5579.8399.2712.675
MAE2.2762.1882.5507.7637.3241.856
MAPE0.0480.0410.0460.1550.1390.037
R 2 0.9820.9820.9790.8360.8550.988
4-step aheadRMSE4.8705.2655.0799.53911.1193.948
MAE3.6123.8893.6667.6178.5702.894
MAPE0.0730.0760.0660.1560.1630.056
R 2 0.9600.9530.9560.8460.7910.974
6-step aheadRMSE5.9866.1606.02810.26611.2775.246
MAE4.5484.6784.4788.0368.7143.943
MAPE0.0850.0900.0840.1640.1980.074
R 2 0.9390.9360.9390.8220.7850.954
8-step aheadRMSE7.5906.1436.96110.47815.5965.995
MAE5.7544.8165.1448.04812.4614.522
MAPE0.1310.0940.0940.1470.2280.088
R 2 0.9030.9360.9180.8140.5890.939
10-step aheadRMSE10.3326.7517.35311.86312.5896.658
MAE7.9275.2525.5589.31910.2935.000
MAPE0.1460.1040.1090.1790.2280.101
R 2 0.8200.9230.9090.7620.7320.925
Table 5. Forecasting performance of Nevada Dataset.
Table 5. Forecasting performance of Nevada Dataset.
Time StepEvaluation Metrics Model
TCNLSTMLSTM-AttentionCNN-LSTMTransformerOurs
1-step aheadRMSE2.8032.8623.6984.7564.8082.508
MAE1.6021.5692.2773.1152.9651.216
MAPE0.0910.0800.1300.1980.1670.063
R 2 0.9520.9500.9170.8620.8590.962
2-step aheadRMSE3.2233.6443.7708.1158.1383.205
MAE1.8132.0072.0315.7965.6931.749
MAPE0.1010.1030.1000.3340.3250.091
R 2 0.9370.9190.9140.5990.5970.938
4-step aheadRMSE5.1884.9325.0086.13010.9504.168
MAE3.0342.9663.0804.1667.2922.090
MAPE0.1550.1580.1700.2450.3790.102
R 2 0.8360.8520.8470.7710.2700.894
6-step aheadRMSE5.3285.3885.3889.2579.5495.122
MAE3.3543.1973.1516.0426.5092.881
MAPE0.1860.1700.1630.3180.3850.142
R 2 0.8270.8230.8230.4780.4440.840
8-step aheadRMSE6.1636.0176.2088.7299.3025.064
MAE3.8643.5883.8545.6266.4633.146
MAPE0.2160.1890.2040.2730.3550.171
R 2 0.7690.7790.7650.5360.4730.844
10-step aheadRMSE6.1116.5376.07211.3179.8365.935
MAE3.8404.0523.6157.1066.8793.472
MAPE0.2210.2140.1820.3350.3980.173
R 2 0.7720.7400.7750.2190.4100.785
Table 6. Forecasting performance of Yunan Dataset.
Table 6. Forecasting performance of Yunan Dataset.
Time StepEvaluation Metrics Model
TCNLSTMLSTM-AttentionCNN-LSTMTransformerOurs
1-step aheadRMSE3.0862.8253.4164.3094.2322.238
MAE2.4051.9572.5203.3113.4341.526
MAPE0.0450.0370.0480.0610.0670.027
R 2 0.9740.9780.9680.9490.9510.986
2-step aheadRMSE3.3293.2233.3826.63410.0252.918
MAE2.4182.2542.3795.1238.0362.074
MAPE0.0460.0410.0430.1000.1550.037
R 2 0.9700.9720.9690.8790.7250.977
4-step aheadRMSE4.5054.3854.1376.79510.4603.762
MAE3.4023.1482.8535.2988.1362.623
MAPE0.0640.0580.0510.1010.1530.047
R 2 0.9440.9470.9530.8740.7000.961
6-step aheadRMSE5.6784.8585.0778.53212.2194.246
MAE4.3853.4363.8756.3579.6882.981
MAPE0.0820.0620.0710.1200.2070.055
R 2 0.9120.9350.9290.8010.5910.951
8-step aheadRMSE6.0445.4295.4396.64310.9764.652
MAE4.4783.9693.7685.0918.4203.270
MAPE0.0830.0750.0690.0960.1710.060
R 2 0.9000.9190.9190.8790.6700.941
10-step aheadRMSE5.8216.3675.8978.90510.7335.604
MAE4.1674.7084.1587.1588.3854.041
MAPE0.0790.0870.0780.1370.1620.077
R 2 0.9070.8890.9050.7830.6850.914
Table 7. Evaluation metrics of ablation experiments.
Table 7. Evaluation metrics of ablation experiments.
Evaluation MetricsModels1-Step2-Step4-Step6-Step8-Step10-Step
RMSEw/o TSAM2.4672.6493.8614.6175.5405.886
w/o RACB2.3132.5943.4504.6584.8365.870
w/o DFEN2.4352.9663.5014.0105.4285.699
w/o LFEN2.5402.7083.4034.2315.1565.674
Ours2.1912.5403.1443.8174.5984.981
MAEw/o TSAM1.7611.8932.8583.4324.1284.395
w/o RACB1.6341.8142.3693.5263.4614.419
w/o DFEN1.6802.2072.4972.8884.1254.279
w/o LFEN1.7931.9232.4103.0463.8884.230
Ours1.5171.7052.1442.6493.2663.554
MAPEw/o TSAM0.0240.0260.0410.0500.0590.065
w/o RACB0.0230.0260.0340.0500.0500.066
w/o DFEN0.0250.0320.0350.0410.0590.061
w/o LFEN0.0270.0280.0340.0440.0550.061
Ours0.0220.0240.0310.0380.0470.051
R 2 w/o TSAM0.9730.9690.9340.9060.8650.847
w/o RACB0.9760.9700.9480.9040.8970.848
w/o DFEN0.9740.9610.9460.9290.8700.857
w/o LFEN0.9720.9680.9490.9210.8830.858
Ours0.9790.9720.9560.9360.9070.891
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, Y.; Tang, Z.; Li, Z.; He, J.; Shi, X.; Zhu, Y. Dual-Path Information Fusion and Twin Attention-Driven Global Modeling for Solar Irradiance Prediction. Sensors 2023, 23, 7469. https://doi.org/10.3390/s23177469

AMA Style

Yang Y, Tang Z, Li Z, He J, Shi X, Zhu Y. Dual-Path Information Fusion and Twin Attention-Driven Global Modeling for Solar Irradiance Prediction. Sensors. 2023; 23(17):7469. https://doi.org/10.3390/s23177469

Chicago/Turabian Style

Yang, Yushi, Zhanjun Tang, Zhiyuan Li, Jianfeng He, Xiaobing Shi, and Yuting Zhu. 2023. "Dual-Path Information Fusion and Twin Attention-Driven Global Modeling for Solar Irradiance Prediction" Sensors 23, no. 17: 7469. https://doi.org/10.3390/s23177469

APA Style

Yang, Y., Tang, Z., Li, Z., He, J., Shi, X., & Zhu, Y. (2023). Dual-Path Information Fusion and Twin Attention-Driven Global Modeling for Solar Irradiance Prediction. Sensors, 23(17), 7469. https://doi.org/10.3390/s23177469

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop