Next Article in Journal
Zeolites: A Theoretical and Practical Approach with Uses in (Bio)Chemical Processes
Next Article in Special Issue
Deep Parallel Optimizations on an LASG/IAP Climate System Ocean Model and Its Large-Scale Parallelization
Previous Article in Journal
An Improved CZT Algorithm for High-Precision Frequency Estimation
Previous Article in Special Issue
Disaster Precursor Identification and Early Warning of the Lishanyuan Landslide Based on Association Rule Mining
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dual-Encoder Transformer for Short-Term Photovoltaic Power Prediction Using Satellite Remote-Sensing Data

1
Computer Network Information Center, Chinese Academy of Sciences, Beijing 100083, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
3
School of Control and Computer Engineering, North China Electric Power University, Beijing 102206, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(3), 1908; https://doi.org/10.3390/app13031908
Submission received: 10 December 2022 / Revised: 19 January 2023 / Accepted: 31 January 2023 / Published: 1 February 2023

Abstract

:
The penetration of photovoltaic (PV) energy has gained a significant increase in recent years because of its sustainable and clean characteristics. However, the uncertainty of PV power affected by variable weather poses challenges to an accurate short-term prediction, which is crucial for reliable power system operation. Existing methods focus on coupling satellite images with ground measurements to extract features using deep neural networks. However, a flexible predictive framework capable of handling these two data structures is still not well developed. The spatial and temporal features are merely concatenated and passed to the following layer of a neural network, which is incapable of utilizing the correlation between them. Therefore, we propose a novel dual-encoder transformer (DualET) for short-term PV power prediction. The dual encoders contain wavelet transform and series decomposition blocks to extract informative features from image and sequence data, respectively. Moreover, we propose a cross-domain attention module to learn the correlation between the temporal features and cloud information and modify the attention modules with the spare form and Fourier transform to improve their performance. The experiments on real-world datasets, including PV station data and satellite images, show that our model achieves better results than other models for short-term PV power prediction.

1. Introduction

Facing the depletion of finite fossil fuels and the demands of carbon emission reduction, photovoltaic (PV) energy has raised widespread interest in recent years [1,2]. As a sustainable and flexible distributed energy source, PV power has been a significant part of the power grid with the rapid growth of its installed capacity worldwide [3]. However, the high intermittency and uncertainty of PV power make accurate prediction difficult and further pose technical challenges for reliable power system operation and control [4,5]. Meteorological factors, e.g., solar irradiance, temperature, wind speed and direction, cloud changes, and air pressure, are the primary causes of this arbitrary fluctuation. Therefore, effective PV prediction places critical demands on the model’s capacity to take into account and handle various meteorological factors.
For PV power prediction, the prediction model and the forecast horizon mainly depend on the data sources. Traditional methods such as ARIMA [6] and exponential smoothing [7] hardly deal with multivariate data and thus usually take historical PV power series as the sole input. However, they are still limited to modeling non-stationary changes. To leverage the meteorological information, many previous studies mainly concentrate on employing local measurements [8] or numerical weather prediction (NWP) data [9]. However, these methods are incapable of capturing the volatility of PV power affected by cloud changes due to ignoring the data sources containing cloud conditions. More specifically, the solar irradiance reaching the PV panels mainly depends on the amount and motion of clouds [10,11] recorded in consecutive cloud images. In light of this, ground-based or satellite remote-sensing images are valuable data sources to support accurate short-term PV power prediction.
Several approaches employ ground-based/sky images for the prediction of irradiance [12,13] and PV power [14]. With sky images, the cloud motion vector (CMV) methods that can calculate the speed and direction of clouds have been developed to improve the irradiance prediction accuracy [15,16]. These methods have a considerably decreasing value beyond a 30 min future horizon and are thus commonly used in ultra-short-term PV power prediction [17]. Moreover, ground-based images are limited in observation range and require equipment installation. In contrast, satellite images offer top-down detection with greater spatial coverage, as well as advantages for prediction up to a few hours ahead [18]. They are not only suitable for CMV methods [11,19,20,21] but also able to provide other valuable information. Cai et al. [22] cluster the interval gray value of satellite images to construct a relationship with PV power. Wang et al. [23] propose a CNN-based fluctuation pattern prediction model with satellite images as input and apply LSTMs to individually forecast different patterns of PV power. Besides these methods that adopt only satellite images, researchers are increasingly interested in coupling satellite data with ground measurements to improve forecasting performance [24]. Si et al. [25] extracted cloud cover factor from satellite images by the modified CNN and then combined it with meteorological information for irradiance prediction. Agoua et al. [26] presented a spatiotemporal model with Lasso to integrate multi-source data for 6-h-ahead prediction. Furthermore, they used the proposed model to assess the impact of each source on forecasting performance, which showed that satellite images improved the accuracy by 3 % . Yao et al. [27] proposed an encoder–decoder model, which includes U-Net and LSTM to extract the spatial features from satellite short-wave radiation data and temporal features from meteorological sequences, respectively. Then, the spatial and temporal features were concatenated for intra-hour PV power prediction.
However, a flexible predictive framework capable of handling data sources with different structures (i.e., image and sequence) is still not well discussed. Due to the insufficient consideration of the characteristics of these two structures, the spatiotemporal features extracted using the above methods are still not informative for PV power prediction up to hours. Although various deep neural networks have been applied in PV power prediction, the majority of existing methods consist of conventional deep learning models such as CNNs, RNNs, and fully connected layers. At present, few studies have followed the state-of-the-art transformer [28] architecture, which has achieved great success in natural language processing [29] and computer vision [30]. Compared with convolutional or recurrent neural networks, transformer-based models with the self-attention mechanism are superior in capturing temporal dependencies. Simeunovic et al. [31] combined a graph neural network with a transformer to propose a graph–convolutional transformer for multi-site PV power prediction based on historical weather data. In addition, most existing models based on hybrid neural networks merely concatenate the spatial and temporal features before passing them as a whole input to the following layer. As a result, this rough treatment is incapable of learning and utilizing the correlation between the features of structured local measurements and unstructured satellite images well.
Motivated by the aforementioned issues, we renovate the transformer architecture and propose a novel transformer-based PV power prediction model that can process the locally measured sequences and satellite remote-sensing images. The main contributions are summarized as follows:
  • We propose a novel dual-encoder transformer (DualET) for short-term PV power prediction. Distinct from the standard transformer architecture, DualET contains dual encoders, including a local seasonal information (LSI) encoder and a remote-sensing information (RSI) encoder, and a single decoder. The dual encoders are designed to handle sequence and image data, while the decoder is to model the joint meteorological features from the encoders and outputs the short-term prediction.
  • To extract informative features from the satellite images and local sequences, we deploy the two-dimensional wavelet transform block and series decomposition block in the encoding stage. The former in the RSI encoder is capable of the frequency feature extraction of image signals to obtain cloud detailed information, while the latter in the LSI encoder conducts series decomposition to improve the learning capacity of temporal patterns.
  • We propose a cross-domain attention module to learn the correlation between the temporal features in sequence data and the cloud information in image data. Furthermore, we enhance the ability to capture dependencies by modifying the attention modules with the sparse form and Fourier transform.
  • Real-world datasets are applied to evaluate the proposed model. The experiments show that our model achieves state-of-the-art results compared with other models including recent transformers.

2. Problem Definition

Two types of data sources are used in this study to predict the PV power, including cloud images processed from the satellite remote sensing data and in situ data from real PV stations. The in situ data contain historical records of several meteorological factors and PV power. The detailed data description is provided in Section 4.1. We denote the historical steps of input data as L in and the prediction steps as L out . The number of in situ data attributes and remote-sensing data channels used is denoted as D LS and C RS , respectively. In summary, this PV power forecasting problem can be formulated as
f ( X RS ; X LS ) y ^
where f ( · ) is the mapping function, i.e., the forecasting model; X RS R L in × H × W × C RS , X LS R L in × D LS are the input remote-sensing data and local sequences (in situ data), respectively; y ^ R L out is the PV power prediction. The hidden dimension of the model is denoted as D.

3. Methodology

PV power generation substantially fluctuates with meteorological factors, such as solar irradiance, temperature, wind speed/direction, and cloud changes. For an accurate PV power prediction, we propose a novel dual-encoder transformer (DualET) to capture context features from these factors. As shown in Figure 1, DualET has an encoder–decoder architecture like most transformers but with dual encoders. One encoder, i.e., local seasonal information encoder (LSI encoder), is to model temporal dynamics from in situ measurements. The other, i.e., remote-sensing information encoder (RSI encoder), is to learn the spatial and temporal features of clouds from satellite images. The output of the LSI encoder and that of the RSI encoder contain fluctuations in the local measurements and clouds, respectively. Both of them can be combined to provide comprehensive fluctuation information for modeling the changes in the PV power series. Further, we design a joint feature decoder to predict future short-term PV power. The details of DualET will be introduced in the following subsections.

3.1. Decomposition Modules

3.1.1. Wavelet Transform Block

The object edges represent abrupt changes around smooth regions and are distributed in the high-frequency signals of an image. For the motion and coverage of clouds, cloud edges offer crucial information. Wavelet transform is a powerful analysis tool widely used in image signal processing [32]. Compared with Fourier transform, it can capture frequency properties without location information loss. Therefore, we build the wavelet transform block (WTBlock) to extract frequency features and provide the RSI encoder with additional cloud details. As shown in Figure 2, the image signals are passed through high-pass and low-pass filters, sequentially from the horizontal and vertical directions. The high-pass filter (HPF) is to extract high-frequency components such as edges, while the low-pass filter (LPF) is to obtain low-frequency components for approximation. We summarize this 2D discrete wavelet transform as
I LL , I LH , I HL , I HH = WTBlock ( I ) ,
where image signal I is decomposed into four components: the approximation I LL (passing through LPFs in both directions), and three details ( I LH , I HL , I HH ) in horizontal, vertical, and diagonal orientations, respectively.

3.1.2. Series Decomposition Block

Real-world sequential signals (e.g., PV power sequence) are often entangled with temporal patterns that are informative for forecasting. Time series decomposition is an effective strategy to decouple knotted patterns from sequences. Among decomposition methods, seasonal-trend decomposition [33] has been widely employed as a feature engineering technique, which separates sequences into seasonal and trend parts. Inspired by Autoformer [34], we apply this decomposition idea as a series decomposition block (SDBlock) to enhance the pattern extraction ability of DualET. Given an input sequence X , the procedure is
S , T = SDBlock ( X ) ,
where T is the moving average result of X and is considered as the trend part; S is the residual part (i.e., the detrend part), which is regarded as the seasonal part. To keep the sequence length unchanged, a padding operation is performed on the input sequence.

3.2. Learning Modules

3.2.1. Residual Connection and Residual Block

The residual network architecture (i.e., residual connection) has become the foundation of deep neural networks, which learn residual mapping to ease the optimization of deep layers [35]. It can be formalized as y = F ( x ) + x , i.e., the input is added to the output of stacked layers (F) as the result. For the RSI encoder, we employ residual blocks to learn the representation of remote-sensing data. As shown in Figure 3, the residual block is stacked using two convolutions, batch normalization [36], and activation (ReLU) layers. The process is summarized as
X R = ResBlock ( X ) .

3.2.2. Attention Mechanism

As one of the most representative identifiers of transformers, the attention mechanism is proposed as a query–key–value (QKV) model to learn long-range dependencies without recurrent structures. Given the matrices Q R L q × D k , K R L k × D k , and V R L k × D v as the projected queries, keys, and values, the single-head version of standard attention mechanism can be formalized as A ( Q , K , V ) = Softmax ( Q K D k ) V , where L q , L k denote the length of queries and keys/values; D k , D v denote the projected dimensions; 1 D k is the scale factor to avoid Softmax ( · ) yielding extremely small gradients. Furthermore, the multi-head version is as follows:
A multi-head ( Q , K , V ) = Concat ( head 1 , , head H ) W O , where head i = A ( Q W i Q , K W i K , V W i V ) .
The queries, keys, and values with dimension D are mapped into H heads (i.e., subspaces) by W i Q , W i K R D × D k , and V W i V R D × D v . Then the outputs of these heads are concatenated and mapped back to D by W O R H D v × D . In most cases, D k = D v = D / H . The standard transformer has two types of multi-head attention: self-attention and cross-attention. For self-attention, its projected queries, keys, and values are shared with the same source, while the key–value pairs of cross-attention are typically from the output of encoder.
In practice, the attention modules used in DualET are modified to improve the performance of capturing dependencies. First, we design an additional cross-domain attention module to discover the correlation between the features of images and sequences. Concretely, the queries are the temporal features of the decoder, and the key–value pairs are the cloud information from the RSI encoder. Furthermore, we perform the fast Fourier transform (FFT) on the input and the inverse FFT on the output:
A FFT ( Q , K , V ) = F 1 ( A ( F ( Q ) , F ( K ) , F ( V ) ) )
where F , F 1 denote the FFT and its inverse, which are also used in the self-attention module of the LSI encoder. The FFT plays a key role in signal processing because it can rapidly convert a signal from the time/space domain to the frequency domain (and vice versa) and describe relationships between these domains [37]. It is defined by
X ¯ k = m = 0 L 1 e 2 π i k · ( m / L ) X m , k = 0 , , L 1
Based on FFT, the attention module can discover the frequency correlation between queries and keys. In addition, we employ the ProbSparse attention mechanism [38] as the self-attention and cross-attention modules of the decoder to improve their performances.

3.2.3. Embedding and Feed-Forward Layer

For sequence modeling, the order information of time steps is crucial. Furthermore, the timestamp records of local sequences (meteorological and PV power series) are instructive for PV power prediction but are hardly utilized in the standard transformer architecture. To introduce this information, we employ timestamp-embedding layers (following Autoformer [34]) for the local sequence inputs.
The feed-forward layer is a position-wise fully connected module, i.e., the learnable parameters are shared with each step. It contains two linear layers ( W 1 , b 1 , W 2 , b 2 ) and a ReLU activation function in between, formulated as
FeedFoward ( X ) = ReLU ( XW 1 + b 1 ) W 2 + b 2 .

3.3. Remote-Sensing Information Encoder

The RSI encoder is designed to learn spatial and temporal features from remote-sensing data. As shown in the top diagram of Figure 1, it is mainly composed of a two-dimensional convolution layer, a wavelet transform block, and a residual block. Given the historical L in steps of cloud images X RS R L in × H × W × C RS as the input of the RSI encoder, the procedure is
I 1 = Conv2D ( X RS ) , I LL , I LH , I HL , I HH = WTBlock ( I 1 ) , I 2 = Conv2D ( ResBlock ( I 1 ) ) , I 3 = ConvFusion ( Concat ( [ I 2 , I LL , I LH + I HL + I HH ] ) ) , Z RS = Linear ( Flatten ( ReLU ( I 3 ) ) ) ,
where ConvFusion is 1 × 1 2D convolution to integrate frequency components with image features; Z RS R L in × D is the output of the RSI encoder.

3.4. Local Seasonal Information Encoder

For weather and PV power series, the seasonal part contains the main volatility feature, which is the key to accurate prediction. Therefore, we introduce series decomposition blocks to the LSI encoder to extract seasonal patterns from local measurement data. As shown in the middle of Figure 1, the LSI encoder is staked with L LSI LSI encoder layers. Given the input of the LSI encoder X LS 0 R L in × D that is embedded from X LS R L in × D DS , the procedure in l-th LSI encoder layer is
S LS , 1 l , _ = SDBlock ( Attention ( X LS l 1 ) + X LS l 1 ) , S LS , 2 l , _ = SDBlock ( FeedFoward ( S LS , 1 l ) + S LS , 1 l ) , X LS l = S LS , 2 l ,
where “_” is the ignored trend part; S LS , i l , i { 1 , 2 } denotes the seasonal part in the l-th layer; Z LS = X LS L LSI denotes the output of the LSI encoder.

3.5. Joint-Feature Decoder

The joint-feature decoder is to model temporal dynamics based on the joint features of local and remote-sensing data and then outputs short-term PV power. As shown in the bottom diagram of Figure 1, the decoder is stacked with L de decoder layers, and each layer contains three attention modules (i.e., self-attention, cross-domain attention, and cross-attention) to determine the correlations from different perspectives. The outputs of two encoders are integrated as the input of cross-attention by ConvFusion: Z en = ConvFusion ( Concat ( [ Z RS , Z LS ] ) ) . The inputs of the decoder contain initialized seasonal part X de 0 and trend part T de 0 . which are decomposed from the latter half of X LS 0 and concatenated with scalar placeholders (zeros for the seasonal part and means for the trend part). The details in the l-th decoder layer are
Z de , 1 l = Attention ( X de l 1 ) , Z de , 2 l = CrossDomianAttention ( X de l 1 , Z RS ) , Z de , 3 l = X de l 1 + ConvFusion ( Concat ( [ Z de , 1 l , Z de , 2 l ] ) ) , S de , 1 l , T de , 1 l = SDBlock ( Attention ( Z de , 3 l , Z en ) + Z de , 3 l ) , X de l , T de , 2 l = SDBlock ( FeedFoward ( S de , 1 l ) + S de , 1 l ) , T de l = T de l 1 + W 1 l T de , 1 l + W 2 l T de , 2 l ,
where Z de , i l , i { 1 , 2 , 3 } is the intermediate feature; S de , 1 l , X de l denote the seasonal parts in l-th layer; T de , 1 l ( i { 1 , 2 } ) , T de l denote the trend parts in l-th layer; W i l { 1 , 2 } ) denote the project functions for the trend parts. After L de decoder layers, the final prediction y ^ is from the sum of two parts: W X de L de + T de L de , where W is the projector for the seasonal part.

4. Experiment

In this section, we evaluate the proposed DualET on satellite images and actual PV station data. We first introduce the datasets and data preprocessing. Then, we describe the experimental setting in detail. Finally, we compare the prediction performance of DualET and the baseline models and conduct several ablation experiments.

4.1. Datasets and Data Preprocessing

Two datasets were used in this study, namely, satellite remote-sensing data and PV station data. The satellite data were the L1 grid data from Himawa-8, a geostationary satellite launched in 2015 by the Japan Meteorological Agency to provide weather forecasts and typhoon and storm reports for Japan, East Asia, and the Western Pacific. The detection range of Himawa-8 is 60 N, 160 E, and 80 W, with a spatial resolution of 0 . 05 , which corresponds to about 5 km on the ground, and the temporal resolution is 10 min. The PV station data contains local measurements from three real PV stations at different latitudes and longitudes in Hebei, China. Each station records meteorological factors (including global and diffuse irradiance, temperature, wind direction and speed, and air pressure) and PV power records at 15 min intervals.
We set the temporal resolution as 30 min to harmonize the time intervals of these two datasets. The satellite remote-sensing data were processed into 40 × 40 cloud images centered on the latitude and longitude of the PV station. The satellite data were sampled from July 2018 to June 2019 to align with that of PV station data. For each day, the data with a time range from 7:00 to 19:00 (UTC +0800) were used. We divided these two datasets into a training set, a validation set, and a test set in the ratio of 8:1:1 after arranging them to ensure the test period covers multiple seasons. Before input to the model, the data were normalized using standardization to eliminate the inconsistency of the magnitude of each dimension.

4.2. Experimental Setting

4.2.1. Baseline Models

We primarily selected seven models as baselines for comparison, namely, a classic statistical model ARIMA [39], an RNN-based model LSTM [40], and three state-of-the-art models, i.e., Transformer [28] and its variants Informer [38] and Autoformer [34].

4.2.2. Hyperparameters and Platform

Our model DualET and transformer baselines, i.e., Transformer, Informer, and Autoformer, were set to the same number of layers, including two encoder layers and one decoder layer. The hidden dimension of model D was set to 512. The number of attention heads was set to 8. The batch size was set to 32, and the training epochs were set to 10 (with early stopping). The loss function was the mean-squared error (MSE) (Equation (12)), and the optimizer was ADAM with an initial learning rate of 1 × 10 4 . The input length of the model was set to 24, i.e., 12 h, and the output prediction length was set to 12, i.e., 6 h. All the models were implemented with PyTorch and conducted on a Ubuntu server with four NVIDIA GeForce RTX 2080Ti 11GB GPUs.
MSE = 1 N t = 1 N ( y t y t ^ ) 2 .

4.2.3. Evaluation Metrics

We evaluated the performance of the model with three widely used metrics, i.e., mean absolute error (MAE), root-mean-squared error (RMSE), and symmetric mean absolute percentage error (SMAPE).
MAE = 1 N t = 1 N | y t y t ^ | , RMSE = 1 N t = 1 N ( y t y ^ t ) 2 , SMAPE = 100 % N t = 1 N | y t y ^ t | ( | y t | + | y ^ t | ) / 2 .

4.3. Results

As shown in Table 1, we evaluated the prediction performance of the proposed DualET on three different PV stations. For short-term (6-h) PV power prediction, DualET achieved the best results on three error metrics: MAE, RMSE, and SMAPE. Compared with the results of other models on all stations, DualET achieved a relative MAE and RMSE reduction of 22.53% and 22.75%, respectively, which is a significant improvement. The average reduction in MAE was even more than 53%, compared with the traditional ARIMA, and 27.72% compared with the popular LSTM. Among these baselines, the transformer-based models, i.e., Transformer, Informer, and Autoformer, were better than ARIMA and LSTM models. Moreover, our DualET still outperformed the competitive transformer-based models and yielded a relative MAE reduction of 10.64%.
The prediction results of different models are presented in Figure 4 and Figure 5. It obviously shows that the number of deviation points predicted by the statistical model ARIMA is much larger than those predicted by other DNN-based models, which indicates that ARIMA is unsuitable for short-term prediction up to several hours, especially for the nonstationary PV power series. As shown in Figure 5, the LSTM model has a more scattered distribution of points than the transformer-based models, which indicates the significance of transformer architecture for sequence modeling. It can also be seen that the proposed DualET presents the fittest curves in Figure 4 and the narrowest band in Figure 5 compared with other baselines, which shows the advantages of PV power prediction.

4.4. Ablation Studies

DualET contains dual encoders, including the LSI encoder to deal with local seasonal information and the RSI encoder to process remote-sensing information, with a shared decoder to combine joint features. In addition, there are different decomposition modules and attention modules employed in DualET to enhance feature extraction. We conducted additional experiments to evaluate the impact of the dual encoders and different functional modules with MAE, RMSE, and SMAPE as evaluation metrics, and we present the results of these experiments in this section.

4.4.1. Different Encoders

We evaluated the impact of different encoders by keeping only one of the encoders from DualET. As shown in Table 2, the proposed model with only the LSI encoder is better than that with only the RSI encoder. As previously mentioned, the LSI encoder is used to learn historical seasonal information from local measurements, while the RSI encoder is to learn cloud information from satellite images. Therefore, this indicates that the fluctuation features of PV power are more contained in local measurements compared with cloud images. Meanwhile, we can see that the model with dual encoders yields the best results compared with that with only one, which shows that cloud images can provide valuable information for performance improvement and further shows the effectiveness of our dual-encoder design.

4.4.2. Decomposition Blocks in Dual Encoders

We evaluated the impact of decomposition blocks for dual encoders by removing the series decomposition (SD) blocks or wavelet transform (WT) blocks. As shown in Table 3, the dual encoders without decomposition blocks will give some performance degradation. This shows that the WT and SD can enhance the learning capacity of dual encoders and provide informative features for the decoder. In addition, the performance of the RSI encoder without WT is worse than the LSI encoder without SD, which indicates that the high- and low-frequency features of cloud images are essential for short-term PV power prediction.

4.4.3. Different Attention Modules

In DualET, the decoder’s self- and cross-attention modules are formed with ProbSparse, while the others are enhanced by fast Fourier transform (FFT). As shown in Table 4, we evaluated the impact of different attention modules including the cross-domain attention module. We can see that if all attention modules are the same type (ProbSparse form or FFT), it results in performance degradation. Therefore, we employed the various attention modules with suitable modifications for better correlation capture. Moreover, the prediction error of the model without cross-domain attention significantly increases. This demonstrates the effectiveness of cross-domain attention to learn the correlation between features in sequence data and image data.

5. Conclusions

To handle satellite images and ground measurements, in this paper, we proposed a novel transformer-based model with dual encoders, named DualET, for short-term PV power prediction. To obtain cloud detailed information from satellite images, a two-dimensional wavelet transform block and a residual block were used in the remote-sensing information encoder. For the local seasonal information encoder, we conducted self-attention and series decomposition to learn the temporal patterns from local sequences. For the decoder, we employed three types of attention modules and series decomposition blocks to model the joint features of local and remote-sensing information and output the prediction. Specifically, a cross-domain attention module was proposed to learn the correlation between the temporal features and cloud information. Finally, the experiments on real-world datasets, including PV station data and satellite images, were presented to show the prediction performance of DualET. In addition, the ablation studies show the effectiveness of our design. In the future, we will attempt to improve the model architecture so that more data sources (e.g., NWP or other satellite remote sensing data) can be utilized to predict a longer horizon.

Author Contributions

Conceptualization, J.W. and H.H.; methodology, H.C. and J.Y.; software, H.C. and J.Y.; validation, X.Z., T.Y. and Y.W.; writing—original draft preparation, H.C. and X.Z.; writing—review and editing, T.Y., J.W. and Y.W.; visualization, J.Y. and X.Z.; supervision, J.W. and H.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Key R&D Program of China (2021ZD0110403).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The remote-sensing data used in this paper were processed from Himawari-8 satellite data supplied by the P-Tree System, Japan Aerospace Exploration Agency (JAXA). We also gratefully acknowledge the support of the MindSpore team.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Kabir, E.; Kumar, P.; Kumar, S.; Adelodun, A.A.; Kim, K.H. Solar Energy: Potential and Future Prospects. Renew. Sustain. Energy Rev. 2018, 82, 894–900. [Google Scholar] [CrossRef]
  2. Armeanu, D.S.; Joldes, C.C.; Gherghina, S.C.; Andrei, J.V. Understanding the Multidimensional Linkages among Renewable Energy, Pollution, Economic Growth and Urbanization in Contemporary Economies: Quantitative Assessments across Different Income Countries’ Groups. Renew. Sustain. Energy Rev. 2021, 142, 110818. [Google Scholar] [CrossRef]
  3. Carriere, T.; Vernay, C.; Pitaval, S.; Kariniotakis, G. A Novel Approach for Seamless Probabilistic Photovoltaic Power Forecasting Covering Multiple Time Frames. IEEE Trans. Smart Grid 2019, 11, 2281–2292. [Google Scholar] [CrossRef]
  4. Sanjari, M.J.; Gooi, H. Probabilistic Forecast of PV Power Generation Based on Higher Order Markov Chain. IEEE Trans. Power Syst. 2016, 32, 2942–2952. [Google Scholar] [CrossRef]
  5. Stein, G.; Letcher, T.M. Integration of PV Generated Electricity into National Grids. In A Comprehensive Guide to Solar Energy Systems; Letcher, T.M., Fthenakis, V.M., Eds.; Academic Press: Cambridge, MA, USA, 2018; pp. 321–332. [Google Scholar] [CrossRef]
  6. Li, Y.; Su, Y.; Shu, L. An ARMAX Model for Forecasting the Power Output of a Grid Connected Photovoltaic System. Renew. Energy 2014, 66, 78–89. [Google Scholar] [CrossRef]
  7. Prema, V.; Rao, K.U. Development of Statistical Time Series Models for Solar Power Prediction. Renew. Energy 2015, 83, 100–109. [Google Scholar] [CrossRef]
  8. Liu, L.; Zhan, M.; Bai, Y. A Recursive Ensemble Model for Forecasting the Power Output of Photovoltaic Systems. Sol. Energy 2019, 189, 291–298. [Google Scholar] [CrossRef]
  9. Böök, H.; Lindfors, A.V. Site-Specific Adjustment of a NWP-Based Photovoltaic Production Forecast. Sol. Energy 2020, 211, 779–788. [Google Scholar] [CrossRef]
  10. Breitkreuz, H.; Schroedter-Homscheidt, M.; Holzer-Popp, T.; Dech, S. Short-Range Direct and Diffuse Irradiance Forecasts for Solar Energy Applications Based on Aerosol Chemical Transport and Numerical Weather Modeling. J. Appl. Meteorol. Climatol. 2009, 48, 1766–1779. [Google Scholar] [CrossRef]
  11. Kato, T.; Manabe, Y.; Funabashi, T.; Yoshiura, K.; Kurimoto, M.; Suzuoki, Y. A Study on Several Hours Ahead Forecasting of Spatial Average Irradiance Using NWP Model and Satellite Infrared Image. In Proceedings of the 2016 IEEE International Conference on Probabilistic Methods Applied to Power Systems (PMAPS), Beijing, China, 16–20 October 2016; pp. 1–8. [Google Scholar]
  12. Zhang, C.; Du, Y.; Chen, X.; Lu, D.D.C. Cloud Motion Tracking System Using Low-Cost Sky Imager for PV Power Ramp-Rate Control. In Proceedings of the 2018 IEEE International Conference on Industrial Electronics for Sustainable Energy Systems (IESES), Hamilton, New Zealand, 31 January–2 February 2018; pp. 493–498. [Google Scholar]
  13. Zhao, X.; Wei, H.; Wang, H.; Zhu, T.; Zhang, K. 3D-CNN-Based Feature Extraction of Ground-Based Cloud Images for Direct Normal Irradiance Prediction. Sol. Energy 2019, 181, 510–518. [Google Scholar] [CrossRef]
  14. Lin, F.; Zhang, Y.; Wang, J. Recent Advances in Intra-Hour Solar Forecasting: A Review of Ground-Based Sky Image Methods. Int. J. Forecast. 2022, 39, 244–265. [Google Scholar] [CrossRef]
  15. Bosch, J.L.; Kleissl, J. Cloud Motion Vectors from a Network of Ground Sensors in a Solar Power Plant. Sol. Energy 2013, 95, 13–20. [Google Scholar] [CrossRef]
  16. Peng, Z.; Yu, D.; Huang, D.; Heiser, J.; Kalb, P. A Hybrid Approach to Estimate the Complex Motions of Clouds in Sky Images. Sol. Energy 2016, 138, 10–25. [Google Scholar] [CrossRef]
  17. Tuohy, A.; Zack, J.; Haupt, S.E.; Sharp, J.; Ahlstrom, M.; Dise, S.; Grimit, E.; Mohrlen, C.; Lange, M.; Casado, M.G.; et al. Solar Forecasting: Methods, Challenges, and Performance. IEEE Power Energy Mag. 2015, 13, 50–59. [Google Scholar] [CrossRef]
  18. Kallio-Myers, V.; Riihelä, A.; Lahtinen, P.; Lindfors, A. Global Horizontal Irradiance Forecast for Finland Based on Geostationary Weather Satellite Data. Sol. Energy 2020, 198, 68–80. [Google Scholar] [CrossRef]
  19. Cros, S.; Liandrat, O.; Sebastien, N.; Schmutz, N. Extracting cloud motion vectors from satellite images for solar power forecasting. In Proceedings of the 2014 IEEE Geoscience and Remote Sensing Symposium, Quebec City, QC, Canada, 13–18 July 2014; pp. 4123–4126. [Google Scholar]
  20. Kebir, N.; Maaroufi, M. Best-effort algorithm for predicting cloud motion impact on solar PV power systems production. In Proceedings of the 2018 6th International Istanbul Smart Grids and Cities Congress and Fair (ICSG), Istanbul, Turkey, 25–26 April 2018; pp. 34–38. [Google Scholar]
  21. Zhou, Z.; Zhang, X.; Zhen, Z.; Mei, S. Cloud Displacement Vector Calculation in Satellite Images Based on Cloud Pixel Spatial Aggregation and Edge Matching for PV Power Forecasting. In Proceedings of the 2020 IEEE Sustainable Power and Energy Conference (iSPEC), Chengdu, China, 23–25 November 2020; pp. 112–119. [Google Scholar]
  22. Cai, Y.; Liu, H.; Hu, P.; Fu, Z.; Wang, Y.; Zhang, D.; Ma, X.; Li, S. Ultra-short-term Photovoltaic Power Prediction Based on Elman Neural Network and Satellite Cloud Images. In Proceedings of the 2021 IEEE 5th Conference on Energy Internet and Energy System Integration (EI2), Taiyuan, China, 22–24 October 2021; pp. 2149–2154. [Google Scholar]
  23. Wang, C.; Lu, X.; Zhen, Z.; Wang, F.; Xu, X.; Ren, H. Ultra-Short-Term Regional PV Power Forecasting Based on Fluctuation Pattern Recognition with Satellite Images. In Proceedings of the 2020 IEEE 3rd Student Conference on Electrical Machines and Systems (SCEMS), Jinan, China, 4–6 December 2020; pp. 970–975. [Google Scholar]
  24. Blanc, P.; Remund, J.; Vallance, L. Short-term solar power forecasting based on satellite images. In Renewable Energy Forecasting; Elsevier: Amsterdam, The Netherlands, 2017; pp. 179–198. [Google Scholar]
  25. Si, Z.; Yu, Y.; Yang, M.; Li, P. Hybrid Solar Forecasting Method Using Satellite Visible Images and Modified Convolutional Neural Networks. IEEE Trans. Ind. Appl. 2020, 57, 5–16. [Google Scholar] [CrossRef]
  26. Agoua, X.G.; Girard, R.; Kariniotakis, G. Photovoltaic Power Forecasting: Assessment of the Impact of Multiple Sources of Spatio-Temporal Data on Forecast Accuracy. Energies 2021, 14, 1432. [Google Scholar] [CrossRef]
  27. Yao, T.; Wang, J.; Wu, H.; Zhang, P.; Li, S.; Xu, K.; Liu, X.; Chi, X. Intra-Hour Photovoltaic Generation Forecasting Based on Multi-Source Data and Deep Learning Methods. IEEE Trans. Sustain. Energy 2021, 13, 607–618. [Google Scholar] [CrossRef]
  28. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  29. Wolf, T.; Debut, L.; Sanh, V.; Chaumond, J.; Delangue, C.; Moi, A.; Cistac, P.; Rault, T.; Louf, R.; Funtowicz, M.; et al. Transformers: State-of-the-Art Natural Language Processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, Online, 16–20 November 2020; pp. 38–45. [Google Scholar]
  30. Han, K.; Wang, Y.; Chen, H.; Chen, X.; Guo, J.; Liu, Z.; Tang, Y.; Xiao, A.; Xu, C.; Xu, Y.; et al. A Survey on Vision Transformer. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 87–110. [Google Scholar] [CrossRef]
  31. Simeunovic, J.; Schubnel, B.; Alet, P.J.; Carrillo, R.E. Spatio-Temporal Graph Neural Networks for Multi-Site PV Power Forecasting. IEEE Trans. Sustain. Energy 2021, 13, 1210–1220. [Google Scholar] [CrossRef]
  32. Walker, J.S. Wavelet-Based Image Processing. Appl. Anal. 2006, 85, 439–458. [Google Scholar] [CrossRef]
  33. Robert, C.; William, C.; Irma, T. STL: A Seasonal-Trend Decomposition Procedure Based on Loess. J. Off. Stat. 1990, 6, 3–73. [Google Scholar]
  34. Wu, H.; Xu, J.; Wang, J.; Long, M. Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting. In Proceedings of the Advances in Neural Information Processing Systems, Virtual, 6–14 December 2021; Curran Associates, Inc.: Red Hook, NY, USA, 2021; Volume 34, pp. 22419–22430. [Google Scholar]
  35. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  36. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In Proceedings of the International Conference on Machine Learning, PMLR, Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
  37. Brigham, E.O. The Fast Fourier Transform and Its Applications; Prentice-Hall, Inc.: Upper Saddle River, NJ, USA, 1988. [Google Scholar]
  38. Zhou, H.; Zhang, S.; Peng, J.; Zhang, S.; Li, J.; Xiong, H.; Zhang, W. Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting. Proc. AAAI Conf. Artif. Intell. 2021, 35, 11106–11115. [Google Scholar] [CrossRef]
  39. Chen, P.; Pedersen, T.; Bak-Jensen, B.; Chen, Z. ARIMA-Based Time Series Model of Stochastic Wind Power Generation. IEEE Trans. Power Syst. 2009, 25, 667–676. [Google Scholar] [CrossRef]
  40. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Model architecture of DualET.
Figure 1. Model architecture of DualET.
Applsci 13 01908 g001
Figure 2. Illustration of 2D wavelet transform. LPF denotes the low-pass filter and HPF denotes the high-pass filter.
Figure 2. Illustration of 2D wavelet transform. LPF denotes the low-pass filter and HPF denotes the high-pass filter.
Applsci 13 01908 g002
Figure 3. Illustration of residual block.
Figure 3. Illustration of residual block.
Applsci 13 01908 g003
Figure 4. Prediction results of different models.
Figure 4. Prediction results of different models.
Applsci 13 01908 g004
Figure 5. Scatter plots of different models.
Figure 5. Scatter plots of different models.
Applsci 13 01908 g005
Table 1. Prediction performance of the proposed DualET.
Table 1. Prediction performance of the proposed DualET.
ModelsMetricsStation0Station1Station2
ARIMAMAE3.69343.77273.6343
RMSE5.36336.38095.4976
SMAPE0.88810.90190.9135
LSTMMAE2.35002.26122.6287
RMSE3.08662.97393.4570
SMAPE0.72180.68880.7635
TransformerMAE2.07802.06781.8413
RMSE2.84682.76592.5875
SMAPE0.71100.64820.6274
InformerMAE1.82741.83931.8564
RMSE2.58742.57732.5940
SMAPE0.65490.62320.6498
AutoformerMAE1.99892.06901.9771
RMSE2.73792.78342.7700
SMAPE0.69050.67640.6810
DualETMAE1.69041.75761.7657
RMSE2.28452.42052.5087
SMAPE0.63350.59310.6258
Table 2. Comparison of different encoders.
Table 2. Comparison of different encoders.
ModelsMAERMSESMAPE
Only LSI encoder1.99482.69640.6784
Only RSI encoder2.02752.74530.6937
DualET1.69042.28450.6335
Table 3. Ablation studies of decomposition structures. SD and WT mean the series decomposition block and wavelet transform block, respectively. W/o means “without”.
Table 3. Ablation studies of decomposition structures. SD and WT mean the series decomposition block and wavelet transform block, respectively. W/o means “without”.
ModelsMAERMSESMAPE
LSI w/o SD1.93782.59270.6935
RSI w/o WT1.99162.64400.7094
DualET1.69042.28450.6335
Table 4. Different attention modules. W/o means “without”.
Table 4. Different attention modules. W/o means “without”.
ModelsMAERMSESMAPE
W/o cross-domain attention2.01922.71001.0466
All attention with FFT1.80602.49650.6630
All attention with ProbSparse1.78482.45070.6657
DualET1.69042.28450.6335
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cao, H.; Yang, J.; Zhao, X.; Yao, T.; Wang, J.; He, H.; Wang, Y. Dual-Encoder Transformer for Short-Term Photovoltaic Power Prediction Using Satellite Remote-Sensing Data. Appl. Sci. 2023, 13, 1908. https://doi.org/10.3390/app13031908

AMA Style

Cao H, Yang J, Zhao X, Yao T, Wang J, He H, Wang Y. Dual-Encoder Transformer for Short-Term Photovoltaic Power Prediction Using Satellite Remote-Sensing Data. Applied Sciences. 2023; 13(3):1908. https://doi.org/10.3390/app13031908

Chicago/Turabian Style

Cao, Haizhou, Jing Yang, Xuemeng Zhao, Tiechui Yao, Jue Wang, Hui He, and Yangang Wang. 2023. "Dual-Encoder Transformer for Short-Term Photovoltaic Power Prediction Using Satellite Remote-Sensing Data" Applied Sciences 13, no. 3: 1908. https://doi.org/10.3390/app13031908

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop