With the rapid development of solar energy plants in recent years, the accurate prediction of solar power generation has become an important and challenging problem in modern intelligent grid systems. To improve the forecasting accuracy of solar energy generation, an effective and robust decomposition-integration method for two-channel solar irradiance forecasting is proposed in this study, which uses complete ensemble empirical mode decomposition with adaptive noise (CEEMDAN), a Wasserstein generative adversarial network (WGAN), and a long short-term memory network (LSTM). The proposed method consists of three essential stages. First, the solar output signal is divided into several relatively simple subsequences using the CEEMDAN method, which has noticeable frequency differences. Second, high and low-frequency subsequences are predicted using the WGAN and LSTM models, respectively. Last, the predicted values of each component are integrated to obtain the final prediction results. The developed model uses data decomposition technology, together with advanced machine learning (ML) and deep learning (DL) models to identify the appropriate dependencies and network topology. The experiments show that compared with many traditional prediction methods and decomposition-integration models, the developed model can produce accurate solar output prediction results under different evaluation criteria. Compared to the suboptimal model, the MAEs, MAPEs, and RMSEs of the four seasons decreased by 3.51%, 6.11%, and 2.25%, respectively.
Keywords: time-series forecasting; deep learning; signal decomposition; solar irradiance
Renewable energy and green infrastructure are widely considered to be the most critical factors in promoting the sustainable development of intelligent cities, both in the current and in the coming decades [[
According to the existing literature, scholars have proposed many data-driven prediction methods to predict solar irradiance, which can be roughly divided into two categories: single and mixed models [[
As an effective and robust method, the decomposition-integrated deep learning framework has attracted attention in various fields and has been applied in diverse areas, including wind speeds [[
Although some studies have confirmed that the hybrid model based on mode-decomposition technology can improve prediction accuracy, some areas still need to be studied and improved. First, considering the combination with mode-decomposition technology, many researchers often use traditional DL prediction models and rarely consider using newer DL prediction models. The second is that most studies only use one model to predict the decomposition subsequences after data decomposition and lack consideration of the spectral differences between the decomposition sequences, that is, the diversity and adaptability of high- and low-frequency data matching the prediction model.
To solve the above problems, this paper combines the methods of previous studies to propose a new decomposition-integrated deep learning framework for solar irradiance prediction, which integrates modern machine learning techniques, including complete ensemble empirical mode decomposition with adaptive noise (CEEMDAN), a Wasserstein GAN (WGAN), and a long short-term memory (LSTM) network. First, the CEEMDAN is leveraged to decompose the original data into several inherent mode functions (IMF) and residual components, dividing the obtained IMFs into high- and low-frequency subsequences for subsequent model prediction. Second, unlike previous similar studies that typically use a single predictor for all IMFs, we consider the spectrum characteristics of the decomposed subsequences in selecting the prediction models. We use a dual channel network structure, the stacked LSTM architecture to predict low-frequency IMFs, and a new stacked WGAN, using a stacked GRU as its generator and an MIP as a discriminator to predict high-frequency IMFs, thus obtaining better modeling. Finally, we combine all the individual results to obtain the final prediction of solar irradiance in the next time step. To sum up, the contributions of this paper are as follows:
- We propose a new decomposition-integrated deep learning framework for solar irradiance forecasting by integrating modern machine learning methods, including CEEMDAN, WGAN, and LSTM.
- The irradiance data were decomposed into high- and low-frequency subsequences using CEEMDAN. We predict high-frequency IMFs using the newly customized WGAN and low-frequency IMFs using the stacked LSTM.
- The proposed model was compared with multiple decomposition integration model and single model structures and the experimental results show that our model has better prediction performance.
Accurate and reliable solar irradiance prediction can significantly benefit the management of power generation and distribution of modern intelligent grids. However, instability, intermittency, and randomness make it challenging to accurately predict solar irradiance. The existing solar correlation prediction methods can be roughly divided into physical-driven and data-driven models based on the mathematical principles adopted [[
The currently popular GAN can effectively overcome the problem of error accumulation in the prediction field, thereby improving prediction accuracy. A GAN has essential applications in capturing the implicit relationships among complex nonlinear sequence data. In [[
The single model structure also has many shortcomings in processing complex data features, including difficulty in determining super parameters, resource consumption, and slow computing speed [[
In addition to the above-mentioned studies, numerous scholars have devoted their research efforts to studying solar irradiance prediction through the decomposition-integration method. This article follows suit by combining CEEMDAN, WGAN, and LSTM into the decomposition-integration method for further exploration and verification.
The general flow chart of the short-term solar irradiance prediction framework proposed in this paper is shown in Figure 1. CEEMDAN is used to decompose and preprocess the irradiance data, and then the decomposed subsequences are divided into high- and low-frequency subsequences. The high-frequency subsets are input to the WGAN and the low-frequency subsets are input to the LSTM network for prediction. Finally, the prediction results are accumulated. This chapter introduces the principles of the relevant models in detail.
Complete ensemble empirical mode decomposition with adaptive noise (CEEMDAN) was developed from EMD, EEMD, and CEEMD. The EMD-based method can adaptively decompose the original data into several IMFs and residuals with different frequencies and scales. They are applied to nonlinear and non-stationary time-series data and are widely used in various fields, including satellite signal analysis, speech recognition, economic data prediction, image processing, etc. EMD has excellent advantages in dealing with non-stationary and nonlinear signals but there is still a problem of "mode mixing." Mode mixing refers to similar oscillations in different modes or amplitudes in one mode. The EEMD algorithm eliminates the mode mixing in the EMD algorithm by adding white Gaussian noise to the signal [[
Here, we define the k-th component obtained from EMD as an operator
- Break down each
1 ) - Calculate the first residual: (
2 ) - Use the decomposition residuals, and the second mode is (
3 ) - Repeat the above steps for each IMF until the residual is obtained. The final residual can be expressed as: (
4 )
where K is the total number of IMFs, and together, the IMFs constitute the characteristics of the original signals on different time scales. The residual clearly shows the trend of the original sequence, which is smoother and the prediction error has been effectively reduced.
Sepp Hochreiter and Jürgen Schmidhuber [[
- Forget gate. We use the forget gate to determine the information that needed to be filtered out in each IMF component after decomposition. We use the sigmoid function to determine whether to filter out the current input and previous state. The formula is shown in Equation (
5 ): (5 ) - Input gate. The sigmoid function determines which input information to retain and the
6 )–(8 ): (6 )7 )8 ) - Output gate. First, the output part of the unit is determined by the sigmoid function, and then the predicted value points of the model are obtained by multiplying the unit state by the output parts of the
9 ) and (10 ): (9 )10 )
In 2014, Goodfellow et al. [[
The traditional GAN model optimizes the training parameters using Jensen Shannon (JS) divergence. The input data of its generator are Gaussian noise and the loss function can be expressed as:
(
where
(
where
(
However, the traditional GAN model uses JS divergence as the model's loss function, which leads to gradient disappearance and pattern collapse, resulting in unsatisfactory data generated by the generator [[
- The sigmoid activation function of the discriminator output layer is no longer used.
- The loss function is no longer logarithmic.
- The value of the discriminator gradient update is controlled between [
- Non-momentum-based algorithms such as random gradient descent are used.
Under ideal conditions, Wasserstein distance
(
where
Based on the above theoretical basis, this study constructs a deep hybrid prediction model, namely CEE-W-L, which combines CEEMDAN, WGAN, and LSTM. The overall flow chart of the hybrid CEE-W-L prediction model proposed in this paper is shown in Figure 4. In this section, we introduce the specific process of data training in the model:
- (
1 ) The single variable solar irradiance data (GHI) are used as input; - (
2 ) The maximum and minimum methods are used to normalize the data; - (
3 ) CEEMDAN is used to decompose the data sequence into several subsequence IMFs. We divide the first half of the IMFs into high-frequency subsequences and the second half into low-frequency subsequences; - (
4 ) The data of each subsequence are divided into a training set and a test set. The training set is used for model training and the test set is used to generate the model prediction results after training; - (
5 ) The divided data are input to the built prediction model for training. The high-frequency subsequences are input to the WGAN for prediction and the low-frequency subsequences are input to the LSTM; - (
6 ) The final prediction results are obtained by combining all of the prediction results obtained from the high- and low-frequency series and using inverse normalization. - (
7 ) Three evaluation indicators (MAE, MAPE, RMSE) are used to evaluate the prediction results.
The dataset used in this study was measured by the National Solar Radiation Data Base (NSRDB) [[
In order to eliminate the dimensional impact of solar irradiance data dimensions and improve the prediction model's calculation speed and prediction accuracy, we normalized the data using the minimum and maximum methods. This article uses the MinMaxScaler function in the Scikit learn 0.24.1 module with Formula (
(
This experiment selected four evaluation indicators, namely the Mean Absolute Error (MAE), Mean Absolute Percentage Error (MAPE), Root Mean Square Error (RMSE), and R-Squared (R2), to evaluate the performance of the prediction model. These four error indicators were selected not only because they are commonly used indicators in time-series prediction research but also because the selection of these evaluation indicators is more symbolic and persuasive, making the results of this article more acceptable in our view. The calculation formulas of the four evaluation criteria are as follows:
(
(
(
(
where
The hardware configuration of the experimental server was as follows: a 12th Gen Intel (R) Core (TM) i7-12700H 2.70 GHz processor, an NVIDIA GeForce RTX3050 Laptop GPU, and a win10 operating system. The relevant version numbers required by the program are Python 3.8, Tensorflow 2.3.0, Torch 1.12.1, and Torch Vision 0.13.0. The parameters we set for all the models are listed in Table 1, Table 2, Table 3 and Table 4.
In order to verify the prediction ability of the mixed prediction model proposed in this paper on the assessment of seasonal irradiance, we forecast the four seasonal irradiances in the same region on the same experimental platform using the CEE-W-L model and eight other prediction models. The models included a single RNN model, a single GRU model, a single Transformer model, a single LSTM model, a single WGAN model, a mixed CEEMDAN-LSTM model, a mixed CEEMDAN-WGAN model, and a mixed CEEMDAN-LSTM-WGAN model. The CEEMDAN-LSTM-WGAN model refers to the proposed model, where the subsequence is decomposed by CEEMDAN, the high-frequency sequence is predicted by LSTM, and the low-frequency sequence is predicted by WGAN, herein referred to as CEE-L-W)
Figure 5 shows the solar irradiance data prediction curve of a randomly selected day from the test dataset (subgraphs represent the prediction results of spring, summer, autumn, and winter, respectively). The red line in the figure represents the final prediction results of the model proposed in this paper, the black line represents the real data curve, and the other colored lines represent the prediction results of each comparison model. It can be seen from the trend of the curve that solar irradiance changes rapidly, and the data contain too much high- and low-frequency noise information. Under constantly changing irradiance, it is difficult for a single model to capture the changing trend, and the curve deviates from the true value curve to a large extent. The mixed model based on CEEMDAN decomposition can achieve relatively good prediction results in this case. It can be seen in the prediction fitting diagram that the proposed CEE-W-L model has a better learning ability for the various fluctuations of the GHI data in the four seasons. It is consistent with the actual data on the overall trend and has an accurate prediction effect under the peak, dramatic changes, and fluctuations of the data.
In addition, Table 5 shows a quantitative comparison of the four evaluation indicators for the proposed CEE-W-L model and the eight models mentioned above. From the data shown in the table, it is evident that the overall performance of the four decomposition-integration models (CEE-WGAN, CEE-LSTM, CEE-L-W, CEE-L-W) was better than that of the five single models (RNN, GRU, LSTM, Transformer, WGAN), which shows that the data decomposed by CEEMDAN were more conducive to function fitting and subsequent model convergence. Furthermore, to more intuitively verify the prediction performance of our proposed model, we converted the quantitative evaluation results (MAE, MAPE, RMSE) of the four decomposition-integration models with better performance into a bar chart. As shown in Figure 6, the overall errors of the CEE-W-L and CEE-L-W models were lower than those of the CEE-LSTM and CEE-WGAN models, which further validates the division of the decomposed subsequence into high and low frequencies for dual-channel prediction. In order to verify the above conclusions, we used a Diebold–Mariano (DM) test, and the results are shown in Table 6. In the case of a significant difference, if the DM value is negative, it means that the performance of model 1 was better than that of model 2, and vice versa if the DM value is positive. According to the results in Table 6, the performance of CEE-L-W was significantly superior to that of CEE-WGAN and CEE-LSTM. CEE-W-L performed significantly better than CEE-L-W. Therefore, passing high-frequency signals through the WGAN and low-frequency signals through the LSTM for the four seasons achieved better performance than passing high-frequency signals through the LSTM and low-frequency signals through the WGAN. Compared to the suboptimal model, our proposed model's MAE, MAPE, and RMSE values decreased by 3.51%, 6.11%, and 2.25%, respectively, for the four seasons. These results sufficiently reflect the advantages of our model.
To sum up, the CEE-W-L method proposed in this paper can achieve the most accurate prediction based on the experimental results shown in the graphs and tables. In the four different statistical verification indicators, the CEE-W-L model outperformed the other benchmark models. Compared with the single LSTM, RNN, GRU, Transformer, WGAN, mixed CEE-LSTM, CEE-WGAN, and CEE-L-W models, we can see that the CEEMDAN-decomposed data were more conducive to function fitting and subsequent model convergence. Compared with the CEEM-LSTM, CEEM-WGAN, and CEE-L-W models, especially after focusing on the CEE-L-W model, we can further conclude that the CEE-W-L model can better preserve and retain the spatiotemporal characteristics of images.
This paper leverages historical data to forecast future patterns of solar irradiance data. The NSRDB public solar irradiance data are used. Based on the study of the data properties and DL models, this paper proposes a combined forecasting model for time-series prediction using data analysis methods and neural networks, which combines the CEEMDAN decomposition model with the WGAN and LSTM prediction models. CEEMDAN is used to decompose the original single variable solar irradiance dataset. Single-column GHI data are converted into multiple subsequence signals and residual signals. Next, the obtained subsequences are divided into high- and low-frequency subsequences, and each subsequence is divided into a training set and a test set for use in the subsequent prediction model. We pass the high-frequency class through the WGAN and the low-frequency class through the LSTM. Then, the prediction results of each subsequence are accumulated to produce the final prediction results. Finally, the prediction efficiency of the CEE-W-L model is evaluated using a fitting curve, error index, and DM test.
According to the experimental results, the proposed method can accurately predict the changes in the real irradiance data, reduce the fitting error to a lower value, and is significantly superior to the other models. In addition, the model in this paper has better and more robust performance than other prediction models and shows superior performance in the four seasons. The prediction methods in this paper are summarized and the following conclusions are drawn. First, for complex, non-stationary data, the waveform decomposition strategy effectively reduces data complexity. Second, the decomposed data can be combined with the updated prediction model. The WGAN has excellent feature extraction ability and can still achieve sound feature extraction efficiency in predicting time-series data with high volatility. Finally, considering the spectral characteristics of the decomposed subsequences, the WGAN has a greater ability to predict high-frequency signals than the LSTM, and the LSTM has higher accuracy and faster efficiency in predicting data with low complexity. We divide their work and use them in cooperation to obtain better modeling. The new proposed prediction framework can effectively support the deployment of photovoltaic power generation systems, which is crucial to developing intelligent grid systems.
The model in this paper uses multiple attempts to select the model parameters, which has limitations in parameter optimization. Future works will include optimizing the model parameters to improve the applicability of the proposed model.
DIAGRAM: Figure 1 The proposed CEE-W-L overall framework flow chart. (a) Flow chart of the signal decomposition module. (b) High- and low-frequency characteristic arrangement of GHI data after decomposition. (c) Detailed diagram of two-channel neural network prediction.
DIAGRAM: Figure 2 The schematic diagram of the working process of the LSTM neural network.
DIAGRAM: Figure 3 The schematic diagram of the working process of the WGAN neural network.
Graph: Figure 4 The overall flow chart of all proposed GHI prediction frameworks.
Graph: Figure 5 The comparison chart of the data of the four seasons using the different methods. The four graphs in the figure represent spring, summer, autumn, and winter, respectively.
Graph: Figure 6 (a) Comparison of the MAE of the proposed model and that of the other decomposition-integration models. (b) Comparison of the MAPE of the proposed model and that of the other decomposition-integration models. (c) Comparison of the RMSE of the proposed model and that of the other decomposition-integration models.
Table 1 Required parameters in RNN and GRU.
Parameters Descriptions Values Ahead-num How long are the historical data used to predict the future 10 Num-layers Number of stack layers 3 Optimizer Minimizes the loss function by training and optimizing the parameters RMSProp Epoch Indicates how many forward calculations and back propagations have been completed 100
Table 2 Required parameters in Transformer.
Parameters Descriptions Values Num-layers Number of stack layers 3 Learning-rate Controls the speed at which we adjust the weights of the neural network based on the loss gradient 0.001 Batch size Number of data passed to the program for training at a time 32 Optimizer Minimizes the loss function by training and optimizing the parameters RMSProp Drop-out In the process of training batches, the phenomenon of overfitting can be significantly reduced by ignoring the general feature detector 0.1
Table 3 Required parameters in LSTM.
Parameters Descriptions Values Num-layers Number of stack layers 2 Hidden-size Number of features in the hidden state 4 Batch size Number of data passed to the program for training at a time 64 Optimizer Minimizes the loss function by training and optimizing the parameters Adam Epoch Indicates how many forward calculations and back propagations have been completed 100 Learning-rate Controls the speed at which we adjust the weights of the neural network based on the loss gradient 0.005
Table 4 Required parameters in WGAN.
Parameters Descriptions Values Time steps How long are the historical data used to predict the future 10 Forecast How long into the future to predict 1 Hidden-size The number of features in the hidden size 128 Learning-rate Controls the speed at which we adjust the weights of the neural network based on the loss gradient Batch size Number of data passed to the program for training at a time 64 Optimizer Minimizes the loss function by training and optimizing the parameters RMSProp Epoch Indicates how many forward calculations and back propagations have been completed 300 Clipping-value After each update of the discriminator parameters, it truncates their absolute values to no more than a fixed constant 0.001 N-critic Number of iterations of the critic per generator iteration 5
Table 5 Quantitative comparison of the four evaluation indicators of the proposed model and eight comparison models for four seasons.
RNN GRU LSTM WGAN Transformer CEE-LSTM CEE-WGAN CEE-L-W CEE-W-L Spring MAE 38.89 38.34 40.10 40.70 40.85 34.43 47.20 32.97 MAPE 31.56 34.26 32.31 32.63 26.35 23.67 25.96 22.71 RMSE 61.43 58.70 51.01 59.57 51.00 49.85 49.60 40.33 R2 0.956 0.959 0.957 0.942 0.956 0.97 0.949 0.958 Summer MAE 28.93 26.09 28.91 37.11 28.16 23.20 30.38 22.01 MAPE 20.79 19.49 23.21 23.89 18.36 17.51 20.23 16.77 RMSE 47.25 44.32 46.48 54.29 45.70 32.52 47.82 31.44 R2 0.979 0.981 0.983 0.975 0.982 0.99 0.977 0.986 Autumn MAE 50.45 40.54 39.00 51.80 49.17 50.48 46.04 34.40 MAPE 30.37 28.68 29.64 36.71 27.49 33.38 24.16 19.93 RMSE 56.13 66.89 67.17 66.82 64.09 64.86 62.25 40.15 R2 0.923 0.94 0.941 0.93 0.928 0.926 0.932 0.935 Winter MAE 34.32 34.61 34.41 31.16 24.20 25.96 34.17 29.65 MAPE 50.08 41.97 45.04 42.08 35.25 35.62 36.38 32.51 RMSE 40.77 44.52 40.77 43.94 38.40 36.52 40.79 36.73 R2 0.953 0.95 0.955 0.952 0.96 0.964 0.943 0.946
Table 6 Results of DM test.
Model 1 Model 2 DM Value CEE-L-W CEE-WGAN −9.85 0.0000 CEE-LSTM −3.192 0.0001 CEE-W-L CEE-WGAN −11.67 0.0000 CEE-LSTM −5.64 0.0000 CEE-L-W −5.37 0.0000
Conceptualization, K.Y.; methodology, K.Y.; software, Q.L.; validation, K.Y.; formal analysis, K.Y.; investigation, D.Z.; resources, D.Z.; data curation, D.Z.; writing—original draft preparation, Q.L.; writing—review and editing, K.Y.; visualization, Q.L.; supervision, K.Y. and D.Z.; project administration, K.Y. and D.Z.; funding acquisition, K.Y. and D.Z. All authors have read and agreed to the published version of the manuscript.
The source code and required data sets of the experiments can be obtained upon requests.
The authors declare no conflict of interest.
The following abbreviations are used in this manuscript:
CEEMDAN Complete ensemble empirical mode decomposition with adaptive noise WGAN Wasserstein generative adversarial network LSTM Long short-term memory network ML Machine learning DL Deep learning EMD Empirical mode decomposition VMD Variational mode decomposition ANN Artificial neural network CNN Convolutional neural network GRU Gated recursive unit IMF Inherent mode functions NSRDB National Solar Radiation Database MAE Mean Absolute Error MAPE Mean Absolute Percentage Error RMSE Root Mean Square Error R2 R-Square DM Diebold–Mariano
By Qianqian Li; Dongping Zhang and Ke Yan
Reported by Author; Author; Author