Zum Hauptinhalt springen

A Solar Irradiance Forecasting Framework Based on the CEE-WGAN-LSTM Model

Li, Qianqian ; Zhang, Dongping ; et al.
In: Sensors, Jg. 23 (2023-03-01), Heft 5, S. 2799-2799
Online academicJournal

A Solar Irradiance Forecasting Framework Based on the CEE-WGAN-LSTM Model 

With the rapid development of solar energy plants in recent years, the accurate prediction of solar power generation has become an important and challenging problem in modern intelligent grid systems. To improve the forecasting accuracy of solar energy generation, an effective and robust decomposition-integration method for two-channel solar irradiance forecasting is proposed in this study, which uses complete ensemble empirical mode decomposition with adaptive noise (CEEMDAN), a Wasserstein generative adversarial network (WGAN), and a long short-term memory network (LSTM). The proposed method consists of three essential stages. First, the solar output signal is divided into several relatively simple subsequences using the CEEMDAN method, which has noticeable frequency differences. Second, high and low-frequency subsequences are predicted using the WGAN and LSTM models, respectively. Last, the predicted values of each component are integrated to obtain the final prediction results. The developed model uses data decomposition technology, together with advanced machine learning (ML) and deep learning (DL) models to identify the appropriate dependencies and network topology. The experiments show that compared with many traditional prediction methods and decomposition-integration models, the developed model can produce accurate solar output prediction results under different evaluation criteria. Compared to the suboptimal model, the MAEs, MAPEs, and RMSEs of the four seasons decreased by 3.51%, 6.11%, and 2.25%, respectively.

Keywords: time-series forecasting; deep learning; signal decomposition; solar irradiance

1. Introduction

Renewable energy and green infrastructure are widely considered to be the most critical factors in promoting the sustainable development of intelligent cities, both in the current and in the coming decades [[1]]. Solar energy is experiencing unprecedented development as a sustainable renewable energy, and climate change and environmental pollution are of serious concern [[3], [5]]. Meanwhile, electricity conversion technology has significantly reduced investment costs, thus greatly improving the market competitiveness of solar energy. According to the literature, in the past ten years, the cost of solar photovoltaic power has decreased by 85%. The average power of PV modules has increased from 250–300 W in 2010 to 400–550 W in 2020 and is expected to reach 800–1200 W by 2030 [[6]]. The continuous improvement of PV module power depends on the continuous development of PV technology. Unlike other manageable energy sources, solar energy output has the characteristics of high nonlinearity, strong randomness, and large intermittency because solar radiation intensity is affected by various complex factors. Iqbal et al. [[7]] found that the environment's dew point temperature and humidity are inversely proportional to the solar radiation intensity. Mustafa et al. [[8]] found that many environmental factors (such as dust accumulation, water droplets, bird droppings, and partial shade conditions) have a significant impact on the performance of photovoltaic systems. Therefore, the accurate and reliable prediction of solar irradiance is crucial for promoting large-scale integration of solar energy and improving the operating efficiency of power systems with multiple energy sources [[9]].

According to the existing literature, scholars have proposed many data-driven prediction methods to predict solar irradiance, which can be roughly divided into two categories: single and mixed models [[10]]. The single-model prediction methods include traditional statistical, classical machine learning, and deep learning methods [[11]]. The research identifies an issue with lag in the prediction task of a single model, which cannot sufficiently reflect the randomness of photovoltaic fluctuations. Mixed-model prediction methods usually combine multiple models to solve the limitations of independent models. Using the characteristics of multiple models to improve performance or combining these methods with feature engineering can solve the problem of poor prediction accuracy of a single model [[12]]. One of the leading hybrid-model prediction methods is the decomposition-integration model.

As an effective and robust method, the decomposition-integrated deep learning framework has attracted attention in various fields and has been applied in diverse areas, including wind speeds [[14]], photovoltaics [[15]], coal prices [[16]], agribusiness [[17]], and electricity prices [[18]]. On one hand, mode-decomposition technology decomposes irradiance data into several frequency subsequences, which reduces the impact of noise on prediction accuracy and makes it easier for prediction models to capture changing modes [[19]]. Empirical mode decomposition (EMD), variational mode decomposition (VMD), and wavelet decomposition (WTD) are among the methods most commonly used by scholars for the decomposition preprocessing of photovoltaic data. On the other hand, the aim of the prediction model is to apply an efficient and simple model to predict the decomposition subsequences. Scholars typically adopt common DL models, such as artificial neural networks (ANNs), convolutional neural networks (CNNs), LSTM, gated recursive units (GRUs), and generative adversarial networks (GANs) [[20]]. GANs have become an important branch of the field of DL with their unique confrontation ideas and unsupervised learning methods. Although GANs have been successfully applied to computer vision and achieved good results, they are rarely used in time-series predictions. Because of the excellent ability of GANs to capture the underlying relationship between data, this paper is committed to applying GANs to photovoltaic time-series predictions [[22]].

Although some studies have confirmed that the hybrid model based on mode-decomposition technology can improve prediction accuracy, some areas still need to be studied and improved. First, considering the combination with mode-decomposition technology, many researchers often use traditional DL prediction models and rarely consider using newer DL prediction models. The second is that most studies only use one model to predict the decomposition subsequences after data decomposition and lack consideration of the spectral differences between the decomposition sequences, that is, the diversity and adaptability of high- and low-frequency data matching the prediction model.

To solve the above problems, this paper combines the methods of previous studies to propose a new decomposition-integrated deep learning framework for solar irradiance prediction, which integrates modern machine learning techniques, including complete ensemble empirical mode decomposition with adaptive noise (CEEMDAN), a Wasserstein GAN (WGAN), and a long short-term memory (LSTM) network. First, the CEEMDAN is leveraged to decompose the original data into several inherent mode functions (IMF) and residual components, dividing the obtained IMFs into high- and low-frequency subsequences for subsequent model prediction. Second, unlike previous similar studies that typically use a single predictor for all IMFs, we consider the spectrum characteristics of the decomposed subsequences in selecting the prediction models. We use a dual channel network structure, the stacked LSTM architecture to predict low-frequency IMFs, and a new stacked WGAN, using a stacked GRU as its generator and an MIP as a discriminator to predict high-frequency IMFs, thus obtaining better modeling. Finally, we combine all the individual results to obtain the final prediction of solar irradiance in the next time step. To sum up, the contributions of this paper are as follows:

  • We propose a new decomposition-integrated deep learning framework for solar irradiance forecasting by integrating modern machine learning methods, including CEEMDAN, WGAN, and LSTM.
  • The irradiance data were decomposed into high- and low-frequency subsequences using CEEMDAN. We predict high-frequency IMFs using the newly customized WGAN and low-frequency IMFs using the stacked LSTM.
  • The proposed model was compared with multiple decomposition integration model and single model structures and the experimental results show that our model has better prediction performance.
2. Related Works

Accurate and reliable solar irradiance prediction can significantly benefit the management of power generation and distribution of modern intelligent grids. However, instability, intermittency, and randomness make it challenging to accurately predict solar irradiance. The existing solar correlation prediction methods can be roughly divided into physical-driven and data-driven models based on the mathematical principles adopted [[23]]. The physical-driven models can use remote sensing and meteorological information to predict future solar power generation. However, when the execution time and computational burden of completing the simulation are unsatisfactory, obtaining the necessary input of physical-driven models on time may be challenging [[24]]. Data-driven models include traditional statistical, emerging machine learning, and deep learning models. They can improve prediction accuracy by finding the internal relationships between data and are the most commonly used technologies in photovoltaic power prediction research. Currently, the most popular single model structures include convolutional neural networks (CNN), LSTM, GAN, GRU, etc. [[25]].

The currently popular GAN can effectively overcome the problem of error accumulation in the prediction field, thereby improving prediction accuracy. A GAN has essential applications in capturing the implicit relationships among complex nonlinear sequence data. In [[26]], a new GAN model for wind power generation forecasting was proposed and the accuracy of the WGAN-GP model for forecasting was verified. Wang et al. [[27]] used a GAN to enhance the training datasets of each weather type. Finally, more accurate results were obtained by training the CNN on the enhanced dataset. However, training the GAN model to achieve Nash equilibrium is challenging. It has problems such as slow convergence, mode collapse, low sample quality, and an unstable training process [[28]].

The single model structure also has many shortcomings in processing complex data features, including difficulty in determining super parameters, resource consumption, and slow computing speed [[29]]. Therefore, scholars focus on hybrid models to improve the accuracy of DL networks [[30]]. Hybrid models combine multiple models to solve the limitations of single models and use the characteristics of multiple models to improve performance. A mainstream hybrid model is a combination of data decomposition methods and prediction models. Signal decomposition is an essential preprocessing step for building these hybrid models. EMD, VMD, and WTD are typically used for irradiance prediction or radiation prediction. Pi et al. [[31]] proposed a model based on WTD and a CNN and combined it with the attention LSTM model for the prediction of solar irradiance. Jin et al. [[32]] proposed a new hybrid artificial intelligence-enhanced prediction model for energy consumption time-series prediction, which combined singular spectrum analysis (SSA) and parallel LSTM neural networks. Zeng et al. [[33]] proposed a PM2.5 air-quality prediction model that combined an extended stationary wavelet transform (ESWT) and nested LSTM neural networks. Lan et al. [[34]] used the EEMD algorithm to preprocess the solar irradiance data and self-organizing mapping and back-propagation neural network models to predict the day irradiance. The experimental results showed that this decomposition method can improve prediction performance. Mohammadi et al. [[35]] proposed combining SVM and WD algorithms to predict solar irradiance. The SVM was used to predict the subsequences of WD decomposition. A hybrid model based on a multi-branch hybrid structure consisting of a 1D convolution and LSTM was used for irradiance prediction. The model used WPD to decompose irradiance and predict subsequences.

In addition to the above-mentioned studies, numerous scholars have devoted their research efforts to studying solar irradiance prediction through the decomposition-integration method. This article follows suit by combining CEEMDAN, WGAN, and LSTM into the decomposition-integration method for further exploration and verification.

3. Methodology

The general flow chart of the short-term solar irradiance prediction framework proposed in this paper is shown in Figure 1. CEEMDAN is used to decompose and preprocess the irradiance data, and then the decomposed subsequences are divided into high- and low-frequency subsequences. The high-frequency subsets are input to the WGAN and the low-frequency subsets are input to the LSTM network for prediction. Finally, the prediction results are accumulated. This chapter introduces the principles of the relevant models in detail.

3.1. CEEMDAN

Complete ensemble empirical mode decomposition with adaptive noise (CEEMDAN) was developed from EMD, EEMD, and CEEMD. The EMD-based method can adaptively decompose the original data into several IMFs and residuals with different frequencies and scales. They are applied to nonlinear and non-stationary time-series data and are widely used in various fields, including satellite signal analysis, speech recognition, economic data prediction, image processing, etc. EMD has excellent advantages in dealing with non-stationary and nonlinear signals but there is still a problem of "mode mixing." Mode mixing refers to similar oscillations in different modes or amplitudes in one mode. The EEMD algorithm eliminates the mode mixing in the EMD algorithm by adding white Gaussian noise to the signal [[36]]. However, the EEMD algorithm cannot eliminate the Gaussian white noise after signal reconstruction, which causes reconstruction errors. In order to overcome the above problems, Flandrin et al. proposed an advanced signal decomposition technology called CEEMDAN [[37]]. It can eliminate the mode mixing more effectively, produce a reconstruction error of almost zero, and significantly reduce the calculation costs.

Here, we define the k-th component obtained from EMD as an operator Ek() and let ωi(t) be the white noise with normal distribution N (0,1). Figure 1a shows the flow chart of the CEEMDAN decomposition and Figure 1b shows the decomposition results of CEEMDAN on the original solar irradiance sequence. The CEEMDAN algorithm steps are as follows:

  • Break down each Xi(t)=X(t)+ε0ωi(t) using EMD. Then, extract the first IMF, where ε0 is the noise coefficient, i=1,2,...,I , and define the first mode as (1) IMF1(t)=1Ii=1IIMF1i(t)
  • Calculate the first residual: (2) r1(t)=X(t)IMF1(t)
  • Use the decomposition residuals, and the second mode is (3) IMF2(t)=1Ii=1IE1(r1(t)+ε1E1(wi(t)))
  • Repeat the above steps for each IMF until the residual is obtained. The final residual can be expressed as: (4) Rk(t)=X(t)k=1KIMFk(t)

where K is the total number of IMFs, and together, the IMFs constitute the characteristics of the original signals on different time scales. The residual clearly shows the trend of the original sequence, which is smoother and the prediction error has been effectively reduced.

3.2. LSTM

Sepp Hochreiter and Jürgen Schmidhuber [[38]] first proposed the long short-term memory network (LSTM), which was carefully designed. Based on an RNN, an input structure is added to each cell state to control the retention of information, thereby addressing the issue of long sequence learning in RNNs. An LSTM network can maintain the long-term memory of neural networks and the model can also be used for solar irradiance prediction. Each cell of an LSTM network has three parts: the forget gate, input gate, and output gate, which, respectively, determine the filtering, preservation, and generation of information. We applied them to predict the low-frequency irradiance-like subsequences following the CEEMDAN decomposition. The complete structure of the model is shown in Figure 2. The LSTM neural network contains multiple gates:

  • Forget gate. We use the forget gate to determine the information that needed to be filtered out in each IMF component after decomposition. We use the sigmoid function to determine whether to filter out the current input and previous state. The formula is shown in Equation (5): (5) ft=σ(Wf·[ht1,xt]+bf)
  • Input gate. The sigmoid function determines which input information to retain and the tanh updates some parts to become new values. Then, update Ct1 to Ct . The formulas are shown in Equations (6)–(8): (6) it=σ(Wi·[ht1,xt]+bi) (7) Ct˜=tanh(Wc·[ht1,xt+bc]) (8) Ct=ftCt1+itCt˜
  • Output gate. First, the output part of the unit is determined by the sigmoid function, and then the predicted value points of the model are obtained by multiplying the unit state by the output parts of the tanh and sigmoid gates. The formulas are shown in Equations (9) and (10): (9) ot=σ(Wo·[ht1,xt]+bo) (10) ht=ot[tanh(ct)]
3.3. WGAN

In 2014, Goodfellow et al. [[39]] proposed the generative adversarial network (GAN) as a breakthrough and a superb idea for machine learning. The GAN is a generative model that is different from a traditional generative model. It avoids learning based on Markov chains and can be trained based on implicit density. The GAN is composed of a generator and a discriminator. The generator generates sample data that conforms to the real data distribution. The discriminator accurately judges and classifies the input information. If the input data are real, the output is 1; If the input data are false, the output is 0. The training of the GAN is divided into two stages: training the discriminator and then training the generator. In training, the two models will constantly update their parameters to minimize their loss functions and output errors.

The traditional GAN model optimizes the training parameters using Jensen Shannon (JS) divergence. The input data of its generator are Gaussian noise and the loss function can be expressed as:

(11) LG=EzPg(z)[D(G(z))]

where LG is the loss function of the generator; E[·] is the expected function; G[·] is the generator function; D[·] is the discriminator function; Pg[·] is the noise data distribution; and Z is the input noise data vector.

(12) LD=(ExPr(x)[lgD(x)]+EzPg(z)[lg(1D(G(z)))])

where LD is the loss function of the discriminator and Pr[·] is the distribution of the target data x. From the above two loss functions, the objective function of the generator–discriminator game can be established, as shown in Equation (3).

(13) L(G,D)=ExPr(x)[lgD(x)]+EzPg(z)[lg(1D(G(z))]

However, the traditional GAN model uses JS divergence as the model's loss function, which leads to gradient disappearance and pattern collapse, resulting in unsatisfactory data generated by the generator [[40]]. Therefore, Arjovsky et al. [[40]] proposed the Wasserstein generative adversarial network (WGAN) to overcome the above problems. The WGAN uses Wasserstein distance to distinguish the gap between two distributions. When the gap between the two distributions is large, the generator can still be updated to overcome the training problem in the GAN, thereby improving the quality of data generated by the generator and the prediction accuracy. Compared to traditional generation countermeasure network models, WGANs mainly improve the following aspects:

  • The sigmoid activation function of the discriminator output layer is no longer used.
  • The loss function is no longer logarithmic.
  • The value of the discriminator gradient update is controlled between [ c , c].
  • Non-momentum-based algorithms such as random gradient descent are used.

Under ideal conditions, Wasserstein distance W(Pr,Pg) is continuously differentiable, as shown in Equation (14).

(14) W(Pr,Pg)=1KsupfK(ExPr(x)[f(x)]EzPr(z)[f(z)])

where sup[·] represents the supremum of the function value; K is the Lipschitz constant; X=G(z) is the generated data; and fK means that function f satisfies the K-Lipschitz continuity and function f can be fitted by the neural network. Therefore, this paper uses the WGAN model to train the high-frequency irradiance subsequences after CEEMDAN decomposition, and its model structure is shown in Figure 3. Our WGAN is newly customized. The generator and discriminator are composed of stacked GRUs and MLPs, which are prevalent network structures in DL. They can effectively overcome the gradient problem in neural networks, help stabilize the WGAN model, and improve the model's prediction performance.

3.4. CEEMDAN-WGAN-LSTM

Based on the above theoretical basis, this study constructs a deep hybrid prediction model, namely CEE-W-L, which combines CEEMDAN, WGAN, and LSTM. The overall flow chart of the hybrid CEE-W-L prediction model proposed in this paper is shown in Figure 4. In this section, we introduce the specific process of data training in the model:

  • (1) The single variable solar irradiance data (GHI) are used as input;
  • (2) The maximum and minimum methods are used to normalize the data;
  • (3) CEEMDAN is used to decompose the data sequence into several subsequence IMFs. We divide the first half of the IMFs into high-frequency subsequences and the second half into low-frequency subsequences;
  • (4) The data of each subsequence are divided into a training set and a test set. The training set is used for model training and the test set is used to generate the model prediction results after training;
  • (5) The divided data are input to the built prediction model for training. The high-frequency subsequences are input to the WGAN for prediction and the low-frequency subsequences are input to the LSTM;
  • (6) The final prediction results are obtained by combining all of the prediction results obtained from the high- and low-frequency series and using inverse normalization.
  • (7) Three evaluation indicators (MAE, MAPE, RMSE) are used to evaluate the prediction results.
4. Experimental

4.1. Data

The dataset used in this study was measured by the National Solar Radiation Data Base (NSRDB) [[41]], and the data can be accessed at the following link: https://nsrdb.nrel.gov/data-viewer (accessed on 1 November 2022). We collected the irradiance data of Texas for one year from this website. The measuring station is located at 118.31 degrees west longitude and 33.98 degrees north latitude. The climate in this area has four distinct seasons and the daily temperature difference is relatively significant. Therefore, the collected data also have prominent climatic characteristics. The irradiance in summer is relatively stable and the irradiance in winter changes violently. This dataset contains the local solar irradiance measurement data from 1 January to 31 December 2018. The time interval is 5 min, the collection length is 365 days, and there are 105,120 pieces of data. The original dataset was preprocessed and normalized before the experiment. Considering that there was no light at night, we carried out zero clearing. The intensity of solar radiation is different throughout the year, so it was necessary to divide the radiation data into four seasons for independent prediction to make more accurate predictions. We divided the data of each season (90 days, including about 12,960 data samples) into training sets (54 days, including about 7776 data samples), verification sets (18 days, including about 2592 data samples), and test sets (18 days, including about 2592 data samples) in a 3:1:1 ratio.

4.2. Experimental Setup

4.2.1. Standardization

In order to eliminate the dimensional impact of solar irradiance data dimensions and improve the prediction model's calculation speed and prediction accuracy, we normalized the data using the minimum and maximum methods. This article uses the MinMaxScaler function in the Scikit learn 0.24.1 module with Formula (15) to normalize the data, where X represents the normalized data; X represents the sample data; and Xmax and Xmin are the maximum and minimum values of the dataset.

(15) X=XXminXmaxXmin

4.2.2. Evaluation Indicators

This experiment selected four evaluation indicators, namely the Mean Absolute Error (MAE), Mean Absolute Percentage Error (MAPE), Root Mean Square Error (RMSE), and R-Squared (R2), to evaluate the performance of the prediction model. These four error indicators were selected not only because they are commonly used indicators in time-series prediction research but also because the selection of these evaluation indicators is more symbolic and persuasive, making the results of this article more acceptable in our view. The calculation formulas of the four evaluation criteria are as follows:

(16) MAE(y,y^)=1ni=1nyiyi^

(17) MAPE(y,y^)=100%ni=1nyiyi^yi

(18) RMSE(y,y^)=1ni=1n(yiyi^)2

(19) R2(y,y^)=1i=1n(yiyi^)2i=1n(yi¯yi^)2

where yi is the result predicted by the model; yi^ is the actual test sample value; and n is the total number of test samples. These evaluation indicators are encapsulated in the sk-learn package. They can be called directly, thereby saving time and ensuring that the experiment is carried out more quickly and efficiently.

4.2.3. Experimental Device

The hardware configuration of the experimental server was as follows: a 12th Gen Intel (R) Core (TM) i7-12700H 2.70 GHz processor, an NVIDIA GeForce RTX3050 Laptop GPU, and a win10 operating system. The relevant version numbers required by the program are Python 3.8, Tensorflow 2.3.0, Torch 1.12.1, and Torch Vision 0.13.0. The parameters we set for all the models are listed in Table 1, Table 2, Table 3 and Table 4.

4.3. Experimental Analysis

In order to verify the prediction ability of the mixed prediction model proposed in this paper on the assessment of seasonal irradiance, we forecast the four seasonal irradiances in the same region on the same experimental platform using the CEE-W-L model and eight other prediction models. The models included a single RNN model, a single GRU model, a single Transformer model, a single LSTM model, a single WGAN model, a mixed CEEMDAN-LSTM model, a mixed CEEMDAN-WGAN model, and a mixed CEEMDAN-LSTM-WGAN model. The CEEMDAN-LSTM-WGAN model refers to the proposed model, where the subsequence is decomposed by CEEMDAN, the high-frequency sequence is predicted by LSTM, and the low-frequency sequence is predicted by WGAN, herein referred to as CEE-L-W)

Figure 5 shows the solar irradiance data prediction curve of a randomly selected day from the test dataset (subgraphs represent the prediction results of spring, summer, autumn, and winter, respectively). The red line in the figure represents the final prediction results of the model proposed in this paper, the black line represents the real data curve, and the other colored lines represent the prediction results of each comparison model. It can be seen from the trend of the curve that solar irradiance changes rapidly, and the data contain too much high- and low-frequency noise information. Under constantly changing irradiance, it is difficult for a single model to capture the changing trend, and the curve deviates from the true value curve to a large extent. The mixed model based on CEEMDAN decomposition can achieve relatively good prediction results in this case. It can be seen in the prediction fitting diagram that the proposed CEE-W-L model has a better learning ability for the various fluctuations of the GHI data in the four seasons. It is consistent with the actual data on the overall trend and has an accurate prediction effect under the peak, dramatic changes, and fluctuations of the data.

In addition, Table 5 shows a quantitative comparison of the four evaluation indicators for the proposed CEE-W-L model and the eight models mentioned above. From the data shown in the table, it is evident that the overall performance of the four decomposition-integration models (CEE-WGAN, CEE-LSTM, CEE-L-W, CEE-L-W) was better than that of the five single models (RNN, GRU, LSTM, Transformer, WGAN), which shows that the data decomposed by CEEMDAN were more conducive to function fitting and subsequent model convergence. Furthermore, to more intuitively verify the prediction performance of our proposed model, we converted the quantitative evaluation results (MAE, MAPE, RMSE) of the four decomposition-integration models with better performance into a bar chart. As shown in Figure 6, the overall errors of the CEE-W-L and CEE-L-W models were lower than those of the CEE-LSTM and CEE-WGAN models, which further validates the division of the decomposed subsequence into high and low frequencies for dual-channel prediction. In order to verify the above conclusions, we used a Diebold–Mariano (DM) test, and the results are shown in Table 6. In the case of a significant difference, if the DM value is negative, it means that the performance of model 1 was better than that of model 2, and vice versa if the DM value is positive. According to the results in Table 6, the performance of CEE-L-W was significantly superior to that of CEE-WGAN and CEE-LSTM. CEE-W-L performed significantly better than CEE-L-W. Therefore, passing high-frequency signals through the WGAN and low-frequency signals through the LSTM for the four seasons achieved better performance than passing high-frequency signals through the LSTM and low-frequency signals through the WGAN. Compared to the suboptimal model, our proposed model's MAE, MAPE, and RMSE values decreased by 3.51%, 6.11%, and 2.25%, respectively, for the four seasons. These results sufficiently reflect the advantages of our model.

To sum up, the CEE-W-L method proposed in this paper can achieve the most accurate prediction based on the experimental results shown in the graphs and tables. In the four different statistical verification indicators, the CEE-W-L model outperformed the other benchmark models. Compared with the single LSTM, RNN, GRU, Transformer, WGAN, mixed CEE-LSTM, CEE-WGAN, and CEE-L-W models, we can see that the CEEMDAN-decomposed data were more conducive to function fitting and subsequent model convergence. Compared with the CEEM-LSTM, CEEM-WGAN, and CEE-L-W models, especially after focusing on the CEE-L-W model, we can further conclude that the CEE-W-L model can better preserve and retain the spatiotemporal characteristics of images.

5. Conclusions and Discussion

This paper leverages historical data to forecast future patterns of solar irradiance data. The NSRDB public solar irradiance data are used. Based on the study of the data properties and DL models, this paper proposes a combined forecasting model for time-series prediction using data analysis methods and neural networks, which combines the CEEMDAN decomposition model with the WGAN and LSTM prediction models. CEEMDAN is used to decompose the original single variable solar irradiance dataset. Single-column GHI data are converted into multiple subsequence signals and residual signals. Next, the obtained subsequences are divided into high- and low-frequency subsequences, and each subsequence is divided into a training set and a test set for use in the subsequent prediction model. We pass the high-frequency class through the WGAN and the low-frequency class through the LSTM. Then, the prediction results of each subsequence are accumulated to produce the final prediction results. Finally, the prediction efficiency of the CEE-W-L model is evaluated using a fitting curve, error index, and DM test.

According to the experimental results, the proposed method can accurately predict the changes in the real irradiance data, reduce the fitting error to a lower value, and is significantly superior to the other models. In addition, the model in this paper has better and more robust performance than other prediction models and shows superior performance in the four seasons. The prediction methods in this paper are summarized and the following conclusions are drawn. First, for complex, non-stationary data, the waveform decomposition strategy effectively reduces data complexity. Second, the decomposed data can be combined with the updated prediction model. The WGAN has excellent feature extraction ability and can still achieve sound feature extraction efficiency in predicting time-series data with high volatility. Finally, considering the spectral characteristics of the decomposed subsequences, the WGAN has a greater ability to predict high-frequency signals than the LSTM, and the LSTM has higher accuracy and faster efficiency in predicting data with low complexity. We divide their work and use them in cooperation to obtain better modeling. The new proposed prediction framework can effectively support the deployment of photovoltaic power generation systems, which is crucial to developing intelligent grid systems.

The model in this paper uses multiple attempts to select the model parameters, which has limitations in parameter optimization. Future works will include optimizing the model parameters to improve the applicability of the proposed model.

Figures and Tables

DIAGRAM: Figure 1 The proposed CEE-W-L overall framework flow chart. (a) Flow chart of the signal decomposition module. (b) High- and low-frequency characteristic arrangement of GHI data after decomposition. (c) Detailed diagram of two-channel neural network prediction.

DIAGRAM: Figure 2 The schematic diagram of the working process of the LSTM neural network.

DIAGRAM: Figure 3 The schematic diagram of the working process of the WGAN neural network.

Graph: Figure 4 The overall flow chart of all proposed GHI prediction frameworks.

Graph: Figure 5 The comparison chart of the data of the four seasons using the different methods. The four graphs in the figure represent spring, summer, autumn, and winter, respectively.

Graph: Figure 6 (a) Comparison of the MAE of the proposed model and that of the other decomposition-integration models. (b) Comparison of the MAPE of the proposed model and that of the other decomposition-integration models. (c) Comparison of the RMSE of the proposed model and that of the other decomposition-integration models.

Table 1 Required parameters in RNN and GRU.

ParametersDescriptionsValues
Ahead-numHow long are the historical data used to predict the future10
Num-layersNumber of stack layers3
OptimizerMinimizes the loss function by training and optimizing the parametersRMSProp
EpochIndicates how many forward calculations and back propagations have been completed100

Table 2 Required parameters in Transformer.

ParametersDescriptionsValues
Num-layersNumber of stack layers3
Learning-rateControls the speed at which we adjust the weights of the neural network based on the loss gradient0.001
Batch sizeNumber of data passed to the program for training at a time32
OptimizerMinimizes the loss function by training and optimizing the parametersRMSProp
Drop-outIn the process of training batches, the phenomenon of overfitting can be significantly reduced by ignoring the general feature detector0.1

Table 3 Required parameters in LSTM.

ParametersDescriptionsValues
Num-layersNumber of stack layers2
Hidden-sizeNumber of features in the hidden state4
Batch sizeNumber of data passed to the program for training at a time64
OptimizerMinimizes the loss function by training and optimizing the parametersAdam
EpochIndicates how many forward calculations and back propagations have been completed100
Learning-rateControls the speed at which we adjust the weights of the neural network based on the loss gradient0.005

Table 4 Required parameters in WGAN.

ParametersDescriptionsValues
Time stepsHow long are the historical data used to predict the future10
ForecastHow long into the future to predict1
Hidden-sizeThe number of features in the hidden size128
Learning-rateControls the speed at which we adjust the weights of the neural network based on the loss gradient

1×104

Batch sizeNumber of data passed to the program for training at a time64
OptimizerMinimizes the loss function by training and optimizing the parametersRMSProp
EpochIndicates how many forward calculations and back propagations have been completed300
Clipping-valueAfter each update of the discriminator parameters, it truncates their absolute values to no more than a fixed constant0.001
N-criticNumber of iterations of the critic per generator iteration5

Table 5 Quantitative comparison of the four evaluation indicators of the proposed model and eight comparison models for four seasons.

RNNGRULSTMWGANTransformerCEE-LSTMCEE-WGANCEE-L-WCEE-W-L
SpringMAE38.8938.3440.1040.7040.8534.4347.2032.9724.98
MAPE31.5634.2632.3132.6326.3523.6725.9622.7115.45
RMSE61.4358.7051.0159.5751.0049.8549.6040.3338.75
R20.9560.9590.9570.9420.9560.970.9490.9580.982
SummerMAE28.9326.0928.9137.1128.1623.2030.3822.0121.39
MAPE20.7919.4923.2123.8918.3617.5120.2316.779.67
RMSE47.2544.3246.4854.2945.7032.5247.8231.4430.60
R20.9790.9810.9830.9750.9820.990.9770.9860.995
AutumnMAE50.4540.5439.0051.8049.1750.4846.0434.4031.27
MAPE30.3728.6829.6436.7127.4933.3824.1619.9316.53
RMSE56.1366.8967.1766.8264.0964.8662.2540.1539.10
R20.9230.940.9410.930.9280.9260.9320.9350.957
WinterMAE34.3234.6134.4131.1624.2025.9634.1729.6521.92
MAPE50.0841.9745.0442.0835.2535.6236.3832.5126.74
RMSE40.7744.5240.7743.9438.4036.5240.7936.7331.01
R20.9530.950.9550.9520.960.9640.9430.9460.976

Table 6 Results of DM test.

Model 1Model 2DM Valuep-Value
CEE-L-WCEE-WGAN−9.850.0000
CEE-LSTM−3.1920.0001
CEE-W-LCEE-WGAN−11.670.0000
CEE-LSTM−5.640.0000
CEE-L-W−5.370.0000

Author Contributions

Conceptualization, K.Y.; methodology, K.Y.; software, Q.L.; validation, K.Y.; formal analysis, K.Y.; investigation, D.Z.; resources, D.Z.; data curation, D.Z.; writing—original draft preparation, Q.L.; writing—review and editing, K.Y.; visualization, Q.L.; supervision, K.Y. and D.Z.; project administration, K.Y. and D.Z.; funding acquisition, K.Y. and D.Z. All authors have read and agreed to the published version of the manuscript.

Data Availability Statement

The source code and required data sets of the experiments can be obtained upon requests.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:

CEEMDANComplete ensemble empirical mode decomposition with adaptive noise
WGANWasserstein generative adversarial network
LSTMLong short-term memory network
MLMachine learning
DLDeep learning
EMDEmpirical mode decomposition
VMDVariational mode decomposition
ANNArtificial neural network
CNNConvolutional neural network
GRUGated recursive unit
IMFInherent mode functions
NSRDBNational Solar Radiation Database
MAEMean Absolute Error
MAPEMean Absolute Percentage Error
RMSERoot Mean Square Error
R2R-Square
DMDiebold–Mariano

Footnotes 1 Disclaimer/Publisher's Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. References Niu Z., Yu Z., Tang W., Wu Q., Reformat M. Wind power forecasting using attention-based gated recurrent unit network. Energy. 2020; 196: 117081. 10.1016/j.energy.2020.117081 2 Wang S., Chen J., Wang H., Zhang D. Degradation evaluation of slewing bearing using HMM and improved GRU. Measurement. 2019; 146: 385-395. 10.1016/j.measurement.2019.06.038 3 Yan M., Wang J., Li J., Zhang K., Yang Z. Traffic scene semantic segmentation using self-attention mechanism and bi-directional GRU to correlate context. Neurocomputing. 2020; 386: 293-304. 10.1016/j.neucom.2019.12.007 4 Yan K., Chen X., Zhou X., Yan Z., Ma J. Physical Model Informed Fault Detection and Diagnosis of Air Handling Units Based on Transformer Generative Adversarial Network. IEEE Trans. Ind. Inform. 2022; 19: 2192-2199. 10.1109/TII.2022.3193733 5 Zhou H., Liu Q., Yan K., Du Y. Deep learning enhanced solar energy forecasting with AI-driven IoT. Wirel. Commun. Mob. Comput. 2021; 2021: 1-11. 10.1155/2021/9249387 6 Feng Z.K., Huang Q.Q., Niu W.J., Yang T., Wang J.Y., Wen S.P. Multi-step-ahead solar output time series prediction with gate recurrent unit neural network using data decomposition and cooperation search algorithm. Energy. 2022; 261: 125217. 10.1016/j.energy.2022.125217 7 Iqbal S., Khan S.N., Sajid M., Khan J., Ayaz Y., Waqas A. Impact and performance efficiency analysis of grid-tied solar photovoltaic system based on installation site environmental factors. Energy Environ. 2022: 0958305X221106618. 10.1177/0958305X221106618 8 Mustafa R.J., Gomaa M.R., Al-Dhaifallah M., Rezk H. Environmental impacts on the performance of solar photovoltaic systems. Sustainability. 2020; 12608. 10.3390/su12020608 9 Tong J., Xie L., Fang S., Yang W., Zhang K. Hourly solar irradiance forecasting based on encoder–decoder model using series decomposition and dynamic error compensation. Energy Convers. Manag. 2022; 270: 116049. 10.1016/j.enconman.2022.116049 Zhou F., Huang Z., Zhang C. Carbon price forecasting based on CEEMDAN and LSTM. Appl. Energy. 2022; 311: 118601. 10.1016/j.apenergy.2022.118601 Yan K., Zhou X., Chen J. Collaborative deep learning framework on IoT data with bidirectional NLSTM neural networks for energy consumption forecasting. J. Parallel Distrib. Comput. 2022; 163: 248-255. 10.1016/j.jpdc.2022.01.012 Yang F., Yan K., Jin N., Du Y. Multiple households energy consumption forecasting using consistent modeling with privacy preservation. Adv. Eng. Inform. 2023; 55: 101846. 10.1016/j.aei.2022.101846 Yan K., Liu L., Xiang Y., Jin Q. Guest editorial: AI and machine learning solution cyber intelligence technologies: New methodologies and applications. IEEE Trans. Ind. Inform. 2020; 16: 6626-6631. 10.1109/TII.2020.2988944 Zhang L., Wang J., Niu X., Liu Z. Ensemble wind speed forecasting with multi-objective Archimedes optimization algorithm and sub-model selection. Appl. Energy. 2021; 301: 117449. 10.1016/j.apenergy.2021.117449 Ribeiro M.H.D.M., da Silva R.G., Moreno S.R., Mariani V.C., dos Santos Coelho L. Efficient bootstrap stacking ensemble learning model applied to wind power generation forecasting. Int. J. Electr. Power Energy Syst. 2022; 136: 107712. 10.1016/j.ijepes.2021.107712 Zhang K., Cao H., Thé J., Yu H. A hybrid model for multi-step coal price forecasting using decomposition technique and deep learning algorithms. Appl. Energy. 2022; 306: 118011. 10.1016/j.apenergy.2021.118011 Ribeiro M.H.D.M., Stefenon S.F., de Lima J.D., Nied A., Mariani V.C., Coelho L.d.S. Electricity price forecasting based on self-adaptive decomposition and heterogeneous ensemble learning. Energies. 2020; 135190. 10.3390/en13195190 Yang W., Sun S., Hao Y., Wang S. A novel machine learning-based electricity price forecasting model based on optimal model selection strategy. Energy. 2022; 238: 121989. 10.1016/j.energy.2021.121989 Li F., Zheng H., Li X. A novel hybrid model for multi-step ahead photovoltaic power prediction based on conditional time series generative adversarial networks. Renew. Energy. 2022; 199: 560-586. 10.1016/j.renene.2022.08.134 Haider S.A., Sajid M., Sajid H., Uddin E., Ayaz Y. Deep learning and statistical methods for short-and long-term solar irradiance forecasting for Islamabad. Renew. Energy. 2022; 198: 51-60. 10.1016/j.renene.2022.07.136 Elsheikh A.H., Panchal H., Ahmadein M., Mosleh A.O., Sadasivuni K.K., Alsaleh N.A. Productivity forecasting of solar distiller integrated with evacuated tubes and external condenser using artificial intelligence model and moth-flame optimizer. Case Stud. Therm. Eng. 2021; 28: 101671. 10.1016/j.csite.2021.101671 Yazdanian P., Sharifian S. E2LG: A multiscale ensemble of LSTM/GAN deep learning architecture for multistep-ahead cloud workload prediction. J. Supercomput. 2021; 77: 11052-11082. 10.1007/s11227-021-03723-6 Liu Z., Liu J. A robust time series prediction method based on empirical mode decomposition and high-order fuzzy cognitive maps. Knowl.-Based Syst. 2020; 203: 106105. 10.1016/j.knosys.2020.106105 Wang J., Cao J., Yuan S., Cheng M. Short-term forecasting of natural gas prices by using a novel hybrid method based on a combination of the CEEMDAN-SE-and the PSO-ALS-optimized GRU network. Energy. 2021; 233: 121082. 10.1016/j.energy.2021.121082 Shi Z., Liang H., Dinavahi V. Direct interval forecast of uncertain wind power based on recurrent neural networks. IEEE Trans. Sustain. Energy. 2017; 9: 1177-1187. 10.1109/TSTE.2017.2774195 Yin H., Ou Z., Zhu Z., Xu X., Fan J., Meng A. A novel asexual-reproduction evolutionary neural network for wind power prediction based on generative adversarial networks. Energy Convers. Manag. 2021; 247: 114714. 10.1016/j.enconman.2021.114714 Wang F., Zhang Z., Liu C., Yu Y., Pang S., Duić N., Shafie-Khah M., Catalão J.P. Generative adversarial networks and convolutional neural networks based weather classification model for day ahead short-term photovoltaic power forecasting. Energy Convers. Manag. 2019; 181: 443-462. 10.1016/j.enconman.2018.11.074 Justin J., Alexandre A., Li F.-F. Perceptual losses for real-time style transfer and super-resolution. Proceedings of the European Conference on Computer Vision. Amsterdam, The Netherlands. 11–14 October 2016; Springer: Berlin/Heidelberg, Germany. 2016: 694-711 Yan K., Dai Y., Xu M., Mo Y. Tunnel surface settlement forecasting with ensemble learning. Sustainability. 2019; 12232. 10.3390/su12010232 Eseye A.T., Zhang J., Zheng D. Short-term photovoltaic solar power forecasting using a hybrid Wavelet-PSO-SVM model based on SCADA and Meteorological information. Renew. Energy. 2018; 118: 357-367. 10.1016/j.renene.2017.11.011 Pi M., Jin N., Ma X., Yan K. Short-term Solar Irradiation Prediction Model Based on WCNN_ALSTM. Proceedings of the 2021 IEEE Intl Conf on Dependable, Autonomic and Secure Computing, Intl Conf on Pervasive Intelligence and Computing, Intl Conf on Cloud and Big Data Computing, Intl Conf on Cyber Science and Technology Congress (DASC/PiCom/CBDCom/CyberSciTech). Calgary, AB, Canada. 25–28 October 2021; IEEE: Piscataway, NJ, USA. 2021: 405-412 Jin N., Yang F., Mo Y., Zeng Y., Zhou X., Yan K., Ma X. Highly accurate energy consumption forecasting model based on parallel LSTM neural networks. Adv. Eng. Inform. 2022; 51: 101442. 10.1016/j.aei.2021.101442 Zeng Y., Chen J., Jin N., Jin X., Du Y. Air quality forecasting with hybrid LSTM and extended stationary wavelet transform. Build. Environ. 2022; 213: 108822. 10.1016/j.buildenv.2022.108822 Lan H., Yin H., Hong Y.Y., Wen S., David C.Y., Cheng P. Day-ahead spatio-temporal forecasting of solar irradiation along a navigation route. Appl. Energy. 2018; 211: 15-27. 10.1016/j.apenergy.2017.11.014 Mohammadi K., Shamshirband S., Tong C.W., Arif M., Petković D., Ch S. A new hybrid support vector machine–wavelet transform approach for estimation of horizontal global solar radiation. Energy Convers. Manag. 2015; 92: 162-171. 10.1016/j.enconman.2014.12.050 Zhaohua Wu N. Ensemble empirical mode decomposition: A noise-assited. Biomed. Technol. 2010; 55: 193-201 Torres M.E., Colominas M.A., Schlotthauer G., Flandrin P. A complete ensemble empirical mode decomposition with adaptive noise. Proceedings of the 2011 IEEE international conference on acoustics, speech and signal processing (ICASSP). Prague, Czech Republic. 22–27 May 2011; IEEE: Piscataway, NJ, USA. 2011: 4144-4147 Hochreiter S., Schmidhuber J. Long short-term memory. Neural Comput. 1997; 9: 1735-1780. 10.1162/neco.1997.9.8.1735. 9377276 Goodfellow I., Pouget-Abadie J., Mirza M., Xu B., Warde-Farley D., Ozair S., Courville A., Bengio Y. Generative adversarial networks. Commun. ACM. 2020; 63: 139-144. 10.1145/3422622 Arjovsky M., Bottou L. Towards principled methods for training generative adversarial networks. arXiv. 2017. 1701.04862 Sengupta M., Xie Y., Lopez A., Habte A., Maclaurin G., Shelby J. The national solar radiation data base (NSRDB). Renew. Sustain. Energy Rev. 2018; 89: 51-60. 10.1016/j.rser.2018.03.003

By Qianqian Li; Dongping Zhang and Ke Yan

Reported by Author; Author; Author

Titel:
A Solar Irradiance Forecasting Framework Based on the CEE-WGAN-LSTM Model
Autor/in / Beteiligte Person: Li, Qianqian ; Zhang, Dongping ; Yan, Ke
Link:
Zeitschrift: Sensors, Jg. 23 (2023-03-01), Heft 5, S. 2799-2799
Veröffentlichung: MDPI AG, 2023
Medientyp: academicJournal
ISSN: 1424-8220 (print)
DOI: 10.3390/s23052799
Schlagwort:
  • time-series forecasting
  • deep learning
  • signal decomposition
  • solar irradiance
  • Chemical technology
  • TP1-1185
Sonstiges:
  • Nachgewiesen in: Directory of Open Access Journals
  • Sprachen: English
  • Collection: LCC:Chemical technology
  • Document Type: article
  • File Description: electronic resource
  • Language: English

Klicken Sie ein Format an und speichern Sie dann die Daten oder geben Sie eine Empfänger-Adresse ein und lassen Sie sich per Email zusenden.

oder
oder

Wählen Sie das für Sie passende Zitationsformat und kopieren Sie es dann in die Zwischenablage, lassen es sich per Mail zusenden oder speichern es als PDF-Datei.

oder
oder

Bitte prüfen Sie, ob die Zitation formal korrekt ist, bevor Sie sie in einer Arbeit verwenden. Benutzen Sie gegebenenfalls den "Exportieren"-Dialog, wenn Sie ein Literaturverwaltungsprogramm verwenden und die Zitat-Angaben selbst formatieren wollen.

xs 0 - 576
sm 576 - 768
md 768 - 992
lg 992 - 1200
xl 1200 - 1366
xxl 1366 -