Abstract

The fluctuations of economic and financial time series are influenced by various kinds of factors and usually demonstrate strong nonstationary and high complexity. Therefore, accurately forecasting economic and financial time series is always a challenging research topic. In this study, a novel multidecomposition and self-optimizing hybrid approach integrating multiple improved complete ensemble empirical mode decompositions with adaptive noise (ICEEMDANs), whale optimization algorithm (WOA), and random vector functional link (RVFL) neural networks, namely, MICEEMDAN-WOA-RVFL, is developed to predict economic and financial time series. First, we employ ICEEMDAN with random parameters to separate the original time series into a group of comparatively simple subseries multiple times. Second, we construct RVFL networks to individually forecast each subseries. Considering the complex parameter settings of RVFL networks, we utilize WOA to search the optimal parameters for RVFL networks simultaneously. Then, we aggregate the prediction results of individual decomposed subseries as the prediction results of each decomposition, respectively, and finally integrate these prediction results of all the decompositions as the final ensemble prediction results. The proposed MICEEMDAN-WOA-RVFL remarkably outperforms the compared single and ensemble benchmark models in terms of forecasting accuracy and stability, as demonstrated by the experiments conducted using various economic and financial time series, including West Texas Intermediate (WTI) crude oil prices, US dollar/Euro foreign exchange rate (USD/EUR), US industrial production (IP), and Shanghai stock exchange composite index (SSEC).

1. Introduction

Economic and financial time series, such as price movements, stock market indices, and exchange rate, are usually characterized by strong nonlinearity and high complexity, since they are influenced by a number of extrinsic and intrinsic factors including economic conditions, political events, and even sudden crises [1, 2]. Economic and financial time series forecasting always play a vital role in social and economic development, which is of great economic importance to both individuals and countries. Therefore, economic and financial time series forecasting is always a very active research area.

In extant research, various forecasting methods were proposed to forecast various economic and financial time series. These forecasting methods mainly include statistical and artificial intelligence (AI) approaches. The frequently used statistical approaches for economic and financial time series forecasting include the error correction model (ECM) [3], hidden Markov model (HMM) [4], random walk (RW) model [5], autoregressive moving average (ARMA) model [6], autoregressive integrated moving average (ARIMA) model [7], and generalized autoregressive conditional heteroskedasticity (GARCH) model [8, 9]. Lanza et al. forecasted the series of crude oil prices in two distinct areas using the ECM [3]. Hassan and Nath developed the HMM approach for forecasting stock price for interrelated markets [4]. Kilian and Taylor analyzed the advantage of RW in exchange rate forecasting [5]. Rout et al. integrated ARMA with differential evolution (DE) to develop a hybrid model for exchange rate forecasting [6]. Mondal et al. conducted a study on the effectiveness of the ARIMA model on the forecasting of 56 Indian stocks from different sectors [7]. Alberg et al. conducted a comprehensive analysis of the stock indices using various GARCH models, and the experimental results showed that the asymmetric GARCH model enhanced the overall prediction performance [9].

Since most economic and financial time series involve the complex characteristics of strong nonlinearity and nonstationarity, it is difficult to obtain satisfactory forecasting accuracy by statistical approaches. Hence, various AI approaches were proposed for economic and financial time series forecasting. These AI forecasting approaches include the artificial neural network (ANN) [10, 11], support vector machine (SVM) [12, 13], extreme learning machine (ELM) [14], random vector functional link (RVFL) neural network [15], and recurrent neural network (RNN) [16]. Pradhan and Kumar utilized ANN to forecast foreign exchange rate in India, and the experimental results indicated that the ANN could effectively forecast the exchange rate [10]. Das and Padhy forecasted the commodity futures contract index using the SVM, and the empirical analysis showed that the proposed model was effective and achieved the satisfactory prediction performance [12]. Li et al. made stock price prediction using the ELM, and the comparison results showed that the ELM with radial basis function (RBF) kernels achieved better prediction performance with faster speed than back propagation neural networks (BPNNs) [14]. Moudiki et al. employed quasirandomized functional link networks for various time series forecasting, and the proposed approach could generate more robust prediction results [15]. Baek and Kim employed long short-term memory (LSTM) for stock index forecasting, and the results confirmed the LSTM model had excellent prediction accuracy [16].

In order to effectively improve prediction accuracy, various hybrid forecasting models were designed for economic and financial time series forecasting. Babu and Reddy combined ARIMA and nonlinear ANN models to develop a novel hybrid ARIMA-ANN model, and the experiments on electricity price and stock index indicated that the proposed ARIMA-ANN had higher prediction accuracy [17]. Kumar and Thenmozhi compared three different hybrid models for the forecasting of stock index returns and concluded that the ARIMA-SVM model could obtain the highest prediction accuracy [18].Hsu built a hybrid model based on a back propagation neural network (BPNN) and genetic programming (GP) for stock/futures price forecasting, and the empirical analysis showed that the proposed hybrid model could effectively improve the prediction accuracy [19]. These hybrid models are able to fully take advantage of the potential of single models and, thus, obtain better prediction accuracy than single models.

Due to the complexity of original economic and financial time series, conducting forecasting on original time series is hard to obtain satisfactory prediction accuracy. To reduce the complexity of original time series, a framework of “decomposition and ensemble” is widely utilized in the field of time series forecasting. The framework includes three stages: decomposition, forecasting, and ensemble. The original time series is firstly separated into a sum of subseries, then a prediction model is used to forecast each subseries, and finally, the predictions of all the subseries are aggregated as the final prediction results. Decomposition, as the first step, is very important for enhancing the performance of the ensemble model. The widely used decomposition approaches include wavelet decomposition (WD), variational mode decomposition (VMD), and empirical mode decomposition (EMD) class methods. Lahmiri combined VMD with a general regression neural network (GRNN) to develop a novel ensemble forecasting model, and the experimental results suggested that VMD outperformed EMD for the prediction of economic and financial time series [20]. Kao et al. integrated WD, support vector regression (SVR), and multivariate adaptive regression splines (Mars) to develop an ensemble forecasting model to forecast stock price, and the proposed model obtained better prediction accuracy [21]. In the second stage of the framework of “decomposition and ensemble,” various optimization approaches were introduced to enhance the performance of predictors. Li et al. proposed a ridge regression (RR) with DE to forecast crude oil prices and obtained excellent forecasting accuracy [22]. Bagheri et al. introduced quantum-behaved particle swarm optimization (QPSO) to tune the adaptive network-based fuzzy inference system (ANFIS) for financial time series forecasting [23]. Wang et al. employed brain storm optimization (BSO) algorithm to optimize SVR, and the results indicated that the developed approach was effective in stock market analysis [24].

In the “decomposition and ensemble” framework, decomposition approaches and prediction approaches influence the final prediction results greatly. Considering the powerful decomposition ability of ICEEMDAN, the excellent search efficiency of WOA, and the accurate forecasting ability of the RVFL network, we develop a novel ensemble prediction model integrating multiple ICEEMDANs, WOA, and the RVFL network, namely, MICEEMDAN-WOA-RVFL, for economic and financial time series forecasting. Firstly, ICEEMDAN with random parameters is utilized to divide the original economic and financial time series into a sum of subseries. Secondly, the RVFL network is applied to forecast each decomposed subseries individually, and WOA is used to optimize the parameter values of the RVFL network simultaneously. Finally, the predictions of all individual subseries are aggregated as the prediction values of one process of decomposition and ensemble. From our observations, we find that the decomposition in the first stage has some disadvantages of the uncertainties with a quite randomness, which can lead to the difference and instability of the prediction results. In addition, extensive literature has shown that combining multiple forecasts can effectively enhance prediction accuracy [25, 26]. Therefore, we randomize the decomposition parameter values of ICEEMDAN in the first stage, repeat the abovementioned processes multiple times, and integrate the results of multiple decompositions and ensembles as the final prediction values. We expect that the multiple decomposition strategy can reduce the randomness of one single decomposition and further improve the ensemble prediction stability and accuracy.

The main contributions of this paper are as follows: (1) we propose a new multidecomposition and self-optimizing ensemble prediction model integrating multiple ICEEMDANs, WOA, and RVFL networks for economic and financial time series forecasting. As far as we know, this is the first time that the novel combination is developed for economic and financial time series forecasting. (2) To further enhance forecasting accuracy and stability, we utilize multiple differentiated ICEEMDANs to decompose original economic and financial time series and, finally, ensemble the predictions of all decompositions as the final predictions. (3) WOA is firstly introduced to optimize various parameters of RVFL networks. (4) The empirical results on four different types of economic and financial time series show that our proposed MICEEMDAN-WOA-RVFL significantly enhances the prediction performance in terms of forecasting accuracy and stability.

The novelty of the proposed MICEEMDAN-WOA-RVFL is three-fold: (1) a novel hybrid model integrating multiple ICEEMDANs, WOA, and RVFL networks is designed for economic and financial time series forecasting; (2) the multiple decomposition strategy is firstly proposed to overcome the randomness of one single decomposition and to improve prediction accuracy and stability; and (3) WOA is first applied to optimizing RVFL networks to improve the performance of individual forecasting.

The remainder of the paper is organized as follows. Section 2 offers a brief introduction to the ICEEMDAN, WOA, and RVFL network. Section 3 provides the architecture and the detailed implementation of the proposed MICEEMDAN-WOA-RVFL. Section 4 analyzes the empirical results on various economic and financial time series forecasting. Section 5 discusses some details of the developed prediction model, and Section 6 concludes this paper.

2. Preliminaries

2.1. Improved Complete Ensemble Empirical Mode Decomposition with Adaptive Noise (ICEEMDAN)

Empirical mode decomposition (EMD), an adaptive time-frequency analysis approach for nonstationary signals, was designed by Huang et al. [27]. EMD separates original time series into a sum of “intrinsic mode functions” (IMFs) and one residue, and thus, it can simplify time series analysis. Due to some drawbacks of EMD, such as mode mixing, EEMD [28] and CEEMDAN [29] have been proposed to improve decomposition performance and applied in various fields [3032]. In spite of that, these decomposition methods still have some new problems. To solve these problems, an improved CEEMDAN (ICEEMDAN) was developed by Colominas et al. [33].

Let Ek(·) be the operator which generates the kth mode using EMD, M(·) be the operator which generates the local mean of the series, and be a realization of zero mean unit variance noise. When x is the original signal, the detailed decomposition process of ICEEMDAN is as follows:(i)Step 1: employ EMD to compute the local means of I realizations x(i) = x + β0E1() to achieve the first residue , and β0 > 0(ii)Step 2: calculate the first IMF1 = x − r1(iii)Step 3: calculate the average of local means of the realizations as the second residue: (iv)Step 4: compute the kth residue for k = 3,…, K: (v)Step 5: go to step 4 for next k

Since ICEEMDAN can effectively decompose the original time series, it has been frequently introduced into various time series forecasting [3436]. In our study, we employ ICEEMDAN to separate the original economic and financial time series into a sum of simpler subseries for subsequent forecasting.

2.2. Whale Optimization Algorithm (WOA)

Whale optimization algorithm (WOA) is a type of optimization method and outperforms particle swarm optimization (PSO), ant colony optimization (ACO), gravitational search algorithm (GSA), and fast evolutionary programming (FEP) in the optimization performance [37, 38]. Simulating the hunting process of whales, WOA includes three main operators: encircling prey, bubble-net foraging, and search for prey. In each iteration, individuals update their positions toward the best individual in the last iteration, which can be formulated as follows:where t represents the tth iteration, A and C are two coefficient vectors, Pbest is the best individual so far, P is the position of an individual, || represents the absolute value, and · indicates an element-by-element multiplication.

In the exploitation (bubble-net foraging) phase, the individual position is updated based on its distance to the best individual by simulating the helix-shaped movement of whales, which is formulated as follows:where D = |Pbest(t) − P(t)| represents the distance between the best individual obtained so far and the ith individual, l is a random number in [−1, 1], and b is a constant which is used to define the shape of logarithmic spiral.

In exploration (search for prey) phase, the individual position is updated using a randomly selected individual. The mathematical model is follows:where Prand represents a randomly selected individual.

The detailed flowchart of WOA is illustrated in Figure 1, where p is a random number in [0, 1]. Due to its very competitive search ability, WOA has been widely applied in various fields [3941]. Therefore, we consider taking advantage of the effective search ability of WOA to seek the optimal parameters for RVFL networks.

2.3. Random Vector Functional Link (RVFL) Neural Network

As a kind of modification of the multilayer perceptron (MLP) model, the random vector functional link (RVFL) neural network was proposed by Pao et al. [42]. The RVFL neural network is able to overcome the slow convergence, overfitting, and local minimum inherently in the traditional gradient-based learning algorithms. Like MLP, the architecture of the RVFL neural network includes three layers, which is illustrated in Figure 2. The main modification in the RVFL network lies in the connection in the network structure. Since the RVFL network has the direct connections from the input layer to the output layer, it can perform better compared to no direct link [43, 44].

The neurons in the hidden layer, known as enhancement nodes, calculate the sum of all the output of the input layer neurons and obtain their output with an activation function:where denotes the weight between in and hm, bm represents the bias of the mth neuron in hidden layer, and represents an activation function.

The output layer neurons integrate all the output from the hidden layer and input layer neurons, and the final output iswhere represents the weight between hm and ol and indicates the weight between in and ol.

To enhance the training efficiency, the RVFL neural network utilizes a given distribution to fix the values of and bm and obtain the weights of and by minimizing the system error:where P indicates the number of training samples and t are the target values, B is the combination of and , and d represents a combined vector.

The RVFL neural network has demonstrated an extremely efficient and fast forecasting ability and has been frequently used in time series forecasting [44, 45].

3. MICEEMDAN-WOA-RVFL: The Proposed Approach for Economic and Financial Time Series Forecasting

Referring to the framework of “decomposition and ensemble,” we design a multidecomposition and self-optimizing hybrid model that integrates multiple ICEEMDANs, WOA, and RVFL networks, termed as MICEEMDAN-WOA-RVFL, to forecast economic and financial time series. The architecture of the proposed hybrid model is illustrated in Figure 3.

Our proposed MICEEMDAN-WOA-RVFL takes advantage of the idea of “divide and conquer” that was frequently used in time series forecasting, image processing, fault diagnosis, and so on [4653]. A procedure of “decomposition and ensemble” in the MICEEMDAN-WOA-RVFL is as follows:(i)Stage 1: decomposition: ICEEMDAN is employed to separate original time series into several subseries (i.e., several IMFs and one residue).(ii)Stage 2: individual forecasting. Each decomposed subseries is divided into a training dataset and a testing dataset, then the RVFL network with WOA optimization is developed on each training dataset independently, and finally, the developed RVFL model is used to each testing dataset. The reason why we select the RVFL network as the predictor is its powerful forecasting ability in extant research [15, 34, 44]. Since the parameter setting of the RVFL network plays an important role in the prediction performance, we introduce WOA to seek the optimal parameter values for the RVFL network in the forecasting stage.(iii)Stage 3: ensemble: the predictions of all the decomposed subseries are aggregated as the final prediction results of one “decomposition and ensemble” using addition aggregation.

In this study, to enhance both accuracy and stability of final prediction, we generate random values for the decomposition parameters of ICEEMDAN in the decomposition stage, including number of realizations (Nr), noise standard deviation (Nsd), and maximum number of sifting iterations (Maxsi), for each decomposition, and repeat the procedure of “decomposition and ensemble” M times and, finally, combine all the results of multiple “decompositions and ensembles” using the RMSE-weighted method as the final prediction results. The corresponding weight of ith forecasting model is as follows:where M denotes the number of individual models and RMSEi indicates the RMSE value of the ith forecasting model in the training process.

Although some recent studies also employ the RVFL network for time series forecasting, they obviously differ from the current study in the decomposition technique and network optimization: (1) they divide the original time series using WD or EMD; (2) they construct RVFL networks using the fixed parameter values. Unlike the previous studies, our study decomposes economic and financial time series using ICEEMDAN and searches the optimal parameter values of RVFL networks based on WOA. Furthermore, the previous research mainly focuses on dividing original time series by one single decomposition or dual decomposition [20, 35, 54]. In dual decomposition, the original signal is first decomposed into several components, and then, the high-frequency components continue to be decomposed into other components using the same or different decomposition method. Essentially, the dual decomposition process belongs to one decomposition, just including two different decomposition stages. Unlike the previous research, one main improvement in this study is the multiple decomposition strategy, which can successfully overcome the randomness of one single decomposition and further improve the prediction accuracy and stability of the developed forecasting approach. To our knowledge, it is the first time that the multiple decomposition strategy is developed for the forecasting of economic and financial time series.

4. Experimental Results

4.1. Data Description

As we know, economic and financial time series are influenced by various factors, sometimes raising and dropping down in a short time. The dramatic fluctuations usually lead to the significant nonlinearity and nonstationarity of the time series. To comprehensively evaluate the effectiveness of the proposed MICEEMDAN-WOA-RVFL, we choose four different time series, including the West Texas Intermediate crude oil spot price (WTI), US dollar/Euro foreign exchange rate (USD/EUR), US industrial production (IP), and Shanghai stock exchange composite index (SSEC), as the experimental datasets. The first three datasets can be accessed via the website of St. Louis Fed Research [55], and the last one can be obtained via the website of NetEase [56].

Each time series is separated into two subdatasets: the first 80% for training and the last 20% for testing. Table 1 shows the divided samples of the abovementioned four economic and financial time series.

We utilize ICEEMDAN to decompose these time series into groups of relatively simple subseries. Figure 4 offers an example of the decomposition of the WTI dataset using ICEEMDAN.

4.2. Evaluation Indices

In this study, we use four evaluation metrics, including the mean absolute percent error (MAPE), the root mean squared error (RMSE), the directional statistic (Dstat), and the Diebold–Mariano (DM) test, to assess the performance of the proposed model. Among them, MAPE and RMSE are used to assess the forecasting error, defined as follows:where N is the size of the evaluated samples, Ot denotes the actual values, and Pt represents the predicted values at time t. The lower the values of RMSE and MAPE, the better the prediction models.

The Dstat indicates the performance of direction prediction, which is formulated as follows:where di = 1 if (Pt+1 − Ot) (Ot+1 − Ot) ≥ 0; otherwise, di = 0. A higher value of Dstat indicates a more accurate direction prediction.

Furthermore, to test the significance of the prediction performance of pairs of models, we employ the Diebold–Mariano (DM) test in this study.

4.3. Experimental Settings

In this study, we compare the proposed MICEEMDAN-WOA-RVFL with several state-of-the-art forecasting models, including the single models and the ensemble models. Among all these models, the single models include one popular statistical model, RW, and two popular AI models. BPNN and least square SVR (LSSVR). The ensemble models derive from the combination of the single models and the decomposition method ICEEMDAN.

The detailed parameters of all prediction models, decomposition approach ICEEMDAN, and optimization method WOA in the experiments are shown in Table 2. The parameter values of BPNN, LSSVR, RVFL, and ICEEMDAN refer to the previous literature [22, 34, 45].

All experiments were conducted using Matlab R2019b on a PC with 64 bit Microsoft Windows 10, 8 GB RAM, and 1.8 GHz i7-8565U CPU.

4.4. Results and Analysis

We compare the forecasting performance of six prediction models, including three single models.

(RW, LSSVR, and BPNN) and three ensemble modes (ICEEMDAN-RW, ICEEMDAN-LSSVR, and ICEEMDAN-BPNN) with that of our proposed MICEEMDAN-WOA-RVFL in terms of MAPE, RMSE, and Dstat. Due to the different horizons, we train different forecasting models separately. That is to say, we use the proposed scheme for different horizons to train different models. Tables 35 report the experimental results in terms of each evaluation index with 1-, 3-, and 6-horizon, respectively.

From Table 3, we can see that the proposed MICEEMDAN-WOA-RVFL obtains the lowest (the best) MAPE values with all the horizons in all the datasets. RW obtains the best MAPE values with all the horizons in all the datasets among all the compared single models, demonstrating it is better than LSSVR and BPNN for the forecasting of economic and financial time series. Of all the ensemble models, the ICEEMDAN-LSSVR model and ICEEMDAN-BPNN model obtain the close MAPE values, obviously better than the ICEEMDAN-RW.

The RMSE values of the four time series datasets are listed in Table 4. From this table, we can find that the proposed MICEEMDAN-WOA-RVFL outperforms all the single and ensemble models with all the horizons in all the datasets. The statistical model RW obtains the best RMSE values in 10 out of 12 cases, demonstrating that it is more suitable for economic and financial time series forecasting than LSSVR and BPNN among all the single models. As to the ensemble models, the proposed MICEEMDAN-WOA-RVFL obtains the lower RMSE values than the compared ensemble models, demonstrating that the former is more effective for economic and financial time series forecasting.

Table 5 shows the directional statistics Dstat, and we can see that the MICEEMDAN-WOA-RVFL achieves the highest Dstat values in all the 12 cases, indicating that it has better performance of direction forecasting. Amongst the single prediction models, LSSVR and RW obtain the best Dstat values in 5 cases, respectively, better than the BPNN. Similarly, the ICEEMDAN-LSSVR model and ICEEMDAN-BPNN model obtain the close Dstat values, obviously better than the ICEEMDAN-RW model in all the 12 cases.

From the all prediction results, we can find that all the ensemble prediction models except ICEEMDAN-RW greatly outperform the corresponding single prediction models in all the 12 cases, showing that the framework of decomposition and ensemble is an effective tool for improving the forecasting performance. Of all the compared models, the proposed MICEEMDAN-WOA-RVFL obtains the highest Dstat values and the lowest MAPE and RMSE values in all the time series datasets, showing that it is completely superior to the benchmark prediction models. Furthermore, for each prediction model, the MAPE and RMSE values increase while the Dstat values decrease with the horizon. This demonstrates that it is easier to forecast time series with a short horizon than with a long one. It is worth noting that the proposed MICEEMDAN-WOA-RVFL still achieves the relatively good MAPE and RMSE with the increase of horizon among the compared models. For example, when the proposed model obtains 0.7737 Dstat with horizon 6 in the WTI dataset, it achieves the relatively low MAPE (0.0113) and RMSE (0.8146), indicating that the prediction values are very close to the real values although the proposed model misses direction about 22.63%. In other words, the proposed MICEEMDAN-WOA-RVFL can still achieve satisfactory forecasting accuracy with a long horizon.

Furthermore, we can find that the multiple decomposition strategy does not improve the forecast for the RW model. One possible explanation is simply that RW infers that the past movement or trend of a time series cannot be used to predict its future movement, and thus, it cannot take advantage of the historical data and the diversity of multiple decomposition to make the future prediction. Therefore, when we aggregate the multiple RW model predictions of all the decomposed subseries, we just integrate several random predictions and, thus, cannot significantly improve the ensemble prediction results. In contrast, the LSSVR and BPNN, as well as RVFL, can fully use all the historical data and the diversity of multiple decomposition to make the future prediction. Specifically, the multiple decomposition using different parameters generates many groups of different decomposed subseries, and the diversity of decomposition can successfully overcome the randomness of one single decomposition and further improve the prediction accuracy and stability of the developed forecasting approach.

In addition, the Diebold–Mariano (DM) test is utilized to evaluate whether the forecasting accuracy of the proposed MICEEMDAN-WOA-RVFL significantly outperforms those of the other compared models. Table 6 shows the statistics and p values (in brackets).

On one hand, the DM statistical values between the ensemble prediction models and their corresponding single predictors are much lower than zero and the corresponding p values are almost equal to zero with all the horizons except for the RW model, showing that the architecture of “decomposition and ensemble” contributes to greatly improving prediction accuracy and the combination of ICEEMDAN and AI predictors is more effective for economic and financial time series forecasting.

On the other hand, DM test results on the prediction of all the four time series datasets indicate that the MICEEMDAN-WOA-RVFL is significantly better than the single models and the other ensemble models with all the horizons, and the corresponding p values are much lower than 0.01 in all the cases.

In summary, the DM test results demonstrate that the combination of multiple ICEEMDANs, RVFL networks, and WOA optimization can significantly enhance the prediction accuracy of economic and financial time series forecasting.

5. Discussion

To better investigate the proposed MICEEMDAN-WOA-RVFL, we further discuss the developed prediction model, including the comparison of single decomposition and multiple decompositions, the optimization effectiveness of WOA, and the impact of ensemble size in this subsection.

5.1. Comparison of Single Decomposition and Multiple Decompositions

One of the main novelties of this study is the multiple decomposition strategy, which can successfully overcome the randomness of a single decomposition and improve the prediction accuracy and stability of the developed forecasting model. To evaluate the effectiveness of the multiple decomposition strategy, we compare the prediction results of MICEEMDAN-WOA-RVFL and ICEEMDAN-WOA-RVFL. The former ensembles the prediction results of M(M = 100) individual ICEEMDAN decompositions with random parameters, while the latter only employs one ICEEMDAN decomposition. We randomly choose 5 out of these 100 decompositions and execute ICEEMDAN-WOA-RVFL for time series forecasting. Tables 79 report the MAPE, RMSE, and Dstat values of the MICEEMDAN-WOA-RVFL and the five ICEEMDAN-WOA-RVFL models using single decomposition and the corresponding mean values of these five models using single decomposition.

On one hand, compared with the prediction results of the five single decompositions and the mean prediction results, the proposed MICEEMDAN-WOA-RVFL achieves the lowest MAPE and the highest Dstat values in all the 12 cases and the lowest RMSE values in 11 out of all the 12 cases, indicating that the multiple decomposition strategy can successfully overcome the randomness of single decomposition and improve the ensemble prediction accuracy.

On the other hand, we can find that the multiple decomposition strategy can greatly improve the stability of the prediction model. For example, the range of MAPE values of the five single decomposition models with horizon 6 in the IP time series dataset is from 0.0034 to 0.1114, indicating that different single decomposition can produce relatively great difference in prediction results. When we employ the multiple decomposition strategy, we can overcome the randomness of single decomposition and, thus, enhance prediction stability.

In summary, the experimental results suggest that the multiple decomposition strategy and prediction ensemble can effectively enhance prediction accuracy and stability. The main reasons for the prediction improvement lie in three aspects: (1) the multiple decomposition can reduce the randomness of one single decomposition and simultaneously generate groups of differential subseries; (2) predictions using these groups of differential subseries can achieve diverse prediction results; and (3) the selection and ensemble of these diverse prediction results can ensure both accuracy and diversity and, thus, improve the final ensemble prediction.

5.2. The Optimization Effectiveness of WOA

When we use RVFL networks to construct predictors, a number of parameters need to be set in advance. In this study, WOA is introduced to search the optimal parameter values for RVFL predictors using its powerful optimization ability. To investigate the optimization effectiveness of WOA for parameter search, we compare the proposed MICEEMDAN-WOA-RVFL with MICEEMDAN-RVFL without WOA optimization. According to the literature [45], we fixed the number of hidden neurons Nhe = 100, activation function Func = sigmoid, and random type Rand = Gaussian in MICEEMDAN-RVFL. The MAPE, RMSE, and Dstat values are reported in Tables 1012, respectively.

In all the four time series datasets, the prediction performance of the proposed MICEEMDAN-WOA-RVFL model is better than or equal to that of the MICEEMDAN-RVFL model without WOA optimization in all the 12 cases except for the RMSE value with horizon 1 in the SSEC dataset in terms of MAPE and RMSE, as listed in Tables 10 and 11. In addition, the MICEEMDAN-WOA-RVFL obtains the higher Dstat values in 10 out of 12 cases, which can be seen in Table 12. The all results indicate that WOA can effectively search the optimal parameter settings for RVFL networks, further improving the overall prediction performance.

5.3. The Impact of Ensemble Size

The previous research has demonstrated that the ensemble strategy of using all individual prediction models is unlikely to work well and the selection of individual prediction models contributes to improving the ensemble prediction performance [57]. In this study, we sort all individual prediction models based on their past performance (RMSE values) and, then, select the top N percent as the ensemble size to construct the ensemble prediction model. To further investigate the impact of ensemble size on ensemble prediction, we use different ensemble sizes (es = 10%, 20%,…, 100%) to select the top N percent of individual forecasting models to develop the ensemble prediction model and conduct the one-step-ahead forecasting experiment on the four time series datasets. The results are demonstrated in Figure 5.

We can see that the MICEEMDAN-WOA-RVFL obtains the best forecasting performance in the WTI, IP, and SSEC datasets when the ensemble size es is in the range of 20%–40% and in the USD/EUR dataset when the ensemble size es is in the range of 10%–40% in terms of MAPE, RMSE, and Dstat. When the es is greater than 40%, the MAPE and RMSE values continue to worsen and become the worst when the ensemble size grows to 100. The experimental results indicate that the ensemble size has an overall significant impact on ensemble prediction, and an ideal range of ensemble size is about 20% to 40%.

6. Conclusions

To better forecast economic and financial time series, we propose a novel multidecomposition and self-optimizing ensemble prediction model MICEEMDAN-WOA-RVFL combining multiple ICEEMDANs, WOA, and RVFL networks. The MICEEMDAN-WOA-RVFL first uses ICEEMDAN to multiply separate original economic and financial time series into groups of subseries many times, and then, RVFL networks are used to individually forecast the decomposed subseries in each decomposition. Simultaneously, WOA is introduced to optimize RVFL networks to further improve the prediction accuracy. Thirdly, the predictions of subseries in each decomposition are integrated into the forecasting results of each decomposition using addition. Finally, the prediction results of each decomposition are selected based on RMSE values and are combined as the final prediction results.

As far as we know, it is the first time that WOA is employed for the optimal parameter search for RVFL networks and the multiple decomposition strategy is introduced in time series forecasting. The empirical results indicate that (1) the proposed MICEEMDAN-WOA-RVFL significantly improves prediction accuracy in various economic and financial time series forecasting; (2) WOA can effectively search optimal parameters for RVFL networks and improve prediction performance of economic and financial time series forecasting; and (3) the multiple decomposition strategy can successfully overcome the randomness of a single decomposition and enhance the prediction accuracy and stability of the developed prediction model.

We will extend our study in two aspects in the future: (1) applying the MICEEMDAN-WOA-RVFL to forecast more economic and financial time series and (2) improving the selection and ensemble method of individual forecasting models to further enhance the prediction performance.

Data Availability

The data used to support the findings of this study are included within the article.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This work was supported by the Fundamental Research Funds for the Central Universities (Grant no. JBK2003001), the Ministry of Education of Humanities and Social Science Project (Grant nos. 19YJAZH047 and 16XJAZH002), and the Scientific Research Fund of Sichuan Provincial Education Department (Grant no. 17ZB0433).