In this paper, we investigate the statistical behaviors of the stock market complex network. A hybrid model is proposed to predict the variations of five stock prices in the securities plate sub-network. This model integrates independent component analysis (ICA) and multivariate long short-term memory (Multi-LSTM) neural network to analyze the trading noise and improve the prediction accuracy of stock prices in the sub-network. Firstly, we apply ICA to deconstruct the original dataset and remove the independent components that represent the trading noise. Secondly, the rest of the independent components are given to Multi-LSTM neural network. Finally, prediction results are reconstructed from the outputs of the Multi-LSTM neural network and the corresponding mixing matrix. The experiment results indicate that the hybrid model outperforms the benchmark approaches, especially in terms of the stock market complex network.
Citation: Hongzeng He, Shufen Dai. A prediction model for stock market based on the integration of independent component analysis and Multi-LSTM[J]. Electronic Research Archive, 2022, 30(10): 3855-3871. doi: 10.3934/era.2022196
Related Papers:
[1]
Ying Li, Xiangrong Wang, Yanhui Guo .
CNN-Trans-SPP: A small Transformer with CNN for stock price prediction. Electronic Research Archive, 2024, 32(12): 6717-6732.
doi: 10.3934/era.2024314
[2]
Yufei Duan, Xian-Ming Gu, Tingyu Lei .
Application of machine learning in quantitative timing model based on factor stock selection. Electronic Research Archive, 2024, 32(1): 174-192.
doi: 10.3934/era.2024009
[3]
Peng Lu, Yuchen He, Wenhui Li, Yuze Chen, Ru Kong, Teng Wang .
An Informer-based multi-scale model that fuses memory factors and wavelet denoising for tidal prediction. Electronic Research Archive, 2025, 33(2): 697-724.
doi: 10.3934/era.2025032
[4]
Yuxia Liu, Qi Zhang, Wei Xiao, Tianguang Chu .
Characteristic period analysis of the Chinese stock market using successive one-sided HP filter. Electronic Research Archive, 2023, 31(10): 6120-6133.
doi: 10.3934/era.2023311
[5]
Xite Yang, Ankang Zou, Jidi Cao, Yongzeng Lai, Jilin Zhang .
Systemic risk prediction based on Savitzky-Golay smoothing and temporal convolutional networks. Electronic Research Archive, 2023, 31(5): 2667-2688.
doi: 10.3934/era.2023135
[6]
Changhai Wang, Jiaxi Ren, Hui Liang .
MSGraph: Modeling multi-scale K-line sequences with graph attention network for profitable indices recommendation. Electronic Research Archive, 2023, 31(5): 2626-2650.
doi: 10.3934/era.2023133
[7]
Suhua Wang, Zhen Huang, Hongjie Ji, Huinan Zhao, Guoyan Zhou, Xiaoxin Sun .
PM2.5 hourly concentration prediction based on graph capsule networks. Electronic Research Archive, 2023, 31(1): 509-529.
doi: 10.3934/era.2023025
[8]
Ju Wang, Leifeng Zhang, Sanqiang Yang, Shaoning Lian, Peng Wang, Lei Yu, Zhenyu Yang .
Optimized LSTM based on improved whale algorithm for surface subsidence deformation prediction. Electronic Research Archive, 2023, 31(6): 3435-3452.
doi: 10.3934/era.2023174
[9]
Peng Lu, Yuze Chen, Ming Chen, Zhenhua Wang, Zongsheng Zheng, Teng Wang, Ru Kong .
An improved stacking-based model for wave height prediction. Electronic Research Archive, 2024, 32(7): 4543-4562.
doi: 10.3934/era.2024206
[10]
Dewang Chen, Xiaoyu Zheng, Ciyang Chen, Wendi Zhao .
Remaining useful life prediction of the lithium-ion battery based on CNN-LSTM fusion model and grey relational analysis. Electronic Research Archive, 2023, 31(2): 633-655.
doi: 10.3934/era.2023031
Abstract
In this paper, we investigate the statistical behaviors of the stock market complex network. A hybrid model is proposed to predict the variations of five stock prices in the securities plate sub-network. This model integrates independent component analysis (ICA) and multivariate long short-term memory (Multi-LSTM) neural network to analyze the trading noise and improve the prediction accuracy of stock prices in the sub-network. Firstly, we apply ICA to deconstruct the original dataset and remove the independent components that represent the trading noise. Secondly, the rest of the independent components are given to Multi-LSTM neural network. Finally, prediction results are reconstructed from the outputs of the Multi-LSTM neural network and the corresponding mixing matrix. The experiment results indicate that the hybrid model outperforms the benchmark approaches, especially in terms of the stock market complex network.
1.
Introduction
It is believed that the stock market could be described by complex networks. The applications of complex networks have provided a new perspective for studying the in-depth mechanisms of the stock market [1]. In the fields of statistics and finance, stock market is regarded as a complex nonlinear dynamic system with numerous variables. The market is affected by economy, politics and society. Consequently, variations in the stock market are difficult to predict accurately. Due to the complexity of the stock market, some prediction models have been proposed to study and analyze the fluctuations of the market prices, mainly through historical data [2,3,4,5,6,7]. In recent studies, the prediction approaches of stock prices can mainly be divided into statistical models, econometric models, neural network models and so on.
Since the financial market data is dynamic, nonlinear, nonstationary, nonparametric and volatile, various techniques are combined with artificial prediction approaches to model the financial time series. For instance, the independent component analysis (ICA) method is applied to reduce the trading noise of the financial data [8,9]. Haifan Liu et al. [9] integrated the ICA method with the neural network model to investigate the statistical behaviors of Shanghai Composite Index fluctuations. Results proved that the proposed model outperforms the conventional single BP model. Kao et al. [10] combined nonlinear ICA with support vector regression (SVR) for stock price forecasting to improve the prediction accuracy of the traditional methods. Jianwei E et al. [11] proposed a variational hybrid method of mode decomposition (VMD), ICA and autoregressive integrated moving average (ARIMA) to analyze the influence factors of crude oil price and predict future prices. Jianwei E et al. [12] offered a novel combination technique based on ICA and gate recurrent unit neural network (GRUNN) to predict the gold price. The combination method achieved higher accuracy than the benchmark methods. Then Jianwei E et al. [8] presented a hybrid model based on extreme-point symmetric mode decomposition (ESMD), kernel independent component analysis (KICA) and least squares support vector regression (LSSVR) to predict the carbon prices from European Union Emissions Trade System (EU ETS).
Recently, neural network models have played a significant role in time series prediction fields. Especially the long short-term memory (LSTM) can memory the information for a longer period to overcome the vanishing gradient problem. It could be summarized that LSTM models are more effective in recognizing trends of the time series than the traditional neural network models [13]. Since the financial time series is nonstationary, nonlinear and with trading noise. Some hybrid prediction models are based on neural networks with historical data, such as ICA, wavelet transform (WT), autoencoder (SAE), complete ensemble empirical mode decomposition (CEEMD), ensemble empirical mode decomposition (EEMD), empirical mode decomposition (EMD), principal component analysis (PCA) and so on. These approaches are combined with LSTM to smooth the original data and further enhance the prediction accuracy. Wq A et al. [14] proposed a combination of SAE and LSTM on gas production and consumption data to improve the prediction performance and solve the gradient disappearance problem of LSTM. Zhang Y et al. [15] applied CEEMDAN and PCA to eliminate redundant information and improve the prediction response speed. Bao W et al. decomposed the stock price series by WT and fed the high-level denoising features into LSTM to forecast the next day's closing price [16]. Akhil Sethia et al. [17] applied LSTM, GRU, ANN and SVM to predict S & P500 index close prices. The ICA is implemented as the data processing to reduce the input attribute dimension from 50 to 12. These 12 attributes are given to the prediction model to optimize the prediction strategy. And these models are among the most popular approaches because of the improvements in accuracy and efficiency.
In all the surveyed studies, ICA was usually combined with predicting methods, such as gate recurrent unit neural network (GRUNN), neural networks (NN), autoregressive integrated moving average (ARIMA), support vector regression (SVR), least squares support vector regression (LSSVR), single back propagation (BP) neural network or implemented to preprocess original attributes. Nevertheless, most of the previous studies focus on the prediction of a single time series without dataset reconstruction. To our knowledge, few studies employ methodologies that take into account the relevance between multiple stocks in a stock market complex network. To fill this gap, we proposed a prediction model for the stock market based on the integration of ICA and Multi-LSTM, including a dataset reconstruction process, to expand the research perspectives on the stock market network and enhance the prediction accuracy.
The innovations of this study are as follows: Firstly, several independent components are decomposed from the same securities plate sub-network dataset by ICA method. Trading noise is recognized and reduced through dimension reduction process. Secondly, remaining independent components are conducted as the input of the Multi-LSTM model. Final predicting results are reconstructed from predicting results of the independent components. Then, experiment results indicate that ICA is suitable for stock market complex network dataset, which could match the input type for Multi-LSTM with less calculation amount. And the proposed ICA-Multi-LSTM model produced better performance than the benchmark approaches.
The organization of this paper is as follows. Section 2 describes the dataset and outlines the proposed model for predicting the stock price in a stock market complex network. The empirical results and comparisons of performance are listed in Section 3. Section 4 concludes the paper. Finally, acknowledgments and references are also presented.
2.
Materials and methods
2.1. Stock market dataset
In this study, we choose the securities plate sub-network of the stock market complex networks in reference [1]. The dataset includes five attributes: daily opening prices, daily closing prices, daily high prices, daily low prices and daily trading volumes of 5 securities companies. The invalid data and empty data are dropped. The dataset is from Jul. 26th, 2015 to Dec. 31st, 2020 (Data sources: Resset.cn) which contains 32,675 data points of 1307 trading days. The samples of the dataset are presented in Table 1.
Table 1.
Trading data samples of 5 stocks on Sept. 7th, 2020.
The network structure of the securities plate sub-network is shown in Figure 1. As shown in Figure 1, stocks are connected according to the correlation coefficient of the stock price series [1]. The correlation relationships of the stocks could be calculated by correlation analysis or clustering methods, such as Pearson correlations, detrended cross-correlation analysis (DCCA) [1], K-means clustering [18] and feature-based clustering [19]. The correlation results can express the historical correlation and trend correlation of stock prices. It is believed that multiple stock price data of the same cluster instead of a single self-data, as input data to train LSTM, helps to improve prediction accuracy [18,19,20]. The overview of the dataset time series is shown in Figure 2. The graph shows that there exists a relevant tendency between the time series of five stocks, which means high correlations of the stocks in the same complex network. In this study, we select the daily closing prices, daily high prices, daily low prices and daily trading volumes as input of the prediction model to predict the daily opening prices and verify the predicting results.
Figure 1.
The network structure of the securities plate sub-network [1].
In this study, we obtain a prediction model for the stock market based on the integration of ICA and Multi-LSTM. For this purpose, we introduce a brief theoretical description of the two methods.
2.2. Independent component analysis
The ICA method is a novel feature extraction technique, which is used to find the hidden influence factors from the mixed source data [10,17]. Since the trading data of the stock market is nonlinear, nonstationary and mixed with noise pollution, the ICA method can separate the independent components of mixed time series data. This technology could be employed to solve the blind source separation problem, time series denoising and so on. The ICA method is usually applied to the field of stock market data denoising. Several independent components (ICs) are separated from the original data, which could be analyzed, selected and reconstructed to the denoising data. Generally, the ICA algorithm consists of the following two steps.
Step 1: Supposing that the m × n trading data time series matrix X=[x1,x2,…,xi,…,xm]T, where xi represents the ith trading data with the length of n and i = 1, …, m. Based on the ICA algorithm, the matrix X could be represented by
X=AS=∑mi=1aisi
(1)
where Am×m=[a1,a2,…,ai,…am]T represents the mixing matrix which includes noise trading data and Sm×n=[s1,s2,…,si,…sm]T represents the separated trading data matrix. The purpose of the ICA algorithm is to find a separating matrix W which could be represented by
Y=[yi]=WX
(2)
where yi is the ith independent component. The independent components must be as independent with each other as possible. If we let the separating matrix W=A−1, and the real trading data matrix S could be reconstructed by
Y=WX=W·A·S=A−1AS=S
(3)
All the yi(i = 1, …, m) are considered independent if they are non-Gaussian distribution. The gaussianity of yi could be given by
J(y)=H(ygauss)−H(y)
(4)
where ygauss is a Gaussian random vector with the same covariance matrix as y and H(y)=−∫p(y)logp(y)dy. Some fast algorithms are applied in engineering on account of the huge amount of calculations [21].
Step 2: Since we have got the independent components from the original data, independent components could be evaluated and analyzed. We could remove the independent components that represent the trading noise and reconstruct the real trading data matrix by
XR=∑i∈S1≤i≤maiyi,1≤i≤m
(5)
where XR=[xR1,xR2,…,xRm]T represents the reconstructed real trading data matrix and the mixing matrix A=[a1,a2,…,am]=W−1 and yi is the ith independent component with less trading noise.
2.3. Multi-LSTM neural network
The long short-term memory (LSTM) neural network is an improved version of the recurrent neural network (RNN). The method is designed to solve the long-term dependency problem and avoid the vanishing gradient problem or the radiant exploding problem of the traditional RNN [14]. LSTM network can predict time series from multivariate input data with higher accuracy. The structure of the Multi-LSTM network is comprised of several LSTM cells. There are input gate (i), output gate (o), forget gate (f) and cell state (C), as shown in Figure 3.
Figure 3.
Structures of the Multi-LSTM neural network.
As is presented in Figure 3, xt−1,xt,xt+1 are the input data at time t-1, t, t+1.ht−1,ht,ht+1 are the output data at time t-1, t, t+1. There are input gate i, output gate o, forget gate f and cell state C in each cell. The σ and tanh are the activation functions. W is the weight coefficient and b is the deviation matrix. The cell state Ct at time t could be adjusted by the forget gate ft and input gate it. The forget gate ft could remember or forget the cell state Ct-1. The input gate it could allow or forbid updating of the cell state. The cell state Ct could be transmitted to the next cell. The structure of the LSTM establishes advantages to predict the long dependent and nonlinear time series.
In this subsection, we describe the details of the proposed prediction model. The daily opening price could be predicted by multivariate factors by this model, such as daily closing prices, daily high prices, daily low prices and daily trading volumes. The hybrid model consists of the following four steps:
Step 1: data preprocessing
We get the historical attributes including daily closing prices, daily high prices, daily low prices and daily trading volumes as the input data to predict daily opening prices. The daily opening price matrix AOppr, daily high price matrix AHipr, daily low price matrix ALopr, daily closing price matrix AClpr and daily trading volume matrix ATrdvol are extracted from the historical trading data by certain data pre-processing techniques such as data purifying and data smoothing. Each row of the matrix represents the stocks in the securities plate sub-network and each column represents the time series data of each stock. The brief example is as follows:
AOppr=[a1,a2,…,ai,…,am]T
(7)
As shown in Eq (7), AOppr is the m × n daily opening price matrix, where m is the number of stocks in the securities plate sub-network and ai is the daily opening price vector of the stock i, where i = 1, …, m. The length of ai is the total trading days n.
Step 2: ICA and dimension reduction
In this step, we begin with the independent component analysis to separate the ICs from AOppr,AHipr,ALopr,AClpr,ATrdvol and calculate the contribution rate of each IC, followed by a dimension reduction process to get rid of the ICs with low contribution rates that represent the trading noise. Finally, the IC matrices with less trading noise are considered as the input of the Multi-LSTM model. The IC matrices are represented as daily opening price IC matrix COppr, daily high price IC matrix CHipr, daily low price IC matrix CLopr, daily closing price IC matrix CClpr, daily trading volume IC matrix CTrdvol, separating matrix W and mixing matrix A. A brief example of the IC matrix COppr is as follows:
COppr=[c1,c2,…,ci,…,cl]T
(8)
where COppr is the l × n IC matrix of the daily opening price, where l is the number of independent components where the noise is eliminated. And ci is the ith IC vector, where i = 1, …, l. The length of ci is the total trading days n.
Step 3: training and predicting independent components with Multi-LSTM
The dataset CHipr,CLopr,CClpr,CTrdvol,COppr are divided into the training set, validation set and test set. And CHipr,CLopr,CClpr,CTrdvol are given to Multi-LSTM to train and predict the COppr. So the training and predicting results COpprpdt could be obtained from the output of Multi-LSTM prediction results:
COpprpdt=[c1_pdt,c2_pdt,…,ci_pdt,…,cl_pdt]T
(9)
where Ci_pdt represents the output results of the ith IC vector. The Ci_pdt spliced by training results validation results and predicting results respectively.
Step 4: final predictions reconstruction
The final prediction results AOppr_pdt could be reconstructed by COppr_pdt and mixing matrix A by
AOppr_pdt=A·COppr_pdt
(10)
where AOppr_pdt=[aOppr_pdt_1,aOppr_pdt_2,…,aOppr_pdt_i,…,aOppr_pdt_m]T and aOppr_pdt_i are the final prediction results of the daily opening price of stock i. Hence, the framework of the proposed model is reflected in Figure 4.
3.1. Independent component analysis and denoising effect
The first stage of the proposed model is independent component separation, in which the original trading data are separated into independent components via the ICA method. Further, the trading noise could be eliminated via the methods mentioned in Section 2.
In this subsection, we get five independent component matrices of the original data AOppr,AHipr,ALopr,AClpr,ATrdvol and calculate the contribution rate of each IC based on eigenvalues of the IC matrices by the method in [22,23]. The eigenvalues and contribution rates of the IC matrices are listed in Tables 2 and 3.
As is illustrated in Tables 2 and 3, the cumulative contribution rates of the first three independent components are 98.07% in the daily opening price IC matrix, 98.22% in the daily high price IC matrix, 98.03% in the daily low price IC matrix, 98.13% in the daily closing price IC matrix, 95.66% in daily trading volume IC matrix. It is now understood that the contribution rate is used to select the independent components representing the real trading information and trading noise [8,9]. In this study, we find that the IC 4 and IC 5 with the lowest contribution rates denote the noise trading data. We guarantee the first three independent components that contain the real trading information are IC 1, IC 2 and IC 3. These ICs are used as the input of the Multi-LSTM. The independent components with less trading noise are in Figure 5. As shown in Figure 5, the fluctuation patterns of the independent components may contain the trading information of periodicity, seasonality or random events. These independent components are selected as the input for the Multi-LSTM model.
Figure 5.
Independent components with less trading noise: opening prices (a), closing prices (b), high prices (c), low prices (d), and trading volumes (e) of the five securities stocks.
3.2. Prediction results of the independent components
The second stage of the proposed model is the prediction module. The independent components with less trading noise are given to the Multi-LSTM model to predict the daily opening price in this stage. To investigate the effectiveness of the proposed model, the dataset structure and the Multi-LSTM network architecture are also illustrated in this subsection.
Dataset structure All the data sets are constructed from the independent components with less trading noise. In this study, there are 1–1052 trading days in the training set, 1053–1300 trading days in the validation set and 1301–1307 trading days in the test set. This study adopts the rolling prediction approach. In other words, we use the previous 60 days to predict the following one day iteratively. For example, the model uses the 1–60 days of daily high price, daily low price, daily closing price, and daily trading volume IC matrices as the input attributes to train and predict the 61st day of the daily opening price, 2–61 days to train and predict the 62nd day and so on. As a result, we could get training predictions of 61–1052 days, validation predictions of 1053–1300 days and test predictions of 1031–1037 days respectively.
Multi-LSTM neural network architecture The architecture of the Multi-LSTM neural network in this article is illustrated in Figure 6. As is shown in Figure 6, the network contains five layers, including the input layer, two hidden layers, the dropout layer and the output layer. We input the IC matrices CHipr,CLopr,CClpr,CTrdvol into the input layer to train and predict COppr. There are two hidden layers in the model according to the computing speed and prediction accuracy. It is practically proved that the model could achieve high precision and good effect if the numbers of the two hidden layers are very close. And the numbers of neural nodes in the two hidden layers are 64 and 10. The dropout layer is set to optimize the neural network and the deactivation rate is 0.25. The optimization algorithm is Adam and the dynamic learning rate is 0.01. To train the Multi-LSTM model, the parameters converge to 30 epochs with Adam Optimizer. The batch size is 1 according to the data size and the architecture of the Multi-LSTM.
Figure 6.
The architecture of the Multi-LSTM neural network.
Figure 7 shows the predicting results of the opening price independent components with less trading noise. The predicting results of the three stock daily opening price independent components contain the train predictions and future predictions. From Figure 7, we consider that the predicting results have the same trends as the actual daily opening price.
Figure 7.
Prediction results of the opening price IC 1, IC 2 and IC 3.
In this stage, we reconstruct the final predicting results AOppr_pdt by the mixing matrix A and COppr_pdt.
AOppr_pdt=A·COppr_pdt
(11)
The COppr_pdt is consisted of 1–60 days of original data, 61–1052 days of training prediction, 1053–1300 days of validation prediction and 1301–1307 days of testing prediction to fit the length of the original data and mixing matricesA in Eq (11). It could be noted that there exists subtle error because the mixing matrix A is an estimated value by the fast ICA algorithm [21].
3.4. Performance criteria
The prediction performance could be evaluated by the following performance measures: Mean Absolute Error (MAE), Root Mean Square Error (RMSE) and Mean Square Error (MSE). The corresponding definitions of N trading days are given as follows:
• Mean Absolute Error (MAE):
MAE=1N∑Ni=1|y'i−yi|
(12)
• Root Mean Square Error (RMSE):
RMSE=√1N∑Ni=1(y'i−yi)2
(13)
• Mean Square Error (MSE):
MSE=1N∑Ni=1(y'i−yi)2
(14)
where y'i is the final prediction result in the trading day i. The yi is the original data in the trading day i, and i = 1, …, N. The N is the number of trading days.
Figure 8.
Final prediction results reconstructed by A and COppr_pdt.
The performance criteria are in three parts: training performance criteria from 61 to 1052 trading days, validation performance criteria from 1053 to 1300 trading days and test performance criteria from 1301 to 1307 trading days. A lower value of the performance criteria means lower prediction error as well as higher accuracy of the prediction model. In Table 4, we can see the fact that the prediction accuracy of the ICA-Multi-LSTM model outperforms the traditional methods in most cases with lower performance criteria, which benefits from the ICA nonlinear and nonstationary data denoising preprocess strategy.
However, since the time series are complicated and changeable. The specific predicting results depend on the structural properties of the time sequence in some cases. In order to compare the overall performance of the proposed model on the securities plate stock market complex network, we introduce M-MAE, M-RMSE and M-MSE which could evaluate the average predicting error. These criteria could indicate the aggregate performance on the stock network. The corresponding definitions are given as follows:
• M-MAE:
M-MAE = 1M×N∑Mj=1∑Ni=1|y'ji−yji|
(15)
• M-RMSE:
M-RMSE=√1M×N∑Mj=1∑Ni=1(y'ji−yji)2
(16)
• M-MSE:
M-MSE = 1M×N∑Mj=1∑Ni=1(y'ji−yji)2
(17)
where y'ji is the final prediction result of stock j in the trading day i,yji is the original data of stock j in the trading day i, and i = 1, …, N; j = 1, …, M. The N is the number of trading days and the M is the total number of stocks in the securities plate sub-network.
The final comparative results are listed in Table 5. From Table 5, the following facts could be yielded. From the perspective of complex networks, the ICA-Multi-LSTM method is superior to the Multi-LSTM method, which shows the advantage of the integration of ICA and Multi-LSTM. The primary reason is that trading noise is reduced via the ICA dimension reduction process which verifies the improvement of the prediction accuracy.
Table 5.
Performance criteria on the stock networks.
To better capture the performance of ICA-Multi-LSTM on the stock market network, we comprehensively introduce the other eight methods and evaluate their predictive criteria in comparison experiments. The former five models are used to check the effectiveness of ICA and other variants. The last three models are used as the performance benchmark. In the CEEMD-Multi -LSTM, EEMD-Multi-LSTM, EMD-Multi-LSTM and WT-Multi-LSTM, every single time series of daily closing prices, daily high prices, daily low prices and daily trading volumes are decomposed by CEEMD [24], EEMD [25], EMD [26] and WT [16] to reduce the trading noise. And the denoised time series are reconstructed from the first 5 intrinsic mode functions (IMFs) according to the cumulative contribution rates to predict the actual daily opening prices [15]. In the WT-Multi-LSTM, the wavelet basis function is Haar and the number of retained principal components for denoising is 3. The CEEMD-Multi-LSTM, EEMD-Multi-LSTM, EMD-Multi-LSTM, WT-Multi-LSTM and Multi-LSTM are with the same LSTM model parameters as ICA-Multi-LSTM. But the CEEMD, EEMD, EMD and WT methods are usually suitable for single signal decomposition. So each denoised time series above are reconstructed from IMFs or wavelet components (WCs) before fed into Multi-LSTM. This is different from ICA-Multi-LSTM in which the final predicting results are reconstructed from predicting results of the independent components. And ICA-Multi-LSTM is suitable for stock market complex network dataset, which could handle the multiple time series with less calculation amount. Prophet is a recent forecasting model introduced by Facebook [27]. This model is proposed to predict the daily time series, including seasonal patterns. Single-LSTM is a single input model with similar parameters as ICA-Multi-LSTM. It uses the single input of historical opening prices to predict the future. 1D-CNN is a traditional deep learning framework that is based on 1D convolutional neural networks [28]. Experimental results are illustrated in Table 5. As shown in Table 5, the final average performance criteria are smaller for ICA-Multi-LSTM.
Ulteriorly, we apply the Diebold-Mariano Test on the M-MAE to interpret the statistical significance and determine whether there is a striking difference among the comparisons of the final predicting results [29]. The p-value of the Diebold-Mariano Test is 0.001, which indicates that there is a striking difference between the final predicting results and the effectiveness of the ICA denoising process could be verified. Meanwhile, we adopt the Wilcoxon Signed-rank Test to evaluate the robustness of the proposed ICA-Multi-LSTM model. The p-value of the Wilcoxon Signed-rank Test is less than 0.001, which indicates that the proposed model with proper dimension reduction denoising process provides better predicting results. On the basis of the above analysis, the proposed model shows better performance not only from the perspective of single stock but also in terms of the whole stock network through comprehensive comparisons. The prediction accuracy could be improved by the ICA noise reduction procedure.
4.
Conclusions
In this paper, a hybrid model is proposed that integrates the ICA and Multi-LSTM neural network to improve the prediction performance on the stock market complex networks. Previous researches have established certain prediction models from the perspective of single time series in economics, such as ESMD-KICA-LSSVR [8], ICA-BP [30], Bayesian-LSTM [31], ICA-SVR [32] and so on, but few studies have taken into account the relevance of the stock market complex network and ICA reconstruction, as well as the nonlinear, nonstationary and noise pollution of the stock market time series. In order to quantify the effectiveness of ICA on the prediction model and prediction performance on the stock market networks, we analyze the ICA results and denoising effect on the original data. And we make an econometric analysis about prediction results and performance comparison with the benchmark approaches. Firstly, the ICA is employed to the original data and identify the independent components. Five independent components are separated from the stock network. Meanwhile, the trading noise is reduced through ICA contribution degree analysis and dimension reduction process. Secondly, the Multi-LSTM prediction model is applied to each IC with less trading noise rather than the original time series. And the final prediction results are reconstructed from predicting results of the independent components. Finally, comparable experiments confirm that the proposed model outperforms the LSTM model without ICA from the perspective of the single stock in most cases. Additionally, since the predicting results depend on the structural properties of the time sequence in some cases, we deploy the securities plate stock market complex network as the target dataset to compare the overall performance of the proposed model. Experimental results show that the proposed model outperforms the benchmark models in terms of the stock market complex network prediction. This model is suitable for stock market complex network dataset, which could handle the multiple time series with less calculation amount. But it still needs to be optimized in practice. We believe that such studies are relevant for a better understanding of the future temporal evolution tendency of the stock market. And such methods could be used to build more effective prediction models in further study.
Conflict of interest
The authors declare that there are no conflicts of interest regarding the publication of this paper.
References
[1]
H. He, S. Dai, Effectiveness of price limit on stock market network: A time-migrated DCCA approach, Complexity, 2021 (2021). https://doi.org/10.1155/2021/3265843 doi: 10.1155/2021/3265843
[2]
S. Kumar Chandar, Hybrid models for intraday stock price forecasting based on artificial neural networks and metaheuristic algorithms, Pattern Recognit. Lett., 147 (2021), 124–133. https://doi.org/10.1016/j.patrec.2021.03.030 doi: 10.1016/j.patrec.2021.03.030
[3]
A. Bose, C. Hsu, S. S. Roy, K. C. Lee, B. Mohammadi-ivatloo, S. Abimannan, Forecasting stock price by hybrid model of cascading Multivariate Adaptive Regression Splines and Deep Neural Network, Comput. Electr. Eng., 95 (2021), 107405. https://doi.org/10.1016/j.compeleceng.2021.107405 doi: 10.1016/j.compeleceng.2021.107405
[4]
A. Thakkar, K. Chaudhari, A comprehensive survey on deep neural networks for stock market: The need, challenges, and future directions, Expert Syst. Appl., 177 (2021), 114800. https://doi.org/10.1016/j.eswa.2021.114800 doi: 10.1016/j.eswa.2021.114800
[5]
H. Na, S. Kim, Predicting stock prices based on informed traders' activities using deep neural networks, Econ. Lett., 204 (2021), 109917. https://doi.org/10.1016/j.econlet.2021.109917 doi: 10.1016/j.econlet.2021.109917
[6]
S. Wang, Z. Li, J. Zhu, Z. Lin, M. Zhong, Stock selection strategy of A-share market based on rotation effect and random forest, AIMS Math., 5 (2020), 4563–4580. https://doi.org/10.3934/math.2020293 doi: 10.3934/math.2020293
[7]
Z. Dai, H. Zhou, X. Dong, Forecasting stock market volatility: the role of gold and exchange rate, AIMS Math., 5 (2020), 5094–5105. https://doi.org/10.3934/math.2020327 doi: 10.3934/math.2020327
[8]
J. E, J. Ye, L. He, H. Jin, A denoising carbon price forecasting method based on the integration of kernel independent component analysis and least squares support vector regression, Neurocomputing, 434 (2021), 67–79. https://doi.org/10.1016/j.neucom.2020.12.086 doi: 10.1016/j.neucom.2020.12.086
[9]
C. Lu, Integrating independent component analysis-based denoising scheme with neural network for stock price prediction, Expert Syst. Appl., 37 (2010), 7056–7064. https://doi.org/10.1016/j.eswa.2010.03.012 doi: 10.1016/j.eswa.2010.03.012
[10]
L. Kao, C. Chiu, C. Lu, J. Yang, Integration of nonlinear independent component analysis and support vector regression for stock price forecasting, Neurocomputing, 99 (2013), 534–542. https://doi.org/10.1016/j.neucom.2012.06.037 doi: 10.1016/j.neucom.2012.06.037
[11]
J. E, Y. Bao, J. Ye, Crude oil price analysis and forecasting based on variational mode decomposition and independent component analysis, Physica A, 484 (2017), 412–427. https://doi.org/10.1016/j.physa.2017.04.160 doi: 10.1016/j.physa.2017.04.160
[12]
J. E, J. Ye, H. Jin, A novel hybrid model on the prediction of time series and its application for the gold price analysis and forecasting, Physica A, 527 (2019), 121454. https://doi.org/10.1016/j.physa.2019.121454 doi: 10.1016/j.physa.2019.121454
[13]
C. Fang, F. Marle, Dealing with project complexity by matrix-based propagation modelling for project risk analysis, J. Eng. Des., 24 (2013), 239–256. https://doi.org/10.1080/09544828.2012.720014 doi: 10.1080/09544828.2012.720014
[14]
W. Qiao, W. Liu, E. Liu, A combination model based on wavelet transform for predicting the difference between monthly natural gas production and consumption of U.S., Energy, 235 (2021), 121216. https://doi.org/10.1016/j.energy.2021.121216 doi: 10.1016/j.energy.2021.121216
[15]
Y. Zhang, B. Yan. M. Aasma, A novel deep learning framework: Prediction and analysis of financial time series using CEEMD and LSTM, Expert Syst. Appl., 159 (2020), 113609. https://doi.org/10.1016/j.eswa.2020.113609 doi: 10.1016/j.eswa.2020.113609
[16]
W. Bao, J. Yue, Y. Rao, A deep learning framework for financial time series using stacked autoencoders and long-short term memory, PLOS ONE, 12 (2017), e0180944. https://doi.org/10.1371/journal.pone.0180944 doi: 10.1371/journal.pone.0180944
[17]
P. Comon, Independent component analysis, A new concept, Signal Process., 36 (1994), 287–314. https://doi.org/10.1016/0165-1684(94)90029-9 doi: 10.1016/0165-1684(94)90029-9
[18]
Y. Chen, J. Wu, Z. Wu, China's commercial bank stock price prediction using a novel K-means-LSTM hybrid approach, Expert Syst. Appl., 202 (2022), 117370. https://doi.org/10.1016/j.eswa.2022.117370 doi: 10.1016/j.eswa.2022.117370
[19]
K. Bandara, C. Bergmeir, S. Smyl, Forecasting across time series databases using recurrent neural networks on groups of similar series: A clustering approach, Expert Syst. Appl., 140 (2020), 112896. https://doi.org/10.1016/j.eswa.2019.112896 doi: 10.1016/j.eswa.2019.112896
[20]
H. G. Seedig, R. Grothmann, T. A. Runkler, Forecasting of clustered time series with recurrent neural networks and a fuzzy clustering scheme, in 2009 International Joint Conference on Neural Networks, IEEE, (2009), 2846–2853. https://doi.org/10.1109/IJCNN.2009.5178775
W. Dai, J. Wu, C. Lu, Combining nonlinear independent component analysis and neural network for the prediction of Asian stock market indexes, Expert Syst. Appl., 39 (2012), 4444–4452. https://doi.org/10.1016/j.eswa.2011.09.145 doi: 10.1016/j.eswa.2011.09.145
[23]
Y. Ouyang, Evaluation of river water quality monitoring stations by principal component analysis, Water. Res., 39 (2005), 2621–2635. https://doi.org/10.1016/j.watres.2005.04.024 doi: 10.1016/j.watres.2005.04.024
[24]
F. Zhou, Z. Huang, C. Zhang, J. Yan, Carbon price forecasting based on CEEMDAN and LSTM, Appl. Energy, 311 (2022), 118601. https://doi.org/10.1016/j.apenergy.2022.118601 doi: 10.1016/j.apenergy.2022.118601
[25]
Y. Wu, Q. Wu, J. Zhu, Improved EEMD-based crude oil price forecasting using LSTM networks, Physica A, 516 (2019), 114–124. https://doi.org/10.1016/j.physa.2018.09.120 doi: 10.1016/j.physa.2018.09.120
[26]
M. A. Colominas, G. Schlotthauer, M. E. Torres, Improved complete ensemble EMD: A suitable tool for biomedical signal processing, Biomed. Signal Process. Control, 14 (2014), 19–29. https://doi.org/10.1016/j.bspc.2014.06.009 doi: 10.1016/j.bspc.2014.06.009
[27]
D. Borges, M. C. V. Nascimento, COVID-19 ICU demand forecasting: A two-stage Prophet-LSTM approach, Appl. Soft Comput., 125 (2022), 109181. https://doi.org/10.1016/j.asoc.2022.109181 doi: 10.1016/j.asoc.2022.109181
[28]
S. Mehrkanoon, Deep shared representation learning for weather elements forecasting, Knowledge Based Syst., 179 (2019), 120–128. https://doi.org/10.1016/j.knosys.2019.05.009 doi: 10.1016/j.knosys.2019.05.009
[29]
F. X. Diebold, R S. Mariano, Comparing predictive accuracy, J. Bus. Econ. Stat., 13 (1995), 134–144. https://doi.org/10.2307/1392185 doi: 10.2307/1392185
[30]
H. Liu, J. Wang, K. Vajravelu, Integrating independent component analysis and principal component analysis with neural network to predict Chinese stock market, Math. Probl. Eng., 2011 (2011), 1–15. https://doi.org/10.1155/2011/382659 doi: 10.1155/2011/382659
[31]
B. Huang, Q. Ding, G. Sun, H. Li, Stock Prediction based on Bayesian-LSTM, in ICMLC 2018: Proceedings of the 2018 10th International Conference on Machine Learning and Computing, (2018), 128–133. https://doi.org/10.1145/3195106.3195170
[32]
C. Lu, T. Lee, C. Chiu, Financial time series forecasting using independent component analysis and support vector regression, Decis. Support Syst., 47 (2009), 115–125. https://doi.org/10.1016/j.dss.2009.02.001 doi: 10.1016/j.dss.2009.02.001
This article has been cited by:
1.
Tingcheng Mo, Chi Xie, Kelong Li, Yingbo Ouyang, Zhijian Zeng,
Transmission effect of extreme risks in China's financial sectors at major emergencies: Empirical study based on the GPD-CAViaR and TVP-SV-VAR approach,
2022,
30,
2688-1594,
4657,
10.3934/era.2022236
2.
Cheng Zhao, Ping Hu, Xiaohui Liu, Xuefeng Lan, Haiming Zhang,
Stock Market Analysis Using Time Series Relational Models for Stock Price Prediction,
2023,
11,
2227-7390,
1130,
10.3390/math11051130
3.
Changhai Wang, Jiaxi Ren, Hui Liang,
MSGraph: Modeling multi-scale K-line sequences with graph attention network for profitable indices recommendation,
2023,
31,
2688-1594,
2626,
10.3934/era.2023133
4.
Mehmet Sarıkoç, Mete Celik,
PCA-ICA-LSTM: A Hybrid Deep Learning Model Based on Dimension Reduction Methods to Predict S&P 500 Index Price,
2024,
0927-7099,
10.1007/s10614-024-10629-x
5.
Deepa Raghunathan, M. Krishnamoorthi,
Optimizing Stock Predictions With Bi-Directional LSTM and Levy Flight Fuzzy Social Spider Optimization (LFFSSO),
2025,
21,
1552-6283,
1,
10.4018/IJSWIS.367280
Hongzeng He, Shufen Dai. A prediction model for stock market based on the integration of independent component analysis and Multi-LSTM[J]. Electronic Research Archive, 2022, 30(10): 3855-3871. doi: 10.3934/era.2022196
Hongzeng He, Shufen Dai. A prediction model for stock market based on the integration of independent component analysis and Multi-LSTM[J]. Electronic Research Archive, 2022, 30(10): 3855-3871. doi: 10.3934/era.2022196
Figure 1. The network structure of the securities plate sub-network [1]
Figure 2. The opening prices (a), closing prices (b), high prices (c), low prices (d) and trading volumes (e) of 5 securities stocks
Figure 3. Structures of the Multi-LSTM neural network
Figure 4. The framework of the proposed model
Figure 5. Independent components with less trading noise: opening prices (a), closing prices (b), high prices (c), low prices (d), and trading volumes (e) of the five securities stocks
Figure 6. The architecture of the Multi-LSTM neural network
Figure 7. Prediction results of the opening price IC 1, IC 2 and IC 3
Figure 8. Final prediction results reconstructed by A and COppr_pdt