Abstract
The stock price changes rapidly and is highly nonlinear in the financial market. One of the common concerns of many scholars and investors is how to accurately predict the stock price and the trend of rising and falling in a short time. Machine learning and deep learning techniques have found their place in financial institutions thanks to the ability of time series data prediction with high precision. However, the prediction accuracy of these models is still far from satisfactory. Most existing studies use original, single prediction algorithms that cannot overcome inherent limitations. This study proposes a hybrid model using principal component analysis (PCA) and backpropagation (BP) neural networks. The historical records of China Merchants Bank are used for data collection from 2015 to 2021. PCA preprocesses the original data to reduce the dimensionality and is then adopted by the BP neural network to predict the stock closing price of China Merchants Bank. We compare and analyze the PCA–BP model with three training algorithms, and the results indicate that the Bayesian regularization algorithm performs best. Besides, we perform the stock prediction using a traditional exponential smoothing approach. The experiment results show that the predicted stock closing price is close to the actual value, and the mean absolute percentage error can reach 0.0130, which is more significant than the traditional approach. Furthermore, A TOPSIS approach is utilized to evaluate the robustness of the proposed model. Finally, we demonstrate the usability of the designed hybrid model by predicting the stock price of another selected stock.
1. Introduction
In the 1929 crash, the Dow fell 82.30%, which means the investor lost 82.30% on average. After China’s stock market crash, the Shanghai Composite Index fell from 6,124.04 points to 1,664.93 points, losing 22 trillion yuan and a per capita loss of 1,30,000 yuan [1]. The stock market crash warns every investor of the importance of risk prevention and investment analysis in the stock market era. Thanks to the real-time update and disclosure of stock market and industry data, researchers can analyze and explore the operation law of stock prices through historical data to predict the stock price trend. The appropriate mathematical model for stock price prediction can reduce investment risk and improve the decision-making efficiency of investors. The traditional stock price prediction method is mainly econometrics. However, the stock price can be affected by many other factors. Consequently, it is difficult for these traditional mathematical models to consider all these factors to make accurate predictions.
The autoregressive integrated moving average (ARIMA) was widely used for time series prediction because of its statistical characteristics [2]. However, this model can only extract linear features from data. It is challenging to value stocks and estimate their future performance as long-term stock values are inherently unpredictable. Recently, various machine learning algorithms, including support vector machine (SVM), gradient-boosted regression trees, and random forecasts, have benefited from integrating statistics and learning models. These techniques may reveal complex patterns with nonlinear properties and certain relations that are difficult to find using linear algorithms. SVM is a hotspot in stock prediction as it can avoid local minima, overfitting, and dimension disasters often encountered in nonlinear models [3]. The use cases of SVM in time series analysis are also called support vector regression (SVR). However, SVR still has problems such as kernel function selection, tuning, and shallow feature extraction [4].
A deep neural network transmits more layers and has a more complex structure. It can transform shallow information in data into more abstract high-feature information [5] with solid performance and broad applicability. A recurrent neural network (RNN) can handle dependencies in time series data, so it is widely used in stock price prediction research [6]. In RNN, the previous layers may stop learning as the gradient disappears. It may forget what it sees in the long-term memory and thus has only a short-term memory [7]. Long short-term memory (LSTM) has a similar control flow to the RNN. It processes the data that transmit information during forward propagation.
The difference between these two lies in the different processing processes within the cell. LSTM comprises three gates: forget, input, and output. During training, these gates can learn what information to save or forget [8]. This is precisely what researchers seek from the vast amount of historical stock trading data. The conventional neural network (CNN) is widely used in image recognition, text recognition, target recognition, and target detection [9]. Due to its ability to extract local and in-depth features from data [10], it can investigate features for predicting the future movement of markets. In recent years, many studies combined financial news with quantitative indicators to improve the performance of stock prediction based on behavioral finance theory [11] with the development of natural language processing.
Despite the widespread use of data mining and machine learning techniques in the financial sector, foreign academics’ research focuses mainly on optimizing specific algorithms and the foreign stock market. Although deep learning methods and attention mechanisms have improved feature representation, the complexity of stock data often leads to the risk of overfitting. Besides, these studies utilize standard performance metrics, such as mean squared errors, to evaluate the performance of the model predictions. There is no consensus on the accuracy of the model predictions as the dataset can be changed. Moreover, few studies investigate the robustness of the prediction models. The market complexity of China is higher, and the local stock market research is currently backward.
Financial markets are complicated, and the stock price can be affected by many inherently complex human factors, including public opinion, the political environment, and news events, which cause noise in stock data. A single model cannot cover all aspects, and the inherent attributes of the model itself have inherent limitations. This paper aims to create a reasonably accurate and reliable stock forecasting valuation model for domestic investors. This paper proposes a hybrid framework based on the combination of principal component analysis (PCA) and backpropagation (BP) neural network to predict the closing price of the stock market in China. PCA is utilized to simplify the data dimension and eliminate redundant information. PCA, however, is unable to uncover the data’s nonlinear connection. The self-learning capability of the BP neural network [12], which can actualize any complicated nonlinear mapping, makes it a suitable model for stock price prediction. The historical stock data of China Merchants Bank are utilized for training and testing the proposed model. Different training algorithms are evaluated along with the PCA–BP neural network. We compare and analyze the performance of the proposed model with a traditional method called the exponential smoothing model. A TOPSIS-based approach is used to analyze the robustness of the prediction model.
The contributions of this paper are summarized as follows:(1)This paper presents a hybrid model using PCA and BP neural networks to forecast the stock closing price of the Chinese stock market.(2)The hybrid model is tested with different training algorithms to determine the optimal model. Except for common performance indicators such as the mean square error (MSE) and mean absolute percentage error (MAPE), this paper adopts a TOPSIS-based approach for ranking the models to evaluate the robustness of different models.(3)The usability of the optimal model has been validated by predicting another stock’s stock price. The findings of this study may be helpful to Chinese investors since they might give them the knowledge that will help them make educated choices about their investments and the diversity of their portfolios.
The rest of this paper is organized as follows: Section 2 overviews the related work of stock price prediction; Section 3 introduces the methodology of this paper; Section 4 introduces the relevant theory; Section 5 presents the experiment results; Section 6 concludes the whole paper and outlines some future research directions.
2. Related Work
The time series problem of stock market forecasting determines potential future direction or price value using historical price data. However, estimating this using typical time series techniques is difficult since the stock market data are not linear and are affected by many aspects. Numerous studies on various time series prediction methods have been carried out for decades. Time series predictions have undergone several stages, including exponential smoothing, autocorrelation, moving averages, and regression prediction [13]. These prediction methods only use autocorrelation or simple linear regression to conduct the prediction process, including the large information granularity. Hence, the accuracy is poor and has limitations by adapting various data types.
To evaluate the movement of stock prices, Ticknor [14] proposes a new method of Bayesian regularization (BR) with an artificial neural networks (ANNs) to predict financial market behavior, which reduces the possibility of overfitting and overtraining. The results indicate that the proposed model can improve the prediction and generalization ability of the ANN. Besides, the error of the prediction results is tiny, even without data preprocessing, seasonal testing, or systematic analysis. The results indicate that neural networks have advantages, such as solid learning ability, better inclusiveness to noisy data, and fine nonlinear mapping ability, which have gradually become a popular model for stock prediction. Many scholars have begun contributing insight into the application effect of neural network models, and various comparative analyses have been conducted. For example, Adebiyi et al. [15] compare and analyze the BP neural network’s and ARIMA’s prediction performance on New York Stock Exchange time series data. The results show that the established models using these two theories can achieve more significant performance on the stock price prediction.
Büyükşahin and Ertekin [16] propose a hybrid method of ARIMA-ANN neural network, which shows that ARIMA has better prediction accuracy in static data while ANN is more suitable for nonstationary data. Hu and Zhu [17] compare and analyze the stepwise regression with BP neural network on short-term stock price prediction. The results show that the prediction errors of these two models do not differ significantly. Besides, the model’s prediction accuracy correlates with the stock denomination and fluctuation range. In contrast to statistical and machine learning methods within stock market prediction, many researchers recently utilized deep learning techniques. Al-Nefaie and Aldhyani [18] use multilayer perceptron (MLP) and LSTM models to predict fluctuations in the Saudi Stock Exchange. The results indicate that LSTM has the best model-fitting capacity among all the algorithms.
Similarly, Yadav et al. [19] propose two LSTM models using stateless or stateful models to predict the Indian stock market. The proposed model is tuned by varying the number of hidden layers. The results show that a stateless LSTM model is more suitable for time series prediction, and the network with fewer hidden layers has better prediction accuracy. Thakkar and Chaudhari [20] designed a cross-reference to an exchange-based stock trend prediction approach using LSTM to predict the stock price and movement of Wipro Limited (WIPRO) company. The usability of the proposed approach is demonstrated by the experiment with two other limited companies. Ammer and Aldhyani [21] present an LSTM algorithm that can forecast the values of four types of cryptocurrencies. The results demonstrate that the LSTM model performs better predicting all forms of cryptocurrencies than existing systems. One of the most critical issues in the field of market prediction is feature extraction from financial data, for which several solutions have been presented. In recent years, CNN has been used for automated feature selection and market forecasting. Hoseinzade and Haratizadeh [22] propose a CNN-based framework, which can collect data from various sources to extract features for predicting the future of those markets. The experiment results indicate that the proposed approach significantly improved prediction accuracy compared to the existing baseline algorithms. Alhazbi et al. [23] propose a CNN model by considering external factors such as oil prices to predict the daily movement of the Qatar Stock Exchange. The results indicate that adding external factors to the stock market data can increase the model performance. Wu et al. [24] present a graph-based CNN-LSTM model to predict the stock price with leading indicators. The experiment results show that the proposed algorithm leads to better results when compared with previous methods. Liu et al. [25] present a four-stage Central European Gas Hub model for intraday stock market forecasting. The results indicate that the proposed model could improve the forecasting performance compared with various baseline methods.
Some other researchers utilize fuzzy-based approaches for stock prediction, such as in [26, 27]. The results indicate that fuzzy-based systems can ensure interpretability and significantly improve stock profitability over traditional artificial intelligence models.
The previous existing studies overviewed in this section are summarized in Table 1. It can be seen from the table that most of the existing studies utilize deep learning methods, which may lead to the risk of overfitting due to the complexity of stock data. Besides, most existing literature uses standard performance metrics such as MSE and MAPE. Unlike these existing studies, this paper utilizes PCA to reduce the dimension of stock data. Furthermore, more performance metrics and a TOPSIS-based approach are used to evaluate the prediction model’s robustness.
3. Methodology
The overall methodology of this paper is relatively straightforward. Figure 1 depicts the methodology at a high level and the flow between modules. This paper explores the hybrid model’s significance for stock price prediction. This work starts with collecting stock market data used as the dataset. The dataset is then passed through the data preprocessing module, including PCA and data normalization using the max–min normalization, and the training and testing dataset is constructed. The training data serve as the prediction models’ input, including the BP neural network and exponential smoothing. The BP neural network is set up by adjusting parameters such as the number of neurons and hidden layers. Then, the neural network is trained with different training algorithms. Multiple tests are performed for exponential smoothing to select the optimal damping coefficient. In the next step, six performance metrics are calculated: MSE, APE, root mean square error (RMSE), MAPE, Accuracy, and Accuracy5. To obtain the best prediction model and analyze the robustness of these models, we use TOPSIS, which ranks all the models under consideration. Finally, another stock’s data are used to verify the optimal prediction model’s usability.

4. Relevant Theory
4.1. PCA
PCA is a common approach used for data dimension reduction [28]. A linear transformation transforms the data into a new coordinate system. The first variance of any data projection is in the first coordinate (called the first principal component), the second variance is in the second coordinate (the second principal component), and so on. PCA is often used to reduce the dimension of a dataset while preserving the characteristics that contribute most to the variance of the dataset. This can be done by keeping the lower-order principal component and ignoring the higher-order principal component. Such lower-order components tend to retain the essential aspects of the data. The input is a dataset with samples and features, that is, sample data Besides, the reduction to the target dimension is k. Thus, the sample data can be represented by the following matrix:
The output is the sample data after dimensionality reduction, that is, .
The steps of PCA are described as follows:(1)Decentralize the matrix to get a new matrix , to perform zero mean normalization on each matrix column. The new matric is represented as follows:(2)Calculate the covariance matrix of the decentralized matrix . The covariance matrix is obtained by using Equation (3):(3)Perform feature composition of the covariance matrix to find the eigenvalue and the related feature vector , that is .(4)Arrange the feature vectors in descending order according to the corresponding eigenvalues, and the first columns are taken to form the matrix .(5)Calculate the sample data after dimension reduction as described in Equation (4):
The linear transformation can transform the sample data into new synthetic variable , which can be represented as the following matrix:
The coefficient is a constant vector, which must meet the following requirements:(1)(2)(3)
are called the principal component. The amount of information extracted from each principal component is measured by Equation (6):
The sum of contribution rates of the first principal components is called cumulative contribution rate, which is calculated by Equation (7):
The larger the variance contribution rate is, the stronger the ability of the corresponding principal component to reflect comprehensive information is. The principal component is generally determined if the cumulative variance contribution rate reaches 85%.
4.2. BP Neural Network
MLP network has played a significant role in developing ANNs, and it is considered an accurate model of ANNs. Its appearance has triggered an upsurge in the study of ANNs. As the original neural network, a single-layer perceptual network (M-P model) has the advantages of a transparent model, a simple structure, and a small amount of computation. However, with the deepening of the research work, people found that it still has some shortcomings, such as being unable to deal with nonlinear problems, even if the function of the computing unit does not use the valve function but other complex nonlinear functions, still can only solve the linear separable problems, cannot achieve some essential functions, thus limiting its application. The only way to enhance the classification and recognition ability of the network and solve the nonlinear problem is to adopt the multilayer feedforward network. That is, the hidden layer is added between the input layer and the output layer to form the multilayer feedforward perceptron network. In the mid-1980s, error BP training [29] was discovered to solve the connection weight learning problem of the hidden layer of a multilayer neural network and gives a complete derivation mathematically. The multilayer feedforward network which uses this algorithm for error correction is called the BP network.
As shown in Figure 2, the structure of a BP neural network generally contains three feedforward network layers: the input layer, the intermediate layer (also known as the hidden layer), and the output layer. The characteristics of the BP neural network are that each layer of neurons is only fully connected with neurons in the adjacent layer, and there is no connection between neurons in the same layer. Besides, there is no feedback connection between neurons in each layer, forming a feedforward neural network system with a hierarchical structure. BP neural network can arbitrarily complex pattern classification and excellent multidimensional function mapping. It can solve the exclusive OR and other problems that simple perceptron cannot solve. In essence, the BP algorithm takes the square of the network error as the objective function, using the gradient descent method to calculate the minimum value of the objective function.

Three training algorithms, including LM (Levenberg–Marquardt), BR, and scaled conjugate gradient (SCG), are used in this study to train the stock market data. The LM algorithm is an iterative technique used primarily in the least squares curve fitting problem. It expresses the minimum multifunction as the sum of squares of real-value nonlinear functions. The BR algorithm can modify the mean sum of square network error to improve the network generalization ability. This algorithm is suitable for overcoming the problem of overfitting. The conjugate gradient algorithm does not require parameters but does not apply to all datasets. As a result, SCG is used since it is effective within its scope, and there is no need to set parameters. SCG can use the step size rather than the line search method in error estimation and minimize the error function.
4.3. Exponential Smoothing
Exponential smoothing is a standard method in production forecasting [30]. It is also used to forecast the middle or short-term economic development trend. Exponential smoothing is the most widely used among all the forecasting methods. Exponential smoothing is a weighted average model that uses the current state’s actual value and predicted value to give different weights calculations as the predicted value of the next state. The purpose of exponential smoothing is to eliminate the irregular changes in the time series to get the general trend that reflects the changes in the time series.
The raw data sequence is presented by starting at time , and the output of the exponential smoothing is commonly written as , which can be regarded as the best estimate of the next value of will be. When the sequence of observations begins at time , the simplest form of exponential smoothing is given by Equation (8):where is the smoothing factor, and , in the market forecast, the method of determining α is generally to make a rough estimate based on experience, and the essential judgment criteria are as follows:(1)When the time series is relatively stable, a small α value of 0.05–0.2 is selected.(2)When the time series fluctuates, the long-term trend does not change significantly. A slightly larger α value (0.1–0.4) can be selected.(3)When the time series fluctuates wildly, and the long-term trend changes have a significant upward or downward trend, a more considerable α value of 0.60–0.80 should be selected.(4)When the time series is ascending or descending, the additive model is satisfied, and α takes a more significant value, 0.6–1.
This calculation process is repeated to compare the standard error of prediction under different α values and then select the optimal α value with a minor error to establish the model.
5. Experiment Results and Discussion
5.1. Performance Metrics
To perform a comprehensive judgment on the prediction ability of the prediction model, that is, the prediction accuracy, for a group of real value and predicted value , the following performance metrics are used in this study:
MSE: MSE is the average squared difference between the estimated and actual values. MSE is calculated using Equation (9):
MSE is sensitive to outliers and varies significantly with different stock prices. Therefore, MSE cannot effectively measure the effectiveness and accuracy of the model if the data vary. As a result, RMSE is used to solve this problem by calculating the square root of MSE. The RMSE is calculated by Equation (10):
APE (absolute percentage error): APE is the ratio of the absolute value of the difference between the actual value and the predicted value to the actual value. APE is calculated using Equation (11):
MAPE: MAPE is the average relative error APE of the n observation days. MAPE is calculated by Equation (12):
Accuracy5: Accuracy5 is the proportion of samples with the APE within 5% of the total number of samples. Accuracy5 can be calculated by Equation (13):where is the number of samples whose APE is within 5%, and count (total) is the total number of samples.
Accuracy: Accuracy is a comprehensive accuracy evaluation from different aspects, combining the MAPE and Accuracy5 to construct the accuracy evaluation standard, which can reflect the prediction accuracy more comprehensively. The formula for Accuracy is described in Equation (14):
TOPSIS: TOPSIS is part of the analytical multicriteria decision-making technique. The basic idea of this approach is to find a feasible scheme that is the closest to the ideal solution and the further to the negative ideal solution. TOPSIS finds the optimal and worst targets among multiple targets through the original data matrix’s normalization. The clustering of an evaluation target, an ideal solution, and a negative ideal solution is calculated to obtain the degree of closeness between each target and the ideal solution. The degree of closeness between each target and the ideal solution is sorted in descending according to the degree of closeness of the ideal solution, which is used as the basis for evaluating the quality of the target. The closeness value ranges from 0 to 1, and the closer the value is to 1, the closer the corresponding evaluation target is to the optimal level. On the contrary, the closer the value is to 0, the closer the evaluation target is to the worst level.
5.2. Experiment Setup and Dataset
As shown in Table 2, all experiments in this paper are performed in a system with an Intel(R) Core (TM) i5-8250U @ 1.60 GHz processor, 12 GB memory, and a Windows 10 64 bit operating system. The implementation and evaluations of all prediction models are conducted using SAS, MATLAB, and SPSS.
The data used for the experiment are China Merchants Bank (600036) stock data from 2015 to 2021, obtained from the iFinD financial data terminal. This dataset involves various technical indicators, such as opening price, highest price, lowest price, closing price, change amount, change rate, etc. The parameters of the raw stock data are listed in Table 3.
The opening price of China Merchants Bank is plotted in Figure 3. The opening price rises with volatility, and the amplitude is significant in some periods, leading to tremendous challenges in predicting the short-term stock price trend.

PCA is performed on the stock market data, and the eigenvalues of the correlation coefficient matrix are presented in Table 4.
Table 5 presents specific feature vectors for each principal component, and the results reveal that the cumulative variance contribution of the first three principal components has reached the cumulative contribution rate of 90.31%. The following principal component score expression can be obtained (where represents the normalized value of the variables ) according to the first three feature values of the correlation coefficient matrix. As a result, these principal components are used as input parameters of the BP neural network to simplify the prediction model.
Finding a model that correctly predicts the output from new input data is one of the critical objectives in machine learning. However, staying away from overfitting and model complexity is also crucial. A model with a high level of complexity could be able to capture more data variations, but it will also be more challenging to train and might be more prone to overfitting. In contrast, a model with a low level of complexity could be simpler to train but might not be able to extract all the pertinent information from the data. In order to avoid overfitting, it is crucial to find the ideal balance between model complexity and overfitting while creating machine learning models.
In order to find the optimal configuration structure for the BP neural network, a comprehensive experiment is performed by varying the number of neurons in the hidden layer, learning rates, and activation functions. Experiments are performed many times for every parameter configuration used for training, with the average results being recorded to investigate the random factor for initializing the weights of the BP neural network. Besides, bias in training is minimized using a fourfold cross-validation approach for each configuration across all tests. This allows the model to be trained on different data and prevents it from being overfitted to a particular dataset. For this experiment, we divide the original dataset into four pieces of equal size (375 instances in every subset). Each test round uses 75% of the data for training and 25% for testing using the predefined arrangement.
The following test evaluates each model to determine the best structure of the BP neural network. The configuration chosen and the related performance evaluated by RMSE are presented in Table 6. We set the maximum number of epochs to 100 for training the BP model in this test. The best network structure consists of 3 inputs, 10 neurons in the hidden layer, and 1 output. We apply the Sigmoid activation function with a learning rate 0.1 and the LM algorithm for training.
5.3. Comparison and Analysis of BP Neural Network with Different Training Algorithms
Figures 4–6 plot the difference between predicted values and actual values of the BP_LM, BP_BR, and BP_SCG, respectively. These figures indicate that the nonlinear fitting ability of the BP_LM and BP_BR models is excellent, which can reflect subtle local changes in the stock price and the overall trend. The BP_SCG model has poor prediction ability in the overall trend but a strong ability of the local changes prediction in the stock prices. On the other hand, this model is more suitable for feature classification, especially in predicting the rise and fall of stock prices.



The evaluation results are compared and summarized in Table 7. The BP_SCG model has the poorest prediction performance. The BP_LM model slightly outperforms the BP_BR model. The MSE is 1.7650, which shows the accuracy and effectiveness of the model. The Accuracy5 and Accuracy values are 94.33% and 95.07%, which means that the prediction accuracy is high while maintaining model stability.
5.4. Comparison and Analysis of PCA–BP Neural Network with Different Training Algorithms
This subsection compares and analyzes the performance of the PCA–BP neural network with different training algorithms. Figures 7–9 plot the difference between predicted values with the actual values of the PCA–BP_LM, PCA–BP_BR, and PCA–BP_SCG, respectively. These figures show that the nonlinear fitting ability of the PCA–BP_LM and PCA–BP_BR models is excellent, and the PCA–BP_SCG model has the poorest prediction ability.



Table 8 compares and analyzes these three models in various performance metrics. The results indicate that the performance of these three models can be sorted as follows: PCA–BP_BR > PCA–BP_LM > PCA–BP_SCG. The PCA–BP_LM model has the best performance among the three models. The PCA–BP_BR model improves the prediction accuracy of MSE and MAPE by 75.8% and 34%, respectively, compared to the BP_LM model.
5.5. Comparison and Analysis of Exponential Smoothing with Different Smoothing Factors
For the exponential smoothing model, selecting the damping coefficient is essential. The damping coefficient reflects the response speed of the model to time series changes and determines the ability to smooth random errors in prediction. In this experiment, we vary the smoothing factor with 0.3, 0.6, and 0.9, respectively. The evaluation results of the exponential smoothing model with three different smoothing factors are represented in Table 9.
The exponential smoothing model obtains the best performance when the smoothing factor is set to 0.3. The MAPE is 0.0154, indicating that the relative error is low on the dataset. The Accuracy5 and Accuracy are 98.67% and 98.63%, respectively, demonstrating that the prediction model performs well in both prediction ability and stability. The summary of all prediction models is represented in Table 10.
We can rank these models in terms of MSE and Accuracy. The ranking ordered by MSE is arranged as follows:
The ranking order by Accuracy is arranged as follows:
The PCA–BP_BR is the top among these models in both MSE and Accuracy, as the BR algorithm can effectively solve the problem of data overfitting. The two BP neural network models with the SCG algorithm have the worst prediction, which is unsuitable for stock price prediction. The exponential smoothing model achieves better prediction performance with a smaller damping coefficient. Table 11 compares the proposed model with some existing studies overviewed in Section 2, and it is observed that the proposed model outperforms existing systems under different performance indexes.
5.6. Model Usability Evaluation
The PCA–BP_BR is selected as the optimal model with the best prediction performance according to the experiment results. In this subsection, another stock, “Wanxiang Denong (600371),” is selected to validate the usability of the optimal model. The process is repeated, including data preprocessing, training, and performance evaluation. The evaluation results are summarized in Table 12.
We can rank these models in terms of MSE and Accuracy. The ranking ordered by MSE is arranged as follows:
The ranking order by Accuracy is arranged as follows:
In this experiment, similar results are obtained; the PCA–BP_BR model has the best prediction performance while the BP_SCG has the worst. These results prove the usability and efficiency of the proposed model.
5.7. TOPSIS Evaluation
This subsection evaluates different prediction models using TOPSIS, and the results are given in Table 13. It can be seen from the table that the PCA–BP_BR is the most robust model, followed by exponential smoothing with a smoothing factor of 0.3. Besides, the exponential smoothing model is more robust than BP_SCG and PCA–BP_SCG. Finally, the advantages and disadvantages of each model tested in this study are discussed in Table 14.
6. Conclusion and Future Research Directions
This paper proposes a hybrid PCA–BP neural network model to predict stock prices in the Chinese stock market. A comprehensive experiment is performed to compare and analyze the model performance by using different training algorithms. TOPSIS has been executed to validate the robustness of all prediction models. An exponential smoothing model is also tested and compared with the proposed model. The following conclusions can be obtained from the experiment results:(1)The hybrid model performs better than a single model, improves prediction accuracy and operation efficiency, and reduces prediction error.(2)The PCA–BP model with the BR training algorithm has the best prediction accuracy.(3)The selection of the damping coefficient is tested in many rounds. The results indicate that the exponential smoothing approach has good prediction performance in time series prediction, exceeding some neural network models.
The novelty of this study is compared with some previously published papers in the same subject area. However, this study has some limitations. While the PCA–BP model, as observed, has provided a quite good prediction ability, it would be interesting to find out which of these algorithms is more accurate for stock price prediction. In future work, the results obtained in this study will be applied in a production environment with a more extensive dataset. Besides, deep learning algorithms instead of BP neural networks will be used for stock price prediction.
Data Availability
All data or codes used to support the findings of this study are available from the corresponding author.
Conflicts of Interest
The authors declare that they have no conflicts of interest.
Authors’ Contributions
LH designed the research and wrote the original manuscript, DL managed the data and conducted the empirical analysis, and FX provided guidance and revised the manuscript.
Acknowledgments
This research was supported by the Shanghai Chenguang Plan (under grant number 21CGB08) and the Yang Fan Project of the Shanghai Science and Technology Committee (under grant number 23YF1431200).