historical data, which may not fully capture future
market conditions, especially during unprecedented
events or structural changes in the market.
Looking ahead to the future, there are several
research directions worth further exploration in
response to these challenges. Firstly, by further
optimizing the attention mechanism, the model's
ability to capture market signals can be enhanced. In
addition, considering integrating alternative data
sources such as sentiment analysis data extracted
from news reports or social media will provide richer
background information for the model's predictions.
Finally, extending the model to more asset classes or
operating across multiple markets not only helps
improve its robustness, but also enhances its
applicability under different market conditions.
4 CONCLUSIONS
To sum up, this study evaluated the effectiveness of
CNN+LSTM+Multi-Head Attention model in stock
price prediction and trading strategy execution. The
research results indicate that the hybrid model
significantly outperforms traditional simple models
in terms of cumulative returns, especially in situations
of high market volatility. By integrating multiple
neural network architectures, this model is able to
effectively capture complex patterns in financial data,
providing traders with powerful support tools.
However, the study also exposed some limitations,
e.g., dependence on historical data and high
computational complexity of the model. Future
research should focus on further optimizing the model
structure and exploring more data sources to enhance
the predictive ability of the model. Overall, the
CNN+LSTM+Multi-Head Attention model
demonstrates broad application prospects in the field
of financial forecasting, with the potential to improve
trading strategies and decision-making processes in
complex market environments.
REFERENCES
Bollerslev, T., 1986. Generalized autoregressive
conditional heteroskedasticity. Journal of econometrics,
31(3), 307-327.
Fischer, T. Krauss, C., 2018. Deep learning with long short-
term memory networks for financial market predictions.
European journal of operational research, 270(2), 654-
669.
Graves, A., 2012. Long short-term memory. Supervised
sequence labelling with recurrent neural networks, 37-
45.
Greff, K., Srivastava, R. K., Koutník, J., Steunebrink, B. R.
Schmidhuber, J., 2016. LSTM: A search space odyssey.
IEEE transactions on neural networks and learning
systems, 28(10), 2222-2232.
Jialin, L., Shanwen, Q., Zhikai, Z., Keyao, L., Jiayong, M.
Toe, T. T., 2022, July. CNN-LSTM Model Stock
Forecasting Based on an Integrated Attention
Mechanism. 2022 3rd International Conference on
Pattern Recognition and Machine Learning (PRML) pp.
403-408.
Kim, T. Kim, H. Y., 2019. Forecasting stock prices with a
feature fusion LSTM-CNN model using different
representations of the same data. PloS one, 14(2),
e0212320.
Li, C., Zhang, X., Qaosar, M., Ahmed, S., Alam, K. M. R.
Morimoto, Y., 2019. Multi-factor based stock price
prediction using hybrid neural networks with attention
mechanism. 2019 IEEE intl conf on dependable,
autonomic and secure computing, intl conf on pervasive
intelligence and computing, intl conf on cloud and big
data computing, intl conf on cyber science and
technology congress
(DASC/PiCom/CBDCom/CyberSciTech), 961-966.
Sezer, O. B., Gudelek, M. U. Ozbayoglu, A. M., 2020.
Financial time series forecasting with deep learning: A
systematic literature review: 2005–2019. Applied soft
computing, 90, 106181.
Siami-Namini, S., Tavakoli, N. Namin, A. S., 2018,
December. A comparison of ARIMA and LSTM in
forecasting time series. 2018 17th IEEE international
conference on machine learning and applications
(ICMLA), 1394-1401.
Thakkar, A. Chaudhari, K., 2021. A comprehensive survey
on deep neural networks for stock market: The need,
challenges, and future directions. Expert Systems with
Applications, 177, 114800.
Vaswani, A., 2017. Attention is all you need. Advances in
Neural Information Processing Systems, 1
Zhang, J., Ye, L. Lai, Y., 2023. Stock price prediction using
CNN-BiLSTM-Attention model. Mathematics, 11(9),
1985.