Conditional Heteroskedasticity (GARCH) model, as
a classical tool for modeling market volatility, can
effectively characterize these time-varying risk
features (Engle and Patton, 2001).
In addition, Value at Risk (VaR) is an important
tool for risk management and provides a standardized
risk measurement framework, but it needs to be
combined with other methods to fully assess risks
(Duffie and Pan, 1997). Moreover, Alexander and
Baptista illustrated that the minimum VaR portfolio
only exists when the confidence level is high enough,
while a low confidence level will lead to irrational
decisions and market imbalance (Alexander and
Baptista, 2002). In extreme risk scenarios, flexible
GARCH models combined with leptokurtic
distributions can significantly improve the accuracy
of VaR predictions. Therefore, by integrating return
forecasting with GARCH volatility modelling, the
VaR calculation can reflect a stronger sensitivity to
risk and enhanced practical applicability.
Current research has been limited to either
comparing the predictive performance of ARIMA
models with machine learning models or focusing
solely on VaR calculation and forecasting using
GARCH models. Although these studies provide
valuable insights into individual model performance,
there is a clear lack of research that integrates all three
modeling approaches into a cohesive framework for
application in real investment portfolios. This
disjointed approach leaves a gap in understanding
how these models can complement each other in
practical, risk-sensitive forecasting environments.
Based on this background, the present study aims
to bridge this gap by conducting a comprehensive
comparison of ARIMA models and typical machine
learning methods in forecasting portfolio returns. The
research will then incorporate GARCH models for
volatility modeling, leveraging the strengths of each
approach to achieve an accurate estimation of
portfolio VaR. In this integrated framework,
comparing the ARIMA model's capacity to capture
linear trends and seasonality with the machine
learning methods' ability to uncover complex
nonlinear patterns, can provide a more robust and
nuanced forecast of returns. Meanwhile, the GARCH
model's proven track record in modelling volatility
clustering in financial time series will be instrumental
in refining the risk measurement process.
This study can provide a deeper understanding of
the adaptability and limitations of these different
models in real-world forecasting scenarios. By
systematically integrating these models, the research
aims to offer both theoretical and quantitative support
for asset allocation and risk control. The findings are
expected to contribute to the development of more
sophisticated risk management tools that can be
applied by financial institutions and portfolio
managers, enhancing decision-making processes in
environments characterized by uncertainty and
market volatility.
2 METHODOLOGY
In this part, the data resources used in this study,
variables involved and specific methods will be
introduced.
2.1 Data Source
The data utilized in this study is sourced from the
Yahoo Finance platform, an international financial
information platform. The database is accessed and
extracted through Python programming language.
The daily stock data set extracted from the Yahoo
Finance platform contains important indicators such
as opening price, closing price, and returns, which can
clearly reflect the daily changes of the stock. To
verify the performance of the ARIMA model and
LSTM model in portfolio returns prediction and risk
measurement, this study selected the historical stock
price data of Apple Inc. (stock code: AAPL) as the
research object, with a time span of January 1, 2015
to February 28, 2025, covering significant market
stages such as Covid-19 and the Russian-Ukrainian
War. As a world-renowned technology company,
Apple's stock has high market liquidity and
representativeness, which can better reflect the
dynamic characteristics of market risk and return.
2.2 Variables Selection and Description
In this study, the daily returns extracted from the
stock data were selected as the basic indicator,
comprising approximately 2,553 observations.
Subsequent data processing on this variable yielded
three additional variables. Ultimately, these variables
were utilized to compute the VaR using two different
prediction models and GARCH model, leading to the
corresponding risk estimates.
2.3 Method Interpretation
In this study, Augmented Dickey-Fuller (ADF) test
was first conducted on the extracted daily returns of
Apple Inc. to examine the presence of unit roots
within the time series. Upon confirming the
stationarity of the daily returns, an optimal ARIMA