Figure 10: Transfer Learning with Fine-Tuning, with the
same setup as Figure 9.
Figure 11: Transfer Learning with Fine-Tuning - Training
behavior.
try. The proposed RL-based agents and methods
outperform some of the state-of-the-art methods in
public benchmarks, demonstrating their usefulness.
Moreover, the framework provides transfer learning
adaptation and an experimentable solution to transfer
knowledge from one market to another. The work has
shown that fine-tuning to the new datasets is neces-
sary, especially when trading rules change.
Although financial applications can be formulated
in multiple ways, the usability of our methods can
be further developed to adapt to multiple scenarios
in the future. Therefore, we plan to report in future
work on the difficulties of developing a solution that
can accommodate all trading or portfolio manage-
ment strategies. Regarding the technical aspects, we
plan to extend our methods by carefully investigating
and implementing other RL methods, such as actor-
critical mechanisms. We are also considering extend-
ing the framework to several types of trading in fi-
nance (e.g., futures, options, CFDs, buy-sell). As sug-
gested in another study in (Bakker, 2001), LSTM net-
works can also enhance the capabilities of RL agents
as an alternative to current knowledge embedding.
REFERENCES
Bakker, B. (2001). Reinforcement learning with long short-
term memory. In Dietterich, T., Becker, S., and
Ghahramani, Z., editors, Advances in Neural Informa-
tion Processing Systems, volume 14. MIT Press.
Beck, N., Rajasekharan, A., and Tran, H. (2022). Transfer
reinforcement learning for differing action spaces via
q-network representations.
Brockman, G., Cheung, V., Pettersson, L., Schneider, J.,
Schulman, J., Tang, J., and Zaremba, W. (2016). Ope-
nai gym.
Czarnecki, W. M., Pascanu, R., Osindero, S., Jayaku-
mar, S., Swirszcz, G., and Jaderberg, M. (2019).
Distilling policy distillation. In Chaudhuri, K. and
Sugiyama, M., editors, Proceedings of the Twenty-
Second International Conference on Artificial Intelli-
gence and Statistics, volume 89 of Proceedings of Ma-
chine Learning Research, pages 1331–1340. PMLR.
Dongrey, S. (2022). Study of market indicators used for
technical analysis. International Journal of Engineer-
ing and Management Research, 12:64–83.
dos Santos Mignon, A. and de Azevedo da Rocha, R. L.
(2017). An adaptive implementation of e-greedy in
reinforcement learning. Procedia Computer Science,
109:1146–1151. 8th International Conference on Am-
bient Systems, Networks and Technologies, ANT-
2017 and the 7th International Conference on Sustain-
able Energy Information Technology, SEIT 2017, 16-
19 May 2017, Madeira, Portugal.
Folk, M., Heber, G., Koziol, Q., Pourmal, E., and Robinson,
D. (2011). An overview of the hdf5 technology suite
and its applications. pages 36–47.
Hochreiter, S. and Schmidhuber, J. (1997). Long short-term
memory. Neural computation, 9(8):1735–1780.
Kipf, T. N. and Welling, M. (2017). Semi-supervised clas-
sification with graph convolutional networks.
Kong, M. and So, J. (2023). Empirical analysis of auto-
mated stock trading using deep reinforcement learn-
ing. Applied Sciences, 13(1).
Koshiyama, A., Blumberg, S., Firoozye, N., Treleaven, P.,
and Flennerhag, S. (2021). Quantnet: transferring
learning across trading strategies. Quantitative Fi-
nance, 22:1–20.
Li, Z., Liu, X.-Y., Zheng, J., Wang, Z., Walid, A., and Guo,
J. (2021). FinRL-podracer. In Proceedings of the Sec-
ond ACM International Conference on AI in Finance.
ACM.
Liang, E., Liaw, R., Nishihara, R., Moritz, P., Fox, R.,
Goldberg, K., Gonzalez, J., Jordan, M., and Stoica,
I. (2018). RLlib: Abstractions for distributed rein-
forcement learning. In Dy, J. and Krause, A., editors,
Proceedings of the 35th International Conference on
Machine Learning, volume 80 of Proceedings of Ma-
chine Learning Research, pages 3053–3062. PMLR.
Liashchynskyi, P. and Liashchynskyi, P. (2019). Grid
search, random search, genetic algorithm: A big com-
parison for nas.
Liu, X.-Y., Xia, Z., Rui, J., Gao, J., Yang, H., Zhu, M.,
Wang, C. D., Wang, Z., and Guo, J. (2022a). Finrl-
Deep Reinforcement Learning and Transfer Learning Methods Used in Autonomous Financial Trading Agents
25