REFERENCES
Ahmadalinezhad, M. and Makrehchi, M. (2020). Basket-
ball lineup performance prediction using edge-centric
multi-view network analysis. Social Network Analysis
and Mining, 10.
Banoth, T. and Hashmi, M. F. (2022). Yolov3-sort: detec-
tion and tracking player/ball in soccer sport. Journal
of Electronic Imaging, 32.
Cao, Z., Hidalgo, G., Simon, T., Wei, S.-E., and Sheikh,
Y. (2019). Openpose: Realtime multi-person 2d pose
estimation using part affinity fields.
Cheng, Y., Fan, Q., Pankanti, S. , and Choudhary, A. (2014).
Temporal sequence modeling for video event detec-
tion. In 2014 IEEE Conference on Computer Vision
and Pattern Recognition, pages 2235–2242.
Ding, G., Sener, F., and Yao, A. (2023). Temporal action
segmentation: An analysis of modern techniques.
Fanuel, M., Yuan, X., N am Kim, H., Qingge, L., and Roy,
K. (2021). A survey on skeleton-based activity recog-
nition using graph convolutional networks (gcn). In
12th International Symposium on Image and Signal
Processing and Analysis (ISPA), pages 177–182.
Furnari, A., Batt iato, S., and Farinella, G. M. (2018).
Personal-location-based temporal segmentation of
egocentric videos for lifelogging applications. Journal
of Visual Communication and Image Representation,
52:1–12.
Garnier, P. and Gregoir, T. (2021). Evaluating soccer player:
from live camera to deep reinforcement learning.
Giancola, S., Cioppa, A., Deli`ege, A., Magera, F., Somers,
V., Kang, L., Zhou, X., Barnich, O., Vleeschouwer,
C., Alahi, A., Ghanem, B., Droogenbroeck, M., Dar-
wish, A., Maglo, A. , Clap´es, A., Luyts, A., Boiarov,
A., Xarles, A. , Orcesi, A., and Li, Z. (2022). Soccer-
net 2022 challenges results. pages 75–86.
Giulietti, N., Caputo, A., Chiariotti, P., and Castellini, P.
(2023). Swimmernet: Underwater 2d swimmer pose
estimation exploiting fully convolutional neural net-
works. Sensors, 23(4).
Hauri, S., Djuric, N., Radosavljevic, V., and Vucetic, S.
(2021). Multi-modal t r aj ectory prediction of nba play-
ers. pages 1639–1648.
Huang, W., He, S., Sun, Y., Evans, J., Song, X., Geng, T.,
Sun, G., and Fu, X. ( 2022). Open dataset recorded
by single cameras for multi-player tracking in soccer
scenarios. Applied Sciences, 12(15).
Khobdeh, S., Yamaghani, M., and Sareshkeh, S. (2023).
Basketball action recognition based on the combina-
tion of yolo and a deep fuzzy lstm network. The Jour-
nal of Supercomputing, pages 1–26.
Kingma, D. P. and Ba, J. (2017). Adam: A method for
stochastic optimization.
Koshkina, M., Pidaparthy, H., and Elder, J. H. (2021). Con-
trastive learning for sports video: Unsupervised player
classification.
Kulkarni, K. M. and Shenoy, S. (2021). Table tennis stroke
recognition using two-dimensional human pose esti-
mation.
Liu, Y., Hafemann, L. G., Jamieson, M., and Javan, M.
(2021). Detecting and matching related objects with
one proposal multiple predictions.
Low, S. (2016). Overcoming Gravity: A Systematic Ap-
proach To Gymnastics And Bodyweight Strength (Sec-
ond Edition). Battleground Creative.
Manafifard, M., Ebadi, H., and Abrishami Moghaddam, H.
(2017). A survey on player tracking in soccer videos.
Computer Vision and Image Understanding, 159:19–
46. Computer Vision in Sports.
Martin, Z., Patel, A., and Hendricks, S. (2021). Automated
tackle injury risk assessment in contact-based sports –
a rugby union example.
McNally, W., Walters, P., Vats, K., Wong, A., and McPhee,
J. (2021). Deepdarts: Modeling keypoints as ob-
jects for automatic scorekeeping in darts using a single
camera.
Munea, T. L., Jembre, Y. Z., Weldegebriel, H. T., Chen,
L., Huang, C., and Yang, C. (2020). The progress of
human pose estimation: A survey and taxonomy of
models applied in 2d human pose estimation. IEEE
Access, 8:133330–133348.
Murthy, P., Taetz, B., Lekhra, A., and Stricker, D. (2023).
Divenet: Dive action localization and physical pose
parameter extraction for high performance training.
IEEE Access, 11:37749–37767.
Naik, B. T., Hashmi, M. F., and Bokde, N. D. (2022). A
comprehensive review of computer vision in sports:
Open issues, future tr ends and research directions. Ap-
plied Sciences, 12(9).
Pidaparthy, H., Dowli ng, M. H., and Elder, J. H. (2021).
Automatic play segmentation of hockey videos. In
2021 IEEE/CVF Conference on Computer Vision
and Pattern Recognition Workshops (CVPRW), pages
4580–4588.
Rahimi, A. M., Lee, K., Agarwal, A., Kwon, H., and
Bhattacharyya, R. (2021). Toward improving the vi-
sual characterization of sport activities with abstracted
scene graphs. In 2021 IEEE/CVF Conference on
Computer Vision and Pattern Recognition Workshops
(CVPRW), pages 4495–4502.
Rahmad, N., As’ari, M. A., Ibrahim, M., S ufri, N. A. J.,
and Rangasamy, K. (2020). Vision Based Automated
Badminton Action Recognition Using the New Local
Convolutional Neural Network Extractor.
Ramanathan, V., Huang, J., Abu-El-Haija, S., Gorban, A.,
Murphy, K., and Fei-Fei, L. (2016). Detecting events
and key actors in multi-person videos.
Ren, B., Liu, M., Ding, R., and Liu, H. (2020). A survey
on 3d skeleton-based action recognition using learn-
ing method.
Richard, A. and Gall, J. (2016). Temporal action detection
using a statistical language model. In 2016 IEEE Con-
ference on Computer Vision and Pattern Recognition
(CVPR), pages 3131–3140.
Shaikh, M. B. and Chai, D. (2021). Rgb-d data-based action
recognition: A review. Sensors, 21(12).
Spagnolo, P., Mazzeo, P. L., Leo, M., N itti, M., Stella, E.,
and Distante, A. (2014). On-field testing and evalu-