
to longer execution times.
This study underscores the importance of prior-
itizing resource allocation, particularly for the most
resource-intensive stages, to minimize bottlenecks
and ensure efficient ML pipeline execution. However,
since many real-world frameworks exhibit greater
hardware variability, including the use of dedicated
accelerators such as GPUs, TPUs, and FPGAs, future
research should investigate whether these findings can
be generalized to such environments. Additionally,
future work should explore dynamic VM provision-
ing and adaptive resource management strategies to
enhance performance in heterogeneous and evolving
computing scenarios.
ACKNOWLEDGEMENTS
This study was funded by the PRR – Plano de
Recuperac¸
˜
ao e Resili
ˆ
encia and by the NextGen-
erationEU funds at University of Aveiro, through
the scope of the Agenda for Business Innovation
“NEXUS: Pacto de Inovac¸
˜
ao – Transic¸
˜
ao Verde e
Digital para Transportes, Log
´
ıstica e Mobilidade”
(Project nº 53 with the application C645112083-
00000059).
REFERENCES
Chen, H.-M., Chen, S.-Y., Hsueh, S.-H., and Wang, S.-
K. (2023). Designing an improved ml task schedul-
ing mechanism on kubernetes. In 2023 Sixth Interna-
tional Symposium on Computer, Consumer and Con-
trol (IS3C), pages 60–63.
Chicco, D. and Jurman, G. (2020). The advantages of the
matthews correlation coefficient (mcc) over f1 score
and accuracy in binary classification evaluation. BMC
genomics, 21(1):1–13.
Dutta, G. Network slicing recognition. https:
//www.kaggle.com/datasets/gauravduttakiit/network-
slicing-recognition.
Faubel, L., Schmid, K., and Eichelberger, H. (2023). Mlops
challenges in industry 4.0. SN Computer Science,
4(6):828.
Google (2024). Mlops: Continuous delivery and
automation pipelines in machine learning.
https://cloud.google.com/architecture/mlops-
continuous-delivery-and-automation-pipelines-
in-machine-learning.
Jayaram Subramanya, S., Arfeen, D., Lin, S., Qiao, A., Jia,
Z., and Ganger, G. R. (2023). Sia: Heterogeneity-
aware, goodput-optimized ml-cluster scheduling. In
Proceedings of the 29th Symposium on Operating Sys-
tems Principles, SOSP ’23, page 642–657, New York,
NY, USA. Association for Computing Machinery.
Kreuzberger, D., K
¨
uhl, N., and Hirschl, S. (2023). Ma-
chine learning operations (mlops): Overview, defini-
tion, and architecture. IEEE access, 11:31866–31879.
Le, T. N., Sun, X., Chowdhury, M., and Liu, Z. (2020).
Allox: compute allocation in hybrid clusters. In
Proceedings of the Fifteenth European Conference
on Computer Systems, EuroSys ’20, New York, NY,
USA. Association for Computing Machinery.
MKhubaiib. IoT attack prediction dataset. https:
//www.kaggle.com/datasets/mkhubaiib/iot-attack-
prediction-dataset.
Moreno, C. and Fischmeister, S. (2017). Accurate measure-
ment of small execution times—getting around mea-
surement errors. IEEE Embedded Systems Letters,
9(1):17–20.
Narayanan, D., Santhanam, K., Kazhamiaka, F., Phan-
ishayee, A., and Zaharia, M.(2020). Heterogeneity-
Aware cluster scheduling policies for deep learning
workloads. In 14th USENIX Symposium on Operat-
ing Systems Design and Implementation (OSDI 20),
pages 481–498. USENIX Association.
Preciado-Velasco, J. E., Gonzalez-Franco, J. D., Anias-
Calderon, C. E., Nieto-Hipolito, J. I., and Rivera-
Rodriguez, R. (2021). 5g/b5g service classifica-
tion using supervised learning. Applied Sciences,
11(11):4942.
Sandeepa, C., Senevirathna, T., Siniarski, B., Nguyen, M.-
D., La, V.-H., Wang, S., and Liyanage, M. (2023).
From opacity to clarity: Leveraging xai for robust net-
work traffic classification. In International Confer-
ence on Asia Pacific Advanced Network, pages 125–
138. Springer.
Smith, J. (2024). Devops and mlops convergence: Improv-
ing collaboration between data science and engineer-
ing teams. Australian Journal of Machine Learning
Research & Applications, 4(2):82–86.
Speedwall10. Iot device network logs. https:
//www.kaggle.com/datasets/speedwall10/iot-device-
network-logs.
Tavallaee, M., Bagheri, E., Lu, W., and Ghorbani, A. A.
(2009). A detailed analysis of the kdd cup 99 data
set. In 2009 IEEE symposium on computational intel-
ligence for security and defense applications, pages
1–6. Ieee.
Thantharate, A., Paropkari, R., Walunj, V., and Beard, C.
(2019). Deepslice: A deep learning approach towards
an efficient and reliable network slicing in 5g net-
works. In 2019 IEEE 10th Annual Ubiquitous Com-
puting, Electronics & Mobile Communication Confer-
ence (UEMCON), pages 0762–0767. IEEE.
Zhang, K. and Wu, B. (2012). Task scheduling for gpu het-
erogeneous cluster. In 2012 IEEE International Con-
ference on Cluster Computing Workshops, pages 161–
169.
Zhang, X. (2024). Mixtran: an efficient and fair scheduler
for mixed deep learning workloads in heterogeneous
gpu environments. Cluster Computing, 27(3):2775–
2784.
Impact of Resource Heterogeneity on MLOps Stages: A Computational Efficiency Study
253