
tem enables users to design, configure and understand
ML pipelines through a conversational interface. Ex-
perimental results demonstrate that this approach can
achieve high model performance on both classifica-
tion and regression tasks while enhancing user inter-
action, interpretability and accessibility.
The system contributes to the democratization of
ML by allowing domain experts and non-specialists
to meaningfully participate in the development of pre-
dictive models. It operationalizes key principles of
HCAI by combining automation with user control,
transparency and contextual support. Furthermore,
its modular and open-source architecture provides a
strong foundation for future enhancements.
Looking ahead, several directions for future work
are identified. First, expanding the system’s multi-
lingual capabilities and fine-tuning LLMs on domain-
specific corpora may improve accuracy in interpret-
ing complex or specialized queries. Second, inte-
grating additional AutoML frameworks beyond Lud-
wig could broaden compatibility and adoption. Third,
introducing support for advanced data manipulation
(e.g., time series decomposition, anomaly detection
or unsupervised learning) would extend the system’s
versatility.
ACKNOWLEDGEMENTS
This work has been partially supported by grant
PID2023-146243OB-I00 funded by MICIU/AEI/
10.13039/501100011033 and by “ERDF/EU”. This
research was also partially developed in the project
FUTCAN - 2023 / TCN / 018 that was co-financed
from the European Regional Development Fund
through the FEDER Operational Program 2021-2027
of Cantabria through the line of grants “Aid for re-
search projects with high industrial potential of excel-
lent technological agents for industrial competitive-
ness TCNIC”.
REFERENCES
Azure Automated ML. Azure Automated ML.
https://azure.microsoft.com/en-us/solutions/
automated-machine-learning. Access: 2 January
2025.
Ben-Kiki, O., Evans, C., and Net, I. d. (2009). Yaml ain’t
markup language (yaml™) version 1.2. https://yaml.
org/spec/1.2/spec.html.
Chang, Y., Wang, X., Wang, J., Wu, Y., Yang, L., Zhu, K.,
Chen, H., Yi, X., Wang, C., Wang, Y., et al. (2024). A
survey on evaluation of large language models. ACM
Transactions on Intelligent Systems and Technology,
15(3):1–45.
Choi, H., Moran, J., Matsumoto, N., Hernandez, M. E., and
Moore, J. H. (2023). Aliro: an automated machine
learning tool leveraging large language models. Bioin-
formatics, 39(10):btad606.
Das, P., Ivkin, N., Bansal, T., Rouesnel, L., Gautier, P.,
Karnin, Z., Dirac, L., Ramakrishnan, L., Peruni-
cic, A., Shcherbatyi, I., Wu, W., Zolic, A., Shen,
H., Ahmed, A., Winkelmolen, F., Miladinovic, M.,
Archembeau, C., Tang, A., Dutt, B., Grao, P., and
Venkateswar, K. (2020). Amazon sagemaker autopi-
lot: a white box automl solution at scale. In Pro-
ceedings of the Fourth International Workshop on
Data Management for End-to-End Machine Learn-
ing, DEEM ’20, New York, NY, USA. Association for
Computing Machinery.
DeepMind, G. (2024). Gemma: Open models by google
deepmind. https://deepmind.google/technologies/
gemma.
Duque, R., T
ˆ
ırn ˇauc ˇa, C., Palazuelos, C., Casas, A., L
´
opez,
A., and P
´
erez, A. (2025). Bridging automl and llms:
Towards a framework for accessible and adaptive ma-
chine learning. In Filipe, J., Smialek, M., Brodsky,
A., and Hammoudi, S., editors, Proceedings of the
27th International Conference on Enterprise Informa-
tion Systems, ICEIS 2025, Porto, Portugal, April 4-6,
2025, Volume 1, pages 959–964. SCITEPRESS.
Fan, L., Li, L., Ma, Z., Lee, S., Yu, H., and Hemphill, L.
(2024). A bibliometric review of large language mod-
els research from 2017 to 2023. ACM Transactions on
Intelligent Systems and Technology, 15(5):1–25.
Gil, Y., Honaker, J., Gupta, S., Ma, Y., D’Orazio, V., Gar-
ijo, D., Gadewar, S., Yang, Q., and Jahanshad, N.
(2019). Towards human-guided machine learning. In
Proceedings of the 24th international conference on
intelligent user interfaces, pages 614–624.
Google AutoML. Google AutoML. https://cloud.google.
com/automl. Access: 2 January 2025.
Karmaker, S. K., Hassan, M. M., Smith, M. J., Xu, L.,
Zhai, C., and Veeramachaneni, K. (2021). Automl to
date and beyond: Challenges and opportunities. ACM
Computing Surveys (CSUR), 54(8):1–36.
LeDell, E. and Poirier, S. (2020). H2O AutoML: Scalable
automatic machine learning. 7th ICML Workshop on
Automated Machine Learning (AutoML).
Liu, S.-C., Wang, S., Lin, W., Hsiung, C.-W., Hsieh, Y.-
C., Cheng, Y.-P., Luo, S.-H., Chang, T., and Zhang,
J. (2023). Jarvix: A llm no code platform for tab-
ular data analysis and optimization. arXiv preprint
arXiv:2312.02213.
Luo, D., Feng, C., Nong, Y., and Shen, Y. (2024). Autom3l:
An automated multimodal machine learning frame-
work with large language models. In Proceedings of
the 32nd ACM International Conference on Multime-
dia, pages 8586–8594.
MindsDB (2018). MindsDB. https://mindsdb.com/. Ac-
cess: 6 December 2024.
Molino, P., Dudin, Y., and Miryala, S. S. (2019). Ludwig: a
type-based declarative deep learning toolbox.
Integrating Large Language Models into Automated Machine Learning: A Human-Centric Approach
471