loading
Papers Papers/2022 Papers Papers/2022

Research.Publish.Connect.

Paper

Authors: Matthias Engelbach 1 ; Dennis Klau 2 ; Felix Scheerer 2 ; Jens Drawehn 1 and Maximilien Kintz 1

Affiliations: 1 Fraunhofer Institute for Industrial Engineering IAO, Nobelstr. 12, 70569 Stuttgart, Germany ; 2 University of Stuttgart, Institute of Human Factors and Technology Management IAT, Allmandring 35, Stuttgart, Germany

Keyword(s): Question-Answering, Language Models, Information Extraction.

Abstract: The emergence of Large Language Models (LLMs) has boosted performance and possibilities in various NLP tasks. While the usage of generative AI models like ChatGPT opens up new opportunities for several business use cases, their current tendency to hallucinate fake content strongly limits their applicability to document analysis, such as information retrieval from documents. In contrast, extractive language models like question answering (QA) or passage retrieval models guarantee query results to be found within the boundaries of an according context document, which makes them candidates for more reliable information extraction in productive environments of companies. In this work we propose an approach that uses and integrates extractive QA models for improved feature extraction of German business documents such as insurance reports or medical leaflets into a document analysis solution. We further show that fine-tuning existing German QA models boosts performance for tailored extract ion tasks of complex linguistic features like damage cause explanations or descriptions of medication appearance, even with using only a small set of annotated data. Finally, we discuss the relevance of scoring metrics for evaluating information extraction tasks and deduce a combined metric from Levenshtein distance, F1-Score, Exact Match and ROUGE-L to mimic the assessment criteria from human experts. (More)

CC BY-NC-ND 4.0

Sign In Guest: Register as new SciTePress user now for free.

Sign In SciTePress user: please login.

PDF ImageMy Papers

You are not signed in, therefore limits apply to your IP address 3.149.251.154

In the current month:
Recent papers: 100 available of 100 total
2+ years older papers: 200 available of 200 total

Paper citation in several formats:
Engelbach, M.; Klau, D.; Scheerer, F.; Drawehn, J. and Kintz, M. (2023). Fine-Tuning and Aligning Question Answering Models for Complex Information Extraction Tasks. In Proceedings of the 15th International Joint Conference on Knowledge Discovery, Knowledge Engineering and Knowledge Management - KDIR; ISBN 978-989-758-671-2; ISSN 2184-3228, SciTePress, pages 196-205. DOI: 10.5220/0012159000003598

@conference{kdir23,
author={Matthias Engelbach. and Dennis Klau. and Felix Scheerer. and Jens Drawehn. and Maximilien Kintz.},
title={Fine-Tuning and Aligning Question Answering Models for Complex Information Extraction Tasks},
booktitle={Proceedings of the 15th International Joint Conference on Knowledge Discovery, Knowledge Engineering and Knowledge Management - KDIR},
year={2023},
pages={196-205},
publisher={SciTePress},
organization={INSTICC},
doi={10.5220/0012159000003598},
isbn={978-989-758-671-2},
issn={2184-3228},
}

TY - CONF

JO - Proceedings of the 15th International Joint Conference on Knowledge Discovery, Knowledge Engineering and Knowledge Management - KDIR
TI - Fine-Tuning and Aligning Question Answering Models for Complex Information Extraction Tasks
SN - 978-989-758-671-2
IS - 2184-3228
AU - Engelbach, M.
AU - Klau, D.
AU - Scheerer, F.
AU - Drawehn, J.
AU - Kintz, M.
PY - 2023
SP - 196
EP - 205
DO - 10.5220/0012159000003598
PB - SciTePress