loading
Papers Papers/2022 Papers Papers/2022

Research.Publish.Connect.

Paper

Authors: Claire Ponciano ; Markus Schaffert and Jean-Jacques Ponciano

Affiliation: i3mainz, University of Applied Sciences, Germany

Keyword(s): Ontology-Grounded Language Modeling, GPT, Knowledge-Enhanced Text Generation, Retrieval-Augmented Generation, Spinoza, Linked Open Data, Historical Text Synthesis, Philosophical Language Modeling, BERTScore Evaluation, Structured Knowledge Integration, Latin Text Generation, Large Language Models, Text Style Transfer, Semantic Conditioning, Canonical Corpus Fine-Tuning.

Abstract: We present an ontology-grounded approach to GPT-based text generation aimed at improving factual grounding, historical plausibility, and stylistic fidelity in a case study: Baruch Spinoza’s Latin writings. We construct a compact ontology from Linked Open Data (Wikidata/DBpedia) augmented with expert-curated facts, serialize triples into natural-language statements, and interleave these with a canonical Latin corpus during fine-tuning of a GPT-2 (124M) model. At inference, retrieval-augmented generation (RAG) prepends ontology-derived facts and lightweight stylistic instructions, guiding the model toward historically consistent continuations in Spinoza’s register. Evaluation follows an 80/20 paragraph split of Ethica: we generate continuations for the 80% of segments retained and measure the semantic similarity (BERTScore) with the 20% omitted. This evaluation is completed by an expert assessment of historical plausibility and cosine similarity scores computation for the stylistic aut henticity. Relative to a GPT-2 baseline trained only on the Latin corpus, our ontology-grounded variant achieves higher BERTScore and produces fewer factual and conceptual errors, preserving Latin rhetorical structure. These results indicate that structured knowledge integration is a feasible and effective way to make generative models more reliable for cultural-heritage text. (More)

CC BY-NC-ND 4.0

Sign In Guest: Register as new SciTePress user now for free.

Sign In SciTePress user: please login.

PDF ImageMy Papers

You are not signed in, therefore limits apply to your IP address 216.73.216.88

In the current month:
Recent papers: 100 available of 100 total
2+ years older papers: 200 available of 200 total

Paper citation in several formats:
Ponciano, C., Schaffert, M. and Ponciano, J.-J. (2025). Ontology-Grounded Language Modeling: Enhancing GPT-Based Philosophical Text Generation with Structured Knowledge. In Proceedings of the 21st International Conference on Web Information Systems and Technologies - WEBIST; ISBN 978-989-758-772-6; ISSN 2184-3252, SciTePress, pages 459-467. DOI: 10.5220/0013864400003985

@conference{webist25,
author={Claire Ponciano and Markus Schaffert and Jean{-}Jacques Ponciano},
title={Ontology-Grounded Language Modeling: Enhancing GPT-Based Philosophical Text Generation with Structured Knowledge},
booktitle={Proceedings of the 21st International Conference on Web Information Systems and Technologies - WEBIST},
year={2025},
pages={459-467},
publisher={SciTePress},
organization={INSTICC},
doi={10.5220/0013864400003985},
isbn={978-989-758-772-6},
issn={2184-3252},
}

TY - CONF

JO - Proceedings of the 21st International Conference on Web Information Systems and Technologies - WEBIST
TI - Ontology-Grounded Language Modeling: Enhancing GPT-Based Philosophical Text Generation with Structured Knowledge
SN - 978-989-758-772-6
IS - 2184-3252
AU - Ponciano, C.
AU - Schaffert, M.
AU - Ponciano, J.
PY - 2025
SP - 459
EP - 467
DO - 10.5220/0013864400003985
PB - SciTePress