
concepts as perceived, which aligns with the idea of
cognitive representation, necessary when we want the
ontology to reflect not only formal structures, but also
how humans conceptualize the domain. In this work,
DOLCE was adopted as the reference top-level ontol-
ogy, anchoring the classes of the domain ontology in
its taxonomy, with the expectation of creating a struc-
tured and interoperable ontology. Note that using a
top-level ontology does not eliminate the need for ad-
justments; contrariwise, it requires careful analysis of
where each domain class fits into the upper hierarchy,
an exercise that also serves as an additional concep-
tual validation.
2.2 LLMs and Ontology Learning
The convergence of ontologies and LLMs has moti-
vated a new wave of research in ontology learning,
or extraction, from text (Lopes et al., 2024). The
task of Ontology Learning (OL) consists in start-
ing from unstructured information and deriving a
structured set of ontological axioms, encompassing
identifying relevant terms, discovering hierarchical
and non-hierarchical relationships between them, and
eventually proposing complex constraints or axioms
(Babaei Giglou et al., 2023). Traditionally, this task
was divided into subtasks handled by specialized Nat-
ural Language Processing (NLP) and machine learn-
ing techniques, such as term extraction, synonym dis-
covery, and hypernym discovery. With the advent of
LLMs, which can understand natural language and
generate coherent text, the possibility has emerged to
treat ontology learning as a language generation prob-
lem, requiring that the model translates raw textual
knowledge into an ontology expressed, for example,
in the OWL language (Schaeffer et al., 2024).
Potential advantages of using LLMs include the
ability to identify implicit concepts and relation-
ships in text without manual work. LLMs have
demonstrated the ability to extract knowledge triples
(subject-predicate-object) from texts, forming basic
knowledge graphs. Recent studies have applied
LLMs to generate complete ontologies: for exam-
ple, (Bakker and Scala, 2024) used GPT-4 to extract
an ontology from a news article, obtaining relevant
classes, individuals, and properties. This and other
researches have shown that LLMs successfully cap-
ture many of the main concepts present in the text and
can propose preliminary hierarchies, indicating an ad-
vance over previous methods that often produced only
flat lists of terms. Furthermore, LLMs offer inter-
action flexibility: it is possible to use sophisticated
prompts, decomposing the task into steps, for exam-
ple, first extracting classes, then relationships, to im-
prove the quality (Bakker and Scala, 2024). Prompt
engineering techniques, such as Chain-of-Thought,
or using CQs are being explored to guide LLMs in
gradual construction processes that potentially im-
prove the consistency of the ontology (Saeedizade and
Blomqvist, 2024).
However, several challenges and limitations of
LLMs for this purpose have been identified. A crit-
ical point is the lack of consistent ontological reason-
ing: LLMs tend to base their responses on statistical
language patterns, without guaranteeing adherence to
the required logical or ontological rules. For example,
(Mai et al., 2025) demonstrated that when confronted
with entirely new terms, pre-trained LLMs failed to
correctly infer semantic relationships, merely repro-
ducing known linguistic structures. This suggests that
language models outside their training domain may
not truly understand the concepts, unless they are fine-
tuned with data from that domain.
Another practical observation is that LLMs may
neglect certain parts of the ontology, particularly re-
lationships. (Bakker and Scala, 2024) noted that, al-
though GPT-4 identified important classes from a text,
it often failed to include properties between classes or
introduced inconsistent properties between instances.
In their evaluations, the raw output of the LLM con-
tained some logical errors and omissions, requiring
manual supplementation. Generally, hallucinations –
inferences not supported by the text – are also a risk:
when generating ontologies, the LLM sometimes in-
vented relationships not mentioned.
Therefore, the literature indicates that LLMs are
useful as assistants, but do not replace human curation
of the learned ontology (Saeedizade and Blomqvist,
2024). They can accelerate knowledge acquisition,
serving as a first draft of the ontology or an exten-
sion of an existing one based on new information.
However, the intervention of an ontology engineer
is necessary to verify and correct errors, add miss-
ing relationships, and ensure axiomatic consistency.
A recommendation is to integrate LLMs into a hy-
brid workflow, whereby the model automates candi-
date proposal steps, and the human performs valida-
tion and fine-tuning. Additionally, adapting LLMs to
the domain via fine-tuning or few-shot learning can
significantly improve quality: (Babaei Giglou et al.,
2023) showed that adapted models achieve signifi-
cantly better performance in tasks such as term typi-
fication, taxonomy discovery, and relationship extrac-
tion, being useful as assistants to alleviate the knowl-
edge acquisition bottleneck in ontology construction.
Advanced prompting techniques are also important
for optimizing the relevance and feasibility of using
LLMs (Schaeffer et al., 2024).
Semi-Automatic Domain Ontology Construction: LLMs, Modularization, and Cognitive Representation
67