ArXiv TLDR

Ontology Generation using Large Language Models

🐦 Tweet
2503.05388

Anna Sofia Lippolis, Mohammad Javad Saeedizade, Robin Keskisärkkä, Sara Zuppiroli, Miguel Ceriani + 3 more

cs.AI

TLDR

This paper explores two new prompting techniques for LLMs to generate OWL ontology drafts from user requirements, showing improvements over state-of-the-art.

Key contributions

  • Introduces two new LLM prompting techniques, Memoryless CQbyCQ and Ontogenia, for automated ontology development.
  • Proposes three structural criteria for multi-dimensional ontology assessment, alongside expert qualitative evaluation.
  • Demonstrates that OpenAI o1-preview with Ontogenia significantly outperforms novice engineers.
  • Shows LLM-generated ontologies meet requirements, improving over current state-of-the-art.

Why it matters

Ontology engineering is complex and error-prone. This work significantly advances the use of LLMs to automate this process, offering a path to more efficient and higher-quality ontology development. It provides practical methods and evaluation criteria to guide future research and application.

Original Abstract

The ontology engineering process is complex, time-consuming, and error-prone, even for experienced ontology engineers. In this work, we investigate the potential of Large Language Models (LLMs) to provide effective OWL ontology drafts directly from ontological requirements described using user stories and competency questions. Our main contribution is the presentation and evaluation of two new prompting techniques for automated ontology development: Memoryless CQbyCQ and Ontogenia. We also emphasize the importance of three structural criteria for ontology assessment, alongside expert qualitative evaluation, highlighting the need for a multi-dimensional evaluation in order to capture the quality and usability of the generated ontologies. Our experiments, conducted on a benchmark dataset of ten ontologies with 100 distinct CQs and 29 different user stories, compare the performance of three LLMs using the two prompting techniques. The results demonstrate improvements over the current state-of-the-art in LLM-supported ontology engineering. More specifically, the model OpenAI o1-preview with Ontogenia produces ontologies of sufficient quality to meet the requirements of ontology engineers, significantly outperforming novice ontology engineers in modelling ability. However, we still note some common mistakes and variability of result quality, which is important to take into account when using LLMs for ontology authoring support. We discuss these limitations and propose directions for future research.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.