Evaluating Large Language Models for RDF Knowledge Graph Related Tasks - The LLM-KG-Bench-Framework 3

Tracking #: 3869-5083

This paper is currently under review
Authors: 
lars-peter meyer
Johannes Frey1
Felix Brei
Desiree Heim
Sabine Gründer-Fahrer
Sara Todorovikj
Claus Stadler
Markus Schröder
Natanael Arndt1
Michael Martin1

Responsible editor: 
Guest Editors 2025 LLM GenAI KGs

Submission type: 
Full Paper
Abstract: 
Current Large Language Models (LLMs) can work with structured information and even assist developing program code, but can they support working with Knowledge Graphs (KGs) as well? Which LLM is offering the best capabilities in the field of Semantic Web and Knowledge Graph Engineering (KGE)? Is it possible to determine this without checking many answers manually? The LLM-KG-Bench framework is designed to answer these questions. It consists of an extensible set of tasks for which the LLM answers are automatically evaluated, and covers different aspects of working with semantic technologies. This article gives a description of the LLM-KG-Bench framework, it's main concepts and the tasks implemented. In a benchmark run, a comprehensive dataset has been generated with it, evaluating more than 40 contemporary open and proprietary LLMs. Finally, this dataset is used for an analysis of the SPARQL related capabilities of the LLMs tested.
Full PDF Version: 
Tags: 
Under Review