Automating SPARQL Query Translations between DBpedia and Wikidata

Research output: Contributions to collected editions/worksArticle in conference proceedingsResearch

Standard

Automating SPARQL Query Translations between DBpedia and Wikidata. / Bartels, Malte Christian; Banerjee, Debayan; Usbeck, Ricardo.
SEMANTiCS Conference 2025. 2025.

Research output: Contributions to collected editions/worksArticle in conference proceedingsResearch

Harvard

APA

Vancouver

Bibtex

@inbook{dd61b91aa9d446f78a1a41a959479bd2,
title = "Automating SPARQL Query Translations between DBpedia and Wikidata",
abstract = " This paper investigates whether state-of-the-art Large Language Models (LLMs) can automatically translate SPARQL between popular Knowledge Graph (KG) schemas. We focus on translations between the DBpedia and Wikidata KG, and later on DBLP and OpenAlex KG. This study addresses a notable gap in KG interoperability research by rigorously evaluating LLM performance on SPARQL-to-SPARQL translation. Two benchmarks are assembled, where the first align 100 DBpedia-Wikidata queries from QALD-9-Plus; the second contains 100 DBLP queries aligned to OpenAlex, testing generalizability beyond encyclopaedic KGs. Three open LLMs: Llama-3-8B, DeepSeek-R1-Distill-Llama-70B, and Mistral-Large-Instruct-2407 are selected based on their sizes and architectures and tested with zero-shot, few-shot, and two chain-of-thought variants. Outputs were compared with gold answers, and resulting errors were categorized. We find that the performance varies markedly across models and prompting strategies, and that translations for Wikidata to DBpedia work far better than translations for DBpedia to Wikidata. ",
keywords = "cs.AI, cs.CL",
author = "Bartels, {Malte Christian} and Debayan Banerjee and Ricardo Usbeck",
note = "18 pages, 2 figues. Paper accepted at SEMANTiCS 2025 conference happening on September 2025",
year = "2025",
month = jul,
day = "14",
language = "English",
booktitle = "SEMANTiCS Conference 2025",

}

RIS

TY - CHAP

T1 - Automating SPARQL Query Translations between DBpedia and Wikidata

AU - Bartels, Malte Christian

AU - Banerjee, Debayan

AU - Usbeck, Ricardo

N1 - 18 pages, 2 figues. Paper accepted at SEMANTiCS 2025 conference happening on September 2025

PY - 2025/7/14

Y1 - 2025/7/14

N2 - This paper investigates whether state-of-the-art Large Language Models (LLMs) can automatically translate SPARQL between popular Knowledge Graph (KG) schemas. We focus on translations between the DBpedia and Wikidata KG, and later on DBLP and OpenAlex KG. This study addresses a notable gap in KG interoperability research by rigorously evaluating LLM performance on SPARQL-to-SPARQL translation. Two benchmarks are assembled, where the first align 100 DBpedia-Wikidata queries from QALD-9-Plus; the second contains 100 DBLP queries aligned to OpenAlex, testing generalizability beyond encyclopaedic KGs. Three open LLMs: Llama-3-8B, DeepSeek-R1-Distill-Llama-70B, and Mistral-Large-Instruct-2407 are selected based on their sizes and architectures and tested with zero-shot, few-shot, and two chain-of-thought variants. Outputs were compared with gold answers, and resulting errors were categorized. We find that the performance varies markedly across models and prompting strategies, and that translations for Wikidata to DBpedia work far better than translations for DBpedia to Wikidata.

AB - This paper investigates whether state-of-the-art Large Language Models (LLMs) can automatically translate SPARQL between popular Knowledge Graph (KG) schemas. We focus on translations between the DBpedia and Wikidata KG, and later on DBLP and OpenAlex KG. This study addresses a notable gap in KG interoperability research by rigorously evaluating LLM performance on SPARQL-to-SPARQL translation. Two benchmarks are assembled, where the first align 100 DBpedia-Wikidata queries from QALD-9-Plus; the second contains 100 DBLP queries aligned to OpenAlex, testing generalizability beyond encyclopaedic KGs. Three open LLMs: Llama-3-8B, DeepSeek-R1-Distill-Llama-70B, and Mistral-Large-Instruct-2407 are selected based on their sizes and architectures and tested with zero-shot, few-shot, and two chain-of-thought variants. Outputs were compared with gold answers, and resulting errors were categorized. We find that the performance varies markedly across models and prompting strategies, and that translations for Wikidata to DBpedia work far better than translations for DBpedia to Wikidata.

KW - cs.AI

KW - cs.CL

M3 - Article in conference proceedings

BT - SEMANTiCS Conference 2025

ER -

Documents

  • 2507.10045v1

    802 KB, PDF document