BERT-Based Transformer Fine-Tuning for Dutch Wikidata Question-Answering

Niels de Jong, Gosse Bouma

Research output: Contribution to journalArticleAcademicpeer-review

101 Downloads (Pure)

Abstract

People rely on data to understand the world and inform their decision-making. However, effective access to data has become more challenging over time: data has increased in volume and velocity, as has its variability in truthfulness, utility, and format. Therefore, improving our interfaces to data has become a pressing issue. One type of interface has lately gained renewed attention, driven by advances in artificial intelligence (AI): natural language (NL) interfaces. As of yet, though, improvements in natural language processing (NLP) have largely concentrated on English. Thus, we propose a text-based Dutch question-answering (QA) interface for accessing information on WikiData, driven by a Dutch-to-SPARQL BERT-based transformer model. Said transformer is a type of encoder-decoder model characterised by use of self-attention. In our application, it is trained to accept sentences in Dutch and to transform these into corresponding SPARQL queries. By subsequently evaluating the obtained queries at a knowledge base, users can retrieve answers to their questions. Since our model learns end-to-end, we need to train it using a dataset consisting of pairs of Dutch questions and SPARQL queries. To this end, we closely follow the procedure of Cui et al. (2021). Particularly, we create a Dutch machine-translated version of LC-QuAD 2.0 (Dubey et al., 2019) and apply entity and relation masking on the NL inputs and SPARQL outputs for increased generality, producing a dataset with 2,648 examples. We then let the transformer model fine-tune on the training subset of this dataset, using system-level BLEU score as the performance measure. Our final transformer configuration obtains a test BLEU score of 51.86, which appears to outperform Cui et al.’s Kannada model (45.6 BLEU) but not their models for Hebrew and Mandarin Chinese (66.0 and 58.6 BLEU, respectively). Additionally, we conduct a qualitative analysis of our model’s outputs, focusing especially on situations where the predicted SPARQL queries are incorrect. Here, we observe that queries involving infrequently-used SPARQL keywords and queries containing literals prove challenging to the transformer, as sometimes do the syntax of SPARQL and the general length of queries. Finally, we conclude our paper by proposing some potential future directions for our Dutch QA system.
Original languageEnglish
Pages (from-to)83-97
Number of pages15
JournalComputational Linguistics in the Netherlands Journal
Volume12
Publication statusPublished - 31-Dec-2022
Event32nd Meeting of Computational Linguistics in the Netherlands (CLIN 32) - Willem II stadium, Tilburg, Netherlands
Duration: 17-Jun-202217-Jun-2022
Conference number: 32
https://clin2022.uvt.nl/

Keywords

  • question answering
  • Sparql
  • DUTCH

Fingerprint

Dive into the research topics of 'BERT-Based Transformer Fine-Tuning for Dutch Wikidata Question-Answering'. Together they form a unique fingerprint.

Cite this