Comparing Neural Meaning-to-Text Approaches for Dutch

Chunliu Wang, Johan Bos

Research output: Contribution to journalArticleAcademicpeer-review

2 Citations (Scopus)
90 Downloads (Pure)

Abstract

The neural turn in computational linguistics has made it relatively easy to build systems for natural language generation, as long as suitable annotated corpora are available. But can such systems deliver the goods? Using Dutch data of the Parallel Meaning Bank, a corpus of (mostly short) texts annotated with language-neutral meaning representations, we investigate what challenges arise and what choices can be made when implementing sequence-to-sequence or graphto- sequence transformer models for generating Dutch texts from formal meaning representations. We compare the performance of linearized input graphs with graphs encoded in various formats and find that stacking encoders obtain the best results for the standard metrics used in natural language generation. A key challenge is dealing with unknown tokens that occur in the input meaning representation. We introduce a new method based on WordNet similarity to deal with out-of-vocab concepts.
Original languageEnglish
Pages (from-to)269-286
Number of pages18
JournalComputational Linguistics in the Netherlands Journal
Volume12
Publication statusPublished - 22-Dec-2022

Fingerprint

Dive into the research topics of 'Comparing Neural Meaning-to-Text Approaches for Dutch'. Together they form a unique fingerprint.

Cite this