Abstract
Recently, sequence-to-sequence models have achieved impressive performance on a number of semantic parsing tasks. However, they often do not exploit available linguistic resources, while these, when employed correctly, are likely to increase performance even further. Research in neural machine translation has shown that employing this information has a lot of potential, especially when using a multi-encoder setup. We employ a range of semantic and syntactic resources to improve performance for the task of Discourse Representation Structure Parsing. We show that (i) linguistic features can be beneficial for neural semantic parsing and (ii) the best method of adding these features is by using multiple encoders.
Original language | English |
---|---|
Title of host publication | Proceedings of the 13th International Conference onComputational Semantics - Short Papers |
Publisher | Association for Computational Linguistics (ACL) |
Pages | 24-31 |
Number of pages | 8 |
Publication status | Published - May-2019 |