Abstract

In the literature, the research on abstract meaning representation (AMR) parsing is much restricted by the size of human-curated dataset which is critical to build an AMR parser with good performance. To alleviate such data size restriction, pre-trained models have been drawing more and more attention in AMR parsing. However, previous pre-trained models, like BERT, are implemented for general purpose which may not work as expected for the specific task of AMR parsing. In this paper, we focus on sequence-to-sequence (seq2seq) AMR parsing and propose a seq2seq pre-training approach to build pre-trained models in both single and joint way on three relevant tasks, i.e., machine translation, syntactic parsing, and AMR parsing itself. Moreover, we extend the vanilla fine-tuning method to a multi-task learning fine-tuning method that optimizes for the performance of AMR parsing while endeavors to preserve the response of pre-trained models. Extensive experimental results on two English benchmark datasets show that both the single and joint pre-trained models significantly improve the performance (e.g., from 71.5 to 80.2 on AMR 2.0), which reaches the state of the art. The result is very encouraging since we achieve this with seq2seq models rather than complex models. We make our code and model available at https://github.com/xdqkid/S2S-AMR-Parser.

Highlights

  • Abstract meaning representation (AMR) parsing aims to translate a textual sentence into a directed and acyclic graph which consists of concept nodes and edges representing the semantic relations between the nodes (Banarescu et al, 2013)

  • Pre-trained models on a single task significantly improve the performance of abstract meaning representation (AMR) parsing, indicating seq2seq pre-training is helpful for seq2seqbased AMR parsing

  • We note that the pre-trained model of neural machine translation (NMT) achieves the best performance, followed by the pre-trained models on AMR parsing and on syntactic parsing

Read more

Summary

Introduction

Abstract meaning representation (AMR) parsing aims to translate a textual sentence into a directed and acyclic graph which consists of concept nodes and edges representing the semantic relations between the nodes (Banarescu et al, 2013). Et al, 2018), graph-based approaches (Flanigan et al, 2014; Werling et al, 2015; Cai and Lam, 2019), transition-based approaches (Damonte et al, 2017; Guo and Lu, 2018), sequence-to-sequence (seq2seq) approaches (Peng et al, 2017; van Noord and Bos, 2017; Konstas et al, 2017; Ge et al, 2019), and sequence-to-graph (seq2graph) approaches (Zhang et al, 2019a,b; Cai and Lam, 2020) Among these approaches, seq2seq-based approaches, which properly transform AMR graphs into sequences, have received much interest, due to the simplicity in implementation and the competitive performance. Linguistic knowledge captured in pre-trained models can be properly

Methods
Results
Discussion
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call