Abstract

The abstractive automatic summarization task is to summarize the main content of the article with short sentences, which is an important research direction in natural language generation. Most abstractive summarization models are based on sequence-to-sequence neural networks. Specifically, they encode input text sequences by Bi-directional Long Short-Term Memory (bi-LSTM), and decode summaries word-by-word by LSTM. However, existing models usually did not consider both the self-attention dependence during the encoding process using bi-LSTM, and deep potential sentence structure information for the decoding process. To tackle these limitations, we propose a Self-Attention based word embedding and Hierarchical Variational AutoEncoders (SA-HVAE) model. The model first introduces self-attention into LSTM to alleviate information decay of encoding, and accomplish summarization with deep structure information inference through hierarchical VAEs. The experimental results on the Gigaword and CNN/Daily Mail datasets validate the superior performance of SA-HVAE, and our model has a significant improvement over the baseline model.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.