Abstract

Comprehensive document encoding and salient information selection are two major difficulties for generating summaries with adequate salient information. To tackle the above difficulties, we propose a Transformer-based encoder-decoder framework with two novel extensions for abstractive document summarization. Specifically, (1) to encode the documents comprehensively, we design a focus-attention mechanism and incorporate it into the encoder. This mechanism models a Gaussian focal bias on attention scores to enhance the perception of local context, which contributes to producing salient and informative summaries. (2) To distinguish salient information precisely, we design an independent saliency-selection network which manages the information flow from encoder to decoder. This network effectively reduces the influences of secondary information on the generated summaries. Experimental results on the popular CNN/Daily Mail benchmark demonstrate that our model outperforms other state-of-the-art baselines on the ROUGE metrics.

Highlights

  • Document summarization is a fundamental task of natural language generation which condenses the given documents and generates fluent summaries with salient information automatically

  • We propose the Extended Transformer model for Abstractive Document Summarization (ETADS) to tackle the above issues

  • We design the focusattention mechanism to improve the capability of capturing the local context information and further encode the document comprehensively

Read more

Summary

Introduction

Document summarization is a fundamental task of natural language generation which condenses the given documents and generates fluent summaries with salient information automatically. Documents: a [duke student] has [admitted to hanging a noose made of rope] from a tree near a student union , [university officials] said thursday . The [student was identified during an investigation] by campus police and the office of student affairs and admitted to placing the noose on the tree early wednesday , the university said . The noose , made of rope , was discovered on campus about 2 a.m. ing improvements have been achieved recently (Li et al, 2018c; Kryscinski et al, 2018), there are still many problems are not studied well, such as the incompetence of salient information modeling. The most essential prerequisite for a practical document summarization model is that the generated summaries should contain adequate salient information of the original documents. Previous seq2seq models are still incapable of achieving convincing performance, which are restricted by the following two difficulties

Results
Discussion
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call