Abstract

While neural sequence learning methods have made significant progress in single-document summarization (SDS), they produce unsatisfactory results on multi-document summarization (MDS). We observe two major challenges when adapting SDS advances to MDS: (1) MDS involves larger search space and yet more limited training data, setting obstacles for neural methods to learn adequate representations; (2) MDS needs to resolve higher information redundancy among the source documents, which SDS methods are less effective to handle. To close the gap, we present RL-MMR, Maximal Margin Relevance-guided Reinforcement Learning for MDS, which unifies advanced neural SDS methods and statistical measures used in classical MDS. RL-MMR casts MMR guidance on fewer promising candidates, which restrains the search space and thus leads to better representation learning. Additionally, the explicit redundancy measure in MMR helps the neural representation of the summary to better capture redundancy. Extensive experiments demonstrate that RL-MMR achieves state-of-the-art performance on benchmark MDS datasets. In particular, we show the benefits of incorporating MMR into end-to-end learning when adapting SDS to MDS in terms of both learning effectiveness and efficiency.

Highlights

  • Text summarization aims to produce condensed summaries covering salient and non-redundant information in the source documents

  • multi-document summarization (MDS) aims at producing summaries from multiple source documents, which exceeds the capacity of neural single-document summarization (SDS) models (See et al, 2017) and sets learning obstacles for adequate representations, especially considering that MDS labeled data is more limited

  • (1) We present an RL-based MDS framework that combines the advances of classical MDS and neural SDS methods via end-to-end learning

Read more

Summary

Introduction

Text summarization aims to produce condensed summaries covering salient and non-redundant information in the source documents. Recent studies on single-document summarization (SDS) benefit from the advances in neural sequence learning (Nallapati et al, 2016; See et al, 2017; Chen and Bansal, 2018; Narayan et al, 2018) as well as pretrained language models (Liu and Lapata, 2019; Lewis et al, 2019; Zhang et al, 2020) and make great progress. MDS aims at producing summaries from multiple source documents, which exceeds the capacity of neural SDS models (See et al, 2017) and sets learning obstacles for adequate representations, especially considering that MDS labeled data is more limited. SDS models adopt attention mechanisms as implicit measures to reduce redundancy (Chen and Bansal, 2018), they fail to handle the much higher redundancy of MDS effectively (Sec. 4.2.3)

Methods
Results
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.