Abstract

Extractive summarization aims to produce a concise summary while retaining the key information through the way of selecting sentences from the original document. Under such background, learning inter-sentence relations has hitherto been the issue of most concern. In this study, we propose a Star architecture based model for extractive summarization (StarSum), that takes advantage of self-attention strategy based Transformer and star-shaped structure, models sentences within a document as satellite nodes and introduces a virtual star node, constructs a star model for each document to learn inter-sentence relations. Based on the constructed star-shaped model, we further develop two sentence representation learning algorithms, namely star guiding satellite (SGS) algorithm and star incorporating satellite (SIS) algorithm, in order to extract summary-worthy sentences. Experimental results on CNN/Daily Mail, New York Times (NYT) and XSum datasets prove that StarSum model achieves advanced performance for extractive summarization and has comparable performance to the state-of-the-art extractive summarization model. The results also demonstrate that the SIS algorithm is more effective than the SGS algorithm.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call