Abstract
Topic models have been widely utilized in Topic Detection and Tracking tasks, which aim to detect, track, and describe topics from a stream of broadcast news reports. However, most existing topic models neglect semantic or syntactic information and lack readable topic descriptions. To exploit semantic and syntactic information, Language Models (LMs) have been applied in many supervised NLP tasks. However, there are still no extensions of LMs for unsupervised topic clustering. Moreover, it is difficult to employ general LMs (e.g., BERT) to produce readable topic summaries due to the mismatch between the pretraining method and the summarization task. In this paper, noticing the similarity between content and summary, first we propose a Language Model-based Topic Model (LMTM) for Topic Clustering by using an LM to generate a deep contextualized word representation. Then, a new method of training a Topic Summarization Model is introduced, where it is not only able to produce brief topic summaries but also used as an LM in LMTM for topic clustering. Empirical evaluations of two different datasets show that the proposed LMTM method achieves better performance over four baselines for JC, FMI, precision, recall and F1-score. Additionally, the generated readable and reasonable summaries also validate the rationality of our model components.
Highlights
With the rapid development of Internet technology, online media has become an important way for the public to publish and obtain information
We propose the use of Language Models (LMs), which have been applied in many supervised NLP tasks but not to unsupervised topic clustering, to overcome the limitation of traditional topic models, i.e., that they do not make full use of semantic and syntactic information, and to explain the details of Language Model-based Topic Model (LMTM) by using a novel language model, e.g., BERT
Considering that general LMs are too large to be flexible and efficient, and the pretraining of them is useless for topic clustering or summarization, we propose a text summarization model whose encoder can be used as a light and flexible LM
Summary
With the rapid development of Internet technology, online media has become an important way for the public to publish and obtain information. Since the number of news reports is very large, and different news articles have different values, as well as gain different amounts of attention, the articles reporting important and hot events may be overshadowed by those with less value. For news related to an ongoing event, it is difficult for readers to link the current news with previous articles, which leads to them having difficulty analyzing the. The associate editor coordinating the review of this manuscript and approving it for publication was Jerry Chun-Wei Lin. development of the events. Automatically extracting hot events from massive news reports and linking them with related topics is an urgent issue [1], [2]
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.