Abstract

The technology of automatic text generation by machine has always been an important task in natural language processing, but the low-quality text generated by the machine seriously affects the user experience due to poor readability and fuzzy effective information. The machine-generated text detection method based on traditional machine learning relies on a large number of artificial features with detection rules. The general method of text classification based on deep learning tends to the orientation of text topics, but logical information between texts sequences is not well utilized. For this problem, we propose an end-to-end model which uses the text sequences self-information to compensate for the information loss in the modeling process, to learn the logical information between the text sequences for machine-generated text detection. This is a text classification task. We experiment on a Chinese question and answer the dataset collected from a biomedical social media, which includes human-written text and machine-generated text. The result shows that our method is effective and exceeds most baseline models.

Highlights

  • Have you ever encountered such a situation: when you ask a question on social media, the answers seem to be substantial and the topic is consistent with the question, but after reading carefully, you may find that the confusing logic makes it hard to read with the worthless contents

  • E technology of automatic text generation can be applied to a question and answer system [8], machine translation [9], and automatic text summarization [10]. e development of this technology will realize more intelligent and natural human-computer interaction

  • We look forward to the day when computers can write like humans, but generating high-quality text sequences is still a challenge

Read more

Summary

Weikuan Wang and Ao Feng

Received 6 November 2020; Revised 17 December 2020; Accepted 8 February 2021; Published 19 February 2021. E general method of text classification based on deep learning tends to the orientation of text topics, but logical information between texts sequences is not well utilized. E technology of automatic text generation by machine has always been an important task in natural language processing, but the low-quality text generated by the machine seriously affects the user experience due to poor readability and fuzzy effective information. For this problem, we propose an end-to-end model which uses the text sequences self-information to compensate for the information loss in the modeling process, to learn the logical information between the text sequences for machine-generated text detection. We experiment on a Chinese question and answer the dataset collected from a biomedical social media, which includes human-written text and machine-generated text. e result shows that our method is effective and exceeds most baseline models

Introduction
Related Work
Lookup table
Experiments
Random initialization parameters Pretrained parameters
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call