Abstract

Neural auto-regressive sequence-to-sequence models have been dominant in text generation tasks, especially the question generation task. However, neural generation models suffer from the global and local semantic semantic drift problems. Hence, we propose the hierarchical encoding–decoding mechanism that aims at encoding rich structure information of the input passages and reducing the variance in the decoding phase. In the encoder, we hierarchically encode the input passages according to its structure at four granularity-levels: [word, chunk, sentence, document]-level. Second, we progressively select the context vector from the document-level representations to the word-level representations at each decoding time step. At each time-step in the decoding phase, we progressively select the context vector from the document-level representations to word-level. We also propose the context switch mechanism that enables the decoder to use the context vector from the last step when generating the current word at each time-step.It provides a means of improving the stability of the text generation process during the decoding phase when generating a set of consecutive words. Additionally, we inject syntactic parsing knowledge to enrich the word representations. Experimental results show that our proposed model substantially improves the performance and outperforms previous baselines according to both automatic and human evaluation. Besides, we implement a deep and comprehensive analysis of generated questions based on their types.

Highlights

  • Question generation (QG) aims to generate appropriate questions for the given passages, it is an important task in natural language processing (NLP) research, QG has many applications for various NLP tasks

  • Evaluating question generation with such discrete metrics is inappropriate since there is only one reference question for each generated question resulting from the common practice of using a question answering (QA) dataset for the QG task

  • We use the following prevalent evaluation metrics to automatically assess the performances of question generation models:

Read more

Summary

Introduction

Question generation (QG) aims to generate appropriate questions for the given passages, it is an important task in natural language processing (NLP) research, QG has many applications for various NLP tasks. It is difficult to accurately parse a sentence and obtain its constituents To overcome such shortcomings, vector-based machine learning models have been introduced into QG tasks with the advent of the neural sequence-to-sequence (seq2seq) framework. Vector-based machine learning models have been introduced into QG tasks with the advent of the neural sequence-to-sequence (seq2seq) framework This adds the advantages of modeling semantics of natural language in vector space and producing more fluent and human-like text [7]. After the deployment of neural networks in the QG tasks, various models were proposed and the quality of generated questions has been significantly improved, especially in terms of readability [8]

Methods
Results
Discussion
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call