Abstract

Text generation is a rapidly evolving field of Natural Language Processing (NLP) with larger Language models proposed very often setting new state-of-the-art. These models are exorbitantly effective in learning the representation of words and their internal coherence in a particular language. However, an established context-driven, end to end text generation model is very rare, even more so for the Bengali language. In this paper, we have proposed a Bidirectional gated recurrent unit (GRU) based architecture that simulates the conditional language model or the decoder portion of the sequence to sequence (seq2seq) model and is further conditioned upon the target context vectors. We have explored several ways of combining multiple context words into a fixed dimensional vector representation that is extracted from the same GloVe language model which is used to generate the embedding matrix. We have used beam search optimization to generate the sentence with the maximum cumulative log probability score. In addition, we have proposed a human scoring based evaluation metric and used it to compare the performance of the model with unidirectional LSTM and GRU networks. Empirical results prove that the proposed model performs exceedingly well in producing meaningful outcomes depicting the target context. The experiment leads to an architecture that can be applied to an extensive domain of context-driven text generation based applications and which is also a key contribution to the NLP based literature of the Bengali language.

Highlights

  • Text generation is the task of producing texts automatically given some contexts or goals, that are indistinguishable from human-written texts

  • It is a subfield of Natural Language Processing (NLP) and a derivative of Computer linguistics and Artificial Intelligence (AI)

  • We have presented context-based Bengali text generation using a conditional language model

Read more

Summary

Introduction

Text generation is the task of producing texts automatically given some contexts or goals, that are indistinguishable from human-written texts. It is a subfield of Natural Language Processing (NLP) and a derivative of Computer linguistics and Artificial Intelligence (AI). Implementations of automatic text generation for structured texts such as codes or URLs have been widely available for a long time. Generation processes for such structured data are relatively straightforward and can be addressed through conventional programming approaches eg.- a starting tag in a markup language will have an ending tag of the same name.

Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call