Abstract

In the recent few years, text analyses with neural models have become more popular due its versatile usages in different software applications. In order to improve the performance of text analytics, there is a huge collection of methods that have been identified and justified by the researchers. Most of these techniques have been efficiently used for text categorization, text generation, text summarization, query formulation, query answering, sentiment analysis and etc. In this review paper, we consolidate a recent literature along with the technical survey on different neural models such as Neural Language Model (NLM), sequence to sequence model (seq2seq), text generation, Bidirectional Encoder Representations from Transformers (BERT), machine translation model (MT), transformation model, attention model from the perception of applying deep machine learning algorithms for text analysis. Applied extensive experiments were conducted on the deep learning model such as Recurrent Neural Network (RNN) / Long Short-Term Memory (LSTM) / Convolutional Neural Network (CNN) and Attentive Transformation model to examine the efficacy of different neural models with the implementation using tensor flow and keras.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call