Abstract

Language model is one of the basic research issues of natural language processing, and which is the premise for realizing more complicated tasks such as speech recognition, machine translation and question answering system. In recent years, neural network language model has become a research hotspot, which greatly enhances the application effect of language model. In this paper, a recurrent neural network language model (RNNLM) based on word embedding is proposed, and the word embedding of each word is generated by pre-training the text data with skip-gram model. The n-gram language model, RNNLM based on one-hot and RNNLM based on word embedding are evaluated on three different public datasets. The experimental results show that the RNNLM based on word embedding performs best, and which can reduce the perplexity of language model significantly.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.