Abstract

One of the most important factors which considerably affects the quality of the neural sequence labeling model is the selection and encoding of input features to generate rich semantic and grammatical word representation vectors. Recently, pre-training text encoders have significantly ameliorated the performance on various Natural Language Processing (NLP) tasks. In this paper, we introduce a deep neural network model to address a challenging task of the sequence labeling problem, the task of named entity recognition. The model consists of three sub-networks to fully exploit current Google’s representations of the most powerful NLP transfer learning BERT and ELMO model, as well as the traditional state-of-the-art BiLSTM-CRF (BLC) model. The model outperforms the state-of-the-art systems on CoNLL-2003 by achieving a 95.56% F1-score and getting results comparable to the previous best-performing system models on OntoNotes 5.0 by achieving an 89.21% F1-score without using any external knowledge.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call