Abstract

Named entity recognition (NER) is a basic but crucial task in the field of natural language processing (NLP) and big data analysis. The recognition of named entities based on Chinese is more complicated and difficult than English, which makes the task of NER in Chinese more challenging. In particular, fine-grained named entity recognition is more challenging than traditional named entity recognition tasks, mainly because fine-grained tasks have higher requirements for the ability of automatic feature extraction and information representation of deep neural models. In this paper, we propose an innovative neural network model named En2BiLSTM-CRF to improve the effect of fine-grained Chinese entity recognition tasks. This proposed model including the initial encoding layer, the enhanced encoding layer, and the decoding layer combines the advantages of pre-training model encoding, dual bidirectional long short-term memory (BiLSTM) networks, and a residual connection mechanism. Hence, it can encode information multiple times and extract contextual features hierarchically. We conducted sufficient experiments on two representative datasets using multiple important metrics and compared them with other advanced baselines. We present promising results showing that our proposed En2BiLSTM-CRF has better performance as well as better generalization ability in both fine-grained and coarse-grained Chinese entity recognition tasks.

Highlights

  • With the vigorous development of big data mining technology, named entity recognition [1,2,3,4] as the most important subtask of natural language processing (NLP) tasks is playing an essential role in many fields driven by data [5,6,7]

  • We propose an innovative neural network model, named En2BiLSTM-conditional random field (CRF), to improve the recognition effect of fine-grained Chinese named entities

  • The contributions of this paper can be summarized as follows: (1) We propose a universal fine-grained named entity recognition model, En2BiLSTM-CRF, which includes three effective modules: the initial encoding module, the enhanced encoding module, and the decoding module

Read more

Summary

Introduction

With the vigorous development of big data mining technology, named entity recognition [1,2,3,4] as the most important subtask of natural language processing (NLP) tasks is playing an essential role in many fields driven by data [5,6,7]. We propose an innovative neural network model, named En2BiLSTM-CRF, to improve the recognition effect of fine-grained Chinese named entities. This end-to-end NER model combines the advantages of the pre-trained model, the dual BiLSTM networks, the residual connection mechanism [26], and the conditional random field. The proposed model uses a structure of three module layers to complete the initial encoding and enhanced encoding as well as decoding of the input data hierarchically, ensuring that sufficient semantic information and context features are used for fine-grained entity recognition tasks. Excellent performance on three metrics proves the generalization ability and robustness of the proposed model

Fine-Grained Named Entity Recognition
Bidirectional LSTM Networks
Pre-Trained Models
Initial
Enhanced Encoding Layer
Decoding Layer
Hyperparameters of the Proposed Model
Dataset Description
Experiments and Settings
CLUENER2020
People’s Daily Ner Corpus
Evaluation Metrics
Baselines
BiLSTM-CRF-NER
BERT-NER
RoBERTa-NER
Human Performance
Experimental Settings
Results and Discussion
Overall Experimental Results and Discussion
Detailed Results and Discussion
Statistical
Conclusions
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call