Abstract

Named entity recognition (NER) is an essential part of natural language processing tasks. Chinese NER task is different from the many European languages due to the lack of natural delimiters. Therefore, Chinese Word Segmentation (CWS) is usually regarded as the first step of processing Chinese NER. However, the word-based NER models relying on CWS are more vulnerable to incorrectly segmented entity boundaries and the presence of out-of-vocabulary (OOV) words. In this paper, we propose a novel character-based Gated Convolutional Recurrent neural network with Attention called GCRA for Chinese NER task. In particular, we introduce a hybrid convolutional neural network with gating filter mechanism to capture local context information and a highway neural network after LSTM to select characters of interest. The additional gated self-attention mechanism is used to capture the global dependencies from different multiple subspaces and arbitrary adjacent characters. We evaluate the performance of our proposed model on three datasets, including SIGHAN bakeoff 2006 MSRA, Chinese Resume, and Literature NER dataset. The experiment results show that our model outperforms other state-of-the-art models without relying on any external resources like lexicons and multi-task joint training.

Highlights

  • Named entity recognition (NER) plays a critical role in the field of natural language processing (NLP)

  • To overcome the shortcomings of the traditional characterbased models, we propose a new neural network, called the GCRA, to improve the performance of Chinese NER task

  • Experiments on several series of datasets show that our proposed GCRA model can significantly improve the performance of the Chinese NER task

Read more

Summary

INTRODUCTION

Named entity recognition (NER) plays a critical role in the field of natural language processing (NLP). After performing word segmentation, we may get completely different recognition boundary information, which leads to two distinct sequence labeling outcomes. Due to the polysemy and polymorphism of Chinese characters, the NER based on the pure character only focuses on the per-character information for losing the latent word and word sequence information For this problem, it is worth exploring how to effectively integrate segmentation information into character-based models for better semantic understanding. For the embedding layer, we apply the label segmentation vector softly concatenating into character embeddings We design a character-level hybrid gated convolutional neural network which combines the dilated gated convolution with the standard gated convolution It can effectively generate local feature information connection and avoid gradient vanishing during training.

RELATED WORK
HYBRID GATED CONVOLUTION LAYER
GATED SELF-ATTENTION LAYER
CRF LAYER
EXPERIMENTS
Findings
CONCLUSION
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call