Abstract

Named entity recognition (NER) is an important task in natural language processing area, which needs to determine entities boundaries and classify them into pre-defined categories. For Chinese NER task, there is only a very small amount of annotated data available. Chinese NER task and Chinese word segmentation (CWS) task have many similar word boundaries. There are also specificities in each task. However, existing methods for Chinese NER either do not exploit word boundary information from CWS or cannot filter the specific information of CWS. In this paper, we propose a novel adversarial transfer learning framework to make full use of task-shared boundaries information and prevent the task-specific features of CWS. Besides, since arbitrary character can provide important cues when predicting entity type, we exploit self-attention to explicitly capture long range dependencies between two tokens. Experimental results on two different widely used datasets show that our proposed model significantly and consistently outperforms other state-of-the-art methods.

Highlights

  • The task of named entity recognition (NER) is to recognize the named entities in given text

  • Numerous methods have been carefully studied for NER task, including Hidden Markov Models (HMMs) (Bikel et al, 1997), Support Vector Machines (SVMs) (Isozaki and Kazawa, 2002) and Conditional Random Fields (CRFs) (Lafferty et al, 2001)

  • To prevent the specific information of Chinese word segmentation (CWS) task from lowering the performance of the Chinese NER task, we introduce adversarial training to ensure that the Chinese NER task only exploits task-shared word boundary information

Read more

Summary

Introduction

The task of named entity recognition (NER) is to recognize the named entities in given text.

Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call