Abstract

Named entity recognition (NER) is one fundamental task in the natural language processing (NLP) community. Supervised neural network models based on contextualized word representations can achieve highly-competitive performance, which requires a large-scale manually-annotated corpus for training. While for the resource-scarce languages, the construction of such as corpus is always expensive and time-consuming. Thus, unsupervised cross-lingual transfer is one good solution to address the problem. In this work, we investigate the unsupervised cross-lingual NER with model transfer based on contextualized word representations, which greatly advances the cross-lingual NER performance. We study several model transfer settings of the unsupervised cross-lingual NER, including (1) different types of the pretrained transformer-based language models as input, (2) the exploration strategies of the multilingual contextualized word representations, and (3) multi-source adaption. In particular, we propose an adapter-based word representation method combining with parameter generation network (PGN) better to capture the relationship between the source and target languages. We conduct experiments on a benchmark ConLL dataset involving four languages to simulate the cross-lingual setting. Results show that we can obtain highly-competitive performance by cross-lingual model transfer. In particular, our proposed adapter-based PGN model can lead to significant improvements for cross-lingual NER.

Highlights

  • Named entity recognition (NER) aims to extract named entities and identify their semantic types from text, which is one of the fundamental tasks in natural language processing (NLP) [1]

  • Our basic model is built according to the system, and our work focuses on the unsupervised cross-lingual setting, studying different exploration methods for the BERTalike word representations

  • We investigated the unsupervised cross-lingual adaption for NER based on the model transfer framework

Read more

Summary

Introduction

Named entity recognition (NER) aims to extract named entities and identify their semantic types (e.g., person, organization and location) from text, which is one of the fundamental tasks in natural language processing (NLP) [1]. The task can be beneficial for a range of applications, including relation extraction [2], coreference resolution [3] and question answering [4], as the extracted named entities are critical elements for these applications.

Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call