Abstract

The goal of Word Sense Disambiguation (WSD) is to identify the correct meaning of a word in the particular context. Traditional supervised methods only use labeled data (context), while missing rich lexical knowledge such as the gloss which defines the meaning of a word sense. Recent studies have shown that incorporating glosses into neural networks for WSD has made significant improvement. However, the previous models usually build the context representation and gloss representation separately. In this paper, we find that the learning for the context and gloss representation can benefit from each other. Gloss can help to highlight the important words in the context, thus building a better context representation. Context can also help to locate the key words in the gloss of the correct word sense. Therefore, we introduce a co-attention mechanism to generate co-dependent representations for the context and gloss. Furthermore, in order to capture both word-level and sentence-level information, we extend the attention mechanism in a hierarchical fashion. Experimental results show that our model achieves the state-of-the-art results on several standard English all-words WSD test datasets.

Highlights

  • Word Sense Disambiguation (WSD) is a crucial task and long-standing problem in Natural Language Processing (NLP)

  • Supervised feature-based methods (Zhi and Ng, 2010; Iacobacci et al, 2016) and neural-based methods (Kageback and Salomonsson, 2016; Raganato et al, 2017a) usually use labeled data to train one or more classifiers. As they often play football together, they know each other quite well g1: participate in games or sports g2: perform music on an instrument g3: behave in a certain way both lexical knowledge and labeled data are of great help for WSD, previous supervised methods rarely take the integration of knowledge into consideration

  • We conduct a series of experiments, which show that our models outperform the state-ofthe-art systems on several standard English all-words WSD test datasets

Read more

Summary

Introduction

Word Sense Disambiguation (WSD) is a crucial task and long-standing problem in Natural Language Processing (NLP). As they often play football together, they know each other quite well g1: participate in games or sports g2: perform music on an instrument g3: behave in a certain way both lexical knowledge (especially gloss) and labeled data are of great help for WSD, previous supervised methods rarely take the integration of knowledge into consideration. To the best of our knowledge, Luo et al (2018) are the first to directly incorporate the gloss knowledge from WordNet into a unified neural network for WSD. This model separately builds the context representation and the gloss representation as distributed vectors and later calculates their similarity in a memory network. We introduce a co-attention mechanism to model the mutual influence between the representations of context and gloss

Methods
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call