Abstract

Corpus-driven approaches can automatically explore is-a relations between the word pairs from corpus. This problem is also called hypernym extraction. Formerly, lexico-syntactic patterns have been used to solve hypernym relations. The language-specific syntactic rules have been manually crafted to build the patterns. On the other hand, recent studies have applied distributional approaches to word semantics. They extracted the semantic relations relying on the idea that similar words share similar contexts. Former distributional approaches have applied one-hot bag-of-word (BOW) encoding. The dimensionality problem of BOW has been solved by various neural network approaches, which represent words in very short and dense vectors, or word embeddings. In this study, we used word embeddings representation and employed the optimized projection algorithm to solve the hypernym problem. The supervised architecture learns a mapping function so that the embeddings (or vectors) of word pairs that are in hypernym relations can be projected to each other. In the training phase, the architecture first learns the embeddings of words and the projection function from a list of word pairs. In the test phase, the projection function maps the embeddings of a given word to a point that is the closest to its hypernym. We utilized the deep learning optimization methods to optimize the model and improve the performances by tuning hyperparameters. We discussed our results by carrying out many experiments based on cross-validation. We also addressed problem-specific loss function, monitored hyperparameters, and evaluated the results with respect to different settings. Finally, we successfully showed that our approach outperformed baseline functions and other studies in the Turkish language.

Highlights

  • Hypenymy indicates an is-a semantic relation between two nouns such as “cat-animal” or “Paris-city”

  • Computational approaches can automatically deduce such relations from raw text by applying corpus-driven approaches. This is called either a hypernym classification problem when classifying if two words given are in hypernym relation or hypernym extraction when pulling out the pairs based on some corpus statistics

  • Experimental results and discussion We assessed the performance of SGD, Nesterov’s accelerated gradient (NAG), Nesterov-accelerated adaptive moment estimation (Nadam), Adagrad, RMSProp, and adaptive moment estimation (Adam) optimizers

Read more

Summary

Introduction

Hypenymy indicates an is-a semantic relation between two nouns such as “cat-animal” or “Paris-city”. Computational approaches can automatically deduce such relations from raw text by applying corpus-driven approaches. This is called either a hypernym classification problem when classifying if two words given are in hypernym relation or hypernym extraction when pulling out the pairs based on some corpus statistics. Many studies have used distributional approaches to extract the pairs having such semantic relations. The distributional hypothesis relies on the idea that similar words share similar contexts and neighbors They do not use predefined sources such as dictionary or linguistics rules. A word is represented by a vector that keeps count of how many times it cooccurs with its nearby words

Methods
Results
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call