Abstract

Word embeddings have introduced a compact and efficient way of representing text for further downstream natural language processing (NLP) tasks. Most word embedding algorithms are optimized at the word level. However, many NLP applications require text representations of groups of words, like sentences or paragraphs. In this paper, we propose a supervised algorithm that produces a task-optimized weighted average of word embeddings for a given task. Our proposed text embedding algorithm combines the compactness and expressiveness of the word-embedding representations with the word-level insights of a BoW-type model, where weights correspond to actual words. Numerical experiments across different domains show the competence of our algorithm.

Highlights

  • Word embeddings, or a learned mapping from a vocabulary to a vector space, are essential tools for state-of-the-art Natural Language Processing (NLP) techniques

  • Our algorithm provides better or comparable performance against unweighted averaged word embedding (UAEm) and WAEm

  • Our paper provides an alternative way of sentence/documentlevel representation for supervised text classification, based on optimization of the weights of words in the corresponding text to be classified

Read more

Summary

INTRODUCTION

A learned mapping from a vocabulary to a vector space, are essential tools for state-of-the-art Natural Language Processing (NLP) techniques. In this paper we propose a supervised algorithm that produces embeddings at the sentence-level that consist on an weighted average of an available pre-trained word-level embedding. On the other hand, identifies “romance” and “action” as two important words in the vocabulary for the supervised task, and assigns weights with high absolute value to these words This leads to shifting of the representation of the two reviews toward their respective important words in the vector space, increasing the distance between them. Our empirical results show that our proposed representation is in general competitive with traditional deep learning based text classification approaches and outperforms them when the training data is relatively small. Our resulting task specific text embedding are as compact as the original word level embedding while providing word level insights similar to a BOW type model.

RELATED WORK
OPTIMAL WORD EMBEDDINGS
Datasets
Word Embeddings
Text Processing
Results
Text Representation
CONCLUSIONS AND FUTURE WORK
DATA AVAILABILITY STATEMENT
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call