Abstract

In recent years, with tremendous progresses of deep learning in multiple disciplines, there are several advanced sequential neural-network (NN) based architectures (e.g., recurrent neural network—RNN, Auto-Encoding—AE, transformer, etc.) have been proposed. Recently, there are several well-known GNN-based architectures like as graph convolutional network (GNN) have been proposed to deal with challenges related to the global representation preservation of text. However, most of recent proposed GNN-based text-embedding models still be unable to integrate the global structure with the semantic sequential representations of words/sentences into the unified textual embedding space. Moreover, they are also considered as unable to learn the rich context-varied representations of words. In order to tackle aforementioned challenges, in this paper we proposed a novel integrated text graph representation learning approach, named as: GOWSeqGCN. Our proposed GOWSeqGCN is an integrated semantic graph-of-words sequential textual representation under the graph convolutional network framework. In order to demonstrate for the effectiveness of our proposed GOWSeqGCN model in comparing with recent state-of-the-art text representation learning baselines, we conducted extensive experiments in benchmark textual datasets. The experimental outputs showed the outperformances and necessary of our proposed ideas in this paper.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call