In recent years, with tremendous progresses of deep learning in multiple disciplines, there are several advanced sequential neural-network (NN) based architectures (e.g., recurrent neural network—RNN, Auto-Encoding—AE, transformer, etc.) have been proposed. Recently, there are several well-known GNN-based architectures like as graph convolutional network (GNN) have been proposed to deal with challenges related to the global representation preservation of text. However, most of recent proposed GNN-based text-embedding models still be unable to integrate the global structure with the semantic sequential representations of words/sentences into the unified textual embedding space. Moreover, they are also considered as unable to learn the rich context-varied representations of words. In order to tackle aforementioned challenges, in this paper we proposed a novel integrated text graph representation learning approach, named as: GOWSeqGCN. Our proposed GOWSeqGCN is an integrated semantic graph-of-words sequential textual representation under the graph convolutional network framework. In order to demonstrate for the effectiveness of our proposed GOWSeqGCN model in comparing with recent state-of-the-art text representation learning baselines, we conducted extensive experiments in benchmark textual datasets. The experimental outputs showed the outperformances and necessary of our proposed ideas in this paper.