Abstract

Deep learning has achieved impressive success in natural language processing. However, most previous models are learned on the specific single tasks, suffering from insufficient training set. Multi-task deep learning can solve this dilemma by sharing the part of parameters, improving generalization. The common multi-task deep learning model consists of the shared layer and the task specific layer. In this paper, we attempt to enhance the performance of shared layer and proposed two variants based on convolutional neural networks. The first model is Agent Model-Direct Concatenate, where each task is assigned with a separate convolutional neural network for extracting the common and task specific features simultaneously. The second model is Agent Model-Gating Concatenation, where the task specific layer could automatically decide the information flow of each element of the output of shared layer. The two networks are trained jointly over three pair-wise groups of movie review data sets. Experiments show the effectiveness of our two networks, inspiring a potential direction for the related research of multi-task deep learning.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.