Abstract
SummaryDeep learning has revolutionized the field of conversation modeling. A lot of the research has been toward making the conversational agent more human‐like. As a result, overall the model size increases. Bigger models require more data and are costly to build and maintain. Often, for some tasks, high‐quality responses are not necessary. In this paper, a model that consumes fewer resources and a way to augment conversation data without increasing the size of the vocabulary is proposed. The proposed model uses a modified version of the GRU instead of the LSTM to encode and decode sequences of text.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
More From: Concurrency and Computation: Practice and Experience
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.