Abstract

Incorporating topic information can help response generation models to produce informative responses for chat-bots. Previous work only considers the individual semantic of each topic, ignoring its specific dialog context, which may result in inaccurate topic representation and hurt response coherence. Besides, as an important feature of multi-turn conversation, dynamic topic transitions have not been well-studied. We propose a Context-Controlled Topic-Aware neural response generation model, i.e., CCTA, which makes dialog context interact with the process of topic representing and transiting to achieve balanced improvements on response informativeness and contextual coherence. CCTA focuses on capturing the semantical relations within topics as well as their corresponding contextual information in conversation, to produce context-dependent topic representations at the word-level and turn-level. Besides, CCTA introduces a context-controlled topic transition strategy, utilizing contextual topics to yield relevant transition words. Extensive experimental results on two benchmark multi-turn conversation datasets validate the superiority of our proposal on generating coherent and informative responses against the state-of-the-art baselines. We also find that topic transition modeling can work as an auxiliary learning task to boost the response generation.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call