Abstract

Intent recognition is a key component of any task-oriented conversational system. The intent recognizer can be used first to classify the user’s utterance into one of several predefined classes (intents) that help to understand the user’s current goal. Then, the most adequate response can be provided accordingly. Intent recognizers also often appear as a form of joint models for performing the natural language understanding and dialog management tasks together as a single process, thus simplifying the set of problems that a conversational system must solve. This happens to be especially true for frequently asked question (FAQ) conversational systems. In this work, we first present an exploratory analysis in which different deep learning (DL) models for intent detection and classification were evaluated. In particular, we experimentally compare and analyze conventional recurrent neural networks (RNN) and state-of-the-art transformer models. Our experiments confirmed that best performance is achieved by using transformers. Specifically, best performance was achieved by fine-tuning the so-called BETO model (a Spanish pretrained bidirectional encoder representations from transformers (BERT) model from the Universidad de Chile) in our intent detection task. Then, as the main contribution of the paper, we analyze the effect of inserting unseen domain words to extend the vocabulary of the model as part of the fine-tuning or domain-adaptation process. Particularly, a very simple word frequency cut-off strategy is experimentally shown to be a suitable method for driving the vocabulary learning decisions over unseen words. The results of our analysis show that the proposed method helps to effectively extend the original vocabulary of the pretrained models. We validated our approach with a selection of the corpus acquired with the Hispabot-Covid19 system obtaining satisfactory results.

Highlights

  • Spoken language understanding (SLU) in conversational systems is traditionally divided into two main subtasks: intent detection and semantic slot filling, both extended with domain recognition for multidomain dialogue systems [1,2]

  • In this paper we present an exploratory analysis in which different deep learning (DL) models for intent detection and classification were evaluated

  • Performance was demonstrated to be significantly better for the model fine-tuned from the cased version of BETO, which confirms the superiority of this bidirectional encoder representations from transformers (BERT)-based approach over the evaluated recurrent neural networks (RNN) models

Read more

Summary

Introduction

Spoken language understanding (SLU) in conversational systems is traditionally divided into two main subtasks: intent detection and semantic slot filling, both extended with domain recognition for multidomain dialogue systems [1,2]. Intent classification can be useful for conversational systems to empower customer services with AI-driven FAQ software [1] Its usefulness in this domain is twofold: first, the intent recognizer may help detecting the main information pieces that are present in the user’s utterances, becoming a solution for completing the natural language understanding task; the dialogue management task could be accomplished by assigning the user utterance to one of the intents defined, the one corresponding to the adequate response from the system. In addition to the comparison between models, we analyze the effect of inserting unseen domain words to extend the vocabulary of the model as part of the fine-tuning or domain-adaptation process In this regard, transformers can be quite inefficient in learning new domain words, those that are not backed up with sufficient domain-specific training data. All the suggested approaches were validated with a selection of the corpus acquired with the Hispabot-COVID-19 conversational system, which was developed by the Spanish government to provide responses to FAQ related to the pandemics originated by the COVID-19

Literature Review
The Hispabot-COVID-19 Dataset
Model Description
Embedding Layer
Bi-LSTM Layer
Attention Layer
Transformer Based Model
Basic Preprocessing of Training Data
General Experimental Setup
RNN Specific Setup
BERT Specific Setup
RNN Model Evaluation
RNN Word Embeddings
RNN Results
BERT Model Evaluation
BERT Results
Analyzing the Effect of the Amount of New Words to Be Included
Analyzing the Effect of Different Data Pre-Processing Methods
Conclusions
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.