Abstract

Abstract In order to be acceptable and able to “camouflage” into their physio-social context in the long run, robots need to be not just functional, but autonomously psycho-affective as well. This motivates a long term necessity of introducing behavioral autonomy in robots, so they can autonomously communicate with humans without the need of “wizard” intervention. This paper proposes a technique to learn robot speech models from human-robot dialog exchanges. It views the entire exchange in the Automated Planning (AP) paradigm, representing the dialog sequences (speech acts) in the form of action sequences that modify the state of the world upon execution, gradually propelling the state to a desired goal. We then exploit intra-action and inter-action dependencies, encoding them in the form of constraints. We attempt to satisfy these constraints using aweighted maximum satisfiability model known as MAX-SAT, and convert the solution into a speech model. This model could have many uses, such as planning of fresh dialogs. In this study, the learnt model is used to predict speech acts in the dialog sequences using the sequence labeling (predicting future acts based on previously seen ones) capabilities of the LSTM (Long Short Term Memory) class of recurrent neural networks. Encouraging empirical results demonstrate the utility of this learnt model and its long term potential to facilitate autonomous behavioral planning of robots, an aspect to be explored in future works.

Highlights

  • In order to be acceptable and able to “camouflage” into their physio-social context in the long run, robots need to be not just functional, but autonomously psycho-affective as well

  • This paper proposes a technique to learn robot speech models from human-robot dialog exchanges

  • Tomated Planning (AP) paradigm, representing the dialog sequences in the form of action sequences that modify the state of the world upon execution, gradually propelling the state to a desired goal

Read more

Summary

Introduction

Abstract: In order to be acceptable and able to “camouflage” into their physio-social context in the long run, robots need to be not just functional, but autonomously psycho-affective as well. This paper proposes a technique to learn robot speech models from human-robot dialog exchanges. Scripting these subtleties is time consuming, impractical and requires a lot of programming In such a complex scenario, learning the underlying speech model of the robot from HRI dialog sequences is a promising alternative. Learning the underlying model comprising action descriptions from action name-signature sequences could save the effort from having to code these action descriptions from scratch, promoting re usability This model can further be fed to an automated planner to generate fresh dialogs, allowing the robot to communicate autonomously in future scenarios. Our contribution in this paper is the following: given a HRI dialog corpus, our approach learns the robot’s behavioral model comprising of the utterances encoded in the form of actions alongwith their signatures, preconditions and effects.

Related work
Definitions and problem formulation
Approach
Annotation and generalization
Intra-operator constraints
Inter-operator constraints
Data encoding for labelling of operator sequences
Evaluation
LSTM based sequence labeling
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.