Abstract

The high request for autonomous human–robot interaction (HRI), combined with the potential of machine learning (ML) techniques, allow us to deploy ML mechanisms in robot control. However, the use of ML can make robots’ behavior unclear to the observer during the learning phase. Recently, transparency in HRI has been investigated to make such interactions more comprehensible. In this work, we propose a model to improve the transparency during reinforcement learning (RL) tasks for HRI scenarios: the model supports transparency by having the robot show nonverbal emotional-behavioral cues. Our model considered human feedback as the reward of the RL algorithm and it presents emotional-behavioral responses based on the progress of the robot learning. The model is managed only by the temporal-difference error. We tested the architecture in a teaching scenario with the iCub humanoid robot. The results highlight that when the robot expresses its emotional-behavioral response, the human teacher is able to understand its learning process better. Furthermore, people prefer to interact with an expressive robot as compared to a mechanical one. Movement-based signals proved to be more effective in revealing the internal state of the robot than facial expressions. In particular, gaze movements were effective in showing the robot's next intentions. In contrast, communicating uncertainty through robot movements sometimes led to action misinterpretation, highlighting the importance of balancing transparency and the legibility of the robot goal. We also found a reliable temporal window in which to register teachers’ feedback that can be used by the robot as a reward.

Highlights

  • T HE more robots become autonomous and flexible, the more their behaviors need to be transparent

  • We found that the learning performances were fairly similar regardless of the showed behavior (two-tailed t-test t(17) = 1.1346, p = 0.2732), with averages and the standard deviations of the number of epochs needed to reach the highest accuracy μm = 8.2 with σm = 3.5, and μt = 9.2 with σt = 3.8

  • We studied robot transparency during reinforcement learning (RL) tasks by providing a temporal difference (TD)-based emotional-behavioral model to a robot

Read more

Summary

Introduction

T HE more robots become autonomous and flexible, the more their behaviors need to be transparent. In the field, one of the interpretations of the term transparency is related to the observability or predictability of a system’s behavior, and the possibility to infer its intentions, e.g., understanding what the robot is doing, why it is doing an action rather than another one, and what it is going to do next. The possibility of interpreting the behaviors of intelligent others, both in case of success and failures, is a fundamental characteristic of successful interactions: it affects human trust in automation [2]. This necessity is evident when dealing with machine learning (ML) algorithms that control robots’ behaviors. People are good at identifying the target of their partner’s gaze and using that information to predict their partner’s actions [24]

Objectives
Methods
Results
Discussion
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call