Abstract

Abstract Trust-based interactions with robots are increasingly common in the marketplace, workplace, on the road, and in the home. However, a valid concern is that people may not trust robots as they do humans. While trust in fellow humans has been studied extensively, little is known about how people extend trust to robots. Here we compare trust-based investments and self-reported emotions from across three nearly identical economic games: human-human trust games, human-robot trust games, and human-robot trust games where the robot decision impacts another human. Robots in our experiment mimic humans: they are programmed to make reciprocity decisions based on previously observed behaviors by humans in analogous situations. We find that people invest similarly in humans and robots. By contrast the social emotions (i.e., gratitude, anger, pride, guilt) elicited by the interactions (but not the non-social emotions) differed across human and robot trust games. Emotional reactions depended on the trust game interaction, and how another person was affected.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call