Abstract

In recent years collaborative robots have become major market drivers in industry 5.0, which aims to incorporate them alongside humans in a wide array of settings ranging from welding to rehabilitation. Improving human–machine collaboration entails using computational algorithms that will save processing as well as communication cost. In this study we have constructed an agent that can choose when to cooperate using an optimal strategy. The agent was designed to operate in the context of divergent interest tacit coordination games in which communication between the players is not possible and the payoff is not symmetric. The agent’s model was based on a behavioral model that can predict the probability of a player converging on prominent solutions with salient features (e.g., focal points) based on the player’s Social Value Orientation (SVO) and the specific game features. The SVO theory pertains to the preferences of decision makers when allocating joint resources between themselves and another player in the context of behavioral game theory. The agent selected stochastically between one of two possible policies, a greedy or a cooperative policy, based on the probability of a player to converge on a focal point. The distribution of the number of points obtained by the autonomous agent incorporating the SVO in the model was better than the results obtained by the human players who played against each other (i.e., the distribution associated with the agent had a higher mean value). Moreover, the distribution of points gained by the agent was better than any of the separate strategies the agent could choose from, namely, always choosing a greedy or a focal point solution. To the best of our knowledge, this is the first attempt to construct an intelligent agent that maximizes its utility by incorporating the belief system of the player in the context of tacit bargaining. This reward-maximizing strategy selection process based on the SVO can also be potentially applied in other human–machine contexts, including multiagent systems.

Highlights

  • Previous research has shown that it is possible to aid an artificial agent in predicting human decisions in the context of coordination problems by incorporating prominent solutions, known as focal points, into learning algorithms [1,2,3,4]

  • This has enabled us to estimate the performance of the adaptive agent without the bias of the prior knowledge of the Social Value Orientation (SVO) lacking in the case of the other two agent classes

  • To the best of our knowledge, this study presents the first attempt to construct an intelligent agent that maximizes its utility by incorporating the belief system of the player in the context of tacit bargaining

Read more

Summary

Introduction

Previous research has shown that it is possible to aid an artificial agent in predicting human decisions in the context of coordination problems by incorporating prominent solutions, known as focal points, into learning algorithms [1,2,3,4]. The framework suggested by [6] was demonstrated by an example taken from rehabilitation robotics By using this example, the authors demonstrate two main principles of artificial agent behavior: adaptability, that is, the ability of the agent to adapt to human behavior, and reliance on a repository containing human-centric strategies. According to [6], these principles may improve the prediction intent of the agent and can diminish the cost of conflict between the user input and the predicted action by the intelligent agent

Objectives
Methods
Results
Discussion
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call