Abstract
In this paper, we describe how agents can deceive within a probabilistic framework for representing their mental state: in doing so, we challenge the so‐called sincerity assumption in Human–Computer interaction (HCI) and multi‐agent systems (MAS). We distinguish deception from its special case of lie and characterize different forms of deception, by identifying several criteria for distinguishing among them. In particular, we propose a model of information impact on the Receiver's mind. As the message Sender must plan its strategy by considering the Receiver's criteria for believing, we also discuss some of these criteria, like content plausibility, source informativity, and information safety. We apply this model to a simplified version of Turing's Imitation Game and describe how we implemented a Simulator of deceptive strategies that we called Mouth of Truth. We conclude the paper by describing an evaluation study that enabled us to verify the validity of our method and to revise it in part.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.