Abstract

Just as our interactions with other people are shaped by our concepts about their beliefs, desires, and goals (i.e., theory of mind), our interactions with intelligent technologies such as robots are shaped by our concepts about their internal operations. Multiple studies have demonstrated that people attribute anthropomorphic features to technological agents in certain contexts, but researchers remain divided on how these attributions arise: What default assumptions do people make about the internal operations of intelligent technology, and what events or additional information cause us to alter those default assumptions? This article explores these open questions and some of their implications for law and policy. First, we review psychological research exploring people's attributions of agency, with particular focus on attributions to technological entities. Next, we define and describe one popular account of this research---a account that assumes a reflexive tendency to broadly attribute humanlike properties to technological agents. We then summarize mounting evidence that people are often more cautious in attributing human properties than the promiscuous agency account suggests. We seek to integrate the mounting evidence for a account with the promiscuous agency account through the transition model of agency. Finally, we explore how selective agency, promiscuous agency, and the transition model relate to a sample of robotics law and policy issues. We address, in turn, issues related to Fourth Amendment protection, copyright law, statutory and regulatory interpretation, and negligence litigation, identifying specific implications of the transition model of agency for each issue.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call