Artificial haptics has the potential to revolutionize the way we integrate physical and virtual technologies in our daily lives, with implications for teleoperation, motor skill acquisition, rehabilitation, gaming, interpersonal communication, and beyond. Here, we delve into the intricate interplay between the somatosensory system and engineered haptic inputs for perception and action. We critically examine the sensory feedback's fidelity and the cognitive demands of interfacing with these systems. We examine how artificial touch interfaces could be redesigned to better align with human sensory, motor, and cognitive systems, emphasizing the dynamic and context-dependent nature of sensory integration. We consider the various learning processes involved in adapting to artificial haptics, highlighting the need for interfaces that support both explicit and implicit learning mechanisms. We emphasize the need for technologies that are not only physiologically biomimetic but also behaviorally and cognitively congruent with the user, affording a range of alternative solutions to users' needs.
Read full abstract