Abstract

Rapid human arm movements often have velocity profiles consisting of several bell-shaped acceleration–deceleration phases, sometimes overlapping in time and sometimes appearing separately. We show how such sub-movement sequences can emerge naturally as an optimal control policy is approximated by a reinforcement learning system in the face of uncertainty and feedback delay. The system learns to generate sequences of pulse-step commands, producing fast initial sub-movements followed by several slow corrective sub-movements that often begin before the initial sub-movement has completed. These results suggest how the nervous system might efficiently control a stochastic motor plant under uncertainty and feedback delay.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.