TY - JOUR
T1 - Human-Like Rewards to Train a Reinforcement Learning Controller for Planar Arm Movement
AU - Jagodnik, Kathleen M.
AU - Thomas, Philip S.
AU - Van Den Bogert, Antonie J.
AU - Branicky, Michael S.
AU - Kirsch, Robert F.
N1 - Funding Information:
This work was supported by the National Institutes of Health fellowship #TRN030167, Veterans Administration Rehabilitation Research and Development predoctoral fellowship "Reinforcement Learning Control for an Upper-Extremity Neuroprosthesis," NIH Training Grant T32 EB004314, and Ardiem Medical Arm Control Device Grant #W81XWH0720044.
Publisher Copyright:
© 2016 IEEE.
PY - 2016/10
Y1 - 2016/10
N2 - High-level spinal cord injury (SCI) in humans causes paralysis below the neck. Functional electrical stimulation (FES) technology applies electrical current to nerves and muscles to restore movement, and controllers for upper extremity FES neuroprostheses calculate stimulation patterns to produce desired arm movement. However, currently available FES controllers have yet to restore natural movements. Reinforcement learning (RL) is a reward-driven control technique; it can employ user-generated rewards, and human preferences can be used in training. To test this concept with FES, we conducted simulation experiments using computer-generated 'pseudo-human' rewards. Rewards with varying properties were used with an actor-critic RL controller for a planar two-degree-of-freedom biomechanical human arm model performing reaching movements. Results demonstrate that sparse, delayed pseudo-human rewards permit stable and effective RL controller learning. The frequency of reward is proportional to learning success, and human-scale sparse rewards permit greater learning than exclusively automated rewards. Diversity of training task sets did not affect learning. Long-term stability of trained controllers was observed. Using human-generated rewards to train RL controllers for upper-extremity FES systems may be useful. Our findings represent progress toward achieving human-machine teaming in control of upper-extremity FES systems for more natural arm movements based on human user preferences and RL algorithm learning capabilities.
AB - High-level spinal cord injury (SCI) in humans causes paralysis below the neck. Functional electrical stimulation (FES) technology applies electrical current to nerves and muscles to restore movement, and controllers for upper extremity FES neuroprostheses calculate stimulation patterns to produce desired arm movement. However, currently available FES controllers have yet to restore natural movements. Reinforcement learning (RL) is a reward-driven control technique; it can employ user-generated rewards, and human preferences can be used in training. To test this concept with FES, we conducted simulation experiments using computer-generated 'pseudo-human' rewards. Rewards with varying properties were used with an actor-critic RL controller for a planar two-degree-of-freedom biomechanical human arm model performing reaching movements. Results demonstrate that sparse, delayed pseudo-human rewards permit stable and effective RL controller learning. The frequency of reward is proportional to learning success, and human-scale sparse rewards permit greater learning than exclusively automated rewards. Diversity of training task sets did not affect learning. Long-term stability of trained controllers was observed. Using human-generated rewards to train RL controllers for upper-extremity FES systems may be useful. Our findings represent progress toward achieving human-machine teaming in control of upper-extremity FES systems for more natural arm movements based on human user preferences and RL algorithm learning capabilities.
KW - Control
KW - functional electrical stimulation (FES)
KW - human-machine teaming
KW - modeling
KW - rehabilitation
KW - reinforcement learning (RL)
KW - simulation
KW - upper extremity
UR - http://www.scopus.com/inward/record.url?scp=84971406219&partnerID=8YFLogxK
U2 - 10.1109/THMS.2016.2558630
DO - 10.1109/THMS.2016.2558630
M3 - Article
AN - SCOPUS:84971406219
VL - 46
SP - 723
EP - 733
JO - IEEE Transactions on Human-Machine Systems
JF - IEEE Transactions on Human-Machine Systems
SN - 2168-2291
IS - 5
M1 - 7478097
ER -