TY - JOUR
T1 - Learning variable impedance control for contact sensitive tasks
AU - Bogdanovic, Miroslav
AU - Khadiv, Majid
AU - Righetti, Ludovic
N1 - Funding Information:
Manuscript received February 24, 2020; accepted July 2, 2020. Date of publication July 23, 2020; date of current version August 4, 2020. This letter was recommended for publication by Associate Editor M. Howard and Editor D. Lee upon evaluation of the reviewers’ comments. This work was supported in part by New York University, in part by the European Union’s Horizon 2020 research and innovation program under Grant agreement 780684, in part by European Research Council under Grant 637935, and in part by Google Faculty Research Award. (Corresponding author: Miroslav Bogdanovic.) Miroslav Bogdanovic and Majid Khadiv are with the Max-Planck Institute for Intelligent Systems, 72076 Tübingen, Germany (e-mail: miroslav. bogdanovic@tuebingen.mpg.de; majid.khadiv@tuebingen.mpg.de).
Publisher Copyright:
© 2016 IEEE.
PY - 2020/10
Y1 - 2020/10
N2 - Reinforcement learning algorithms have shown great success in solving different problems ranging from playing video games to robotics. However, they struggle to solve delicate robotic problems, especially those involving contact interactions. Though in principle a policy directly outputting joint torques should be able to learn to perform these tasks, in practice we see that it has difficulty to robustly solve the problem without any given structure in the action space. In this letter, we investigate how the choice of action space can give robust performance in presence of contact uncertainties. We propose learning a policy giving as output impedance and desired position in joint space and compare the performance of that approach to torque and position control under different contact uncertainties. Furthermore, we propose an additional reward term designed to regularize these variable impedance control policies, giving them interpretability and facilitating their transfer to real systems. We present extensive experiments in simulation of both floating and fixed-base systems in tasks involving contact uncertainties, as well as results for running the learned policies on a real system (accompanying videos can be seen here: https://youtu.be/AQuuQ-h4dBM).
AB - Reinforcement learning algorithms have shown great success in solving different problems ranging from playing video games to robotics. However, they struggle to solve delicate robotic problems, especially those involving contact interactions. Though in principle a policy directly outputting joint torques should be able to learn to perform these tasks, in practice we see that it has difficulty to robustly solve the problem without any given structure in the action space. In this letter, we investigate how the choice of action space can give robust performance in presence of contact uncertainties. We propose learning a policy giving as output impedance and desired position in joint space and compare the performance of that approach to torque and position control under different contact uncertainties. Furthermore, we propose an additional reward term designed to regularize these variable impedance control policies, giving them interpretability and facilitating their transfer to real systems. We present extensive experiments in simulation of both floating and fixed-base systems in tasks involving contact uncertainties, as well as results for running the learned policies on a real system (accompanying videos can be seen here: https://youtu.be/AQuuQ-h4dBM).
KW - Reinforcement learning
KW - compliance and impedance control
KW - motion control
UR - http://www.scopus.com/inward/record.url?scp=85090364765&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85090364765&partnerID=8YFLogxK
U2 - 10.1109/LRA.2020.3011379
DO - 10.1109/LRA.2020.3011379
M3 - Article
AN - SCOPUS:85090364765
SN - 2377-3766
VL - 5
SP - 6129
EP - 6136
JO - IEEE Robotics and Automation Letters
JF - IEEE Robotics and Automation Letters
IS - 4
M1 - 9146673
ER -