TY - GEN
T1 - Learning to push by grasping
T2 - 2017 IEEE International Conference on Robotics and Automation, ICRA 2017
AU - Pinto, Lerrel
AU - Gupta, Abhinav
N1 - Funding Information:
This work was supported by ONR MURI N000141612007, NSF IIS-1320083 and Google Focused Award. AG was supported in part by Sloan Research Fellowship.
Publisher Copyright:
© 2017 IEEE.
PY - 2017/7/21
Y1 - 2017/7/21
N2 - Recently, end-to-end learning frameworks are gaining prevalence in the field of robot control. These frameworks input states/images and directly predict the torques or the action parameters. However, these approaches are often critiqued due to their huge data requirements for learning a task. The argument of the difficulty in scalability to multiple tasks is well founded, since training these tasks often require hundreds or thousands of examples. But do end-to-end approaches need to learn a unique model for every task? Intuitively, it seems that sharing across tasks should help since all tasks require some common understanding of the environment. In this paper, we attempt to take the next step in data-driven end-to-end learning frameworks: move from the realm of task-specific models to joint learning of multiple robot tasks. In an astonishing result we show that models with multi-task learning tend to perform better than task-specific models trained with same amounts of data. For example, a deep-network learned with 2.5K grasp and 2.5K push examples performs better on grasping than a network trained on 5K grasp examples.
AB - Recently, end-to-end learning frameworks are gaining prevalence in the field of robot control. These frameworks input states/images and directly predict the torques or the action parameters. However, these approaches are often critiqued due to their huge data requirements for learning a task. The argument of the difficulty in scalability to multiple tasks is well founded, since training these tasks often require hundreds or thousands of examples. But do end-to-end approaches need to learn a unique model for every task? Intuitively, it seems that sharing across tasks should help since all tasks require some common understanding of the environment. In this paper, we attempt to take the next step in data-driven end-to-end learning frameworks: move from the realm of task-specific models to joint learning of multiple robot tasks. In an astonishing result we show that models with multi-task learning tend to perform better than task-specific models trained with same amounts of data. For example, a deep-network learned with 2.5K grasp and 2.5K push examples performs better on grasping than a network trained on 5K grasp examples.
UR - http://www.scopus.com/inward/record.url?scp=85027997155&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85027997155&partnerID=8YFLogxK
U2 - 10.1109/ICRA.2017.7989249
DO - 10.1109/ICRA.2017.7989249
M3 - Conference contribution
AN - SCOPUS:85027997155
T3 - Proceedings - IEEE International Conference on Robotics and Automation
SP - 2161
EP - 2168
BT - ICRA 2017 - IEEE International Conference on Robotics and Automation
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 29 May 2017 through 3 June 2017
ER -