TY - GEN
T1 - Diffusion-based learning of contact plans for agile locomotion
AU - Dhédin, Victor
AU - Ravi, Adithya Kumar Chinnakkonda
AU - Jordana, Armand
AU - Zhu, Huaijiang
AU - Meduri, Avadesh
AU - Righetti, Ludovic
AU - Schölkopf, Bernhard
AU - Khadiv, Majid
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Legged robots have become capable of performing highly dynamic maneuvers in the past few years. However, agile locomotion in highly constrained environments such as stepping stones is still a challenge. In this paper, we propose a combination of model-based control, search, and learning to design efficient control policies for agile locomotion on stepping stones. In our framework, we use nonlinear model predictive control (NMPC) to generate whole-body motions for a given contact plan. To efficiently search for an optimal contact plan, we propose to use Monte Carlo tree search (MCTS). While the combination of MCTS and NMPC can quickly find a feasible plan for a given environment (a few seconds), it is not yet suitable to be used as a reactive policy. Hence, we generate a dataset for optimal goal-conditioned policy for a given scene and learn it through supervised learning. In particular, we leverage the power of diffusion models in handling multi-modality in the dataset. We test our proposed framework on a scenario where our quadruped robot Solo12 successfully jumps to different goals in a highly constrained environment (video).
AB - Legged robots have become capable of performing highly dynamic maneuvers in the past few years. However, agile locomotion in highly constrained environments such as stepping stones is still a challenge. In this paper, we propose a combination of model-based control, search, and learning to design efficient control policies for agile locomotion on stepping stones. In our framework, we use nonlinear model predictive control (NMPC) to generate whole-body motions for a given contact plan. To efficiently search for an optimal contact plan, we propose to use Monte Carlo tree search (MCTS). While the combination of MCTS and NMPC can quickly find a feasible plan for a given environment (a few seconds), it is not yet suitable to be used as a reactive policy. Hence, we generate a dataset for optimal goal-conditioned policy for a given scene and learn it through supervised learning. In particular, we leverage the power of diffusion models in handling multi-modality in the dataset. We test our proposed framework on a scenario where our quadruped robot Solo12 successfully jumps to different goals in a highly constrained environment (video).
UR - http://www.scopus.com/inward/record.url?scp=85212762177&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85212762177&partnerID=8YFLogxK
U2 - 10.1109/Humanoids58906.2024.10769875
DO - 10.1109/Humanoids58906.2024.10769875
M3 - Conference contribution
AN - SCOPUS:85212762177
T3 - IEEE-RAS International Conference on Humanoid Robots
SP - 637
EP - 644
BT - 2024 IEEE-RAS 23rd International Conference on Humanoid Robots, Humanoids 2024
PB - IEEE Computer Society
T2 - 23rd IEEE-RAS International Conference on Humanoid Robots, Humanoids 2024
Y2 - 22 November 2024 through 24 November 2024
ER -