Welcome to the new version of CaltechAUTHORS. Login is currently restricted to library staff. If you notice any issues, please email coda@library.caltech.edu
Published March 16, 2023 | Submitted
Report Open

MimicPlay: Long-Horizon Imitation Learning by Watching Human Play

Abstract

Imitation Learning from human demonstrations is a promising paradigm to teach robots manipulation skills in the real world, but learning complex long-horizon tasks often requires an unattainable amount of demonstrations. To reduce the high data requirement, we resort to human play data - video sequences of people freely interacting with the environment using their hands. We hypothesize that even with different morphologies, human play data contain rich and salient information about physical interactions that can readily facilitate robot policy learning. Motivated by this, we introduce a hierarchical learning framework named MimicPlay that learns latent plans from human play data to guide low-level visuomotor control trained on a small number of teleoperated demonstrations. With systematic evaluations of 14 long-horizon manipulation tasks in the real world, we show that MimicPlay dramatically outperforms state-of-the-art imitation learning methods in task success rate, generalization ability, and robustness to disturbances. More details and video results could be found at https://mimic-play.github.io/.

Additional Information

We are extremely grateful to Yifeng Zhu, Ajay Mandlekar for their efforts in developing the robot control library - Deoxys1 and RoboTurk[30]. Stanford provides the necessary computing resource and infrastructure for this project. L. FF is partially supported by the Stanford HAI Hoffman-Yee Research Grant. This work is done during Chen Wang's internship at NVIDIA.

Attached Files

Submitted - 2302.12422.pdf

Files

2302.12422.pdf
Files (11.2 MB)
Name Size Download all
md5:54c51c331de587939fdc21900e6e0522
11.2 MB Preview Download

Additional details

Created:
August 20, 2023
Modified:
October 25, 2023