Catching objects in flight (i.e., thrown objects) is a common daily skill for humans, yet it presents a significant challenge for robots. This task requires a robot with agile and accurate motion, a large spatial workspace, and the ability to interact with diverse objects. In this paper, we build a mobile manipulator composed of a mobile base, a 6-DoF arm, and a 12-DoF dexterous hand to tackle such a challenging task. We propose a two-stage reinforcement learning framework to efficiently train a whole-body-control catching policy for this high-DoF system in simulation. The throwing configurations are randomized during the training to improve policy adaptivity for different trajectories of objects in flight. The results show that our trained policy catches randomly thrown trajectories of diverse objects at a high success rate of about 80% in simulation, with a significant improvement over the baselines. The policy trained in simulation can be directly deployed in the real world with onboard sensing and computation, which achieves catching flying objects thrown by humans.
Catch It! is a method using reinforcement learning (RL) to learn a whole-body control policy for catching objects in simulation, with Sim2Real transfer to real robots.
The beginning of training
We first train the base and arm control policies in a tracking task parallelly. During the initial phase of training, it is evident that the robot is unable to follow the object's trajectory.
After training for a while
After a period of training, the robot has learned to track the trajectory of the object and make contact with it using the palm.
The beginning of training
We continue to train the robot's catching policy. At the early stage of training, the robot tracks the object easily but fails to grasp it.
After training for catching
Through the two-stage training approach, the robot ultimately achieves the ability to track and catch objects.
To validate the policy's generalization in simulation, we test with 5 unseen object shapes. And our trained policy demonstrates robust generalization to unseen objects.
When deploying the trained policy in the real world, our system may struggle with elastic objects, which tend to rebound off contact surfaces, particularly when both of the object and robot move at high speeds.
@article{zhang2024catchitlearningcatch,
title={Catch It! Learning to Catch in Flight with Mobile Dexterous Hands},
author={Zhang, Yuanhang and Liang, Tianhai and Chen, Zhenyang and Ze, Yanjie and Xu, Huazhe},
journal={arXiv preprint arXiv:2409.10319},
year={2024}
}