Physics-Driven Data Generation for Contact-Rich Manipulation via Trajectory Optimization

1MIT, 2Robotics and AI Institute


An efficient pipeline that augments human demonstrations into large-scale contact-rich datasets for cross-embodiments robot learning.

Abstract

We present a low-cost data generation pipeline that integrates physics-based simulation, human demonstrations, and model-based planning to efficiently generate large-scale, high-quality datasets for contact-rich robotic manipulation tasks. Starting with a small number of embodiment-flexible human demonstrations collected in a virtual reality simulation environment, the pipeline refines these demonstrations using optimization-based kinematic retargeting and trajectory optimization to adapt them across various robot embodiments and physical parameters. This process yields a diverse, physically consistent dataset that enables cross-embodiment data transfer, and offers the potential to reuse legacy datasets collected under different hardware configurations or physical parameters. We validate the pipeline's effectiveness by training diffusion policies from the generated datasets for challenging contact-rich manipulation tasks across multiple robot embodiments, including a floating Allegro hand and bimanual robot arms. The trained policies are deployed zero-shot on hardware for bimanual iiwa arms, achieving high success rates with minimal human input.

VR-Based Human Demonstration

Trajectory optimization is effective for generating dynamically feasible trajectories

Trajectory optimization is effective for generating diverse trajectories

Video

BibTeX

@article{yang2025physics,
      title={Physics-Driven Data Generation for Contact-Rich Manipulation via Trajectory Optimization},
      author={Yang, Lujie and Suh, H.J. Terry and Zhao, Tong and Græsdal, Bernhard Paus and Kelestemur, Tarik and Wang, Jiuguang and Pang, Tao and Tedrake, Russ},
      journal={arXiv preprint arXiv:2502.20382},
      year={2025}
    }