Michelle J. Wang
MIT EECS | Boeing Undergraduate Research and Innovation Scholar
SoftMimic: Learning Compliant Whole-body Control from Examples
2025–2026
Electrical Engineering and Computer Science
- Robotics
Pulkit Agrawal
We introduce SoftMimic, a framework for learning compliant whole-body control policies for humanoid robots from example motions. Imitating human motions with reinforcement learning allows humanoids to quickly learn new skills, but existing methods incentivize stiff control that aggressively corrects deviations from a reference motion, leading to brittle and unsafe behavior when the robot encounters unexpected contacts. In contrast, SoftMimic enables robots to respond compliantly to external forces while maintaining balance and posture. Our approach leverages an inverse kinematics solver to generate an augmented dataset of feasible compliant motions, which we use to train a reinforcement learning policy. By rewarding the policy for matching compliant responses rather than rigidly tracking the reference motion, SoftMimic learns to absorb disturbances and generalize to varied tasks from a single motion clip. We validate our method through simulations and real-world experiments, demonstrating safe and effective interaction with the environment.
I am participating in SuperUROP to gain hands-on experience in robotics and AI research. I am excited to learn more about reinforcement learning, modeling the physical world, and humanoid robots over the course of a longer-term project. I hope that this project will contribute towards safer, more functional robots.
