Open Box
Operating smoothly in contact-rich environments is crucial for robots to effectively perform daily tasks. Leveraging contact force information can enhance the smoothness of interactive operations, force control should be addressed throughout the entire process of autonomous policy development, from data collection, training to deployment.
we propose an admittance imitation learning visuomotor policy framework to reduce mean contact force and force fluctuations. Our framework utilizes RGB images, robot joint positions, end-effector poses, and contact force as inputs, employing a diffusion model to generate future end-effector trajectories and contact force.
An admittance controller is employed to track this trajectories, enabling effective force control for various tasks. Furthermore, a low-cost hand-arm teleoperation system with interactive feedback is designed for data collection. A evaluation of our teleoperation system and policy framework was conducted on five contact-rich manipulation tasks, each representing an action primitive. Results show that our framework achieves the highest success rate and demonstrates smoother contact compared to other methods. Particularly for door opening task, the average contact force is reduced by 53.92%, while the standard deviation of contact force fluctuations is diminished by 76.51%.
AdmitDiff Policy Framework. Left: During inference, the previous two steps’ observations are encoded as inputs for noise estimation, while the student model outputs actions for the next 8 time steps, the number K represents the denoising iteration required by the diffuser. The arm’s force-position trajectory is used in the admittance controller to compute the desired pose. Middle: The teacher model is trained for 100 denoising steps, then its parameters are frozen to train the student model with a consistency loss for single-step denoising. Right: Data collection, including contact force information, is performed using the teleoperation system designed in this work.
There are three main components that provide critical data for the teleoperation task.
Open Box
Rotate Box
Drag
AdmitDiff Policy consistently demonstrates the most stable contact force control, with lower variance and more controlled median force, especially in tasks like Insertion and Wiping that require precision. In contrast, Diffusion Policy and Consistency Policy exhibit higher variance, particularly in force-intensive tasks like DoorOpening, indicating less stable control. The AdmitDiff Policy shows an average reduction of approximately 48.8% in mean contact force and 52.0% in standard deviation compared to the other methods in the five tasks.
@article{bo@2024admitdiff,
author = {Bo Zhou, Ruixuan Jiao, Yi Li, Xiaogang Yuan, Fang Fang, Shihua Li,},
title = {Admittance Visuomotor Policy Learning for General-Purpose Contact-Rich Manipulations},
journal = {arXiv preprint arXiv:2409.14440},
year = {2024},
}