OmniDexGrasp: Generalizable Dexterous Grasping via Foundation Model and Force Feedback

Yi-Lin Wei, Zhexi Luo, Yuhao Lin, Mu Lin, Zhizhao Liang, Shuoyu Chen, Wei-Shi Zheng*
Sun Yat-sen University
Equal contribution, *Corresponding author

Teaser

Overview

OmniDexGrasp teaser image

OmniDexGrasp can achieve generalizable dexterous grasping with omni capabilities in user prompting, dexterous embodiment, scenes, and grasping tasks, by leveraging the foundation model and the propose transfer and grasp strategy.

Abstract

Framework

Framework of OmniDexGrasp

(a) Using a foundation generative model, a human grasp image is generated based on the given grasp instruction and the initial scene image. (b) Relying solely on foundation visual models, the human-image-to-robot-action transfer module reconstructs the 3D hand–object interaction from the generated grasp image, retargets the human grasp to the robot’s dexterous hand, and aligns the grasp with the real-world object 6D pose to obtain an executable dexterous grasp action. (c) A force-sensing adaptive grasping strategy executes the grasp by dynamically adjusting finger motions according to force feedback, ensuring stable and reliable grasp execution.

Interactive Visualization

Generated Human Grasp

Real World Grasping

Generated grasp for Watering Can

BibTeX

@article{wei2025omnidexgrasp,
  author    = {Yi-Lin Wei and Zhexi Luo and Yuhao Lin and Mu Lin and Zhizhao Liang and Shuoyu Chen and Wei-Shi Zheng},
  title     = {OmniDexGrasp: Generalizable Dexterous Grasping via Foundation Model and Force Feedback},
  journal   = {arXiv},
  year      = {2025},
}

The source code for this website is adapted from the template provided by nerfies.github.io.