Skip to the content.

In-Hand Following of Deformable Linear Objects Using Dexterous Fingers with Tactile Sensing

The paper has been accepted by IROS 2024 (oral presentation, top 12%).

[arXiv] [Code]

Video

Abstract

Most research on deformable linear object (DLO) manipulation assumes rigid grasping. However, beyond rigid grasping and re-grasping, in-hand following is also an essential skill that humans use to dexterously manipulate DLOs, which requires continuously changing the grasp point by in-hand sliding while holding the DLO to prevent it from falling. Achieving such a skill is very challenging for robots without using specially designed but not versatile end-effectors. Previous works have attempted using generic parallel grippers, but their robustness is unsatisfactory owing to the conflict between following and holding, which is hard to balance with a one-degree-of-freedom gripper. In this work, inspired by how humans use fingers to follow DLOs, we explore the usage of a generic dexterous hand with tactile sensing to imitate human skills and achieve robust in-hand DLO following. To enable the hardware system to function in the real world, we develop a framework that includes Cartesian-space arm-hand control, tactile-based in-hand 3-D DLO pose estimation, and task-specific motion design. Experimental results demonstrate the significant superiority of our method over using parallel grippers, as well as its great robustness, generalizability, and efficiency.

Something not Included in the Paper

Due to the page limit, we cannot include everything in the conference paper. Here we share more details and discussions.

More details

Limitations

In this work, we pioneeringly explore the usage of dexterous hands to enhance the DLO following. We acknowledge that the current version contains some limitations and there is a lot of room for improvement.

Potential future work

Contact

If you have any question, feel free to contact the authors: Mingrui Yu, mingruiyu98@gmail.com .

Mingrui Yu’s Homepage is at mingrui-yu.github.io.