One-Shot Learning for Task-Oriented Grasping

Valerija Holomjova*, Andrew J. Starkey, Bruno Yun, Pascal Meisner

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Task-oriented grasping models aim to predict a suitable grasp pose on an object to fulfill a task. These systems have limited generalization capabilities to new tasks, but have shown the ability to generalize to novel objects by recognizing affordances. This object generalization comes at the cost of being unable to recognize the object category being grasped, which could lead to unpredictable or risky behaviors. To overcome these generalization limitations, we contribute a novel system for task-oriented grasping called the One-Shot Task-Oriented Grasping (OS-TOG) framework. OS-TOG comprises four interchangeable neural networks that interact through dependable reasoning components, resulting in a single system that predicts multiple grasp candidates for a specific object and task from multi-object scenes. Embedded one-shot learning models leverage references within a database for OS-TOG to generalize to novel objects and tasks more efficiently than existing alternatives. Additionally, the paper presents suitable candidates for the framework's neural components, covering essential adjustments for their integration and evaluative comparisons to state-of-the-art. In physical experiments with novel objects, OS-TOG recognizes 69.4% of detected objects correctly and predicts suitable task-oriented grasps with 82.3% accuracy, having a physical grasp success rate of 82.3%.

Original languageEnglish
Pages (from-to)8232-8238
Number of pages7
JournalIEEE Robotics and Automation Letters
Volume8
Issue number12
Early online dateNov 2023
DOIs
Publication statusPublished - 1 Dec 2023

Keywords

  • computer vision for automation
  • Deep learning in grasping and manipulation
  • grasping
  • recognition

Fingerprint

Dive into the research topics of 'One-Shot Learning for Task-Oriented Grasping'. Together they form a unique fingerprint.

Cite this