Tziafas, Georgios;
Xu, Yucheng;
Goel, Arushi;
Kasaei, Mohammadreza;
Li, Zhibin;
Kasaei, Hamidreza;
(2023)
Language-guided Robot Grasping: CLIP-based Referring Grasp Synthesis in Clutter.
In: Lawrence, Neil, (ed.)
Proceedings of The 7th Conference on Robot Learning.
(pp. pp. 1450-1466).
PMLR: Atlanta, USA.
Preview |
Text
tziafas23a.pdf - Published Version Download (19MB) | Preview |
Abstract
Robots operating in human-centric environments require the integration of visual grounding and grasping capabilities to effectively manipulate objects based on user instructions. This work focuses on the task of referring grasp synthesis, which predicts a grasp pose for an object referred through natural language in cluttered scenes. Existing approaches often employ multi-stage pipelines that first segment the referred object and then propose a suitable grasp, and are evaluated in private datasets or simulators that do not capture the complexity of natural indoor scenes. To address these limitations, we develop a challenging benchmark based on cluttered indoor scenes from OCID dataset, for which we generate referring expressions and connect them with 4-DoF grasp poses. Further, we propose a novel end-to-end model (CROG) that leverages the visual grounding capabilities of CLIP to learn grasp synthesis directly from image-text pairs. Our results show that vanilla integration of CLIP with pretrained models transfers poorly in our challenging benchmark, while CROG achieves significant improvements both in terms of grounding and grasping. Extensive robot experiments in both simulation and hardware demonstrate the effectiveness of our approach in challenging interactive object grasping scenarios that include clutter.
Type: | Proceedings paper |
---|---|
Title: | Language-guided Robot Grasping: CLIP-based Referring Grasp Synthesis in Clutter |
Event: | 7th Conference on Robot Learning (CoRL 2023) |
Open access status: | An open access version is available from UCL Discovery |
Publisher version: | https://proceedings.mlr.press/v229/ |
Language: | English |
Additional information: | © The authors and PMLR 2023. This is an Open Access article distributed under the terms of the Creative Commons Attribution Licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. https://creativecommons.org/licenses/by/4.0/ |
Keywords: | Language-Guided Robot Grasping, Referring Grasp Synthesis, Visual Grounding |
UCL classification: | UCL UCL > Provost and Vice Provost Offices > UCL BEAMS UCL > Provost and Vice Provost Offices > UCL BEAMS > Faculty of Engineering Science UCL > Provost and Vice Provost Offices > UCL BEAMS > Faculty of Engineering Science > Dept of Computer Science |
URI: | https://discovery.ucl.ac.uk/id/eprint/10188331 |
Archive Staff Only
View Item |