Combining touch and vision for the estimation of an object's pose during manipulation

Joao Bimbo, Lakmal D. Seneviratne, Kaspar Althoefer, Hongbin Liu*

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference paperpeer-review

47 Citations (Scopus)

Abstract

Robot grasping and manipulation relies mainly on two types of sensory data: vision and tactile sensing. Localisation and recognition of the object is typically done through vision alone, while tactile sensors are commonly used for grasp control. Vision performs reliably in uncluttered environments, but its performance may deteriorate when the object is occluded, which is often the case during a manipulation task, when the object is in-hand and the robot fingers stand between the camera and the object. This paper presents a method to use the robot's sense of touch to refine the knowledge of a manipulated object's pose from an initial estimate provided by vision. The objective is to find a transformation on the object's location that is coherent with the current proprioceptive and tactile sensory data. The method was tested with different object geometries and proposes applications where this method can be used to improve the overall performance of a robotic system. Experimental results show an improvement of around 70% on the estimate of the object's location when compared to using only vision.
Original languageEnglish
Title of host publicationIEEE International Conference on Intelligent Robots and Systems
PublisherIEEE
Pages4021-4026
Number of pages6
ISBN (Print)9781467363587
DOIs
Publication statusPublished - 2013
Event2013 26th IEEE/RSJ International Conference on Intelligent Robots and Systems: New Horizon, IROS 2013 - Tokyo, Japan
Duration: 3 Nov 20138 Nov 2013

Conference

Conference2013 26th IEEE/RSJ International Conference on Intelligent Robots and Systems: New Horizon, IROS 2013
Country/TerritoryJapan
CityTokyo
Period3/11/20138/11/2013

Fingerprint

Dive into the research topics of 'Combining touch and vision for the estimation of an object's pose during manipulation'. Together they form a unique fingerprint.

Cite this