Smart Objects and Technologies for Social Good. Second International Conference, GOODTECHS 2016, Venice, Italy, November 30 – December 1, 2016, Proceedings

Research Article

Object Detection and Spatial Coordinates Extraction Using a Monocular Camera for a Wheelchair Mounted Robotic Arm

Download
335 downloads
  • @INPROCEEDINGS{10.1007/978-3-319-61949-1_24,
        author={Alessandro Palla and Alessandro Frigerio and Gabriele Meoni and Luca Fanucci},
        title={Object Detection and Spatial Coordinates Extraction Using a Monocular Camera for a Wheelchair Mounted Robotic Arm},
        proceedings={Smart Objects and Technologies for Social Good. Second International Conference, GOODTECHS 2016, Venice, Italy, November 30 -- December 1, 2016, Proceedings},
        proceedings_a={GOODTECHS},
        year={2017},
        month={7},
        keywords={Robotic arm Power wheelchair Visual Servoing PBVS Eye-in-hand Computer Vision SIFT Features extraction PTAM ROS Human machine interface Assistive technology Open-source},
        doi={10.1007/978-3-319-61949-1_24}
    }
    
  • Alessandro Palla
    Alessandro Frigerio
    Gabriele Meoni
    Luca Fanucci
    Year: 2017
    Object Detection and Spatial Coordinates Extraction Using a Monocular Camera for a Wheelchair Mounted Robotic Arm
    GOODTECHS
    Springer
    DOI: 10.1007/978-3-319-61949-1_24
Alessandro Palla1,*, Alessandro Frigerio1,*, Gabriele Meoni1,*, Luca Fanucci1,*
  • 1: University of Pisa
*Contact email: alessandro.palla@for.unipi.it, a.frigerio@studenti.unipi.it, gabriele.meoni@ing.unipi.it, luca.fanucci@unipi.it

Abstract

In the last decades, smart power wheelchairs have being used by people with motor skill impairment in order to improve their autonomy, independence and quality of life. The most recent power wheelchairs feature many technological devices, such as laser scanners to provide automatic obstacle detection or robotic arms to perform simple operations like pick and place. However, if a motor skill impaired user was able to control a very complex robotic arm, paradoxically he would not need it. For that reason, in this paper we present an autonomous control system based on Computer Vision algorithms which allows the user to interact with buttons or elevator panels via a robotic arm in a simple and easy way. Scale-Invariant Feature Transform (SIFT) algorithm has been used to detect and track buttons. Objects detected by SIFT are mapped in a tridimensional reference system collected with Parallel and Tracking Mapping (PTAM) algorithm. Real word coordinates are obtained using a Maximum-Likelihood estimator, fusing the PTAM coordinates with distance information provided by a proximity sensor. The visual servoing algorithm has been developed in Robotic Operative System (ROS) Environment, in which the previous algorithms are implemented as different nodes. Performances have been analyzed in a test scenario, obtaining good results on the real position of the selected objects.