z-logo
Premium
A LoCATe‐based visual place recognition system for mobile robotics and GPGPUs
Author(s) -
Bampis Loukas,
Chatzichristofis Savvas,
Iakovidou Chryssanthi,
Amanatiadis Angelos,
Boutalis Yiannis,
Gasteratos Antonios
Publication year - 2017
Publication title -
concurrency and computation: practice and experience
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.309
H-Index - 67
eISSN - 1532-0634
pISSN - 1532-0626
DOI - 10.1002/cpe.4146
Subject(s) - computer science , artificial intelligence , pipeline (software) , computer vision , mobile device , feature (linguistics) , scalability , search engine indexing , general purpose computing on graphics processing units , graphics , pattern recognition (psychology) , computer graphics (images) , philosophy , linguistics , database , programming language , operating system
Summary In this paper, a novel visual Place Recognition approach is evaluated based on a visual vocabulary of the Color and Edge Directivity Descriptor (CEDD) to address the loop closure detection task. Even though CEDD was initially designed so as to globally describe the color and texture information of an input image addressing Image Indexing and Retrieval tasks, its scalability on characterizing single feature points has already been proven. Thus, instead of using CEDD as a global descriptor, we adopt a bottom‐up approach and use its localized version, Local Color And Texture dEscriptor, as an input to a state‐of‐the‐art visual Place Recognition technique based on Visual Word Vectors. Also, we use a parallel execution pipeline based on a previous work of ours using the well established General Purpose Graphics Processing Unit (GPGPU) computing. Our experiments show that the usage of CEDD as a local descriptor produces high accuracy visual Place Recognition results, while the parallelization used allows for a real‐time implementation even in the case of a low‐cost mobile device.

This content is not available in your region!

Continue researching here.

Having issues? You can contact us here